• Ramin Honary@lemmy.ml
    link
    fedilink
    English
    arrow-up
    2
    ·
    10 months ago

    Some of those DALL-E renders looked pretty different from the original.

    You have to always keep in mind that the result you see is heavily biased by the dataset that was used to train that AI, so the result you see may actually look nothing like the real person.

  • keepthepace@slrpnk.net
    link
    fedilink
    arrow-up
    2
    ·
    10 months ago

    I see several people trying to generate images on historical pictures/wikipedia articles and while it is technically interesting, I am thinking, what is the point? These pictures are informative: they give you hints about how these people were depicted/dressed at the time. “Making it look stadnard-good” kinda defeats the purpose…

    • lanolinoil@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      1
      ·
      10 months ago

      I built this for myself because I kept putting in 18th century etchings or ruined frescoes into chat GPT to see what they looked like in real life. I’ve always been interested to see marble statues in flesh too.

      This is not AI generated or the same thing, but partially what I think put the idea in my brain – https://tenochtitlan.thomaskole.nl/

      How cool is that? Feels like you’re there almost. I’ll take that over some timeworn painting, which is not truthful either – Ceci n’est pas une pipe

      • keepthepace@slrpnk.net
        link
        fedilink
        arrow-up
        1
        ·
        10 months ago

        My problem is that reconstructions by architects are produced differently, using more information, than Hollywood-like DALL-E outputs. The first one is made to be informative, about the size and aspect of the buildings, the urban organization, etc. The second one is just designed to look good.

        • lanolinoil@lemmy.worldOP
          link
          fedilink
          English
          arrow-up
          1
          ·
          10 months ago

          Not entirely – This uses controlnet to mimic the lines and shapes in the original image – It’s finicky getting the settings right so it works well in all cases, but this is more than just a straight img2img conversion or DALLE txt2img generation.

          This is a good example:

          original wiki image

          SDXL Controlnet image

          • keepthepace@slrpnk.net
            link
            fedilink
            arrow-up
            1
            ·
            10 months ago

            This is indeed a good example. You literally just give it squares to fill with whatever the model feels makes sense and as a result you have christian looking fresco probably very far from the likely reconstructions archaeologists would produce (the red parts were likely totally red, without characters in them).

            I feel that applied like that it is closer to misinformation than enhancement.

            • lanolinoil@lemmy.worldOP
              link
              fedilink
              English
              arrow-up
              1
              ·
              10 months ago

              Well, not exactly again – The original image is passed to GPT4 vision along with the article title and the image’s description, so it’s attempting to be contextual in its prompts and contextual to the shapes in the original image.