manucardoen 1 day ago

It's not good at making ASCII art. This, for example, is when I ask it for a realistic depiction of the Eiffel tower on fire:

                       .
                      .'.
                      |o|
                     .'o'.
                     |.-.|
                     '   '
                    /     \
                   |       |
                __/_________\__
               |###############|
               |###############|
               |###############|
               |###############|
               |###############|
              /###############/|
             /###############/ |
            /###############/ /|
           /###############/ / |
          /###############/ / /|
         /###############/ / / |
        /###############/ / / /|
       /###############/ / / / |
      /###############/ / / / /|
     '-----------------' / / / /
      |   FIRE & ASH   |/ / / /
       '---------------' / / /
          ~ ~ ~ ~ ~ ~ ~ ~ / /
         ~~ /
          ~~ 
           ~~     ⬆ SMOKE
            ~~

5
cucubeleza 1 day ago

well, I asked to GPT to draw Arnold Schwarzenegger:

        ______
       /      \
      |  O  O  |
      |   __   |
       \______/
         ||||
       /--||--\ 
     /    ||    \
    |     ||     | 
    |    /  \    |
     \__/    \__/ 
       ||    ||
       ||    || 
      / |    | \
     /_/      \_\

DuncanCoffee 1 day ago

Reminds me of the abomination of nature you could build in Monkey Island 4

https://www.youtube-nocookie.com/embed/RiwL9awO5y0

cucubeleza 1 day ago

jeeeeeeesus christ that's horrible, but it's cool that you can do that

KyleBerezin 1 day ago

That's amazing. It really captured the likeness of ol' Arnold.

FrostAutomata 1 day ago

ASCII art is extremely difficult for LLMs due to how characters are converted into tokens without preserving their relative positions.

beklein 1 day ago

Great point, but you need to have a good understanding in how LLMs work to understand this limitation. If you don't have an intuitive understanding think like it's one of these draw on my back games, just each new token is a new human in the loop, known words are like simple shapes you felt before 100 times on your back and are easy for you to reproduce and change, random ASCII strings are harder to _grasp_ and will produce a fuzzy output... all models are wrong, but some are useful.

https://www.youtube.com/watch?v=bA_DQHoxj34

light_hue_1 1 day ago

This isn't the reason. Models are pretty good at understanding relative positions. We put that in them and reward it a lot.

The issue is the same as why we don't use LLMs for image generation. Even though they can nominally do that.

Image generation seems to need some amount of ability to revise the output in place. And it needs a big picture view to make local decisions. It doesn't lend itself to outputting pixel by pixel or character by character.

yesbabyyes 1 day ago

This is something I and a few of my colleagues have noticed, as we asked several models to draw ASCII art of a wasp, which is one of our logos. The results are hilarious, and only seem to get worse as you ask it to do better.

bezbac 1 day ago

I've read that the results improve if you ask them to write a program that creates the desired ASCII art. Haven't tried it myself yet so far.

bn-l 1 day ago

Art is highly subjective

ohgr 1 day ago

I subjectively judge that as shit.