• tal
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    4 hours ago

    Not yet! One thing that AI generated images right now are not so good at is maintaining a consistent portrayal of a character from image to image, which is something you want for illustrating a story.

    You might be able to do something like that with a 3d modeler to pose characters, generate a wireframe, and then feed that wireframe into ControlNet. Or if you have a huge corpus of existing images of a particular character portrayed in a particular way, you could maybe create new images with them in new situations. But without that, it’s hard to go from a text description to many images portrayed in a consistent way. For one image, it works, and for some things, that’s fine. But you’d have a hard time doing, say, a graphic novel that way.

    I suspect that doing something like that is going to require having models that are actually working with 3D internal representations of the world, rather than 2D, at a bare minimum.