I suspect these are using additional tools to guide the AI beyond a simple prompt. For example the spiraling medieval village was generated with stable diffusion and controlnet.
Not sure what this prompt was, since I didn’t make this one. I linked the site I used above, and they’re pretty simple to do but need a few tries to get a good one.
I think the prompt is not much other than “puppies” and “kittens”. Major, middle and minor features of the image can be controlled individually in some AIs (they can be differentiated using a Fourier transform or Gauss convolutions and fed into different discriminators) so I think:
major features (scenery) are controlled by the prompt (grass or couch)
middle features (text) are a source image that the AI is punished for straying from
minor features (details) are controlled by the prompt (faces and fur)
Or it’s just Stable Diffusion that starts with a text rather than random noise.
What’s the prompt to get this stuff done? This looks wild and interesting.
I suspect these are using additional tools to guide the AI beyond a simple prompt. For example the spiraling medieval village was generated with stable diffusion and controlnet.
https://arstechnica.com/information-technology/2023/09/dreamy-ai-generated-geometric-scenes-mesmerize-social-media-users/
Not sure what this prompt was, since I didn’t make this one. I linked the site I used above, and they’re pretty simple to do but need a few tries to get a good one.
I think the prompt is not much other than “puppies” and “kittens”. Major, middle and minor features of the image can be controlled individually in some AIs (they can be differentiated using a Fourier transform or Gauss convolutions and fed into different discriminators) so I think:
Or it’s just Stable Diffusion that starts with a text rather than random noise.