thing 12: Generating Images

Have you ever thought, I’d like to find an image of a sunset beach? A quick internet search will usually do the trick. But what if what you really wanted was:

An impressionist painting of a sunset beach attended by Victorian-era ghosts having a tea party on jellyfish-shaped beach towels?

While a typical internet search is ideal for finding real or existing images or photographs of specific people or places, if your goal is to create custom illustrations, conceptual art or visualizing ideas of things that may not even exist, you’ll need a different tool (besides your own talent and competence of course).

AI image generators really do put the generative in generative AI. AI image generators don’t search the web or refer to a database of images. Instead, they generate entirely new visual content by drawing on patterns and relationships learned from millions of examples. The image doesn’t exist until your prompt activates a generative process that brings your idea to life.

So how does it do that?

Read

How does the technology work? (Univ. of Arizona Libraries)

When it comes to AI image generation you don’t find the image. You write it into existence. The prompt is the blueprint.

Image Prompting 101 (Microsoft)

Activity

AI Image Prompting Walkthrough: Yarn Garden Edition

We’ll be using Microsoft Copilot for this exercise which uses the Dall-E model from openAI.
Heads up: If you’re seeing a message about daily limits while using Copilot’s image generator, it just means you’ve hit the usage cap for the day. This also may impact the complexity of images Copilot delivers. You’re still doing the activity right! Try another image generator like ChatGPT, or Ideogram, or Gemini, or Krea!

Optional: Download this companion document for some hints on how you might remix the prompts in this activity. There’s also an additional and optional activity and some more prompts to try out.

Prompting Tip

If using Copilot, to make sure Copilot generates an image and not a paragraph of text, start your prompt with a visual action verb, such as:

  • Create, Draw, Generate, Illustrate, Design, or Visualize

Example:

❌ A beautiful mountain made of books → (Might return text)

✅ Generate an oil painting of a beautiful mountain made of books → (Image result!)

Prompt Formula

Here is a rough formula we’ll use to craft increasingly descriptive image prompts:

Create a [image type or medium] of [subject(s)] in [setting or environment], [composition or layout], [style or mood], [lighting or texture], and optionally [camera or rendering detail].

Example: Create a watercolor painting of two golden retrievers lounging on a front porch in autumn, framed by a wooden railing and scattered leaves, in a cozy and nostalgic mood, with soft afternoon lighting and gentle brush textures.

Activity

Prompting in Action: The Yarn Garden

Before you start, head over to Microsoft Copilot (or ChatGPT, or Ideogram, or Gemini, or Krea!). You can copy and paste the prompts below and see what happens as you change the descriptive language of your prompt. These prompts will get you started, but feel free to try out your own versions and variations!

PROMPT 1: Basic PROMPT

Create a digital illustration of a garden in the countryside with rows of vegetables and a sunset.

PROMPT 2: Add Texture and Lighting

Generate a whimsical digital illustration of a garden made entirely from yarn, set during golden hour, showing colorful vegetables, swaying sunflowers, and soft yarn textures.

PROMPT 3: Focus on Scene Details

Create an image of a scene where rows of tomatoes, peppers, and carrots made from colorful yarn are surrounded by yarn sunflowers and scattered yarn seeds, captured in a wide-angle view, lit by a glowing yarn sunset, evoking a cozy and enchanting atmosphere.

PROMPT: Advanced Variation

Design a high-resolution digital illustration of a magical yarn garden at sunset, with vegetables and flowers crafted from vibrant thread, styled like a stop-motion children’s book, using soft natural lighting and tactile fiber textures to create a handmade, heartwarming tone.

To see the prompt used to create the images at the top of this page click the small triangle. The images were all generated with the exact same prompt using different tools.

A high-resolution digital illustration in a stop-motion children’s book style depicts a whimsical yarn garden at sunset. At the heart of the scene sits a plump, vibrant tomato constructed entirely of scarlet thread, with delicate green tendrils woven to resemble leaves and supported by a spool base. Around the tomato are crocheted sunflowers with bright yellow petals and fuzzy brown centers, alongside miniature carrot sculptures crafted from orange yarn, all nestled within a bed of moss-like textured thread. Soft, warm sunlight filters through the scene, casting long shadows and illuminating the tactile fiber textures, creating a heartwarming and inviting atmosphere.

Discussion

Which part of your prompt had the greatest impact? Was it the the subject, the material, the lighting, or the style? Or something else? How do you know?

17 replies on “thing 12: Generating Images”

In copilot it only created the first prompt. It errored every prompt after that. Very disappointing. I used ChatGPT after that, and it created a very similar first image oddly enough. It took the prompt very literally, as it just had the listed vegetables placed on the ground, and not them on a plant like you would expect. The third and fourth images look extremely similar. The greatest impact for my responses were specific defined details. It took it very literally, and the “heartwarming tone” did not seem to mean anything to it- the AI didn’t know how to visualize heartwarming.

Went off on a totally different direction and ended up doing my own prompt. I asked: “Create an impressionist painting of Victorian ghosts drinking tea in a poison garden.” I wanted to see if it knew what a poison garden was or if it would totally go sideways in its interpretation. It didn’t. It knew what a poison garden was. I asked for the same thing, except to give the painting a gothic fairytale feel, which then took the impressionist style out for some reason. I then tried a third time to give the original prompt a golden hour glow. It started to give me error messages after that.

The image Copilot created, after following the prompts provided above, was very two-dimensional and didn’t look like the image above that was generated by Copilot. I asked to add farm animals and Copilot added a cute kitten, dog and some rabbits. And then when I tried additional prompts, I was told I had reached my daily limit. I would try again and try to be more specific and creative with prompts.

I think that the last three photos Copilot gave looked almost identical, especially the final two photos. The instructions on lighting seemed to have made the biggest difference, as there was a distinct difference in warmth once that was specified. I think it quickly hit a wall with how much it was going to change, and it got in the groove of making tiny, almost imperceptible changes.

I think ChatGPT (or other GenAI) could just write their own prompts for this stuff. I asked ChatGPT to define to two concepts “self compassion” and “psychological richness.” Then I asked it to “Describe a piece of visual art that would inspire the audience to imagine a reinforcing relationship between self compassion and psychological richness.” It came up with a description that was much richer and more beautiful than I could have composed (given that I am not much of an artist). Then I asked it to “translate that description to a large format acrylic painting,” where it outlined all the details about the visual structure, color palette, and interpretative statement for the gallery wall. The end product was actually quite nice.

I think the phrase “lit by a glowing yarn sunset” had the most impact. It resulted in the only image that had rays extending from the sun and a brighter, lighter image. It was interesting that the image created from the prompt using the word “glowing” was the only “yarn” image that had sunflowers but no purple flowers.

The material and style made the most difference from the first prompt. However, those differences weren’t THAT impactful. If I were going to use this image in a presentation, the differences were subtle. This activity demonstrated how the tool just reuses information and how human originality is crucial.

The last two images were quite similar. Interestingly, the first two images both had houses in them. I thought those might be generated by the “in the countryside” prompt, since the house was barn-like (except for the chimney), but it was in the second prompt as well. Adding the “from yarn” portion of the prompt flattened the images — the third three look almost like needlepoint to me.

I was only able to do the first prompt with Copilot before I hit the image generating cap. I tried the second prompt in ChatGPT, and the image seemed to me to convey malicious sunflowers moving in to attack unsuspecting vegetables– which I thought was kind of cool. Nothing changed in the following prompts except the sky was a bit darker. Since I’m not an illustrator, I found this exercise to be an impressive display of what GenAI can do. If I was an illustrator, I’d be worried about losing my job and sad about the flattening of creativity that is “good enough” for non-illustrators.

I tried rewording this prompt several ways: “Create a digital illustration of a forest setting with a river running through it. In the forest are unicorns and in the river are mermaids.” Each time it started to craft an image and then gave me an error: “Sorry, I can’t do that. However, I can help you create a detailed written description of the scene for an artist or suggest tools you can use to bring your vision to life. Would you like a description or some creative tool recommendations?”

I ended up having to change it to a 2 part prompt, where I first asked for: “Create a storybook image of unicorns in a forest at sunset. ” and then asked it to “Include mermaids in a river.” Interestingly it was able to add the mermaids but then it lost the sunset portion.

The images were fine but it took a lot of time and prompting. Not sure I’ll do that again unless I need something super specific.

I tried Prompt 1 and then went to Prompt 4. The first pic resembled an oil painting and was very like something you would see in a children’s book ( I feel like). The 4th prompt let to an image very similar to the one shown at the top of the site, however the sunflowers were standing up straight and I had a square pattern on the ground (possibly resembling gardening beds). I would love to see the results from others and what differences there are in the results when using the same prompt and the same AI tool. I feel the more detail you can give the better and tested that theory by taking some detail away from Prompt 4 (Create A high-resolution digital illustration in children’s book style of a yarn garden at sunset. At the heart of the scene sits a tomato constructed entirely of thread, with tendrils woven to resemble leaves and supported by a spool base. Around the tomato are crocheted sunflowers and miniature carrot sculptures. Sunlight filters through the scene) and the results were very similar (however it made a mistake and it looks like a sunflower grows out of a carrot lol). I then took that prompt and put it in ChatGPT as comparison. The image was quite different. I used the longer, more detailed prompt then in ChatGPT to see if the image would be similar then. The result was almost identical to the first one. So more detail does not result in changes once we reached a certain level of information already it seems.

I used the example prompts and the biggest changes were form prompt one to prompt two and prompt three to prompt four. Prompt one was an actual realistic farm with rows of vegetables and realistic lighting while three, four, and five were just plants and a sun behind them. Adding the “digital illustration” modifier in prompt four made the image go from two-dimensional to three-dimensional. The shadows were still unrealistic and there was a weird plant that grew a Sunflower, Tomato, and Bell pepper on the same plant.

I created a different image of a dog dancing as a ballerina on a beach. The picture was just ok, easily detected as a generated photo; but I did appreciate the shadow and the ocean waves crashing in the background.

I asked Copilot to “Create a pointillist painting of a garden in the countryside with rows of vegetables during golden hour” and then to shift the focus of the painting to the foreground. It was quirky–it shifted the frame of the ‘painting’ as you might expect, and it changed the size of the dots somewhat–larger in the background more precise and pixilated in the foreground like a camera (not like pointillism). I am also not sure it understands what visual focus is either–the garden rows in the image while “close up” draw the eye out into the distance of the second image creating something where your eye is dwelling on the upper third of the frame and not the foreground.

The first picture created by CoPilot looked like a painting, but it had huge rows of carrots laying on their sides. It was quite comical. Then, it just got too tired to give me more for the next prompt. I do think mentioning style helps a lot. I have used CoPilot to create pictures for presentations. It was gotten much faster than a year ago. I do notice that it really doesn’t do human hands well. Usually too many fingers.

The first picture created by CoPilot looked like a painting, but it had huge rows of carrots laying on their sides. It was quite comical. Then, it just got too tired to give me more for the next prompt. I do think mentioning style helps a lot. I have used CoPilot to create pictures for presentations. It has gotten much faster than a year ago. I do notice that it really doesn’t do human hands well. Usually too many fingers.

To me, step 3 impacted the most because I asked chat gpt how to set up the specific scene (sunflowers, how they surround other things etc). It’s really really slow and every single time, it creates a somewhat different picture. Instead of adding new things on an existing photo, it just kept creating completely new things

Leave a Reply