Creative direction: The secret to great AI images

How to art direct prompts with taste, judgment, and intent

A clear glass bowl sits on a paint-splattered surface. Both are filled with creative tools, including paint tubes, color swatches, fabric samples, brushes, a camera, film, thread, and notebooks. Several hands reach in from the edges of the frame toward the bowl; one of the hands pours liquid into it, and another sprinkles in blue pigment.

AI image generated in Firefly Boards using Gemini 3.

If you’re a designer, you know the feeling of sitting through a presentation with no visual anchors. Slide after slide of Myriad Pro can drain energy from even the best ideas. It’s one of the reasons images have become a practical entry point into generative AI.

But anyone who’s spent time generating images knows how quickly outputs can turn strange, generic, or just unmistakably AI. Generative output alone doesn’t guarantee quality. That still depends on human taste—along with intent and judgment. Learning how to communicate those qualities clearly is what transforms AI from a slot machine to a creative partner.

Translating the image in your head into a prompt, though, is easier said than done. And I don’t know about you, but I’m not interested in becoming a prompt engineer. Instead, I lean into art direction—taking the creative language I already use and applying it to define generative outcomes.

The model doesn’t know what a good image looks like. You do. One or two iterations can be the difference between something you’d never use and something that feels right.

Over the past year, I’ve tested image models and methods shared by online communities and sharpened a process that improves quality and coherence. It’s a process grounded in the same things designers already think about: tone, style, and composition.

Use LLMs to write prompts

Prompt bars, with their single empty text field and blinking cursor, are deceptively simple. But writing a detailed prompt from scratch can take a surprising amount of time. And if you’re visually inclined like me, it’s not always the most satisfying part of the process.

One of the biggest unlocks I had, while trying to generate better images, was to separate prompt writing from image generation by using a large language model (LLM), like ChatGPT, to write my prompts. It sounds simple, but it can change the trajectory of the process: An LLM can get you past the blank-prompt barrier and into something actionable, fast.

Three images stacked. Each shows a moderately different version of a mint-green stand mixer, on a paint-splattered table, blending vivid pink, blue, yellow, and green art materials. A hand reaches in from the edge of the frame to steady the bowl or the mixer as the contents swirl into a multicolored batter.
Iterative image generation (Gemini 3) to achieve a clean, close-up shot from above.

But prompting is rarely one‑and‑done. While first results can be technically correct, visually coherent, and otherwise “fine,” they can also simultaneously feel lifeless and “close, but not quite right.” This is where working with an LLM really shines: Don’t automatically accept what an LLM gives you. Instead, think of it as a drafting partner. Read what it produces, then instead of starting over, change things up, override what feels wrong, and iterate with instructions like: “Make it feel more editorial,” or “Less realism and more imagination.”

The model doesn’t know what a good image looks like. You do. One or two iterations can be the difference between something you’d never use and something that feels right.

Talk like an art director

AI responds best when you’re explicit about the things designers already care about. When you break prompts into familiar categories, images become more intentional and less generic. AI creator Ohneis says that some of the most valuable dimensions to include are:

The list of possibilities is limited only by imagination, but Images generated without cues like those listed above often look fine but basic. Add them, and things start to feel more real and intentional.

Two images stacked. Each shows a slightly different version of a studio workspace with jars of pigment and tools nearby. In each, hands reach in from the edge of the frame to cut stacked crayons and textured art materials on a wooden cutting board.
Art direction adds dimension and visual interest to a photo (bottom). Without it, results feel flat and generic (top).

Support prompting with reference images

Starting from scratch works well when you have a clear mental image, but sometimes it’s easier to show than to tell. Reference images can dramatically speed things up, especially when you’re defining a visual direction. One to three somewhat similar images are a great starting point.

Upload those references to the LLM and ask it to, “Translate the image into a JSON context profile.” This formatted JavaScript Object Notation profile extracts visual characteristics—lighting, mood, textures, composition—and turns them into a structured description. From there, you can generate new prompts that carry the same DNA, even as the subject changes.

Two images stacked. Each shows a different version of trays being pulled from lighted ovens. In the top image gloved hands are removing freshly baked chocolate chip cookies and in the bottom image paint-splattered oven mitts are removing neatly piped, colorful meringues.
Using a reference image helped me achieve the angle and close-up perspective of a person putting a baking sheet in the oven.

Brand alignment is one of the trickier aspects of generative imagery, but it’s not impossible—especially when brand constraints are treated as part of the art direction, not an afterthought:

If your output is for customer work, you can extend your art direction by describing their mission, values, and identity directly to the LLM. That context will influence the prompt's tone in subtle yet meaningful ways. As with most things in prompting, specificity wins.

Experiment: Different models = different outcomes

Different image models also interpret the same prompt differently. One model might feel more photographic, while another might feel more stylized. With a strong prompt structure, the core elements stay intact while the aesthetic shifts. Choosing between models becomes an act of art direction. There’s no universally “best” model—only the one that’s right for the outcome you want. There are many ways to experiment. My favorite is to do a side-by-side comparison using partner models in Adobe Firefly Boards.

But experimentation only works if you’re not reinventing the prompt every time. Instead of starting from scratch for each model or variation, making what works reusable creates a stable foundation you can test across models without losing your intent.

Three images stacked. Each shows a slightly different version of a plated dessert-like form on a satin-covered surface. Hands, reaching in from the edge of the frame, use tweezers to adjust the dessert which consists of a piped, colorful meringue atop a stack of translucent and opaque sheets surrounded by marbles.
Using Firefly Boards, I experimented with one prompt in different image models: Firefly (top); GPT (middle); Gemini 3 (bottom). Each model preserves the core elements of the prompt with a slightly different interpretation.

Experimenting with different models is where JSON context profiles become valuable: They enable you to carry a visual style from a successful prompt into any number of future generations. Think of them as style containers that include the essential details that define an image's look and feel—composition, lighting, mood, materials—without locking you into a single subject

Once you have that, you can generate new prompts using the same blueprint, swap subjects, and try different models without losing consistency.

Generative AI doesn’t work without your taste

When you treat prompting as a form of art direction, the results start to feel thoughtfully curated and designed. Once you learn the formula, the process moves quickly, and your creativity can multiply. The approach comes down to a few repeatable habits:

  1. Use LLMs like ChatGPT to write prompts. They save time and generate highly detailed starting points.
  2. Understand design terminology and the language of art direction so you can use them in your prompts.
  3. Start from scratch or use reference images to accelerate prompt generation.
  4. Experiment with models and choose intentionally to find the look that best suits the ideas in your head.


At the end of the day, your creativity is what sets you apart. Use generative AI to explore your imagination, and bring life to your presentation decks, but let your taste, humanity, and instincts set its direction.

Header copy
Design your career at Adobe.
Button copy
View all jobs
Button link
/jobs