×

Dynamic AI
Co-Creation

A Human-Centered Approach
by Will Luers

Created through the Digtal Pubishing Initiative at The Creative Media and Digital Culture program, with support of the OER Grants at Washington State University Vancouver.

publisher logo

Chapter 4: AI Images

1. Image Generation

AI image generation tools, such as Midjourney, DALL-E, and Stable Diffusion, have profoundly changed how we can create visual images. These tools synthesize visual works based on text prompts, effectively remixing open source, public domain and copyrighted training data. For instance, a prompt specifying "a Van Gogh landscape painting" creates a dataset of the artists' work online that is then used for analysis and generation of a new image without directly copying any specific painting. But what value does a newly genetrated "Van Gogh" painting have? Such an exercise of copying an artist style provides useful insights into the technology, but is otherwise of little artistic value. On the other hand, a prompt that twists references might be the start of a more interesting creative project. What is "a Van Gogh sculpture?" While AI image generation has raised concerns about copying and copyright infringement, AI image generation extends far beyond mere imitation.

AI tools offer vast potential across numerous fields by providing a powerful means of visual exploration and rapid ideation. For interior designers and architects, AI-generated images can help visualize and refine design concepts, allowing clients to see different styles, color schemes, and layouts before any physical work begins. This can facilitate better communication and decision-making, ensuring that the final design aligns with the client's vision.

Filmmakers can use AI to explore production elements such as location, costume, and lighting ideas before committing significant resources. By generating visual mock-ups, directors can experiment with different atmospheres and aesthetics, aiding in pre-visualization and enhancing creative planning. This iterative process can lead to more informed and effective production decisions, ultimately saving time and money.

In education, AI-generated images can be used to create engaging and illustrative materials for teaching complex concepts. In research, scientists might visualize data or create conceptual visualizations in novel ways, enabling new insights and discoveries. Artists can collaborate with AI to push the boundaries of their creativity, exploring unique styles and compositions that might not be possible through traditional means.

AI image generation tools have catalyzed numerous legal and ethical quandaries around copyright ownership and the nature of intellectual property in the age of machine intelligence. Some argue that the transformative process of an AI generating a new image from semantic prompts constitutes fair use. Others contend that commercial interests driving AI companies necessitate licensing agreements with rights holders for any training data usage. Beyond copyright, there are philosophical questions about whether AI-generated images can be granted copyright protections, as they are computational outputs lacking a clear human author. There are also concerns about generative AI's potential for enabling new forms of misinformation and manipulation, such as deepfake images and videos.

As generative AI becomes more ubiquitous across creative industries, there is a need for an evolution in legal frameworks to address these issues. At the same time, there is great cultural value in exploring the creative possibilities of these tools in ethical and human-centered ways.

30X40 Design Workshop | Using AI as a Design Tool in My Architecture Practice
Figma | AI and the future of design: Designing with AI
Script to Storyboard AI

2. GAN Technology

IBM Cloud: What are GANs (Generative Adversarial Networks)?

Generative Adversarial Networks (GANs) are a type of AI used to create images. Central to how GANs work is something called latent space. You can think of latent space as a hidden map, where each point represents the potential for creating a unique image. The generator in a GAN is like an artist that takes a set of coordinates (called a latent vector) from this map and transforms them into an image based on what it has learned from existing pictures.

latent space
Latent space

As the generator improves over time, it gets better at turning these coordinates into more realistic images. The discriminator acts like an art critic, judging the results and helping the generator refine its skills.

In simpler terms, latent space is where the GAN explores different possibilities. By moving through this space, the AI can create new and varied images that resemble the data it has been trained on. The closer two points are in latent space, the more similar the images they produce will be.

Generative Adversarial Networks (GANs) work through a back-and-forth process between two parts: one that creates images and one that evaluates them. This interaction leads to a kind of "synthetic imagination," where the AI generates new visuals that imitate the styles and details of the images it was trained on but with unique variations. Even small changes in the input can produce very different results, balancing between order and creativity. However, GANs do have limitations, such as mode collapse (when the AI keeps generating the same or similar images) and training instability (when the learning process breaks down), which can cause issues like repeated patterns or flawed images.

"a tree in the city"

3. Creative Strategies

Machine Art

latent space
Edmond de Belamy by Obvious art collective

The first AI artwork to be sold at auction, Edmond de Belamy, marked a groundbreaking moment in the art world, showcasing the novel capabilities of machine-generated creativity. Created by the Paris-based art collective Obvious, this portrait was not painted by a human hand but by a machine. Utilizing a Generative Adversarial Network (GAN), the AI was trained on a dataset of 15,000 portraits spanning six centuries. The resulting piece, characterized by its hauntingly abstract features, reflects the essence of classic portraiture while simultaneously challenging traditional notions of artistry. When Edmond de Belamy was auctioned at Christie’s in 2018, it fetched an astonishing $432,500, far exceeding initial estimates. This sale highlighted the profound potential of AI in art, demonstrating that, with the right data and algorithms, machines can produce works that are both innovative and evocative, without direct human intervention beyond their initial programming.

While GANs and diffusion models are powerful visual creation tools, many artists use AI as a creative assistant rather than an autonomous image generator. Key strategies include:

  • Image-to-Image Translation: Start with an initial image like a rough sketch, painting, or photo. For example, an artist might use a hand-drawn sketch of a futuristic cityscape as the input for an AI model. The AI can then generate detailed, stylistically coherent versions of the sketch, allowing the artist to maintain control over the composition while exploring different visual styles and details. This technique helps guide the generative process into a desired visual direction.
  • Textual Inversion: Artists can give semantic meaning to their specific training images, enabling AI models to better understand and recreate unique contents, styles, and attributes. For instance, a photographer might upload a series of images with distinct lighting and compositional styles. By training the AI on these images, the artist can generate new "photos" that adhere to their specific aesthetic, such as replicating the mood of lighting.
  • Crafted Prompts: Carefully crafting text descriptions and instructions provided to an AI image generator allows artists to synthesize visual concepts grounded in rich contextual cues from language. For example, a filmmaker could use descriptive prompts to generate storyboard frames for a scene set in a dystopian future, specifying elements like "neon-lit streets," "crumbling buildings," and "hovering drones." This helps in visualizing scenes and refining ideas before production.
  • Iterative Curation: Rather than accepting an AI's first output, artists iterate by feeding the generated image back into the model as a new input. For instance, a designer working on a new logo might use AI to generate initial concepts, select the most promising ones, and then refine them through multiple iterations. By repeatedly curating and refining, the designer can guide the AI to produce a polished final design that aligns with their vision.

These methods emphasize the human's central creative role in priming, curating, and combining the AI's generative capabilities with their own aesthetics and visions. AI becomes a collaborator and tutor, expanding the toolset for ideation and realization while respecting the artist's orchestration of the final artifacts.

How This Guy Uses A.I. to Create Art | Obsessed | WIRED

4. Generative Imaging Tools

The past few years have witnessed an explosive proliferation of platforms and software tools enabling AI-assisted image creation and manipulation:

The rapid pace of development in this space means that incredibly capable new tools are emerging constantly. With many offered through accessible web interfaces or applications, AI-powered visual creation is quickly becoming open to everyone, not just skilled artists and developers.

5. AI Artists

While the AI art tools themselves are impressive technological marvels, equally essential are the pioneering human artists pushing the boundaries of how this tech can expand modes of creative expression and communication. Notable AI artists include:

AI Artists

Sofia Crespo

latent space
Sofia Crespo - LINK

Sofia Crespo is a pioneering artist whose work focuses on the intersection of biology and artificial intelligence. She uses AI models, particularly GANs and neural networks, to create intricate digital art pieces that explore the relationship between natural and artificial life forms. Her work often features organic shapes and textures reminiscent of biological entities, reflecting on how AI can mimic and interpret the complexities of the natural world. Crespo’s notable projects include "Neural Zoo," where she generates images of speculative creatures and plants that do not exist in reality but appear convincingly organic, challenging our perceptions of nature and machine-generated art.


Refik Anadol

latent space
Refik Anadol - LINK

Refik Anadol is an artist known for his immersive installations that transform data into visually stunning and thought-provoking art pieces. He leverages AI and machine learning algorithms to process large datasets, such as urban landscapes, social media interactions, and cultural archives, turning them into dynamic visualizations and media sculptures. Anadol's work often involves projecting these data-driven visuals onto architectural surfaces, creating a seamless blend of the physical and digital realms. His projects like "Infinity Room" and "Melting Memories" push the boundaries of how data can be experienced aesthetically, offering a glimpse into the future of media art and the potential of AI to reshape our interaction with information.


latent space
Stephanie Dinkins - LINK

Stephanie Dinkins is an artist and educator whose work critically examines the intersections of AI, race, gender, and social equity. She is known for her long-term project "Conversations with Bina48," in which she engages with an advanced social robot to explore issues of consciousness, bias, and the potential for AI to embody diverse perspectives. Dinkins’ work often involves participatory and community-based practices, aiming to democratize AI technologies and make them accessible to marginalized communities. Through her art, she seeks to foster dialogues about the ethical implications of AI and advocate for the inclusion of diverse voices in the development of AI systems.

6. Unit Exercise: AI-Assisted Comic Creation

This exercise explores how comic artists can leverage AI image generation tools to enhance their creative workflow, from initial sketches to final colored panels:

  1. Character Sketch Generation: Begin by sketching on paper characters for your comic. Sketch multiple variations for each character. Take photos of your favorites with a smartphone.
  2. Panel Composition: Upload the photos of sketches into an AI image generator to generate panels from your sketches. Many image generators allow you to upload an image as a reference. Create additional prompt descriptions that describe key scenes from your comic, such as: "Low-angle shot of pirate captain dramatically silhouetted against twin moons." Experiment with different panel compositions, camera angles, and character interactions. Generate at least three distinct panels for a three panel layout.
  3. Style Application: Choose a cohesive art style for your comic. Use AI tools to apply this style across your character designs and panel layouts. Iterate on the results, adjusting prompts to refine the style application.
  4. AI-Human Collaboration: Take your AI-generated panels into your preferred digital art software, such as Adobe Illustrator or Photoshop. Arrange the panels in a grid or sequence. Refine character designs, adding personal touches and ensuring consistency. Adjust panel compositions, adding additional details and background elements. Iterate with other image generations. Or generate backgrounds for your characters. Complete at least one fully realized comic strip of multiple panels.
  5. Creative Reflection: Analyze your completed AI-assisted comic panel. How did AI tools enhance or challenge your usual comic creation process? In what ways did AI-generated elements inspire new creative directions? Where did you find it necessary to intervene with your artistic skills?

7. Discussion Questions

  • How do AI-generated artworks challenge traditional notions of creativity, artistic expression, and the distinction between human and machine input?
  • As with other breakthrough technolgies in human history, how can AI expand the boundaries of human creativity?
  • How should the art world respond to AI-generated outputs that are virtually indistinguishable from human artworks? What implications does this have for the value and authenticity of human-created art?
  • What ethical guidelines are necessary to ensure that AI-generated art does not perpetuate harmful cultural stereotypes, misappropriate underrepresented artistic styles, or enable misinformation and manipulation?
  • How might economic models surrounding AI art, such as royalties from image sales and commissions, impact the livelihoods of human artists and content creators? What new models could emerge?
  • As AI tools become integrated into visual arts education and practice, how can we instill strong ethical frameworks in the next generation of AI-augmented artists? What should these frameworks include?
  • How can different fields like interior design, architecture, and filmmaking leverage AI image generation to enhance their creative processes? Provide specific examples of how these industries might use AI tools effectively.

8. Bibliography

  • Arrieta, et al. "Creativity and the Arts with Artificial Intelligence." ArXiv, 2021.
  • Audry, Sofian. Art in the Age of Machine Learning. Leonardo, 2021.
  • Du Sautoy, Marcus. The Creativity Code: Art and Innovation in the Age of AI. Harvard University Press, 2019.
  • Elgammal, Ahmed, et al. "CAN: Creative Adversarial Networks, Generating 'Art' by Learning About Styles and Deviating from Style Norms." ArXiv, 2017.
  • McCormack, Jon, and Mark d’Inverno, editors. Computers and Creativity. Springer, 2012.
  • Miller, Arthur I. The Artist in the Machine: The World of AI-Powered Creativity. MIT Press, 2019.
  • Zylinska, Joanna. AI Art: Machine Visions and Warped Dreams. Open Humanities Press, 2020.
Dynamic AI Co-Creation: A Human-Centered Approach
by Will Luers | Sept. 2024