×

Dynamic AI
Co-Creation

A Human-Centered Approach
by Will Luers

Created through the Digtal Pubishing Initiative at The Creative Media and Digital Culture program, with support of the OER Grants at Washington State University Vancouver.

publisher logo

Chapter 4: AI Images

1. Image Generation

AI image generation tools, such as Midjourney, DALL-E, and Stable Diffusion, have profoundly changed how we can create visual images. These tools synthesize visual works based on text prompts, effectively remixing open source, public domain and copyrighted training data. For instance, a prompt specifying "a Van Gogh landscape painting" creates a dataset of the artists' work online that is then used for analysis and generation of a new image without directly copying any specific painting. But what value does a newly genetrated "Van Gogh" painting have? Such an exercise of copying an artist style provides useful insights into the technology, but is otherwise of little artistic value. On the other hand, a prompt that twists references might be the start of a more interesting creative project. What is "a Van Gogh sculpture?" While AI image generation has raised concerns about copying and copyright infringement, AI image generation extends far beyond mere imitation.

AI tools offer vast potential across numerous fields by providing a powerful means of visual exploration and rapid ideation. For interior designers and architects, AI-generated images can help visualize and refine design concepts, allowing clients to see different styles, color schemes, and layouts before any physical work begins. This can facilitate better communication and decision-making, ensuring that the final design aligns with the client's vision.

Filmmakers can use AI to explore production elements such as location, costume, and lighting ideas before committing significant resources. By generating visual mock-ups, directors can experiment with different atmospheres and aesthetics, aiding in pre-visualization and enhancing creative planning. This iterative process can lead to more informed and effective production decisions, ultimately saving time and money.

In education, AI-generated images can be used to create engaging and illustrative materials for teaching complex concepts. In research, scientists might visualize data or create conceptual visualizations in novel ways, enabling new insights and discoveries. Artists can collaborate with AI to push the boundaries of their creativity, exploring unique styles and compositions that might not be possible through traditional means.

AI image generation tools have catalyzed numerous legal and ethical quandaries around copyright ownership and the nature of intellectual property in the age of machine intelligence. Some argue that the transformative process of an AI generating a new image from semantic prompts constitutes fair use. Others contend that commercial interests driving AI companies necessitate licensing agreements with rights holders for any training data usage. Beyond copyright, there are philosophical questions about whether AI-generated images can be granted copyright protections, as they are computational outputs lacking a clear human author. There are also concerns about generative AI's potential for enabling new forms of misinformation and manipulation, such as deepfake images and videos.

As generative AI becomes more ubiquitous across creative industries, there is a need for an evolution in legal frameworks to address these issues. At the same time, there is great cultural value in exploring the creative possibilities of these tools in ethical and human-centered ways.

30X40 Design Workshop | Using AI as a Design Tool in My Architecture Practice
Figma | AI and the future of design: Designing with AI
Script to Storyboard AI

2. GAN Technology

IBM Cloud: What are GANs (Generative Adversarial Networks)?

Generative Adversarial Networks (GANs) are a type of AI used to create images. Central to how GANs work is something called latent space. You can think of latent space as a hidden map, where each point represents the potential for creating a unique image. The generator in a GAN is like an artist that takes a set of coordinates (called a latent vector) from this map and transforms them into an image based on what it has learned from existing pictures.

latent space
Latent space

As the generator improves over time, it gets better at turning these coordinates into more realistic images. The discriminator acts like an art critic, judging the results and helping the generator refine its skills.

In simpler terms, latent space is where the GAN explores different possibilities. By moving through this space, the AI can create new and varied images that resemble the data it has been trained on. The closer two points are in latent space, the more similar the images they produce will be.

Generative Adversarial Networks (GANs) work through a back-and-forth process between two parts: one that creates images and one that evaluates them. This interaction leads to a kind of "synthetic imagination," where the AI generates new visuals that imitate the styles and details of the images it was trained on but with unique variations. Even small changes in the input can produce very different results, balancing between order and creativity. However, GANs do have limitations, such as mode collapse (when the AI keeps generating the same or similar images) and training instability (when the learning process breaks down), which can cause issues like repeated patterns or flawed images.

"a tree in the city"

3. Creative Strategies

Machine Art

latent space
Edmond de Belamy by Obvious art collective

The first AI artwork to be sold at auction, Edmond de Belamy, marked a groundbreaking moment in the art world, showcasing the novel capabilities of machine-generated creativity. Created by the Paris-based art collective Obvious, this portrait was not painted by a human hand but by a machine. Utilizing a Generative Adversarial Network (GAN), the AI was trained on a dataset of 15,000 portraits spanning six centuries. The resulting piece, characterized by its hauntingly abstract features, reflects the essence of classic portraiture while simultaneously challenging traditional notions of artistry. When Edmond de Belamy was auctioned at Christie’s in 2018, it fetched an astonishing $432,500, far exceeding initial estimates. This sale highlighted the profound potential of AI in art, demonstrating that, with the right data and algorithms, machines can produce works that are both innovative and evocative, without direct human intervention beyond their initial programming.

While GANs and diffusion models are powerful visual creation tools, many artists use AI as a creative assistant rather than an autonomous image generator. Key strategies include:

  • Image-to-Image Translation: Start with an initial image like a rough sketch, painting, or photo. For example, an artist might use a hand-drawn sketch of a futuristic cityscape as the input for an AI model. The AI can then generate detailed, stylistically coherent versions of the sketch, allowing the artist to maintain control over the composition while exploring different visual styles and details. This technique helps guide the generative process into a desired visual direction.
  • Textual Inversion: Artists can give semantic meaning to their specific training images, enabling AI models to better understand and recreate unique contents, styles, and attributes. For instance, a photographer might upload a series of images with distinct lighting and compositional styles. By training the AI on these images, the artist can generate new "photos" that adhere to their specific aesthetic, such as replicating the mood of lighting.
  • Crafted Prompts: Carefully crafting text descriptions and instructions provided to an AI image generator allows artists to synthesize visual concepts grounded in rich contextual cues from language. For example, a filmmaker could use descriptive prompts to generate storyboard frames for a scene set in a dystopian future, specifying elements like "neon-lit streets," "crumbling buildings," and "hovering drones." This helps in visualizing scenes and refining ideas before production.
  • Iterative Curation: Rather than accepting an AI's first output, artists iterate by feeding the generated image back into the model as a new input. For instance, a designer working on a new logo might use AI to generate initial concepts, select the most promising ones, and then refine them through multiple iterations. By repeatedly curating and refining, the designer can guide the AI to produce a polished final design that aligns with their vision.

These methods emphasize the human's central creative role in priming, curating, and combining the AI's generative capabilities with their own aesthetics and visions. AI becomes a collaborator and tutor, expanding the toolset for ideation and realization while respecting the artist's orchestration of the final artifacts.

How This Guy Uses A.I. to Create Art | Obsessed | WIRED

4. Generative Imaging Tools

The past few years have witnessed an explosive proliferation of platforms and software tools enabling AI-assisted image creation and manipulation:

Generative Models

  • Stable Diffusion - An open-source AI model that generates detailed images from text prompts, allowing for customization and fine-tuning by users.
  • DALL-E 2 - OpenAI's image generator that creates realistic or artistic images from text descriptions, known for its high-quality and creative outputs.
  • Midjourney - A powerful AI tool for generating imaginative and artistic visuals from text prompts, known for its unique stylistic capabilities.
  • RunwayML - A multimodal creative platform that offers AI-powered tools for image, video, and audio generation, with a user-friendly interface for artists.

Image Editing/Transformation Tools

  • Adobe Firefly - Adobe's generative AI tool that enables image generation and editing with advanced features integrated into the Adobe Creative Cloud suite.
  • RunwayML - Besides image generation, RunwayML provides advanced editing and transformation features, allowing users to create animations, apply filters, and more.

Sketch-to-Render Tools

  • Vizcom - A sketch-to-render tool that allows users to upload rough sketches and convert them into polished 3D renders or detailed concept art using AI.
  • PromeAI - An intuitive AI tool that turns hand-drawn sketches into fully rendered images with impressive detail, often used for creative and design projects.
  • Clipdrop Stable Doodle - This tool combines basic sketches with AI's ability to transform them into detailed, high-quality images, leveraging Stable Diffusion technology.
  • Scribble Diffusion - A fun, easy-to-use sketch-to-image tool that allows users to convert hand-drawn doodles into detailed images using the power of AI.

Practical AI Image Generation Tools

  • Remove.bg - A practical tool for quickly removing backgrounds from images using AI, useful for photo editing and graphic design.
  • Let's Enhance - An AI-based image enhancer that improves image resolution and quality, often used for upscaling photos or making low-resolution images clearer.
  • Topaz Gigapixel AI - A tool designed to enlarge images without losing detail, using AI to improve image sharpness and clarity, making it ideal for photographers and designers.
  • Deep Image AI - An online AI service for improving image quality, noise reduction, and upscaling, tailored for commercial use in marketing, printing, and e-commerce.

The rapid pace of development in this space means that incredibly capable new tools are emerging constantly. With many offered through accessible web interfaces or applications, AI-powered visual creation is quickly becoming open to everyone, not just skilled artists and developers.

5. AI Artists

While the AI art tools themselves are impressive technological marvels, equally essential are the pioneering human artists pushing the boundaries of how this tech can expand modes of creative expression and communication. Notable AI artists include:

AI Artists

Sofia Crespo

latent space
Sofia Crespo - LINK

Sofia Crespo is a pioneering artist whose work focuses on the intersection of biology and artificial intelligence. She uses AI models, particularly GANs and neural networks, to create intricate digital art pieces that explore the relationship between natural and artificial life forms. Her work often features organic shapes and textures reminiscent of biological entities, reflecting on how AI can mimic and interpret the complexities of the natural world. Crespo’s notable projects include "Neural Zoo," where she generates images of speculative creatures and plants that do not exist in reality but appear convincingly organic, challenging our perceptions of nature and machine-generated art.


Sougwen Chung

latent space
Sougwen Chung - LINK

Sougwen Chung is an interdisciplinary artist who explores the relationship between humans and machines through collaborative drawings and installations. Her work frequently involves drawing alongside robotic arms controlled by AI systems that learn from her style. This ongoing human-AI collaboration, as seen in her project "Drawing Operations," highlights the evolving nature of co-creativity and questions the role of autonomy in artistic creation. Chung’s practice bridges art, performance, and technology, offering a poetic exploration of how humans and machines can create together.


Refik Anadol

latent space
Refik Anadol - LINK

Refik Anadol is an artist known for his immersive installations that transform data into visually stunning and thought-provoking art pieces. He leverages AI and machine learning algorithms to process large datasets, such as urban landscapes, social media interactions, and cultural archives, turning them into dynamic visualizations and media sculptures. Anadol's work often involves projecting these data-driven visuals onto architectural surfaces, creating a seamless blend of the physical and digital realms. His projects like "Infinity Room" and "Melting Memories" push the boundaries of how data can be experienced aesthetically, offering a glimpse into the future of media art and the potential of AI to reshape our interaction with information.


Anna Ridler

latent space
Anna Ridler - LINK

Anna Ridler is an artist who works with both AI and data to explore themes of narrative, storytelling, and ownership of information. She often creates hand-annotated datasets, which are then used to train AI models, allowing her to maintain creative control over the outputs. One of her most well-known works is "Mosaic Virus," where she used a dataset of tulip images to comment on historical tulip mania and contemporary financial bubbles. Ridler’s work offers a thoughtful examination of how humans can shape the datasets that AI learns from, turning the creative process into a more personal and intentional act.


Ben Snell

latent space
Ben Snell - LINK

Ben Snell is an artist who uses artificial intelligence to explore the intersection of human creativity and machine learning. He trains custom AI models on datasets of his own sculptures, letting the machine generate new forms that he then brings into physical existence. By transforming AI-generated concepts into tangible objects, Snell's work investigates the evolving relationship between creator and tool, and the ways in which machines can co-author the artistic process. His project "Dio" involved feeding a computer's physical body back into the machine, creating a self-referential loop of artistic creation.


Roma Lipski

latent space
Roma Lipski - LINK

Roman Lipski is a Polish painter who collaborates with artificial intelligence to push the boundaries of his artistic practice. Through a constant exchange between the artist and an AI system, Lipski's work has evolved into a process of co-creation where the machine suggests novel ideas based on his past works. His "Unfinished" series is an ongoing dialogue between human intuition and machine learning, where AI not only inspires but also influences the creative process. This partnership explores the potential of AI to expand human imagination and challenges traditional notions of authorship in the arts.

6. Unit Exercise: AI-Assisted Comic Creation

This exercise explores how comic artists can leverage AI image generation tools to enhance their creative workflow, from initial sketches to final colored panels:

  1. Character Sketch Generation: Begin by sketching on paper characters for your comic. Sketch multiple variations for each character. Take photos of your favorites with a smartphone.
  2. Panel Composition: Upload the photos of sketches into an AI image generator to generate panels from your sketches, such as Vizcom. Many image generators allow you to upload an image as a reference. Create additional prompt descriptions that describe key scenes from your comic, such as: "Low-angle shot of pirate captain dramatically silhouetted against moon." Experiment with different panel compositions, camera angles, and character interactions. Generate at least three distinct panels for a three panel layout.
  3. Style Application: Choose a cohesive art style for your comic. Use AI tools to apply this style across your character designs and panel layouts. Iterate on the results, adjusting prompts to refine the style application.
  4. AI-Human Collaboration: Take your AI-generated panels into your preferred digital art software, such as Adobe Illustrator or Photoshop. Arrange the panels in a grid or sequence. Refine character designs, adding personal touches and ensuring consistency. Adjust panel compositions, adding additional details and background elements. Iterate with other image generations. Or generate backgrounds for your characters. Complete at least one fully realized comic strip of multiple panels.
  5. Creative Reflection: Analyze your completed AI-assisted comic panel. How did AI tools enhance or challenge your usual comic creation process? In what ways did AI-generated elements inspire new creative directions? Where did you find it necessary to intervene with your artistic skills?

7. Discussion Questions

  • How do AI-generated artworks challenge traditional notions of creativity, artistic expression, and the distinction between human and machine input?
  • As with other breakthrough technolgies in human history, how can AI expand the boundaries of human creativity?
  • How should the art world respond to AI-generated outputs that are virtually indistinguishable from human artworks? What implications does this have for the value and authenticity of human-created art?
  • What ethical guidelines are necessary to ensure that AI-generated art does not perpetuate harmful cultural stereotypes, misappropriate underrepresented artistic styles, or enable misinformation and manipulation?
  • How might economic models surrounding AI art, such as royalties from image sales and commissions, impact the livelihoods of human artists and content creators? What new models could emerge?
  • As AI tools become integrated into visual arts education and practice, how can we instill strong ethical frameworks in the next generation of AI-augmented artists? What should these frameworks include?
  • How can different fields like interior design, architecture, and filmmaking leverage AI image generation to enhance their creative processes? Provide specific examples of how these industries might use AI tools effectively.

8. Bibliography

  • Arrieta, et al. "Creativity and the Arts with Artificial Intelligence." ArXiv, 2021.
  • Audry, Sofian. Art in the Age of Machine Learning. Leonardo, 2021.
  • Du Sautoy, Marcus. The Creativity Code: Art and Innovation in the Age of AI. Harvard University Press, 2019.
  • Elgammal, Ahmed, et al. "CAN: Creative Adversarial Networks, Generating 'Art' by Learning About Styles and Deviating from Style Norms." ArXiv, 2017.
  • McCormack, Jon, and Mark d’Inverno, editors. Computers and Creativity. Springer, 2012.
  • Miller, Arthur I. The Artist in the Machine: The World of AI-Powered Creativity. MIT Press, 2019.
  • Zylinska, Joanna. AI Art: Machine Visions and Warped Dreams. Open Humanities Press, 2020.
Dynamic AI Co-Creation: A Human-Centered Approach
by Will Luers | Sept. 2024