Dynamic AI
Co-Creation
A Human-Centered ApproachCreated through the Digtal Pubishing Initiative at The Creative Media and Digital Culture program, with support of the OER Grants at Washington State University Vancouver.
AI image generation tools, such as Midjourney, DALL-E, and Stable Diffusion, have profoundly changed how we can create visual images. These tools synthesize visual works based on text prompts, effectively remixing open source, public domain and copyrighted training data. For instance, a prompt specifying "a Van Gogh landscape painting" creates a dataset of the artists' work online that is then used for analysis and generation of a new image without directly copying any specific painting. But what value does a newly genetrated "Van Gogh" painting have? Such an exercise of copying an artist style provides useful insights into the technology, but is otherwise of little artistic value. On the other hand, a prompt that twists references might be the start of a more interesting creative project. What is "a Van Gogh sculpture?" While AI image generation has raised concerns about copying and copyright infringement, AI image generation extends far beyond mere imitation.
AI tools offer vast potential across numerous fields by providing a powerful means of visual exploration and rapid ideation. For interior designers and architects, AI-generated images can help visualize and refine design concepts, allowing clients to see different styles, color schemes, and layouts before any physical work begins. This can facilitate better communication and decision-making, ensuring that the final design aligns with the client's vision.
Filmmakers can use AI to explore production elements such as location, costume, and lighting ideas before committing significant resources. By generating visual mock-ups, directors can experiment with different atmospheres and aesthetics, aiding in pre-visualization and enhancing creative planning. This iterative process can lead to more informed and effective production decisions, ultimately saving time and money.
In education, AI-generated images can be used to create engaging and illustrative materials for teaching complex concepts. In research, scientists might visualize data or create conceptual visualizations in novel ways, enabling new insights and discoveries. Artists can collaborate with AI to push the boundaries of their creativity, exploring unique styles and compositions that might not be possible through traditional means.
AI image generation tools have catalyzed numerous legal and ethical quandaries around copyright ownership and the nature of intellectual property in the age of machine intelligence. Some argue that the transformative process of an AI generating a new image from semantic prompts constitutes fair use. Others contend that commercial interests driving AI companies necessitate licensing agreements with rights holders for any training data usage. Beyond copyright, there are philosophical questions about whether AI-generated images can be granted copyright protections, as they are computational outputs lacking a clear human author. There are also concerns about generative AI's potential for enabling new forms of misinformation and manipulation, such as deepfake images and videos.
As generative AI becomes more ubiquitous across creative industries, there is a need for an evolution in legal frameworks to address these issues. At the same time, there is great cultural value in exploring the creative possibilities of these tools in ethical and human-centered ways.
Generative Adversarial Networks (GANs) are a type of AI used to create images. Central to how GANs work is something called latent space. You can think of latent space as a hidden map, where each point represents the potential for creating a unique image. The generator in a GAN is like an artist that takes a set of coordinates (called a latent vector) from this map and transforms them into an image based on what it has learned from existing pictures.
As the generator improves over time, it gets better at turning these coordinates into more realistic images. The discriminator acts like an art critic, judging the results and helping the generator refine its skills.
In simpler terms, latent space is where the GAN explores different possibilities. By moving through this space, the AI can create new and varied images that resemble the data it has been trained on. The closer two points are in latent space, the more similar the images they produce will be.
Generative Adversarial Networks (GANs) work through a back-and-forth process between two parts: one that creates images and one that evaluates them. This interaction leads to a kind of "synthetic imagination," where the AI generates new visuals that imitate the styles and details of the images it was trained on but with unique variations. Even small changes in the input can produce very different results, balancing between order and creativity. However, GANs do have limitations, such as mode collapse (when the AI keeps generating the same or similar images) and training instability (when the learning process breaks down), which can cause issues like repeated patterns or flawed images.
The first AI artwork to be sold at auction, Edmond de Belamy, marked a groundbreaking moment in the art world, showcasing the novel capabilities of machine-generated creativity. Created by the Paris-based art collective Obvious, this portrait was not painted by a human hand but by a machine. Utilizing a Generative Adversarial Network (GAN), the AI was trained on a dataset of 15,000 portraits spanning six centuries. The resulting piece, characterized by its hauntingly abstract features, reflects the essence of classic portraiture while simultaneously challenging traditional notions of artistry. When Edmond de Belamy was auctioned at Christie’s in 2018, it fetched an astonishing $432,500, far exceeding initial estimates. This sale highlighted the profound potential of AI in art, demonstrating that, with the right data and algorithms, machines can produce works that are both innovative and evocative, without direct human intervention beyond their initial programming.
While GANs and diffusion models are powerful visual creation tools, many artists use AI as a creative assistant rather than an autonomous image generator. Key strategies include:
These methods emphasize the human's central creative role in priming, curating, and combining the AI's generative capabilities with their own aesthetics and visions. AI becomes a collaborator and tutor, expanding the toolset for ideation and realization while respecting the artist's orchestration of the final artifacts.
The past few years have witnessed an explosive proliferation of platforms and software tools enabling AI-assisted image creation and manipulation:
The rapid pace of development in this space means that incredibly capable new tools are emerging constantly. With many offered through accessible web interfaces or applications, AI-powered visual creation is quickly becoming open to everyone, not just skilled artists and developers.
While the AI art tools themselves are impressive technological marvels, equally essential are the pioneering human artists pushing the boundaries of how this tech can expand modes of creative expression and communication. Notable AI artists include:
Sofia Crespo is a pioneering artist whose work focuses on the intersection of biology and artificial intelligence. She uses AI models, particularly GANs and neural networks, to create intricate digital art pieces that explore the relationship between natural and artificial life forms. Her work often features organic shapes and textures reminiscent of biological entities, reflecting on how AI can mimic and interpret the complexities of the natural world. Crespo’s notable projects include "Neural Zoo," where she generates images of speculative creatures and plants that do not exist in reality but appear convincingly organic, challenging our perceptions of nature and machine-generated art.
Refik Anadol is an artist known for his immersive installations that transform data into visually stunning and thought-provoking art pieces. He leverages AI and machine learning algorithms to process large datasets, such as urban landscapes, social media interactions, and cultural archives, turning them into dynamic visualizations and media sculptures. Anadol's work often involves projecting these data-driven visuals onto architectural surfaces, creating a seamless blend of the physical and digital realms. His projects like "Infinity Room" and "Melting Memories" push the boundaries of how data can be experienced aesthetically, offering a glimpse into the future of media art and the potential of AI to reshape our interaction with information.
Stephanie Dinkins is an artist and educator whose work critically examines the intersections of AI, race, gender, and social equity. She is known for her long-term project "Conversations with Bina48," in which she engages with an advanced social robot to explore issues of consciousness, bias, and the potential for AI to embody diverse perspectives. Dinkins’ work often involves participatory and community-based practices, aiming to democratize AI technologies and make them accessible to marginalized communities. Through her art, she seeks to foster dialogues about the ethical implications of AI and advocate for the inclusion of diverse voices in the development of AI systems.
This exercise explores how comic artists can leverage AI image generation tools to enhance their creative workflow, from initial sketches to final colored panels: