{"id":183,"date":"2025-08-06T11:17:34","date_gmt":"2025-08-06T10:17:34","guid":{"rendered":"https:\/\/generic.wordpress.soton.ac.uk\/genai\/?p=183"},"modified":"2025-11-04T17:48:54","modified_gmt":"2025-11-04T17:48:54","slug":"using-genai-for-image-generation","status":"publish","type":"post","link":"https:\/\/generic.wordpress.soton.ac.uk\/genai\/2025\/08\/06\/using-genai-for-image-generation\/","title":{"rendered":"Using GenAI for image generation"},"content":{"rendered":"\n<figure class=\"wp-block-image size-large\"><img decoding=\"async\" src=\"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-content\/uploads\/sites\/534\/2025\/09\/GettyImages-1483786049-1024x590.jpg\" alt=\"Creative humanoid robot working in the art studio, she is painting on canvas\" class=\"wp-image-184\" \/><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\">How does AI image generation work?&nbsp;<\/h2>\n\n\n\n<p>Many generative AI tools now include the ability to create brand-new images based on a text prompt. Some tools even allow you to work with original images that you can then prompt to be edited or added to with AI.&nbsp;&nbsp;<\/p>\n\n\n\n<p>The first thing to understand is that the new image created is not a collage of pre-existing images, it is a new image that has never been created before. But how does it work?&nbsp;&nbsp;<\/p>\n\n\n\n<p>In a lot of instances, including Microsoft\u2019s Co-pilot, the model used for creating images is DALL-E 3 (from Open AI). First DALL-E 3 was given a dataset of images and importantly their associated text captions, these sets are either from scraping images directly off the internet or by using curated datasets like LAION-5B, these data sets are typically very large for example LAION-5B has around 5.85 billion image\u2013text pairs.&nbsp;&nbsp;<\/p>\n\n\n\n<p>Nearly all the models use a diffusion technique to learn how to create new images, the model (DALL-E 3) adds \u2018noise\u2019 to all the images in the dataset and then attempts to reverse that process, back to the original image, the model then checks to see how well it has done. This training often takes months and requires a lot of computing power.&nbsp;<\/p>\n\n\n\n<p>Once the training is complete the model can then use that noise to create new images.&nbsp;<\/p>\n\n\n\n<p>Here is an example of the stages used in this instance by Stable Diffusion (this one was trained on a subset of the LAION-5B database.) to create a new image from noise.&nbsp;<\/p>\n\n\n\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"936\" height=\"633\" src=\"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-content\/uploads\/sites\/534\/2025\/09\/image-5.png\" alt=\"10 images showing the gradual formation of a castle. Each image gets sharper and more detailed as they go on.\" class=\"wp-image-185\" srcset=\"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-content\/uploads\/sites\/534\/2025\/09\/image-5.png 936w, https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-content\/uploads\/sites\/534\/2025\/09\/image-5-300x203.png 300w, https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-content\/uploads\/sites\/534\/2025\/09\/image-5-768x519.png 768w\" sizes=\"auto, (max-width: 936px) 100vw, 936px\" \/><\/figure>\n\n\n\n<p><em>By Benlisquare &#8211; Own work, CC BY-SA 4.0, https:\/\/commons.wikimedia.org\/w\/index.php?<\/em>curid=124800742 &nbsp;<\/p>\n\n\n\n<p>So, with this information to hand you may want to use this type of image making to enhance a document or a presentation. Adding images can really help to not only make your work more visually appealing and to create specific images for your content but is proven to help <a href=\"https:\/\/elearningindustry.com\/cone-of-experience-what-really-is\" target=\"_blank\" rel=\"noreferrer noopener\">improve retention of your message and content within education<\/a>.&nbsp;<\/p>\n\n\n\n<p><\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Considerations&nbsp;<\/h2>\n\n\n\n<p>There are ongoing concerns from designers, artists and photographers including legal challenges of AI-generated images and issues related to the use of copyrighted materials to train AI models. Primarily these issues focus on whether there was unauthorized data collection, whether the collection was a copyright infringement and whether the training of AI on copyrighted material is acceptable under a \u2018fair use\u2019 argument.&nbsp;&nbsp;<\/p>\n\n\n\n<p>The AI-generated images that are output are dependent on their models\u2019 training and so this means that you may find images can have various biases, for example, if one model was only trained with images of cities in the northern hemisphere, when asking it to create cities of Africa it would be wildly wrong or create clearly cliched images. If you ask AI to generate an image of a wristwatch for example the time will often be set to 10:10, most commercial images of watches are set to 10:10, so this time becomes statistically overrepresented in the training data.&nbsp;<\/p>\n\n\n\n<p>Trying different tools with different training sets and the same prompts will provide very different results.&nbsp;<\/p>\n\n\n\n<p>Here are two images from one prompt only using two different tools: <strong><em>Create a photorealistic image in the style of Brian Skerry of a large humpback whale with an emphasis on biodiversity, and themes of protecting the planet<\/em><\/strong><\/p>\n\n\n\n<div class=\"wp-block-group is-nowrap is-layout-flex wp-container-core-group-is-layout-ad2f72ca wp-block-group-is-layout-flex\">\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"439\" height=\"439\" src=\"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-content\/uploads\/sites\/534\/2025\/09\/image-6.png\" alt=\"A humpback whale swimming underwater, surrounded by underwater plants. \" class=\"wp-image-186\" srcset=\"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-content\/uploads\/sites\/534\/2025\/09\/image-6.png 439w, https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-content\/uploads\/sites\/534\/2025\/09\/image-6-300x300.png 300w, https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-content\/uploads\/sites\/534\/2025\/09\/image-6-150x150.png 150w\" sizes=\"auto, (max-width: 439px) 100vw, 439px\" \/><\/figure>\n\n\n\n<figure class=\"wp-block-image size-full is-resized\"><img loading=\"lazy\" decoding=\"async\" width=\"474\" height=\"367\" src=\"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-content\/uploads\/sites\/534\/2025\/09\/image-7.png\" alt=\"A humpback whale breaching out of the water with snow-capped mountains in the background.\" class=\"wp-image-187\" style=\"width:405px;height:auto\" srcset=\"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-content\/uploads\/sites\/534\/2025\/09\/image-7.png 474w, https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-content\/uploads\/sites\/534\/2025\/09\/image-7-300x232.png 300w\" sizes=\"auto, (max-width: 474px) 100vw, 474px\" \/><\/figure>\n<\/div>\n\n\n\n<p><em>(Image one Perplexity Labs, Image two Adobe Firefly)<\/em>&nbsp;&nbsp;<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Bite-sized task&nbsp;<\/h2>\n\n\n\n<div class=\"wp-block-group is-vertical is-layout-flex wp-container-core-group-is-layout-8cf370e7 wp-block-group-is-layout-flex\">\n<p><strong>Step 1 &#8211; learn<\/strong>&nbsp;<\/p>\n\n\n\n<p><a href=\"https:\/\/www.linkedin.com\/learning\/openai-api-image-generation-with-dall-e\/craft-effective-image-prompts?resume=false&amp;u=35146660\" data-type=\"link\" data-id=\"https:\/\/www.linkedin.com\/learning\/openai-api-image-generation-with-dall-e\/craft-effective-image-prompts?resume=false&amp;u=35146660\" target=\"_blank\" rel=\"noreferrer noopener\">Watch this two-min chapter on LinkedIn Learning <\/a>to get some tips on how to write good image generation prompts, writing a good prompt is key to gaining as much control over the output as possible.<\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-vertical is-layout-flex wp-container-core-group-is-layout-8cf370e7 wp-block-group-is-layout-flex\">\n<p><strong>Step 2 &#8211; do<\/strong>&nbsp;<\/p>\n\n\n\n<p>Now using Microsoft Co-pilot, find some current learning materials, for example, a presentation, and see where you could replace text with an image to help illustrate your main points. Don\u2019t forget you can ask for improvements and edits with follow-up prompts to Copilot.\u00a0<\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-vertical is-layout-flex wp-container-core-group-is-layout-8cf370e7 wp-block-group-is-layout-flex\">\n<p><strong>Step 3 \u2013 reflect&nbsp;<\/strong><\/p>\n\n\n\n<p>Has the use of images in your presentation enhanced content and presentation? How might you evaluate the effectiveness of these changes to your presentation for your students?&nbsp;<\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-group is-vertical is-layout-flex wp-container-core-group-is-layout-8cf370e7 wp-block-group-is-layout-flex\">\n<div class=\"wp-block-group has-background is-vertical is-layout-flex wp-container-core-group-is-layout-8cf370e7 wp-block-group-is-layout-flex\" style=\"background-color:#caebfa\">\n<h2 class=\"wp-block-heading\">Join the conversation&nbsp;<\/h2>\n\n\n\n<p><a href=\"https:\/\/teams.microsoft.com\/l\/channel\/19%3Aa6bcd01693ee48e5b0402b091da0219d%40thread.tacv2\/GenAI%20Essentials%20-%20Quick%20Tips%20for%20Educators?groupId=fff5f73d-b455-4784-afa1-8fe78cfcde3a&amp;tenantId=4a5378f9-29f4-4d3e-be89-669d03ada9d8\" data-type=\"link\" data-id=\"https:\/\/teams.microsoft.com\/l\/channel\/19%3Aa6bcd01693ee48e5b0402b091da0219d%40thread.tacv2\/GenAI%20Essentials%20-%20Quick%20Tips%20for%20Educators?groupId=fff5f73d-b455-4784-afa1-8fe78cfcde3a&amp;tenantId=4a5378f9-29f4-4d3e-be89-669d03ada9d8\">Post your thoughts on the weekly Teams post to join the conversation.<\/a><\/p>\n<\/div>\n<\/div>\n\n\n\n<h2 class=\"wp-block-heading\">Further links&nbsp;<\/h2>\n\n\n\n<p><a href=\"https:\/\/www.computer.org\/publications\/tech-news\/community-voices\/ethics-of-ai-image-generation\" target=\"_blank\" rel=\"noreferrer noopener\">Ethical Pros and Cons of AI Image Generation &#8211; IEEE computing society<\/a>&nbsp;&nbsp;<\/p>\n\n\n\n<p><a href=\"https:\/\/stablediffusionweb.com\/\" target=\"_blank\" rel=\"noreferrer noopener\">Stable Diffusion<\/a>&nbsp;&nbsp;<\/p>\n\n\n\n<p><a href=\"https:\/\/www.midjourney.com\/home\" target=\"_blank\" rel=\"noreferrer noopener\">Midjourney<\/a>&nbsp;<\/p>\n\n\n\n<p>The LAION 5b dataset <a href=\"https:\/\/laion.ai\/blog\/laion-5b\/\" target=\"_blank\" rel=\"noreferrer noopener\">https:\/\/laion.ai\/blog\/laion-5b\/<\/a>&nbsp;&nbsp;<\/p>\n\n\n\n<p>For a deep dive: <a href=\"https:\/\/simonwillison.net\/\" target=\"_blank\" rel=\"noreferrer noopener\">Simon Willison writes insightful pieces<\/a> on using AI, LLMs with a high level of technical detail.&nbsp;<\/p>\n\n\n\n<div class=\"wp-block-group\"><div class=\"wp-block-group__inner-container is-layout-constrained wp-block-group-is-layout-constrained\">\n<div class=\"wp-block-group has-background\" style=\"background-color:#ffe9a6\"><div class=\"wp-block-group__inner-container is-layout-constrained wp-block-group-is-layout-constrained\">\n<h2 class=\"wp-block-heading\">Contributor biography&nbsp;<\/h2>\n\n\n\n<p>Dr. Adam Procter is a Principal Teaching Fellow of Games and Interaction Design at Winchester School of Art, University of Southampton. He leads the BA (Hons) Games Design &amp; Art programme within the Department of Art and Media Technology. He recently worked with two post graduate researchers from the Creative Computing Institute firstly to review the AI image generation landscape as it emerged and then he created a series of workshops using AI image generation techniques and built and tested some bespoke AI image tools using small, curated data sets with Dr Christina Mamakos and Winchester School of Art, Fine Art Painting Students.&nbsp;&nbsp;<\/p>\n<\/div><\/div>\n<\/div><\/div>\n\n\n\n<p><em>\u00a9 2025. This work is openly licensed via<\/em> <a href=\"https:\/\/creativecommons.org\/licenses\/by-nc-sa\/4.0\/\">CC BY-NC-SA<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>How does GenAI image generation work?\u00a0<\/p>\n","protected":false},"author":6063,"featured_media":184,"comment_status":"closed","ping_status":"closed","sticky":true,"template":"","format":"standard","meta":{"_monsterinsights_skip_tracking":false,"_monsterinsights_sitenote_active":false,"_monsterinsights_sitenote_note":"","_monsterinsights_sitenote_category":0,"footnotes":""},"categories":[1],"tags":[],"class_list":["post-183","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-uncategorised"],"jetpack_featured_media_url":"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-content\/uploads\/sites\/534\/2025\/09\/GettyImages-1483786049-e1758189297650.jpg","_links":{"self":[{"href":"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-json\/wp\/v2\/posts\/183","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-json\/wp\/v2\/users\/6063"}],"replies":[{"embeddable":true,"href":"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-json\/wp\/v2\/comments?post=183"}],"version-history":[{"count":15,"href":"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-json\/wp\/v2\/posts\/183\/revisions"}],"predecessor-version":[{"id":766,"href":"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-json\/wp\/v2\/posts\/183\/revisions\/766"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-json\/wp\/v2\/media\/184"}],"wp:attachment":[{"href":"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-json\/wp\/v2\/media?parent=183"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-json\/wp\/v2\/categories?post=183"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/generic.wordpress.soton.ac.uk\/genai\/wp-json\/wp\/v2\/tags?post=183"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}