{"id":331,"date":"2025-03-12T17:50:19","date_gmt":"2025-03-12T17:50:19","guid":{"rendered":"https:\/\/blogs.dickinson.edu\/academic-technology\/?p=331"},"modified":"2025-03-12T18:09:54","modified_gmt":"2025-03-12T18:09:54","slug":"ai-generated-art","status":"publish","type":"post","link":"https:\/\/blogs.dickinson.edu\/academic-technology\/2025\/03\/12\/ai-generated-art\/","title":{"rendered":"AI Generated Art"},"content":{"rendered":"\n<p>Written by: William Milberry (milberrw@dickinson.edu)<\/p>\n\n\n\n<p class=\"x_MsoNormal\">I\u2019ve been an avid photographer for over 20 years and consider myself a little bit of an artist.&nbsp; I think there is a lot of cross-over between using generative AI and photography or even other forms of visual art.&nbsp; Long ago I read the books of legendary photographer Ansel Adams and his words about pre-visualizing your image and using all the tools and techniques to bring it into reality stuck with me.&nbsp; Today anyone can touch the shutter button on their iPhone and take a clean snapshot of something.&nbsp; The difference between that and a beautiful photograph is that the photographer has an idea and chooses the subject, angle, lenses, settings, and editing to make that image in his mind into a reality.&nbsp; I think AI is very much like this where you can ask it for a picture of something and get a passable result (the proverbial snapshot), or you can work all of the levers and get a beautiful original image.<\/p>\n\n\n\n<p class=\"x_MsoNormal\">I\u2019ve never studied the prompting techniques used by graphic professionals who use AI in their work.&nbsp; They probably know a lot more than I&#8217;ve discovered.&nbsp; But, here is what I work by\u2026<\/p>\n\n\n\n<div class=\"wp-block-media-text is-stacked-on-mobile\"><figure class=\"wp-block-media-text__media\"><a href=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image001.jpg\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"1024\" src=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image001.jpg\" alt=\"Fireman\" class=\"wp-image-332 size-full\" srcset=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image001.jpg 1024w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image001-300x300.jpg 300w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image001-150x150.jpg 150w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image001-768x768.jpg 768w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/a><\/figure><div class=\"wp-block-media-text__content\">\n<p>If you ask an AI like <a title=\"https:\/\/nam12.safelinks.protection.outlook.com\/?url=https%3A%2F%2Fm365.cloud.microsoft%2Fchat%3F&amp;data=05%7C02%7Cburker%40dickinson.edu%7C020ff1d41c7546ff999708dd61864f3b%7C6232b05576b94c139b88b562ae7db6fb%7C0%7C0%7C638773951785576027%7CUnknown%7CTWFpbGZsb3d8eyJFbXB0eU1hcGkiOnRydWUsIlYiOiIwLjAuMDAwMCIsIlAiOiJXaW4zMiIsIkFOIjoiTWFpbCIsIldUIjoyfQ%3D%3D%7C0%7C%7C%7C&amp;sdata=U0s1F8%2Fv0IZ2PIw2sM7I9J50P1we0fnrD3NTePcgw4k%3D&amp;reserved=0\" data-outlook-id=\"c744f425-5903-4e30-9fa9-c801cda47543\" href=\"https:\/\/nam12.safelinks.protection.outlook.com\/?url=https%3A%2F%2Fm365.cloud.microsoft%2Fchat%3F&amp;data=05%7C02%7Cburker%40dickinson.edu%7C020ff1d41c7546ff999708dd61864f3b%7C6232b05576b94c139b88b562ae7db6fb%7C0%7C0%7C638773951785576027%7CUnknown%7CTWFpbGZsb3d8eyJFbXB0eU1hcGkiOnRydWUsIlYiOiIwLjAuMDAwMCIsIlAiOiJXaW4zMiIsIkFOIjoiTWFpbCIsIldUIjoyfQ%3D%3D%7C0%7C%7C%7C&amp;sdata=U0s1F8%2Fv0IZ2PIw2sM7I9J50P1we0fnrD3NTePcgw4k%3D&amp;reserved=0\">Microsoft Copilot Chat<\/a> something like \u201c<b><i>generate a photo of a fireman<\/i><\/b>\u201d It will give you a representative image that resembles the typical photos of firemen that it\u2019s been trained on.&nbsp; I consider this the equivalent of an iPhone snapshot (see image below).<\/p>\n<\/div><\/div>\n\n\n\n<div class=\"wp-block-media-text has-media-on-the-right is-stacked-on-mobile\"><div class=\"wp-block-media-text__content\">\n<p>To get something better in my opinion, you have to feed it detail about the subject and how you wanted it presented. &nbsp;A prompt like \u201c<b><i>Generate a three quarters body shot of a fireman, looking at the camera, and standing in front of a burning building. He should have sweat on his face, look dirty and have a determined look in his eyes.<\/i><\/b>\u201d<span class=\"x_apple-converted-space\"> will start to be less generic and more relatable as it starts to reflect the creator\u2019s imagination.<\/span><\/p>\n<\/div><figure class=\"wp-block-media-text__media\"><a href=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image002.jpg\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"1024\" src=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image002.jpg\" alt=\"Fireman 2\" class=\"wp-image-335 size-full\" srcset=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image002.jpg 1024w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image002-300x300.jpg 300w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image002-150x150.jpg 150w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image002-768x768.jpg 768w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/a><\/figure><\/div>\n\n\n\n<p class=\"x_MsoNormal\">Beyond the content of the image, AI understands a lot of prompts about presentation.&nbsp; To make better images you can take control and specify how you want it presented with prompts like&nbsp;<em>wide-angle, portrait, cowboy shot, full-body shot<\/em>, etc.&nbsp; You can take it a step further with prompts about lighting such as&nbsp;<em>diffuse light, direct sunlight, backlighting<\/em>, etc.&nbsp; On top of that you can also specify if you want it to have the look of a particular film stock, medium, era, or even well known artist\/director\u2019s style.&nbsp; Prompts like&nbsp;<em>1960\u2019s cinema, 1980\u2019s TV, Kodachrome film, anime, in the style of Hayao Miyazaki&nbsp;<\/em>can all flavor and stylize an image.<\/p>\n\n\n\n<p>Not every AI is equally good at these details.&nbsp; I\u2019ve seen some AMAZING things from Midjourney.&nbsp; I haven\u2019t used Microsoft too much, but it seems to be tuned to give somewhat generic images.&nbsp; I use Stable Diffusion locally and it can give some great results, but it needs some technical understanding to use.&nbsp; The key point in this is that you don\u2019t just want to describe the scene or subject but also describe the presentation of it.<\/p>\n\n\n\n<p>Most generative AIs accept natural language like my simple examples above, but they also have syntax and parameters that can be used such as weights attached to different parts of the prompt to affect how much or how little the AI incorporates that particular detail.<\/p>\n\n\n\n<p class=\"x_MsoNormal\">Below is a Stable Diffusion XL prompt I used (in conjunction with a model tuned for illustration &amp; animation.)&nbsp;&nbsp;I did this when I was still new to it and copied details from a lot of different example prompts that I liked.&nbsp;&nbsp;Because the prompt is quite cluttered, you can see how Stable Diffusion incorporated some of what I asked for but not all of it.&nbsp; The numbers after words in the prompt are weights.&nbsp;&nbsp;<strong><em>(silhouette:1.25)<\/em><\/strong>&nbsp;means silhouette should be given 125% weight because it was a very important aspect of what I wanted to create.<\/p>\n\n\n\n<div class=\"wp-block-group has-tertiary-background-color has-background is-vertical is-layout-flex wp-container-core-group-is-layout-274f5b33 wp-block-group-is-layout-flex\" style=\"padding-top:var(--wp--preset--spacing--30);padding-right:var(--wp--preset--spacing--30);padding-bottom:var(--wp--preset--spacing--30);padding-left:var(--wp--preset--spacing--30)\">\n<figure data-wp-context=\"{&quot;imageId&quot;:&quot;69e7ee6556900&quot;}\" data-wp-interactive=\"core\/image\" data-wp-key=\"69e7ee6556900\" class=\"wp-block-image size-large wp-lightbox-container\"><img loading=\"lazy\" decoding=\"async\" width=\"768\" height=\"1024\" data-wp-class--hide=\"state.isContentHidden\" data-wp-class--show=\"state.isContentVisible\" data-wp-init=\"callbacks.setButtonStyles\" data-wp-on--click=\"actions.showLightbox\" data-wp-on--load=\"callbacks.setButtonStyles\" data-wp-on-window--resize=\"callbacks.setButtonStyles\" src=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image003-768x1024.png\" alt=\"\" class=\"wp-image-336\" srcset=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image003-768x1024.png 768w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image003-225x300.png 225w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image003-1152x1536.png 1152w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image003.png 1200w\" sizes=\"auto, (max-width: 768px) 100vw, 768px\" \/><button\n\t\t\tclass=\"lightbox-trigger\"\n\t\t\ttype=\"button\"\n\t\t\taria-haspopup=\"dialog\"\n\t\t\taria-label=\"Enlarge\"\n\t\t\tdata-wp-init=\"callbacks.initTriggerButton\"\n\t\t\tdata-wp-on--click=\"actions.showLightbox\"\n\t\t\tdata-wp-style--right=\"state.imageButtonRight\"\n\t\t\tdata-wp-style--top=\"state.imageButtonTop\"\n\t\t>\n\t\t\t<svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" width=\"12\" height=\"12\" fill=\"none\" viewBox=\"0 0 12 12\">\n\t\t\t\t<path fill=\"#fff\" d=\"M2 0a2 2 0 0 0-2 2v2h1.5V2a.5.5 0 0 1 .5-.5h2V0H2Zm2 10.5H2a.5.5 0 0 1-.5-.5V8H0v2a2 2 0 0 0 2 2h2v-1.5ZM8 12v-1.5h2a.5.5 0 0 0 .5-.5V8H12v2a2 2 0 0 1-2 2H8Zm2-12a2 2 0 0 1 2 2v2h-1.5V2a.5.5 0 0 0-.5-.5H8V0h2Z\" \/>\n\t\t\t<\/svg>\n\t\t<\/button><\/figure>\n\n\n\n<p><strong><em>ultra wide<\/em><\/strong><strong><em>&nbsp;<\/em><\/strong><strong><em>angle, (silhouette:1.25),1girl, 15 year old girl, in skirt, (short feathered hair), wearing boots and black backpack, standing on mountain overlooking a distant giant and fat Buddha statue sitting in lotus position on a hill facing the girl, hazy distance, dark background with two merging galaxies in the sky, blacklight, wide shot, full body, somber expression, looking up, dark energy, vibrant magenta, portal to another world, flat color, flat shading,&nbsp;ultra realistic,highres,superb,8k wallpaper, extremely detailed, intricate, limited palette, pink, yellow<\/em><\/strong><\/p>\n\n\n\n<p><strong><em>Negative prompt:&nbsp;easynegative,(badv2:0.8),(badhandv4:1.18),(bad quality:1.4),(low quality, worst quality:1.14),watermark,(blurry),(cropped),(nsfw:1.18),(animal)<\/em><\/strong><\/p>\n\n\n\n<p><strong><em>Steps: 30, Sampler: Euler, CFG scale: 7, Seed: 804295955, Size: 768&#215;1024, Model hash: 65298be5b1, Model: realcartoonXL_v5, Denoising strength: 0.45, Clip skip: 2, Hires upscale: 2, Hires steps: 20, Hires&nbsp;upscaler: R-ESRGAN 4x+, Version: v1.6.1<\/em><\/strong><\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-group has-tertiary-background-color has-background is-vertical is-layout-flex wp-container-core-group-is-layout-274f5b33 wp-block-group-is-layout-flex\" style=\"padding-top:var(--wp--preset--spacing--30);padding-right:var(--wp--preset--spacing--30);padding-bottom:var(--wp--preset--spacing--30);padding-left:var(--wp--preset--spacing--30)\">\n<p><strong><em>If I trim off some of the Stable Diffusion specific stuff and ask Microsoft \u201cgenerate an image using this Stable Diffusion prompt\u2026\u201d, it will give me something similar.<\/em><\/strong><\/p>\n\n\n\n<figure data-wp-context=\"{&quot;imageId&quot;:&quot;69e7ee65572f5&quot;}\" data-wp-interactive=\"core\/image\" data-wp-key=\"69e7ee65572f5\" class=\"wp-block-image size-full wp-lightbox-container\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"1024\" data-wp-class--hide=\"state.isContentHidden\" data-wp-class--show=\"state.isContentVisible\" data-wp-init=\"callbacks.setButtonStyles\" data-wp-on--click=\"actions.showLightbox\" data-wp-on--load=\"callbacks.setButtonStyles\" data-wp-on-window--resize=\"callbacks.setButtonStyles\" src=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image004.png\" alt=\"\" class=\"wp-image-337\" srcset=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image004.png 1024w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image004-300x300.png 300w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image004-150x150.png 150w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image004-768x768.png 768w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><button\n\t\t\tclass=\"lightbox-trigger\"\n\t\t\ttype=\"button\"\n\t\t\taria-haspopup=\"dialog\"\n\t\t\taria-label=\"Enlarge\"\n\t\t\tdata-wp-init=\"callbacks.initTriggerButton\"\n\t\t\tdata-wp-on--click=\"actions.showLightbox\"\n\t\t\tdata-wp-style--right=\"state.imageButtonRight\"\n\t\t\tdata-wp-style--top=\"state.imageButtonTop\"\n\t\t>\n\t\t\t<svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" width=\"12\" height=\"12\" fill=\"none\" viewBox=\"0 0 12 12\">\n\t\t\t\t<path fill=\"#fff\" d=\"M2 0a2 2 0 0 0-2 2v2h1.5V2a.5.5 0 0 1 .5-.5h2V0H2Zm2 10.5H2a.5.5 0 0 1-.5-.5V8H0v2a2 2 0 0 0 2 2h2v-1.5ZM8 12v-1.5h2a.5.5 0 0 0 .5-.5V8H12v2a2 2 0 0 1-2 2H8Zm2-12a2 2 0 0 1 2 2v2h-1.5V2a.5.5 0 0 0-.5-.5H8V0h2Z\" \/>\n\t\t\t<\/svg>\n\t\t<\/button><\/figure>\n<\/div>\n\n\n\n<p>Prompts are only one part of achieving stunning results.&nbsp; Midjourney has different models tuned to different things like animation (the \u201cNiji\u201d model) and photo-realism and&nbsp;<a href=\"https:\/\/nam12.safelinks.protection.outlook.com\/?url=https%3A%2F%2Fwww.tutorialspoint.com%2Fmidjourney%2Fmidjourney-model-versions.htm&amp;data=05%7C02%7Cburker%40dickinson.edu%7C020ff1d41c7546ff999708dd61864f3b%7C6232b05576b94c139b88b562ae7db6fb%7C0%7C0%7C638773951785590720%7CUnknown%7CTWFpbGZsb3d8eyJFbXB0eU1hcGkiOnRydWUsIlYiOiIwLjAuMDAwMCIsIlAiOiJXaW4zMiIsIkFOIjoiTWFpbCIsIldUIjoyfQ%3D%3D%7C0%7C%7C%7C&amp;sdata=%2BKsIomOdbxFF%2Bx3E%2F0XA2uEc%2Bd5vUZygo3P4P71%2FGGM%3D&amp;reserved=0\">you can specify in your prompt which model it uses<\/a>.&nbsp; Likewise, there is a large community who custom train models that you can download for Stable Diffusion allowing you to do really high-quality work with a specific subjects.&nbsp; There is a popular animated series out of France called Miraculous featuring a female superhero named Ladybug.&nbsp; Using a model tuned for general illustration\/animation and an add-on (called a LORA) specifically trained on images from the show, I made a really nice, semi-realistic portrait of the main character Marinette (left is my portrait, right is a screenshot from the show for reference).&nbsp; This type of result can\u2019t really be achieved by prompting alone.&nbsp; You either need customized models OR to use an Image-to-image feature where you give an AI an image and a prompt and it mimics\/works from the source image.<br><\/p>\n\n\n\n<div class=\"wp-block-cover alignfull\" style=\"padding-top:var(--wp--preset--spacing--30);padding-right:var(--wp--preset--spacing--30);padding-bottom:var(--wp--preset--spacing--30);padding-left:var(--wp--preset--spacing--30);min-height:66vh;aspect-ratio:unset;\"><span aria-hidden=\"true\" class=\"wp-block-cover__background has-black-background-color has-background-dim-100 has-background-dim\"><\/span><div class=\"wp-block-cover__inner-container is-layout-flow wp-block-cover-is-layout-flow\">\n<div class=\"wp-block-group is-content-justification-center is-nowrap is-layout-flex wp-container-core-group-is-layout-3eb4be21 wp-block-group-is-layout-flex\">\n<figure data-wp-context=\"{&quot;imageId&quot;:&quot;69e7ee65576ca&quot;}\" data-wp-interactive=\"core\/image\" data-wp-key=\"69e7ee65576ca\" class=\"wp-block-image size-full is-resized wp-lightbox-container\"><img loading=\"lazy\" decoding=\"async\" width=\"640\" height=\"768\" data-wp-class--hide=\"state.isContentHidden\" data-wp-class--show=\"state.isContentVisible\" data-wp-init=\"callbacks.setButtonStyles\" data-wp-on--click=\"actions.showLightbox\" data-wp-on--load=\"callbacks.setButtonStyles\" data-wp-on-window--resize=\"callbacks.setButtonStyles\" src=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image005.png\" alt=\"\" class=\"wp-image-338\" style=\"width:447px;height:auto\" srcset=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image005.png 640w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image005-250x300.png 250w\" sizes=\"auto, (max-width: 640px) 100vw, 640px\" \/><button\n\t\t\tclass=\"lightbox-trigger\"\n\t\t\ttype=\"button\"\n\t\t\taria-haspopup=\"dialog\"\n\t\t\taria-label=\"Enlarge\"\n\t\t\tdata-wp-init=\"callbacks.initTriggerButton\"\n\t\t\tdata-wp-on--click=\"actions.showLightbox\"\n\t\t\tdata-wp-style--right=\"state.imageButtonRight\"\n\t\t\tdata-wp-style--top=\"state.imageButtonTop\"\n\t\t>\n\t\t\t<svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" width=\"12\" height=\"12\" fill=\"none\" viewBox=\"0 0 12 12\">\n\t\t\t\t<path fill=\"#fff\" d=\"M2 0a2 2 0 0 0-2 2v2h1.5V2a.5.5 0 0 1 .5-.5h2V0H2Zm2 10.5H2a.5.5 0 0 1-.5-.5V8H0v2a2 2 0 0 0 2 2h2v-1.5ZM8 12v-1.5h2a.5.5 0 0 0 .5-.5V8H12v2a2 2 0 0 1-2 2H8Zm2-12a2 2 0 0 1 2 2v2h-1.5V2a.5.5 0 0 0-.5-.5H8V0h2Z\" \/>\n\t\t\t<\/svg>\n\t\t<\/button><\/figure>\n\n\n\n<figure data-wp-context=\"{&quot;imageId&quot;:&quot;69e7ee6557992&quot;}\" data-wp-interactive=\"core\/image\" data-wp-key=\"69e7ee6557992\" class=\"wp-block-image size-large wp-lightbox-container\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"576\" data-wp-class--hide=\"state.isContentHidden\" data-wp-class--show=\"state.isContentVisible\" data-wp-init=\"callbacks.setButtonStyles\" data-wp-on--click=\"actions.showLightbox\" data-wp-on--load=\"callbacks.setButtonStyles\" data-wp-on-window--resize=\"callbacks.setButtonStyles\" src=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image006-1024x576.png\" alt=\"\" class=\"wp-image-339\" srcset=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image006-1024x576.png 1024w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image006-300x169.png 300w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image006-768x432.png 768w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image006.png 1280w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><button\n\t\t\tclass=\"lightbox-trigger\"\n\t\t\ttype=\"button\"\n\t\t\taria-haspopup=\"dialog\"\n\t\t\taria-label=\"Enlarge\"\n\t\t\tdata-wp-init=\"callbacks.initTriggerButton\"\n\t\t\tdata-wp-on--click=\"actions.showLightbox\"\n\t\t\tdata-wp-style--right=\"state.imageButtonRight\"\n\t\t\tdata-wp-style--top=\"state.imageButtonTop\"\n\t\t>\n\t\t\t<svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" width=\"12\" height=\"12\" fill=\"none\" viewBox=\"0 0 12 12\">\n\t\t\t\t<path fill=\"#fff\" d=\"M2 0a2 2 0 0 0-2 2v2h1.5V2a.5.5 0 0 1 .5-.5h2V0H2Zm2 10.5H2a.5.5 0 0 1-.5-.5V8H0v2a2 2 0 0 0 2 2h2v-1.5ZM8 12v-1.5h2a.5.5 0 0 0 .5-.5V8H12v2a2 2 0 0 1-2 2H8Zm2-12a2 2 0 0 1 2 2v2h-1.5V2a.5.5 0 0 0-.5-.5H8V0h2Z\" \/>\n\t\t\t<\/svg>\n\t\t<\/button><\/figure>\n<\/div>\n<\/div><\/div>\n\n\n\n<p>Beyond selecting the model, there is a host of settings that dramatically affect the results.&nbsp; Midjourney allows you access to these through parameters you include in the prompt.&nbsp; I use a GUI called Automatic1111 for Stable Diffusion (screenshot below) that gives me access to everything.&nbsp; I\u2019m not certain what level of control Microsoft, Adobe, Grok, and Google Gemini offer (if any).<\/p>\n\n\n\n<p>An AI\u2019s neural network is like taking a road trip from Carlisle to Los Angeles.&nbsp; If you take different turns or different streets on such a long road-trip then what you will see on that trip will vary dramatically.&nbsp; Changing one parameter or even one word in your prompt will cause a different traversal of the AI model\u2019s data resulting in shifts in the image or even a very different image (just like your road trip would be different if you took a different route.)&nbsp; Unlike with graphic software or 3D animation where you can fix a single element, changing a tiny thing in an AI image you otherwise like is very difficult.&nbsp; For that reason, creating AI images for me is an iterative process where I produce dozens of images with prompt and setting variations until I get extremely close to what I want.&nbsp; I then usually polish the little details in Photoshop.<\/p>\n\n\n\n<p>Below is a screenshot of the Automatic1111 GUI for Stable Diffusion.&nbsp; A more capable and complex UI called ComfyUI has become popular, but I still use this older one.<\/p>\n\n\n\n<figure data-wp-context=\"{&quot;imageId&quot;:&quot;69e7ee6557de8&quot;}\" data-wp-interactive=\"core\/image\" data-wp-key=\"69e7ee6557de8\" class=\"wp-block-image size-large is-resized wp-lightbox-container\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"611\" data-wp-class--hide=\"state.isContentHidden\" data-wp-class--show=\"state.isContentVisible\" data-wp-init=\"callbacks.setButtonStyles\" data-wp-on--click=\"actions.showLightbox\" data-wp-on--load=\"callbacks.setButtonStyles\" data-wp-on-window--resize=\"callbacks.setButtonStyles\" src=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image007-1024x611.png\" alt=\"\" class=\"wp-image-341\" style=\"aspect-ratio:1.5;object-fit:contain;width:650px;height:auto\" srcset=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image007-1024x611.png 1024w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image007-300x179.png 300w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image007-768x458.png 768w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image007-1536x916.png 1536w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image007-2048x1221.png 2048w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><button\n\t\t\tclass=\"lightbox-trigger\"\n\t\t\ttype=\"button\"\n\t\t\taria-haspopup=\"dialog\"\n\t\t\taria-label=\"Enlarge\"\n\t\t\tdata-wp-init=\"callbacks.initTriggerButton\"\n\t\t\tdata-wp-on--click=\"actions.showLightbox\"\n\t\t\tdata-wp-style--right=\"state.imageButtonRight\"\n\t\t\tdata-wp-style--top=\"state.imageButtonTop\"\n\t\t>\n\t\t\t<svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" width=\"12\" height=\"12\" fill=\"none\" viewBox=\"0 0 12 12\">\n\t\t\t\t<path fill=\"#fff\" d=\"M2 0a2 2 0 0 0-2 2v2h1.5V2a.5.5 0 0 1 .5-.5h2V0H2Zm2 10.5H2a.5.5 0 0 1-.5-.5V8H0v2a2 2 0 0 0 2 2h2v-1.5ZM8 12v-1.5h2a.5.5 0 0 0 .5-.5V8H12v2a2 2 0 0 1-2 2H8Zm2-12a2 2 0 0 1 2 2v2h-1.5V2a.5.5 0 0 0-.5-.5H8V0h2Z\" \/>\n\t\t\t<\/svg>\n\t\t<\/button><\/figure>\n\n\n\n<p>If you read the prompt below you might wonder about some parts of it like \u201c11 year old\u201d and \u201csmall breasts\u201d.&nbsp; As I was working on this image, I wanted a younger, somewhat anime-style character but I kept getting a more adult looking character with a more mature figure.&nbsp; So, I started adding terms to counter that until a balance was struck and the character resembled what I was going for.&nbsp; Also, since my Jellymaid characters are anthropomorphic jellyfish, I want them to look natural like they live in the ocean.&nbsp; So, when recognizable clothes showed up on them it didn\u2019t look right.&nbsp; That\u2019s when I put \u201cclothes, bikini, shoes\u201d into the negative prompt (the text box below the prompt), to stop those items from showing up.<\/p>\n\n\n\n<div class=\"wp-block-cover\" style=\"margin-top:var(--wp--preset--spacing--30);margin-bottom:var(--wp--preset--spacing--30);padding-top:0;padding-bottom:0;min-height:816px;aspect-ratio:unset;\"><span aria-hidden=\"true\" class=\"wp-block-cover__background has-base-background-color has-background-dim-0 has-background-dim\"><\/span><img loading=\"lazy\" decoding=\"async\" width=\"1440\" height=\"1808\" class=\"wp-block-cover__image-background wp-image-342\" alt=\"\" src=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image008.png\" data-object-fit=\"cover\" srcset=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image008.png 1440w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image008-239x300.png 239w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image008-816x1024.png 816w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image008-768x964.png 768w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image008-1223x1536.png 1223w\" sizes=\"auto, (max-width: 1440px) 100vw, 1440px\" \/><div class=\"wp-block-cover__inner-container is-layout-flow wp-block-cover-is-layout-flow\">\n<p class=\"has-text-align-center has-large-font-size\" style=\"padding-top:600px;padding-bottom:0\">I actually used the AI generative fill in Photoshop to expand the borders of this image<\/p>\n<\/div><\/div>\n\n\n\n<p>If I had to try to distill what I\u2019ve learned down into some bullet-point tips I would say:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Specify details of the subject:&nbsp;&nbsp;<\/strong>Don\u2019t just say a subject like \u201cfireman\u201d, \u201csports car\u201d, or \u201ccity road\u201d, but include a judicious description of any details you want it to have like expression, position, features, colors, age, etc.<br>* Note \u2013 many AIs are NOT yet good at applying different details to different people or objects within the same image, so describing two different people will have mixed results<\/li>\n\n\n\n<li><strong>Specify presentation:<\/strong>&nbsp; Specify how you want the image to look ie.) wide-angle, portrait, close-up, film look, strong contrast, faded colors, 1980\u2019s look, blurred background, diffuse lighting, a particular artist\u2019s style, etc., etc.<\/li>\n\n\n\n<li><strong>If an AI allows negative prompts, make use of it: &nbsp;<\/strong>The negative prompt is what you don\u2019t want in the image, and you should make use of it.&nbsp; If you are getting funky hands you can add \u201cextra fingers\u201d, \u201cdeformed hands\u201d, etc. to the negative prompt to help give the hands special attention.&nbsp; Once I described a character with \u201cfeathered hair\u201d and wound up with feathers in the image, by adding \u201cfeathers, wings\u201d to the negative prompt I was able to use \u201cfeathered hair\u201d without actual feathers appearing in the image.<\/li>\n\n\n\n<li><strong>Use weights if supported<\/strong>.&nbsp; For example, in the Buddha image above \u201c<strong><em>(silhouette:1.25)\u201d<\/em><\/strong>&nbsp;specifies 125% importance that it be a silhouette because that aspect more important than other details in the image.<\/li>\n\n\n\n<li><strong>Use custom models if available:&nbsp;&nbsp;<\/strong>If you are using Midjourney or another AI where you can select from different models, don\u2019t use the generic base model but choose the one most tuned to what you are making (such as photo-realistic or animated.)<\/li>\n\n\n\n<li><strong>Iterate!<\/strong>&nbsp;\u2013 Small changes to prompts or parameters can cause the image to shift and change or even produce vastly different images.&nbsp; Try many variations until you hone in on what you want.<\/li>\n\n\n\n<li><strong>AI safety &amp; censorship:&nbsp;&nbsp;<\/strong>Most online and commercial generative AI has strong safety and censorship measures in place.&nbsp; They will not accept prompts containing violence, sexuality, drugs, other harmful topics, and sometimes certain copyrighted materials.&nbsp; Some also check their output before giving it to you to ensure nothing that goes against their terms was generated and err heavily on the side of caution.&nbsp; If you get a prompt rejected or a false alarm, consider what might have triggered it in the prompt and try again with slightly different wording.&nbsp; Adobe is the worst when it comes to this.<\/li>\n<\/ul>\n\n\n\n<figure data-wp-context=\"{&quot;imageId&quot;:&quot;69e7ee655831b&quot;}\" data-wp-interactive=\"core\/image\" data-wp-key=\"69e7ee655831b\" class=\"wp-block-image size-full wp-lightbox-container\"><img loading=\"lazy\" decoding=\"async\" width=\"1017\" height=\"521\" data-wp-class--hide=\"state.isContentHidden\" data-wp-class--show=\"state.isContentVisible\" data-wp-init=\"callbacks.setButtonStyles\" data-wp-on--click=\"actions.showLightbox\" data-wp-on--load=\"callbacks.setButtonStyles\" data-wp-on-window--resize=\"callbacks.setButtonStyles\" src=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image009.jpg\" alt=\"\" class=\"wp-image-343\" srcset=\"https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image009.jpg 1017w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image009-300x154.jpg 300w, https:\/\/blogs.dickinson.edu\/academic-technology\/files\/2025\/03\/image009-768x393.jpg 768w\" sizes=\"auto, (max-width: 1017px) 100vw, 1017px\" \/><button\n\t\t\tclass=\"lightbox-trigger\"\n\t\t\ttype=\"button\"\n\t\t\taria-haspopup=\"dialog\"\n\t\t\taria-label=\"Enlarge\"\n\t\t\tdata-wp-init=\"callbacks.initTriggerButton\"\n\t\t\tdata-wp-on--click=\"actions.showLightbox\"\n\t\t\tdata-wp-style--right=\"state.imageButtonRight\"\n\t\t\tdata-wp-style--top=\"state.imageButtonTop\"\n\t\t>\n\t\t\t<svg xmlns=\"http:\/\/www.w3.org\/2000\/svg\" width=\"12\" height=\"12\" fill=\"none\" viewBox=\"0 0 12 12\">\n\t\t\t\t<path fill=\"#fff\" d=\"M2 0a2 2 0 0 0-2 2v2h1.5V2a.5.5 0 0 1 .5-.5h2V0H2Zm2 10.5H2a.5.5 0 0 1-.5-.5V8H0v2a2 2 0 0 0 2 2h2v-1.5ZM8 12v-1.5h2a.5.5 0 0 0 .5-.5V8H12v2a2 2 0 0 1-2 2H8Zm2-12a2 2 0 0 1 2 2v2h-1.5V2a.5.5 0 0 0-.5-.5H8V0h2Z\" \/>\n\t\t\t<\/svg>\n\t\t<\/button><\/figure>\n\n\n\n<p>This is what I know and the conclusions I\u2019ve come to over the past year.&nbsp; I hope it\u2019s useful.&nbsp; If you have any questions or would ever like to get together and talk about this topic more, just let me know!<\/p>\n\n\n\n<p>~ William Milberry (milberrw@dickinson.edu)<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Written by: William Milberry (milberrw@dickinson.edu) I\u2019ve been an avid photographer for over 20 years and consider myself a little bit of an artist.&nbsp; I think there is a lot of cross-over between using generative AI and photography or even other forms of visual art.&nbsp; Long ago I read the books of legendary photographer Ansel Adams [&hellip;]<\/p>\n","protected":false},"author":126,"featured_media":336,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1],"tags":[],"class_list":["post-331","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-uncategorized"],"_links":{"self":[{"href":"https:\/\/blogs.dickinson.edu\/academic-technology\/wp-json\/wp\/v2\/posts\/331","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/blogs.dickinson.edu\/academic-technology\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/blogs.dickinson.edu\/academic-technology\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/blogs.dickinson.edu\/academic-technology\/wp-json\/wp\/v2\/users\/126"}],"replies":[{"embeddable":true,"href":"https:\/\/blogs.dickinson.edu\/academic-technology\/wp-json\/wp\/v2\/comments?post=331"}],"version-history":[{"count":0,"href":"https:\/\/blogs.dickinson.edu\/academic-technology\/wp-json\/wp\/v2\/posts\/331\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/blogs.dickinson.edu\/academic-technology\/wp-json\/wp\/v2\/media\/336"}],"wp:attachment":[{"href":"https:\/\/blogs.dickinson.edu\/academic-technology\/wp-json\/wp\/v2\/media?parent=331"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/blogs.dickinson.edu\/academic-technology\/wp-json\/wp\/v2\/categories?post=331"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/blogs.dickinson.edu\/academic-technology\/wp-json\/wp\/v2\/tags?post=331"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}