{"id":3895,"date":"2022-11-29T17:12:52","date_gmt":"2022-11-29T17:12:52","guid":{"rendered":"https:\/\/itp.nyu.edu\/lowres\/critex-monika\/?p=3895"},"modified":"2022-11-29T17:14:42","modified_gmt":"2022-11-29T17:14:42","slug":"topic-2-final-reflection-3","status":"publish","type":"post","link":"https:\/\/itp.nyu.edu\/lowres\/critex-monika\/2022\/11\/29\/topic-2-final-reflection-3\/","title":{"rendered":"Topic 2 Final Reflection"},"content":{"rendered":"<ul>\n<li style=\"font-weight: 300\"><span style=\"font-weight: 300\">What did you learn?<br \/>\n<\/span>Disco Diffusion, Midjourney and Stable Diffusion are three pieces of open source software that generate images in response to a language prompt. A prompt, which is a series of words, is a key to tuning the tools to translate your thoughts into images.<\/p>\n<p>Here at the gallery, the three installations with user interfaces using local GPUs aim to make the creation process more accessible and more efficient. We encourage you to explore these three different tools and to learn to talk to these text-to-image models, we first start with \u201cprompt engineering\u201d. A helpful analogy to understand this process is to think of a prompt as a search query just like how we use the Google search engine. Similarly, we give these image generators a search query to search among a structured representation of all the images it was trained on. We then evaluate the result, and refine the input text until the output image is the closest to our expectations among all possible outputs.<\/p>\n<p>For example, if you use \u201c\/imagine prompt: colorful butterfly\u201d, the AI will generate an image of what it thinks a colorful butterfly is. However, using one adjective like \u201ccolorful\u201d is usually not sufficient. If you use one adjective, the AI will usually generate something simple. Hence, it\u2019s recommended that you use multiple adjectives and descriptions for the AI to generate an accurate image. Instead of \u201ccolorful butterfly\u201d, you can use \u201ca rainbow-colored butterfly flying across a field of flowers during a sunset\u201d. The more descriptive you are, the better the results.<\/p>\n<p>Also, the copyright of who owns the AI-Generated art still need to be discussed, So once you\u2019ve created your AI-generated masterpiece, what\u2019s stopping you from claiming it as your own, using it commercially or preventing others from using it? On top of existentially threatening the very concept of artists and creatives, Ryan says that AI-generated content raises many new legal issues.<\/li>\n<li style=\"font-weight: 300\"><span style=\"font-weight: 300\">What feedback did you receive? Any reflections on critique itself?<br \/>\nThe topic doesn&#8217;t have a certain answer, people would love to see the debate between humans and AI. I also received feedback on my website, which can provide a better user experience and make the copyright problem prominent.\u00a0<\/span><\/li>\n<li style=\"font-weight: 300\"><span style=\"font-weight: 300\">What might you do differently in terms of process or content?<br \/>\nI might create an interactive artwork using different kinds of digital software and AI-Generated art platforms instead of creating a website.<\/span><\/li>\n<li style=\"font-weight: 300\"><span style=\"font-weight: 300\">What was inspiring? What parts?<br \/>\nTrying to use different creative tools is inspiring, I can experience the differences between these popular AI-Generated Art platforms such as Disco Diffusion, Midjourney and Stable Diffusion, and also I have learned a lot about the training data, data biases, terms of use, policy, rules, and even The US Copyright Office.<\/span><\/li>\n<li style=\"font-weight: 300\"><span style=\"font-weight: 300\"><span style=\"font-weight: 300\">Revisit the assignment prompts: how did your project relate to the original prompts, in terms of critical lens, audience, tone, etc\u2026<br \/>\nBecause my audiences are those <\/span><\/span>people who haven\u2019t used AI-generated art before, which may have some contradiction that if someone hasn&#8217;t used it so far, they probably don&#8217;t have an interest to use these kinds of tools, so my project may not drive people&#8217;s awareness to fulfill my goal.<\/li>\n<li style=\"font-weight: 300\"><span style=\"font-weight: 300\">How did you balance research and experimentation? Which is easier for you? How can you focus more on the areas that you shy away from<br \/>\nI do like the daily practice part, which forced me to dig deeper every day, and also because of that I can narrow down my topic and find my critical lens.\u00a0<\/span><\/li>\n<\/ul>\n","protected":false},"excerpt":{"rendered":"<p>What did you learn? Disco Diffusion, Midjourney and Stable Diffusion are three pieces of open source software that generate images in response to a language prompt. A prompt, which is a series of words, is a key to tuning the tools to translate your thoughts into images. Here at the gallery, the three installations with&hellip;&nbsp;<a href=\"https:\/\/itp.nyu.edu\/lowres\/critex-monika\/2022\/11\/29\/topic-2-final-reflection-3\/\" rel=\"bookmark\">Read More &raquo;<span class=\"screen-reader-text\">Topic 2 Final Reflection<\/span><\/a><\/p>\n","protected":false},"author":149,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"neve_meta_sidebar":"","neve_meta_container":"","neve_meta_enable_content_width":"off","neve_meta_content_width":70,"neve_meta_title_alignment":"","neve_meta_author_avatar":"","neve_post_elements_order":"","neve_meta_disable_header":"","neve_meta_disable_footer":"","neve_meta_disable_title":"","footnotes":""},"categories":[15],"tags":[],"_links":{"self":[{"href":"https:\/\/itp.nyu.edu\/lowres\/critex-monika\/wp-json\/wp\/v2\/posts\/3895"}],"collection":[{"href":"https:\/\/itp.nyu.edu\/lowres\/critex-monika\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/itp.nyu.edu\/lowres\/critex-monika\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/itp.nyu.edu\/lowres\/critex-monika\/wp-json\/wp\/v2\/users\/149"}],"replies":[{"embeddable":true,"href":"https:\/\/itp.nyu.edu\/lowres\/critex-monika\/wp-json\/wp\/v2\/comments?post=3895"}],"version-history":[{"count":3,"href":"https:\/\/itp.nyu.edu\/lowres\/critex-monika\/wp-json\/wp\/v2\/posts\/3895\/revisions"}],"predecessor-version":[{"id":3901,"href":"https:\/\/itp.nyu.edu\/lowres\/critex-monika\/wp-json\/wp\/v2\/posts\/3895\/revisions\/3901"}],"wp:attachment":[{"href":"https:\/\/itp.nyu.edu\/lowres\/critex-monika\/wp-json\/wp\/v2\/media?parent=3895"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/itp.nyu.edu\/lowres\/critex-monika\/wp-json\/wp\/v2\/categories?post=3895"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/itp.nyu.edu\/lowres\/critex-monika\/wp-json\/wp\/v2\/tags?post=3895"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}