{"id":17172,"date":"2024-02-23T18:05:45","date_gmt":"2024-02-23T18:05:45","guid":{"rendered":"http:\/\/scannn.com\/what-happened-with-gemini-image-generation\/"},"modified":"2024-02-23T18:05:45","modified_gmt":"2024-02-23T18:05:45","slug":"what-happened-with-gemini-image-generation","status":"publish","type":"post","link":"https:\/\/scannn.com\/lv\/what-happened-with-gemini-image-generation\/","title":{"rendered":"What happened with Gemini image generation"},"content":{"rendered":"<p> <br \/>\n<\/p>\n<div>\n<p data-block-key=\"p0ud1\">Three weeks ago, we launched a new image generation feature for the Gemini conversational app (formerly known as Bard), which included the ability to create images of people.  <\/p>\n<p data-block-key=\"ekmh6\">It\u2019s clear that this feature missed the mark. Some of the images generated are inaccurate or even offensive. We\u2019re grateful for users\u2019 feedback and are sorry the feature didn&#8217;t work well. <\/p>\n<p data-block-key=\"1jpih\">We\u2019ve <a href=\"https:\/\/twitter.com\/Google_Comms\/status\/1760603321944121506\" rt-link-type=\"external\">acknowledged the mistake<\/a> and temporarily paused image generation of people in Gemini while we work on an improved version.<\/p>\n<h2 data-block-key=\"5spnq\">What happened<\/h2>\n<p data-block-key=\"7jef9\">The Gemini conversational app is a specific product that is separate from Search, our underlying AI models, and our other products. Its image generation feature was built on top of an AI model called Imagen 2.<\/p>\n<p data-block-key=\"asqer\">When we built this feature in Gemini, we tuned it to ensure it doesn\u2019t fall into some of the traps we\u2019ve seen in the past with image generation technology \u2014 such as creating violent or sexually explicit images, or depictions of real people. And because our users come from all over the world, we want it to work well for everyone. If you ask for a picture of football players, or someone walking a dog, you may want to receive a range of people. You probably don\u2019t just want to only receive images of people of just one type of ethnicity (or any other characteristic). <\/p>\n<p data-block-key=\"495j9\">However, if you prompt Gemini for images of a specific type of person \u2014 such as \u201ca Black teacher in a classroom,\u201d or \u201ca white veterinarian with a dog\u201d \u2014 or people in particular cultural or historic contexts, you should absolutely get a response that accurately reflects what you ask for. <\/p>\n<p data-block-key=\"ps25\">So what went wrong? In short, two things. First, our tuning to ensure that Gemini showed a range of people failed to account for cases that should clearly <i>not<\/i> show a range. And second, over time, the model became way more cautious than we intended and refused to answer certain prompts entirely \u2014 wrongly interpreting some very anodyne prompts as sensitive. <\/p>\n<p data-block-key=\"49bbq\">These two things led the model to overcompensate in some cases, and be over-conservative in others, leading to images that were embarrassing and wrong. <\/p>\n<h2 data-block-key=\"5m637\">Next steps and lessons learned<\/h2>\n<p data-block-key=\"8fuoj\">This wasn\u2019t what we intended. We did not want Gemini to refuse to create images of any particular group. And we did not want it to create inaccurate historical \u2014 or any other \u2014 images. So we turned the image generation of people off and will work to improve it significantly before turning it back on. This process will include extensive testing.<\/p>\n<p data-block-key=\"8ff8n\">One thing to bear in mind: Gemini is built as a creativity and productivity tool, and it may not always be reliable, especially when it comes to generating images or text about current events, evolving news or hot-button topics. It will make mistakes. As we\u2019ve said from the beginning, hallucinations are a known challenge with all LLMs \u2014 there are instances where the AI just gets things wrong. This is something that we\u2019re constantly working on improving. <\/p>\n<p data-block-key=\"6bjp0\">Gemini tries to give factual responses to prompts \u2014 and our double-check feature helps evaluate whether there\u2019s content across the web to substantiate Gemini\u2019s responses \u2014 but we recommend relying on Google Search, where separate systems surface fresh, high-quality information on these kinds of topics from sources across the web. <\/p>\n<p data-block-key=\"d2ie8\">I can\u2019t promise that Gemini won\u2019t occasionally generate embarrassing, inaccurate or offensive results \u2014 but I can promise that we will continue to take action whenever we identify an issue. AI is an emerging technology which is helpful in so many ways, with huge potential, and we\u2019re doing our best to roll it out safely and responsibly.<\/p>\n<\/div>\n<p><script async src=\"\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script><br \/>\n<br \/><br \/>\n<br \/><a href=\"https:\/\/blog.google\/products\/gemini\/gemini-image-generation-issue\/\">Source link <\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Three weeks ago, we launched a new image generation feature for the Gemini conversational app (formerly known as Bard), which included the ability to create images of people. It\u2019s clear that this feature missed the mark. Some of the images generated are inaccurate or even offensive. We\u2019re grateful for users\u2019 feedback and are sorry the [&hellip;]<\/p>\n","protected":false},"author":16,"featured_media":14097,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[100],"tags":[],"class_list":["post-17172","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-google"],"_links":{"self":[{"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/posts\/17172","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/users\/16"}],"replies":[{"embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/comments?post=17172"}],"version-history":[{"count":0,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/posts\/17172\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/media\/14097"}],"wp:attachment":[{"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/media?parent=17172"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/categories?post=17172"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/scannn.com\/lv\/wp-json\/wp\/v2\/tags?post=17172"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}