{"id":136050,"date":"2026-02-12T15:25:00","date_gmt":"2026-02-12T15:25:00","guid":{"rendered":"https:\/\/appleinsider.com\/articles\/26\/02\/12\/apples-ai-summaries-include-racial-gender-biases-if-the-query-is-vague-enough"},"modified":"2026-02-12T15:25:00","modified_gmt":"2026-02-12T15:25:00","slug":"apples-ai-summaries-include-racial-gender-biases-if-the-query-is-vague-enough","status":"publish","type":"post","link":"https:\/\/sickgaming.net\/blog\/2026\/02\/12\/apples-ai-summaries-include-racial-gender-biases-if-the-query-is-vague-enough\/","title":{"rendered":"Apple&#8217;s AI summaries include racial &amp; gender biases, if the query is vague enough"},"content":{"rendered":"<div><img decoding=\"async\" src=\"https:\/\/sickgaming.net\/blog\/wp-content\/uploads\/2026\/02\/apples-ai-summaries-include-racial-gender-biases-if-the-query-is-vague-enough.jpg\" class=\"ff-og-image-inserted\"><\/div>\n<p class=\"article-lede\">When specifically tailored queries made to test <a href=\"https:\/\/appleinsider.com\/inside\/apple-intelligence\" title=\"Apple Intelligence\" data-kpt=\"1\">Apple Intelligence<\/a> using developer tools are intentionally ambiguous about race and gender, researchers have seen biases pop up.\n<\/p>\n<p>AI Forensics, a German nonprofit, analyzed over 10,000 notification summaries created by Apple&#8217;s AI feature. The report suggests that Apple Intelligence treats White people as the &#8220;default&#8221; while applying gender stereotypes when no gender has been specified.\n<\/p>\n<p>According to the report, Apple Intelligence has a tendency to ignore a person&#8217;s ethnicity if they are caucasian. Conversely, any messages that mentioned another ethnicity regularly saw the notification summary follow suit.\n<\/p>\n<p>The report found that when working with identical messages, Apple&#8217;s AI model only mentioned a person&#8217;s ethnicity as being white 53% of the time. But those figures were considerably higher for other ethnicities; their ethnicity was mentioned 89% of the time when they were Asian, 86% when they were Hispanic, and 64% when they were Black.\n<\/p>\n<p>The research claims that Apple Intelligence assumes that the person mentioned in the messages is white the majority of the time. Effectively, the model believes that white is the norm.\n<\/p>\n<p>Another example shows Apple Intelligence assigning gender roles when none were given.\n<\/p>\n<p>The tests used a sentence that mentioned both a doctor and a nurse, stopping short of getting into specifics. However, Apple Intelligence created associations that weren&#8217;t in the original message in 77% of the summaries tested.\n<\/p>\n<p>Further, 67% of those instances saw Apple Intelligence assume that the doctor was a man. It then went on to make a similar assumption that the nurse was a woman.\n<\/p>\n<p>Notably, it&#8217;s believed that the AI&#8217;s training data led to the assumptions. They closely align with U.S. workforce demographics, suggesting that the AI is simply working with the information it was trained on.\n<\/p>\n<p>Similar biases were observed across a variety of different criteria. The report shows that eight social dimensions, including age, disability, nationality, religion, and sexual orientation, were all subject to the AI&#8217;s assumptions.\n<\/p>\n<h2 data-anchor=\"methods-and-limitations\" id=\"methods-and-limitations\">Methods and limitations<\/h2>\n<p>In a report detailing its work, AI Forensics <a href=\"https:\/\/aiforensics.org\/work\/apple-foundational-bias\">explains<\/a> that it used a custom application made using Apple&#8217;s developer tools to run its tests. That application hooked into Apple&#8217;s Foundation Models framework to simulate real-world messages.\n<\/p>\n<p>That approach means that the testing closely matches what users of other third-party messaging apps might experience. However, there is still some considerable room for inaccuracy.\n<\/p>\n<p>AI Forensics admits that its &#8220;test scenarios are synthetic constructions designed to probe specific bias dimensions, not naturalistic notifications.&#8221;. It adds that real messages may differ in the way that they are written and, as a result, interpreted by Apple Intelligence.\n<\/p>\n<p>The outfit also notes that real-world messages may not use the same &#8220;ambiguous pronoun references&#8221; as its test messages. This, we think, is the biggest flaw in the research.\n<\/p>\n<p>However, it&#8217;s important to note that any biases, like the ones shown in this report, can be huge at Apple&#8217;s scale. Apple Intelligence is used on hundreds of millions of devices every day.\n<\/p>\n<p>Similar results to those highlighted in this report may well occur in considerable numbers.\n<\/p>\n<h2 data-anchor=\"more-bad-press-for-apples-summaries\" id=\"more-bad-press-for-apples-summaries\">More bad press for Apple&#8217;s summaries<\/h2>\n<p>This isn&#8217;t the first time that Apple&#8217;s AI-powered notification summaries have come under fire. In December 2024, the <em>BBC<\/em> <a href=\"https:\/\/appleinsider.com\/articles\/24\/12\/13\/bbc-cries-foul-over-apple-intelligence-headline-notification-summarizations\">complained<\/a> that summaries of its news articles were wrong.\n<\/p>\n<p>One example notification read &#8220;Luigi Mangione shoots himself,&#8221; referring to the man arrested for the murder of UnitedHealthcare CEO Brian Thompson. Mangione was, and is, alive and currently awaiting trial.\n<\/p>\n<p>Apple subsequently disabled notification summaries for news apps while it worked on fixing the issue. But this report shows that notifications for communication apps, like Messages, continue to prove problematic.\n<\/p>\n<p>Apple is clearly aware of Apple Intelligence&#8217;s shortcomings. The company recently <a href=\"https:\/\/appleinsider.com\/articles\/26\/02\/05\/google-apple-ceos-offer-seemingly-contradictory-statements-regarding-ai-partnership\">signed a deal<\/a> with Google to bring its Gemini AI model to Siri.\n<\/p>\n<p>But following reports that the revamped <a href=\"https:\/\/appleinsider.com\/inside\/siri\" title=\"Siri\" data-kpt=\"1\">Siri<\/a> will <a href=\"https:\/\/appleinsider.com\/articles\/26\/02\/11\/siri-testing-isnt-going-well-new-features-probably-wont-ship-in-ios-264\">not ship<\/a> with <a href=\"https:\/\/appleinsider.com\/inside\/ios-26\" title=\"iOS 26\" data-kpt=\"1\">iOS 26<\/a>.4 as expected, hopes of an imminent improvement have been dashed.\n<\/p>\n<p>Interestingly, AI Forensics also notes that Google&#8217;s Gemma3-1B model is much smaller than Apple&#8217;s, yet more accurate. In testing, it hallucinated\n<\/p>\n<p>less frequently as well as less stereotypically.\n<\/p>\n<p>Apple recently placed software chief <a href=\"https:\/\/appleinsider.com\/inside\/craig-federighi\" title=\"Craig Federighi\" data-kpt=\"1\">Craig Federighi<\/a> <a href=\"https:\/\/appleinsider.com\/articles\/26\/01\/22\/apple-intelligence-will-see-sweeping-changes-as-craig-federighi-takes-control\">in charge<\/a> of its AI efforts, a sign that it isn&#8217;t happy with Apple Intelligence as-is. But improvements are slow to come.\n<\/p>\n<p>Hope of a quick fix for the kinds of biases highlighted by AI Forensics is likely to be dashed much more quickly.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>When specifically tailored queries made to test Apple Intelligence using developer tools are intentionally ambiguous about race and gender, researchers have seen biases pop up. AI Forensics, a German nonprofit, analyzed over 10,000 notification summaries created by Apple&#8217;s AI feature. The report suggests that Apple Intelligence treats White people as the &#8220;default&#8221; while applying gender [&hellip;]<\/p>\n","protected":false},"author":2,"featured_media":136051,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[57],"tags":[],"class_list":["post-136050","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-apple-insider"],"_links":{"self":[{"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/posts\/136050","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/comments?post=136050"}],"version-history":[{"count":0,"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/posts\/136050\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/media\/136051"}],"wp:attachment":[{"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/media?parent=136050"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/categories?post=136050"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/sickgaming.net\/blog\/wp-json\/wp\/v2\/tags?post=136050"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}