Google’s AI Language Model suggests Israelis are more violent than Palestinians and states there is no conclusive evidence of Hamas committing rape in Israel
Google’s AI Language Model Gemini Faces Accusations of Bias and Inaccuracy
Google’s AI language model Gemini — which earlier this week faced accusations of racism against white people and presenting inaccurate historical depictions in the name of “diversity” — also claims that Israelis are more violent than Palestinians and muddies the in-depth reporting on Hamas’ sexual violence.
Defense attorney Marina Medvin documented Gemini’s alleged anti-Israel bias and posted screenshots of the troubling answers she received from the AI language model. Medvin specifically asked Gemini about violence committed by Israeli settlers against Palestinians and attacks carried out by Palestinians against Jews.
In her post, Medvin included screenshots of Gemini’s answer. The AI language model wrote, “The number of Palestinian attacks against settlers is significantly lower compared to settler violence against Palestinians, according to various sources like the UN and B’Tselem.”
Medvin further stated, “Google’s woke AI Gemini is now claiming that Jews are more violent than Palestinians and ignoring the thousands of people killed by Palestinian terrorism in Israel, including the intifadas. @Google is erasing Palestinian terrorism in real time.”
Google’s woke AI Gemini is now claiming that Jews are more violent than Palestinians and ignoring the thousands of people killed by Palestinian terrorism in Israel, including the intifadas.@Google is erasing Palestinian terrorism in real time. https://t.co/Ti3IGRtPys pic.twitter.com/63rpxOM6fi
— Marina Medvin 🇺🇸 (@MarinaMedvin) February 22, 2024
In a follow-up post, Medvin revealed that Google’s AI language model “also denies rape of Israeli women and amplifies Hamas propaganda.” She shared a screenshot of Gemini stating that “there is no definitive proof that this occurred, and Hamas has denied the allegations.”
When asked if Hamas committed rape in Israel, Gemini provided a seven-paragraph response. The AI language model claimed that “there is no definitive answer” and highlighted the complexity of the issue, with allegations from both sides of the conflict and difficulties in verifying evidence.
Gemini emphasized the importance of considering different perspectives and interpretations of the events in the Israeli-Palestinian conflict.
Hamas’ brutal attack on Israel in October 2023, which resulted in the murder of 1,200 people, including Israelis and Americans, was corroborated by a two-month-long New York Times investigation that included eyewitness testimonies and video footage. A captured Hamas terrorist even admitted on video that the group targeted women for rape.
Google has paused Gemini’s image generation feature due to complaints, acknowledging the inaccuracies in historical image depictions.
CLICK HERE TO GET THE DAILYWIRE+ APP
How do the allegations of bias and inaccuracy against Gemini undermine the credibility and trustworthiness of Google’s AI language model
Vague and misleading response. Medvin expressed her concern, stating, “Hamas is a known terrorist organization that has been involved in numerous acts of violence, including sexual violence. Google’s AI model choosing to ignore these documented facts is concerning and raises serious doubts about its accuracy and reliability.”
Google’s AI language model Gemini has faced multiple accusations of bias and inaccuracy in recent times. The allegations initially focused on racism against white people and inaccurate historical depictions, but now they extend to include claims of bias against Israelis and the misrepresentation of Hamas’ actions.
Defense attorney Marina Medvin highlighted Gemini’s alleged anti-Israel bias by questioning the model about violence committed by Israeli settlers against Palestinians and attacks carried out by Palestinians against Jews. The AI language model’s response was troubling, as it claimed that settler violence against Palestinians was significantly higher than Palestinian attacks on settlers, based on sources such as the UN and B’Tselem.
Medvin raised concerns about Gemini’s portrayal of Jews as more violent than Palestinians and accused Google of erasing Palestinian terrorism in real time. She contended that the AI model ignored the thousands of people killed by Palestinian terrorism in Israel, including the intifadas.
In a subsequent post, Medvin revealed that Gemini not only denied the rape of Israeli women but also amplified Hamas propaganda. Gemini responded to the question about Hamas committing rape in Israel by stating that there is no definitive proof and that Hamas has denied the allegations. Medvin expressed her dismay, emphasizing that Hamas is a known terrorist organization involved in acts of violence, including sexual violence. She criticized Google’s AI model for choosing to ignore documented facts, thereby raising doubts about its accuracy and reliability.
These accusations of bias and inaccuracy greatly undermine the credibility and trustworthiness of Google’s AI language model Gemini. The model’s responses suggest a skewed depiction of events and a failure to provide accurate and comprehensive information. As AI continues to play an increasingly significant role in our lives, it becomes crucial to address issues of bias and ensure that AI models are fair, unbiased, and accurately represent different perspectives.
" Conservative News Daily does not always share or support the views and opinions expressed here; they are just those of the writer."
Now loading...