Skip to Content
logologo
AI Incident Database
Open TwitterOpen RSS FeedOpen FacebookOpen LinkedInOpen GitHub
Open Menu
Discover
Submit
  • Welcome to the AIID
  • Discover Incidents
  • Spatial View
  • Table View
  • List view
  • Entities
  • Taxonomies
  • Submit Incident Reports
  • Submission Leaderboard
  • Blog
  • AI News Digest
  • Risk Checklists
  • Random Incident
  • Sign Up
Collapse
Discover
Submit
  • Welcome to the AIID
  • Discover Incidents
  • Spatial View
  • Table View
  • List view
  • Entities
  • Taxonomies
  • Submit Incident Reports
  • Submission Leaderboard
  • Blog
  • AI News Digest
  • Risk Checklists
  • Random Incident
  • Sign Up
Collapse

Report 53

Associated Incidents

Incident 147 Report
Biased Sentiment Analysis

Loading...
Google Sentiment Analysis API gives a biased output
techleer.com · 2017

Google developed its Cloud Natural Language API to allows the developers to work with language analyzer to reveal the actual meaning of the text. The system decided that whether the text was said in the positive sentiment or the negative. According to recent data released by Google, its API considers a word like ‘Homosexual” as negative.

We all know that API judges based on the information fed to it but what might surprise you that the API can also be biased just like humans. The users have to enter whole sentences gives predictive analysis on each word as well as the statement as a whole. In the output, one can see that the API gauges certain words to have negative or positive sentiment. These AI systems are trained using text, books.

According to a recent revelation, Cloud Natural Language API tend to bias its analysis toward negative attribution of certain descriptive terms. It is similar to how humans behave in the world. We all start our life with the good thoughts and memories which starts to get polluted negativity from the world.

Google also confirmed that its cloud API is giving the biased output. Google issued a statement apologizing the developers for the fault in the software.

Google said that “We dedicate a lot of efforts to making sure the NLP API avoids bias, but we don't always get it right. This is an example of one of those times, and we are sorry. We take this seriously and are working on improving our models. We will correct this specific case, and, more broadly, build more inclusive algorithms is crucial to bringing the benefits of machine learning to everyone.”

This incident is similar to that of what happened with Microsoft’s AI chatbot Tay. Microsoft had to close Tay in March 2016 after Twitter users taught it to be extremely a hideously racist and sexist conspiracy theorist.

Google has to come up with some solution to get rid of the biased output otherwise it may have to pull back its API.

Read the Source

Research

  • Defining an “AI Incident”
  • Defining an “AI Incident Response”
  • Database Roadmap
  • Related Work
  • Download Complete Database

Project and Community

  • About
  • Contact and Follow
  • Apps and Summaries
  • Editor’s Guide

Incidents

  • All Incidents in List Form
  • Flagged Incidents
  • Submission Queue
  • Classifications View
  • Taxonomies

2024 - AI Incident Database

  • Terms of use
  • Privacy Policy
  • Open twitterOpen githubOpen rssOpen facebookOpen linkedin
  • e1b50cd