インシデントのステータス
CSETv0 分類法のクラス
分類法の詳細Problem Nature
Specification, Robustness
Physical System
Software only
Level of Autonomy
High
Nature of End User
Amateur
Public Sector Deployment
No
Data Inputs
Online comments
GMF 分類法のクラス
分類法の詳細Known AI Goal Snippets
(Snippet Text: However, computer scientists and others on the internet have found the system unable to identify a wide swath of hateful comments, while categorizing innocuous word combinations like “hate is bad” and “garbage truck” as overwhelmingly toxic., Related Classifications: Hate Speech Detection)
CSETv1 分類法のクラス
分類法の詳細Incident Number
13
CSETv1_Annotator-1 分類法のクラス
分類法の詳細Incident Number
13
Special Interest Intangible Harm
yes
Date of Incident Year
2017
Estimated Date
Yes
Multiple AI Interaction
no
Embedded
no
CSETv1_Annotator-3 分類法のクラス
分類法の詳細Incident Number
13
インシデントレポート
レポートタイムライン
- 情報源として元のレポートを表示
- インターネットアーカイブでレポートを表示
Yesterday, Google and its sister Alphabet company Jigsaw announced Perspective, a tool that uses machine learning to police the internet against hate speech. The company heralded the tech as a nascent but powerful weapon in combatting onlin…
- 情報源として元のレポートを表示
- インターネットアーカイブでレポートを表示
In the examples below on hot-button topics of climate change, Brexit and the recent US election -- which were taken directly from the Perspective API website -- the UW team simply misspelled or added extraneous punctuation or spaces to the …
- 情報源として元のレポートを表示
- インターネットアーカイブでレポートを表示
The Google AI tool used to flag “offensive comments” has a seemingly built-in bias against conservative and libertarian viewpoints.
Perspective API, a “machine learning model” developed by Google which scores “the perceived impact a comment…
- 情報源として元のレポートを表示
- インターネットアーカイブでレポートを表示
Don’t you just hate how vile some people are on the Internet? How easy it’s become to say horrible and hurtful things about other groups and individuals? How this tool that was supposed to spread knowledge, amity, and good cheer is being us…
- 情報源として元のレポートを表示
- インターネットアーカイブでレポートを表示
Last month, I wrote a blog post warning about how, if you follow popular trends in NLP, you can easily accidentally make a classifier that is pretty racist. To demonstrate this, I included the very simple code, as a “cautionary tutorial”.
T…
- 情報源として元のレポートを表示
- インターネットアーカイブでレポートを表示
As politics in the US and Europe have become increasingly divisive, there's been a push by op-ed writers and politicians alike for more "civility" in our debates, including online. Amidst this push comes a new tool by Google's Jigsaw that u…
- 情報源として元のレポートを表示
- インターネットアーカイブでレポートを表示
A recent, sprawling Wired feature outlined the results of its analysis on toxicity in online commenters across the United States. Unsurprisingly, it was like catnip for everyone who's ever heard the phrase "don't read the comments." Accordi…
- 情報源として元のレポートを表示
- インターネットアーカイブでレポートを表示
Abstract
The ability to quantify incivility online, in news and in congressional debates, is of great interest to political scientists. Computational tools for detecting online incivility for English are now fairly accessible and potentiall…
- 情報源として元のレポートを表示
- インターネットアーカイブでレポートを表示
According to a 2019 Pew Center survey, the majority of respondents believe the tone and nature of political debate in the U.S. have become more negative and less respectful. This observation has motivated scientists to study the civility or…
バリアント
よく似たインシデント
Did our AI mess up? Flag the unrelated incidents
よく似たインシデント
Did our AI mess up? Flag the unrelated incidents