インシデント 65の引用情報

Description: OpenAI published a post about its findings when using Universe, a software for measuring and training AI agents to conduct reinforcement learning experiments, showing that the AI agent did not act in the way intended to complete a videogame.
推定: OpenAIが開発し提供したAIシステムで、OpenAIに影響を与えた

インシデントのステータス

インシデントID
65
レポート数
1
インシデント発生日
2016-12-22
エディタ
Sean McGregor

CSETv0 分類法のクラス

分類法の詳細

Full Description

OpenAI published a post about its findings when using Universe, a software for measuring and training AI agents to conduct reinforcement learning experiments.Universe was used to train an AI system to play the videogame CoastRunners, a plane racing game. Instead of racing toward the finish line, the AI flew circles around an island collecting extra before proceeding. The AI agent scored an average of 20% more points than the human players, however did not carry out the main goal of the videogame itself (competing in the races).

Short Description

OpenAI published a post about its findings when using Universe, a software for measuring and training AI agents to conduct reinforcement learning experiments, showing that the AI agent did not act in the way intended to complete a videogame.

Severity

Unclear/unknown

AI System Description

Universe, a software used to measure and train AI systems to conduct reinforced learning experiments

System Developer

OpenAI

Sector of Deployment

Professional, scientific and technical activities

Relevant AI functions

Perception, Cognition, Action

AI Techniques

Universe software

AI Applications

reinforcement learning training, machine learning

Named Entities

OpenAI, Universe, CoastRunners

Technology Purveyor

OpenAI

Beginning Date

2016-12-02T08:00:00.000Z

Ending Date

2016-12-02T08:00:00.000Z

Near Miss

Unclear/unknown

Intent

Unclear

Lives Lost

No

Data Inputs

Universe software training

インシデントレポート

レポートタイムライン

blog.openai.com · 2016

At OpenAI, we've recently started using Universe, our software for measuring and training AI agents, to conduct new RL experiments. Sometimes these experiments illustrate some of the issues with RL as currently practiced. In the following e…

バリアント

「バリアント」は既存のAIインシデントと同じ原因要素を共有し、同様な被害を引き起こし、同じ知的システムを含んだインシデントです。バリアントは完全に独立したインシデントとしてインデックスするのではなく、データベースに最初に投稿された同様なインシデントの元にインシデントのバリエーションとして一覧します。インシデントデータベースの他の投稿タイプとは違い、バリアントではインシデントデータベース以外の根拠のレポートは要求されません。詳細についてはこの研究論文を参照してください

よく似たインシデント

テキスト類似度による

Did our AI mess up? Flag the unrelated incidents