Toward Trustworthy AI Development: Mechanisms for Supporting Verifiable Claims
release_3nn6ov5hpbbjbo55xgeg6rm2gu
by
Miles Brundage, Shahar Avin, Jasmine Wang, Haydn Belfield, Gretchen Krueger, Gillian Hadfield, Heidy Khlaaf, Jingying Yang, Helen Toner, Ruth Fong, Tegan Maharaj, Pang Wei Koh (+47 others)
2020
Abstract
With the recent wave of progress in artificial intelligence (AI) has come a
growing awareness of the large-scale impacts of AI systems, and recognition
that existing regulations and norms in industry and academia are insufficient
to ensure responsible AI development. In order for AI developers to earn trust
from system users, customers, civil society, governments, and other
stakeholders that they are building AI responsibly, they will need to make
verifiable claims to which they can be held accountable. Those outside of a
given organization also need effective means of scrutinizing such claims. This
report suggests various steps that different stakeholders can take to improve
the verifiability of claims made about AI systems and their associated
development processes, with a focus on providing evidence about the safety,
security, fairness, and privacy protection of AI systems. We analyze ten
mechanisms for this purpose--spanning institutions, software, and hardware--and
make recommendations aimed at implementing, exploring, or improving those
mechanisms.
In text/plain
format
Archived Files and Locations
application/pdf 880.5 kB
file_br6bis66lnchhfzpehlwhuofom
|
arxiv.org (repository) web.archive.org (webarchive) |
2004.07213v1
access all versions, variants, and formats of this works (eg, pre-prints)