The Montreal AI Ethics Institute (MAIEI) co-hosted two public consultations with the Partnership on AI in May 2020. These meetups examined potential publication norms for responsible AI. MAIEI provides six initial recommendations, these include: create tools to navigate publication decisions.
The history of science and technology shows that seemingly innocuous
developments in scientific theories and research have enabled real-world
applications with significant negative consequences for humanity. In order to
ensure that the science and technology of AI is developed in a humane manner,
we must develop research publication norms that are informed by our growing
understanding of AI's potential threats and use cases. Unfortunately, it's
difficult to create a set of publication norms for responsible AI because the
field of AI is currently fragmented in terms of how this technology is
researched, developed, funded, etc. To examine this challenge and find
solutions, the Montreal AI Ethics Institute (MAIEI) co-hosted two public
consultations with the Partnership on AI in May 2020. These meetups examined
potential publication norms for responsible AI, with the goal of creating a
clear set of recommendations and ways forward for publishers.
In its submission, MAIEI provides six initial recommendations, these include:
1) create tools to navigate publication decisions, 2) offer a page number
extension, 3) develop a network of peers, 4) require broad impact statements,
5) require the publication of expected results, and 6) revamp the peer-review
process. After considering potential concerns regarding these recommendations,
including constraining innovation and creating a "black market" for AI
research, MAIEI outlines three ways forward for publishers, these include: 1)
state clearly and consistently the need for established norms, 2) coordinate
and build trust as a community, and 3) change the approach.