Back to articles
AIHugging Face Blog

Comments on U.S. National AI Research Resource Interim Report

Back to Articles AI Policy @馃: Comments on U.S. National AI Research Resource Interim Report Published August 1, 2022 Update on GitHub Upvote - Irene Solaiman irenesolaiman Follow In late June 2022, Huggin...

The RSS feed only provided an excerpt. FlowMarket recovered the public content available from the original page without bypassing restricted content.

AI Policy @馃: Comments on U.S. National AI Research Resource Interim Report

Irene Solaiman

In late June 2022, Hugging Face submitted a response to the White House Office of Science and Technology Policy and National Science Foundation鈥檚 Request for Information on a roadmap for implementing the National Artificial Intelligence Research Resource (NAIRR) Task Force鈥檚 interim report findings. As a platform working to democratize machine learning by empowering all backgrounds to contribute to AI, we strongly support NAIRR鈥檚 efforts.

In our response, we encourage the Task Force to:

  • Appoint Technical and Ethical Experts as Advisors Technical experts with a track record of ethical innovation should be prioritized as advisors; they can calibrate NAIRR on not only what is technically feasible, implementable, and necessary for AI systems, but also on how to avoid exacerbating harmful biases and other malicious uses of AI systems. Dr. Margaret Mitchell , one of the most prominent technical experts and ethics practitioners in the AI field and Hugging Face鈥檚 Chief Ethics Scientist, is a natural example of an external advisor.

Resource (Model and Data) Documentation Standards

  • NAIRR-provided standards and templates for system and dataset documentation will ease accessibility and function as a checklist. This standardization should ensure readability across audiences and backgrounds. Model Cards are a vastly adopted structure for documentation that can be a strong template for AI models.

Make ML Accessible to Interdisciplinary, Non-Technical Experts

  • NAIRR should provide education resources as well as easily understandable interfaces and low- or no-code tools for all relevant experts to conduct complex tasks, such as training an AI model. For example, Hugging Face鈥檚 AutoTrain empowers anyone regardless of technical skill to train, evaluate, and deploy a natural language processing (NLP) model.

Monitor for Open-Source and Open-Science for High Misuse and Malicious Use Potential

  • Harm must be defined by NAIRR and advisors and continually updated, but should encompass egregious and harmful biases, political disinformation, and hate speech. NAIRR should also invest in legal expertise to craft Responsible AI Licenses to take action should an actor misuse resources.

Empower Diverse Researcher Perspectives via Accessible Tooling and Resources

  • Tooling and resources must be available and accessible to different disciplines as well as the many languages and perspectives needed to drive responsible innovation. This means at minimum providing resources in multiple languages, which can be based on the most spoken languages in the U.S. The BigScience Research Workshop , a community of over 1000 researchers from different disciplines hosted by Hugging Face and the French government, is a good example of empowering perspectives from over 60 countries to build one of the most powerful open-source multilingual language models.

Our memo goes into further detail for each recommendation. We are eager for more resources to make AI broadly accessible in a responsible manner.

More Articles from our Blog

AI and the Future of Cybersecurity: Why Openness Matters

AI Policy @馃: Response to the U.S. NTIA's Request for Comment on AI Accountability

Community

路 Sign up or log in to comment

Need an n8n workflow or help installing it?

After the briefing, move to execution: find an n8n template or a creator who can adapt it to your tools.

Source

Hugging Face Blog - huggingface.co

View original publication