NIST introduces program to analyze real-world AI risks
It is critical to evaluate the societal implications of using artificial intelligence, especially in situations where people regularly engage with AI.
The National Institute of Standards and Technology (NIST) has introduced a new initiative, the Assessing Risks and Impacts of AI (ARIA) program, designed to assess the interactions between AI systems and human users to understand how these systems impact and are influenced by social factors.
According to NIST, the program will offer measurable metrics for assessing the performance of these systems, with a focus on aspects such as validity, reliability, safety, security, privacy, and fairness. The objective is to support the U.S. AI Safety Institute‘s endeavors to establish reliable AI systems.
“This new effort will support the U.S. AI Safety Institute, expand NIST’s already broad engagement with the research community, and help establish reliable methods for testing and evaluating AI’s functionality in the real world,” says Laurie E. Locascio, NIST director and under secretary of Commerce for Standards and Technology.
The initiative includes three tiers of assessment: model testing, red teaming, and field testing. Model testing aims to verify the capabilities of AI models, while red teaming focuses on stress testing applications to identify vulnerabilities. Field testing is used to examine how individuals interact with AI in everyday scenarios. Some specific instances where AI models will be evaluated include TV spoilers, meal planning, and pathfinding.
Looking ahead, NIST anticipates expanding the ARIA evaluation to add generative AI technologies, such as text-to-image models, as well as other AI methods like recommender systems or decision support tools.
One of NIST’s most well-known contributions to the field of AI is the ongoing evaluation of face biometrics and analysis models.
“ARIA will consider AI beyond the model and assess systems in context, including what happens when people interact with AI technology in realistic settings under regular use. This gives a broader, more holistic view of the net effects of these technologies,” says Reva Schwartz, NIST Information Technology Lab’s ARIA program lead.
The ARIA program builds on the AI Risk Management Framework released in January 2023. The framework outlines methods for analyzing and monitoring the risks and impacts of AI systems using both quantitative and qualitative techniques. The framework was developed to guide companies in managing the risks associated with AI systems.
Article Topics
AI | Assessing Risks and Impacts of AI (ARIA) | NIST | research and development | responsible AI | standards
Comments