Advances in AI Risk Management and Governance

The field of AI research is shifting towards a greater emphasis on risk management and governance, with a focus on developing systematic approaches to identifying and mitigating potential risks associated with advanced AI systems. This is evident in the development of new frameworks and tools for probabilistic risk assessment, human reliability analysis, and security steerability. Noteworthy papers include:

  • Adapting Probabilistic Risk Assessment for AI, which introduces a framework for assessing risks in AI systems using established techniques from high-reliability industries.
  • A Cognitive-Mechanistic Human Reliability Analysis Framework, which proposes a cognitive-mechanistic framework for human reliability analysis in nuclear power plants.
  • Security Steerability is All You Need, which defines a novel security measure for large language models and presents a methodology for measuring security steerability.

Sources

Adapting Probabilistic Risk Assessment for AI

A Cognitive-Mechanistic Human Reliability Analysis Framework: A Nuclear Power Plant Case Study

Navigating AI Policy Landscapes: Insights into Human Rights Considerations Across IEEE Regions

Security Steerability is All You Need

The EU AI Act in Development Practice: A Pro-justice Approach

Understanding and Mitigating Risks of Generative AI in Financial Services

When Testing AI Tests Us: Safeguarding Mental Health on the Digital Frontlines

Real-World Gaps in AI Governance Research

Out of the Loop Again: How Dangerous is Weaponizing Automated Nuclear Systems?

Catastrophic Liability: Managing Systemic Risks in Frontier AI Development

Built with on top of