Check out the paper here:
partnershiponai.org/resource/pri...
Thanks to my co-authors and @partnershipai.bsky.social especially for leading the charge on this timely work!
Check out the paper here:
partnershiponai.org/resource/pri...
Thanks to my co-authors and @partnershipai.bsky.social especially for leading the charge on this timely work!
π€β¨ New report with @partnershipai.bsky.social!
AI agents pose new risks. Monitoring is essential to ensure effective oversight and intervention when needed. Our paper presents a framework for real-time failure detection that takes into account stakes, reversibility and affordances of agent actions.
β¨New Analysisβ¨
Can the new EU AI Code of Practice change the global AI safety landscape?
As companies like Anthropic, OpenAI, and Google sign on, CSETβs @miahoffmann.bsky.social explores the codeβs Safety and Security chapter. cset.georgetown.edu/article/eu-a...
Yesterday's new AI Action Plan has a lot worth discussing!
One interesting aspect is its statement that the federal government should withhold AI-related funding from states with "burdensome AI regulations."
This could be cause for concern.
βοΈ New Explainer!
Effectively evaluating AI models is more crucial than ever. But how do AI evaluations actually work?
In their new explainer,
@jessicaji.bsky.social, @vikramvenkatram.bsky.social &
@stephbatalis.bsky.social break down the different fundamental types of AI safety evaluations.
π‘Funding opportunityβshare with your AI research networksπ‘
Internal deployments of frontier AI models are an underexplored source of risk. My program at @csetgeorgetown.bsky.social just opened a call for research ideasβEOIs due Jun 30.
Full details β‘οΈ cset.georgetown.edu/wp-content/u...
Summary β¬οΈ
11) And if youβre now curious about CSETβs other recommendations for the AI Action Plan, you can check out the full response to the RFI here: cset.georgetown.edu/publication/...
10) If youβre still doubting the benefits of AI incident tracking, come by the Massive Data Instituteβs event on "AI Hazards: Understanding AI Incidents" today at 3pm, and let me and my fabulous co-panelists convince you in person! mdi.georgetown.edu/events/tswee...
Finally, and critically: central data collection and dissemination of lessons learned means that harms only have to occur once for everyone to mitigate their risk. This prevents recurrence and builds user and consumer confidence, which is essential for widespread AI adoption.
Incident tracking also reveals new, unexpected AI failure modes that we arenβt yet mitigating against. Over time, systematic data collection can help detect emerging risks and new types of harms, a critical benefit given the fast pace of AI innovation and deployment.
Over time, incident data can be used to evaluate the effectiveness of new safety policies and regulation through before and after comparisons. This helps refine governance policies through a direct feedback loop.
Using real-world data on what works and what doesnβt to guide AI safety research will help us innovate quicker and build reliable systems that are safe to deploy faster. In this way, incident reporting can help prioritize and direct AI safety research to where it is most effective.
AI incidents also shed light on the effectiveness of existing safety efforts. We might learn where current technical standards or risk management processes are insufficient to protect people from harm, revealing critical gaps that can be addressed by AI safety research.
For instance, we can learn about *how* the use of AI results in harm, e.g. through misuse, user error or AI failure. This information helps channel resources to the right kinds of safety efforts, since preventing misuse requires different measures than addressing operator error.
Why should the government do this?
What makes AI risk management so tricky is predicting how deploying an AI system can go wrong. AI incidents are a rich source of information about AI harms, harm mechanisms, AI failure modes and more. Leveraging those insights can make AI use safer.
Broadly speaking, an AI incident reporting regime has 4 core parts:
1) Incident detection;
2) Reporting to oversight bodies and inclusion in incident database;
3) Performance of impact assessments and root cause analyses; and
4) Dissemination of lessons learned
First, a definition. AI incidents are situations in which a deployed AI system is implicated in harm, e.g. when an AI recruiting tool makes a biased hiring decision. Incidents are varied and often take unexpected forms, so go check out the AIID for more real-world examples! incidentdatabase.ai
Today, @csetgeorgetown.bsky.social published our recommendations for the U.S. AI Action Plan. One of them is a CSET evergreen: implement an AI incident reporting regime for AI used by the federal government. Why? Short answer: because we can learn a ton from incidents! Long answer: π
π¨We're hiring β only a few days left to apply!π¨
CSET is looking for a Media Engagement Specialist to amplify our research. If you're a strategic communicator who can craft press releases, media pitches, & social content, apply by March 17, 2025! cset.georgetown.edu/job/media-en...
What: CSET Webinar πΊ
When: Tuesday, 3/25 at 12PM ET π
Whatβs next for AI red-teaming? And how do we make it more useful?
Join Tori Westerhoff, Christina Liaghati, Marius Hobbhahn, and CSET's @dr-bly.bsky.social * @jessicaji.bsky.social for a great discussion: cset.georgetown.edu/event/whats-...
What does the EU's shifting strategy mean for AI?
CSET's @miahoffmann.bsky.social & @ojdaniels.bsky.social have a new piece out for @techpolicypress.bsky.social.
Read it now π
Mia Hoffmann and Owen J. Daniels from Georgetownβs Center for Security and Emerging Technology say Europe's apparent shift on AI policy could change the global landscape for AI governance.
If youβve ever wondered what the EU and elephants have in common - or are wondering now- read my latest piece with @ojdaniels.bsky.social! We take a look what the EUβs new innovation-friendly regulatory approach might mean for the global AI policy ecosystem www.techpolicy.press/out-of-balan...
CSET is hiring π’
Weβre hiring a software engineer to support @emergingtechobs.bsky.social. Help build high-quality public tools and datasets to inform critical decisions on emerging tech issues.
Interested or know someone who would be? Learn more and apply π cset.georgetown.edu/job/software...
Thirdly, and most importantly, this decision reveals that the new European Commission is buying into the false narrative of innovation versus regulation which already dominates - and paralyzes - US tech policy.
Secondly, these questions will now be relegated to the national legal systems, which means uneven rules across the EU. Because what opponents to EU regulation need to understand is that the alternative to EU rules is not No Rules, it is 27 different sets of rules. Howβs that for simplification?
So what does this mean?
First, substantively, the questions on how to deal with liability for fundamental rights violations from AI, and liability across the value chains will remain open at the EU level.
For example, the PLD covers material harms from AI, but violations of fundamental rights - which are covered in the EU AI Act - would have fallen in the domain of the AILD. Similarly, the AILD was going to address the question of how liability should be distributed along the AI value chain.
Now, the AILD was not a flawless proposal. There was a lot of overlap with the Product Liability Directive (PLD), which already deals with software, including AI. But at the same time, it dealt with important aspects the PLD did not.
It appears the decision was political, and is a reflection of the new orientation of the new Commission: industry-friendly and anti-regulation. This is mirrored in statements made by EU leaders at the AI summit, claiming that EU rules would be βsimplifiedβ and applied in βbusiness-friendly waysβ.