Mia Hoffmann's Avatar

Mia Hoffmann

@miahoffmann

AI governance, harms and assessment | Research fellow @csetgeorgetown.bsky.social

162
Followers
242
Following
42
Posts
26.11.2024
Joined
Posts Following

Latest posts by Mia Hoffmann @miahoffmann

Preview
Prioritizing Real-Time Failure Detection in AI Agents - Partnership on AI A new PAI report argues that we need real-time failure detection to ensure AI agents can be monitored and stopped when needed.

Check out the paper here:
partnershiponai.org/resource/pri...

Thanks to my co-authors and @partnershipai.bsky.social especially for leading the charge on this timely work!

11.09.2025 16:35 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

πŸ€–βœ¨ New report with @partnershipai.bsky.social!
AI agents pose new risks. Monitoring is essential to ensure effective oversight and intervention when needed. Our paper presents a framework for real-time failure detection that takes into account stakes, reversibility and affordances of agent actions.

11.09.2025 16:35 πŸ‘ 1 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0
Preview
AI Safety under the EU AI Code of Practice β€” A New Global Standard? | Center for Security and Emerging Technology To protect Europeans from the risks posed by artificial intelligence, the EU passed its AI Act last year. This month, the EU released a Code of Practice to help providers of general purpose AI comply ...

✨New Analysis✨

Can the new EU AI Code of Practice change the global AI safety landscape?

As companies like Anthropic, OpenAI, and Google sign on, CSET’s @miahoffmann.bsky.social explores the code’s Safety and Security chapter. cset.georgetown.edu/article/eu-a...

30.07.2025 14:00 πŸ‘ 1 πŸ” 2 πŸ’¬ 0 πŸ“Œ 0

Yesterday's new AI Action Plan has a lot worth discussing!

One interesting aspect is its statement that the federal government should withhold AI-related funding from states with "burdensome AI regulations."

This could be cause for concern.

24.07.2025 18:55 πŸ‘ 6 πŸ” 3 πŸ’¬ 1 πŸ“Œ 0
Post image

βš–οΈ New Explainer!

Effectively evaluating AI models is more crucial than ever. But how do AI evaluations actually work?

In their new explainer,
@jessicaji.bsky.social, @vikramvenkatram.bsky.social &
@stephbatalis.bsky.social break down the different fundamental types of AI safety evaluations.

28.05.2025 14:02 πŸ‘ 4 πŸ” 2 πŸ’¬ 1 πŸ“Œ 2
Post image

πŸ’‘Funding opportunityβ€”share with your AI research networksπŸ’‘

Internal deployments of frontier AI models are an underexplored source of risk. My program at @csetgeorgetown.bsky.social just opened a call for research ideasβ€”EOIs due Jun 30.

Full details ➑️ cset.georgetown.edu/wp-content/u...

Summary ⬇️

19.05.2025 16:59 πŸ‘ 9 πŸ” 5 πŸ’¬ 1 πŸ“Œ 1
Preview
CSET's Recommendations for an AI Action Plan | Center for Security and Emerging Technology In response to the Office of Science and Technology Policy's request for input on an AI Action Plan, CSET provides key recommendations for advancing AI research, ensuring U.S. competitiveness, and max...

11) And if you’re now curious about CSET’s other recommendations for the AI Action Plan, you can check out the full response to the RFI here: cset.georgetown.edu/publication/...

17.03.2025 14:30 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Preview
Tech & Society Week 2025 β€” AI Hazards: Understanding AI Incidents - Massive Data Institute On Monday, March 17, 2025 from 3:00 to 4:00pm in Fisher Colloquium in Hariri Building on the Hilltop Campus, we will be hosting a panel discussion on AI incidents during Tech & Society Week 2025. The ...

10) If you’re still doubting the benefits of AI incident tracking, come by the Massive Data Institute’s event on "AI Hazards: Understanding AI Incidents" today at 3pm, and let me and my fabulous co-panelists convince you in person! mdi.georgetown.edu/events/tswee...

17.03.2025 14:30 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Finally, and critically: central data collection and dissemination of lessons learned means that harms only have to occur once for everyone to mitigate their risk. This prevents recurrence and builds user and consumer confidence, which is essential for widespread AI adoption.

17.03.2025 14:30 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Incident tracking also reveals new, unexpected AI failure modes that we aren’t yet mitigating against. Over time, systematic data collection can help detect emerging risks and new types of harms, a critical benefit given the fast pace of AI innovation and deployment.

17.03.2025 14:30 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Over time, incident data can be used to evaluate the effectiveness of new safety policies and regulation through before and after comparisons. This helps refine governance policies through a direct feedback loop.

17.03.2025 14:30 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Using real-world data on what works and what doesn’t to guide AI safety research will help us innovate quicker and build reliable systems that are safe to deploy faster. In this way, incident reporting can help prioritize and direct AI safety research to where it is most effective.

17.03.2025 14:30 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

AI incidents also shed light on the effectiveness of existing safety efforts. We might learn where current technical standards or risk management processes are insufficient to protect people from harm, revealing critical gaps that can be addressed by AI safety research.

17.03.2025 14:30 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

For instance, we can learn about *how* the use of AI results in harm, e.g. through misuse, user error or AI failure. This information helps channel resources to the right kinds of safety efforts, since preventing misuse requires different measures than addressing operator error.

17.03.2025 14:30 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Why should the government do this?
What makes AI risk management so tricky is predicting how deploying an AI system can go wrong. AI incidents are a rich source of information about AI harms, harm mechanisms, AI failure modes and more. Leveraging those insights can make AI use safer.

17.03.2025 14:30 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Broadly speaking, an AI incident reporting regime has 4 core parts:
1) Incident detection;
2) Reporting to oversight bodies and inclusion in incident database;
3) Performance of impact assessments and root cause analyses; and
4) Dissemination of lessons learned

17.03.2025 14:30 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Preview
Welcome to the Artificial Intelligence Incident Database The starting point for information about the AI Incident Database

First, a definition. AI incidents are situations in which a deployed AI system is implicated in harm, e.g. when an AI recruiting tool makes a biased hiring decision. Incidents are varied and often take unexpected forms, so go check out the AIID for more real-world examples! incidentdatabase.ai

17.03.2025 14:30 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Today, @csetgeorgetown.bsky.social published our recommendations for the U.S. AI Action Plan. One of them is a CSET evergreen: implement an AI incident reporting regime for AI used by the federal government. Why? Short answer: because we can learn a ton from incidents! Long answer: πŸ‘‡

17.03.2025 14:30 πŸ‘ 4 πŸ” 2 πŸ’¬ 1 πŸ“Œ 0
Preview
Media Engagement Specialist | Center for Security and Emerging Technology The Center for Security and Emerging Technology, under the School of Foreign Service, is a research organization focused on studying the security impacts of emerging technologies, supporting academic ...

🚨We're hiring β€” only a few days left to apply!🚨

CSET is looking for a Media Engagement Specialist to amplify our research. If you're a strategic communicator who can craft press releases, media pitches, & social content, apply by March 17, 2025! cset.georgetown.edu/job/media-en...

14.03.2025 14:27 πŸ‘ 0 πŸ” 2 πŸ’¬ 0 πŸ“Œ 0
Preview
What’s Next for AI Red-Teaming? | Center for Security and Emerging Technology On March 25, CSET will host an in-depth discussion about AI red-teaming β€” what it is, how it works in practice, and how to make it more useful in the future.

What: CSET Webinar πŸ“Ί
When: Tuesday, 3/25 at 12PM ET πŸ“…

What’s next for AI red-teaming? And how do we make it more useful?

Join Tori Westerhoff, Christina Liaghati, Marius Hobbhahn, and CSET's @dr-bly.bsky.social * @jessicaji.bsky.social for a great discussion: cset.georgetown.edu/event/whats-...

12.03.2025 15:11 πŸ‘ 4 πŸ” 4 πŸ’¬ 0 πŸ“Œ 1

What does the EU's shifting strategy mean for AI?

CSET's @miahoffmann.bsky.social & @ojdaniels.bsky.social have a new piece out for @techpolicypress.bsky.social.

Read it now πŸ‘‡

10.03.2025 14:17 πŸ‘ 4 πŸ” 4 πŸ’¬ 0 πŸ“Œ 0
Preview
Out of Balance: What the EU's Strategy Shift Means for the AI Ecosystem | TechPolicy.Press Mia Hoffmann and Owen J. Daniels from Georgetown’s Center for Security and Emerging Technology say Europe's movements could change the global landscape.

Mia Hoffmann and Owen J. Daniels from Georgetown’s Center for Security and Emerging Technology say Europe's apparent shift on AI policy could change the global landscape for AI governance.

10.03.2025 13:28 πŸ‘ 6 πŸ” 3 πŸ’¬ 0 πŸ“Œ 0
Preview
Out of Balance: What the EU's Strategy Shift Means for the AI Ecosystem | TechPolicy.Press Mia Hoffmann and Owen J. Daniels from Georgetown’s Center for Security and Emerging Technology say Europe's movements could change the global landscape.

If you’ve ever wondered what the EU and elephants have in common - or are wondering now- read my latest piece with @ojdaniels.bsky.social! We take a look what the EU’s new innovation-friendly regulatory approach might mean for the global AI policy ecosystem www.techpolicy.press/out-of-balan...

10.03.2025 13:42 πŸ‘ 2 πŸ” 1 πŸ’¬ 0 πŸ“Œ 2
Preview
Software Engineer | Center for Security and Emerging Technology The Center for Security and Emerging Technology (CSET), under the School of Foreign Service, is hiring a Software Engineer. The Software Engineer will be a generalist who can flex between full-stack w...

CSET is hiring πŸ“’

We’re hiring a software engineer to support @emergingtechobs.bsky.social. Help build high-quality public tools and datasets to inform critical decisions on emerging tech issues.

Interested or know someone who would be? Learn more and apply πŸ‘‡ cset.georgetown.edu/job/software...

03.03.2025 20:09 πŸ‘ 3 πŸ” 1 πŸ’¬ 0 πŸ“Œ 3

Thirdly, and most importantly, this decision reveals that the new European Commission is buying into the false narrative of innovation versus regulation which already dominates - and paralyzes - US tech policy.

13.02.2025 15:35 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Secondly, these questions will now be relegated to the national legal systems, which means uneven rules across the EU. Because what opponents to EU regulation need to understand is that the alternative to EU rules is not No Rules, it is 27 different sets of rules. How’s that for simplification?

13.02.2025 15:35 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

So what does this mean?
First, substantively, the questions on how to deal with liability for fundamental rights violations from AI, and liability across the value chains will remain open at the EU level.

13.02.2025 15:35 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

For example, the PLD covers material harms from AI, but violations of fundamental rights - which are covered in the EU AI Act - would have fallen in the domain of the AILD. Similarly, the AILD was going to address the question of how liability should be distributed along the AI value chain.

13.02.2025 15:35 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Now, the AILD was not a flawless proposal. There was a lot of overlap with the Product Liability Directive (PLD), which already deals with software, including AI. But at the same time, it dealt with important aspects the PLD did not.

13.02.2025 15:35 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

It appears the decision was political, and is a reflection of the new orientation of the new Commission: industry-friendly and anti-regulation. This is mirrored in statements made by EU leaders at the AI summit, claiming that EU rules would be β€œsimplified” and applied in β€œbusiness-friendly ways”.

13.02.2025 15:35 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0