Eugene Bagdasarian's Avatar

Eugene Bagdasarian

@ebagdasa

Assistant Professor at UMass Amherst CICS, Research Scientist at Google Research. https://people.cs.umass.edu/~eugene/

32
Followers
43
Following
7
Posts
01.02.2025
Joined
Posts Following

Latest posts by Eugene Bagdasarian @ebagdasa

Trustworthy AI Talk Series

Hey everyone, @ebagdasa.bsky.social and I are rebooting and widening the old prisec-ml seminar series. Give us a follow at @trust-ai-sem.bsky.social and check out our website at: trust-ai-sem.github.io

We already have a couple of amazing speakers booked!

27.02.2026 02:32 πŸ‘ 3 πŸ” 3 πŸ’¬ 0 πŸ“Œ 0
Preview
Trustworthy AI Seminar Series Trustworthy AI Seminar Series

πŸ“£ The first seminar of the Trustworthy AI Seminars will be given by Peter Kairouz on "Context Matters: A New Paradigm for Differentially Private Synthetic Text Generation."

πŸ•°οΈ Friday, March 6, at 9am PT / 12pm ET / 5pm UK time

🌎 Details, Zoom link, YouTube stream, etc. at trust-ai-sem.github.io

02.03.2026 20:08 πŸ‘ 2 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0
Preview
AISec'25: Context Rules! Privacy and Security for Future Trustworthy AI Agents Context Rules! Privacy and Security for Future AI Agents Eugene Bagdasarian AISec Workshop at CCS’25 Notes: Paper links are clickable Leave constructive comments and questions Spread the word if you f...

Here are my slides: docs.google.com/presentation... . Learn more and comment!
#AISec #CCS2025 #Agents #AISecurity #AISafety #ContextualIntegrity 6/6

06.11.2025 14:53 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

Future problems:
* HCI/UX designs: How to design new interactions?
* Evaluations and datasets: What datasets will enable to accelerate progress?
* Social norms: How do we know what is appropriate and not?
* Multi-agent systems: When multiple agents interact, who is right? 5/6

06.11.2025 14:53 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

While these problems are unsolvable in general, Contextual Privacy and Security for Agents offer new way to define policies for each situation to mitigate these threats. We can use both Model-level (reasoning) and System-level (reference monitors) designs to operationalize the policies. 4/6

06.11.2025 14:53 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

We can break this into three questions for agents:
* Subjectivity of outputs: Does AI-generated text or image have spin or bias?
* Reality/ambiguity of inputs: What part of inputs can we trust?
* System complexity: What new issues complex reasoning and action plans (research agents) will bring 3/5

06.11.2025 14:53 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

Core idea: agent security isn’t just "can the model do X?", but "should it do X here?" and that answer depends on context. Without contextual policies we either lose in utility or open agents to new attacks. 2/6

06.11.2025 14:53 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

I just gave a keynote at the 18th(!) AI Security workshop CCS'25 in beautiful Taipei. I talked about challenges that future AI agents will face and argued that for defenses we must rely on the context and generate dynamic policies that define what is appropriate to share and do in each context! 1/6

06.11.2025 14:53 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Preview
Schmidt Sciences awards $18M to researchers working to ensure AI benefits society - Schmidt Sciences Media Contact: Carlie Wiener; cwiener@schmidtsciences.orgΒ  Β  Β  Β  Β  Β  Β  Β  Β Β  NEW YORKβ€”Schmidt Sciences announced today that 28 scholars studying how to fulfill AI’s potential to dramatically benefit…

We're excited to welcome 28 new AI2050 Fellows! This 4th cohort of researchers are pursuing projects that include building AI scientists, designing trustworthy models, and improving biological and medical research, among other areas.

05.11.2025 15:43 πŸ‘ 15 πŸ” 2 πŸ’¬ 0 πŸ“Œ 2

Wanted to get into Multi-Agent System Safety research? Look no more! Terrarium 🐍 by @masonnaka.bsky.social enables different types of attacks: from misalignment to prompt injection to privacy leakage. Realistic tasks, clear metrics, nice abstractions!

31.10.2025 20:07 πŸ‘ 3 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0