Trustworthy AI Talk Series
Hey everyone, @ebagdasa.bsky.social and I are rebooting and widening the old prisec-ml seminar series. Give us a follow at @trust-ai-sem.bsky.social and check out our website at: trust-ai-sem.github.io
We already have a couple of amazing speakers booked!
27.02.2026 02:32
π 3
π 3
π¬ 0
π 0
Trustworthy AI Seminar Series
Trustworthy AI Seminar Series
π£ The first seminar of the Trustworthy AI Seminars will be given by Peter Kairouz on "Context Matters: A New Paradigm for Differentially Private Synthetic Text Generation."
π°οΈ Friday, March 6, at 9am PT / 12pm ET / 5pm UK time
π Details, Zoom link, YouTube stream, etc. at trust-ai-sem.github.io
02.03.2026 20:08
π 2
π 1
π¬ 1
π 0
Future problems:
* HCI/UX designs: How to design new interactions?
* Evaluations and datasets: What datasets will enable to accelerate progress?
* Social norms: How do we know what is appropriate and not?
* Multi-agent systems: When multiple agents interact, who is right? 5/6
06.11.2025 14:53
π 1
π 0
π¬ 1
π 0
While these problems are unsolvable in general, Contextual Privacy and Security for Agents offer new way to define policies for each situation to mitigate these threats. We can use both Model-level (reasoning) and System-level (reference monitors) designs to operationalize the policies. 4/6
06.11.2025 14:53
π 0
π 0
π¬ 1
π 0
We can break this into three questions for agents:
* Subjectivity of outputs: Does AI-generated text or image have spin or bias?
* Reality/ambiguity of inputs: What part of inputs can we trust?
* System complexity: What new issues complex reasoning and action plans (research agents) will bring 3/5
06.11.2025 14:53
π 1
π 0
π¬ 1
π 0
Core idea: agent security isnβt just "can the model do X?", but "should it do X here?" and that answer depends on context. Without contextual policies we either lose in utility or open agents to new attacks. 2/6
06.11.2025 14:53
π 0
π 0
π¬ 1
π 0
I just gave a keynote at the 18th(!) AI Security workshop CCS'25 in beautiful Taipei. I talked about challenges that future AI agents will face and argued that for defenses we must rely on the context and generate dynamic policies that define what is appropriate to share and do in each context! 1/6
06.11.2025 14:53
π 2
π 0
π¬ 1
π 0
Wanted to get into Multi-Agent System Safety research? Look no more! Terrarium π by @masonnaka.bsky.social enables different types of attacks: from misalignment to prompt injection to privacy leakage. Realistic tasks, clear metrics, nice abstractions!
31.10.2025 20:07
π 3
π 0
π¬ 0
π 0