ExplainableML's Avatar

ExplainableML

@eml-munich

Institute for Explainable Machine Learning at @www.helmholtz-munich.de and Interpretable and Reliable Machine Learning group at Technical University of Munich and part of @munichcenterml.bsky.social

499
Followers
25
Following
168
Posts
20.11.2024
Joined
Posts Following

Latest posts by ExplainableML @eml-munich

Post image

BayesVLM provides reliable uncertainty for pretrained vision-language models without retraining or inference-time sampling. It improves zero-shot calibration, reduces overconfident errors under domain shift, and enables more sample-efficient active learning with negligible overhead.

12.02.2026 13:08 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

BayesVLM places a Laplace posterior over the final projection layers and analytically propagates uncertainty to cosine similarities.
This avoids Monte Carlo sampling while enabling efficient uncertainty-aware inference and active learning.

12.02.2026 13:08 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

We introduce BayesVLM, a training-free post-hoc Bayesian method for uncertainty estimation in pretrained VLMs.
BayesVLM yields interpretable, well-calibrated uncertainty with virtually no inference overhead.

12.02.2026 13:08 πŸ‘ 2 πŸ” 2 πŸ’¬ 1 πŸ“Œ 0

[Paper]: arxiv.org/pdf/2412.06014
[Project]: aaltoml.github.io/BayesVLM/
[Code]: github.com/AaltoML/Baye...

12.02.2026 13:08 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

3/
Post-hoc Probabilistic Vision-Language Models
@antonbaumann.bsky.social, @ruili-pml.bsky.social, @marcusklasson.bsky.social, @smentu.bsky.social, @shyamgopal.bsky.social, @zeynepakata.bsky.social, @arnosolin.bsky.social , @trappmartin.bsky.social

12.02.2026 13:08 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

While models are largely robust, recovery is inefficient and doubt expression plays a crucial role in recovery. Models are also not style-invariant, and suppression of doubt in a reasoning trace can lead to performance degradation.

12.02.2026 13:08 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

To evaluate this, we developed a method to alter the chain of thought of a reasoning model at certain fixed time steps of the reasoning process. We then modified the reasoning by introducing various interventions and evaluated how the models reacted to them.

12.02.2026 13:08 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

Robust reasoning is becoming ever more important as we deploy LLMs in critical settings. But how robust is their ability to recover from noisy or incorrect reasoning steps? Which recovery mechanisms do they employ? And can we potentially make this recovery process more robust?

12.02.2026 13:08 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

2/
Are Reasoning LLMS Robust to Interventions
on Their Chain-of-thought?
Alexander von Recum, @lgirrbach.bsky.social, @zeynepakata.bsky.social
[Paper]: arxiv.org/pdf/2602.07470

12.02.2026 13:08 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

This establishes the first large-scale empirical link proving that dataset composition is a primary driver of model bias, and also creates the foundation to study complex dynamics like second-order bias transfer and amplification through model architecture.

12.02.2026 13:08 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

We demonstrate that a simple linear fit predicts 60-70% of the gender bias found in CLIP and Stable Diffusion directly from co-occurrences in the training data .

12.02.2026 13:08 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

We use an ensemble of MLLMs and custom classifiers to generate 276M+ person-centric annotations across the full dataset . 🦾 With these labels, we measured "dataset bias" via co-occurrence frequencies, correlating them with "model bias" in CLIP and Stable Diffusion to see if data predicts the model .

12.02.2026 13:08 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

We set out to change that! By auditing the massive LAION-400M dataset, we finally enable researchers to empirically test how well dataset statistics actually predict downstream model behavior .

12.02.2026 13:08 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Do foundation models merely reflect the bias in their data, or do they amplify it? So far, the link between dataset imbalances and model bias has been an assumption rather than a measurement .

12.02.2026 13:08 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

1/
Person-Centric Annotations of LAION-400M: Auditing Bias and Its Transfer to Models
@lgirrbach.bsky.social, Stephan Alaniz, Genevieve Smith,Trevor Darrell, @zeynepakata.bsky.social
[Paper]: arxiv.org/pdf/2510.03721?

12.02.2026 13:08 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

πŸ₯³Happy to share that we have three papers accepted to #ICLR2026. Congrats to our authors and see you in RioπŸŒ΄πŸ‡§πŸ‡·. Check the thread for highlightsπŸ‘‡

12.02.2026 13:08 πŸ‘ 6 πŸ” 2 πŸ’¬ 1 πŸ“Œ 0

4/
Invited talk: The Asymmetry of Adaptation: Reverse-Engineering Multimodal In-Context Learning
Yiran Huang & @zeynepakata.bsky.social
πŸ“[πŸ‡ΊπŸ‡Έ NeurIPS CCFM Workshop]: Sunday, December 7th 2025, 8:15 AM - 9:00 AM, San Diego Convention Center, Upper Level Room 25ABC

03.12.2025 11:52 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

3/
Concept-Guided Interpretability via Neural Chunking
Shuchen Wu , Stephan Alaniz , @shyamgopal.bsky.social , Peter Dayan, Eric Schulz, @zeynepakata.bsky.social
πŸ“[πŸ‡ΊπŸ‡ΈNeurIPS]: Fri, Dec 5, 2025 β€’ 11:00 AM – 2:00 PM PST, Exhibit Hall C,D,E #2113

03.12.2025 11:52 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

2/
Noise Hypernetworks: Amortizing Test-Time Compute in Diffusion Models
@lucaeyring.bsky.social , @shyamgopal.bsky.social , Alexey Dosovitskiy, @natanielruiz.bsky.social , @zeynepakata.bsky.social
πŸ“[πŸ‡ΊπŸ‡ΈNeurIPS]: Thu, Dec 4 β€’ 11:00 AM – 2:00 PM PST, Exhibit Hall C,D,E #3605

03.12.2025 11:52 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

1/
Sparse Autoencoders Learn Monosemantic Features in Vision-Language Models
Mateusz Pach, @shyamgopal.bsky.social , @qbouniot.bsky.social , Serge Belongie, @zeynepakata.bsky.social
πŸ“[πŸ‡ΊπŸ‡ΈNeurIPS]: Wed, Dec 3 β€’ 4:30 PM – 7:30 PM PST, Exhibit Hall C,D,E #1007
πŸ“[πŸ‡©πŸ‡°EurIPS]: Thu, Dec 4, #98

03.12.2025 11:52 πŸ‘ 5 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0

Our EML members have arrived in beautiful San Diego for #NeurIPS2025! 🌴
We’re excited to share our latest research β€” three poster presentations and one workshop presentation.
Check out the thread below πŸ‘‡ and come say hi to our authors!

03.12.2025 11:52 πŸ‘ 4 πŸ” 0 πŸ’¬ 1 πŸ“Œ 1

πŸŽ“PhD application season is back!

We’re hiring ONLY through the ELLIS @ellis.eu and the MCML
@munichcenterml.bsky.social

πŸ“ŒPlease denote Prof. Zeynep Akata @zeynepakata.bsky.social as your preferred supervisor!
πŸ‘‰ Link to ELLIS (ellis.eu/news/ellis-p...) and MCML (mcml.ai/opportunitie...)

24.10.2025 10:21 πŸ‘ 5 πŸ” 1 πŸ’¬ 0 πŸ“Œ 1
Post image

Results. GenEval: SDXL 0.55β†’0.61 (notable gains in two objects, counting, color attribution). T2I-CompBench: broad boosts (esp. Color/Texture). DPG-Bench (SDXL): DSG 74.65β†’79.26, Q-Align 0.72β†’0.81; user study: RankDPO wins over SDXL & DPO-SDXL.

20.10.2025 12:34 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

We propose RankDPOβ€”a listwise preference objective that weights pairwise denoising comparisons using DCG-style gains/discounts, optimizing the entire ranking per prompt rather than isolated pairs.

20.10.2025 12:34 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

Direct Preference Optimization is strong for T2Iβ€”but human labels are pricey/outdated. We build Syn-Pic: a fully synthetic ranked preference dataset by ensembling 5 reward models to remove humans from the loop.

20.10.2025 12:34 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

2/
Scalable Ranked Preference Optimization for Text-to-Image Generation
@shyamgopal.bsky.social , Huseyin Coskun, @zeynepakata.bsky.social , Sergey Tulyakov, Jian Ren, Anil Kag
[Paper]: arxiv.org/pdf/2410.18013
πŸ“Hall I #1702
πŸ•‘Oct 22, Poster Session 4

20.10.2025 12:34 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

SUB enables rigorous stress-testing of interpretable models. We find that CBMs fail to generalize to these novel combinations of known concepts.

20.10.2025 12:34 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

To generate precise variations, we propose Tied Diffusion Guidance (TDG) β€” sharing noise across two parallel denoising processes to ensure correct class and attribute generation.

20.10.2025 12:34 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

We introduce SUB, a fine-grained image & concept benchmark with 38,400 synthetic bird images 🦀.
Using 33 classes & 45 concepts (e.g., wing color, belly pattern), SUB tests how robust CBMs are to targeted concept variations.

20.10.2025 12:34 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Concept Bottleneck Models (CBMs) hold huge promise for making AI more transparentβ€”especially in high-stakes fields like medicine. But how well do they hold up under distribution shifts? 🧠

20.10.2025 12:34 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0