Excited to share that 2/2 papers from our Lab @AreaSciencePark were accepted to #NeurIPS2025 (one spotlight ๐)
Great work everyone!
@alexpietroserra.bsky.social @francescortu.bsky.social @lbasile.bsky.social @lvaleriani.bsky.social @diegodoimo.bsky.social @maiorca.xyz @locatelf.bsky.social
22.09.2025 08:55
๐ 7
๐ 2
๐ฌ 0
๐ 0
Nice start of @neuripsconf.bsky.social!
Our work with @francescortu.bsky.social and @diegodoimo.bsky.social on the Competition of Mechanisms to understand counterfactuality in LLMs featured in the "Causality for LLMs" workshop :-)
Check out our ACL2024 paper aclanthology.org/2024.acl-long.โฆ
10.12.2024 20:19
๐ 9
๐ 1
๐ฌ 0
๐ 0
Thanks again, @diegodoimo.bsky.social and @albecazzaniga.bsky.social , for the fantastic mentorship and support! ๐๐ They are also attending #NeurIPS, so feel free to reach out to them to discuss our results. Iโm excited to keep pushing forward on these topics! ๐
10.12.2024 20:10
๐ 1
๐ 0
๐ฌ 0
๐ 0
Thanks to the amazing team at LADE @areasciencepark: @lvaleriani.bsky.social @lbasile.bsky.social @AlessioAnsuini @diegodoimo.bsky.social @albecazzaniga.bsky.social ๐
10.12.2024 20:10
๐ 2
๐ 0
๐ฌ 1
๐ 0
It was super fun to take our first step in interpreting multimodal LLMs, working closely with the brilliant @alexpietroserra.bsky.social and @EmanuelePanizon
10.12.2024 20:10
๐ 0
๐ 0
๐ฌ 1
๐ 0
โ
This shows that, starting from the mid-layers, a single token effectively summarizes all 1024 image tokens!
โ This does not occur in models fine-tuned for visual understanding (such as Pixtral).
10.12.2024 20:10
๐ 1
๐ 0
๐ฌ 1
๐ 0
Additionally, blocking communication from this token significantly disrupts performance on standard benchmarks, while blocking image-text communication does not
10.12.2024 20:10
๐ 1
๐ 0
๐ฌ 1
๐ 0
๐ฏ Key finding: In these models the hidden representations of images and text form disjoint clusters and the communication between modalities is mediated by the special token <end-of-image>!
10.12.2024 20:10
๐ 1
๐ 0
๐ฌ 1
๐ 0
๐ Check out our code and data at: ritareasciencepark.github.io/Narrow-gate
10.12.2024 20:10
๐ 0
๐ 0
๐ฌ 1
๐ 0
๐จ ๐จ Excited to share our latest paper, now on #arXiv!
๐ผ๏ธ We studied how unified VLMs, trained to generate both text and images (e.g., Meta's Chameleon), exchange information between modalities, comparing them to standard VLMs.
๐ Paper: arxiv.org/abs/2412.06646
Deep dive: ๐
10.12.2024 20:10
๐ 11
๐ 2
๐ฌ 1
๐ 3
Screenshot of the paper.
Even as an interpretable ML researcher, I wasn't sure what to make of Mechanistic Interpretability, which seemed to come out of nowhere not too long ago.
But then I found the paper "Mechanistic?" by
@nsaphra.bsky.social and @sarah-nlp.bsky.social, which clarified things.
20.11.2024 08:00
๐ 230
๐ 26
๐ฌ 7
๐ 2
Thanks for creating the starter pack! I'd love to be added as well! ๐
20.11.2024 10:41
๐ 2
๐ 0
๐ฌ 0
๐ 0