Hubble is finally out! We used 200k GPU hours from NAIRR and NVIDIA to build a comprehensive resource for the scientific study of LLM memorization. Fully open-source models & data up to 8B params + 500B tokens with controlled data insertion to study memorization risks πβ¨
24.10.2025 18:36
π 7
π 1
π¬ 0
π 0
Hubble Suite logo (cloth patch with names of key organizations involved: USC, MPI, NVIDIA)
Announcing πHubble, a suite of open-source LLMs to advance the study of memorization!
Pretrained 1B/8B param models, with controlled insertion of texts designed to emulate key memorization risks: copyright (e.g., book passages), privacy (e.g., synthetic biographies), and test set contamination
24.10.2025 18:21
π 7
π 4
π¬ 1
π 2
I had a lot of fun contemplating about memorization questions at the @l2m2workshop.bsky.social panel yesterday together with Niloofar Mireshghallah and Reza Shokri, moderated by
@pietrolesci.bsky.social who did a fantastic job!
#ACL2025
02.08.2025 15:04
π 12
π 2
π¬ 1
π 1
Automatic metrics for assessing factuality are easy to run and commonly used, but do they work? In < 1 hour, come find the answer at poster 349 in Hall X4, where Iβll be presenting @ameyagodbole.bsky.social βs work uncovering inconsistencies, errors, and biases of factuality metrics!
30.07.2025 08:15
π 2
π 0
π¬ 1
π 0
Iβll be at ACL 2025 next week where my group has papers on evaluating evaluation metrics, watermarking training data, and mechanistic interpretability. Iβll also be co-organizing the first Workshop on LLM Memorization @l2m2workshop.bsky.social on Friday. Hope to see lots of folks there!
25.07.2025 16:36
π 2
π 0
π¬ 0
π 0
LLMs can propose plans and generate action semantics, but struggle with state tracking. Symbolic planners leverage specialized search algorithms, but require predefined action semantics for the environment.
PSALM integrates the strengths of both.
Come by @naaclmeeting.bsky.social Poster 6 in Hall 3 from 4-530pm today to see @billzhu.bsky.social's and Ishika Singh's work with me and @robinjia.bsky.social on PSALM: autonomously inducing symbolic pre- and post-conditions of actions with LLMs, symbolic planning, and text environment interaction!
01.05.2025 17:39
π 6
π 1
π¬ 1
π 0
Check out @billzhu.bsky.social βs excellent work on combining LLMs with symbolic planners at NAACL on Thursday! I will also be at NAACL Friday-Sunday, looking forward to chatting about LLM memorization, interpretability, evaluation, and more
30.04.2025 19:46
π 3
π 0
π¬ 0
π 0
At @naaclmeeting.bsky.social this week! Iβll be presenting our work on LLM domain induction with @thomason.bsky.social on Thu (5/1) at 4pm in Hall 3, Section I.
Would love to connect and chat about LLM planning, reasoning, AI4Science, multimodal stuff, or anything else. Feel free to DM!
30.04.2025 18:38
π 4
π 3
π¬ 0
π 1
Excited to share that my intern work at Meta GenAI is accepted to @iclr-conf.bsky.social #ICLR2025
Introducing TLDR: Token-Level Detective Reward Model For Large Vision Language Models.
TLDR provides fine-grained annotations to
each text token.
πarXiv: arxiv.org/abs/2410.04734
08.02.2025 05:29
π 5
π 1
π¬ 1
π 1
Our workshop on LLM Memorization is coming to ACL 2025! The call for papers is out, please submit both archival and non-archival (work in progress or already published) papers
27.01.2025 23:23
π 8
π 3
π¬ 0
π 0
I'll be at #NeurIPS2024! My group has papers analyzing how LLMs use Fourier Features for arithmetic and how TFs learn higher-order optimization for ICL (led by @deqing.bsky.social), plus workshop papers on backdoor detection and LLMs + PDDL (led by @billzhu.bsky.social)
09.12.2024 22:21
π 23
π 3
π¬ 1
π 1
A starter pack for #NLP #NLProc researchers! π
go.bsky.app/SngwGeS
04.11.2024 10:01
π 251
π 99
π¬ 45
π 13
USC NLP folks are on Bluesky!
Follow my amazing colleagues here
go.bsky.app/KUwSZ6W
12.11.2024 17:44
π 17
π 5
π¬ 3
π 2
Started a SoCal AI/ML/NLP researchers starter pack! It's a bit sparse right now, and perhaps more NLP heavy, but hey, nominate yourself and others! go.bsky.app/6QckPj9
19.11.2024 15:28
π 43
π 8
π¬ 17
π 1