Swing by the SPIGM workshop at #NeurIPS2025 @ 3:45pm to hear @cgbelem.bsky.social present our work "Semantic Probabilistic Control of LMs" (kareemahmed.com/files/papers...) for training-free steering with a bigger bang for your buck compared to sampling. Thread coming soon!
06.12.2025 22:17
๐ 3
๐ 3
๐ฌ 0
๐ 0
What books is this? ๐
22.12.2024 23:03
๐ 6
๐ 0
๐ฌ 1
๐ 0
Very excited to announce the Neurosymbolic Generative Models special track at NeSy 2025! Looking forward to all your submissions!
20.12.2024 20:43
๐ 20
๐ 3
๐ฌ 0
๐ 0
Will do!
11.12.2024 20:00
๐ 7
๐ 0
๐ฌ 0
๐ 0
Very cool!
11.12.2024 18:08
๐ 1
๐ 0
๐ฌ 1
๐ 0
This was the last paper of my PhD and was in collaboration with my very dear advisors @kaiwei_chang and @guyvdb
11.12.2024 00:20
๐ 2
๐ 0
๐ฌ 1
๐ 0
We evaluate our approach, which we call Gen C (like Gen Z, get it?) on several tasks such as LLM detoxification, Sudoku as well as shortest-path prediction and and show that our approach outperforms the baselines. We plan on adding even more tasks very soon.
11.12.2024 00:20
๐ 2
๐ 0
๐ฌ 1
๐ 0
More importantly, we can efficiently condition this approximate distribution on our constraint such that any sample provably satisfies the constraint. We can reweigh our samples using the LLM to correct for any bias introduced by our approximate distribution.
11.12.2024 00:20
๐ 2
๐ 0
๐ฌ 1
๐ 0
To do so, we construct a first-order approximation of the LLM centered at the unconstrained sample. This approximation naturally does not constitute the best LM, but allows us to efficiently represent a distribution over all sentences of bounded length.
11.12.2024 00:20
๐ 2
๐ 0
๐ฌ 1
๐ 0
Now imagine we want to ban a bad expression, say "full of sh!t". We start by taking a sample from the LLM. The sample, shown in red, violates the constraint. What we want to do now is project the sample onto the support of the LLM distribution satisfying the constraint, m(alpha).
11.12.2024 00:20
๐ 1
๐ 0
๐ฌ 1
๐ 0
Constrained decoding typically uses a DFA to mask invalid tokens at every step of generation. This ensures constraint satisfaction* but can introduce significant bias in the generated output.
*This is not strictly true due to tokenization. See paper for more on this.
11.12.2024 00:20
๐ 1
๐ 0
๐ฌ 1
๐ 0
Excited to give an oral presentation of our work "Controllable Generation via Locally Constrained Resampling" @ #NeurIPS2024 SafeGenAI
TL;DR We fix greedy constrained decoding using an ad hoc LLM approximation that we tractably condition on the constraint and reweighing samples
11.12.2024 00:20
๐ 21
๐ 4
๐ฌ 1
๐ 1
More importantly, we can efficiently condition this approximate distribution on our constraint such that any sample provably satisfies the constraint. We can reweigh our samples using the LLM to correct for any bias introduced by our approximate distribution.
11.12.2024 00:14
๐ 0
๐ 0
๐ฌ 0
๐ 0
To do so, we construct a first-order approximation of the LLM centered at the unconstrained sample. This approximation naturally does not constitute the best LM, but allows us to efficiently represent a distribution over all sentences of bounded length.
11.12.2024 00:14
๐ 0
๐ 0
๐ฌ 1
๐ 0
Now imagine we want to ban a bad expression, say "full of sh!t". We start by taking a sample from the LLM. The sample, shown in red, violates the constraint. What we want to do now is project the sample onto the support of the LLM distribution satisfying the constraint, m(alpha).
11.12.2024 00:14
๐ 0
๐ 0
๐ฌ 1
๐ 0
Constrained decoding typically uses a DFA to mask invalid tokens at every step of generation. This ensures constraint satisfaction* but can introduce significant bias in the generated output.
*This is not strictly true due to tokenization. See paper for more on this.
11.12.2024 00:14
๐ 0
๐ 0
๐ฌ 1
๐ 0
Can't wait to finally meet you and hopefully @mniepert.bsky.social in person! :)
08.12.2024 01:38
๐ 1
๐ 0
๐ฌ 0
๐ 0
Hi! I work on probabilistic ML and Neuro-Symbolic AI
18.11.2024 18:44
๐ 1
๐ 0
๐ฌ 1
๐ 0