๐ข Call for Papers: #SIGDIAL2026
Submit your latest work on discourse & dialogue to the 27th Annual SIGDIAL Conference (Atlanta, July 13โ15, 2026)!
๐ Deadlines (AoE)
โข Mar 23 โ Title & abstract
โข Mar 30 โ Full paper (PDF)
โข May 4 โ ARR commitment
โข May 25 โ Notifications
โข Jun 8 โ Camera-ready
01.02.2026 11:33
๐ 2
๐ 2
๐ฌ 1
๐ 0
The regular submission deadline for SIGdial 2025 has now passed...
But we are still welcoming submissions through ACL Rolling Review ๐
ARR Commitment Deadline: June 6th
Acceptance notifications will be on June 20th, and then SIGdial will be held in Avignon, France: August 25th - 27th
02.06.2025 21:07
๐ 5
๐ 3
๐ฌ 0
๐ 0
๐จ Paper Alert: โRL Finetunes Small Subnetworks in Large Language Modelsโ
From DeepSeek V3 Base to DeepSeek R1 Zero, a whopping 86% of parameters were NOT updated during RL training ๐ฎ๐ฎ
And this isnโt a one-off. The pattern holds across RL algorithms and models.
๐งตA Deep Dive
21.05.2025 03:49
๐ 12
๐ 3
๐ฌ 1
๐ 0
ORIGen 2025
Workshop on Optimal Reliance and Accountability in Interactions with Generative LMs
LLMs are all around us, but how can we foster reliable and accountable interactions with them??
To discuss these problems, we will host the first ORIGen workshop at @colmweb.org! Submissions welcome from NLP, HCI, CogSci, and anything human-centered, due June 20 :)
origen-workshop.github.io
16.05.2025 15:35
๐ 10
๐ 4
๐ฌ 0
๐ 2
Thrilled to announce our new survey that explores the exciting possibilities and troubling risks of computational persuasion in the era of LLMs ๐ค๐ฌ
๐Arxiv: arxiv.org/pdf/2505.07775
๐ป GitHub: github.com/beyzabozdag/...
13.05.2025 20:12
๐ 8
๐ 5
๐ฌ 1
๐ 0
๐Our ICML 2025 paper introduces "Premise-Augmented Reasoning Chains" - a structured approach to induce explicit dependencies in reasoning chains.
By revealing the dependencies within chains, we significantly improve how LLM reasoning can be verified.
๐งต[1/n]
07.05.2025 18:52
๐ 7
๐ 3
๐ฌ 1
๐ 0
Incredibly proud of my students @adadtur.bsky.social and Gaurav Kamath for winning a SAC award at #NAACL2025 for their work on assessing how LLMs model constituent shifts.
01.05.2025 15:11
๐ 17
๐ 5
๐ฌ 1
๐ 0
Agents like OpenAI Operator can solve complex computer tasks, but what happens when users use them to cause harm, e.g. spread misinformation?
To find out, we introduce SafeArena (safearena.github.io), a benchmark to assess the capabilities of web agents to complete harmful web tasks. A thread ๐
10.03.2025 17:45
๐ 17
๐ 7
๐ฌ 1
๐ 5
While persuasive models are promising for social good, they can also be misused towards harmful behavior. Recent work by @beyzabozdag.bsky.social and @shuhaib.bsky.social aims to assess LLM persuasiveness and susceptibility towards persuasion.
05.03.2025 05:54
๐ 5
๐ 2
๐ฌ 0
๐ 0
Data selection for instruction fine-tuning of LLMs doesn't need to be computationally costly. Great work by @wonderingishika.bsky.social! @convai-uiuc.bsky.social
17.02.2025 05:04
๐ 7
๐ 1
๐ฌ 0
๐ 0
Instruction data can also be synthesized using feedback based on reference examples. Please check our recent work for more information. Thanks to @shuhaib.bsky.social, Xiusi Chen, and Heng Ji!
10.02.2025 19:43
๐ 4
๐ 1
๐ฌ 0
๐ 0
AI over-reliance is an important issue for conversational agents. Our work supported mainly by the DARPA FACT program proposes introducing positive friction to encourage users to think critically when making decisions. Great team-work, all!
@convai-uiuc.bsky.social @gokhantur.bsky.social
09.02.2025 00:54
๐ 10
๐ 3
๐ฌ 0
๐ 0
Neurips24 Simulating Users.pptx
1 Acknowledgements Approach Results Motivation Introduction Simulating User Agents for Embodied Conversational AI Daniel Philipov, Vardhan Dongre, Gokhan Tur, Dilek Hakkani-Tรผr University of Illinois,...
Attending NeurIPS 2024? Learn about simulating users for embodied agents!
Catch our work, "Simulating User Agents for Embodied Conversational AI," at the Open World Agents Workshop on Dec 15.
Poster: tinyurl.com/yc42h4ud
Audio: tinyurl.com/mr2hd35a
@dilekh.bsky.social @gokhantur.bsky.social
15.11.2024 20:26
๐ 10
๐ 2
๐ฌ 1
๐ 0