๐ฏ Submission deadline: July 25, 2026
๐ฅ Organizers: JinYeong Bak, Rob van der Goot, Hyeju Jang, Weerayut Buaphet, Alan Ramponi, Wei Xu, Alan Ritter
๐ฏ Submission deadline: July 25, 2026
๐ฅ Organizers: JinYeong Bak, Rob van der Goot, Hyeju Jang, Weerayut Buaphet, Alan Ramponi, Wei Xu, Alan Ritter
๐ฃ The 11th Workshop on Natural User-generated Text #W-NUT will be held at #EMNLP2026!
We welcome original research on noisy data, informal texts, and natural variation in language use ๐ฃ๏ธ and host the MultiLexNorm2 task ๐
More info at:
๐ noisy-text.github.io/2026/
@emnlpmeeting.bsky.social #NLProc
Image showing the text "FadeIT: Fallacy Detection in Italian Social Media Texts, a Shared Task at EVALITA 2026".
โจ Shared task alert! โจ
We are organizing FadeIT, the first shared task on fallacy detection accounting for genuine disagreement. FadeIT is part of EVALITA, whose workshop will be held in beautiful Bari, Italy in February 2026.
Learn more from our website!
๐ sites.google.com/fbk.eu/fadei...
Stiamo vivendo una rivoluzione tecnologica, ma quanto conosciamo realmente l'uso dell'IA nella nostra quotidianitร ? Aiuta a scoprirlo compilando questo breve questionario (10 min): bit.ly/sondaggio_ai...
๐ Stiamo studiando come l'AI viene usata in Italia e per farlo abbiamo costruito un sondaggio!
๐ bit.ly/sondaggio_ai...
(รจ anonimo, richiede ~10 minuti, e se partecipi o lo fai girare ci aiuti un sacco๐)
Ci interessa anche raggiungere persone che non si occupano e non sono esperte di AI!
Further information:
๐ Website: noisy-text.github.io/2025/
๐ Proceedings: aclanthology.org/volumes/2025...
Organizers: JinYeong Bak, Rob van der Goot, Hyeju Jang, Weerayut Buaphet, Alan Ramponi, Wei Xu, Alan Ritter
See you tomorrow!
Abstract of Su Lin Blodgett's keynote talk
Keynote talk 2๏ธโฃ
๐ฃ๏ธ Su Lin Blodgett (Microsoft Research Montrรฉal)
๐ค May 3rd, 16:00 (UTC-6 time)
โจ What Can We Learn from Perspectives on Noisy
User-Generated Text?
Abstract of Verena Blaschke's keynote talk
Keynote talk 1๏ธโฃ
๐ฃ๏ธ Verena Blaschke (LMU Munich & MCML)
๐ค May 3rd, 09:30 (UTC-6 time)
โจ Beyond โnoisyโ text: How (and why) to process dialect data
๐ฃ Join us tomorrow May 3rd for the 10th Workshop on Noisy and User-generated Text #W-NUT at #NAACL2025 (๐ Room Navajo/Nambe)!
The workshop features 16 paper presentations and 2 exciting keynote talks by @verenablaschke.bsky.social and Su Lin Blodgett (titles+abstracts below)! #NLProc #NAACL
๐
I'll be presenting our work today ๐ Apr 30th, 16:15 (UTC-6 time) at #NAACL2025 during the R&E.2 oral session (Ballroom A)! Come say hi ๐
๐ aclanthology.org/2025.naacl-l...
#NAACL #NLProc #NLP
๐ great initiative!
We release data, code, and the full annotation guidelines to encourage extensions to cover new languages, topics, and additional perspectives ๐ฃ๏ธ
See you in Albuquerque! ๐๏ธ
A manual analysis of LLMsโ outputs unveils and quantifies different types of issues that call for future research to make generated responses less brittle in complex setups such as ours ๐ต๏ธโโ๏ธ
Check the paper for full results, analyses, discussion and insights! ๐
8/8
Our results show that fallacy detection, which involves capturing lexical, semantic, and even pragmatic aspects of communication, is still far from being addressed with LLMs in a zero-shot setup, especially if we aim at embracing human label variation
7/๐งต
We design multi-task fallacy detection baselines and assess LLMs in a zero-shot setting in four fallacy detection setups of increasing complexity: at the post- or the span-level, and using either fallacy macro-categories or the full inventory
6/๐งต
In the paper, we provide in-depth analyses and insights into the full annotation process ๐
We also conducted experiments by simultaneously accounting for multiple test sets (beyond โsingle ground truthโ), partial span matches, overlaps, and the varying severity of labeling errors
5/๐งต
Image showing inter-annotator agreement (IAA) scores for both span identification (gamma) and classification (gamma-cat) at each annotation round, before and after discussion
Due to the complexity of the task, we avoided crowdsourcing and instead devised multiple rounds of annotation and discussion among two expert annotators. We minimize annotation errors whilst keeping signals of human label variation on the whole dataset
โ ๏ธ Natural disagreement is not noise!
4/๐งต
Faina covers public discourse on ๐ migration, ๐ฑ climate change, and ๐ฅ public health over a โ๏ธ 4-year time frame (2019-22). It opens opportunities for modeling multiple ground truths at a the fine-grained level of text segments and benchmarking fallacy detection methods across topics and time
3/๐งต
We introduce Faina, the first fallacy detection dataset that embraces multiple plausible answers and natural disagreement. Faina includes >11K human-labeled span annotations with overlaps across 20 fallacy types on social media posts in Italian
*Faina (en: โbeech martenโ) ๐
2/๐งต
Example showing multiple plausible span annotations provided by annotators A1 and A2 due to different interpretations for the text "American study: mutation spreads four times faster, but ๐ are needed" in Italian
Happy to share that โFine-grained Fallacy Detection with Human Label Variationโ with @agnesedaff.bsky.social and @satonelli.bsky.social was accepted to #NAACL2025 main conference ๐
๐ arxiv.org/abs/2502.13853
#NLProc #NLP #NAACL
1/๐งต