Miriam Schirmer's Avatar

Miriam Schirmer

@miriamschirmer

CSS Postdoc @ Northwestern University NLP for Violence Research & Mental Health / Misinformation in Science 🌐 https://miriamschirmer.github.io

336
Followers
511
Following
22
Posts
29.09.2023
Joined
Posts Following

Latest posts by Miriam Schirmer @miriamschirmer

Post image

The computational psych preconference is back @spspnews.bsky.social for a full day! This year's lineup:

πŸ‘‰theory-driven modeling: Hyowon Gweon
πŸ‘‰data-driven discovery: @clemensstachl.bsky.social
πŸ‘‰application: me
πŸ‘‰ panel: @steveread.bsky.social Sandra Matz, @markthornton.bsky.social Wil Cunningham

29.09.2025 16:57 πŸ‘ 24 πŸ” 8 πŸ’¬ 1 πŸ“Œ 1
We present our new preprint titled "Large Language Model Hacking: Quantifying the Hidden Risks of Using LLMs for Text Annotation".
We quantify LLM hacking risk through systematic replication of 37 diverse computational social science annotation tasks.
For these tasks, we use a combined set of 2,361 realistic hypotheses that researchers might test using these annotations.
Then, we collect 13 million LLM annotations across plausible LLM configurations.
These annotations feed into 1.4 million regressions testing the hypotheses. 
For a hypothesis with no true effect (ground truth $p > 0.05$), different LLM configurations yield conflicting conclusions.
Checkmarks indicate correct statistical conclusions matching ground truth; crosses indicate LLM hacking -- incorrect conclusions due to annotation errors.
Across all experiments, LLM hacking occurs in 31-50\% of cases even with highly capable models.
Since minor configuration changes can flip scientific conclusions, from correct to incorrect, LLM hacking can be exploited to present anything as statistically significant.

We present our new preprint titled "Large Language Model Hacking: Quantifying the Hidden Risks of Using LLMs for Text Annotation". We quantify LLM hacking risk through systematic replication of 37 diverse computational social science annotation tasks. For these tasks, we use a combined set of 2,361 realistic hypotheses that researchers might test using these annotations. Then, we collect 13 million LLM annotations across plausible LLM configurations. These annotations feed into 1.4 million regressions testing the hypotheses. For a hypothesis with no true effect (ground truth $p > 0.05$), different LLM configurations yield conflicting conclusions. Checkmarks indicate correct statistical conclusions matching ground truth; crosses indicate LLM hacking -- incorrect conclusions due to annotation errors. Across all experiments, LLM hacking occurs in 31-50\% of cases even with highly capable models. Since minor configuration changes can flip scientific conclusions, from correct to incorrect, LLM hacking can be exploited to present anything as statistically significant.

🚨 New paper alert 🚨 Using LLMs as data annotators, you can produce any scientific result you want. We call this **LLM Hacking**.

Paper: arxiv.org/pdf/2509.08825

12.09.2025 10:33 πŸ‘ 303 πŸ” 106 πŸ’¬ 6 πŸ“Œ 23
Preview
Disparities by design: Toward a research agenda that links science misinformation and socioeconomic marginalization in the age of AI | HKS Misinformation Review Misinformation research often draws optimistic conclusions, with fact-checking, for example, being established as an effective means of reducing false beliefs. However, it rarely considers the details...

New Commentary in @misinforeview.bsky.social

We argue that recent technological disruptions provide an opportune moment for self-reflection, bringing together #AI, #science #misinformation, and #social #disparities within one research agenda. 🌎 πŸ€–

misinforeview.hks.harvard.edu/article/disp...

12.08.2025 16:48 πŸ‘ 3 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0
Post image

And that’s a wrap on the day! Our workshop keeps growing, and so do our social events.

Huge thanks to all the authors, speakers, attendees, and of course, our amazing organisers πŸ‘

The drinks were a blast!

#ACL2025NLP

01.08.2025 21:05 πŸ‘ 12 πŸ” 3 πŸ’¬ 0 πŸ“Œ 1

The 9th edition of the Workshop for Online Abuse and Harms is kicking off in Room 2.17 at #ACL2025NLP! We have a full program today, including 3 invited speakers, a panel discussion and a poster session. This year's workshop theme is Harms Beyond Hate Speech.

01.08.2025 07:10 πŸ‘ 14 πŸ” 4 πŸ’¬ 0 πŸ“Œ 1

Our paper on #ChildObjectification on TikTok is finally out! Check out my poster presentation tomorrow at the #Workshop on #Online #Abuse and #Harms @ #ACL.
⬇️ Read more about our findings in the post below!

aclanthology.org/2025.woah-1....

31.07.2025 12:14 πŸ‘ 6 πŸ” 2 πŸ’¬ 0 πŸ“Œ 0
Post image Post image

What an inspiring first day at the Summer School for Women* in Political Methodology here in Bremen yesterday! We were honored to be welcomed by the president of @unibremen.bsky.social‬, Prof. Jutta GΓΌnther, (1)

21.07.2025 13:55 πŸ‘ 49 πŸ” 4 πŸ’¬ 1 πŸ“Œ 4
Post image

The #SocialMedia track is starting at #IC2S2 with a talk on data donations, followed by presentations on polarization, engagement patterns, and community rules! 🌐

23.07.2025 09:06 πŸ‘ 6 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

Join our Centre for Sociology of Humans & Machines (SOHAM) at Trinity College Dublin as a Postdoctoral Research Fellow, exploring how AI can enhance collective intelligence. We're looking for someone with strong computational skills & a passion for human-machine systems
my.corehr.com/pls/trrecrui...

07.07.2025 15:58 πŸ‘ 4 πŸ” 2 πŸ’¬ 0 πŸ“Œ 2

The paper will be out with the ACL proceedings soon, I’ll keep you posted !

23.07.2025 08:17 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
A photo of Olstugan Tullen

A photo of Olstugan Tullen

I’ve never seen so many people at my instituteβ€˜s favorite bar β€” #ic2s2 is going strong day and night. Amazing people, amazing vibes!

22.07.2025 21:05 πŸ‘ 20 πŸ” 1 πŸ’¬ 1 πŸ“Œ 1
Topic Overview of Comments on Videos with Children

Topic Overview of Comments on Videos with Children

Join my talk on #ChildObjectification on TikTok at #IC2S2 today at 2:30 PM (πŸ“ Social Good & Ethics).
I built classification models to detect objectifying language and found: 10% of comments refer to appearance, 3% are objectifying. Models struggle with this task, with #RoBERTa outperforming GPT-4.

22.07.2025 08:27 πŸ‘ 10 πŸ” 2 πŸ’¬ 3 πŸ“Œ 1

If you're @ic2s2.bsky.social #IC2S2 and are looking for a faculty position, get in touch. @uni-saarland.de is hiring.
Right now, there's an open call in Computer Science @saarland-informatics-campus.de: www.uni-saarland.de/fileadmin/up...
Soon, there'll be open calls in the Social Sciences.

22.07.2025 06:02 πŸ‘ 16 πŸ” 12 πŸ’¬ 1 πŸ“Œ 0

We just kicked off our tutorial at #IC2S2 in beautiful NorrkΓΆping. @dustinbwright.com is currently presenting our hands-on part on automated fact checking - πŸ“stop by at Vingen 3-4!

21.07.2025 08:15 πŸ‘ 18 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0

I had a great time at GSO's #LeadershipAcademy in Berlin! Leading other people is not always straightforward, especially in academia.
My takeaway: merging #values around who I want to be as a mentor with a strategy for a lab in #ComputationalViolenceResearch I hope to build.

19.07.2025 18:28 πŸ‘ 8 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0

Thanks for inviting me and for the mug! β˜•οΈ

11.07.2025 16:34 πŸ‘ 4 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Being a good researcher goes beyond academic publishing; to me, it also means fostering a research environment grounded in integrity and continuous learning. That’s why I’m excited to be part of this year’s GSO #Leadership Academy. Looking forward to meeting this year’s cohort! #ResponsibleResearch

22.05.2025 16:09 πŸ‘ 4 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0
Post image

Have you ever asked yourself about the overall extent of TikTok? Here some numbers from "Just Another Hour on TikTok" - Great compliment to @bendavidsteel.bsky.social for this data collection effort!
w/ @miriamschirmer.bsky.social & Derek Ruths
arxiv.org/abs/2504.13279

21.04.2025 17:09 πŸ‘ 79 πŸ” 32 πŸ’¬ 4 πŸ“Œ 4

Join us online for my #TaDa talk on child exposure on TikTok next week on Wednesday! πŸ€“

28.03.2025 14:02 πŸ‘ 2 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Preview
Post-Doctoral Research Fellow at The University of Edinburgh Looking for a new job opportunity in academia? Check out this job opening for a Post-Doctoral Research Fellow on jobs.ac.uk!

Job offer! 3-year PDRA studying archaeological content on social media from a cultural evolution perspective. Based at Edinburgh, supervised by @chiara-bonacchi.bsky.social and myself -project "Weaponised Pasts: The Evolution Of Heritage-based Hostility On Social Media’. Please share!
bit.ly/3QCXUFS

25.02.2025 10:54 πŸ‘ 18 πŸ” 23 πŸ’¬ 1 πŸ“Œ 1

🀝 Human-AI synergy remains essential: AI speeds up analysis and identifies overarching themes. Humans provide interpretative depth, socio-political and historical awareness, and ideological sensitivity crucial in discourse analysis.

14.03.2025 14:08 πŸ‘ 2 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

πŸ’‘ Our Suggestion: To maximize the potential of LLMs, combine AI’s scalability with human expertise through culturally-informed training and theory-driven prompting.

14.03.2025 14:08 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

❗️ The Problem: AI neutrality often obscures deeper socio-political meanings. The model often misclassifies ideologically charged statements, underplaying their socio-political implications.

14.03.2025 14:07 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

πŸ” Key Findings:
GPT-4 efficiently identifies broad themes in large datasets but struggles with ideologically charged and culturally sensitive discourse. The model tends to default to neutral categories, failing to explicitly recognize hate speech or exclusionary narratives.

14.03.2025 14:06 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Preview
Large Language Models and the challenge of analyzing discriminatory discourse: human-AI synergy in researching hate speech on social media This article explores the potential and limitations of Large Language Models (LLMs) in conducting qualitative thematic analysis (TA). By comparing GPT-4’s analysis with human-led analysis of a YouT...

🧡New paper out! How effective is GPT-4 for analyzing hate speech on social media? Our research in the Journal of Multicultural Discourses explores this important question. www.tandfonline.com/doi/full/10....

14.03.2025 14:06 πŸ‘ 5 πŸ” 2 πŸ’¬ 4 πŸ“Œ 0

Great to be part of this spring's #TaDa Speaker Series - looking forward to the talks this season!

04.03.2025 18:38 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
mediaTUM - Medien- und Publikationsserver

πŸ“• My dissertation on #NLP for #Violence Studies has been published: mediatum.ub.tum.de?id=1751256

I've been looking at #abusive behavior online, as well as sharing of personal experiences with violence, incl. psychological #trauma.

Excited to push this research forward and connect with others 🌐

17.01.2025 18:12 πŸ‘ 7 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0
Post image

JOIN 2025 MethodsNET Summer School with Central European University (CEU)

πŸ“ Central European University, Vienna
πŸ—“οΈ 30 June to 18 July 2025

The full course list is online!!
More details: bit.ly/4gLbv9V
Early Bird Registration: Early February, 2025

#research #researchmethods #methodology #methods

10.01.2025 18:30 πŸ‘ 22 πŸ” 12 πŸ’¬ 0 πŸ“Œ 4
Preview
Junior Professorship (W1, non tenured) in Computational Social Science Deadline: 7th of January 2025

Job announcement: Junior Professorship in Computational Social Science at the University of Konstanz.
Join us at the Center for Data and Methods! Feel free to reach out if you have questions.
stellen.uni-konstanz.de/jobposting/2...

10.12.2024 14:34 πŸ‘ 36 πŸ” 31 πŸ’¬ 0 πŸ“Œ 0
Post image

We are hosting the 11th International Conference on Computational Social Science in Sweden
πŸš€The IC2S2'25 website is LIVE, and submissions are OPEN!
πŸ“NorrkΓΆping | July 21-24, 2025
Call for Abstracts (until Feb 24)
Call for Tutorials (until Jan 17)
πŸ”—Explore details & submit: ic2s2-2025.org

25.10.2024 12:10 πŸ‘ 74 πŸ” 47 πŸ’¬ 0 πŸ“Œ 1