Martin Tutek's Avatar

Martin Tutek

@mtutek

Postdoc @ TakeLab, UniZG | previously: Technion; TU Darmstadt | PhD @ TakeLab, UniZG Faithful explainability, controllability & safety of LLMs. πŸ”Ž On the academic job market πŸ”Ž https://mttk.github.io/

330
Followers
389
Following
76
Posts
24.11.2024
Joined
Posts Following

Latest posts by Martin Tutek @mtutek

Post image

πŸ“’ PhD position in the NeuroAI of Language

Why can LLMs predict brain activity so well? We're hiring a PhD student to find out -- AI interpretability meets neuroimaging
Deadline March 20
Please RT πŸ™
πŸ‘‡
mpi.nl/career-education/vacancies/vacancy/fully-funded-4-year-phd-position-neuroai-language

05.03.2026 13:34 πŸ‘ 45 πŸ” 35 πŸ’¬ 2 πŸ“Œ 1
Preview
Arithmetic Without Algorithms: Language Models Solve Math With a Bag of Heuristics Do large language models (LLMs) solve reasoning tasks by learning robust generalizable algorithms, or do they memorize training data? To investigate this question, we use arithmetic reasoning as a rep...

Great test for anyone learning mech interp is reading Nikankin et al's "Arithmetic Without Algorithms" which uses activation patching / circuits, probing, logit lens, describing max activating examples.. If you you follow along while reading, you'll realize you know a lot!

arxiv.org/abs/2410.21272

04.03.2026 06:07 πŸ‘ 15 πŸ” 4 πŸ’¬ 1 πŸ“Œ 0
Post image

We were thrilled to host @mtutek.bsky.social at our lab last week.
His talk "From Internals to Integrity: How Insights into Transformer LMs Improve Safety, Interpretability, and Explanation Faithfulness" led to great discussions! πŸ‘
#Transformers #AISafety #ExplainableAI #MLResearch #NLProc

24.02.2026 14:05 πŸ‘ 18 πŸ” 3 πŸ’¬ 0 πŸ“Œ 0
Post image

Can LLMs figure out who you are from your anonymous posts?

From a handful of comments, LLMs can infer where you live, what you do, and your interests; then search for you on the web.

New πŸ“„ w/ @SimonLermenAI, @joshua_swans, @AerniMichael, Nicholas Carlini, @florian_tramer 🧡

20.02.2026 17:03 πŸ‘ 122 πŸ” 44 πŸ’¬ 8 πŸ“Œ 14
Code examples for the Python package elfen:

# initializing extractor
extractor = elfen.Extractor(
data = df,
language = "en",
text_column = "text")
# extracting a single feature: ttr
extractor.extract("ttr")
# extracting a feature area/group: readability
extractor.extract_feature_group("readability")
# extracting all available features
extractor.extract_features()

Code examples for the Python package elfen: # initializing extractor extractor = elfen.Extractor( data = df, language = "en", text_column = "text") # extracting a single feature: ttr extractor.extract("ttr") # extracting a feature area/group: readability extractor.extract_feature_group("readability") # extracting all available features extractor.extract_features()

Extracting structural/linguistic properties for large text datasets can be annoying. Existing tools either are not maintained, do not scale, or do not cover extensive sets of linguistic features.

For this reason, I implemented 🧝elfen, a Python package for efficient linguistic feature extraction

18.02.2026 15:55 πŸ‘ 10 πŸ” 5 πŸ’¬ 1 πŸ“Œ 0

🚨 Emergency reviewer needed for ARR Resources and Evaluation track! Please ping me if you could review one paper by Friday. Topic is AI hallucinations, broadly speaking.

11.02.2026 11:10 πŸ‘ 1 πŸ” 3 πŸ’¬ 1 πŸ“Œ 0

Hello #NLProc #ACL2026NLP people. I am looking for **two emergency reviewers** in the Safety and Alignment in LLMs track for ACL/ARR.

Reviews are due Feb 15th. Please DM if interested and available.

Happy to offer drinks/food if you live in/pass by Lisbon β˜€οΈ

10.02.2026 14:59 πŸ‘ 6 πŸ” 10 πŸ’¬ 0 πŸ“Œ 0

Hello #NLProc #ACL2026NLP community, I'm looking for an emergency reviewer for an ARR submission on LLM interpretability.

If you're available to complete a review before Feb 15, please reply or DM πŸ™

10.02.2026 14:41 πŸ‘ 2 πŸ” 6 πŸ’¬ 0 πŸ“Œ 0

I need one too!

10.02.2026 14:35 πŸ‘ 2 πŸ” 2 πŸ’¬ 0 πŸ“Œ 0

I could use an emergency reviewer for an ACL submission involving interpretability and syntax. Please DM me if you might be able to provide an emergency review before February 15!

10.02.2026 07:38 πŸ‘ 4 πŸ” 4 πŸ’¬ 1 πŸ“Œ 0

I am looking for 2 emergency reviewers for the ARR Ethics, Bias & Fairness track. Please DM me if you are available πŸ™

10.02.2026 09:27 πŸ‘ 6 πŸ” 6 πŸ’¬ 0 πŸ“Œ 0

I'm looking for two emergency reviewers πŸ§‘β€πŸš’πŸ‘©β€πŸš’ for the ARR January Generalizability and Transfer track.

Please reach out if you have time & qualify for review or RT for visibilityπŸ™πŸ™

10.02.2026 11:43 πŸ‘ 2 πŸ” 6 πŸ’¬ 0 πŸ“Œ 0

Seems to be a common situation for ACs this round, but I'm also looking for two emergency reviewers for the January #ARR Evaluation and Resources track. I'd appreciate any help (reposts, encouragement, black magic...)

10.02.2026 11:15 πŸ‘ 3 πŸ” 6 πŸ’¬ 0 πŸ“Œ 0

Looking for emergency reviewers for ARR Special Track "Explainability of NLP Models". Topics: Faithfulness, mechanistic interpretability, surveys and position papers. Deadline Feb 14 AoE. #ACL2026NLP

09.02.2026 17:33 πŸ‘ 8 πŸ” 7 πŸ’¬ 1 πŸ“Œ 1

I'm sorry the guy changing the rules on the fly is named what

07.02.2026 14:34 πŸ‘ 17869 πŸ” 4468 πŸ’¬ 250 πŸ“Œ 110

We already know prompt repetition is a handy hack to improve a decoder-only LM’s performance as it allows the model to β€œsee” bidirectionally, an ability otherwise suppressed by the causal mask.

But what happens if we increase the number of repetitions? πŸ€”πŸ§΅ @eaclmeeting.bsky.social #EACL2026

02.02.2026 12:04 πŸ‘ 5 πŸ” 4 πŸ’¬ 1 πŸ“Œ 1

ManagerBench was accepted to ICLR! @iclr-conf.bsky.social #ICLR2026

LLMs are still either unsafe, or completely harm avoidant - even when the harm affects furniture πŸ›‹οΈ

Check out our benchmark, online or in Rio πŸ‡§πŸ‡·

04.02.2026 19:45 πŸ‘ 3 πŸ” 1 πŸ’¬ 0 πŸ“Œ 1

Matija Luka Kuki\'c, Marko \v{C}uljak, David Duki\'c, Martin Tutek, Jan \v{S}najder
Sequence Repetition Enhances Token Embeddings and Improves Sequence Labeling with Decoder-only Language Models
https://arxiv.org/abs/2601.17585

28.01.2026 01:15 πŸ‘ 2 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0
simplified overview of our aligned probing setup, where we join the behavioral and internal evaluation of LMs' toxicity

simplified overview of our aligned probing setup, where we join the behavioral and internal evaluation of LMs' toxicity

LMs that "know more" about toxicity are less toxic!
Our #TACL πŸ“„ connects behavior and internals:
πŸ’  LMs amplify toxicity beyond humans
πŸ’  Information about toxicity peaks in lower layers
πŸ’  Bypassing these layers increases toxicity
More detailsπŸ‘‡ #NLProc #interpretability (1/🧡)

27.01.2026 13:01 πŸ‘ 11 πŸ” 5 πŸ’¬ 1 πŸ“Œ 0
Post image

Can you solve this algebra puzzle? 🧩

cb=c, ac=b, ab=?

A small transformer can learn to solve problems like this!

And since the letters don't have inherent meaning, this lets us study how context alone imparts meaning. Here's what we found:πŸ§΅β¬‡οΈ

22.01.2026 16:09 πŸ‘ 48 πŸ” 10 πŸ’¬ 2 πŸ“Œ 2
Post image

⏳ Deadline approaching! We’re hiring 2 fully funded postdocs in #NLP.

Join the MilaNLP team and contribute to our upcoming research projects (SALMON & TOLD)

πŸ”— Details + how to apply: milanlproc.github.io/open_positio...

⏰ Deadline: Jan 31, 2026

19.01.2026 17:24 πŸ‘ 11 πŸ” 10 πŸ’¬ 0 πŸ“Œ 1

I'd like a link as well!

15.01.2026 09:20 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Nathan Stringham, Fateme Hashemi Chaleshtori, Xinyuan Yan, Zhichao Xu, Bei Wang, Ana Marasovi\'c
Teaching People LLM's Errors and Getting it Right
https://arxiv.org/abs/2512.21422

29.12.2025 07:45 πŸ‘ 4 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0
Post image

πŸš€ We’re opening 2 fully funded postdoc positions in #NLP!

Join the MilaNLP team and contribute to our upcoming research projects.

πŸ”— More details: milanlproc.github.io/open_positio...

⏰ Deadline: Jan 31, 2026

18.12.2025 15:29 πŸ‘ 19 πŸ” 13 πŸ’¬ 0 πŸ“Œ 2
Llama enjoying a mug of hot cocoa in an office with Tuesday, March 31 circled on a calendar behind them

Llama enjoying a mug of hot cocoa in an office with Tuesday, March 31 circled on a calendar behind them

COLM 2026 is just around the corner! Mark your calendars for:

πŸ’‘ Abstract deadline: Thursday, March 26, 2026
πŸ“„ Full paper submission deadline: Tuesday, March 31, 2026

Call for papers (website coming soon):
docs.google.com/document/d/1...

16.12.2025 15:31 πŸ‘ 9 πŸ” 4 πŸ’¬ 1 πŸ“Œ 1
The Doge of Venice visits a Murano glassworks in the 17th century. I will talk about why glassmaking in this era has some similarities to AI research today.

The Doge of Venice visits a Murano glassworks in the 17th century. I will talk about why glassmaking in this era has some similarities to AI research today.

At the #Neurips2025 mechanistic interpretability workshop I gave a brief talk about Venetian glassmaking, since I think we face a similar moment in AI research today.

Here is a blog post summarizing the talk:

davidbau.com/archives/202...

11.12.2025 15:02 πŸ‘ 17 πŸ” 3 πŸ’¬ 2 πŸ“Œ 2

I’m recruiting a postdoc to work on algorithms for cancer genome reconstruction. We have access to a rich set of tumour samples sequenced across multiple technologies. If interested, feel free to DM. Please share.

11.12.2025 03:04 πŸ‘ 13 πŸ” 12 πŸ’¬ 0 πŸ“Œ 1
Post image

πŸ§‘β€πŸ”¬I’m recruiting PhD students in Natural Language Processing @unileipzig.bsky.social Computer Science, together with @scadsai.bsky.social!

Topics include, but aren’t limited to:

πŸ”ŽLinguistic Interpretability
🌍Multilingual Evaluation
πŸ“–Computational Typology

Please share!

#NLProc #NLP

11.12.2025 13:36 πŸ‘ 41 πŸ” 25 πŸ’¬ 1 πŸ“Œ 3

I will be @euripsconf.bsky.social this week to present our paper as non-archival at the PAIG workshop (Beyong Regulation:
Private Governance & Oversight Mechanisms for AI). Very much looking forward to the discussions!

If you are at #EurIPS and want to chat about LLM's training data. Reach out!

02.12.2025 21:47 πŸ‘ 9 πŸ” 4 πŸ’¬ 0 πŸ“Œ 0
Post image

πŸ“’ Postdoc position πŸ“’

I’m recruiting a postdoc for my lab at NYU! Topics include LM reasoning, creativity, limitations of scaling, AI for science, & more! Apply by Feb 1.

(Different from NYU Faculty Fellows, which are also great but less connected to my lab.)

Link in 🧡

02.12.2025 16:04 πŸ‘ 21 πŸ” 12 πŸ’¬ 2 πŸ“Œ 1