Nicholas Popovič's Avatar

Nicholas Popovič

@nicpopovic.com

PhD student at TU Dresden and ScaDS.AI (formerly at KIT) doing research on language models interacting with structured information.

275
Followers
1,010
Following
23
Posts
21.11.2024
Joined
Posts Following

Latest posts by Nicholas Popovič @nicpopovic.com

Extractive Fact Decomposition for Interpretable Natural Language Inference in One Forward Pass

Find me at my poster in Hall C3 and/or find out more here: jedi.nicpopovic.com

07.11.2025 03:19 👍 0 🔁 0 💬 0 📌 0
Post image

Fact decomposition for interpretable and robust NLI without the need for an LLM?

Let me tell you how!

At 2pm today, I will be presenting “Extractive Fact Decomposition for Interpretable Natural Language Inference in One Forward Pass” at #EMNLP2025!

07.11.2025 03:19 👍 0 🔁 0 💬 1 📌 0
Post image

🚀 We are excited to introduce Kaleidoscope, the largest culturally-authentic exam benchmark.

📌 Most VLM benchmarks are English-centric or rely on translations—missing linguistic & cultural nuance. Kaleidoscope expands in-language multilingual 🌎 & multimodal 👀 VLMs evaluation

10.04.2025 20:24 👍 18 🔁 7 💬 1 📌 2
Post image

A bit of a mess around the conflict of COLM with the ARR (and to lesser degree ICML) reviews release. We feel this is creating a lot of pressure and uncertainty. So, we are pushing our deadlines:

Abstracts due March 22 AoE (+48hr)
Full papers due March 28 AoE (+24hr)

Plz RT 🙏

20.03.2025 18:20 👍 37 🔁 31 💬 3 📌 2

fyi: links seem to be broken

22.01.2025 12:53 👍 3 🔁 0 💬 1 📌 0

While you're not wrong, it felt like the paper in question had involved manual work. For example, the figures did not look generated or just modified. They were remade and unfortunately less clear than the original. Also, the paraphrasing wasn't great, I think ChatGPT would've done a better job 😅

10.01.2025 13:55 👍 2 🔁 0 💬 0 📌 0

That's a good guess :)

09.01.2025 16:45 👍 2 🔁 0 💬 0 📌 0

I reviewed a paper last year where the approach section turned out to be a sentence-for-sentence copy (+ simple paraphrasing) of another paper. Figures were (poorly) redrawn, too. Of course the evaluation section had new results, beating SOTA by miles... Still amazes me that someone would try that.

09.01.2025 15:10 👍 4 🔁 1 💬 2 📌 0

thanks for the insight! the whole process is such a black box. nice to hear that applications are still being read by humans, even if the workload is so high :)

17.12.2024 08:57 👍 1 🔁 0 💬 0 📌 0

Just wondering how many applications even make it through the initial filter and to your desk..

17.12.2024 05:18 👍 0 🔁 0 💬 1 📌 0
Video thumbnail

"Sora is a data-driven physics engine."
x.com/chrisoffner3...

10.12.2024 12:42 👍 137 🔁 16 💬 12 📌 10

The best part about beating sota is that you'll finally get to find the bug in your eval code.

06.12.2024 15:34 👍 5 🔁 0 💬 0 📌 0

Nice, would be interesting to see how it performs on various tasks depending on the language used for CoT...

02.12.2024 18:29 👍 1 🔁 0 💬 0 📌 0

Anybody else notice Qwen models occasionally (really not very often) switching to Chinese mid-sentence?

02.12.2024 17:07 👍 1 🔁 0 💬 1 📌 0

The answer to life's problems is simple:

sudo reboot

02.12.2024 01:31 👍 155 🔁 22 💬 20 📌 2

Depending on how funny you think it is that you just typed "meat-llama" instead of "meta-llama", it might be time for a break 🥩🦙

29.11.2024 17:29 👍 1 🔁 0 💬 0 📌 0

Cool, thanks for the link!

I wonder how much the results would be affected if the user messages include telling the assistant that it did a great job 😅

kind of like the whole “gpt will do better if you offer a tip” thing

28.11.2024 19:36 👍 0 🔁 0 💬 1 📌 0

For ICL, is it better to put examples in the system prompt or as user/assistant messages?

28.11.2024 18:57 👍 0 🔁 0 💬 1 📌 0

Now we just need to know which one to set to one for AGI :D

28.11.2024 15:08 👍 1 🔁 0 💬 0 📌 0

👋

23.11.2024 13:45 👍 1 🔁 0 💬 0 📌 0
Preview
Marco-o1: Towards Open Reasoning Models for Open-Ended Solutions Currently OpenAI o1 has sparked a surge of interest in the study of large reasoning models (LRM). Building on this momentum, Marco-o1 not only focuses on disciplines with standard answers, such as mat...

Alibaba has their own version on GPT-o1. This might be the best description of “o1-type”systems so far arxiv.org/abs/2411.14405

22.11.2024 12:18 👍 267 🔁 38 💬 8 📌 2

👋

22.11.2024 07:56 👍 1 🔁 0 💬 0 📌 0
Post image

👀

22.11.2024 07:54 👍 2 🔁 0 💬 0 📌 0
Post image

Has GPT-4o always done this kind of thing, or is it trained on "self-correction" chains from o1 now?

21.11.2024 13:40 👍 2 🔁 0 💬 1 📌 0
Post image

TIL. And I thought I was being silly by suggesting its probably "all clear" in a strong german accent...

21.11.2024 11:01 👍 0 🔁 0 💬 0 📌 0

Thank you :)

21.11.2024 08:51 👍 0 🔁 0 💬 0 📌 0
Video thumbnail

Huh, so do gifs not work here...?

21.11.2024 08:38 👍 0 🔁 0 💬 0 📌 0
Post image

First post, just to fill the void on my profile :)

Check out my recent EMNLP paper on how to use probing classifiers for streaming named entity recognition!

Link to paper and demo (please try the demo, I'm really proud of it 😅): ember.nicpopovic.com

21.11.2024 08:32 👍 0 🔁 0 💬 1 📌 0