Hazel Doughty's Avatar

Hazel Doughty

@hazeldoughty

Assistant Professor at Leiden University, NL. Computer Vision, Video Understanding. https://hazeldoughty.github.io

275
Followers
131
Following
33
Posts
19.11.2024
Joined
Posts Following

Latest posts by Hazel Doughty @hazeldoughty

Preview
DataCV 2026 @ CVPR 2026 Announcements The 5th DataCV Workshop and Challenge will be held as a half-day workshop in Denver, Colorado, USA, in conjunction with CVPR 2026. Paper submission deadline: March 10th, 2026 (23:59 Any...

We’re organizing the 5th DataCV Workshop @ #CVPR2026 .

If your work focuses on data, such as bias, robustness, distribution shifts, synthetic data, or dataset analysis, we’d love to see it.

Proceedings + DataCV Challenge.

Deadline: March 10, 2026 (AOE)
sites.google.com/view/datacv-...

23.02.2026 13:51 👍 2 🔁 1 💬 0 📌 0
Preview
Let's Split Up: Zero-Shot Classifier Edits for Fine-Grained Video Understanding Video recognition models are typically trained on fixed taxonomies which are often too coarse, collapsing distinctions in object, manner or outcome under a single label. As tasks and definitions evolv...

Paper: arxiv.org/abs/2602.16545
Code and benchmarks coming soon.

19.02.2026 13:57 👍 1 🔁 0 💬 0 📌 0
Post image

Beyond improved accuracy, our editing method enables richer refinements to new action variants and deeper splits of already fine-grained categories.

19.02.2026 13:57 👍 0 🔁 0 💬 1 📌 0

If a small number of examples is available, the zero-shot edit provides a strong initialization for low-shot refinement, without retraining the backbone.

19.02.2026 13:57 👍 0 🔁 0 💬 1 📌 0
Post image

The edit enables post-hoc category splitting and outperforms strong vision-language baselines, while preserving the rest of the label space.

19.02.2026 13:57 👍 0 🔁 0 💬 1 📌 0
Post image

Yes it can. Our approach derives a zero-shot edit to the classifier head by decomposing and reusing structure already encoded in the model.

No new video data required.

19.02.2026 13:57 👍 1 🔁 0 💬 1 📌 0

At first glance, splitting a category sounds like a data problem: collect more videos, retrain.

We ask whether that’s actually necessary.

Can a trained classifier be refined without retraining and without video data?

19.02.2026 13:57 👍 1 🔁 0 💬 1 📌 0
Post image

We introduce a new problem: category splitting.

Given a trained classifier, the goal is to replace one category with finer-grained subcategories, while preserving performance on all others.

19.02.2026 13:57 👍 1 🔁 0 💬 1 📌 0
Preview
Let's Split Up: Zero-Shot Classifier Edits for Fine-Grained Video Understanding Video recognition models are typically trained on fixed taxonomies which are often too coarse, collapsing distinctions in object, manner or outcome under a single label. As tasks and definitions evolv...

How flexible is a trained video classifier after training?

Our new #ICLR2026 paper investigates whether a category can be split into finer ones without retraining and without any videos.

arxiv.org/abs/2602.16545

19.02.2026 13:57 👍 3 🔁 1 💬 1 📌 0

Excited about detailed visual reasoning and subtle distinctions in #ComputerVision?

Only 1 week left to apply 👇

13.02.2026 12:27 👍 0 🔁 0 💬 0 📌 0
PhD Candidate in Fine-Grained Visual Understanding PhD Candidate in Fine-Grained Visual Understanding

🏹 Job alert: PhD Candidate in Fine-Grained Visual Understanding at @unileiden.bsky.social

📍 Leiden 🇳🇱
📅 Apply by Feb 20th
🔗 https://careers.universiteitleiden.nl/job/PhD-Candidate-in-Fine-Grained-Visual-Understanding/16323-en_US/

06.02.2026 13:45 👍 4 🔁 2 💬 0 📌 0

✨PhD vacancy alert✨ Joost Batenburg and I are looking for someone that wants to work on fine-grained visual understanding in #ComputerVision

Apply here before 20 Feb:
careers.universiteitleiden.nl/job/PhD-Cand...

22.01.2026 16:47 👍 1 🔁 0 💬 0 📌 1

Tomorrow, I’ll give a talk about future predictions in egocentric vision at the #CVPR2025 precognition workshop, in room 107A at 4pm.

I’ll retrace some history and show how precognition enables assistive downstream tasks and representation learning for procedural understanding.

11.06.2025 19:35 👍 5 🔁 1 💬 0 📌 0

Excited to be giving a keynote at the #CVPR2025 Workshop on Interactive Video Search and Exploration (IViSE) tomorrow. I'll be sharing our efforts working towards detailed video understanding.
📅 09:45 Thursday 12th June
📍 208 A
👉 sites.google.com/view/ivise2025

11.06.2025 16:36 👍 4 🔁 1 💬 0 📌 0
Post image

Have you heard about HD-EPIC?
Attending #CVPR2025
Multiple opportunities to know about the most highly-detailed video dataset with a digital twin, long-term object tracks, VQA,…
hd-epic.github.io

1. Find any of the 10 authors attending @cvprconference.bsky.social
– identified by this badge.

🧵

10.06.2025 21:48 👍 5 🔁 3 💬 1 📌 0
HD-EPIC: A Highly-Detailed Egocentric Video Dataset A Highly-Detailed Egocentric Video Dataset

Do you want to prove your Video-Language Model understands fine-grained, long-video, 3D world or anticipates interactions?
Be the 🥇st to win HD-EPIC VQA challenge
hd-epic.github.io/index#vqa-be...
DL 19 May
Winners announced @cvprconference.bsky.social #EgoVis workshop

06.05.2025 10:27 👍 8 🔁 3 💬 0 📌 0

Object masks &tracks for HD-EPIC have been released.. This completes our highly-detailed annotations.

Also, HD-EPIC VQA challenge is open [Leaderboard closes 19 May]... can you be 1st winner?
codalab.lisn.upsaclay.fr/competitions...

Btw, HD-EPIC was accepted @cvprconference.bsky.social #CVPR2025

03.04.2025 19:06 👍 9 🔁 4 💬 0 📌 0

The HD-EPIC VQA challenge for CVPR 2025 is now live: codalab.lisn.upsaclay.fr/competitions...

See how your model stacks up against Gemini and LLaVA Video on a wide range of video understanding tasks.

05.03.2025 15:55 👍 0 🔁 0 💬 0 📌 0
Post image

#CVPR2025 PRO TIP: To get a discount on your registration, join the Computer Vision Foundation (CVF). It’s FREE and makes @wjscheirer smile 😉

CVF: thecvf.com

28.02.2025 01:50 👍 14 🔁 6 💬 0 📌 1

HD-EPIC - hd-epic.github.io
Egocentric videos 👩‍🍳 with very rich annotations: the perfect testbed for many egocentric vision tasks 👌

07.02.2025 15:26 👍 6 🔁 2 💬 0 📌 0
https://hd-epic.github.io/

This was a monumental effort from a large team across Bristol, Leiden Singapore and Bath.

The VQA benchmark only scratches the surfaces of what is possible to evaluate with this detail of annotations.

Check out the website if you want to know more: hd-epic.github.io

07.02.2025 12:27 👍 1 🔁 0 💬 0 📌 0
Post image

VQA Benchmark

Our benchmark tests understanding in recipes, ingredients, nutrition, fine-grained actions, 3D perception, object movement and gaze. Current models have a long way to go with a best performance of 38% vs. 90% human baseline.

07.02.2025 12:27 👍 1 🔁 0 💬 1 📌 1
Post image

Scene & Object Movements

We reconstruct participants kitchens and annotate every time an object is moved.

07.02.2025 12:27 👍 2 🔁 0 💬 1 📌 0
Post image Post image

Fine-grained Actions

Every action has a dense description not only describing what happens in detail, but also how and why it happens.

07.02.2025 12:27 👍 1 🔁 0 💬 1 📌 0
Post image

As well as annotating temporal segments corresponding to each step we also annotate all the preparation needed to complete each step.

07.02.2025 12:27 👍 1 🔁 0 💬 1 📌 0
Post image

Recipe & Nutrition

We collect details of all the recipes participants chose to perform over 3 days in their own kitchen. Alongside ingredient weights and nutrition.

07.02.2025 12:27 👍 1 🔁 0 💬 1 📌 0
Video thumbnail

📢 Today we're releasing a new highly detailed dataset for video understanding: HD-EPIC

arxiv.org/abs/2502.04144

hd-epic.github.io

What makes the dataset unique is the vast detail contained in the annotations with 263 annotations per minute over 41 hours of video.

07.02.2025 12:27 👍 16 🔁 4 💬 1 📌 1
Video thumbnail

🛑📢
HD-EPIC: A Highly-Detailed Egocentric Video Dataset
hd-epic.github.io
arxiv.org/abs/2502.04144
New collected videos
263 annotations/min: recipe, nutrition, actions, sounds, 3D object movement &fixture associations, masks.
26K VQA benchmark to challenge current VLMs
1/N

07.02.2025 11:45 👍 33 🔁 6 💬 2 📌 4

We propose a simple baseline using phrase-level negatives and visual prompting to balance coarse- and fine-grained performance. This can easily combined with existing approaches. However, there is much potential for future work.

10.12.2024 06:46 👍 1 🔁 0 💬 1 📌 0
Post image

Incorporating fine-grained negatives into training does improve fine-grained performance, however it comes at the cost of coarse-grained performance.

10.12.2024 06:46 👍 1 🔁 0 💬 1 📌 0