EgoExoBench: A Benchmark for First- and Third-person View Video
Understanding in MLLMs
Baoqi Pei, Guo Chen et al.
Paper
Details
#EgoExoBench #MLLMs #VideoUnderstanding
📰🚨TwelveLabs video understanding models are now available in Amazon Bedrock by Channy Yun (윤석찬)
#TwelveLabs #AmazonBedrock #VideoUnderstanding #AIModels #VideoSearch
𝗠𝗖𝗠𝗟 𝗕𝗹𝗼𝗴: Finding moments in long videos? Easy for humans, tough for AI.
ReVisionLLM — by MCML Members Tanveer Hannan, Thomas Seidl & team — learns to scan like we do: look wide, zoom in and spot interesting segments.
🔗 mcml.ai/news/2025-06...
#AI #LLM #VideoUnderstanding #MCML
Next, "Re-thinking Temporal Search for Long-Form Video Understanding" #CVPR2025
🗓️ Fri Jun 13, 4PM-6PM
📍 ExHall D Poster #306
🔗 Paper: arxiv.org/abs/2504.02259
🌐 Website: longvideohaystack.github.io
💻 Code: github.com/LongVideoHay...
📊 Data: huggingface.co/datasets/LVH...
#VideoUnderstanding
Check out our pioneering paper on Video and Audiovisual Understanding with LLMs! Dive into the future of AI with us: #VideoUnderstanding #LargeLanguageModels #AIResearch
🕸️ github.com/yunlong10/Aw...
Apollo’s Multimodal Models Outperform Larger AI Rivals 🚀📹✨ www.azoai.com/news/2025010... #AI #MachineLearning #VideoUnderstanding #DeepLearning #Innovation #Research #Multimodal #Technology #Scalability #MetaGenAI @arxiv-stat-ml.bsky.social
Really interesting workshop by my colleagues at Surrey, don't miss it if you're at #BMVC in Glasgow #BMVC2024 #videounderstanding #computervision