Understanding Long Videos in One Multimodal Language Model Pass

25 Mar 2024  ·  Kanchana Ranasinghe, Xiang Li, Kumara Kahatapitiya, Michael S. Ryoo ·

Large Language Models (LLMs), known to contain a strong awareness of world knowledge, have allowed recent approaches to achieve excellent performance on Long-Video Understanding benchmarks, but at high inference costs. In this work, we first propose Likelihood Selection, a simple technique that unlocks faster inference in autoregressive LLMs for multiple-choice tasks common in long-video benchmarks. In addition to faster inference, we discover the resulting models to yield surprisingly good accuracy on long-video tasks, even with no video specific information. Building on this, we inject video-specific object-centric information extracted from off-the-shelf pre-trained models and utilize natural language as a medium for information fusion. Our resulting Multimodal Video Understanding (MVU) framework demonstrates state-of-the-art performance across long-video and fine-grained action recognition benchmarks. Code available at: https://github.com/kahnchana/mvu

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Zero-Shot Video Question Answer EgoSchema (fullset) MVU (13B) Accuracy 0.376 # 12
Zero-Shot Video Question Answer EgoSchema (subset) MVU (13B) Accuracy 60.3 # 2
Inference Speed (s) 2.42 # 1
Zero-Shot Video Question Answer NExT-QA MVU (13B) Accuracy 55.2 # 11

Methods


No methods listed for this paper. Add relevant methods here