skip to content

Tags #video

  • Overview

    Our contributions are threefold:

    (1) LongVT: An End-to-End Agentic Framework for “Thinking with Long Videos”
    We introduce a novel paradigm that natively interleaves multimodal tool-augmented Chain-of-Thought (CoT) with on-demand clip inspection over hours-long videos, thereby enabling large multimodal models (LMMs) to perform more effective and reliable long-video reasoning.

    (2) VideoSIAH: A Fine-Grained Data Suite for Evidence-Sparse Long-Video Reasoning
    We construct a scalable data pipeline that produces diverse and high-quality question-answering (QA) data and tool-integrated reasoning traces, and a dedicated benchmark under a video segment-in-a-haystack setting.

    (3) LongVT-7B-RFT: A State-of-the-Art Baseline with Invaluable Insights
    Through extensive quantitative comparisons, systematic ablations on data recipes, training strategies, and design choices, as well as in-depth analyses of training dynamics, we establish and open-source a powerful baseline model with “thinking with long videos” capabilities.

    LongVT Interleaved Multimodal Chain-of-Tool-Thought

    Interleaved Multimodal Chain-of-Tool-Thought (iMCoTT). Compared to prior text-based CoT reasoning, iMCoTT in our proposed LongVT can natively perform self-reflection via calling crop_video(start_time, end_time) tool. It proposes a time window after a global preview, proactively fetches the corresponding short clip, rethinks based on the new evidence, and determines whether to refine or answer directly. Such tool-augmented reasoning behaviors ground each step in what is actually seen rather than blindly rephrasing in text-only CoT, which mitigates hallucination and leads to enhanced temporal localization and answer correctness.


    Motivation of VideoSIAH

    Long-video reasoning presents a fundamentally different challenge from previous video QA settings: LMMs must locate sparse, fine-grained, and causally decisive moments embedded within hours-long content. However, existing LMMs are mostly trained with coarse-grained and clip-level data. This mismatch leaves modern LMMs lacking the supervision needed to learn how temporal hypotheses are formed, verified, or revised—a critical yet underexplored capability for agentic long-video reasoning.

    Moreover, most existing video understanding benchmarks only offer multiple-choice QAs, which can be solved without genuine temporal grounding and are vulnerable to dataset leakage or shortcut exploitation. To fill this gap, we introduce VideoSIAH, a large-scale, diverse, and high-quality data suite that serves collectively as a training dataset capturing the reasoning dynamics required for video segment-in-a-haystack QA, and a fine-grained evaluation benchmark, VideoSIAH-Eval, with human-in-the-loop validation for long-video open-ended question-answering.

    We conduct a rigorous contamination study on the Qwen-VL series across two probing settings: (1) No Visual, where we feed the text prompt without video frames to test for direct memorization; (2) Rearranged Choices, where we randomize the mapping between option labels and their textual content for multiple-choice questions to detect label memorization. Our experimental results reveal significant vulnerabilities in existing benchmarks and highlight the necessity of our proposed VideoSIAH-Eval.

    SettingVideoMME (w/o sub)VideoMMMU adapt.VideoMMMU comp.VideoMMMU perc.VideoSIAH-Eval
    Qwen2.5-VL-7B-Instruct
    Original64.335.744.356.733.8
    No Visual40.125.738.339.312.7
    Rearranged Choices56.029.740.367.0-
    Qwen3-VL-8B-Instruct
    Original69.340.760.371.346.6
    No Visual44.133.739.346.70.00
    Rearranged Choices69.036.347.769.3-

    Contamination Tests for Qwen-VL Series on Long Video Understanding and Reasoning Benchmarks. The VideoSIAH-Eval column shows ”-” entries for Rearranged Choices since our proposed benchmark is fully open-ended QA, where random option-answer mapping is not applicable.


    Data Pipeline

    VideoSIAH Data Pipeline

    Data Pipeline of VideoSIAH. We construct a semi-automatic data pipeline that integrates several state-of-the-art LMMs to sequentially perform long video segmentation, video clip captioning, segment-in-a-haystack QA generation, cross-modal QA filtering, and iMCoTT generation. Icons with human silhouettes denote human-in-the-loop validation, where annotators inspect a small set of representative failures to refine prompting rules for QA generation, QA filtering, and iMCoTT generation. Note that iMCoTT traces are generated only for the cold-start supervised fine-tuning (SFT) stage, whereas reinforcement learning (RL) operates solely on the filtered QA pairs.


    Dataset Statistics

    SplitSourcePurposeSamplesTotal
    SFT (w/o tool)LongVideo-Reason CoTReasoning-augmented Open-ended QA5,238228,835
    Video-R1 CoTReasoning-augmented Video QA165,575
    Image-based CoTReasoning-augmented Image QA58,022
    SFT (w/ tool)Gemini-distilled iMCoTTTool-augmented Open-ended QA12,76619,161
    Qwen-distilled iMCoTTTool-augmented Temporal Grounding6,395
    RLGemini-distilled QAsOpen-ended QA over Long Videos1,66717,020
    RFTSelf-distilled iMCoTTAgentic Behaviors15,353

    Dataset Statistics of VideoSIAH. Our proposed dataset contains a large-scale of non-tool SFT data, tool-augmented SFT data, RL QAs, and self-distilled reinforcement fine-tuning (RFT) traces.

    Video Category Distribution
    Question Category Distribution

    Category Distribution of VideoSIAH-Eval. We present the distribution of video types (left) and question types (right), highlighting the diversity of our proposed benchmark.


    Quantitative Comparisons

    We compare our LongVT models against proprietary LMMs and state-of-the-art open-source video reasoning models across various long video understanding and reasoning benchmarks.

    ModelReasoningToolVideoMMEVideoMMMULVBenchVideoSIAH-EvalAvg
    PromptCallingw/ subadapt.comp.perc.
    Proprietary LMMs
    GPT-4o77.266.062.055.730.817.451.5
    Gemini 1.5 Pro81.359.053.349.333.1-55.2
    Open-Source (Sparse Sampling)
    Qwen2.5-VL-7B62.637.328.036.730.728.137.2
    Video-R1-7B61.036.340.752.337.227.942.6
    VideoRFT-7B60.936.742.053.034.726.542.3
    Video-Thinker-7B61.034.344.753.052.210.442.6
    LongVT-7B-SFT (Ours)12.537.746.058.336.026.836.2
    LongVT-7B-RL (Ours)66.132.744.750.037.831.043.7
    Open-Source (Dense Sampling)
    Qwen2.5-VL-7B64.335.744.356.740.933.846.0
    Video-R1-7B60.537.338.746.340.133.142.7
    VideoRFT-7B49.237.740.748.718.726.937.0
    Video-Thinker-7B60.837.742.755.354.36.642.9
    LongVT-7B-SFT (Ours)64.932.342.049.741.134.844.1
    LongVT-7B-RL (Ours)66.137.742.356.341.435.946.6
    LongVT-7B-RFT (Ours)67.035.743.756.741.342.047.7

    Performance Comparison with Existing Video-Centric LMMs across Various Long Video Understanding and Reasoning Benchmarks. The best and second-best results among open-source models in each column are marked in bold and underlined, respectively.


    Ablation Studies

    We conduct comprehensive ablation studies to examine the impact of data recipes, training stages, and reward design on model performance.

    Data Recipe

    SettingVideoMMEVideoMMMULVBenchVideoSIAH-EvalAvg
    w/ subadapt.comp.perc.
    SFT w/o self-curated iMCoTT8.433.641.646.015.14.124.8
    SFT w/ self-curated iMCoTT64.932.342.049.741.134.844.1
    RL w/o self-curated QAs55.130.642.045.638.430.840.4
    RL w/ self-curated QAs66.137.742.356.341.435.946.6

    Training Stage

    SettingVideoMMEVideoMMMULVBenchVideoSIAH-EvalAvg
    w/ subadapt.comp.perc.
    SFT only64.932.342.049.741.134.844.1
    RL only52.735.343.055.137.128.241.9
    SFT+RL66.137.742.356.341.435.946.6
    SFT+RL+RFT67.035.743.756.741.342.047.7

    Training Dynamics

    Training Dynamics and Ablations on Reward Design

    (a) shows training dynamics under different accuracy and time rewards, and (b) shows the effect of tool-call reward on tool usage.

    Recall encourages coverage; IoU demands precision. Using Recall as the reward function during RL presents a drawback: the policy can enlarge the predicted span to envelop the ground-truth interval, which monotonically raises the Recall-based score while ignoring boundary quality. This plateau in the curve of Recall Accuracy Score validates our hypothesized reward hacking. In contrast, IoU explicitly penalizes span inflation via the union term, yielding better-aligned boundaries and more disciplined tool use.

    Is tool reward really necessary? The Qwen2.5-VL-7B baseline collapses to near-zero tool calls after training in both configurations (w/ and w/o tool reward), indicating that the model does not internalize the tool’s function. After performing cold-start SFT to obtain LongVT-7B-SFT, tool-call frequency rises during training under both configurations and accuracy improves in tandem. Hence, the tool reward is not required for basic competence: once SFT grounds the tool’s semantics, the model learns when and how to invoke the tool.


    Open-Source Resources
    We open-source LongVT to facilitate future development of long-video reasoning with tool calling in the community
    Model Checkpoints
    Pre-trained models with SFT, RL, and RFT optimization
    Training Datasets
    VideoSIAH data suite for long-video reasoning
  • Video-MMMU Overview
    Video-MMMU: A comprehensive benchmark for evaluating knowledge acquisition from educational videos across multiple disciplines
    WebsitePaperDataset

    Video-MMMU asks a fundamental question: If a model ‘goes to class,’ can the model learn from the lecture and apply what it learned to MMMU-style exam problems?

    🎯 Motivation

    Our original goal was to build a video reasoning benchmark, motivated by the observation that the most demanding forms of reasoning arise in academic settings—for example, MMMU-style university exam questions.

    Online lectures create an ideal environment for evaluating video reasoning. They effectively convey knowledge and naturally test a model’s ability to learn from video. These videos have three key attributes:

    1. High information density (heavy OCR/ASR signals)
    2. Advanced knowledge requirements (college-level knowledge)
    3. Temporal structure (concepts unfolding over time)

    These properties make reasoning from lecture video notably harder. This leads to our core question:

    When a model watches an online lecture, can it learn like a student—understand the content, acquire the knowledge, and then solve related problems?

    Therefore, we introduce Video-MMMU, a video reasoning benchmark that evaluates knowledge acquisition from video.

    🏆 Video-MMMU Leaderboard

    ModelOverall \ΔknowledgePerceptionComprehensionAdaptation
    GPT-5-thinking84.6 \
    Gemini-2.5-Pro83.6 \
    OpenAI O383.3 \
    Claude-3.5-Sonnet65.78 \🟢 +11.472.0069.6755.67
    Kimi-VL-A3B-Thinking-250665.22 \🟢 +3.575.0066.3354.33
    GPT-4o61.22 \🟢 +15.666.0062.0055.67
    Qwen-2.5-VL-72B60.22 \🟢 +9.769.3361.0050.33

    See full leaderboard with 20+ models in our paper and website

    📚 Overview

    We introduce Video-MMMU, a multi-modal, multi-disciplinary, multi-track benchmark designed to evaluate how effectively large multimodal models (LMMs) acquire knowledge from educational videos.

    1) Video: Knowledge Source

    Traditional VideoQA benchmarks focus on scene understanding. Video-MMMU treats video as a source of knowledge, evaluating whether LMMs can actually learn from instructional content.

    Dataset Composition:

    • 300 college-level, lecture-style videos
    • 30 subjects across 6 disciplines: Art, Business, Science, Medicine, Humanities, and Engineering
    • High-quality educational content from university-level courses

    2) QA Design: Three Stages of Knowledge Acquisition

    Each video is paired with three questions, designed to reflect a progression in knowledge acquisition:

    • 🔍 Perception – Identifying relevant surface information
    • 🧠 Comprehension – Understanding underlying concepts or strategies
    • 🎯 Adaptation – Applying learned knowledge to new scenarios
    Knowledge Acquisition Categories
    Figure 2: Examples for each knowledge acquisition category across different disciplines. Perception (ASR/OCR-based), Comprehension (concept/strategy understanding), and Adaptation (application to new scenarios).
    Benchmark Structure
    Figure 3: Video-MMMU benchmark structure showing the progression from video content to three-tier evaluation framework.

    3) In-Context Knowledge Acquisition: Learning Like Humans

    Humans consistently learn from the world around them. For models to operate effectively in real-world environments, the same principle should apply: they must be able to learn from the world, because unlike humans, they cannot be endlessly re-trained after deployment.

    In this sense, videos provide a natural proxy for the world. For a model, the video becomes its world. The ability to learn from video therefore becomes more than a technical benchmark—it is a measure of true, dynamic intelligence. It marks the shift from simply solving a task to demonstrating the ability to learn how to solve the task.

    4) Metric: From Absolute Accuracy to Learning Efficiency (Δknowledge)

    A core innovation in Video-MMMU is its shift from measuring only final performance to measuring learning.

    Δknowledge Formula

    Δknowledge = (Acc_after_video - Acc_before_video) / (100% - Acc_before_video) × 100%
    

    Evaluation Process

    1. Initial Test: The model attempts to answer a question without seeing the video.

    2. Re-Test after video viewing: We provide the corresponding lecture video. The model is asked the same question again.

    3. Performance Gain: If the model succeeds after watching, it demonstrates successful knowledge acquisition from video.

    This setup mirrors a human’s natural educational process:

    Don't know → Learn by watching → Apply the knowledge
    

    🔍 Key Insights

    Performance Analysis
    Figure 4: Comprehensive analysis showing progressive performance decline and the human-model gap in knowledge acquisition from videos.

    Progressive Performance Decline

    Model performance decreases as cognitive demands increase. While models perform relatively better on Perception tasks, accuracy drops on Comprehension and declines further on Adaptation.

    Knowledge Acquisition Challenge

    The Δknowledge metric reveals a significant human–model gap:

    • Humans: Substantial improvement (Δknowledge ≈ 33.1%)
    • Top Models: Smaller gains (GPT-4o: 15.6%, Claude-3.5-Sonnet: 11.4%)

    This highlights a current limitation: LMMs still struggle to learn from videos in the way humans do.

    📊 Case Studies

    Failure Case: Method Adaptation Error

    Failure Case Study
    Figure 5: Example of method adaptation failure - the model failed to adapt the method from video to solve the Adaptation question.

    Success Case: Learning from Video

    Success Case Study
    Figure 6: Example of successful learning from video - transforming an initial wrong answer into a correct one after watching the educational content.

    🚀 Research Impact

    Paradigm Shift

    Video-MMMU represents a paradigm shift from traditional video understanding to knowledge acquisition evaluation:

    • From Scene Understanding to Learning - Moving beyond visual comprehension to knowledge acquisition
    • From Static Evaluation to Dynamic Learning - Measuring improvement rather than just final performance
    • From Task Solving to Learning Capability - Evaluating the ability to learn new skills

    Implications for AI Development

    1. Real-World Deployment - Models must learn continuously after deployment
    2. Educational AI - Critical for AI tutoring and educational applications
    3. Knowledge Transfer - Understanding how models generalize learned concepts
    4. Human-AI Alignment - Bridging the gap in learning capabilities

    📈 Future Directions

    Benchmark Extensions

    • Multimodal Knowledge Sources - Incorporating diverse educational formats
    • Long-term Learning - Evaluating knowledge retention over time
    • Interactive Learning - Adding feedback loops and iterative improvement

    Model Development

    • Learning-Optimized Architectures - Designing models specifically for knowledge acquisition
    • Memory Integration - Better mechanisms for knowledge storage and retrieval
    • Transfer Learning - Improving cross-domain knowledge application

    🎯 Getting Started

    1. Download the Video-MMMU dataset from Hugging Face
    2. Set up the evaluation environment using our GitHub repository
    3. Run baseline evaluations on your models
    4. Analyze Δknowledge metrics to understand learning capabilities
    5. Compare results with our comprehensive leaderboard

    Video-MMMU challenges the current state of multimodal AI by shifting focus from static performance to dynamic learning capability - a critical step toward truly intelligent and adaptive AI systems.

  • The development of video large multimodal models (LMMs) has been hindered by the difficulty of curating large amounts of high-quality raw data from the web. To address this, we consider an alternative approach, creating a high-quality synthetic dataset specifically for video instruction-following, namely LLaVA-Video-178K. This dataset includes key tasks such as detailed captioning, open-ended question-answering (QA), and multiple-choice QA. By training on this proposed dataset, in combination with existing visual instruction tuning data, we introduce LLaVA-Video, a new video LMM. Our experiments demonstrate that LLaVA-Video achieves strong performance across various video benchmarks, highlighting the effectiveness of our dataset. We plan to release the dataset, its generation pipeline, and the model checkpoints.

    Video Instruction-Following Data Synthesis

    A high-quality dataset for video instruction-tuning is crucial for developing effective video-language models. We identify a key factor in building such datasets: ensuring richness and diversity in both video content and its language annotations. We perform comprehensive survey on the existing video benchmarks, covering across various public video captioning and question-answering datasets, then identify ten unique video sources that contribute to over 40 video-language benchmarks. From each source, we select videos that exhibit significant temporal dynamics. To maintain diversity in the annotations, we establish a pipeline capable of generating detailed captions for videos of any length. Additionally, we define 16 types of questions that guide GPT-4o in creating question-answer pairs to assess the perceptual and reasoning skills of the video-language models.

    Video Sources

    Video Sources for LLaVA-Video
    Video sources in the proposed LLaVA-Video-178K. The relationship between 10 video sources we have utilized and other existing video-language datasets.

    We noticed that although different video-language datasets focus on various video understanding tasks , most are sourced from ten main video sources, which offer a wide range of video data from different websites, viewpoints, and domains. The relationship between these ten selected video datasets and others is shown in figure below. We select the dynamic video from these source, we detail the video selection logic in the paper.

    Automated Generation for Video Detail Description

    LLaVA-Video Data Creation
    The video detail description creation pipeline. A three-level creation pipeline is considered, with each level developed via a recurrent approach. Note that t is the index of time internal at its own level, and T is the last time internal index

    For selected videos, we use GPT-4o to systematically describe their content. We start by sampling video frames at one frame per second (fps). However, due to the input size constraints of GPT-4o, we cannot use all sampled frames. Instead, we describe the videos sequentially, as shown in figure below. We create descriptions at three distinct levels, detailed below.

    Automated Generation for Video Question Answering

    In addition to detailed video descriptions, our dataset includes a variety of question-answer pairs designed for complex interactions. This setup improves the video understanding model’s ability to handle real-life queries. We refer to public video question-answering benchmarks to organize these questions into 16 specific categories, as shown in Figure 3. Given a detailed video description, we use GPT-4o to generate at most one question-answer pair for each type of question. Please refer to the paper for more details of the question types and the generation process.

    Dataset Statistics

    We carefully select from our collected data sources to form a balanced and comprehensive collection, resulting in a total of 178K videos and 1.3M instruction-following samples. This includes 178K captions, 960K open-ended QAs, and 196K multiple-choice QAs.

    Dataset Comparison

    We provide a comparison of high-quality instruction-following video-language datasets, with a focus on synthetic data created with strong AI models, as shown in Table 1.

    A broad collection of dynamic videos. In terms of video sources, although LLaVA-Hound contains the largest number of videos, 44% of its video data are sourced from WebVid, where most videos are static. ShareGPT4Video includes 30% of its videos from Pexels, ,Pixabay, and Mixkit, which are aesthetically good but also mostly static. Additionally, the majority of its videos come from Panda-70M, which are short clips from longer videos, suggesting simpler plots. In contrast, we carefully select video sources that offer dynamic, untrimmed videos with complex plots, which are crucial for developing a powerful video understanding model.

    High frames per second. Regarding frame sampling in language annotations, the proposed dataset considers 1 FPS, while other datasets consider much lower FPS. LLaVA-Hound uniformly samples 10 frames from videos of any length. The average FPS is 0.008, which may miss some fine details. ShareGPT4Video picks key frames using CLIP based on frame uniqueness. This method might also miss subtle changes in the video because CLIP embeddings do not capture fine-grained dynamics well. Our method samples FPS=1 without using key frame selection algorithms, ensuring that detailed temporal information can be expressed in annotations with high coverage.

    Diverse tasks. The proposed dataset considers three common task types, including caption, free-form, and closed-form QA, while existing datasets only consider a subset. Meanwhile, the quality and number of samples in our dataset is higher.

  • LLaVA-OneVision
    LLaVA-OneVision: A unified model for single-image, multi-image, and video understanding

    Overview

    We present LLaVA-OneVision, a family of open large multimodal models (LMMs) developed by consolidating our insights into data, models, and visual representations in the LLaVA-NeXT blog series. LLaVA-OneVision is the first single model that can simultaneously push the performance boundaries of open LMMs in three important computer vision scenarios: single-image, multi-image, and video scenarios.

    Key Features

    Unified Architecture

    LLaVA-OneVision is designed to have a similar maximum visual token count across different scenarios, enabling flexible extension to multiple visual signal types while maintaining consistent performance.

    Model Sizes

    • 0.5B parameters - Lightweight deployment
    • 7B parameters - Balanced performance
    • 72B parameters - State-of-the-art capabilities

    Emerging Capabilities

    The design of LLaVA-OneVision enables strong transfer learning across different modalities and scenarios, yielding impressive emerging capabilities:

    1. Cross-Scenario Understanding

    Seamlessly process and understand content across single images, multiple images, and videos within a unified framework.

    2. Advanced Visual Analysis

    • Diagram and table interpretation - Understanding complex visual structures
    • Multi-screenshot interaction - Analyzing relationships across multiple screens
    • Set-of-mark object referencing - Precise object identification and tracking

    3. Video Capabilities

    • Image-to-video generation understanding - Comprehending temporal transitions
    • Video analysis and comparison - Deep understanding of video content
    • Multi-camera video interpretation - Processing footage from multiple viewpoints
    • Detailed video subject description - Rich, contextual video narration

    Strong Transfer Learning

    Importantly, the design of LLaVA-OneVision allows strong transfer learning across different modalities/scenarios. In particular, strong video understanding and cross-scenario capabilities are demonstrated through task transfer from images to videos, showcasing the model’s ability to generalize learned representations across visual domains.

    Development Roadmap

    LLaVA-OneVision represents a significant milestone in our iterative improvements through the LLaVA-NeXT series, focusing on:

    • Enhanced reasoning capabilities
    • Improved OCR performance
    • Expanded world knowledge
    • Advanced multimodal understanding
  • LongVA Visual Needle-in-a-Haystack Heatmap
    LongVA's performance on Visual Needle-In-A-Haystack benchmark showing accurate retrieval across long video sequences

    Overview

    Gemini has amazed the world with its capability to understand hour-long videos. However, we still lack an open-source alternative with similar capabilities. Our latest research presents an innovative solution towards long video LMM, shifting the focus from reducing visual tokens per frame to leveraging the long context capabilities of language models.

    Here, we present our state-of-the-art video model, Long Video Assistant (LongVA), and our novel benchmark, Visual Needle-In-A-Haystack (V-NIAH).

    Key Innovations

    🔄 Long Context Transfer

    We discovered and verified that the long context capability of language models can be directly transferred to the video domain in modality-aligned multi-modal models. On V-NIAH, LongVA is the only open-source model capable of accurately retrieving visual information from inputs with:

    • 2000+ frames
    • 200K+ visual tokens

    🎯 UniRes: Unified Visual Encoding

    We proposed UniRes, a unified visual encoding scheme that encodes both images and videos. In UniRes, a video is encoded the same as multiple image crops in a sequence.

    Key Benefits:

    • Leverages the Long Context Transfer property
    • Enables superior zero-shot performance in video tasks
    • No video-specific training data required

    Performance Highlights

    🏆 State-of-the-Art Results

    LongVA achieves state-of-the-art performance on the comprehensive Video-MME benchmarks among 7B models.

    Key Performance Features:

    • Performance increases with denser sampling of video frames
    • Superior zero-shot capabilities on video understanding tasks
    • Comprehensive ablation studies validating improvement sources

    📊 V-NIAH Benchmark

    Our novel Visual Needle-In-A-Haystack (V-NIAH) benchmark provides:

    • Rigorous evaluation of long-context visual understanding
    • Testing retrieval accuracy across extended video sequences
    • Open-source evaluation framework for the community

    Technical Architecture

    Multi-Modal Alignment

    LongVA demonstrates that language models’ inherent long-context capabilities can be effectively transferred to visual domains through proper modality alignment.

    Scalable Design

    The architecture scales efficiently with:

    • Increased frame sampling rates
    • Extended sequence lengths
    • Larger visual token counts

    Research Impact

    Open-Source Alternative

    LongVA provides the first viable open-source alternative to proprietary long-video understanding systems, enabling:

    • Academic research advancement
    • Commercial application development
    • Community-driven improvements

    Methodology Innovation

    The long context transfer approach opens new research directions in:

    • Cross-modal capability transfer
    • Efficient video processing
    • Unified multi-modal architectures

    Future Directions

    LongVA establishes a foundation for:

    1. Extended Context Models - Pushing beyond current frame limits
    2. Multi-Modal Transfer Learning - Applying insights to other modalities
    3. Efficient Video Processing - Optimizing computational requirements
    4. Benchmark Development - Creating more comprehensive evaluation metrics
    LongVA Resources
    Complete resources for LongVA including source code, evaluation benchmark, and pre-trained models