VedioLLava / README.md
Mitzi4132's picture
Update README.md
ed96310 verified
metadata
configs:
  - config_name: light
    data_files: /light_yes.xlsx
    features:
      - name: question
        dtype: string
      - name: candidates
        dtype: string
      - name: answer
        dtype: string
      - name: video
        dtype: string
language:
  - en
size_categories:
  - 1K<n<10K

MVBench

Dataset Description

images

We introduce a novel static-to-dynamic method for defining temporal-related tasks. By converting static tasks into dynamic ones, we facilitate systematic generation of video tasks necessitating a wide range of temporal abilities, from perception to cognition. Guided by task definitions, we then automatically transform public video annotations into multiple-choice QA for task evaluation. This unique paradigm enables efficient creation of MVBench with minimal manual intervention while ensuring evaluation fairness through ground-truth video annotations and avoiding biased LLM scoring. The 20 temporal task examples are as follows.

images

Evaluation

An evaluation example is provided in mvbench.ipynb. Please follow the pipeline to prepare the evaluation code for various MLLMs.

  • Preprocess: We preserve the raw video (high resolution, long duration, etc.) along with corresponding annotations (start, end, subtitles, etc.) for future exploration; hence, the decoding of some raw videos like Perception Test may be slow.
  • Prompt: We explore effective system prompts to encourage better temporal reasoning in MLLM, as well as efficient answer prompts for option extraction.

Leadrboard

While an Online leaderboard is under construction, the current standings are as follows:

images