wangyueqian commited on
Commit
2366917
·
verified ·
1 Parent(s): b6413e7

add paper and video demo to REAME.md

Browse files
Files changed (1) hide show
  1. README.md +11 -3
README.md CHANGED
@@ -23,14 +23,22 @@ This is the model checkpoint of **MMDuet**, a VideoLLM you can interact with in
23
 
24
 
25
  ## Related Resources
 
26
  - **Github:** [MMDuet](https://github.com/yellow-binary-tree/MMDuet)
27
- - **Paper:** TODO
28
- - **Demo:** [Video Demo](https://www.youtube.com/watch?v=n1OybwhQvtk)
29
  - **Data:** [MMDuetIT](https://huggingface.co/datasets/wangyueqian/MMDuetIT)
30
 
31
 
32
  ## Citation
33
  If you use this work in your research, please consider cite:
34
  ```bibtex
35
-
 
 
 
 
 
 
 
 
36
  ```
 
23
 
24
 
25
  ## Related Resources
26
+ - **Paper:** [VideoLLM Knows When to Speak: Enhancing Time-Sensitive Video Comprehension with Video-Text Duet Interaction Format](https://arxiv.org/abs/2411.17991)
27
  - **Github:** [MMDuet](https://github.com/yellow-binary-tree/MMDuet)
28
+ - **Video Demo:** [On Youtube](https://www.youtube.com/watch?v=n1OybwhQvtk) and [On Bilibili](https://www.bilibili.com/video/BV1nwzGYBEPE)
 
29
  - **Data:** [MMDuetIT](https://huggingface.co/datasets/wangyueqian/MMDuetIT)
30
 
31
 
32
  ## Citation
33
  If you use this work in your research, please consider cite:
34
  ```bibtex
35
+ @misc{wang2024mmduet,
36
+ title={VideoLLM Knows When to Speak: Enhancing Time-Sensitive Video Comprehension with Video-Text Duet Interaction Format},
37
+ author={Yueqian Wang and Xiaojun Meng and Yuxuan Wang and Jianxin Liang and Jiansheng Wei and Huishuai Zhang and Dongyan Zhao},
38
+ year={2024},
39
+ eprint={2411.17991},
40
+ archivePrefix={arXiv},
41
+ primaryClass={cs.CV},
42
+ url={https://arxiv.org/abs/2411.17991},
43
+ }
44
  ```