Spaces:
Running
on
A10G
Running
on
A10G
舟勤
commited on
Commit
·
89361cc
1
Parent(s):
6370ed2
update
Browse files
app.py
CHANGED
@@ -157,7 +157,7 @@ title = """
|
|
157 |
|
158 |
Thank you for using the Video-LLaMA Demo Page! If you have any questions or feedback, feel free to contact us.
|
159 |
|
160 |
-
If you
|
161 |
|
162 |
Current online demo uses the 7B version of Video-LLaMA due to resource limitations. We have released \
|
163 |
the 13B version on our GitHub repository.
|
@@ -173,6 +173,19 @@ The output results may be influenced by input quality, limitations of the datase
|
|
173 |
**Copyright 2023 Alibaba DAMO Academy.**
|
174 |
""")
|
175 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
176 |
#TODO show examples below
|
177 |
|
178 |
with gr.Blocks() as demo:
|
@@ -225,7 +238,8 @@ with gr.Blocks() as demo:
|
|
225 |
[f"examples/birthday.mp4", "What is the boy doing? "],
|
226 |
[f"examples/Iron_Man.mp4", "Is the guy in the video Iron Man? "],
|
227 |
], inputs=[video, text_input])
|
228 |
-
|
|
|
229 |
upload_button.click(upload_imgorvideo, [video, image, text_input, chat_state], [video, image, text_input, upload_button, chat_state, img_list])
|
230 |
|
231 |
text_input.submit(gradio_ask, [text_input, chatbot, chat_state], [text_input, chatbot, chat_state]).then(
|
|
|
157 |
|
158 |
Thank you for using the Video-LLaMA Demo Page! If you have any questions or feedback, feel free to contact us.
|
159 |
|
160 |
+
If you find Video-LLaMA interesting, please give us a star on GitHub.
|
161 |
|
162 |
Current online demo uses the 7B version of Video-LLaMA due to resource limitations. We have released \
|
163 |
the 13B version on our GitHub repository.
|
|
|
173 |
**Copyright 2023 Alibaba DAMO Academy.**
|
174 |
""")
|
175 |
|
176 |
+
cite_markdown = ("""
|
177 |
+
## Citation
|
178 |
+
If you find our project useful, hope you can star our repo and cite our paper as follows:
|
179 |
+
```
|
180 |
+
@article{damonlpsg2023videollama,
|
181 |
+
author = {Zhang, Hang and Li, Xin and Bing, Lidong},
|
182 |
+
title = {Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding},
|
183 |
+
year = 2023,
|
184 |
+
journal = {arXiv preprint arXiv:2306.02858}
|
185 |
+
url = {https://arxiv.org/abs/2306.02858}
|
186 |
+
}
|
187 |
+
""")
|
188 |
+
|
189 |
#TODO show examples below
|
190 |
|
191 |
with gr.Blocks() as demo:
|
|
|
238 |
[f"examples/birthday.mp4", "What is the boy doing? "],
|
239 |
[f"examples/Iron_Man.mp4", "Is the guy in the video Iron Man? "],
|
240 |
], inputs=[video, text_input])
|
241 |
+
|
242 |
+
gr.Markdown(cite_markdown)
|
243 |
upload_button.click(upload_imgorvideo, [video, image, text_input, chat_state], [video, image, text_input, upload_button, chat_state, img_list])
|
244 |
|
245 |
text_input.submit(gradio_ask, [text_input, chatbot, chat_state], [text_input, chatbot, chat_state]).then(
|