Image-Text-to-Text
xtuner
LZHgrla commited on
Commit
3f2e96f
·
1 Parent(s): 018f102

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -17,7 +17,7 @@ pipeline_tag: visual-question-answering
17
 
18
  ## Model
19
 
20
- llava-internlm-chat-7b-clip-vit-large-p14-336 is a LLaVA model fine-tuned from [InternLM-Chat-7B](https://huggingface.co/internlm/internlm-chat-7b) and [CLIP-ViT-Large-patch14-336](https://huggingface.co/openai/clip-vit-large-patch14-336) with [LLaVA-Pretrain](https://huggingface.co/datasets/liuhaotian/LLaVA-Pretrain) and [LLaVA-Instruct](https://huggingface.co/datasets/liuhaotian/LLaVA-Instruct-150K) by [XTuner](https://github.com/InternLM/xtuner).
21
 
22
 
23
  ## Quickstart
 
17
 
18
  ## Model
19
 
20
+ **LLaVA-InternLM-7B** (llava-internlm-chat-7b-clip-vit-large-p14-336) is a LLaVA model fine-tuned from [InternLM-Chat-7B](https://huggingface.co/internlm/internlm-chat-7b) and [CLIP-ViT-Large-patch14-336](https://huggingface.co/openai/clip-vit-large-patch14-336) with [LLaVA-Pretrain](https://huggingface.co/datasets/liuhaotian/LLaVA-Pretrain) and [LLaVA-Instruct](https://huggingface.co/datasets/liuhaotian/LLaVA-Instruct-150K) by [XTuner](https://github.com/InternLM/xtuner).
21
 
22
 
23
  ## Quickstart