GunaKoppula commited on
Commit
deeed66
·
verified ·
1 Parent(s): 6837ed1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -17,7 +17,7 @@ This is a multimodal implementation of [Phi2](https://huggingface.co/microsoft/p
17
  2. Vision Tower: [clip-vit-large-patch14-336](https://huggingface.co/openai/clip-vit-large-patch14-336)
18
  4. Pretraining Dataset: [LAION-CC-SBU dataset with BLIP captions(200k samples)](https://huggingface.co/datasets/liuhaotian/LLaVA-Pretrain)
19
  5. Finetuning Dataset: [Instruct 150k dataset based on COCO](https://huggingface.co/datasets/liuhaotian/LLaVA-Instruct-150K)
20
- 6. Finetuned Model: [RaviNaik/Llava-Phi2](https://huggingface.co/RaviNaik/Llava-Phi2)
21
 
22
 
23
  ### Model Sources
@@ -46,8 +46,8 @@ pip install -e .
46
  ```
47
  3. Run the Model
48
  ```bash
49
- python llava_phi/eval/run_llava_phi.py --model-path="RaviNaik/Llava-Phi2" \
50
- --image-file="https://huggingface.co/RaviNaik/Llava-Phi2/resolve/main/people.jpg?download=true" \
51
  --query="How many people are there in the image?"
52
  ```
53
 
 
17
  2. Vision Tower: [clip-vit-large-patch14-336](https://huggingface.co/openai/clip-vit-large-patch14-336)
18
  4. Pretraining Dataset: [LAION-CC-SBU dataset with BLIP captions(200k samples)](https://huggingface.co/datasets/liuhaotian/LLaVA-Pretrain)
19
  5. Finetuning Dataset: [Instruct 150k dataset based on COCO](https://huggingface.co/datasets/liuhaotian/LLaVA-Instruct-150K)
20
+ 6. Finetuned Model: [RaviNaik/Llava-Phi2](https://huggingface.co/GunaKoppula/Llava-Phi2)
21
 
22
 
23
  ### Model Sources
 
46
  ```
47
  3. Run the Model
48
  ```bash
49
+ python llava_phi/eval/run_llava_phi.py --model-path="GunaKoppula/Llava-Phi2" \
50
+ --image-file="https://huggingface.co/GunaKoppula/Llava-Phi2/resolve/main/people.jpg?download=true" \
51
  --query="How many people are there in the image?"
52
  ```
53