GunaKoppula
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -17,7 +17,7 @@ This is a multimodal implementation of [Phi2](https://huggingface.co/microsoft/p
|
|
17 |
2. Vision Tower: [clip-vit-large-patch14-336](https://huggingface.co/openai/clip-vit-large-patch14-336)
|
18 |
4. Pretraining Dataset: [LAION-CC-SBU dataset with BLIP captions(200k samples)](https://huggingface.co/datasets/liuhaotian/LLaVA-Pretrain)
|
19 |
5. Finetuning Dataset: [Instruct 150k dataset based on COCO](https://huggingface.co/datasets/liuhaotian/LLaVA-Instruct-150K)
|
20 |
-
6. Finetuned Model: [RaviNaik/Llava-Phi2](https://huggingface.co/
|
21 |
|
22 |
|
23 |
### Model Sources
|
@@ -46,8 +46,8 @@ pip install -e .
|
|
46 |
```
|
47 |
3. Run the Model
|
48 |
```bash
|
49 |
-
python llava_phi/eval/run_llava_phi.py --model-path="
|
50 |
-
--image-file="https://huggingface.co/
|
51 |
--query="How many people are there in the image?"
|
52 |
```
|
53 |
|
|
|
17 |
2. Vision Tower: [clip-vit-large-patch14-336](https://huggingface.co/openai/clip-vit-large-patch14-336)
|
18 |
4. Pretraining Dataset: [LAION-CC-SBU dataset with BLIP captions(200k samples)](https://huggingface.co/datasets/liuhaotian/LLaVA-Pretrain)
|
19 |
5. Finetuning Dataset: [Instruct 150k dataset based on COCO](https://huggingface.co/datasets/liuhaotian/LLaVA-Instruct-150K)
|
20 |
+
6. Finetuned Model: [RaviNaik/Llava-Phi2](https://huggingface.co/GunaKoppula/Llava-Phi2)
|
21 |
|
22 |
|
23 |
### Model Sources
|
|
|
46 |
```
|
47 |
3. Run the Model
|
48 |
```bash
|
49 |
+
python llava_phi/eval/run_llava_phi.py --model-path="GunaKoppula/Llava-Phi2" \
|
50 |
+
--image-file="https://huggingface.co/GunaKoppula/Llava-Phi2/resolve/main/people.jpg?download=true" \
|
51 |
--query="How many people are there in the image?"
|
52 |
```
|
53 |
|