prince-canuma commited on
Commit
6e2bc13
1 Parent(s): 9521592

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -7,11 +7,12 @@ tags:
7
  - multimodal
8
  - qwen
9
  - mlx
 
10
  ---
11
 
12
  # mlx-community/nanoLLaVA-1.5-8bit
13
- This model was converted to MLX format from [`qnguyen3/nanoLLaVA-1.5`]() using mlx-vlm version **0.0.11**.
14
- Refer to the [original model card](https://huggingface.co/qnguyen3/nanoLLaVA-1.5) for more details on the model.
15
  ## Use with mlx
16
 
17
  ```bash
 
7
  - multimodal
8
  - qwen
9
  - mlx
10
+ - mlx
11
  ---
12
 
13
  # mlx-community/nanoLLaVA-1.5-8bit
14
+ This model was converted to MLX format from [`mlx-community/nanoLLaVA-1.5-bf16`]() using mlx-vlm version **0.1.6**.
15
+ Refer to the [original model card](https://huggingface.co/mlx-community/nanoLLaVA-1.5-bf16) for more details on the model.
16
  ## Use with mlx
17
 
18
  ```bash
config.json CHANGED
@@ -21,48 +21,6 @@
21
  "mm_hidden_size": 1152,
22
  "mm_projector_lr": null,
23
  "mm_projector_type": "mlp2x_gelu",
24
- "vision_config": {
25
- "return_dict": true,
26
- "tie_word_embeddings": true,
27
- "max_length": 20,
28
- "min_length": 0,
29
- "do_sample": false,
30
- "early_stopping": false,
31
- "num_beams": 1,
32
- "num_beam_groups": 1,
33
- "diversity_penalty": 0.0,
34
- "temperature": 1.0,
35
- "model_type": "siglip_vision_model",
36
- "hidden_size": 1152,
37
- "intermediate_size": 4304,
38
- "num_hidden_layers": 27,
39
- "num_attention_heads": 16,
40
- "num_channels": 3,
41
- "patch_size": 14,
42
- "image_size": 384,
43
- "attention_dropout": 0.0,
44
- "layer_norm_eps": 1e-06,
45
- "hidden_act": "gelu_pytorch_tanh"
46
- },
47
- "text_config": {
48
- "vocab_size": 151936,
49
- "max_position_embeddings": 32768,
50
- "hidden_size": 1024,
51
- "intermediate_size": 2816,
52
- "num_hidden_layers": 24,
53
- "num_attention_heads": 16,
54
- "use_sliding_window": false,
55
- "sliding_window": "None",
56
- "max_window_layers": 21,
57
- "num_key_value_heads": 16,
58
- "hidden_act": "silu",
59
- "initializer_range": 0.02,
60
- "rms_norm_eps": 1e-06,
61
- "use_cache": false,
62
- "rope_theta": 1000000.0,
63
- "attention_dropout": 0.0,
64
- "model_type": "qwen2"
65
- },
66
  "model_type": "llava-qwen2",
67
  "num_attention_heads": 16,
68
  "num_hidden_layers": 24,
@@ -74,14 +32,56 @@
74
  "rms_norm_eps": 1e-06,
75
  "rope_theta": 1000000.0,
76
  "sliding_window": 4096,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
77
  "tie_word_embeddings": false,
78
  "tokenizer_model_max_length": 4096,
79
  "tokenizer_padding_side": "right",
80
  "torch_dtype": "bfloat16",
81
- "transformers_version": "4.39.2",
82
  "tune_mm_mlp_adapter": false,
83
- "use_cache": false,
84
  "use_mm_proj": true,
85
  "use_sliding_window": false,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
86
  "vocab_size": 151936
87
  }
 
21
  "mm_hidden_size": 1152,
22
  "mm_projector_lr": null,
23
  "mm_projector_type": "mlp2x_gelu",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  "model_type": "llava-qwen2",
25
  "num_attention_heads": 16,
26
  "num_hidden_layers": 24,
 
32
  "rms_norm_eps": 1e-06,
33
  "rope_theta": 1000000.0,
34
  "sliding_window": 4096,
35
+ "text_config": {
36
+ "vocab_size": 151936,
37
+ "max_position_embeddings": 32768,
38
+ "hidden_size": 1024,
39
+ "intermediate_size": 2816,
40
+ "num_hidden_layers": 24,
41
+ "num_attention_heads": 16,
42
+ "use_sliding_window": false,
43
+ "sliding_window": "None",
44
+ "max_window_layers": 21,
45
+ "num_key_value_heads": 16,
46
+ "hidden_act": "silu",
47
+ "initializer_range": 0.02,
48
+ "rms_norm_eps": 1e-06,
49
+ "use_cache": false,
50
+ "rope_theta": 1000000.0,
51
+ "attention_dropout": 0.0,
52
+ "model_type": "qwen2"
53
+ },
54
  "tie_word_embeddings": false,
55
  "tokenizer_model_max_length": 4096,
56
  "tokenizer_padding_side": "right",
57
  "torch_dtype": "bfloat16",
58
+ "transformers_version": "4.41.1",
59
  "tune_mm_mlp_adapter": false,
60
+ "use_cache": true,
61
  "use_mm_proj": true,
62
  "use_sliding_window": false,
63
+ "vision_config": {
64
+ "return_dict": true,
65
+ "tie_word_embeddings": true,
66
+ "max_length": 20,
67
+ "min_length": 0,
68
+ "do_sample": false,
69
+ "early_stopping": false,
70
+ "num_beams": 1,
71
+ "num_beam_groups": 1,
72
+ "diversity_penalty": 0.0,
73
+ "temperature": 1.0,
74
+ "model_type": "siglip_vision_model",
75
+ "hidden_size": 1152,
76
+ "intermediate_size": 4352,
77
+ "num_hidden_layers": 27,
78
+ "num_attention_heads": 16,
79
+ "num_channels": 3,
80
+ "patch_size": 14,
81
+ "image_size": 384,
82
+ "attention_dropout": 0.0,
83
+ "layer_norm_eps": 1e-06,
84
+ "hidden_act": "gelu_pytorch_tanh"
85
+ },
86
  "vocab_size": 151936
87
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:76cae428eb06e3ef36ddc9ebddee0c01a2030b1d1dd7f00e84a117068a3f959f
3
- size 1138138539
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0a1d1517ad9e6e810bc6ac99956643c66e4b87a2f82bd5d1b5cb0966e5c5476
3
+ size 1120283029
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 1137941384
4
  },
5
  "weight_map": {
6
  "language_model.model.embed_tokens.biases": "model.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 1120085896
4
  },
5
  "weight_map": {
6
  "language_model.model.embed_tokens.biases": "model.safetensors",
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -35,6 +35,7 @@
35
  "clean_up_tokenization_spaces": false,
36
  "eos_token": "<|im_end|>",
37
  "errors": "replace",
 
38
  "model_max_length": 4096,
39
  "pad_token": "<|endoftext|>",
40
  "padding_side": "right",
 
35
  "clean_up_tokenization_spaces": false,
36
  "eos_token": "<|im_end|>",
37
  "errors": "replace",
38
+ "extra_special_tokens": {},
39
  "model_max_length": 4096,
40
  "pad_token": "<|endoftext|>",
41
  "padding_side": "right",