sanali209 commited on
Commit
68264c5
·
1 Parent(s): 2b7a1e9

commit files to HF hub

Browse files
README.md CHANGED
@@ -1,61 +1,31 @@
1
  ---
2
- library_name: transformers
3
  tags:
4
- - generated_from_trainer
5
- datasets:
6
- - imagefolder
 
 
 
7
  model-index:
8
- - name: comixBF
9
- results: []
 
 
 
 
 
 
 
10
  ---
11
 
12
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
- should probably proofread and complete it, then remove this comment. -->
14
-
15
- # comixBF
16
-
17
- This model was trained from scratch on the imagefolder dataset.
18
- It achieves the following results on the evaluation set:
19
- - Loss: 0.3232
20
-
21
- ## Model description
22
-
23
- More information needed
24
-
25
- ## Intended uses & limitations
26
-
27
- More information needed
28
-
29
- ## Training and evaluation data
30
-
31
- More information needed
32
-
33
- ## Training procedure
34
-
35
- ### Training hyperparameters
36
 
37
- The following hyperparameters were used during training:
38
- - learning_rate: 2e-05
39
- - train_batch_size: 32
40
- - eval_batch_size: 16
41
- - seed: 42
42
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
- - lr_scheduler_type: linear
44
- - num_epochs: 4
45
 
46
- ### Training results
47
 
48
- | Training Loss | Epoch | Step | Validation Loss |
49
- |:-------------:|:-----:|:----:|:---------------:|
50
- | No log | 1.0 | 202 | 0.3801 |
51
- | No log | 2.0 | 404 | 0.3245 |
52
- | 0.4916 | 3.0 | 606 | 0.3441 |
53
- | 0.4916 | 4.0 | 808 | 0.3232 |
54
 
 
55
 
56
- ### Framework versions
57
 
58
- - Transformers 4.45.1
59
- - Pytorch 2.4.1+cu121
60
- - Datasets 3.0.1
61
- - Tokenizers 0.20.0
 
1
  ---
 
2
  tags:
3
+ - image-classification
4
+ - pytorch
5
+ - huggingpics
6
+ metrics:
7
+ - accuracy
8
+
9
  model-index:
10
+ - name: sanali209/comixBF
11
+ results:
12
+ - task:
13
+ name: Image Classification
14
+ type: image-classification
15
+ metrics:
16
+ - name: Accuracy
17
+ type: accuracy
18
+ value: 0.8955321907997131
19
  ---
20
 
21
+ # sanali209/comixBF
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
 
 
 
 
 
 
 
 
 
23
 
24
+ Autogenerated by HuggingPics🤗🖼️
25
 
26
+ Create your own image classifier for **anything** by running [the demo on Google Colab](https://colab.research.google.com/github/nateraw/huggingpics/blob/main/HuggingPics.ipynb).
 
 
 
 
 
27
 
28
+ Report any issues with the demo at the [github repo](https://github.com/nateraw/huggingpics).
29
 
 
30
 
31
+ ## Example Images
 
 
 
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "sanali209/comixBF",
3
  "architectures": [
4
  "ViTForImageClassification"
5
  ],
@@ -7,7 +7,7 @@
7
  "encoder_stride": 16,
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.0,
10
- "hidden_size": 1024,
11
  "id2label": {
12
  "0": "comix bw",
13
  "1": "comix color",
@@ -17,22 +17,22 @@
17
  },
18
  "image_size": 224,
19
  "initializer_range": 0.02,
20
- "intermediate_size": 4096,
21
  "label2id": {
22
- "comix bw": 0,
23
- "comix color": 1,
24
- "comix cower": 2,
25
- "image set": 3,
26
- "other": 4
27
  },
28
  "layer_norm_eps": 1e-12,
29
  "model_type": "vit",
30
- "num_attention_heads": 16,
31
  "num_channels": 3,
32
- "num_hidden_layers": 24,
33
  "patch_size": 16,
34
  "problem_type": "single_label_classification",
35
  "qkv_bias": true,
36
  "torch_dtype": "float32",
37
- "transformers_version": "4.45.1"
38
  }
 
1
  {
2
+ "_name_or_path": "google/vit-base-patch16-224",
3
  "architectures": [
4
  "ViTForImageClassification"
5
  ],
 
7
  "encoder_stride": 16,
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.0,
10
+ "hidden_size": 768,
11
  "id2label": {
12
  "0": "comix bw",
13
  "1": "comix color",
 
17
  },
18
  "image_size": 224,
19
  "initializer_range": 0.02,
20
+ "intermediate_size": 3072,
21
  "label2id": {
22
+ "comix bw": "0",
23
+ "comix color": "1",
24
+ "comix cower": "2",
25
+ "image set": "3",
26
+ "other": "4"
27
  },
28
  "layer_norm_eps": 1e-12,
29
  "model_type": "vit",
30
+ "num_attention_heads": 12,
31
  "num_channels": 3,
32
+ "num_hidden_layers": 12,
33
  "patch_size": 16,
34
  "problem_type": "single_label_classification",
35
  "qkv_bias": true,
36
  "torch_dtype": "float32",
37
+ "transformers_version": "4.44.2"
38
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f88965fdb9655c79c62192df1e6ffcf20f1c8026c24863e8b5dc57ef882eae5
3
- size 1213273572
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8b17bd792ad2330bdd6a3d2fbdbfee5139cd0e4f74da6228846f3ee44fba427
3
+ size 343233204
preprocessor_config.json CHANGED
@@ -7,7 +7,7 @@
7
  0.5,
8
  0.5
9
  ],
10
- "image_processor_type": "ViTImageProcessor",
11
  "image_std": [
12
  0.5,
13
  0.5,
 
7
  0.5,
8
  0.5
9
  ],
10
+ "image_processor_type": "ViTFeatureExtractor",
11
  "image_std": [
12
  0.5,
13
  0.5,
runs/events.out.tfevents.1729171301.6a860356268b.1471.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5958c1f9afe3e34408ad01bafcb3c7a912cda0430ba1df83020c78661f009db9
3
+ size 4936