Vipitis commited on
Commit
657d724
·
1 Parent(s): 6bf0547

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +54 -3
README.md CHANGED
@@ -1,16 +1,67 @@
1
  ---
 
 
2
  license: bigcode-openrail-m
3
  datasets:
4
  - bigcode/the-stack-dedup
5
- - Vipitis/Shadertoys
6
  pipeline_tag: text-generation
7
  tags:
8
  - code
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  ---
10
 
11
  [Santacoder](https://huggingface.co/bigcode/santacoder) finetuned on [Shadertoys](https://huggingface.co/datasets/Vipitis/Shadertoys) for 1000 steps with a batch size of 2 and full sequence length of 2048.
12
- Origianl finetuning script from found [here](https://github.com/loubnabnl/santacoder-finetuning), adapted version to follow (soon^^).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
 
14
- Main purpose of this model is to explore if finetuning models improves performance on [ShaderEval](https://huggingface.co/spaces/Vipitis/ShaderEval), results to follow (sooner).
 
15
 
16
  License carried over from model, however training data has an undefied license. Check details in [Shadertoys](https://huggingface.co/datasets/Vipitis/Shadertoys).
 
1
  ---
2
+ language:
3
+ - code
4
  license: bigcode-openrail-m
5
  datasets:
6
  - bigcode/the-stack-dedup
7
+ - Vipitis/Shadertoys-fine
8
  pipeline_tag: text-generation
9
  tags:
10
  - code
11
+ - shader
12
+ widget:
13
+ - text: void mainImage( out vec4 fragColor, in vec2 fragCoord )
14
+ example_title: mainImage
15
+ group: Shadertoy
16
+ model-index:
17
+ - name: santacoder-finetuned-the-stack-glsl
18
+ results:
19
+ - task:
20
+ type: text-generation
21
+ name: ShaderEval
22
+ dataset:
23
+ type: Vipitis/Shadertoys-fine
24
+ name: Shadertoys-fine
25
+ config: return_completion
26
+ revision: 0.0.2
27
+ metrics:
28
+ - type: exact_match
29
+ value: 0.550
30
+ name: 300 samples, greedy decoding
31
+ verified: false
32
  ---
33
 
34
  [Santacoder](https://huggingface.co/bigcode/santacoder) finetuned on [Shadertoys](https://huggingface.co/datasets/Vipitis/Shadertoys) for 1000 steps with a batch size of 2 and full sequence length of 2048.
35
+ adapted finetuning script found [here](./train.py)
36
+
37
+ Try model in the [ShaderCoder](https://huggingface.co/spaces/Vipitis/ShaderCoder) demo space
38
+
39
+ ### Finetuning parameters
40
+ ```sh
41
+ python3 train.py --model_path "bigcode/santacoder" \
42
+ --dataset_name "Vipitis/Shadertoys" \
43
+ --data_column "code" \
44
+ --split "train" \
45
+ --seq_length 2048 \
46
+ --max_steps 1000 \
47
+ --batch_size 2 \
48
+ --gradient_accumulation_steps 4 \
49
+ --learning_rate 5e-5 \
50
+ --num_warmup_steps 100 \
51
+ --eval_freq 100 \
52
+ --save_freq 100 \
53
+ --log_freq 1 \
54
+ --output_dir "checkpoint_dir" \
55
+ --no_fp16
56
+
57
+
58
+ ```
59
+
60
+ Main purpose of this model is to explore if finetuning models improves performance on [ShaderEval](https://huggingface.co/spaces/Vipitis/ShaderEval), which reached 0.550 with 300 samples.
61
+
62
+ ### Disclaimer
63
 
64
+ While the train/test split is held out, there is a lot of data contamination. The model results can't be trusted for this simple benchmark.
65
+ Better tasks for the benchmark will be developed and tested against these models.
66
 
67
  License carried over from model, however training data has an undefied license. Check details in [Shadertoys](https://huggingface.co/datasets/Vipitis/Shadertoys).