davidgaofc commited on
Commit
e73a45c
·
1 Parent(s): b4c65a4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -13
README.md CHANGED
@@ -7,28 +7,28 @@ model-index:
7
  results: []
8
  ---
9
 
10
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
- should probably proofread and complete it, then remove this comment. -->
12
-
13
  # training
14
 
15
- This model is a fine-tuned version of [huggingface/CodeBERTa-small-v1](https://huggingface.co/huggingface/CodeBERTa-small-v1) on an unknown dataset.
16
- It achieves the following results on the evaluation set:
17
- - Loss: 0.1370
18
 
19
  ## Model description
20
 
21
- More information needed
 
 
22
 
23
  ## Intended uses & limitations
24
 
25
- More information needed
26
 
27
  ## Training and evaluation data
28
 
29
- More information needed
 
 
30
 
31
  ## Training procedure
 
32
 
33
  ### Training hyperparameters
34
 
@@ -41,10 +41,6 @@ The following hyperparameters were used during training:
41
  - lr_scheduler_type: linear
42
  - num_epochs: 1
43
 
44
- ### Training results
45
-
46
-
47
-
48
  ### Framework versions
49
 
50
  - Transformers 4.35.0
 
7
  results: []
8
  ---
9
 
 
 
 
10
  # training
11
 
12
+ This model is a fine-tuned version of [huggingface/CodeBERTa-small-v1](https://huggingface.co/huggingface/CodeBERTa-small-v1) on an [my own techdebt dataset](https://huggingface.co/datasets/davidgaofc/techdebt).
 
 
13
 
14
  ## Model description
15
 
16
+ Classifies cleaned diffs of code.
17
+ * 1: exhibits possible technical debt
18
+ * 0: is probably clean
19
 
20
  ## Intended uses & limitations
21
 
22
+ Limited by many things probably, use with caution. Improvements in progress.
23
 
24
  ## Training and evaluation data
25
 
26
+ ~95% accurate on the test split of dataset above
27
+ ~.94 F1 score on test split of dataset above.
28
+
29
 
30
  ## Training procedure
31
+ One epoch of training done on the dataset above.
32
 
33
  ### Training hyperparameters
34
 
 
41
  - lr_scheduler_type: linear
42
  - num_epochs: 1
43
 
 
 
 
 
44
  ### Framework versions
45
 
46
  - Transformers 4.35.0