thatdramebaazguy commited on
Commit
f462fe3
·
1 Parent(s): 9be3344

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +57 -0
README.md CHANGED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ datasets:
3
+ - imdb
4
+ - cornell_movie_dialogue
5
+
6
+ language:
7
+ - English
8
+
9
+ thumbnail:
10
+
11
+ tags:
12
+ - roberta
13
+ - roberta-base
14
+ - masked-language-modeling
15
+ - masked-lm
16
+
17
+ license: cc-by-4.0
18
+
19
+ ---
20
+ # roberta-base for MLM
21
+
22
+ ```
23
+ model_name = "thatdramebaazguy/movie-roberta-base"
24
+ pipeline(model=model_name, tokenizer=model_name, revision="v1.0", task="Fill-Mask")
25
+ ```
26
+ ## Overview
27
+ **Language model:** roberta-base
28
+ **Language:** English
29
+ **Downstream-task:** Fill-Mask
30
+ **Training data:** imdb, polarity movie data, cornell_movie_dialogue, 25mlens movie names
31
+ **Eval data:** imdb, polarity movie data, cornell_movie_dialogue, 25mlens movie names
32
+ **Infrastructure**: 4x Tesla v100
33
+ **Code:** See [example](https://github.com/adityaarunsinghal/Domain-Adaptation/blob/master/scripts/shell_scripts/train_movie_roberta.sh)
34
+
35
+ ## Hyperparameters
36
+ ```
37
+ Num examples = 4767233
38
+ Num Epochs = 2
39
+ Instantaneous batch size per device = 20
40
+ Total train batch size (w. parallel, distributed & accumulation) = 80
41
+ Gradient Accumulation steps = 1
42
+ Total optimization steps = 119182
43
+ eval_loss = 1.6153
44
+ eval_samples = 20573
45
+ perplexity = 5.0296
46
+ learning_rate=5e-05
47
+ n_gpu = 4
48
+
49
+ ```
50
+ ## Performance
51
+
52
+ perplexity = 5.0296
53
+
54
+ Some of my work:
55
+ - [Domain-Adaptation Project](https://github.com/adityaarunsinghal/Domain-Adaptation/)
56
+
57
+ ---