|
--- |
|
library_name: transformers |
|
tags: [] |
|
--- |
|
|
|
This is a GPT-2 model trained in llm.c for 330K steps (of 1M batch size) on FineWeb-EDU. |
|
|
|
A lot more detailed information is here: https://github.com/karpathy/llm.c/discussions/677 . |
|
|
|
This model has a bit of a complicated history. I wanted to train it for 400K steps, i.e. (`-x 400000`), but it became unstable later in training and exploded around step 330K. Because I was losing my computing quota shortly, I decided to just rewind back to checkpoint 300K, and then instead of going all the way to 400K I started annealing linearly down to 330K. This went without incident and produced this model. |
|
|
|
This is the longest I've trained a GPT-2 model for, and it reaches HellaSwag of 62.7 by the end. |
|
|