| --- |
| datasets: HuggingFaceFW/fineweb-edu |
| widget: |
| - example_title: Example interaction |
| text: During photosynthesis in green plants |
| inference: |
| parameters: |
| repetition_penalty: 1.3 |
| language: |
| - en |
| library_name: transformers |
| license: mit |
| --- |
| |
| # Model Card for gpt2-124M-edu-fineweb-10B |
|
|
| A 124M parameter GPT2 model trained with the 10B fineweb-edu dataset using [https://github.com/karpathy/llm.c](https://github.com/karpathy/llm.c) |
|
|
| Training took 20 hours on a single 4090 GPU (limited to 350W), giving the following graphs: |
|
|
|  |
|
|
| ## Training |
|
|
| The training parameters where: |
| ``` |
| ./train_gpt2cu \ |
| -i "dev/data/edu_fineweb10B/edu_fineweb_train_*.bin" \ |
| -j "dev/data/edu_fineweb10B/edu_fineweb_val_*.bin" \ |
| -o log124M \ |
| -e "d12" \ |
| -b 56 -t 1024 \ |
| -d 458752 \ |
| -r 1 \ |
| -z 1 \ |
| -c 0.1 \ |
| -l 0.002 \ |
| -q 0.0 \ |
| -u 700 \ |
| -n 5000 \ |
| -v 250 -s 20000 \ |
| -h 1 |
| ``` |
|
|
| The model has had no further finetuning. |
|
|
| ## Evaluation |
|
|
| Evals using [Eleuther AI Harness](https://github.com/EleutherAI/lm-evaluation-harness/tree/b281b0921b636bc36ad05c0b0b0763bd6dd43463) |
| as described in the [open_llm_leaderboard](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard) |
| and comparing with those published for [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) |
|
|
|  |
|
|
| | Eval Test | Score | |
| | --------- | ----- | |
| | arc_challenge (25 shot) | 24.49 | |
| | gsm8k (5 shot) | 0.08 | |
| | hellaswag (10 shot) | 32.64 | |
| | mmlu (5 shot) | 26.06 | |
| | truthfulqa (0 shot) | 42.45 | |
| | winogrande (5 shot) | 52.17 | |
| | **Overall Score** | **29.65** | |