Commit
·
3fef353
1
Parent(s):
1f16131
fix checkpoint count and shorten intro section
Browse files
README.md
CHANGED
|
@@ -15,8 +15,8 @@ interpretability research. It contains two sets of eight models of sizes
|
|
| 15 |
70M, 160M, 410M, 1B, 1.4B, 2.8B, 6.9B, and 12B. For each size, there are two
|
| 16 |
models: one trained on the Pile, and one trained on the Pile after the dataset
|
| 17 |
has been globally deduplicated. All 8 model sizes are trained on the exact
|
| 18 |
-
same data, in the exact same order.
|
| 19 |
-
|
| 20 |
|
| 21 |
The Pythia model suite was deliberately designed to promote scientific
|
| 22 |
research on large language models, especially interpretability research.
|
|
@@ -24,20 +24,25 @@ Despite not centering downstream performance as a design goal, we find the
|
|
| 24 |
models <a href="#evaluations">match or exceed</a> the performance of
|
| 25 |
similar and same-sized models, such as those in the OPT and GPT-Neo suites.
|
| 26 |
|
|
|
|
|
|
|
|
|
|
| 27 |
Previously, we released an early version of the Pythia suite to the public.
|
| 28 |
However, we decided to retrain the model suite to address a few hyperparameter
|
| 29 |
discrepancies. This model card <a href="#changelog">lists the changes</a>;
|
| 30 |
see appendix B in the Pythia paper for further discussion. We found no
|
| 31 |
difference in benchmark performance between the two Pythia versions.
|
| 32 |
The old models are
|
| 33 |
-
[still available](https://huggingface.co/models?other=pythia_v0)
|
| 34 |
-
|
| 35 |
**This is the current release.**
|
| 36 |
|
| 37 |
Please note that all models in the *Pythia* suite were renamed in January
|
| 38 |
2023. For clarity, a <a href="#naming-convention-and-parameter-count">table
|
| 39 |
comparing the old and new names</a> is provided in this model card, together
|
| 40 |
with exact parameter counts.
|
|
|
|
|
|
|
| 41 |
|
| 42 |
# Pythia-12B
|
| 43 |
|
|
@@ -80,11 +85,12 @@ non-embedding parameters.</figcaption>
|
|
| 80 |
|
| 81 |
The primary intended use of Pythia is research on the behavior, functionality,
|
| 82 |
and limitations of large language models. This suite is intended to provide
|
| 83 |
-
a controlled setting for performing scientific experiments.
|
| 84 |
-
|
| 85 |
-
143 evenly
|
| 86 |
-
hosted on Hugging Face as branches. Note
|
| 87 |
-
exactly to the model checkpoint on the `main`
|
|
|
|
| 88 |
|
| 89 |
You may also further fine-tune and adapt Pythia-12B for deployment,
|
| 90 |
as long as your use is in accordance with the Apache 2.0 license. Pythia
|
|
@@ -108,7 +114,7 @@ language models are commonly deployed, such as writing genre prose,
|
|
| 108 |
or commercial chatbots. This means Pythia-12B will **not**
|
| 109 |
respond to a given prompt the way a product like ChatGPT does. This is because,
|
| 110 |
unlike this model, ChatGPT was fine-tuned using methods such as Reinforcement
|
| 111 |
-
Learning from Human Feedback (RLHF) to better “
|
| 112 |
|
| 113 |
### Limitations and biases
|
| 114 |
|
|
@@ -181,7 +187,9 @@ The Pile was **not** deduplicated before being used to train Pythia-12B.
|
|
| 181 |
|
| 182 |
All models were trained on the exact same data, in the exact same order. Each
|
| 183 |
model saw 299,892,736,000 tokens during training, and 143 checkpoints for each
|
| 184 |
-
model are saved every 2,097,152,000 tokens, spaced evenly throughout training
|
|
|
|
|
|
|
| 185 |
This corresponds to training for just under 1 epoch on the Pile for
|
| 186 |
non-deduplicated models, and about 1.5 epochs on the deduplicated Pile.
|
| 187 |
|
|
@@ -198,7 +206,7 @@ Pythia uses the same tokenizer as [GPT-NeoX-
|
|
| 198 |
All 16 *Pythia* models were evaluated using the [LM Evaluation
|
| 199 |
Harness](https://github.com/EleutherAI/lm-evaluation-harness). You can access
|
| 200 |
the results by model and step at `results/json/*` in the [GitHub
|
| 201 |
-
repository](https://github.com/EleutherAI/pythia/tree/main/results/json/
|
| 202 |
Expand the sections below to see plots of evaluation results for all
|
| 203 |
Pythia and Pythia-deduped models compared with OPT and BLOOM.
|
| 204 |
|
|
|
|
| 15 |
70M, 160M, 410M, 1B, 1.4B, 2.8B, 6.9B, and 12B. For each size, there are two
|
| 16 |
models: one trained on the Pile, and one trained on the Pile after the dataset
|
| 17 |
has been globally deduplicated. All 8 model sizes are trained on the exact
|
| 18 |
+
same data, in the exact same order. We also provide 154 intermediate
|
| 19 |
+
checkpoints per model, hosted on Hugging Face as branches.
|
| 20 |
|
| 21 |
The Pythia model suite was deliberately designed to promote scientific
|
| 22 |
research on large language models, especially interpretability research.
|
|
|
|
| 24 |
models <a href="#evaluations">match or exceed</a> the performance of
|
| 25 |
similar and same-sized models, such as those in the OPT and GPT-Neo suites.
|
| 26 |
|
| 27 |
+
<details>
|
| 28 |
+
<summary style="font-weight: 600">Past early release and naming convention.</summary>
|
| 29 |
+
|
| 30 |
Previously, we released an early version of the Pythia suite to the public.
|
| 31 |
However, we decided to retrain the model suite to address a few hyperparameter
|
| 32 |
discrepancies. This model card <a href="#changelog">lists the changes</a>;
|
| 33 |
see appendix B in the Pythia paper for further discussion. We found no
|
| 34 |
difference in benchmark performance between the two Pythia versions.
|
| 35 |
The old models are
|
| 36 |
+
[still available](https://huggingface.co/models?other=pythia_v0), but we
|
| 37 |
+
suggest the retrained suite if you are just starting to use Pythia.<br>
|
| 38 |
**This is the current release.**
|
| 39 |
|
| 40 |
Please note that all models in the *Pythia* suite were renamed in January
|
| 41 |
2023. For clarity, a <a href="#naming-convention-and-parameter-count">table
|
| 42 |
comparing the old and new names</a> is provided in this model card, together
|
| 43 |
with exact parameter counts.
|
| 44 |
+
</details>
|
| 45 |
+
<br>
|
| 46 |
|
| 47 |
# Pythia-12B
|
| 48 |
|
|
|
|
| 85 |
|
| 86 |
The primary intended use of Pythia is research on the behavior, functionality,
|
| 87 |
and limitations of large language models. This suite is intended to provide
|
| 88 |
+
a controlled setting for performing scientific experiments. We also provide
|
| 89 |
+
154 checkpoints per model: initial `step0`, 10 log-spaced checkpoints
|
| 90 |
+
`step{1,2,4...512}`, and 143 evenly-spaced checkpoints from `step1000` to
|
| 91 |
+
`step143000`. These checkpoints are hosted on Hugging Face as branches. Note
|
| 92 |
+
that branch `143000` corresponds exactly to the model checkpoint on the `main`
|
| 93 |
+
branch of each model.
|
| 94 |
|
| 95 |
You may also further fine-tune and adapt Pythia-12B for deployment,
|
| 96 |
as long as your use is in accordance with the Apache 2.0 license. Pythia
|
|
|
|
| 114 |
or commercial chatbots. This means Pythia-12B will **not**
|
| 115 |
respond to a given prompt the way a product like ChatGPT does. This is because,
|
| 116 |
unlike this model, ChatGPT was fine-tuned using methods such as Reinforcement
|
| 117 |
+
Learning from Human Feedback (RLHF) to better “follow” human instructions.
|
| 118 |
|
| 119 |
### Limitations and biases
|
| 120 |
|
|
|
|
| 187 |
|
| 188 |
All models were trained on the exact same data, in the exact same order. Each
|
| 189 |
model saw 299,892,736,000 tokens during training, and 143 checkpoints for each
|
| 190 |
+
model are saved every 2,097,152,000 tokens, spaced evenly throughout training,
|
| 191 |
+
from `step1000` to `step143000` (which is the same as `main`). In addition, we
|
| 192 |
+
also provide frequent early checkpoints: `step0` and `step{1,2,4...512}`.
|
| 193 |
This corresponds to training for just under 1 epoch on the Pile for
|
| 194 |
non-deduplicated models, and about 1.5 epochs on the deduplicated Pile.
|
| 195 |
|
|
|
|
| 206 |
All 16 *Pythia* models were evaluated using the [LM Evaluation
|
| 207 |
Harness](https://github.com/EleutherAI/lm-evaluation-harness). You can access
|
| 208 |
the results by model and step at `results/json/*` in the [GitHub
|
| 209 |
+
repository](https://github.com/EleutherAI/pythia/tree/main/results/json/).<br>
|
| 210 |
Expand the sections below to see plots of evaluation results for all
|
| 211 |
Pythia and Pythia-deduped models compared with OPT and BLOOM.
|
| 212 |
|