Update README.md
Browse files
README.md
CHANGED
|
@@ -10,6 +10,10 @@ pipeline_tag: text-generation
|
|
| 10 |
|
| 11 |
# PRRC-Reasoning Language Model (1.3B Parameters, 30B Tokens)
|
| 12 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 13 |
## Model Description
|
| 14 |
|
| 15 |
This is a 1.3B parameter transformer-based decoder-only language model trained from scratch on 30B tokens selected from SlimPajama dataset using the **Reasoning** dimension of the PRRC framework. The training data was curated by selecting text with high reasoning complexity, focusing on content that requires multi-step logical analysis and critical thinking.
|
|
|
|
| 10 |
|
| 11 |
# PRRC-Reasoning Language Model (1.3B Parameters, 30B Tokens)
|
| 12 |
|
| 13 |
+
This repository contains the model described in the paper [Meta-rater: A Multi-dimensional Data Selection Method for Pre-training Language Models](https://huggingface.co/papers/2504.14194).
|
| 14 |
+
|
| 15 |
+
Code: https://github.com/opendatalab/Meta-rater
|
| 16 |
+
|
| 17 |
## Model Description
|
| 18 |
|
| 19 |
This is a 1.3B parameter transformer-based decoder-only language model trained from scratch on 30B tokens selected from SlimPajama dataset using the **Reasoning** dimension of the PRRC framework. The training data was curated by selecting text with high reasoning complexity, focusing on content that requires multi-step logical analysis and critical thinking.
|