Update README.md
Browse files
README.md
CHANGED
|
@@ -121,6 +121,16 @@ solution = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
|
| 121 |
print(solution)
|
| 122 |
```
|
| 123 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 124 |
## Citation
|
| 125 |
|
| 126 |
```bibtex
|
|
|
|
| 121 |
print(solution)
|
| 122 |
```
|
| 123 |
|
| 124 |
+
## 🙏 Acknowledgements
|
| 125 |
+
|
| 126 |
+
We are grateful to the open-source community for their invaluable contributions. Special thanks to:
|
| 127 |
+
|
| 128 |
+
- **[Qwen3](https://huggingface.co/collections/Qwen/qwen3)** - for providing the foundational base models that powered our research
|
| 129 |
+
- **[slime](https://github.com/THUDM/slime)** - for their innovative work on efficient reinforcement learning framework that powered our training pipeline
|
| 130 |
+
- **[verl](https://github.com/volcengine/verl)** - for the versatile reinforcement learning framework that enabled our training pipeline
|
| 131 |
+
- **[sglang](https://github.com/sgl-project/sglang)** - for the efficient LLM serving and inference infrastructure
|
| 132 |
+
- **[Megatron-LM](https://github.com/NVIDIA/Megatron-LM)** - for the large-scale model training framework
|
| 133 |
+
|
| 134 |
## Citation
|
| 135 |
|
| 136 |
```bibtex
|