XiaodongChen commited on
Commit
6b5e304
·
verified ·
1 Parent(s): e0e2dd8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -3
README.md CHANGED
@@ -1,3 +1,12 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ base_model:
4
+ - meta-llama/Llama-2-7b-hf
5
+ ---
6
+
7
+ The model is derived from Llama-2-7b-hf through pruning using LLM-Streamline **(Streamlining Redundant Layers to Compress Large Language Models, ICLR 2025 Spotlight)**. The entire training process required only 0.06B tokens.
8
+ Below are the results of the evaluation using lm-eval:
9
+ | | arc_c | arc_e | boolq | hellaswag | openbookqa | rte | winogrande | Avg |
10
+ |--------------|-------|-------|-------|-----------|------------|------|------------|------|
11
+ | Llama-2-7b | 43.3 | 76.4 | 77.7 | 57.2 | 31.4 | 62.8 | 69.1 | 59.7 |
12
+ | Llama-2-4.7b | 34.0 | 64.6 | 74.7 | 49.8 | 27.4 | 61.7 | 66.4 | 54.1 |