jzhang533 ckl117 commited on
Commit
b565cf6
·
verified ·
1 Parent(s): 9ee3601

config.json add "tie_word_embeddings": true (#7)

Browse files

- config.json add "tie_word_embeddings": true (3a2dba6387cabdaa7b895ff944ceef408232b5f2)
- minor update to readme (2112044c45d8a784e5eec33d02b95b5d22222aa8)


Co-authored-by: ckl <[email protected]>

Files changed (2) hide show
  1. README.md +2 -3
  2. config.json +1 -0
README.md CHANGED
@@ -68,8 +68,7 @@ ERNIE-4.5-0.3B is a text dense Post-trained model. The following are the model c
68
 
69
  ### Using `transformers` library
70
 
71
- **Note**: Before using the model, please ensure you have the `transformers` library installed
72
- (upcoming version 4.54.0 or [the latest version](https://github.com/huggingface/transformers?tab=readme-ov-file#installation))
73
 
74
  The following contains a code snippet illustrating how to use the model generate content based on given inputs.
75
 
@@ -116,7 +115,7 @@ print("generate_text:", generate_text)
116
  [vllm](https://github.com/vllm-project/vllm/tree/main) github library. Python-only [build](https://docs.vllm.ai/en/latest/getting_started/installation/gpu.html#set-up-using-python-only-build-without-compilation).
117
 
118
  ```bash
119
- vllm serve baidu/ERNIE-4.5-0.3B-PT --trust-remote-code
120
  ```
121
 
122
  ## License
 
68
 
69
  ### Using `transformers` library
70
 
71
+ **Note**: You'll need the `transformers` library (version 4.54.0 or newer) installed to use this model.
 
72
 
73
  The following contains a code snippet illustrating how to use the model generate content based on given inputs.
74
 
 
115
  [vllm](https://github.com/vllm-project/vllm/tree/main) github library. Python-only [build](https://docs.vllm.ai/en/latest/getting_started/installation/gpu.html#set-up-using-python-only-build-without-compilation).
116
 
117
  ```bash
118
+ vllm serve baidu/ERNIE-4.5-0.3B-PT
119
  ```
120
 
121
  ## License
config.json CHANGED
@@ -18,6 +18,7 @@
18
  "rms_norm_eps": 1e-05,
19
  "rope_scaling": null,
20
  "rope_theta": 500000.0,
 
21
  "torch_dtype": "bfloat16",
22
  "transformers_version": "4.54.0.dev0",
23
  "use_bias": false,
 
18
  "rms_norm_eps": 1e-05,
19
  "rope_scaling": null,
20
  "rope_theta": 500000.0,
21
+ "tie_word_embeddings": true,
22
  "torch_dtype": "bfloat16",
23
  "transformers_version": "4.54.0.dev0",
24
  "use_bias": false,