Commit
·
6ef4a98
1
Parent(s):
e4dbadc
Update README.md
Browse files
README.md
CHANGED
|
@@ -9,17 +9,7 @@ tags:
|
|
| 9 |
license: gpl-3.0
|
| 10 |
---
|
| 11 |
|
| 12 |
-
|
| 13 |
-
|
| 14 |
-
Advance notice regarding the deletion of Qwen:
|
| 15 |
-
|
| 16 |
-
**I remain unaware as to the reasons behind Qwen's deletion. Should this repository be found in violation of any terms stipulated by Qwen that necessitate its removal, I earnestly request you to establish contact with me. I pledge to expunge all references to Qwen and maintain the tokenizer and associated weights as an autonomous model, inherently distinct from Qwen. I will then proceed to christen this model with a new identifier.**
|
| 17 |
-
|
| 18 |
-
对于通义千问删除的事先说明:
|
| 19 |
-
|
| 20 |
-
**我尚不清楚通义千问被删除的原因。如果此仓库违反了任何由通义千问提出的需要进行移除的条款,我真诚地请求您与我取得联系。我承诺将清除所有与通义千问/Qwen的相关引用,并将分词器和相关权重作为与通义千问本质上不同的独立模型进行维护。然后,我将为这个模型取一个新的名字。**
|
| 21 |
-
|
| 22 |
-
This is the LLaMAfied replica of [Qwen/Qwen-VL-Chat](https://huggingface.co/Qwen/Qwen-VL-Chat), recalibrated to fit the original LLaMA/LLaMA-2-like model structure.
|
| 23 |
|
| 24 |
You can use LlamaForCausalLM for model inference, which is the same as LLaMA/LLaMA-2 models (using GPT2Tokenizer converted from the original tiktoken, by [vonjack](https://huggingface.co/vonjack)).
|
| 25 |
|
|
|
|
| 9 |
license: gpl-3.0
|
| 10 |
---
|
| 11 |
|
| 12 |
+
This is the LLaMAfied replica of [Qwen/Qwen-VL-Chat](https://huggingface.co/Qwen/Qwen-VL-Chat) (Original Version before 25.09.2023), recalibrated to fit the original LLaMA/LLaMA-2-like model structure.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 13 |
|
| 14 |
You can use LlamaForCausalLM for model inference, which is the same as LLaMA/LLaMA-2 models (using GPT2Tokenizer converted from the original tiktoken, by [vonjack](https://huggingface.co/vonjack)).
|
| 15 |
|