Update librechat.yaml
Browse files- librechat.yaml +2 -22
librechat.yaml
CHANGED
|
@@ -187,27 +187,6 @@ endpoints:
|
|
| 187 |
capabilities: ["file_search", "actions", "tools", "artifacts", "ocr", "chain"]
|
| 188 |
custom:
|
| 189 |
- name: 'Tiny-DEV' # Unique name for the endpoint
|
| 190 |
-
# For `apiKey` and `baseURL`, you can use environment variables that you define.
|
| 191 |
-
# recommended environment variables:
|
| 192 |
-
apiKey: '77ea72d89b98d279c1848389cd027a51c408'
|
| 193 |
-
baseURL: 'https://akiko19191-backend2.hf.space/'
|
| 194 |
-
# baseURL: 'https://e9e7-182-69-179-206.ngrok-free.app/'
|
| 195 |
-
|
| 196 |
-
|
| 197 |
-
# Models configuration
|
| 198 |
-
models:
|
| 199 |
-
# List of default models to use. At least one value is required.
|
| 200 |
-
default: ['gemini-2.5-pro-exp-03-25','gemini-2.5-flash-preview-04-17','gemini-2.0-flash-thinking-exp-01-21' ,'deepseek.r1','deepseek-reasoner','MAI-DS-R1-FP8','DeepSeek-V3-0324','deepseek-chat','gpt4.1','gpt-4o','o4-mini','gemini-2.0-pro-exp-02-05','GLM-4-32B-0414','GLM-Z1-32B-0414','deepseek-r1-distill-llama-70b', 'qwq-32b','llama-3.3-70b-versatile','meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8','microsoft/Phi-4-multimodal-instruct','google/gemma-3-27b-it','meta-llama/Llama-4-Scout-17B-16E-Instruct']
|
| 201 |
-
# Fetch option: Set to true to fetch models from API.
|
| 202 |
-
# fetch: false # Defaults to false.
|
| 203 |
-
|
| 204 |
-
# Optional configurations
|
| 205 |
-
|
| 206 |
-
# Title Conversation setting
|
| 207 |
-
titleConvo: true # Set to true to enable title conversation
|
| 208 |
-
|
| 209 |
-
modelDisplayLabel: 'Tiny' # Default is "AI" when not set.
|
| 210 |
-
- name: 'Tiny-DEBUGGING' # Unique name for the endpoint
|
| 211 |
# For `apiKey` and `baseURL`, you can use environment variables that you define.
|
| 212 |
# recommended environment variables:
|
| 213 |
apiKey: '77ea72d89b98d279c1848389cd027a51c408'
|
|
@@ -218,7 +197,7 @@ endpoints:
|
|
| 218 |
# Models configuration
|
| 219 |
models:
|
| 220 |
# List of default models to use. At least one value is required.
|
| 221 |
-
default: ['gemini-2.5-pro-exp-03-25','gemini-2.5-flash-preview-04-17','gemini-2.0-flash-thinking-exp-01-21' ,'deepseek.r1','deepseek-reasoner','MAI-DS-R1-FP8','DeepSeek-V3-0324','deepseek-chat','gpt4.1','gpt-4o','o4-mini','gemini-2.0-pro-exp-02-05','GLM-4-32B-0414','GLM-Z1-32B-0414','deepseek-r1-distill-llama-70b', 'qwq-32b','llama-3.3-70b-versatile','meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8','microsoft/Phi-4-multimodal-instruct','google/gemma-3-27b-it','meta-llama/Llama-4-Scout-17B-16E-Instruct']
|
| 222 |
# Fetch option: Set to true to fetch models from API.
|
| 223 |
# fetch: false # Defaults to false.
|
| 224 |
|
|
@@ -228,6 +207,7 @@ endpoints:
|
|
| 228 |
titleConvo: true # Set to true to enable title conversation
|
| 229 |
|
| 230 |
modelDisplayLabel: 'Tiny' # Default is "AI" when not set.
|
|
|
|
| 231 |
ocr:
|
| 232 |
mistralModel: "mistral-ocr-latest" # Optional: Specify Mistral model, defaults to "mistral-ocr-latest"
|
| 233 |
apiKey: '${OCR_API_KEY}' # Optional: Defaults to OCR_API_KEY env variable
|
|
|
|
| 187 |
capabilities: ["file_search", "actions", "tools", "artifacts", "ocr", "chain"]
|
| 188 |
custom:
|
| 189 |
- name: 'Tiny-DEV' # Unique name for the endpoint
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 190 |
# For `apiKey` and `baseURL`, you can use environment variables that you define.
|
| 191 |
# recommended environment variables:
|
| 192 |
apiKey: '77ea72d89b98d279c1848389cd027a51c408'
|
|
|
|
| 197 |
# Models configuration
|
| 198 |
models:
|
| 199 |
# List of default models to use. At least one value is required.
|
| 200 |
+
default: ['gemini-2.5-pro-exp-03-25','gemini-2.5-flash-preview-04-17','gemini-2.0-flash-thinking-exp-01-21' ,'deepseek.r1','deepseek-reasoner','MAI-DS-R1-FP8', 'DeepSeek-R1-Zero', 'DeepSeek-R1-Chimera' , 'QwQ-32B-ArliAI-RpR-v1' , 'DeepSeek-V3-0324','deepseek-chat','gpt4.1','gpt-4o','o4-mini','gemini-2.0-pro-exp-02-05','GLM-4-32B-0414','GLM-Z1-32B-0414','deepseek-r1-distill-llama-70b', 'qwq-32b','llama-3.3-70b-versatile','meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8','microsoft/Phi-4-multimodal-instruct','google/gemma-3-27b-it','meta-llama/Llama-4-Scout-17B-16E-Instruct']
|
| 201 |
# Fetch option: Set to true to fetch models from API.
|
| 202 |
# fetch: false # Defaults to false.
|
| 203 |
|
|
|
|
| 207 |
titleConvo: true # Set to true to enable title conversation
|
| 208 |
|
| 209 |
modelDisplayLabel: 'Tiny' # Default is "AI" when not set.
|
| 210 |
+
|
| 211 |
ocr:
|
| 212 |
mistralModel: "mistral-ocr-latest" # Optional: Specify Mistral model, defaults to "mistral-ocr-latest"
|
| 213 |
apiKey: '${OCR_API_KEY}' # Optional: Defaults to OCR_API_KEY env variable
|