#!/bin/bash # Configuración MODEL_PATH="/home/agent_h/data/starvector-1b-im2svg" CHAT_TEMPLATE="configs/chat-template.jinja" MAX_MODEL_LEN=8192 # SSH Tunnels (opcionales) bash -c "$SSH_TUNNEL_CMD_1" & bash -c "$SSH_TUNNEL_CMD_2" & bash -c "$SSH_TUNNEL_CMD_3" & bash -c "$SSH_TUNNEL_CMD_4" & # Controller python -m starvector.serve.vllm_api_gradio.controller --host 0.0.0.0 --port 10000 & # Esperar a que el controller inicie sleep 5 # ====== AÑADIR ESTOS COMANDOS - Iniciar servidores vLLM ====== echo "Iniciando servidores vLLM..." vllm serve $MODEL_PATH \ --chat-template $CHAT_TEMPLATE \ --trust-remote-code \ --port 8000 \ --max-model-len $MAX_MODEL_LEN & vllm serve $MODEL_PATH \ --chat-template $CHAT_TEMPLATE \ --trust-remote-code \ --port 8001 \ --max-model-len $MAX_MODEL_LEN & vllm serve $MODEL_PATH \ --chat-template $CHAT_TEMPLATE \ --trust-remote-code \ --port 8002 \ --max-model-len $MAX_MODEL_LEN & vllm serve $MODEL_PATH \ --chat-template $CHAT_TEMPLATE \ --trust-remote-code \ --port 8003 \ --max-model-len $MAX_MODEL_LEN & # Esperar a que vLLM inicie (puede tardar varios minutos) echo "Esperando a que vLLM cargue el modelo (esto puede tardar 2-3 minutos)..." sleep 180 # ====== FIN DE LA MODIFICACIÓN ====== # Model Workers python -m starvector.serve.vllm_api_gradio.model_worker --host 0.0.0.0 --controller http://localhost:10000 --port 40000 --worker http://localhost:40000 --model-name $MODEL_PATH --vllm-base-url http://localhost:8000 & python -m starvector.serve.vllm_api_gradio.model_worker --host 0.0.0.0 --controller http://localhost:10000 --port 40001 --worker http://localhost:40001 --model-name $MODEL_PATH --vllm-base-url http://localhost:8001 & python -m starvector.serve.vllm_api_gradio.model_worker --host 0.0.0.0 --controller http://localhost:10000 --port 40002 --worker http://localhost:40002 --model-name $MODEL_PATH --vllm-base-url http://localhost:8002 & python -m starvector.serve.vllm_api_gradio.model_worker --host 0.0.0.0 --controller http://localhost:10000 --port 40003 --worker http://localhost:40003 --model-name $MODEL_PATH --vllm-base-url http://localhost:8003 & # Gradio Web Server python -m starvector.serve.vllm_api_gradio.gradio_web_server --controller http://localhost:10000 --model-list-mode reload --port 7860