HunyuanVideo_MLLM_text_encoder

This model is adapted from the llava-llama-3-8b-v1_1-transformers model, retaining only its model.language_model and processor.tokenizer. This organization is not the provider of the original model and has no legal affiliation with it.

Intended for use as a text encoder for the HunyuanVideo open-source model.

You can download the model using the following git clone command or the ModelScope SDK:

SDK Download

# Install ModelScope
pip install modelscope
# Download model via SDK
from modelscope import snapshot_download
model_dir = snapshot_download('DiffSynth-Studio/HunyuanVideo_MLLM_text_encoder')

Git Download

# Git model download
git clone https://www.modelscope.cn/DiffSynth-Studio/HunyuanVideo_MLLM_text_encoder.git
Downloads last month
5
Safetensors
Model size
8B params
Tensor type
F16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Spaces using DiffSynth-Studio/HunyuanVideo_MLLM_text_encoder 5