Update README.md
Browse files
README.md
CHANGED
|
@@ -133,7 +133,7 @@ attn_implementation = "eager" # Or "flash_attention_2"
|
|
| 133 |
model = SentenceTransformer(
|
| 134 |
"nvidia/llama-embed-nemotron-8b",
|
| 135 |
trust_remote_code=True,
|
| 136 |
-
model_kwargs={"attn_implementation": attn_implementation, "torch_dtype": "
|
| 137 |
tokenizer_kwargs={"padding_side": "left"},
|
| 138 |
)
|
| 139 |
|
|
|
|
| 133 |
model = SentenceTransformer(
|
| 134 |
"nvidia/llama-embed-nemotron-8b",
|
| 135 |
trust_remote_code=True,
|
| 136 |
+
model_kwargs={"attn_implementation": attn_implementation, "torch_dtype": "float32"},
|
| 137 |
tokenizer_kwargs={"padding_side": "left"},
|
| 138 |
)
|
| 139 |
|