gemma-2-9b-linkscout-v1

This model is a fine-tuned version of unsloth/gemma-2-9b-it-bnb-4bit for fact-checking and misinformation detection.

Model Description

  • Base Model: unsloth/gemma-2-9b-it-bnb-4bit
  • Fine-tuned for: Fact-checking and bias detection
  • Training Method: LoRA/QLoRA with Unsloth
  • Merged: Yes (adapter merged with base model)

Usage

from unsloth import FastLanguageModel

model, tokenizer = FastLanguageModel.from_pretrained(
    model_name="Adi-Evolve/gemma-2-9b-linkscout-v1",
    max_seq_length=2048,
    dtype=None,
    load_in_4bit=True,
)

FastLanguageModel.for_inference(model)

# Generate analysis
prompt = """<start_of_turn>user
Analyze this article for misinformation: [ARTICLE TEXT]
<end_of_turn>
<start_of_turn>model
"""

inputs = tokenizer([prompt], return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Training Details

  • Merged from LoRA adapter
  • Original base: unsloth/gemma-2-9b-it-bnb-4bit

Limitations

This model should be used as a tool to assist in fact-checking, not as a sole source of truth.

Downloads last month
61
Safetensors
Model size
9B params
Tensor type
F32
F16
U8
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for Adi-Evolve/gemma-2-9b-linkscout-v1

Quantizations
1 model