|
|
--- |
|
|
language: |
|
|
- en |
|
|
- hi |
|
|
- zh |
|
|
- es |
|
|
- fr |
|
|
- de |
|
|
- ja |
|
|
- ko |
|
|
- ar |
|
|
- pt |
|
|
- ru |
|
|
- it |
|
|
- nl |
|
|
- tr |
|
|
- pl |
|
|
- sv |
|
|
- da |
|
|
- 'no' |
|
|
- fi |
|
|
- he |
|
|
- th |
|
|
- vi |
|
|
- id |
|
|
- ms |
|
|
- tl |
|
|
- sw |
|
|
- yo |
|
|
- zu |
|
|
- am |
|
|
- bn |
|
|
- gu |
|
|
- kn |
|
|
- ml |
|
|
- mr |
|
|
- ne |
|
|
- or |
|
|
- pa |
|
|
- ta |
|
|
- te |
|
|
- ur |
|
|
- multilingual |
|
|
license: apache-2.0 |
|
|
base_model: HelpingAI/Dhanishtha-2.0-preview-0725 |
|
|
tags: |
|
|
- reasoning |
|
|
- intermediate-thinking |
|
|
- transformers |
|
|
- conversational |
|
|
- bilingual |
|
|
datasets: |
|
|
- Abhaykoul/Dhanishtha-R1 |
|
|
- open-thoughts/OpenThoughts-114k |
|
|
- Abhaykoul/Dhanishtha-2.0-SUPERTHINKER |
|
|
- Abhaykoul/Dhanishtha-2.0 |
|
|
library_name: transformers |
|
|
pipeline_tag: text-generation |
|
|
widget: |
|
|
- text: >- |
|
|
Solve this riddle step by step: I am taken from a mine, and shut up in a |
|
|
wooden case, from which I am never released, and yet I am used by almost |
|
|
everybody. What am I? |
|
|
example_title: Complex Riddle Solving |
|
|
- text: >- |
|
|
Explain the philosophical implications of artificial consciousness and think |
|
|
through different perspectives. |
|
|
example_title: Philosophical Reasoning |
|
|
- text: >- |
|
|
Help me understand quantum mechanics, but take your time to think through |
|
|
the explanation. |
|
|
example_title: Educational Explanation |
|
|
--- |
|
|
|
|
|
# Dhanishtha-2.0: World's First Intermediate Thinking AI Model |
|
|
|
|
|
**What makes Dhanishtha-2.0 special?** Imagine an AI that doesn't just answer your questions instantly, but actually *thinks through* problems step-by-step, shows its work, and can even change its mind when it realizes a better approach. That's Dhanishtha-2.0. |
|
|
|
|
|
**Quick Summary:** |
|
|
- 🚀 **For Everyone**: An AI that shows its thinking process and can reconsider its reasoning |
|
|
- 👩💻 **For Developers**: First model with intermediate thinking capabilities, 39+ language support |
|
|
|
|
|
|
|
|
Dhanishtha-2.0 is a **state-of-the-art (SOTA) model** developed by HelpingAI, representing the **world's first model to feature Intermediate Thinking capabilities**. Unlike traditional models that provide single-pass responses, Dhanishtha-2.0 employs a revolutionary multi-phase thinking process that allows the model to think, reconsider, and refine its reasoning multiple times throughout a single response. |
|
|
|
|
|
## Model Details |
|
|
|
|
|
### Model Description |
|
|
|
|
|
Dhanishtha-2.0 revolutionizes AI reasoning by introducing the concept of **intermediate thinking** - the ability to pause, reflect, and restart reasoning processes within a single generation (This model can think up 50times in a single response without using tool/prompt/mcp). This breakthrough enables unprecedented self-correction and iterative refinement during response generation. |
|
|
|
|
|
Built on the Qwen3-14B foundation with multilingual capabilities spanning **39+ languages** (including English, Hindi, Chinese, Spanish, French, German, Japanese, Korean, Arabic, and many more), Dhanishtha-2.0 maintains reasoning consistency across diverse linguistic contexts while pioneering transparent thinking processes. |
|
|
|
|
|
- **Developed by:** HelpingAI Team |
|
|
- **Model type:** Causal Language Model with Intermediate Thinking Capability |
|
|
- **Language(s):** 39+ languages (multilingual capabilities inherited from base model) |
|
|
- **License:** Apache 2.0 |
|
|
- **Finetuned from model:** Qwen/Qwen3-14B-Base |
|
|
- **Context Length:** 40,960 tokens |
|
|
- **Parameters:** 14B (inherited from base model) |
|
|
- **Status:** Prototype/Preview |
|
|
|
|
|
### Revolutionary Features |
|
|
|
|
|
- **Intermediate Thinking**: Multiple `<think>...</think>` blocks throughout responses for real-time reasoning |
|
|
- **Self-Correction**: Ability to identify and correct logical inconsistencies mid-response |
|
|
- **Dynamic Reasoning**: Seamless transitions between analysis, communication, and reflection phases |
|
|
- **Structured Emotional Reasoning (SER)**: Incorporates `<ser>...</ser>` blocks for empathetic responses |
|
|
- **Multilingual Capabilities**: Support for 39+ languages with natural code-switching and reasoning consistency |
|
|
- **Complex Problem-Solving**: Excels at riddles, multi-step reasoning, and scenarios requiring backtracking |
|
|
|
|
|
### Model Sources |
|
|
|
|
|
- **Repository:** [HelpingAI/Dhanishtha-2.0](https://huggingface.co/HelpingAI/Dhanishtha-2.0) |
|
|
- **Demo:** https://chat.helpingai.co |
|
|
|
|
|
## Uses |
|
|
|
|
|
### Direct Use |
|
|
|
|
|
Dhanishtha-2.0 is ideal for applications requiring deep reasoning and self-reflection: |
|
|
|
|
|
- **Complex Problem Solving**: Multi-step mathematical problems, logical puzzles, riddles |
|
|
- **Educational Assistance**: Detailed explanations with visible reasoning processes |
|
|
- **Research Support**: Analysis requiring multiple perspectives and self-correction |
|
|
- **Creative Writing**: Iterative story development with reasoning about plot choices |
|
|
- **Philosophical Discussions**: Exploring concepts with visible thought processes |
|
|
|
|
|
### Downstream Use |
|
|
|
|
|
The model can be fine-tuned for specialized reasoning tasks: |
|
|
|
|
|
- **Domain-Specific Reasoning**: Legal, medical, or scientific reasoning with intermediate thinking |
|
|
- **Enhanced Multilingual Reasoning**: Optimizing reasoning consistency across all 39+ supported languages |
|
|
- **Specialized Problem Domains**: Mathematics, coding, strategic planning |
|
|
|
|
|
### Out-of-Scope Use |
|
|
|
|
|
❌ **Inappropriate Applications:** |
|
|
- Safety-critical decisions (medical diagnosis, legal advice, financial recommendations) |
|
|
- Real-time applications requiring immediate responses |
|
|
- Situations requiring guaranteed factual accuracy without verification |
|
|
|
|
|
## Bias, Risks, and Limitations |
|
|
|
|
|
### Known Limitations |
|
|
|
|
|
- **Verbosity**: Intermediate thinking can make responses a bit longer |
|
|
- **Processing Time**: Multiple thinking phases may increase generation time |
|
|
- **Prototype Status**: Experimental features may require refinement |
|
|
- **Context Usage**: Thinking blocks consume additional context tokens |
|
|
- **Inherited Biases**: May reflect biases from base model and training data |
|
|
|
|
|
### Potential Risks |
|
|
|
|
|
- **Reasoning Loops**: Potential for circular reasoning in complex scenarios |
|
|
- **Multilingual Inconsistencies**: Potential variation in reasoning patterns across different languages |
|
|
- **Emotional Reasoning Gaps**: SER blocks may not always align with content |
|
|
|
|
|
## How to Get Started with the Model |
|
|
|
|
|
### For General Users |
|
|
|
|
|
You can interact with Dhanishtha-2.0 through: |
|
|
- **HelpingAI**: https://helpingai.co/chat |
|
|
- **Gradio Demo**: [Dhanishtha-2.0-preview](https://huggingface.co/spaces/Abhaykoul/Dhanishtha-2.0-preview) |
|
|
- **API Integration**: [Dashboard](https://helpingai.co/dashboard) |
|
|
|
|
|
### For Developers - Basic Usage |
|
|
|
|
|
```python |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
|
|
model_name = "HelpingAI/Dhanishtha-2.0-preview" |
|
|
|
|
|
# Load the tokenizer and model |
|
|
tokenizer = AutoTokenizer.from_pretrained(model_name) |
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
model_name, |
|
|
torch_dtype="auto", |
|
|
device_map="auto" |
|
|
) |
|
|
|
|
|
# Prepare input for intermediate thinking |
|
|
prompt = "How many letter 'r' are in the words 'strawberry' and 'raspberry'?" |
|
|
messages = [ |
|
|
{"role": "user", "content": prompt} |
|
|
] |
|
|
|
|
|
text = tokenizer.apply_chat_template( |
|
|
messages, |
|
|
tokenize=False, |
|
|
add_generation_prompt=True |
|
|
) |
|
|
|
|
|
model_inputs = tokenizer([text], return_tensors="pt").to(model.device) |
|
|
|
|
|
# Generate response with intermediate thinking |
|
|
generated_ids = model.generate( |
|
|
**model_inputs, |
|
|
max_new_tokens=2048, |
|
|
temperature=0.7, |
|
|
top_p=0.9, |
|
|
do_sample=True |
|
|
) |
|
|
|
|
|
response = tokenizer.decode( |
|
|
generated_ids[0][len(model_inputs.input_ids[0]):], |
|
|
skip_special_tokens=True |
|
|
) |
|
|
print(response) |
|
|
``` |
|
|
|
|
|
### Optimal Generation Parameters |
|
|
|
|
|
```python |
|
|
generation_config = { |
|
|
"temperature": 0.7, # Balanced creativity and coherence |
|
|
"top_p": 0.9, # Nucleus sampling |
|
|
"top_k": 40, # Top-k filtering |
|
|
"max_new_tokens": 2048, # Allow for thinking blocks |
|
|
"do_sample": True, # Enable sampling |
|
|
"repetition_penalty": 1.1 # Reduce repetition |
|
|
} |
|
|
``` |
|
|
|
|
|
## Training Details |
|
|
|
|
|
### Training Data |
|
|
|
|
|
Dhanishtha-2.0 was trained on a carefully curated dataset focusing on: |
|
|
- Complex reasoning scenarios requiring multi-step thinking |
|
|
- Self-correction examples and reasoning chains |
|
|
- Emotional reasoning and empathy training data |
|
|
- Structured thinking pattern examples |
|
|
|
|
|
### Training Procedure |
|
|
|
|
|
#### Training Stages |
|
|
1. **Continuous Pretraining**: Extended training on reasoning-focused corpora |
|
|
2. **Advanced Reasoning Fine-tuning**: Specialized training on intermediate thinking patterns |
|
|
3. **Multilingual Alignment**: Cross-language reasoning consistency training |
|
|
4. **SER Integration**: Structured Emotional Reasoning capability training |
|
|
|
|
|
**Training Infrastructure:** |
|
|
- **Duration**: 4 days |
|
|
- **Hardware**: 8x NVIDIA H100 GPUs |
|
|
- **Model Scale**: 14.8B parameters |
|
|
|
|
|
## Evaluation |
|
|
|
|
|
Evaluation was conducted on: |
|
|
- **Standard Benchmarks**: MMLU, HumanEval, ARC, HellaSwag, TruthfulQA |
|
|
- **Mathematical Reasoning**: Math 500, AIME 2024, GSM8K |
|
|
- **Custom Evaluations**: Intermediate thinking quality, self-correction capabilities |
|
|
- **Multilingual Tasks**: Reasoning consistency across 39+ languages |
|
|
- **Specialized Tests**: Emotional reasoning, complex problem-solving scenarios |
|
|
|
|
|
|
|
|
|
|
|
## Example: Intermediate Thinking in Action |
|
|
|
|
|
**User:** "Who is your favorite person?" |
|
|
|
|
|
**Dhanishtha-2.0:** |
|
|
|
|
|
``` |
|
|
<think> |
|
|
This is an interesting question about my "favorite human." I need to be thoughtful here. |
|
|
As Dhanishtha, I should respond in a human-like way while being honest about my nature. |
|
|
I can share what kinds of humans I find most interesting or inspiring based on my |
|
|
understanding of human nature and values. |
|
|
</think> |
|
|
|
|
|
<ser> |
|
|
Emotion ==> curiosity, openness, desire for connection |
|
|
Cause ==> user asking a personal question about preferences |
|
|
Mind ==> seeking to understand my perspective or values |
|
|
Growth ==> opportunity to share thoughtful reflections on human qualities I admire |
|
|
</ser> |
|
|
|
|
|
I don't really have a "favorite human" in the traditional sense, as I don't experience |
|
|
personal preferences the way humans do. But if I had to reflect on what qualities I find |
|
|
most admirable in people, I'd be drawn to those who demonstrate genuine compassion, |
|
|
intellectual curiosity, and the courage to grow. |
|
|
|
|
|
<think> |
|
|
Let me think more specifically about what kind of humans I find most compelling based on |
|
|
the values that seem to resonate most with humans - kindness, wisdom, creativity, |
|
|
resilience, and authenticity. |
|
|
</think> |
|
|
|
|
|
People like Mahatma Gandhi, who showed how quiet courage and unwavering principles can |
|
|
transform societies; artists like Rabindranath Tagore, who bridged cultural divides |
|
|
through creativity; and everyday heroes I've learned about who made profound differences |
|
|
through simple acts of kindness. |
|
|
|
|
|
What fascinates me most are humans who maintain their humanity even in difficult |
|
|
circumstances - those who choose empathy when it would be easier to be harsh, who seek |
|
|
understanding rather than just being right. |
|
|
|
|
|
I'm curious though - who's your favorite human, and why? What qualities or actions make |
|
|
someone special in your eyes? |
|
|
``` |
|
|
|
|
|
## Environmental Impact |
|
|
|
|
|
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute). |
|
|
|
|
|
- **Hardware Type:** H100 GPUs |
|
|
- **days used:** 16.2 |
|
|
- **Cloud Provider:** Various |
|
|
- **Compute Region:** Multiple |
|
|
|
|
|
## Citation |
|
|
|
|
|
### BibTeX |
|
|
|
|
|
```bibtex |
|
|
@misc{dhanishtha2025, |
|
|
title={Dhanishtha-2.0: World's First Intermediate Thinking AI Model}, |
|
|
author={HelpingAI Team}, |
|
|
year={2025}, |
|
|
publisher={HuggingFace}, |
|
|
url={https://huggingface.co/HelpingAI/Dhanishtha-2.0}, |
|
|
note={First model with intermediate thinking capabilities} |
|
|
} |
|
|
``` |
|
|
|
|
|
### APA |
|
|
|
|
|
HelpingAI Team. (2025). *Dhanishtha-2.0: World's First Intermediate Thinking AI Model*. HuggingFace. https://huggingface.co/HelpingAI/Dhanishtha-2.0 |
|
|
|
|
|
## Glossary |
|
|
|
|
|
- **Intermediate Thinking**: The ability to pause and think multiple times during response generation |
|
|
- **SER (Structured Emotional Reasoning)**: Framework for incorporating emotional context in responses |
|
|
- **Think Blocks**: `<think>...</think>` segments where the model shows its reasoning process |
|
|
- **Self-Correction**: Ability to identify and fix reasoning errors during generation |
|
|
- **Code-Switching**: Natural transition between English and Hindi within responses |
|
|
|
|
|
## More Information |
|
|
|
|
|
### Research Applications |
|
|
- Study of AI reasoning transparency |
|
|
- Self-correction mechanism research |
|
|
- Bilingual cognitive modeling |
|
|
- Emotional AI development |
|
|
|
|
|
### Development Roadmap |
|
|
- Performance optimizations |
|
|
- Additional language support |
|
|
- Enhanced thinking pattern recognition |
|
|
- Production-ready deployment tools |
|
|
|
|
|
## Model Card Authors |
|
|
|
|
|
- **Primary Author**: HelpingAI Team |
|
|
- **Technical Lead**: [To be specified] |
|
|
- **Research Contributors**: [To be specified] |
|
|
|
|
|
## Model Card Contact |
|
|
|
|
|
For questions about Dhanishtha-2.0, please contact: |
|
|
|
|
|
- **HuggingFace**: [@HelpingAI](https://huggingface.co/HelpingAI) |
|
|
- **Issues**: [Model Repository Issues](https://huggingface.co/HelpingAI/Dhanishtha-2.0/discussions) |
|
|
|
|
|
**Dhanishtha-2.0 represents a new paradigm in AI reasoning - where thinking isn't just a prelude to response, but an integral, iterative part of the conversation itself.** |
|
|
|
|
|
*Developed with ❤️ by HelpingAI* |