Learn about HelpingAI's available models, their capabilities, and how to choose the right model for your use case.
Our flagship model with revolutionary Chain of Recursive Thoughts capabilities and token efficiency.
Model ID: Dhanishtha-2.0-preview
hideThink parameter| Feature | Specification |
|---|---|
| Context Length | 32,768 tokens |
| Training Data | Up to April 2024 |
| Languages | Optimized for English, supports 50+ languages |
| Input Cost | $0.50 per 1M tokens |
| Output Cost | $1.50 per 1M tokens |
| Max Output | 4,000 tokens |
| Feature | Dhanishtha-2.0-preview | GPT-4 | Deepseek |
|---|---|---|---|
| Chain of Recursive Thoughts | ✅ Mid-response reasoning | ❌ Pre-processing only | ❌ Pre-processing only |
| Token Efficiency | ✅ 5x fewer tokens | ❌ Standard | ❌ Standard |
| Inference Speed | ✅ 4x faster | ❌ Standard | ❌ Standard |
| Context Length | 32K tokens | 128K tokens | 200K tokens |
| Tool Calling | ✅ Advanced | ✅ Standard | ✅ Standard |
| Streaming | ✅ Yes | ✅ Yes | ✅ Yes |
| Cost per Quality Unit | ✅ Lower | ❌ Higher | ❌ Higher |
| hideThink Parameter | ✅ Unique | ❌ Not available | ❌ Not available |
import requests
url = "https://api.helpingai.co/v1/chat/completions"
headers = {
"Authorization": "Bearer YOUR_API_KEY",
"Content-Type": "application/json"
}
data = {
"model": "Dhanishtha-2.0-preview",
"messages": [
{"role": "user", "content": "Can you help me solve this step by step?"}
]
}
response = requests.post(url, headers=headers, json=data)
print(response.json())from openai import OpenAI
client = OpenAI(
base_url="https://api.helpingai.co/v1",
api_key="YOUR_API_KEY"
)
response = client.chat.completions.create(
model="Dhanishtha-2.0-preview",
messages=[
{"role": "user", "content": "Can you help me solve this step by step?"}
]
)
print(response.choices[0].message.content)from helpingai import HelpingAI
client = HelpingAI(api_key="YOUR_API_KEY")
response = client.chat.completions.create(
model="Dhanishtha-2.0-preview",
messages=[
{"role": "user", "content": "Can you help me solve this step by step?"}
]
)
print(response.choices[0].message.content)const axios = require("axios");
const response = await axios.post(
"https://api.helpingai.co/v1/chat/completions",
{
model: "Dhanishtha-2.0-preview",
messages: [
{
role: "user",
content: "Can you help me solve this step by step?",
},
],
},
{
headers: {
Authorization: "Bearer YOUR_API_KEY",
"Content-Type": "application/json",
},
}
);
console.log(response.data.choices[0].message.content);import OpenAI from "openai";
const openai = new OpenAI({
baseURL: "https://api.helpingai.co/v1",
apiKey: "YOUR_API_KEY",
});
const response = await openai.chat.completions.create({
model: "Dhanishtha-2.0-preview",
messages: [
{
role: "user",
content: "Can you help me solve this step by step?",
},
],
});
console.log(response.choices[0].message.content);import { HelpingAI } from "helpingai";
const client = new HelpingAI({
apiKey: "YOUR_API_KEY",
});
const response = await client.chat.completions.create({
model: "Dhanishtha-2.0-preview",
messages: [
{
role: "user",
content: "Can you help me solve this step by step?",
},
],
});
console.log(response.choices[0].message.content);Dhanishtha-2.0-preview uses 5x fewer tokens than Deepseek while maintaining quality:
# Efficient reasoning without redundant chains {#efficient-reasoning-without-redundant-chains}
response = client.chat.completions.create(
model="Dhanishtha-2.0-preview",
messages=[
{"role": "user", "content": "Explain quantum computing in simple terms"}
]
)
# Gets high-quality response with fewer tokens {#gets-high-quality-response-with-fewer-tokens}
print(response.choices[0].message.content)
print(f"Tokens used: {response.usage.total_tokens}")See how the model thinks with the hideThink parameter:
response = client.chat.completions.create(
model="Dhanishtha-2.0-preview",
messages=[
{"role": "user", "content": "What's 15 * 24? Show your work."}
],
hideThink=False # Show reasoning process
)
# Response will include <think>...</think> tags with step-by-step reasoning {#response-will-include-thinkthink-tags-with-step-by-step-reasoning}
print(response.choices[0].message.content)Dhanishtha-2.0-preview excels at intelligent tool usage:
tools = [{
"type": "function",
"function": {
"name": "calculate_math",
"description": "Perform mathematical calculations",
"parameters": {
"type": "object",
"properties": {
"expression": {"type": "string"},
"show_steps": {"type": "boolean", "default": True}
}
}
}
}]
response = client.chat.completions.create(
model="Dhanishtha-2.0-preview",
messages=[
{"role": "user", "content": "Can you calculate 15 * 24 and show the steps?"}
],
tools=tools
)Recommended: Dhanishtha-2.0-preview
Why: Built-in emotional intelligence with 98 EQ score provides natural, empathetic responses that understand and validate human emotions.
Example Use Cases:
Recommended: Dhanishtha-2.0-preview with hideThink=false
Why: Chain of Recursive Thoughts engine shows step-by-step thinking process, making it ideal for educational and analytical applications.
Example Use Cases:
Recommended: Dhanishtha-2.0-preview
Why: 5x more token-efficient than comparable models while maintaining high quality output.
Example Use Cases:
Recommended: Dhanishtha-2.0-preview with streaming
Why: 4x faster inference with optimized streaming support for responsive user experiences.
Example Use Cases:
Dhanishtha-2.0-preview follows responsible AI guidelines:
max_tokens# Efficient approach {#efficient-approach}
response = client.chat.completions.create(
model="Dhanishtha-2.0-preview",
messages=[
{"role": "user", "content": "Summarize climate change in 50 words"}
],
max_tokens=75, # Limit response length
temperature=0.3 # Lower temperature for factual content
)
# Less efficient approach {#less-efficient-approach}
response = client.chat.completions.create(
model="Dhanishtha-2.0-preview",
messages=[
{"role": "user", "content": "Please provide me with a very detailed and comprehensive explanation about climate change including all the scientific details, historical context, current impacts, and future predictions with extensive examples and case studies"}
],
max_tokens=2000, # Very long response
temperature=0.8 # Higher creativity (more tokens)
)We're continuously developing new models and capabilities:
Stay updated on new releases: