Available Models
Select a model to view details and integration examples
Code Examples
Loading code examples...
Select a model to view details and integration examples
Loading code examples...
Select a model to view details and integration examples
Loading code examples...
Choose from our available models and get started with API integration
Select a model to view details and integration examples
DeepSeek R1 Distill Llama 70B is a distilled large language model based on Llama-3.3-70B-Instruct, using outputs from DeepSeek R1. The model combines advanced distillation techniques to achieve high performance across multiple benchmarks, including: AIME 2024 pass@1: 70., MATH-500 pass@1: 94.5, CodeForces Rating: 1633.
Qwen3-32B is a dense 32.8B parameter causal language model from the Qwen3 series, optimized for both complex reasoning and efficient dialogue. It supports seamless switching between a thinking mode for tasks like math, coding, and logical inference, and a non-thinking mode for faster, general-purpose conversation.
Devstral-Small-2507 is a 24B parameter agentic LLM fine-tuned from Mistral-Small-3.1, jointly developed by Mistral AI and All Hands AI for advanced software engineering tasks. It is optimized for codebase exploration, multi-file editing, and integration into coding agents, achieving state-of-the-art results on SWE-Bench Verified (46.8%).
The Meta Llama 3.3 multilingual large language model (LLM) is a pretrained and instruction tuned generative model in 70B (text in/text out). The Llama 3.3 instruction tuned text only model is optimized for multilingual dialogue use cases and outperforms many of the available open source and closed chat models on common industry benchmarks.
DeepSeek R1 Distill Llama 70B is a distilled large language model based on Llama-3.3-70B-Instruct, using outputs from DeepSeek R1. The model combines advanced distillation techniques to achieve high performance across multiple benchmarks, including: AIME 2024 pass@1: 70., MATH-500 pass@1: 94.5, CodeForces Rating: 1633.
Direct API endpoints for DeepSeek R1 Distilled Llama 70B (AWQ)
curl --location 'https://tandemn-frontend.vercel.app/api/v1/chat/completions' \
--header 'Authorization: Bearer YOUR_API_KEY' \
--header 'Accept: text/event-stream' \
--header 'Cache-Control: no-cache' \
--header 'Content-Type: application/json' \
--data '{
"model": "casperhansen/deepseek-r1-distill-llama-70b-awq",
"messages": [
{
"role": "system",
"content": "You are a helpful assistant"
},
{
"role": "user",
"content": "Hello! Can you explain quantum computing?"
}
],
"stream": true,
"temperature": 0.7,
"max_completion_tokens": 2000
}'