Key Highlights
Novita AI has introduced DeepSeek R1 Turbo, offering 3x throughput and limited-time 60% discount. Moreover, this version fully supports function calling.
You can start a free trail on Novita AI!

Even More Exciting: Novita AI Ranks 1 for DeepSeek R1 API on OpenRouter

What is DeepSeek R1?
- Release Date: January 21, 2025
- Model Scale:
- Key Features:
- Model Size: 671B parameters (37B active/token)
- Tokenizer: Enhanced tokenizer with self-reflection tags
- Supported Languages: Multilingual with cultural adaptation
- Multimodal: Text-only
- Context Window: 128K tokens
- Storage Formats: Q8/Q5 quantization support
- Architecture: Mixture of Experts (MoE) + RL-enhanced training pipeline
- Training Method: Built on V3 base with RL pipeline (SFT → RL → SFT → RL)
- Training Data: V3 base + RL optimization data

Benchmark

DeepSeek-R1 demonstrates significant performance improvements in most tasks, especially achieving a 9.0% and 5.7% increase in Codeforces and GPQA Diamond tasks, respectively, showcasing its strong capabilities in handling complex tasks.
Deepseek R1 Hardware Requirements
| Model | precision | GPU | Quantity |
|---|---|---|---|
| DeepSeek-R1-Distill-Qwen-1.5B | BF16 | RTX 4090 | 1 |
| DeepSeek-R1-Distill-Qwen-7B | BF16 | RTX 4090 | 1 |
| DeepSeek-R1-Distill-Llama-8B | BF16 | RTX 4090 | 1 |
| DeepSeek-R1-Distill-Qwen-14B | BF16 | RTX 4090 | 2 |
| DeepSeek-R1-Distill-Qwen-32B | BF16 | RTX 4090 | 4 |
| DeepSeek-R1-Distill-Llama-70B | BF16 | RTX 4090 | 8 |
| DeepSeek-R1 (671B) | BF16 | H100 | 8 |
The hardware requirements of DeepSeek-R1 are clearly stratified:
- Distilled models have relatively low hardware requirements, especially for small and medium scales, making them suitable for high-end consumer-grade hardware (e.g., RTX 4090).
- Full-scale models have extremely high hardware demands, especially ultra-large models, which rely on professional-grade hardware (e.g., H100 GPUs).
- Overall, distillation technology significantly reduces hardware requirements and enhances the flexibility of model deployment, but full-scale models still require very high-performance hardware to achieve peak performance.
If you find purchasing GPUs or deploying them yourself too complicated, you can use Novita AI’s GPU and API services, which offer a fully optimized version of DeepSeek-R1. It’s extremely convenient and charges based on the number of tokens.
Choosing Novita AI for Cloud GPU Services
Novita AI is an AI cloud platform that offers developers an easy way to deploy AI models using our simple API, while also providing the affordable and reliable GPU cloud for building and scaling.
Step1:Register an account
If you’re new to Novita AI, begin by creating an account on our website. Once you’re registered, head to the “GPUs” tab to explore available resources and start your journey.

Step2:Exploring Templates and GPU Servers
Start by selecting a template that matches your project needs, such as PyTorch, TensorFlow, or CUDA. Choose the version that fits your requirements, like PyTorch 2.2.1 or CUDA 11.8.0. Then, select the A100 GPU server configuration, which offers powerful performance to handle demanding workloads with ample VRAM, RAM, and disk capacity.

Step3:Tailor Your Deployment
After selecting a template and GPU, customize your deployment settings by adjusting parameters like the operating system version (e.g., CUDA 11.8). You can also tweak other configurations to tailor the environment to your project’s specific requirements.

Step4:Launch an instance
Once you’ve finalized the template and deployment settings, click “Launch Instance” to set up your GPU instance. This will start the environment setup, enabling you to begin using the GPU resources for your AI tasks.

Choosing Novita AI for API Services
Novita AI is an AI cloud platform that offers developers an easy way to deploy AI models using our simple API, while also providing the affordable and reliable GPU cloud for building and scaling.
Step 1: Log In and Access the Model Library
Log in to your account and click on the Model Library button.

Step 2: Choose Your Model
Browse through the available options and select the model that suits your needs.

Step 3: Start Your Free Trial
Begin your free trial to explore the capabilities of the selected model.

Step 4: Get Your API Key
To authenticate with the API, we will provide you with a new API key. Entering the “Settings“ page, you can copy the API key as indicated in the image.

Step 5: Install the API
Install API using the package manager specific to your programming language.

After installation, import the necessary libraries into your development environment. Initialize the API with your API key to start interacting with Novita AI LLM. This is an example of using chat completions API for python users.
from openai import OpenAI
client = OpenAI(
base_url="https://api.novita.ai/v3/openai",
api_key="<YOUR Novita AI API Key>",
)
model = "deepseek/deepseek_r1"
stream = True # or False
max_tokens = 2048
system_content = """Be a helpful assistant"""
temperature = 1
top_p = 1
min_p = 0
top_k = 50
presence_penalty = 0
frequency_penalty = 0
repetition_penalty = 1
response_format = { "type": "text" }
chat_completion_res = client.chat.completions.create(
model=model,
messages=[
{
"role": "system",
"content": system_content,
},
{
"role": "user",
"content": "Hi there!",
}
],
stream=stream,
max_tokens=max_tokens,
temperature=temperature,
top_p=top_p,
presence_penalty=presence_penalty,
frequency_penalty=frequency_penalty,
response_format=response_format,
extra_body={
"top_k": top_k,
"repetition_penalty": repetition_penalty,
"min_p": min_p
}
)
if stream:
for chunk in chat_completion_res:
print(chunk.choices[0].delta.content or "", end="")
else:
print(chat_completion_res.choices[0].message.content)
Novita AI revolutionizes AI deployment with DeepSeek R1 Turbo, offering 3x throughput and support for advanced features like function calling, now available with a limited-time 20% discount. Whether you’re running distilled models on consumer-grade GPUs like RTX 4090 or deploying full-scale models with enterprise-grade H100 GPUs, Novita AI provides unmatched flexibility. For those seeking a hassle-free solution, Novita AI’s GPU and API services deliver a fully optimized version of DeepSeek R1 Turbo, paired with token-based pricing and a free trial. Experience next-level AI performance with DeepSeek R1 Turbo—start with Novita AI today!
Frequently Asked Question
DeepSeek R1 Turbo is an optimized version of DeepSeek R1, offering 3x throughput, full support for function calling, and a limited-time 20% discount on Novita AI!
Full-scale models: Require enterprise-grade GPUs like H100 for optimal performance.
Yes, Novita AI offers a free trial for DeepSeek R1 via its API platform.
Novita AI is the All-in-one cloud platform that empowers your AI ambitions. Integrated APIs, serverless, GPU Instance — the cost-effective tools you need. Eliminate infrastructure, start free, and make your AI vision a reality.
Recommend Reading
- Why LLaMA 3.3 70B VRAM Requirements Are a Challenge for Home Servers?
- Qwen 2.5 72b vs Llama 3.3 70b: Which Model Suits Your Needs?
- Qwen 2.5 vs Llama 3.2 90B: A Comparative Analysis of Coding and Image Reasoning Capabilities
Discover more from Novita
Subscribe to get the latest posts sent to your email.





