Refer your friends today and both of you get $10 in LLM API credits—that’s up to $500 in total rewards waiting for you!
Llama 3.2 1B, Qwen2.5 7B, Qwen 3 (0.6B, 1.7B, 4B) ,GLM 4 — all available now on Novita AI to supercharge your projects without spending a dime!
DeepSeek R1 0528 offers cutting-edge AI capabilities with its 685B parameter Mixture-of-Experts architecture, excelling in reasoning, coding, and multilingual tasks.
However, its significant hardware requirements make local deployment challenging. For smaller-scale needs, DeepSeek R1 0528 Qwen 3 8B provides a compact and efficient alternative.
Alternatively, cloud-based solutions like Novita AI eliminate infrastructure challenges, offering scalable and cost-effective access to DeepSeek models.
How Many Types Does Deepseek R1 0528 Have?
Deepseek R1 0528
Model Size: 685 billion parameters
Open Source: Yes
Architecture: Mixture of Experts (MoE)
Language Support: Multilingual, excels in English and Chinese
Supported Modalities: Text-to-Text
Training Method: In the latest update, the model’s depth of reasoning and inference capabilities was significantly enhanced using increased computational resources and algorithmic optimizations during post-training.
Deepseek R1 0528 Qwen 3 8B
Model Size: 8.19 billion parameters
Open Source: Yes
Architecture: Transformer
Language Support: Multilingual, excels in English and Chinese
Supported Modalities: Text-to-Text
Training Method: Post-trained with the chain-of-thought distilled from DeepSeek-R1-0528, resulting in DeepSeek-R1-0528-Qwen3-8B.
Deepseek R1 0528 Benchmark

How Much to Run Deepseek R1 0528?
Here is an overview of the hardware requirements for Deepseek r1 0528 and Deepseek r1 0528 Qwen 3 8B, highlighting their respective configurations and system needs:
Hardware Requirements
Deepseek r1 0528 Full Version
- Model Size: ~1900GB
- Hardware Configuration:
- 24 × NVIDIA H100 GPUs (80GB memory each); 8 x H200 SXM 141GB
- Total GPU Memory: 1920GB
- System RAM:
- Recommended: ≥512GB
- Optimal: 1TB (for GPU offload, KV cache, parallel tasks)
- Storage:
- High-speed NVMe SSD
- Capacity: ≥500GB
- CPU:
- Multi-core, high-frequency processors (e.g., Dual Intel Xeon or AMD EPYC)
- Cooling & Power:
- Enterprise-grade cooling and power systems
- Typical power consumption: several kW
Deepseek r1 0528 Qwen 3 8B
- Model Size: 18.72GB
- Hardware Configuration:
- 1× NVIDIA RTX 4090 GPU (24GB memory)
While DeepSeek R1 Qwen 3 8B provides a viable option for local or resource-constrained deployments, the larger DeepSeek R1 configurations deliver superior performance across all benchmarks, particularly in demanding tasks like coding and reasoning.
DeepSeek R1 0528 Locally: Efficient Yet Challenging
1. Hardware and Cost Constraints
- High GPU Requirements: 24× H100 GPUs are prohibitively expensive and require a large-scale data center. Each H100 GPU costs tens of thousands of dollars.
- Large System RAM: A minimum of 512GB RAM, ideally 1TB, is far beyond standard consumer-grade hardware.
- Storage Needs: High-speed NVMe SSDs with large capacities are essential, adding significant cost.
2. Power and Cooling
- Power Consumption: The system requires several kW of power, which exceeds the capabilities of a typical home or office setup.
- Cooling: Enterprise-grade cooling systems (e.g., water cooling) are needed to prevent overheating, which is difficult to achieve locally.
3. Physical Space
- Size of the System: Rack-mounted servers for 24 GPUs require significant physical space, likely unavailable in a home or small office.
4. Expertise and Software
- Maintenance: Managing such a powerful system involves ongoing maintenance, which may be challenging without a dedicated IT team.
- System Setup: Setting up distributed training or inference on 24 GPUs requires expertise in cluster management and software like PyTorch, NCCL, or DeepSpeed.
Accessing DeepSeek R1 0528 Alternative: API Like Novita AI
- Cloud-Based Access
Novita AI leverages powerful cloud infrastructure, eliminating the need for expensive local hardware. This allows users to access advanced AI capabilities from any device with an internet connection.
- Easy to Use
With Novita AI, there’s no need for complex installations or dependency management. Users can seamlessly access its features via a web interface or API, avoiding the technical challenges associated with deploying DeepSeek V3.
- Cost-Effective
Instead of investing in costly GPUs and incurring high power consumption, Novita AI offers a pay-as-you-go model, making it a more affordable option for a wide range of use cases.
Novita AI is an AI cloud platform that offers developers an easy way to deploy AI models using our simple API, while also providing the affordable and reliable GPU cloud for building and scaling.
Llama 3.2 1B, Qwen2.5 7B, Qwen 3 (0.6B, 1.7B, 4B) ,GLM 4 — all available now on Novita AI to supercharge your projects without spending a dime!
Step 1: Log In and Access the Model Library
Log in to your account and click on the Model Library button.

Step 2: Choose Your Model
Browse through the available options and select the model that suits your needs.

Step 3: Start Your Free Trial
Begin your free trial to explore the capabilities of the selected model.

Step 4: Get Your API Key
To authenticate with the API, we will provide you with a new API key. Entering the “Settings“ page, you can copy the API key as indicated in the image.

Step 5: Install the API
Install API using the package manager specific to your programming language.
After installation, import the necessary libraries into your development environment. Initialize the API with your API key to start interacting with Novita AI LLM. This is an example of using chat completions API for python users.
from openai import OpenAI
client = OpenAI(
base_url="https://api.novita.ai/v3/openai",
api_key="session_H_85jwhkUyBsRipBTIU9n_adbP5B9Qvu0wxGGMN4Vq-BpFVKntQQXOAJF4IpkuDJh2e-NQkoJkcwMhus4t81PQ==",
)
model = "deepseek/deepseek-r1-0528-qwen3-8b"
stream = True # or False
max_tokens = 16000
system_content = ""Be a helpful assistant""
temperature = 1
top_p = 1
min_p = 0
top_k = 50
presence_penalty = 0
frequency_penalty = 0
repetition_penalty = 1
response_format = { "type": "text" }
chat_completion_res = client.chat.completions.create(
model=model,
messages=[
{
"role": "system",
"content": system_content,
},
{
"role": "user",
"content": "Hi there!",
}
],
stream=stream,
max_tokens=max_tokens,
temperature=temperature,
top_p=top_p,
presence_penalty=presence_penalty,
frequency_penalty=frequency_penalty,
response_format=response_format,
extra_body={
"top_k": top_k,
"repetition_penalty": repetition_penalty,
"min_p": min_p
}
)
if stream:
for chunk in chat_completion_res:
print(chunk.choices[0].delta.content or "", end="")
else:
print(chat_completion_res.choices[0].message.content)
Due to the high hardware requirements of DeepSeek R1, we encourage everyone to use Novita AI, a cloud-based platform that provides cost-effective and scalable access to advanced AI models without the need for expensive infrastructure.
Frequently Asked Questions
DeepSeek R1 0528: 685B parameters, Mixture-of-Experts architecture, requires 24× H100 GPUs.
DeepSeek R1 0528 Qwen 3 8B: 8.19B parameters, Transformer architecture, runs on a single RTX 4090 GPU.
MoE dynamically activates subsets of parameters (“experts”) for specific tasks, improving computational efficiency for high-complexity tasks, but it demands advanced hardware.
Local deployment is possible but requires enterprise-grade hardware, including 1920GB GPU memory and several kW of power. Cloud platforms like Novita AI provide a practical alternative.
Novita AI is the All-in-one cloud platform that empowers your AI ambitions. Integrated APIs, serverless, GPU Instance — the cost-effective tools you need. Eliminate infrastructure, start free, and make your AI vision a reality.
Recommend Reading
- DeepSeek V3: Advancing Open-Source Code Models, Now Available on Novita AI
- Deepseek v3 vs Llama 3.3 70b: Language Tasks vs Code & Math
- Llama 3.2 3B vs DeepSeek V3: Comparing Efficiency and Performance.
Discover more from Novita
Subscribe to get the latest posts sent to your email.





