DeepSeek-R1-0528 and DeepSeek-R1-0528-Qwen3-8B Now Available on Novita AI

DeepSeek-R1-0528-Qwen3-8B

DeepSeek has released two groundbreaking models—DeepSeek-R1-0528 and DeepSeek-R1-0528-Qwen3-8B — both now available on Novita AI, an AI cloud platform that offers developers easy AI model deployment through our straightforward API. These models deliver exceptional reasoning capabilities at competitive prices, offering significantly better performance than their predecessors.

Novita AI offers both DeepSeek-R1-0528 models at competitive rates with transparent, pay-per-use pricing:

DeepSeek-R1-0528 (Full Model):

  • Input tokens: $0.7 per million tokens
  • Output tokens: $2.5 per million tokens

DeepSeek-R1-0528-Qwen3-8B (Efficient 8B Model):

  • Input tokens: $0.06 per million tokens
  • Output tokens: $0.09 per million tokens

What is DeepSeek-R1-0528?

The DeepSeek R1 model has undergone a minor version upgrade, with the current version being DeepSeek-R1-0528. In this latest update, DeepSeek R1 has significantly improved its depth of reasoning and inference capabilities by leveraging increased computational resources and introducing algorithmic optimization mechanisms during post-training.

The model demonstrates outstanding performance across various benchmark evaluations, including mathematics, programming, and general logic. Its overall performance now approaches that of leading models, such as O3 and Gemini 2.5 Pro.

Key Improvements of DeepSeek-R1-0528

🔹 Improved benchmark performance across mathematics, coding, and reasoning tasks
🔹 Enhanced front-end capabilities for better user experience
🔹 Reduced hallucinations for more reliable outputs
🔹 Supports JSON output & function calling for seamless integration

Enhanced Reasoning Depth

Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks. For instance, in the AIME 2025 test, the model’s accuracy increased from 70% in the previous version to 87.5% in the current version.

This advancement stems from enhanced thinking depth during the reasoning process: in the AIME test set, the previous model used an average of 12K tokens per question, whereas the new version averages 23K tokens per question.

Competitive Performance Against Industry Leaders

DeepSeek-R1-0528 demonstrates remarkable performance when compared to industry-leading models across multiple challenging benchmarks:

deepseek r1 0528 benchmark

DeepSeek-R1 vs DeepSeek-R1-0528

Mathematical Reasoning Performance

BenchmarkDeepSeek R1DeepSeek R1-0528Improvement
AIME 2024 (Pass@1)79.891.4+11.6
AIME 2025 (Pass@1)70.087.5+17.5
HMMT 2025 (Pass@1)41.779.4+37.7
CNMO 2024 (Pass@1)78.886.9+8.1

Coding Performance Enhancement

BenchmarkDeepSeek R1DeepSeek R1-0528Improvement
LiveCodeBench (Pass@1)63.573.3+9.8
Codeforces-Div1 (Rating)15301930+400
SWE Verified (Resolved)49.257.6+8.4
Aider-Polyglot (Acc.)53.371.6+18.3

General Reasoning Tasks

BenchmarkDeepSeek R1DeepSeek R1-0528Improvement
MMLU-Redux (EM)92.993.4+0.5
MMLU-Pro (EM)84.085.0+1.0
GPQA-Diamond (Pass@1)71.581.0+9.5
SimpleQA (Correct)30.127.8-2.3
FRAMES (Acc.)82.583.0+0.5
Humanity’s Last Exam (Pass@1)8.517.7+9.2

New Tool Integration Capabilities

DeepSeek-R1-0528 introduces enhanced capabilities in function calling and tool integration:

  • BFCL_v3_MultiTurn (Acc): 37.0%
  • Tau-Bench (Pass@1): 53.5% (Airline) / 63.9% (Retail)

DeepSeek-R1-0528-Qwen3-8B: Compact Excellence

DeepSeek distilled DeepSeek-R1-0528’s chain-of-thought reasoning to create DeepSeek-R1-0528-Qwen3-8B from Qwen3 8B Base. This model achieves state-of-the-art (SOTA) performance among open-source models on AIME 2024, surpassing Qwen3-8B by +10.0% and matching the performance of Qwen3-235B-thinking.

This demonstrates that the chain-of-thought from DeepSeek-R1-0528 holds significant importance for both academic research on reasoning models and industrial development focused on small-scale models.

Competitive Analysis with Leading Models

ModelAIME 24AIME 25HMMT Feb 25GPQA DiamondLiveCodeBench
Qwen3-235B-A22B85.781.562.571.166.5
Qwen3-32B81.472.968.4
Qwen3-8B76.067.362.0
Phi-4-Reasoning-Plus-14B81.378.053.669.3
Gemini-2.5-Flash-Thinking-052082.372.064.282.862.3
o3-mini (medium)79.676.753.376.865.9
DeepSeek-R1-0528-Qwen3-8B86.076.361.561.160.5

Usage Improvements and Technical Details

Enhanced Usability Features

Compared to previous versions of DeepSeek-R1, the usage recommendations for DeepSeek-R1-0528 include the following improvements:

  • System prompt support: Unlike earlier versions, you can now use system prompts to provide consistent context and instructions
  • Automatic thinking pattern: No manual activation required—the model automatically engages its reasoning capabilities without needing “<think>\n” at the beginning of outputs

DeepSeek-R1-0528-Qwen3-8B Architecture

The model architecture of DeepSeek-R1-0528-Qwen3-8B is identical to that of Qwen3-8B, but it shares the same tokenizer configuration as DeepSeek-R1-0528. This model can run in the same manner as Qwen3-8B, but ensure all configuration files come from DeepSeek’s repository, not the original Qwen3 project.

Why Developers Choose Novita AI

Simplified AI Model Deployment

  • No Infrastructure Hassle: Focus on building applications, not managing servers
  • Instant Model Access: Deploy state-of-the-art models with a single API call
  • Reliable Performance: Enterprise-grade infrastructure with 99.9% uptime
  • Global Availability: Low-latency access from anywhere in the world

Perfect for Any Scale

  • Rapid Prototyping: Test ideas quickly with immediate model access
  • Production Ready: Scale seamlessly from prototype to production
  • Cost Effective: Pay only for what you use with transparent pricing
  • Enterprise Support: Dedicated support for high-volume applications

Ideal Use Cases

  • Mathematical AI Applications: Build tutoring platforms achieving 87.5% accuracy on AIME 2025
  • Code Generation Tools: Create development assistants with 73.3% success on LiveCodeBench
  • Research Platforms: Deploy reasoning models that compete with industry leaders
  • Educational Software: Develop step-by-step problem-solving applications

Getting Started on Novita AI

  1. Sign up in seconds at Novita AI and claim your $10 free credits
  2. Explpre the DeepSeek-R1-0528 (full capability) or DeepSeek-R1-0528-Qwen3-8B (efficient)
  3. Get your API key from the dashboard
  4. Make your first API call using our OpenAI-compatible endpoints
  5. Start building with industry-leading reasoning capabilities

For python users:

from openai import OpenAI

base_url = "https://api.novita.ai/v3/openai"
api_key = "<Your API Key>"
model = "deepseek/deepseek-r1-0528"

client = OpenAI(
    base_url=base_url,
    api_key=api_key,
)

stream = True # or False
max_tokens = 1000

response_format = { "type": "text" }

chat_completion_res = client.chat.completions.create(
    model=model,
    messages=[
        
        {
            "role": "user",
            "content": "Hi there!",
        }
    ],
    stream=stream,
    extra_body={
    }
  )

if stream:
    for chunk in chat_completion_res:
        print(chunk.choices[0].delta.content or "", end="")
else:
    print(chat_completion_res.choices[0].message.content)
  
  

Conclusion

DeepSeek-R1-0528 series models deliver industry-leading performance at competitive prices—achieving 91.4% on AIME 2024 and 73.3% on LiveCodeBench while rivaling OpenAI-o3 and Gemini-2.5-Pro capabilities.

Through Novita AI’s developer-friendly platform, you get instant access to these cutting-edge models without infrastructure complexity. With transparent pricing, $10 free credits, and our intuitive API, integrating world-class reasoning into your applications has never been easier.

Novita AI is an AI cloud platform that offers developers an easy way to deploy AI models using our simple API, while also providing the affordable and reliable GPU cloud for building and scaling.


Discover more from Novita

Subscribe to get the latest posts sent to your email.

Leave a Comment

Scroll to Top

Discover more from Novita

Subscribe now to keep reading and get access to the full archive.

Continue reading