bge-reranker-v2-m3 Now Available on Novita AI

bge-reranker-v2-m3 on Novita AI

Key Hightlights

Reranking Models: Reranking models are essential for optimizing search result quality by intelligently reordering candidate documents based on precise relevance scoring, ensuring users see the most pertinent information first.

BAAI/bge-reranker-v2-m3: The bge-reranker-v2-m3 model delivers exceptional cross-language reranking capabilities, supports 18+ languages, and excels at processing query-document pairs for accurate relevance assessment. Its precision and multilingual support make it indispensable for high-quality search experiences.

The powerful bge-reranker-v2-m3 model is available on Novita AI, offering cost-effective reranking solutions. Start your free trial on Novita AI!

Reranking is a cornerstone of modern search systems, empowering applications to deliver highly relevant results through sophisticated relevance analysis. This guide explores what rerankers (reranking models) are, why they are crucial for search quality, and how BAAI/bge-reranker-v2-m3 perform on elevating search capabilities to drive superior user experiences.

Understanding Rerankers

A reranker is a specialized AI model designed to refine and reorder search results by evaluating the semantic relevance between queries and candidate documents. Unlike traditional retrieval systems that focus on speed and recall, rerankers prioritize precision by computing detailed relevance scores for query-document pairs.

In modern AI applications, rerankers play a particularly crucial role in RAG (Retrieval-Augmented Generation) systems, where the quality of retrieved documents directly impacts the final generated response. The reranking stage acts as a precision filter, ensuring that only the most relevant documents proceed to the content generation phase.

Rag Pipeline with Reranking

Modern information retrieval systems typically employ a two-stage architecture as shown above. The first stage uses initial fast retrieval methods like dual-tower models, BM25, or vector databases to quickly filter candidate documents from large collections. The second stage employs rerankers to perform fine-grained sorting of these candidate results, focusing on precise relevance judgment.

Rerankers use a cross-encoder architecture that processes query and document jointly, enabling deeper interaction modeling compared to traditional retrieval methods. This architecture allows rerankers to identify complex semantic associations, synonyms, and hierarchical relationships that keyword-based matching might miss.

Applications of Rerankers

Rerankers excel in various application scenarios where semantic understanding and precise ranking are crucial:

Critical Domain Applications

Medical information systems: Retrieving relevant medical literature where accuracy is life-critical
Legal document retrieval: Finding precise case law and statutes for legal professionals
Scientific research: Surfacing the most relevant studies from thousands of papers

AI-Powered Systems

Question-Answering Systems: In RAG systems, rerankers ensure that the most semantically relevant documents are selected as context for answer generation, directly impacting response quality and accuracy
Recommendation Systems: Sort candidate items by relevance to user preferences, improving engagement and personalization

Enterprise Solutions

Enterprise Search: Large organizations use rerankers to optimize document retrieval across vast knowledge bases, making information discovery more efficient and accurate
Search Engines: Perform secondary ranking of search results to improve relevance and user satisfaction beyond traditional keyword matching

Cross-Platform Applications

Multilingual Platforms: Handle cross-language retrieval scenarios effectively, providing consistent search quality across different languages

Understanding BAAI/bge-reranker-v2-m3

BAAI/bge-reranker-v2-m3 is a lightweight multilingual reranker developed by the Beijing Academy of Artificial Intelligence (BAAI). As part of the BGE (Beijing General Embedding) series, this model is specifically optimized for reranking tasks across multiple languages.

Key Features:

  • Lightweight design: 568M parameters for efficient deployment
  • Multilingual support: 18+ languages with cross-language capabilities
  • Fast inference: Millisecond-level response times
  • Easy integration: Multiple APIs and libraries available

Technical Specifications:

  • Architecture: Cross-encoder sequence classification
  • Input format: [query text, document text]
  • Maximum length: 512 tokens
  • Output: Direct relevance scores with optional normalization
  • Acceleration: fp16/bf16 support for GPU inference

Supported Languages: English, Chinese, Japanese, Korean, Spanish, French, German, Russian, Arabic, Hindi, Bengali, Persian, Finnish, Indonesian, Thai, Telugu, Swahili, Yoruba

The model is trained on diverse datasets including bge-m3-data (multilingual retrieval), Quora train data (question-answer pairs), and FEVER train data (fact verification), ensuring robust performance across various domains and use cases.

llama-index.

llama-index.

BEIR.

BEIR.
rerank the top 100 results from bge-en-v1.5 large.
rerank the top 100 results from bge-en-v1.5 large.
BEIR. 
rerank the top 100 results from e5 mistral 7b instruct.
rerank the top 100 results from e5 mistral 7b instruct.

CMTEB-retrieval.

CMTEB-retrieval.
rerank the top 100 results from bge-zh-v1.5 large.

miracl (multi-language).

miracl (multi-language).
rerank the top 100 results from bge-m3.

How to Access the BAAI/bge-reranker-v2-m3 on Novita AI

Novita AI is an AI cloud platform that offers developers an easy way to deploy AI models using our simple API, while also providing the affordable and reliable GPU cloud for building and scaling.

Step 1: Log In and Access the Model Console

Log in to your account and access the model console.

Novita Model Console

Step 2: Choose Your Model and Start a Free Trail

Browse through the available options and search for the model that suits your needs.

choose your model

Step 3: Get Your API Key

To authenticate with the API, we will provide you with a new API key. Entering the “Settings“ page, you can copy the API key as indicated in the image.

get your API key

Step 4: Install the API

Install API using the package manager specific to your programming language.

install the API

After installation, import the necessary libraries into your development environment. Initialize the API with your API key to start interacting with Novita AI rerank service. This example demonstrates how to use the rerank API.

curl --request POST \
  --url "https://api.novita.ai/v3/openai/rerank" \
  --header "Authorization: Bearer <Your API Key>" \
  --header "Content-Type: application/json" \
  --data @- << 'EOF'
{
  "model": "baai/bge-reranker-v2-m3",
  "query": "Who is Novita",
  "documents": [
    "Novita AI is an all-in-one AI cloud solution that empowers businesses with open-source model APIs, serverless GPUs, and on-demand GPU instances. Drive innovation and gain a competitive edge with the power of Novita AI.",
    "Access Novita AI's Model Library with 200+ APIs for language, image, audio, and video. Simplify AI deployment with powerful, scalable solutions.",
    "Optimize your AI with Novita's GPU instances and serverless GPU cloud. Save up to 50%, auto-scale, and access high-capacity storage for global deployment."
  ],
  "top_n": 3
}
EOF
 

Upon registration, Novita AI provides a $0.5 credit to get you started!

If the free credits is used up, you can pay to continue using it.

Rerankers play a critical role in information retrieval and AI applications, allowing systems to effectively understand semantic relevance and improve search accuracy. Advanced models like BAAI/bge-reranker-v2-m3 elevate these capabilities through superior performance and extensive language support, making them indispensable tools for driving modern AI applications.

Frequently Asked Questions

What is the difference between BGE Reranker Large and BGE Reranker v2 m3?

BGE Reranker Large is based on xlm-roberta-large and supports Chinese and English, while BGE Reranker v2 m3 is based on bge-m3 and offers strong multilingual capabilities supporting multiple languages beyond just Chinese and English.

Is BGE-m3 good?

Yes, BGE-m3 is considered one of the top-performing multilingual embedding models with excellent results on various benchmarks. It excels in cross-lingual retrieval and provides strong performance across multiple languages.

What is BGE-m3?

BGE-m3 (BAAI General Embedding Multilingual-3) is a multilingual embedding model developed by BAAI that supports over 100 languages. It’s designed for various tasks including text retrieval, semantic similarity, and cross-lingual applications.

Novita AI is an AI cloud platform that offers developers an easy way to deploy AI models using our simple API, while also providing the affordable and reliable GPU cloud for building and scaling.


Discover more from Novita

Subscribe to get the latest posts sent to your email.

Leave a Comment

Scroll to Top

Discover more from Novita

Subscribe now to keep reading and get access to the full archive.

Continue reading