Fusion-aware ordering
Intelligently blends dense and sparse candidates instead of relying on fixed heuristics.
Fine-tuned on your data
Learns from historical dense and sparse outcomes so every blend reflects your domain.
Evidence-based scoring
Ranks by learned relevance, not raw similarity, giving the reranker a stronger shortlist.
Calibrated confidence
Produces scores with clear thresholds so downstream systems know when to trust the set.
How it works
We replace Reciprocal Rank Fusion (RRF)—a fixed, query-agnostic heuristic—with a learned model that produces a calibrated fused score before reranking. Concretely, for a query q
and candidate d
, we learn:
s_fuse(q, d) = g([s_dense(q, d), s_sparse(q, d), Δrank, lex_overlap, meta, φ(q)])
Inputs to g
include:
- Dense score `s_dense(q, d)`
- Sparse score `s_sparse(q, d)`
- Rank deltas `Δrank` and reciprocal features
- Lexical overlap and metadata features
- Query representation φ(q) capturing intent and difficulty
Training optimizes pairwise or listwise losses (e.g., NDCG, MRR) so the fusion score mirrors real customer satisfaction outcomes while keeping downstream latency and token budgets in check.