@@ -187,9 +187,180 @@ def score_for_task(self, task: TaskType) -> float:
187187 api_provider = "serper" ,
188188 best_for = ["real-time data" , "current events" , "web search" ],
189189 ),
190+
191+ # === DIVER DENSE RETRIEVERS ===
192+ # method="diver-dense", use model_path as the model_id argument
193+ # Valid model_ids: bge, sbert, contriever_st, nomic, diver, inst-l, inst-xl, sf, e5, rader, m2, contriever, grit
194+ "diver-bge" : ModelMetadata (
195+ name = "Diver (BGE Large)" ,
196+ method = "diver-dense" ,
197+ description = "BAAI/bge-large-en-v1.5 via the Diver dense retrieval framework." ,
198+ speed = Speed .MEDIUM ,
199+ accuracy = Accuracy .STATE_OF_THE_ART ,
200+ gpu_required = True ,
201+ memory_mb = 3000 ,
202+ best_for = ["semantic search" , "high accuracy" , "BEIR benchmarks" ],
203+ model_path = "bge" ,
204+ ),
205+ "diver-sbert" : ModelMetadata (
206+ name = "Diver (SBERT all-mpnet-base-v2)" ,
207+ method = "diver-dense" ,
208+ description = "sentence-transformers/all-mpnet-base-v2 via the Diver framework." ,
209+ speed = Speed .FAST ,
210+ accuracy = Accuracy .VERY_GOOD ,
211+ gpu_required = True ,
212+ memory_mb = 1500 ,
213+ best_for = ["sentence similarity" , "semantic search" ],
214+ model_path = "sbert" ,
215+ ),
216+ "diver-nomic" : ModelMetadata (
217+ name = "Diver (Nomic Embed)" ,
218+ method = "diver-dense" ,
219+ description = "nomic-ai/nomic-embed-text-v1 via the Diver framework." ,
220+ speed = Speed .FAST ,
221+ accuracy = Accuracy .VERY_GOOD ,
222+ gpu_required = True ,
223+ memory_mb = 1500 ,
224+ best_for = ["long context" , "semantic search" , "document retrieval" ],
225+ model_path = "nomic" ,
226+ ),
227+ "diver-e5" : ModelMetadata (
228+ name = "Diver (E5-Mistral-7B)" ,
229+ method = "diver-dense" ,
230+ description = "intfloat/e5-mistral-7b-instruct — instruction-tuned LLM encoder in the Diver framework." ,
231+ speed = Speed .SLOW ,
232+ accuracy = Accuracy .STATE_OF_THE_ART ,
233+ gpu_required = True ,
234+ memory_mb = 16000 ,
235+ best_for = ["zero-shot retrieval" , "instruction following" , "complex queries" ],
236+ model_path = "e5" ,
237+ ),
238+ "diver-sf" : ModelMetadata (
239+ name = "Diver (SFR-Embedding-Mistral)" ,
240+ method = "diver-dense" ,
241+ description = "Salesforce/SFR-Embedding-Mistral — Salesforce Mistral-based bi-encoder in the Diver framework." ,
242+ speed = Speed .SLOW ,
243+ accuracy = Accuracy .STATE_OF_THE_ART ,
244+ gpu_required = True ,
245+ memory_mb = 16000 ,
246+ best_for = ["high accuracy retrieval" , "complex queries" , "BEIR benchmarks" ],
247+ model_path = "sf" ,
248+ ),
249+ "diver-rader" : ModelMetadata (
250+ name = "Diver (RaDeR)" ,
251+ method = "diver-dense" ,
252+ description = "Raderspace/RaDeR_Qwen_25_7B — reasoning-aware dense retriever in the Diver framework." ,
253+ speed = Speed .SLOW ,
254+ accuracy = Accuracy .EXCELLENT ,
255+ gpu_required = True ,
256+ memory_mb = 16000 ,
257+ best_for = ["reasoning-intensive queries" , "multi-hop QA" , "math-related retrieval" ],
258+ model_path = "rader" ,
259+ ),
260+ "diver-grit" : ModelMetadata (
261+ name = "Diver (GritLM-7B)" ,
262+ method = "diver-dense" ,
263+ description = "GritLM/GritLM-7B — generative representation model in the Diver framework." ,
264+ speed = Speed .VERY_SLOW ,
265+ accuracy = Accuracy .STATE_OF_THE_ART ,
266+ gpu_required = True ,
267+ memory_mb = 16000 ,
268+ best_for = ["generative retrieval" , "LLM-quality embeddings" , "long context" ],
269+ model_path = "grit" ,
270+ ),
271+ "diver-model" : ModelMetadata (
272+ name = "Diver Retriever-4B" ,
273+ method = "diver-dense" ,
274+ description = "AQ-MedAI/Diver-Retriever-4B — the flagship Diver diverse-evidence retrieval model." ,
275+ speed = Speed .SLOW ,
276+ accuracy = Accuracy .STATE_OF_THE_ART ,
277+ gpu_required = True ,
278+ memory_mb = 8000 ,
279+ best_for = ["diverse evidence retrieval" , "BEIR benchmarks" , "medical QA" ],
280+ model_path = "diver" ,
281+ ),
282+ "diver-inst-l" : ModelMetadata (
283+ name = "Diver (Instructor-Large)" ,
284+ method = "diver-dense" ,
285+ description = "hkunlp/instructor-large — instruction-following encoder in the Diver framework." ,
286+ speed = Speed .MEDIUM ,
287+ accuracy = Accuracy .VERY_GOOD ,
288+ gpu_required = True ,
289+ memory_mb = 3000 ,
290+ best_for = ["instruction following" , "domain-specific retrieval" ],
291+ model_path = "inst-l" ,
292+ ),
293+ "diver-m2" : ModelMetadata (
294+ name = "Diver (M2-BERT-32K)" ,
295+ method = "diver-dense" ,
296+ description = "togethercomputer/m2-bert-80M-32k-retrieval — long-context retrieval in the Diver framework." ,
297+ speed = Speed .MEDIUM ,
298+ accuracy = Accuracy .VERY_GOOD ,
299+ gpu_required = True ,
300+ memory_mb = 2000 ,
301+ best_for = ["long-context retrieval" , "32k sequence length" ],
302+ model_path = "m2" ,
303+ ),
304+
305+ # === REASONING-AUGMENTED RETRIEVERS ===
306+ "reasonir" : ModelMetadata (
307+ name = "ReasonIR-8B" ,
308+ method = "reasonir" ,
309+ description = "reasonir/ReasonIR-8B — SOTA reasoning-intensive retriever on the BRIGHT benchmark. No model_id needed." ,
310+ speed = Speed .VERY_SLOW ,
311+ accuracy = Accuracy .STATE_OF_THE_ART ,
312+ gpu_required = True ,
313+ memory_mb = 16000 ,
314+ best_for = ["reasoning-intensive queries" , "BRIGHT benchmark" , "complex multi-hop QA" , "science queries" ],
315+ ),
316+ "reason-embed-qwen3-8b" : ModelMetadata (
317+ name = "ReasonEmbed Qwen3-8B" ,
318+ method = "reason-embed" ,
319+ description = "hanhainebula/reason-embed-qwen3-8b-0928 — Qwen3-8B for reasoning retrieval. Use model_id='qwen3-8b'." ,
320+ speed = Speed .VERY_SLOW ,
321+ accuracy = Accuracy .STATE_OF_THE_ART ,
322+ gpu_required = True ,
323+ memory_mb = 16000 ,
324+ best_for = ["reasoning-intensive retrieval" , "complex queries" ],
325+ model_path = "qwen3-8b" ,
326+ ),
327+ "reason-embed-qwen3-4b" : ModelMetadata (
328+ name = "ReasonEmbed Qwen3-4B" ,
329+ method = "reason-embed" ,
330+ description = "hanhainebula/reason-embed-qwen3-4b-0928 — balanced Qwen3-4B for reasoning retrieval. Use model_id='qwen3-4b'." ,
331+ speed = Speed .SLOW ,
332+ accuracy = Accuracy .EXCELLENT ,
333+ gpu_required = True ,
334+ memory_mb = 8000 ,
335+ best_for = ["reasoning retrieval" , "balanced accuracy/speed" ],
336+ model_path = "qwen3-4b" ,
337+ ),
338+ "reason-embed-llama-8b" : ModelMetadata (
339+ name = "ReasonEmbed LLaMA-3.1-8B" ,
340+ method = "reason-embed" ,
341+ description = "hanhainebula/reason-embed-llama-3.1-8b-0928 — LLaMA-3.1-8B for reasoning retrieval. Use model_id='llama-8b'." ,
342+ speed = Speed .SLOW ,
343+ accuracy = Accuracy .EXCELLENT ,
344+ gpu_required = True ,
345+ memory_mb = 16000 ,
346+ best_for = ["reasoning retrieval" , "open-source LLaMA backbone" ],
347+ model_path = "llama-8b" ,
348+ ),
349+ "bge-reasoner-embed" : ModelMetadata (
350+ name = "BGE Reasoner Embed (Qwen3-8B)" ,
351+ method = "bge-reasoner-embed" ,
352+ description = "BAAI/bge-reasoner-embed-qwen3-8b-0923 — BGE reasoning-augmented retriever. No model_id needed." ,
353+ speed = Speed .SLOW ,
354+ accuracy = Accuracy .EXCELLENT ,
355+ gpu_required = True ,
356+ memory_mb = 16000 ,
357+ best_for = ["reasoning-augmented retrieval" , "BEIR benchmarks" , "complex queries" ],
358+ ),
190359}
191360
192361
362+
363+
193364# =============================================================================
194365# RERANKER REGISTRY
195366# =============================================================================
0 commit comments