Skip to content

A Visual Document Benchmark for Scientific Retrieval and Question Answering

License

Notifications You must be signed in to change notification settings

weaviate/IRPAPERS

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

6 Commits
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

IRPAPERS

A Visual Document Benchmark for Scientific Retrieval and Question Answering.

Retrieval Leaderboard ๐Ÿ”Ž

Rank Retriever Type Recall@1 Recall@5 Recall@20
1 Multimodal Hybrid (Cohere Embed v4.0 + Voyage 3 Large + BM25) Hybrid 58% 91% 98%
2 Cohere Embed v4.0 Image 58% 87% 97%
3 Voyage 3 Large Text 52% 86% 95%
4 ColQwen2 Image 49% 81% 94%
5 Multimodal Hybrid (ColModernVBERT + Arctic 2.0 + BM25) Hybrid 49% 81% 95%
6 Hybrid Text Search (Arctic 2.0 + BM25) Text 46% 78% 91%
7 ColPali Image 45% 79% 93%
8 BM25 Text 45% 71% 90%
9 Arctic 2.0 Text 44% 76% 88%
10 ColModernVBERT Image 43% 78% 93%
11 ColModernVBERT + MUVERA (ef=1024) Image 41% 75% 88%
12 ColModernVBERT + MUVERA (ef=512) Image 37% 68% 78%
13 ColModernVBERT + MUVERA (ef=256) Image 35% 61% 66%

Question Answering Leaderboard ๐Ÿ’ฌ

Rank System Type Alignment Score Avg Input Tokens Avg Output Tokens
1 TextRAG (k=5) Text 0.82 6,022 243
2 Oracle Text Retrieval (k=1) Text 0.74 1,294 155
3 ImageRAG (k=5) Image 0.71 5,200 178
4 Oracle Image Retrieval (k=1) Image 0.68 1,208 125
5 TextRAG (k=1) Text 0.62 1,366 160
6 ImageRAG (k=1) Image 0.40 1,228 124
7 Hard Negative Text Context (k=1) Text 0.39 1,304 162
8 No Retrieval Baseline โ€” 0.16 173 135
9 Hard Negative Image Context (k=1) Image 0.12 1,233 134

IRPAPERS-MCP (Coming Soon) โš ๏ธ

News ๐Ÿ“ฏ

About

A Visual Document Benchmark for Scientific Retrieval and Question Answering

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published