Skip to content

Run benchmarking with the supported models on BEIR MSMARCO #177

@HAKSOAT

Description

@HAKSOAT

We need to run benchmarking on the BEIR MSMARCO dataset, to have a better understanding of how the models are performing for retrieval tasks.

We can use the test split available on Hugging Face hub:

QRels
Corpus

Proposed metrics:

  • NDCG@10
  • Precision@10
  • Recall@100

Considering non-judged documents as non-relevant.

Metadata

Metadata

Assignees

No one assigned

    Labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions