From e1f456861c4ba62db1cf7fc31093d04e66b5040f Mon Sep 17 00:00:00 2001 From: Xiao Date: Tue, 2 Jul 2024 15:13:57 +0000 Subject: [PATCH] Update MIRACL evaluation results (#68) - Update MIRACL evaluation results (0c6f0d0ea8f284b9070c3ffaa50677440943f984) Co-authored-by: Jianlv Chen --- README.md | 7 +++++++ 1 file changed, 7 insertions(+) diff --git a/README.md b/README.md index 9382959..e5a3201 100644 --- a/README.md +++ b/README.md @@ -33,6 +33,13 @@ Utilizing the re-ranking model (e.g., [bge-reranker](https://github.com/FlagOpen ## News: +- 2024/7/1: **We update the MIRACL evaluation results of BGE-M3**. To reproduce the new results, you can refer to: [bge-m3_miracl_2cr](https://huggingface.co/datasets/hanhainebula/bge-m3_miracl_2cr). We have also updated our [paper](https://arxiv.org/pdf/2402.03216) on arXiv. +
+ Details + + The previous test results were lower because we mistakenly removed the passages that have the same id as the query from the search results. After correcting this mistake, the overall performance of BGE-M3 on MIRACL is higher than the previous results, but the experimental conclusion remains unchanged. The other results are not affected by this mistake. To reproduce the previous lower results, you need to add the `--remove-query` parameter when using `pyserini.search.faiss` or `pyserini.search.lucene` to search the passages. + +
- 2024/3/20: **Thanks Milvus team!** Now you can use hybrid retrieval of bge-m3 in Milvus: [pymilvus/examples /hello_hybrid_sparse_dense.py](https://github.com/milvus-io/pymilvus/blob/master/examples/hello_hybrid_sparse_dense.py). - 2024/3/8: **Thanks for the [experimental results](https://towardsdatascience.com/openai-vs-open-source-multilingual-embedding-models-e5ccb7c90f05) from @[Yannael](https://huggingface.co/Yannael). In this benchmark, BGE-M3 achieves top performance in both English and other languages, surpassing models such as OpenAI.**