Datasets:
Tasks:
Question Answering
Modalities:
Text
Formats:
parquet
Languages:
English
Size:
10M - 100M
DOI:
import glob | |
import json | |
import pickle | |
import sys | |
from typing import Dict | |
import numpy as np | |
from beir.retrieval.evaluation import EvaluateRetrieval | |
def load_qrels(filename: str) -> Dict: | |
with open(filename, "r") as f: | |
qrels = json.load(f) | |
return qrels | |
def merge_retrieved_shards( | |
suffix: str, output_file: str, top_n: int, qrels: dict, metric: str | |
) -> None: | |
shard_files = glob.glob(f"*{suffix}") | |
print(f"There are {len(shard_files)} shards found") | |
merged_results = {} | |
print("Loading All shards") | |
for shard_file in shard_files: | |
print(f"Loading shard {shard_file} ") | |
with open(shard_file, "rb") as f: | |
shard_results = pickle.load(f) | |
for query_id, doc_scores in shard_results.items(): | |
if query_id not in merged_results: | |
merged_results[query_id] = [] | |
merged_results[query_id].extend(doc_scores.items()) | |
print("Shards all loaded, merging results and sorting by score") | |
run = {} | |
per_query = [] | |
for query_id, doc_scores in merged_results.items(): | |
if query_id in qrels: | |
doc_score_dict = {} | |
for passage_id, score in doc_scores: | |
doc_id = passage_id.split("#")[ | |
0 | |
] # everything after # is the passage idenfitier withing a doc | |
if doc_id not in doc_score_dict: | |
doc_score_dict[doc_id] = ( | |
-1 | |
) # scores are in range -1 to 1 on similairty so starting at -1 is floor | |
if score > doc_score_dict[doc_id]: | |
doc_score_dict[doc_id] = score | |
top_docs = sorted(doc_score_dict.items(), key=lambda x: x[1], reverse=True)[ | |
:top_n | |
] | |
run[query_id] = { | |
doc_id: round(score * 100, 2) for doc_id, score in top_docs | |
} | |
scores = EvaluateRetrieval.evaluate( | |
qrels, {query_id: run[query_id]}, k_values=[1, 3, 5, 10, 100, 1000] | |
) | |
scores = {k: v for d in scores for k, v in d.items()} | |
per_query.append(scores[metric]) | |
print("Done merging and sorting results, Evaluating and saving run") | |
print(f"There are {len(run)} queries being evaled agaisnt qrels") | |
print(f"There were {len(shard_files)} shards found") | |
print( | |
f"Per Query Score average: {np.array(per_query).mean()} for {metric}. Individual scores{per_query}" | |
) | |
print("Overall Score Numbers:") | |
print(EvaluateRetrieval.evaluate(qrels, run, k_values=[1, 3, 5, 10, 100, 1000])) | |
with open(output_file, "wb") as w: | |
pickle.dump(run, w) | |
if __name__ == "__main__": | |
suffix = sys.argv[1] | |
output_file = sys.argv[2] | |
top_n = int(sys.argv[3]) | |
qrel_filename = sys.argv[4] | |
metric = sys.argv[5] | |
merge_retrieved_shards( | |
suffix, output_file, top_n, load_qrels(qrel_filename), metric | |
) | |