import re import string from lifelines.utils import concordance_index def keep_integers_commas_spaces(input_string): cleaned_string = re.sub(r'[^0-9\s,]', '', str(input_string)) return cleaned_string def normalize_answer(s): """Lower text and remove punctuation, articles and extra whitespace.""" def remove_punc(text): exclude = set(string.punctuation) return "".join(ch for ch in text if ch not in exclude) normalized_list = keep_integers_commas_spaces(s).replace(",", " ").strip(string.punctuation).split() try: normalized_list = [int(remove_punc(x).strip()) for x in normalized_list] except ValueError: return [] return normalized_list def concordant_index_score(prediction, ground_truth): normalized_prediction = normalize_answer(prediction) normalized_ground_truth = normalize_answer(ground_truth) if sorted(normalized_ground_truth) != sorted(normalized_prediction): return 0.0 pred_order = summ_id_per_location_to_pos_of_id(normalized_prediction) gold_order = summ_id_per_location_to_pos_of_id(normalized_ground_truth) return concordance_index(gold_order, pred_order) def summ_id_per_location_to_pos_of_id(id_per_location): order = [-1] * len(id_per_location) for i, id_ in enumerate(id_per_location, 1): order[id_ - 1] = i return order def metric_max_over_ground_truths(metric_fn, prediction, ground_truths): scores_for_ground_truths = [] for ground_truth in ground_truths: score = metric_fn(prediction, ground_truth) scores_for_ground_truths.append(score) return max(scores_for_ground_truths) def compute_concordance_index(predictions, references): concordant_index = 0 for prediction, ground_truths in zip(predictions, references): concordant_index += metric_max_over_ground_truths(concordant_index_score, prediction, ground_truths) return 100.0 * concordant_index / len(predictions)