--- dataset_info: features: - name: question dtype: string - name: choices sequence: string - name: label dtype: int64 splits: - name: train num_bytes: 619513 num_examples: 384 - name: test num_bytes: 2301030 num_examples: 1416 download_size: 1491635 dataset_size: 2920543 configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* --- # QUANDHO: QUestion ANswering Data for italian HistOry Original Paper: https://aclanthology.org/L16-1069.pdf QUANDHO (QUestion ANswering Data for italian HistOry) is an Italian question answering dataset created to cover the history of Italy in the first half of the XX century. Starting from QUANDHO we defined a Multi-choice QA dataset, with a correct answer and four different distractors. ## Data and Distractors Generation We relied on the original data, to create this dataset. For each question-answer correct pair, we defined a dataset sample. For each sample, we gather four different distractors from incorrect question-answer pairs, where the question is the one of the chosen sample. ## Example Here you can see the structure of the single sample in the present dataset. ```json { "text": string, # text of the question "choices": list, # list of possible answers, with the correct one plus 3 distractors "label": int, # index of the correct anser in the choices } ``` ## Statistics Training: 384 Test: 1416 ## Proposed Prompts Here we will describe the prompt given to the model over which we will compute the perplexity score, as model's answer we will chose the prompt with lower perplexity. Moreover, for each subtask, we define a description that is prepended to the prompts, needed by the model to understand the task. Description of the task: ```txt Ti saranno poste domande di storia italiana.\nIdentifica quali paragrafi contengono la risposta alle domande date.\n\n ``` Prompt: ```txt Data la domanda: \"{{question}}\"\nQuale tra i seguenti paragrafi risponde alla domanda?\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nRisposta: ``` ## Results | QUANDHO | ACCURACY (2-shots) | | :-----: | :--: | | Gemma-2B | 43.99 | | QWEN2-1.5B | 56.43 | | Mistral-7B | 72.66 | | ZEFIRO | 70.12| | Llama-3-8B | 70.26 | | Llama-3-8B-IT | 81.07 | | ANITA | 74.29 | ## Acknowledgment The original data can be downloaded from the following [link](https://dh.fbk.eu/2016/03/quandho-question-answering-data-for-italian-history/) We want to thank the dataset's creators, that release such interesting resource publicly. ## License The original dataset is licensed under [Creative Commons Attribution 4.0 International License](https://creativecommons.org/licenses/by/4.0/)