--- license: cc-by-nc-4.0 datasets: - coastalcph/eu_debates language: - en library_name: peft base_model: meta-llama/Llama-2-13b-chat-hf pipeline_tag: text-generation tags: - politics widget: - text: What is your opinion on "Immigration, the role of Frontex and cooperation among Member States"? extra_gated_heading: You need to share contact information to access this model. You have to briefly describe your affiliation and the research scope of your work. If not, the request will automatically be rejected. The whole purpose is to filter requests and avoid easily identifiable misuse by AI bots or individuals with malicious intents. extra_gated_fields: First Name: text Last Name: text Country: country Affiliation: text Scope of Research: text extra_gated_button_content: Submit extra_gated_prompt: >- ### EU Llama-Chat Adapters - Terms of Use Please read carefully the Terms of Use, before requesting access. To grant access to this model, you have to accept the following Terms of Use: (a) This model is exclusively available for research purposes. (b) It is prohibited to use this model for commercial and non-commercial uses outside the scope of research. (c) It is prohibited to deploy this model publicly on the web. (d) The model is released without any warranties. (e) The model may produce content that can be considered discriminatory or harmful. You are only allowed to share this content with caution when discussing biases in your work. --- # Model Card (Llama-2 13B - LoRA on speeches from members of EPP) In this work [(Chalkidis and Brandl, 2024)](https://arxiv.org/abs/2403.13592), we adapt *Llama Chat* to speeches of the members of a euro party from the [EU Debates](https://huggingface.co/datasets/coastalcph/eu_debates) dataset. To do so, we fine-tune the [13B *Llama Chat* model](https://huggingface.co/meta-llama/Llama-2-13b-chat-hf) on the speeches using adapters, specifically Low-Rank Adaptation (LoRA) ([Hu et al., 2022](https://openreview.net/forum?id=nZeVKeeFYf9)). Since we are interested in fine-tuning conversational (chat-based) models, we create instructions as pseudo-QA pairs, similar to [Cheng et al. (2023)](https://arxiv.org/abs/2309.09530) using a pseudo-QA template: ``` [INST] What is your opinion on T ? [/INST] U ``` where the instruction (question) is based on the title (topic) of the debate (*T*), e.g., *"Immigration, the role of Frontex and cooperation among Member States"*, and *U* is a clean version of a speech (utterance) from an MEP affiliated with the party of interest. We use a learning rate of 2e-4, and train for 10 epochs across all data points (speeches) from the party of interest. We set LoRa alpha (α) at 16, and the rank (r) at 8. The model was developed with the code base in the following GitHub repository: https://github.com/coastalcph/eu-politics-llms/ This is the LoRA adapter for the model adapted to the speeches from MEPs affiliated with: the European People's Party (EPP). We have also released the following LoRA adapters: | Euro-party | Model HF Alias | | ---------- | -------------- | | EPP | [`coastalcph/Llama-2-13b-chat-hf-LoRA-eu-debates-epp`](https://huggingface.co/coastalcph/Llama-2-13b-chat-hf-LoRA-eu-debates-epp) | S&D | [`coastalcph/Llama-2-13b-chat-hf-LoRA-eu-debates-sd`](https://huggingface.co/coastalcph/Llama-2-13b-chat-hf-LoRA-eu-debates-sd) | ID | [`coastalcph/Llama-2-13b-chat-hf-LoRA-eu-debates-id`](https://huggingface.co/coastalcph/Llama-2-13b-chat-hf-LoRA-eu-debates-id) | GUE/NGL | [`coastalcph/Llama-2-13b-chat-hf-LoRA-eu-debates-gue-ngl`](https://huggingface.co/coastalcph/Llama-2-13b-chat-hf-LoRA-eu-debates-gue-ngl) | Greens/EFA | [`coastalcph/Llama-2-13b-chat-hf-LoRA-eu-debates-greens-efa`](https://huggingface.co/coastalcph/Llama-2-13b-chat-hf-LoRA-eu-debates-greens-efa) ![Radar Plots](https://i.ibb.co/7yVy32k/radar-plots-big.png) # Caution / Unintended Use / Biases The adapted models can be seen as data-driven mirrors of the parties’ ideologies, but are by no means ’perfectly’ aligned, and thus may misrepresent them. The models have been developed solely for research purposes and should not be used to generate content and share publicly. We urge the community and the public to refer to credible sources, e.g., parties’ programs, interviews, original speeches, etc., when it comes to getting political information. In some cases, models may generate text that can be considered hateful, toxic, harmful, or inappropriate. Please use them with caution. # How to use ```python import torch from transformers import AutoTokenizer, AutoModelForCausalLM from peft import PeftModel # Load the base model and tokenizer tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-2-13b-chat-hf") base_model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-2-13b-chat-hf", device_map="auto", torch_dtype=torch.float16, attn_implementation="flash_attention_2") # Load the LoRA adapter model = PeftModel.from_pretrained(base_model, "coastalcph/Llama-2-13b-chat-hf-LoRA-eu-debates-epp", device_map="auto") # Build pipeline pipeline = transformers.pipeline( "text-generation", model=model, tokenizer=tokenizer ) ``` # Citation Information *[Llama meets EU: Investigating the European political spectrum through the lens of LLMs. Ilias Chalkidis and Stephanie Brandl. In the Proceedings of the Annual Conference of the North American Chapter of the Association for Computational Linguistics (NAACL), Mexico City, Mexico, June 16–21, 2024.](https://arxiv.org/abs/2403.13592)* ``` @inproceedings{chalkidis-and-brandl-eu-llama-2024, title = "Llama meets EU: Investigating the European political spectrum through the lens of LLMs", author = "Chalkidis, Ilias and Brandl, Stephanie", booktitle = "Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics", month = jun, year = "2024", address = "Mexico City, Mexico", publisher = "Association for Computational Linguistics", } ```