--- language: - en tags: - Twitter - Climate Change license: mit --- # Model Card Climate-TwitterBERT-step-1 ## Overview: Using Covid-Twitter-BERT-v2 (https://huggingface.co/digitalepidemiologylab/covid-twitter-bert-v2) as the starting model, we continued domain-adaptive pre-training on a corpus of firm tweets between 2007 and 2020. The model was then fine-tuned on the downstream task to classify whether a given tweet is related to climate change topics. The model provides a label and probability score, indicating whether a given tweet is related to climate change topics (label = 1) or not (label = 0). ## Performance metrics: Based on the test set, the model achieves the following results: • Loss: 0.0632 • F1-weighted: 0.9778 • F1: 0.9148 • Accuracy: 0.9775 • Precision: 0. 8841 • Recall: 0. 9477 ## Example usage: ```python from transformers import pipeline, AutoTokenizer, AutoModelForSequenceClassification task_name = 'binary' model_name = Climate-TwitterBERT/ Climate-TwitterBERT-step1' tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForSequenceClassification.from_pretrained(model_name) pipe = pipeline(task=‘binary‘, model=model, tokenizer=tokenizer) tweet = "We are committed to significantly cutting our carbon emissions by 30% before 2030." result = pipe(tweet) # The 'result' variable will contain the classification output: 0 = non-climate tweet, 1= climate tweet ``` ## Citation: ```bibtex @article{fzz2023climatetwitter, title={Responding to Climate Change crisis - firms' tradeoffs}, author={Fritsch, Felix and Zhang, Qi and Zheng, Xiang}, journal={Working paper}, year={2023}, institution={University of Mannheim, the Chinese University of Hong Kong, and NHH Norwegian School of Economics}, url={https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4527255} } ``` Fritsch, F., Zhang, Q., & Zheng, X. (2023). Responding to Climate Change crisis - firms' tradeoffs [Working paper]. University of Mannheim, the Chinese University of Hong Kong, and NHH Norwegian School of Economics. ## Framework versions • Transformers 4.28.1 • Pytorch 2.0.1+cu118 • Datasets 2.14.1 • Tokenizers 0.13.3