--- library_name: transformers tags: - climate-change - flan-t5 - qlora - instruction-tuning --- # Model Card for FLAN-T5 Climate Action QLoRA This is a QLoRA-finetuned version of FLAN-T5 specifically trained for climate action content analysis and generation. The model is optimized for processing and analyzing text related to climate change, sustainability, and environmental policies. ## Model Details ### Model Description - **Developed by:** Kshitiz Khanal - **Shared by:** kshitizkhanal7 - **Model type:** Instruction-tuned Language Model with QLoRA fine-tuning - **Language(s):** English - **License:** Apache 2.0 - **Finetuned from model:** google/flan-t5-base ### Model Sources - **Repository:** https://huggingface.co/kshitizkhanal7/flan-t5-climate-qlora - **Training Data:** FineWeb dataset (climate action filtered) ## Uses ### Direct Use The model is designed for: - Analyzing climate policies and initiatives - Summarizing climate action documents - Answering questions about climate change and environmental policies - Evaluating sustainability measures - Processing climate-related research and reports ### Downstream Use The model can be integrated into: - Climate policy analysis tools - Environmental reporting systems - Sustainability assessment frameworks - Climate research applications - Educational tools about climate change ### Out-of-Scope Use The model should not be used for: - Critical policy decisions without human oversight - Generation of climate misinformation - Technical climate science research without expert validation - Commercial deployment without proper testing - Medical or legal advice ## Bias, Risks, and Limitations - Limited to climate-related content analysis - May not perform well on general domain tasks - Potential biases from web-based training data - Should not be the sole source for critical decisions - Performance varies on technical climate science topics ### Recommendations - Always verify model outputs with authoritative sources - Use human expert oversight for critical applications - Consider the model as a supplementary tool, not a replacement for expert knowledge - Regular evaluation of outputs for potential biases - Use in conjunction with other data sources for comprehensive analysis ## Training Details ### Training Data - Source: FineWeb dataset filtered for climate content - Selection criteria: Climate-related keywords and quality metrics - Processing: Instruction-style formatting with climate focus ### Training Procedure #### Preprocessing - Text cleaning and normalization - Instruction templates for climate context - Maximum input length: 512 tokens - Maximum output length: 128 tokens #### Training Hyperparameters - Training regime: QLoRA 4-bit fine-tuning - Epochs: 3 - Learning rate: 2e-4 - Batch size: 4 - Gradient accumulation steps: 4 - LoRA rank: 16 - LoRA alpha: 32 - Target modules: Query and Value matrices - LoRA dropout: 0.05 ## Environmental Impact - **Hardware Type:** Single GPU - **Hours used:** ~4 hours - **Cloud Provider:** Local - **Carbon Emitted:** Minimal due to QLoRA efficiency ## Technical Specifications ### Model Architecture and Objective - Base architecture: FLAN-T5 - Objective: Climate-specific text analysis - QLoRA adaptation for efficient fine-tuning - 4-bit quantization for reduced memory usage ### Compute Infrastructure - Python 3.8+ - PyTorch - Transformers library - bitsandbytes for quantization - PEFT for LoRA implementation ### Hardware Minimum requirements: - 16GB GPU memory for inference - 24GB GPU memory recommended for training - CPU inference possible but slower ## Citation If you use this model, please cite: ```bibtex @misc{khanal2024climate, title={FLAN-T5 Climate Action QLoRA}, author={Khanal, Kshitiz}, year={2024}, publisher={HuggingFace}, howpublished={\url{https://huggingface.co/kshitizkhanal7/flan-t5-climate-qlora}} }