--- base_model: microsoft/Phi-3-mini-4k-instruct datasets: - AlignmentLab-AI/alpaca-cot-collection language: - en library_name: peft license: apache-2.0 pipeline_tag: text-generation --- # Xenith-3B Xenith-3B is a fine-tuned language model based on the microsoft/Phi-3-mini-4k-instruct model. It has been specifically trained on the AlignmentLab-AI/alpaca-cot-collection dataset, which focuses on chain-of-thought reasoning and instruction following. # Model Overview - Model Name: Xenith-3B - Base Model: microsoft/Phi-3-mini-4k-instruct - Fine-Tuned On: AlignmentLab-AI/alpaca-cot-collection - Model Size: 3 Billion parameters - Architecture: Transformer-based LLM # Training Details - Objective: Fine-tune the base model to enhance its performance on tasks requiring complex reasoning and multi-step problem-solving. - Training Duration: 10 epochs - Batch Size: 8 - Learning Rate: 3e-5 - Optimizer: AdamW - Hardware Used: 2x NVIDIA L4 GPUs # Performance Xenith-3B excels in tasks that require: - Chain-of-thought reasoning - Instruction following - Contextual understanding - Complex problem-solving - The model has shown significant improvements in these areas compared to the base model.