--- datasets: - nur-dev/kaz-for-lm language: - kk library_name: transformers pipeline_tag: text-generation license: afl-3.0 --- # GPT-J-3.48B-Kazakh
Llama Model Logo

Kazakh Language GPT-J-3.48B

General-purpose Kazakh Language Model

Architecture: GPTJForCausalLM
Tokenizer: retrained GPT2Tokenizer (Vocabulary size: 50,400, Model Max Length: 2048)
## Overview This model is a Kazakh language variant of the GPT-J-3.48B architecture, designed for general-purpose language modeling tasks. It has been trained on a diverse set of Kazakh language texts and is intended to support various natural language processing applications in the Kazakh language. ## Usage Example The model can be used with the Hugging Face Transformers library: ```python from transformers import AutoModelForCausalLM, AutoTokenizer import torch model = AutoModelForCausalLM.from_pretrained("nur-dev/gpt-j-3.4B-kaz") tokenizer = AutoTokenizer.from_pretrained("nur-dev/gpt-j-3.4B-kaz") model.eval() ``` ### Training Details The model is being trained using the DeepSpeed library with Zero Optimization Stage 2. During the training process, zero optimization is applied at stage 2, with the optimizer offloaded to the CPU and pin memory enabled. The training also includes allgather partitions with a bucket size of 200M, overlap communication, reduce scatter, an automatic reduce bucket size, and the use of contiguous gradients. Hardware: 4 NVIDIA A100 GPUs (40GB each) Training Steps: Approximately 180,000 (ongoing) Epochs: 1(ongoing) Batch Size: 2 per device (for both training and evaluation) Gradient Accumulation Steps: 4 Learning Rate: 5e-5 Weight Decay: 0.05 Learning Rate Scheduler: Cosine with Restarts Warmup Steps: 15,000 Checkpointing Steps: Every 10,000 steps ## Model Authors **Name:** Kadyrbek Nurgali - **Email:** nurgaliqadyrbek@gmail.com - **LinkedIn:** [Kadyrbek Nurgali](https://www.linkedin.com/in/nurgali-kadyrbek-504260231/)