Skip to content

aynursusuz/build-nanogpt

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

GPT-2 Training Script

GPT-2 Training

This repository contains a PyTorch implementation for training and fine-tuning GPT-2 models. The main script train_gpt2.py provides functionality for training GPT-2 models with various configurations and supports both single-GPU and distributed training setups.

Simple Launch (Single GPU)

python train_gpt2.py

Distributed Training (Multiple GPUs)

torchrun --standalone --nproc_per_node=8 train_gpt2.py

Training Parameters

  • Learning Rate: 6e-4
  • Weight Decay: 0.1
  • Block Size: 1024
  • Vocabulary Size: 50,257 (GPT-2 tokenizer)

License

This project is part of the build-nanogpt repository.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published