Skip to content

rygx/simple-gpt

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

A PyTorch implementation of generative pre-trained transformers (GPT)

This is a personal exercise to build, train and use GPT, inspired by minGPT.

Prerequisite

Installation

git clone https://github.com/rygx/simple-gpt.git && cd simple-gpt
python -m venv .venv
pip install pip-tools
./update_deps.sh

Usage

Data

Any text-fromat data file should be usable, given the size can fit with training and inference environment.

Train

Use train/train_gpt.py. After training, model state dictionary and hyper parameters will be stored in models directory.

Generate

Use train/generate.py. Under models/ directory there is already a coarsely pre-trained model (using minGPT's tinyshakespeare sample) with ID 9cdb42ed-0b16-4a3a-88e2-fffa61fa4f50. Generate texts using this model with the following command:

python train/generate.py --dir "models" -u "9cdb42ed-0b16-4a3a-88e2-fffa61fa4f50" --prompt "QUEEN: "

Generation length (in number of tokens) and temperature can also be tuned using --length/-l and --temp/-t options.

TODOs/plans

  • Easier setup (e.g., using setuptool)
  • Unit tests :P
  • Implement own tokenizer

About

A PyTorch implementation of GPT

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published