Generative Pre-trained Transformer in PyTorch from scratch
python src/train.py
Options:
--batch_size 64
--num-epochs 100
--lr 0.0001
--from-checkpoint checkpoint_path.pth
Model is checkpointed after each epoch and stored in checkpoints/
directory
from train import train
train()
python src/run.py --from-checkpoint checkpoint_path.pth
from run import run
run(model_path="checkpoint_path.pth", prompt="Rick:\nMorty, where are you?)
GPL v3
Jędrzej Maczan, 2024