Skip to content

Latest commit

 

History

History
26 lines (19 loc) · 1.21 KB

README.md

File metadata and controls

26 lines (19 loc) · 1.21 KB

nanoGPT

The simplest, fastest repository for training/finetuning medium-sized GPTs. This repository contains code for a Transformer-based language model, specifically the Generative Pre-trained Transformer (GPT) model. GPT is a state-of-the-art language model architecture that has achieved impressive results in various natural language processing tasks, including text generation and language understanding. It is directly inspired by Andrej Karpathy's gpt video.

This repository contains the code that summed up from the Karpathy's makemore series, ultimately leading to this. I have implemented the code from makemore series too. You can check it out here: Makemore series

Requirements

simply, install the dependencies using

pip install -r requirements.txt

Training

To train the GPT model, follow these steps:

  1. Install the required dependencies: pip install -r requirements.txt
  2. Prepare your training data in a text file (data/train.txt).
  3. Run the training script:
cd nanoGPT
python gpt.py

You will also get sample outputs once the training script is completed.