Skip to content

Hands-on implementation of transformer and related models

License

Notifications You must be signed in to change notification settings

hash-ir/transformer-lab

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

18 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

transformer-lab 🤖

A WIP project featuring implementation of GPT and related transformer models from scratch. Inspired by Andrej Karpathy's famous "Let's Build GPT" tutorial. This repository is actively evolving.

Features ✨

  • Decoder implementation of the Transformer
    • Multi-head self-attention (parallel processing)
    • The Transformer block: connection followed by computation
    • Text generator based on a context
  • Byte Pair Encoding (BPE) algorithm for tokenization, popularized by the GPT-2 paper

Roadmap 🎯

  • Custom dataset training
  • Pretraining and fine-tuning experiments
  • Implementation of "Encoder" block

Contributions and feedback are welcome!

About

Hands-on implementation of transformer and related models

Topics

Resources

License

Stars

Watchers

Forks

Languages