Skip to content

As barebones as you can get the GPT, now accelerated on Macs thanks to tinygrad.

Notifications You must be signed in to change notification settings

plugyawn/tinything

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

21 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

tinythings: tiny implementations in tinygrad

Currently a small(er) nanoGPT, written in tinygrad, from scratch. Tinygrad allows acceleration on virtually any backend: probably the simplest way to write accelerated training on modern Macs.

  • CausalAttention with RoPE [layers/attention.py]
  • Basic FFNs [layers/feedforward.py]
  • SwiGLU FFNs [layers/feedforward.py]
  • Mixture of Experts [layers/moe.py]
  • SGD [optimizers/sgd.py]
  • Adam [optimizers/sgd.py]
  • Muon [optimizers/muon.py]
  • LayerNorm [utils/transformer_methods.py]
  • Cross Entropy [utils/loss_functions.py]
  • Naive character-level tokenization [utils/dataloader.py]
  • Byte-pair encoding [utils/dataloader.py]
  • Diffusion text modelling

About

As barebones as you can get the GPT, now accelerated on Macs thanks to tinygrad.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages