Skip to content

plugyawn/tinything

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

21 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

tinythings: tiny implementations in tinygrad

Currently a small(er) nanoGPT, written in tinygrad, from scratch. Tinygrad allows acceleration on virtually any backend: probably the simplest way to write accelerated training on modern Macs.

  • CausalAttention with RoPE [layers/attention.py]
  • Basic FFNs [layers/feedforward.py]
  • SwiGLU FFNs [layers/feedforward.py]
  • Mixture of Experts [layers/moe.py]
  • SGD [optimizers/sgd.py]
  • Adam [optimizers/sgd.py]
  • Muon [optimizers/muon.py]
  • LayerNorm [utils/transformer_methods.py]
  • Cross Entropy [utils/loss_functions.py]
  • Naive character-level tokenization [utils/dataloader.py]
  • Byte-pair encoding [utils/dataloader.py]
  • Diffusion text modelling

About

As barebones as you can get the GPT, now accelerated on Macs thanks to tinygrad.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages