You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
A clean, ground-up implementation of the Transformer architecture in PyTorch, including positional encoding, multi-head attention, encoder-decoder layers, and masking. Great for learning or building upon the core model.
Mini Transformer (Implemented From Scratch) A from-scratch implementation of a Transformer in Python using only NumPy. This repository is designed for educational purposes to understand the inner workings of the Transformer architecture, including positional encoding, attention mechanisms, and basic decoding — without relying on frameworks like Py