Implements a GPT-style Transformer model from scratch using our modular deep learning library, . Trains small transformers to perform arithmetic tasks and larger ones to generate Shakespearean text, showcasing attention mechanisms and positional encoding.
-
Notifications
You must be signed in to change notification settings - Fork 0
treytuscai/Transformers
Folders and files
| Name | Name | Last commit message | Last commit date | |
|---|---|---|---|---|
Repository files navigation
About
Implements a GPT‑style Transformer architecture from Vaswani et al. (2017) to generate text and solve arithmetic tasks.
Topics
Resources
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published