nanogpt
π―Skillfrom ovachiever/droid-tings
Trains compact GPT models from scratch on Shakespeare or OpenWebText, enabling deep learning enthusiasts to understand transformer architectures through minimalist, hackable code.
Part of
ovachiever/droid-tings(370 items)
Installation
pip install torch numpy transformers datasets tiktoken wandb tqdmpython data/shakespeare_char/prepare.pypython train.py config/train_shakespeare_char.pypython sample.py --out_dir=out-shakespeare-charpython data/openwebtext/prepare.py+ 3 more commands
Skill Details
Educational GPT implementation in ~300 lines. Reproduces GPT-2 (124M) on OpenWebText. Clean, hackable code for learning transformers. By Andrej Karpathy. Perfect for understanding GPT architecture from scratch. Train on Shakespeare (CPU) or OpenWebText (multi-GPU).
More from this repository10
nextjs-shadcn-builder skill from ovachiever/droid-tings
security-auditor skill from ovachiever/droid-tings
threejs-graphics-optimizer skill from ovachiever/droid-tings
api-documenter skill from ovachiever/droid-tings
secret-scanner skill from ovachiever/droid-tings
readme-updater skill from ovachiever/droid-tings
applying-brand-guidelines skill from ovachiever/droid-tings
Configures Tailwind v4 with shadcn/ui, automating CSS variable setup, dark mode, and preventing common initialization errors.
deep-reading-analyst skill from ovachiever/droid-tings
dependency-auditor skill from ovachiever/droid-tings