Founder, Eureka Labs |Ex-OpenAI/Tesla
OpenAI founding member, former Tesla AI Director, and creator of nanoGPT/micrograd/llm.c. Founded Eureka Labs for AI-native education. Coined 'vibe coding' and 'Software 3.0.' His YouTube 'Zero to Hero' series has 1M+ subscribers. Released nanochat, autoresearch, and LLM101n.
GitHub
27 repositories · 344.5k total stars
The simplest, fastest repository for training/finetuning medium-sized GPTs.
The best ChatGPT that $100 can buy.
AI agents running research on single-GPU nanochat training automatically
LLM101n: Let's build a Storyteller
LLM training in simple, raw C/CUDA
A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training
Neural Networks: Zero to Hero
Inference Llama 2 in one file of pure C
Languages
Episodes
Biography
Andrej Karpathy is a Slovakian-Canadian AI researcher, educator, and engineer who has shaped modern deep learning practice across research, industry, and education. Born in Bratislava, he moved to Toronto at 15, earned a BSc in Computer Science and Physics from the University of Toronto (2009), an MSc from UBC (UBC) (2011), and a PhD from Stanford under Fei-Fei Li (2015) with the thesis 'Connecting Images and Natural Language.' He co-designed and was lead instructor of Stanford's CS231n, the first deep learning course at Stanford. As a founding member of OpenAI (2015-2017) he worked on deep learning, generative models, and reinforcement learning. As Director of AI at Tesla (2017-2022) he led the Autopilot computer vision team. After a brief return to OpenAI (2023-2024) working on midtraining and synthetic data, he founded Eureka Labs in July 2024, an AI-native education company. Karpathy is the creator of iconic open-source projects including char-rnn, convnetjs, nanoGPT, minGPT, micrograd, llm.c, llama2.c, minbpe, nanochat, autoresearch, and LLM101n. His YouTube channel 'Neural Networks: Zero to Hero' has over 1 million subscribers. He coined the terms 'Software 2.0' and 'vibe coding,' and popularized 'context engineering' as a replacement for 'prompt engineering.' His GitHub has 149,000+ followers, making him one of the most followed developers on the platform.
The simplest, fastest repository for training/finetuning medium-sized GPTs. Arguably the most influential educational ML codebase, with 55k+ GitHub stars. Made transformer training accessible to individual developers.
Full-stack ChatGPT clone: pretraining, SFT, RLHF, and inference in ~8,000 lines of PyTorch. Trainable end-to-end on a single GPU for under $100. 49k+ stars.
AI agents that autonomously run ML experiments on a single GPU overnight. 630 lines of Python, built on the nanochat training core. 43k+ stars.
LLM training in simple, raw C/CUDA with no PyTorch or Python dependency. Achieved multi-GPU bfloat16 training with flash attention 7% faster than PyTorch nightly. 29k+ stars.
Eureka Labs' first course: 'Let's Build a Storyteller.' An undergraduate-level guide to training your own AI from scratch, designed to be guided by an AI Teaching Assistant. 36k+ stars for the course repo alone.
A tiny scalar-valued autograd engine and neural net library with PyTorch-like API. The canonical educational resource for understanding backpropagation. 15k+ stars.
YouTube lecture series covering backpropagation, language modeling, tokenization, attention, and GPT training from scratch. Over 1 million subscribers. Required viewing for aspiring ML engineers.
Co-designed and was primary instructor for Stanford's first deep learning course, Convolutional Neural Networks for Visual Recognition. Grew from 150 to 750 students and became one of the most popular CS courses at Stanford.
As Director of AI at Tesla (2017-2022), led the computer vision team responsible for Autopilot's vision-based self-driving pipeline, moving Tesla from radar+camera to a vision-only approach.
Coined 'Software 2.0' (neural networks as code written by optimization) and later 'Software 3.0' (natural language as the programming interface for LLMs), reshaping how the industry thinks about software development paradigms.
Context engineering is the delicate art and science of filling the context window with just the right information for the next step.
I've never felt this much behind as a programmer. The profession is being dramatically refactored as the bits contributed by the programmer are increasingly sparse and between.
Vibe coding is a new kind of coding where you fully give in to the vibes, embrace exponentials, and forget that the code even exists.
The most important quality in a startup founder right now is taste. You need to know what is good because AI can produce infinite slop.
Research generated March 19, 2026