Curated Resources
The best books, courses, research papers, tools, and communities for learning deep learning — hand-picked and verified.
📚 Textbooks
These books range from beginner-friendly introductions to graduate-level references. All are available free online or have free editions.
Deep Learning
The definitive graduate-level textbook. Covers mathematical foundations, modern deep networks, and research topics including autoencoders, representation learning, and generative models.
Dive into Deep Learning
Interactive, code-first textbook adopted by 500+ universities. Every concept comes with runnable PyTorch, TensorFlow, and JAX code. Covers CNNs, RNNs, attention, transformers, GANs, and optimization.
Neural Networks and Deep Learning
Beginner-focused free online book. Explains core concepts — backpropagation, gradient descent, regularization — with clear writing and interactive visualizations. Great first book.
Understanding Deep Learning
Modern, comprehensive textbook covering everything from linear regression to transformers and diffusion models. Includes 400+ figures and exercises. Freely available as PDF.
The Little Book of Deep Learning
A concise 170-page overview of deep learning designed to fit in a shirt pocket. Covers architectures, training, and applications with remarkable clarity and density.
🎓 University Courses
Full lecture series from Stanford, MIT, and other top universities — all freely available online. These are the same courses taken by researchers at leading AI labs.
CS231n: Deep Learning for Computer Vision
The gold standard for computer vision. Covers image classification, CNNs, object detection, segmentation, generative models, and visualizing learned features. Assignments in Python / NumPy / PyTorch.
CS224N: NLP with Deep Learning
Comprehensive NLP course covering word vectors, dependency parsing, RNNs, attention, transformers, pretraining (BERT, GPT), question answering, and text generation.
MIT 6.S191: Introduction to Deep Learning
MIT's official introductory deep learning course. Covers foundations, CNNs, RNNs, generative models, reinforcement learning, and AI for science. Updated annually with labs in TensorFlow.
Practical Deep Learning for Coders
Top-down, code-first approach to deep learning. Teaches how to train state-of-the-art models in computer vision and NLP within the first lesson. Uses PyTorch and the fastai library.
Neural Networks: Zero to Hero
Build neural networks from scratch in Python. Starts with micrograd (autograd engine), builds makemore (character-level language model), and culminates in a GPT from scratch. No libraries needed.
Deep Learning Specialization
Five-course specialization covering neural networks, hyperparameter tuning, structuring ML projects, CNNs, and sequence models. Assignments in Python / TensorFlow. Audit for free.
📄 Landmark Research Papers
The 10 most influential papers in deep learning history. Reading these will give you a solid understanding of where the field came from and where it's going.
Attention Is All You Need
Introduced the Transformer architecture — self-attention replacing recurrence entirely. Foundation of GPT, BERT, T5, and virtually all modern language models.
Deep Residual Learning for Image Recognition
Introduced skip connections (ResNet), enabling training of 152+ layer networks. Won ILSVRC 2015 and remains the backbone of most vision systems.
Generative Adversarial Networks
Introduced the GAN framework — two networks (generator and discriminator) competing in a minimax game. Pioneered neural image generation.
BERT: Pre-training of Deep Bidirectional Transformers
Showed that pretraining a bidirectional transformer on masked language modeling then fine-tuning produced SOTA on 11 NLP tasks. Sparked the pretraining revolution.
Gradient-Based Learning Applied to Document Recognition
Introduced LeNet-5 and demonstrated end-to-end CNN training for handwritten digit recognition. Established the convolutional neural network architecture used today.
Long Short-Term Memory
Solved the vanishing gradient problem in RNNs with gated memory cells. LSTM became the dominant architecture for sequence modeling for two decades until transformers.
ImageNet Classification with Deep Convolutional Neural Networks
AlexNet won ILSVRC 2012 by a huge margin using GPU-trained deep CNNs with ReLU and dropout. The result that reignited the deep learning revolution.
Neural Machine Translation by Jointly Learning to Align and Translate
Introduced the attention mechanism for seq2seq models, allowing the decoder to focus on relevant source words. The crucial precursor to the Transformer.
Denoising Diffusion Probabilistic Models
Showed that iterative denoising can generate high-quality images rivaling GANs. Foundation of Stable Diffusion, DALL·E 2, and Midjourney.
Improving Language Understanding by Generative Pre-Training
The original GPT paper. Showed that unsupervised pretraining of a transformer decoder followed by supervised fine-tuning achieves strong NLP performance. Led to GPT-2, 3, 4.
🛠️ Tools & Frameworks
The essential software for building, training, and deploying deep learning models. All free and open-source.
PyTorch
Dynamic computation graphs, Pythonic API. The most popular framework in research and increasingly in production.
pytorch.org →TensorFlow
Google's production-ready ML platform. TensorFlow 2.x with Keras API for rapid prototyping and TF Serving for deployment.
tensorflow.org →Hugging Face
500K+ pretrained models, datasets, and Spaces. The hub for NLP, vision, and audio models with the Transformers library.
huggingface.co →Google Colab
Free GPU/TPU Jupyter notebooks in the cloud. Zero setup — import data, train models, and share results instantly.
colab.google →Weights & Biases
Experiment tracking, hyperparameter sweeps, and model versioning. Free for personal and academic use.
wandb.ai →Papers With Code
Browse state-of-the-art results, find code implementations for any paper, and compare model performance on benchmarks.
paperswithcode.com →🤝 Communities
Where to ask questions, share progress, read discussions, and find collaborators.
r/MachineLearning
3M+ members. Research paper discussions, industry news, and AMA sessions with top researchers.
Visit →r/learnmachinelearning
Beginner-friendly subreddit. Study groups, resource recommendations, project feedback, and career advice.
Visit →Hugging Face Forums
Official community for Transformers, Datasets, and model hosting. Great for NLP and vision model questions.
Visit →PyTorch Forums
Official PyTorch discussion forum. Best place for framework-specific questions, debugging, and feature requests.
Visit →