Machine Learning – Fundamentals and Applications

Neural Networks Demystified - Part 1: Data and Architecture – An introduction to neural networks, covering basic structure and how data flows through them.
Neural Networks Demystified - Part 2: Forward Propagation – Explains forward propagation with intuitive visuals.
Neural Networks Demystified - Part 3: Gradient Descent – Introduction to gradient descent for training neural networks.
Neural Networks Demystified - Part 4: Backpropagation – Detailed look at backpropagation.
Neural Networks Demystified - Part 5: Numerical Gradient Checking – Validating gradients using numerical methods.
Neural Networks Demystified - Part 6: Training – Training process explained.
Neural Networks Demystified - Part 7: Overfitting and Regularization – Handling overfitting in neural networks.


Deep Learning Specialization

Deep Learning Specialization - Intro – Overview of deep learning concepts.
Deep Learning Specialization - Neural Networks Basics – Fundamentals of neural networks.
Deep Learning Specialization - Improving Neural Networks – Hyperparameter tuning and optimization.
Deep Learning Specialization - Structuring ML Projects – How to organize ML projects.
Deep Learning Specialization - CNNs – Convolutional Neural Networks explained.
Deep Learning Specialization - Sequence Models – Understanding sequence models and RNNs.


Deep Learning – Intuitions

Why Deep Learning Works Unreasonably Well – A geometric and intuitive explanation of why deep learning performs so effectively.


AI and Society

How DeepSeek Rewrote the Transformer (MLA) – Explains the Multi-Head Latent Attention (MLA) technique introduced by DeepSeek, showcasing how it reengineers the Transformer architecture.


Visualizing Transformers

I Visualised Attention in Transformers – A simple and intuitive visual explanation of why attention is needed in Transformer models, illustrating what it actually does.
Visualizing transformers and attention | Talk for TNG Big Tech Day ‘24 – A comprehensive talk that dives into how Transformers work, including the attention mechanism, tokenization, parallelization, and practical applications across domains.