โ† Explore Courses |
Ai Models From Scratch Beginners Edition

Start building with us today.

Buy this course โ€” $199.00

Ai Models From Scratch Beginners Edition

๐Ÿ“Š beginner ๐Ÿ‘จโ€๐Ÿซ Expert Instructor

AI Models From Scratch

Course 1 of 3 โ€” Beginner Edition

๐ŸŸข Beginner Level๐Ÿ“ฆ 30 Lessons๐Ÿ–ฅ๏ธ CPU Only๐Ÿ Python + NumPy
LevelHands-On AppsAny LaptopNo GPU Needed

Build every piece of an AI model from mathematical primitives. No black boxes. No frameworks. Every neuron, every gradient, every decision boundary โ€” coded and visualised by you.


Table of Contents

  1. Why This Course?

  2. What You Will Build

  3. Who Should Take This Course?

  4. What Makes This Course Different?

  5. Key Topics Covered

  6. Prerequisites

  7. Learning Outcomes

  8. Course Structure

  9. Full Curriculum โ€” All 30 Lessons

  10. Lesson-Level Learning Objectives

  11. Section Deep Dives

  12. Appendix: Complete Setup Guide


Why This Course?

Most AI courses hand you a model.fit() call and call it education. That is not learning โ€” that is using someone else's understanding.

This course exists because the engineers who build truly reliable AI systems โ€” at Google, Meta, DeepMind, and Anthropic โ€” are not the ones who learned the most frameworks. They are the ones who understand what every line of a neural network is actually doing. This course gives you that understanding, built brick by brick through working software.

Every single lesson ends with a running Streamlit application. Not a notebook. Not pseudocode. A real app with inputs, processing, model, and visual output โ€” something you can show anyone and explain completely because you built every piece of it from scratch.

"The engineers who really understand AI are the ones who could rebuild it from scratch if they had to. That is the bar this course sets โ€” and clears."
โ€” Design principle behind this curriculum

Five Reasons to Take This Course

  1. You will never be fooled by a framework again โ€” Once you have written forward pass and backprop yourself, PyTorch's autograd is transparent โ€” not magical.

  2. Debugging AI becomes systematic, not guesswork โ€” When your loss curves behave strangely, you will know exactly which layer, which operation, or which weight is misbehaving.

  3. Your Intermediate and Expert courses unlock fully โ€” Every concept in Part 2 (PyTorch) and Part 3 (LLMs, RAG, MLOps) references foundations you built here. Nothing will feel foreign.

  4. You build a portfolio of 30 working apps โ€” Each lesson is a self-contained Streamlit app. Thirty lessons means thirty portfolio items, each explaining a core AI concept visually.

  5. Industry engineering habits start on Day 1 โ€” Modular code structure, evaluation rigor, reproducibility, and documentation โ€” the same standards that big tech teams enforce.


What You Will Build

30 self-contained Streamlit applications โ€” each one a live, interactive demonstration of a core AI concept.

AppWhat It Demonstrates
NumPy PlaygroundA live math workbench โ€” operations, shapes, broadcasting
Gradient Descent LabWatch an optimizer converge frame-by-frame on any function
The NeuronSingle neuron with sliders โ€” see it fire or stay silent
XOR SolverA 2-layer network solving the classic non-linear problem
Backprop TracerClick 'Next' and watch gradients flow backward, step by step
Data CleanerUpload messy CSV, clean it, download production-ready data
Feature EngineerAdd polynomial/interaction features, live correlation heatmap
Spam DetectorLogistic regression classifier built with only 60 lines of NumPy
KNN VisualizerPlot points, click test input, watch K neighbors vote live
Metrics DashboardPrecision, Recall, F1, AUC-ROC with interactive gauges
Fit VisualizerPolynomial degree slider reveals overfitting/underfitting live
Digit Guesser (Capstone)Draw a digit โ†’ pure NumPy model predicts 0-9 โ€” no libraries

Every app shares the same four-file structure so the pattern becomes muscle memory:

Code
lesson_XX/
โ”œโ”€โ”€ app.py       # Streamlit UI โ€” launch with: streamlit run app.py
โ”œโ”€โ”€ model.py     # Core AI logic โ€” the piece you build from scratch
โ”œโ”€โ”€ train.py     # Training script with logging and checkpointing
โ””โ”€โ”€ README.md    # Run it ยท Break it ยท Extend it

Who Should Take This Course?

Built for anyone who writes, designs, builds, or decides on software โ€” not just data scientists.

This course was deliberately designed for a wide technical audience. AI is no longer a specialisation โ€” it is a layer of every product, every system, every engineering decision.

AudienceWhat This Course Gives You
Software Engineers / DevelopersUnderstand what AI models are actually doing so you can integrate, debug, and optimize them confidently.
Software Architects & DesignersMake informed decisions about where AI fits in a system design and what its failure modes look like.
Product ManagersUnderstand AI tradeoffs at a technical level to set realistic expectations and evaluate feasibility.
QA / SRE / DevOps EngineersKnow how to write tests, monitor outputs, and build confidence intervals around AI system behavior.
Data EngineersUnderstand how the data you build pipelines for is actually consumed by downstream models.
Fresh CS / Engineering GraduatesClose the gap between university theory and the practical engineering work that teams actually do.
Engineering ManagersEvaluate AI work rigorously, understand what 'good' looks like, and spot when something is being cargo-culted.
UI/UX DesignersUnderstand model confidence, uncertainty, and edge cases that directly affect how AI-powered UIs should behave.
Technical Writers & IT ConsultantsWrite and advise with the depth that comes from having actually built the systems you document.

What Makes This Course Different?

Six concrete differences from every other AI course on the market.

  1. Every Lesson Ships a Working App โ€” Not a notebook, not a tutorial walkthrough โ€” a real Streamlit application with UI, model logic, and visual output. Thirty lessons = thirty portfolio items that run on anyone's computer.

  2. Built on CPU. On Purpose. โ€” Every lesson runs on any laptop with 4GB RAM and no GPU. This forces every app to be lean, well-optimized, and practical. The discipline transfers directly to production engineering.

  3. Scratch-First, Framework-Second โ€” You implement neurons, backpropagation, gradient descent, and loss functions yourself before ever touching PyTorch. This is how engineers who actually understand AI work.

  4. Production File Structure From Lesson 1 โ€” Every lesson follows the same modular four-file layout used by AI engineering teams at production companies. You are learning a work structure, not a course structure.

  5. Four-Step Learning Loop per Lesson โ€” Run it (60 seconds to a working app) โ†’ Understand it (read model.py line by line) โ†’ Break it (change one thing intentionally) โ†’ Extend it (ship a new feature). This loop builds real engineering intuition.

  6. Designed for Non-Data-Scientists โ€” The curriculum was deliberately structured so that a software architect, product manager, QA engineer, or designer gets as much value as a data scientist. AI literacy is the goal, not AI specialisation.


Key Topics Covered

Six interlocking domains, each with its own section and dedicated Streamlit applications.

๐Ÿ“ Mathematical Engine

  • N-dimensional arrays & broadcasting

  • Matrix multiplication & linear transforms

  • Derivatives, partial derivatives & gradients

  • Probability distributions & the CLT

  • Vectorized computation patterns

โšก Neuron & Learning

  • Weighted sum, bias, threshold mechanics

  • Activation functions: 6 types compared

  • Loss function geometry & convexity

  • Gradient descent convergence mechanics

  • Learning rate sensitivity & optimal ranges

๐Ÿง  Network Architecture

  • Forward pass computation graphs

  • Backpropagation via chain rule, by hand

  • Depth vs width capacity tradeoffs

  • Weight initialization: He, Xavier, zero

  • Custom training loop anatomy

๐Ÿ—‚๏ธ Data Engineering

  • Null imputation & outlier detection

  • Feature engineering: poly, log, bins

  • Stratified train/val/test splitting

  • Min-Max, Standard, Robust normalization

  • Augmentation as regularization

๐ŸŒฒ Classical ML Models

  • Logistic regression from 60 lines

  • Linear regression with residual analysis

  • KNN: distance metrics & K sensitivity

  • Decision tree splits: Gini vs entropy

  • Random forests & out-of-bag error

๐Ÿ“Š Evaluation & Shipping

  • Precision, Recall, F1, AUC-ROC suite

  • Confusion matrices for multi-class

  • Bias-variance tradeoff via learning curves

  • Model serialization & version metadata

  • End-to-end digit recognition system


Prerequisites

The shortest prerequisite list in any serious AI course โ€” intentionally.

AreaWhat You Need to KnowLevel Required
Python ProgrammingWrite functions, loops, list comprehensions, and import librariesComfortable
Basic MathFractions, percentages, basic algebra (e.g., solve for x)High school
Command LineRun commands in a terminal (cd, pip install, python file.py)Basic
AI / ML KnowledgeNone required โ€” this course builds the foundations from zeroNone
GPU / Cloud AccountNot required โ€” every lesson runs on CPU on any 4GB RAM laptopNone
Advanced MathCalculus and linear algebra are taught through code in Lessons 1โ€“5None
Prior ML FrameworksPyTorch, TensorFlow โ€” not used in this course (those are Parts 2 and 3)None

"The only prerequisite that truly matters is curiosity about how things actually work. If you have ever looked at a model's output and wondered 'but what is it actually doing?' โ€” this course is for you."
โ€” Course design rationale


Learning Outcomes

What you will be able to do โ€” concretely โ€” after completing all 30 lessons.

  • โœ… Explain exactly what a neuron computes and why it needs activation

  • โœ… Implement gradient descent from scratch on any differentiable function

  • โœ… Write a training loop with loss tracking, batching, and early stopping

  • โœ… Debug a model by reading its loss curve and weight histograms

  • โœ… Implement backpropagation through a 3-layer network with no libraries

  • โœ… Build logistic regression, KNN, and decision trees from raw NumPy

  • โœ… Design a data preprocessing pipeline that prevents leakage

  • โœ… Evaluate a model with precision, recall, F1, and AUC-ROC correctly

  • โœ… Identify overfitting/underfitting and apply the right remedy

  • โœ… Ship a working Streamlit AI application from a blank file

  • โœ… Explain bias-variance tradeoff to a non-technical stakeholder

  • โœ… Prepare a model for hand-off to Part 2 (PyTorch) and Part 3 (MLOps)


Course Structure

Six sections, each targeting a distinct layer of understanding, each ending with a working application.

SectionTitleLessons
๐Ÿ“ Section 1Math EngineL1โ€“5
โšก Section 2The NeuronL6โ€“10
๐Ÿง  Section 3NetworksL11โ€“15
๐Ÿ—‚๏ธ Section 4Data EngineeringL16โ€“20
๐ŸŒฒ Section 5Classic MLL21โ€“25
๐Ÿ“Š Section 6EvaluationL26โ€“30

Section Overviews

01 โ€” The Mathematical Engine (Lessons 1โ€“5)
Master the linear algebra, calculus, and probability that every AI model is built on โ€” through code, not textbooks.

02 โ€” The Neuron: AI's Atom (Lessons 6โ€“10)
Build the core computational unit of every neural network from scratch and understand exactly why it learns.

03 โ€” Your First Network (Lessons 11โ€“15)
Stack neurons into layers, implement backpropagation by hand, and write your own training loop from zero.

04 โ€” Data Engineering (Lessons 16โ€“20)
Learn why data prep makes or breaks models โ€” cleaning, feature engineering, splitting, scaling, and augmentation.

05 โ€” Classical ML Models (Lessons 21โ€“25)
Build logistic regression, KNN, decision trees and random forests from scratch, understanding every algorithm decision.

06 โ€” Evaluation & Graduation (Lessons 26โ€“30)
Master the full evaluation toolkit and ship a production-ready end-to-end digit recognition system.


Full Curriculum โ€” All 30 Lessons

Every lesson is a standalone Streamlit app. Each row below is a deliverable, not a topic.


Section 01: The Mathematical Engine ยท Lessons 1โ€“5

#App NameWhat You BuildConcepts MasteredTools
1NumPy Playground โ€” Python & NumPy for AIAn interactive calculator that performs vector/matrix operations and explains each step with live outputN-dimensional arrays, broadcasting rules, vectorization vs loops, dtype castingNumPy, Streamlit
2Tensor Shape Visualizer โ€” Data as TensorsA shape explorer: upload any CSV or image and see its 0Dโ€“4D tensor structure, slices, and axis meaningsTensor rank, shape, axes, indexing, slicing, reshape vs viewNumPy, Matplotlib, Streamlit
3Matrix Transformer โ€” Linear Algebra in CodeA live 2D transformation app: type a matrix, watch unit vectors stretch/rotate on a grid in real timeDot product, matrix multiply, transpose, inverse, determinant, identityNumPy, Plotly, Streamlit
4Derivative Driller โ€” Calculus Intuition for AIA function plotter with tangent-line animation: drag a point, see slope update instantly for any f(x)Derivative, partial derivative, gradient direction, slope as rate of changeNumPy, Matplotlib, Streamlit
5Probability Sandbox โ€” Stats & Probability FoundationsA dice/coin simulator that tracks empirical vs theoretical distributions as sample size grows livePDF, CDF, mean, variance, standard deviation, Central Limit Theorem, samplingNumPy, Plotly, Streamlit

Section 02: The Neuron โ€” AI's Atom ยท Lessons 6โ€“10

#App NameWhat You BuildConcepts MasteredTools
6The Neuron โ€” Single Neuron from ScratchA neuron with sliders for each input weight and bias โ€” fire/don't fire decision visualized in real timeWeighted sum, bias, threshold, perceptron model, binary decisionNumPy, Streamlit
7Activation Explorer โ€” Activation Functions Deep DiveSide-by-side plots of 6 activation functions with gradient curves; sliders distort input range liveSigmoid saturation, dying ReLU, Tanh symmetry, GELU smoothness, non-linearityNumPy, Plotly, Streamlit
8Loss Lens โ€” Loss Functions DemystifiedPaste predicted vs actual values โ€” app computes MSE, MAE, Cross-Entropy with penalty breakdown barsRegression vs classification loss, convex surface, squared penalty intuitionNumPy, Plotly, Streamlit
9Gradient Descent Lab โ€” Gradient Descent Step by StepPick any 1D function, set learning rate, watch the optimizer ball roll to the minimum frame by frameGradient direction, step size, convergence, local vs global minima, overshootingNumPy, Matplotlib, Streamlit
10Learning Rate Finder โ€” The Learning Rate EffectTrains a tiny model at 8 different LRs simultaneously โ€” overlaid loss curves show diverge/convergeLR too high diverges, LR too low stalls, optimal range, LR sensitivityNumPy, Plotly, Streamlit

Section 03: Your First Network ยท Lessons 11โ€“15

#App NameWhat You BuildConcepts MasteredTools
11XOR Solver โ€” First Neural Network from ScratchA 2-layer network solves XOR, coded in 80 lines of pure NumPy โ€” shows decision boundary liveForward pass, non-linear separability, hidden layer role, XOR impossibility for linearNumPy, Plotly, Streamlit
12Backprop Tracer โ€” Backpropagation Step by StepStep through a 3-node network: click Next to watch gradients flow backward layer by layer, color-codedChain rule, upstream/downstream gradient, accumulation, vanishing gradient warningNumPy, Graphviz, Streamlit
13Layer Stack Builder โ€” Depth vs Width ExperimentsUI to add/remove layers and neurons, train on toy data, see accuracy + decision boundary change liveModel capacity, universal approximation theorem, depth vs width tradeoffNumPy, Plotly, Streamlit
14Weight Watcher โ€” Weight Initialization MattersCompare zero-init vs random vs He vs Xavier: histogram of weights as training evolves each epochSymmetry breaking, exploding/vanishing gradients, He/Xavier derivation intuitionNumPy, Plotly, Streamlit
15Mini Trainer โ€” Writing Your Own Training LoopA configurable training loop with live loss/accuracy chart, epoch counter, and early-stop toggleEpoch, batch, iteration, loss plateau, early stopping, training loop anatomyNumPy, Streamlit, Plotly

Section 04: Data Engineering ยท Lessons 16โ€“20

#App NameWhat You BuildConcepts MasteredTools
16Data Cleaner โ€” Data Preprocessing PipelinesUpload any CSV โ€” auto-detect nulls, outliers, skewed columns, apply fixes, download cleaned versionNull imputation, z-score outliers, skew detection, column typing, pipeline orderPandas, NumPy, Streamlit
17Feature Engineer โ€” Feature Engineering StudioRaw columns in โ€” add polynomial, log, bin, interaction features โ€” live correlation heatmap updatesFeature crosses, polynomial expansion, log transform, binning, VIF/collinearityPandas, NumPy, Plotly, Streamlit
18Dataset Splitter โ€” Train / Val / Test Splits Done RightDrag sliders to set split ratios, toggle stratification โ€” shows class distribution per split liveData leakage, stratified split, temporal split, holdout discipline, why val existsPandas, Scikit-learn, Streamlit
19Normalizer Lab โ€” Scaling & NormalizationToggle between MinMax, Standard, Robust, L2 โ€” see how each changes the histogram and model convergenceFeature scale sensitivity, normalization vs standardization, robust to outliersPandas, Scikit-learn, Plotly, Streamlit
20Image Augmenter โ€” Data Augmentation TechniquesUpload any image, apply 8 augmentation types, preview a 4ร—4 augmented batch grid in real timeFlipping, rotation, crop, noise, brightness, cutout โ€” regularization via dataPIL, Albumentations, Streamlit

Section 05: Classical ML Models ยท Lessons 21โ€“25

#App NameWhat You BuildConcepts MasteredTools
21Spam Detector โ€” Logistic Regression from ScratchType an email โ€” a scratch-built logistic classifier scores spam probability with word contribution chartSigmoid output, log-odds, MLE, decision threshold, binary cross-entropyNumPy, Streamlit
22House Price Predictor โ€” Linear Regression Deep DiveEnter house features (sqft, rooms, age) โ€” model predicts price with confidence interval plotLeast squares, R-squared, residual analysis, multicollinearity, prediction intervalNumPy, Plotly, Streamlit
23KNN Visualizer โ€” K-Nearest Neighbors InternalsPlot 2D points, click to add test point, watch K neighbors vote live, adjust K with sliderDistance metrics (L1/L2/cosine), curse of dimensionality, K sensitivity, lazy learningNumPy, Plotly, Streamlit
24Tree Builder โ€” Decision Trees from ScratchUpload tabular data โ€” tree grows step-by-step showing Gini/entropy split at each node, pruning toggleInformation gain, Gini impurity, recursive splitting, depth control, overfittingScikit-learn, Graphviz, Streamlit
25Forest Ensemble โ€” Random Forests & BaggingTrain one tree vs random forest โ€” compare accuracy, OOB error, feature importance bar chart liveBagging, feature sampling, OOB error, variance reduction, feature importanceScikit-learn, Plotly, Streamlit

Section 06: Evaluation & Graduation ยท Lessons 26โ€“30

#App NameWhat You BuildConcepts MasteredTools
26Metrics Dashboard โ€” Evaluation Metrics SuiteUpload predictions CSV โ€” auto-computes Accuracy, Precision, Recall, F1, AUC-ROC with gauge dialsPrecision-recall tradeoff, threshold selection, macro/micro/weighted averages, AUCScikit-learn, Plotly, Streamlit
27Confusion Analyzer โ€” Confusion Matrix Deep DiveMulti-class interactive confusion matrix with per-class drill-down and error pattern analysisTP/FP/TN/FN per class, class imbalance impact, per-class precision vs recallScikit-learn, Seaborn, Streamlit
28Fit Visualizer โ€” Overfitting, Underfitting, Just RightAdjust polynomial degree live โ€” model fits training vs val data, bias-variance curve updates liveBias-variance tradeoff, model capacity, generalization gap, learning curvesNumPy, Plotly, Streamlit
29Model Vault โ€” Saving, Loading & Versioning ModelsTrain model, save with metadata (accuracy, params, date), load old versions and compare side-by-sideSerialization, joblib/pickle, reproducibility, version discipline, model cardsJoblib, Pandas, Streamlit
30Digit Guesser โ€” CAPSTONE: Full AI Pipeline in 200 LinesDraw a digit on canvas โ†’ pure NumPy 3-layer net (trained live) predicts 0โ€“9 with confidence barsEnd-to-end pipeline: preprocess โ†’ train โ†’ infer, all scratch, no ML libsNumPy only, Streamlit, canvas

Lesson-Level Learning Objectives

Each lesson has a single, measurable outcome. You know exactly what you built and why.

LessonBy the end of this lesson you can...
L01Vectorize a computation using NumPy broadcasting and explain why it runs 100x faster than a Python loop
L02Reshape a CSV, image, and audio file into named tensors and explain shape, rank, and axis semantics
L03Apply a 2ร—2 matrix to a 2D point cloud and predict the transformed shape before running the code
L04Compute the derivative of any smooth function numerically and analytically, and verify they match
L05Explain the Central Limit Theorem by running a simulation that shows it converging in real time
L06Build a single neuron from scratch and explain every arithmetic operation in a forward pass
L07Choose the correct activation function for a given problem by comparing saturation and gradient properties
L08Implement MSE and Cross-Entropy loss from scratch and explain geometrically why they penalise errors
L09Run gradient descent on a custom function and predict convergence behaviour given a learning rate
L10Identify the optimal learning rate range for a model by interpreting a loss-vs-LR curve
L11Build a 2-layer network that solves XOR and explain why a single neuron cannot
L12Trace the backward pass of a 3-neuron network manually, then verify with the app's gradient display
L13Predict the effect of adding a hidden layer on decision boundary complexity before running the training
L14Explain why zero-initialisation fails and select the correct initialisation scheme for a given activation
L15Write a training loop from a blank file including loss logging, epoch tracking, and early stopping
L16Detect and fix three data quality issues (nulls, outliers, skew) in an uploaded CSV using the app
L17Add polynomial and interaction features to a dataset and measure the change in model correlation
L18Split a dataset with stratification and confirm class balance across all three splits using the app's chart
L19Select the correct normalisation method for a given feature distribution and justify the choice
L20Build an 8-augmentation pipeline for images and explain how each transform prevents overfitting
L21Implement logistic regression from scratch in 60 lines of NumPy and classify spam with >90% accuracy
L22Fit a linear regression model, analyse residuals, and identify whether the relationship is actually linear
L23Explain the K sensitivity of KNN by watching accuracy change as K moves from 1 to 30 in the app
L24Build a decision tree that splits on information gain and explain every decision node in plain English
L25Quantify how random forests reduce variance over a single tree using OOB error on a real dataset
L26Compute Precision, Recall, F1, and AUC-ROC for a multi-class model and interpret each metric correctly
L27Identify error patterns in a confusion matrix and propose a class-specific improvement strategy
L28Diagnose overfitting vs underfitting by reading a learning curve and prescribe the correct remedy
L29Save and reload a model with full metadata, confirm reproducibility, and compare two saved versions
L30Ship a digit recogniser using only NumPy โ€” data loading, training, inference, and Streamlit UI included

Section Deep Dives

What you actually learn inside each of the six sections โ€” the non-obvious insights that make this curriculum different.


Section 01: The Mathematical Engine (Lessons 1โ€“5)

Most AI resources tell you 'you need linear algebra and calculus.' This section shows you exactly which pieces โ€” and only those pieces โ€” and hands you a running tool that makes each concept visceral before moving on.

Broadcasting is not magic: Lesson 1 forces you to predict the output shape before running the code. Broadcasting is a deterministic rule, and understanding it prevents a class of bugs that haunt production AI code.

Derivatives as sensitivity: Lesson 4 treats derivatives not as exam equations but as slope-at-a-point tools. You drag a slider and watch the tangent line update. That intuition is the entire foundation of how gradient descent knows which way to step.

The Central Limit Theorem in practice: Lesson 5 lets you run sampling experiments live. You will watch non-normal distributions converge to normal as sample size grows โ€” the reason validation metrics are reliable even from noisy data.


Section 02: The Neuron โ€” AI's Atom (Lessons 6โ€“10)

The neuron is the unit of understanding in all of AI. Engineers who cannot explain a single neuron's computation have no foundation for debugging anything deeper. This section eliminates that gap entirely.

Activation functions have tradeoffs: Lesson 7 plots 6 activation functions side-by-side with their gradient curves. You see exactly why sigmoid saturates, why ReLU dies with negative inputs, and why GELU is preferred in modern transformers. These tradeoffs appear in real architecture decisions.

Loss surfaces determine trainability: Lesson 8 shows you the penalty landscape of different loss functions. Cross-entropy's log penalty creates a steeper gradient far from truth โ€” that is why it trains faster than MSE for classification, and you can see it.

Learning rate is the single most important hyperparameter: Lesson 10 trains the same model at 8 learning rates simultaneously and overlays the loss curves. No textbook explanation of divergence is as clear as watching it happen in real time.


Section 03: Your First Network (Lessons 11โ€“15)

Building backprop by hand is the defining exercise of this section. Every AI engineer should be able to trace a gradient from loss to weight without a framework doing it for them. Lesson 12 makes this non-optional.

XOR reveals why depth matters: Lesson 11 shows you that a single layer literally cannot draw the decision boundary that separates XOR inputs. Adding one hidden layer solves it instantly. That geometry is why deep networks are more powerful than wide ones.

Initialization determines whether training starts: Lesson 14 compares zero initialization (weights never differentiate), random initialization (weights explode), and He/Xavier (weights start at the right scale). The histograms make the difference impossible to ignore.

The training loop is architecture, not boilerplate: Lesson 15 walks through every line of a training loop: why you zero gradients, why you call .backward(), why batch size changes the loss landscape. Engineers who skip this explanation treat training as a ritual rather than an engineering decision.


Section 04: Data Engineering (Lessons 16โ€“20)

The model you build is only as good as the data you feed it. This section covers data engineering decisions that change model accuracy by 10โ€“30 percentage points โ€” decisions that most courses treat as preprocessing trivia.

Data leakage is the most common production failure mode: Lesson 18 makes data leakage visible: splitting after normalization, letting val/test statistics influence training features. You toggle the leaky version and watch evaluation metrics become falsely optimistic.

Normalization choice changes convergence speed: Lesson 19 lets you compare MinMax, Standard, and Robust scaling on the same dataset and training run. Standard scaling converges 3ร— faster on most datasets โ€” but Robust is essential when outliers are present.

Augmentation is regularization, not data generation: Lesson 20 reframes augmentation: it is not about making your dataset bigger โ€” it is about making your model invariant to transforms it will encounter in production. That conceptual shift changes how engineers design augmentation pipelines.


Section 05: Classical ML Models (Lessons 21โ€“25)

Classical models are not 'old' models โ€” they are production workhorses that outperform neural networks on tabular data at many companies today. Understanding their internals makes you a more complete AI engineer.

Logistic regression in 60 lines: Lesson 21 builds a logistic classifier from the sigmoid function up, in 60 lines of NumPy. If you cannot explain every line, you cannot debug a neural network layer that does the same computation at scale.

Decision trees are interpretable by design: Lesson 24 builds a tree step-by-step and shows the Gini calculation at every split node. That interpretability is not just pedagogically useful โ€” it is often a production requirement in regulated industries (finance, healthcare, legal).

Random forests quantify their own uncertainty: Lesson 25 introduces out-of-bag error: the random forest evaluates itself on samples not used in each tree's training. This gives you a validation signal without touching your held-out test set โ€” a technique used in production feature selection.


Section 06: Evaluation & Graduation (Lessons 26โ€“30)

Model evaluation is where AI projects win or lose in production. A model with 94% accuracy on an imbalanced dataset may be completely useless. This section teaches you to evaluate models the way senior ML engineers do.

Accuracy is almost never the right metric: Lesson 26 shows you a spam classifier with 96% accuracy that catches no spam at all โ€” because 96% of email is legitimate. Precision, Recall, and F1 tell the truth. Accuracy hides it.

The confusion matrix is a debugging interface: Lesson 27 turns the confusion matrix into an error analysis tool. Which classes are getting confused for each other, and why? That analysis drives the next experiment โ€” it is not just a reporting artifact.

The capstone is a complete engineering system: Lesson 30 requires you to build every component โ€” data loading, preprocessing, the network, the training loop, the inference path, and the Streamlit UI โ€” from a blank file using only NumPy. If you can do this, you understand AI.


Appendix: Complete Setup Guide

Every lesson uses one or more of the following tools. All are open-source, free, and run without any cloud account or GPU.

LibraryRoleInstall CommandUsed In
Python 3.9+Runtime environmentpip install --upgrade pipAll 30 lessons
StreamlitApp frameworkpip install streamlitAll 30 lessons
NumPyArray computingpip install numpyAll 30 lessons
PandasData manipulationpip install pandasLessons 16โ€“30
MatplotlibStatic chartingpip install matplotlibLessons 1โ€“15
PlotlyInteractive chartingpip install plotlyLessons 6โ€“30
Scikit-learnSplitting & metricspip install scikit-learnLessons 15โ€“30
GraphvizTree visualizationpip install graphvizLessons 12, 24
Pillow (PIL)Image handlingpip install PillowLessons 20, 30
AlbumentationsImage augmentationpip install albumentationsLesson 20
JoblibModel serializationpip install joblibLesson 29
SeabornStatistical visualizationpip install seabornLesson 27

Quick Start โ€” Lesson 1 in 90 Seconds

bash
# Clone the course repository
git clone https://github.com/your-org/ai-from-scratch-beginner

# Navigate to Lesson 1
cd lesson_01

# Install dependencies
pip install -r requirements.txt

# Launch the app
streamlit run app.py

"Thirty lessons from now you will build a digit recognition system from a completely blank Python file โ€” no imports except NumPy and Streamlit. That is not a party trick. It is proof that you understand AI at the level that makes every framework, every model, and every debugging session more tractable for the rest of your career."
โ€” What this course delivers

Pricing
$199.00
one-time ยท lifetime access
Or access with monthly subscription โ†’
Level
beginner
Need help?
๐ŸŒ Country:

Showing international pricing ($)