The mathematical evolution behind the journey of multi layer perceptron from historic perceptron algorirthm.
🎉 Presentation Shared: The Journey of the Artificial Neuron is Here!
Super excited to share my presentation on the incredible mathematical journey that built the foundations of modern AI! This repository contains the PDF titled "The Calculus of Cognition: From Biological Neuron to Backpropagation."
It was an awesome experience exploring how a simple mathematical idea, the Perceptron, evolved into the powerful structure (the MLP) that underpins today's Deep Learning revolution.
For context, this was a special seminar delivered By GiGi Koneti (me) during our 'Introduction to Artificial Intelligence' class.
- Date: Tuesday, October 28, 2025
- Duration: A deep dive session that ran for well over an hour, starting at 1:00 PM.
- Setting: Room No. 312, Building Number 13.
- Course: Semester 1, Department of Automobile Engineering.
Our journey followed the story of these computational neurons and the brilliance needed to make them truly "learn":
- The Perceptron (1957): This was the first spark! Introduced by Frank Rosenblatt, it modeled a biological neuron perfectly—a weighted sum of inputs that "fires" if a threshold is crossed. It was AI's first true linear classifier.
- The Problem Child (The XOR Flaw): AI hit its first major roadblock. As proven by Minsky & Papert, the single Perceptron could only draw a single straight line to classify data. It failed miserably at simple non-linear problems like the XOR gate, leading to the first "AI Winter."
- The Breakthrough (The Multi-Layer Perceptron - MLP): The solution was surprisingly simple: stack the neurons! By adding one or more Hidden Layers, the MLP gained the power to create incredibly complex, non-linear decision boundaries.
- The Calculus Secret Sauce (Backpropagation): The big question was how to train the weights in those hidden layers. Backpropagation (popularized by Geoffrey Hinton) is the elegant mathematical answer. It's simply the Chain Rule of differential calculus applied to efficiently spread the error backward from the output, telling every single weight how much to adjust itself.
If you want the full details, the presentation walks through the math in a clear, step-by-step way. Here’s the core of the evolution:
Think of a Perceptron as a simple decision maker. It takes inputs (
The Decision Formula (Weighted Sum):
It calculates the total signal
If
The MLP stacked these neurons, making it non-linear. But how do we teach the inside (the hidden layers)? This is where Backpropagation comes in—it's the algorithm that learns.
Our goal is to minimize the Loss (
The Learning Rule (Gradient Descent): To find the fastest way down the error "hill," we move against the gradient (the slope) for every single weight.
-
$\alpha$ is the Learning Rate (how big a step we take). -
$\frac{\partial L}{\partial w}$ is the Partial Derivative, calculated by skillfully applying the Chain Rule across the whole network. This tells us exactly how much the error changes with respect to that specific weight.
This single equation, powered by the calculus of the Chain Rule, is what allows complex, multi-layered networks to learn from data and is the real 'secret' behind the AI we see today!
You can download the full presentation slides here:
