<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://qbase.texpertssolutions.com/index.php?action=history&amp;feed=atom&amp;title=Backpropagation</id>
	<title>Backpropagation - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://qbase.texpertssolutions.com/index.php?action=history&amp;feed=atom&amp;title=Backpropagation"/>
	<link rel="alternate" type="text/html" href="https://qbase.texpertssolutions.com/index.php?title=Backpropagation&amp;action=history"/>
	<updated>2026-05-14T15:36:34Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.43.1</generator>
	<entry>
		<id>https://qbase.texpertssolutions.com/index.php?title=Backpropagation&amp;diff=259&amp;oldid=prev</id>
		<title>Thakshashila: Created page with &quot;== Backpropagation ==  &#039;&#039;&#039;Backpropagation&#039;&#039;&#039; (short for &quot;backward propagation of errors&quot;) is a fundamental algorithm used to train neural networks. It calculates how much each weight in the network contributed to the total error and updates them to reduce this error.  === 🧠 Purpose ===  The main goal of backpropagation is to: * Minimize the &#039;&#039;&#039;loss function&#039;&#039;&#039; (error) 📉 * Improve model accuracy over time by adjusting weights 🔧  === 🔁 How It Works (Step-by-Ste...&quot;</title>
		<link rel="alternate" type="text/html" href="https://qbase.texpertssolutions.com/index.php?title=Backpropagation&amp;diff=259&amp;oldid=prev"/>
		<updated>2025-06-11T11:12:10Z</updated>

		<summary type="html">&lt;p&gt;Created page with &amp;quot;== Backpropagation ==  &amp;#039;&amp;#039;&amp;#039;Backpropagation&amp;#039;&amp;#039;&amp;#039; (short for &amp;quot;backward propagation of errors&amp;quot;) is a fundamental algorithm used to train neural networks. It calculates how much each weight in the network contributed to the total error and updates them to reduce this error.  === 🧠 Purpose ===  The main goal of backpropagation is to: * Minimize the &amp;#039;&amp;#039;&amp;#039;loss function&amp;#039;&amp;#039;&amp;#039; (error) 📉 * Improve model accuracy over time by adjusting weights 🔧  === 🔁 How It Works (Step-by-Ste...&amp;quot;&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;== Backpropagation ==&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;&amp;#039;Backpropagation&amp;#039;&amp;#039;&amp;#039; (short for &amp;quot;backward propagation of errors&amp;quot;) is a fundamental algorithm used to train neural networks. It calculates how much each weight in the network contributed to the total error and updates them to reduce this error.&lt;br /&gt;
&lt;br /&gt;
=== 🧠 Purpose ===&lt;br /&gt;
&lt;br /&gt;
The main goal of backpropagation is to:&lt;br /&gt;
* Minimize the &amp;#039;&amp;#039;&amp;#039;loss function&amp;#039;&amp;#039;&amp;#039; (error) 📉&lt;br /&gt;
* Improve model accuracy over time by adjusting weights 🔧&lt;br /&gt;
&lt;br /&gt;
=== 🔁 How It Works (Step-by-Step) ===&lt;br /&gt;
&lt;br /&gt;
Neural network training has two main steps:&lt;br /&gt;
&lt;br /&gt;
# &amp;#039;&amp;#039;&amp;#039;Forward pass&amp;#039;&amp;#039;&amp;#039;: Inputs go through the network to make a prediction.&lt;br /&gt;
# &amp;#039;&amp;#039;&amp;#039;Backward pass (Backpropagation)&amp;#039;&amp;#039;&amp;#039;:&lt;br /&gt;
## Calculate the error (loss)&lt;br /&gt;
## Compute the gradient (how much each weight affects the loss)&lt;br /&gt;
## Update weights using gradient descent&lt;br /&gt;
&lt;br /&gt;
=== 🧮 Mathematical Explanation ===&lt;br /&gt;
&lt;br /&gt;
Let:&lt;br /&gt;
* &amp;lt;math&amp;gt;L&amp;lt;/math&amp;gt; = Loss function&lt;br /&gt;
* &amp;lt;math&amp;gt;y&amp;lt;/math&amp;gt; = Actual output&lt;br /&gt;
* &amp;lt;math&amp;gt;\hat{y}&amp;lt;/math&amp;gt; = Predicted output&lt;br /&gt;
* &amp;lt;math&amp;gt;w&amp;lt;/math&amp;gt; = Weights&lt;br /&gt;
* &amp;lt;math&amp;gt;x&amp;lt;/math&amp;gt; = Inputs&lt;br /&gt;
&lt;br /&gt;
Loss:&lt;br /&gt;
:&amp;lt;math&amp;gt;L = \frac{1}{2}(y - \hat{y})^2&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Gradient of loss w.r.t. weights:&lt;br /&gt;
:&amp;lt;math&amp;gt;\frac{\partial L}{\partial w}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The weights are updated as:&lt;br /&gt;
:&amp;lt;math&amp;gt;w = w - \eta \cdot \frac{\partial L}{\partial w}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Where:&lt;br /&gt;
* &amp;lt;math&amp;gt;\eta&amp;lt;/math&amp;gt; = learning rate 🔧&lt;br /&gt;
&lt;br /&gt;
This update rule is applied to each layer using the chain rule from calculus.&lt;br /&gt;
&lt;br /&gt;
=== 📊 Example Workflow ===&lt;br /&gt;
&lt;br /&gt;
Let’s say we have:&lt;br /&gt;
* A network with one hidden layer&lt;br /&gt;
* Sigmoid activation&lt;br /&gt;
* Mean squared error loss&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Step&lt;br /&gt;
! Description&lt;br /&gt;
|-&lt;br /&gt;
| 1&lt;br /&gt;
| Do a forward pass to get predicted output &amp;lt;math&amp;gt;\hat{y}&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| 2&lt;br /&gt;
| Calculate the error &amp;lt;math&amp;gt;L = (y - \hat{y})^2&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| 3&lt;br /&gt;
| Compute the derivative of loss with respect to each weight&lt;br /&gt;
|-&lt;br /&gt;
| 4&lt;br /&gt;
| Update weights: &amp;lt;math&amp;gt;w = w - \eta \cdot \frac{\partial L}{\partial w}&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| 5&lt;br /&gt;
| Repeat this process for many epochs (passes over data)&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== 🔧 Backpropagation Uses ===&lt;br /&gt;
&lt;br /&gt;
* Deep learning (CNNs, RNNs, Transformers)&lt;br /&gt;
* Supervised learning tasks (image classification, NLP, etc.)&lt;br /&gt;
* Any task where you need to minimize a loss function&lt;br /&gt;
&lt;br /&gt;
=== 💡 Key Concepts ===&lt;br /&gt;
&lt;br /&gt;
* Chain Rule: Used to pass the gradient from the output layer back to the input layer&lt;br /&gt;
* Gradient Descent: Optimizer that uses gradients to minimize loss&lt;br /&gt;
* Learning Rate: Controls how big the weight updates are&lt;br /&gt;
&lt;br /&gt;
=== 🚫 Challenges ===&lt;br /&gt;
&lt;br /&gt;
* Can suffer from [[Vanishing Gradient Problem]]&lt;br /&gt;
* Can also face [[Exploding Gradient Problem]]&lt;br /&gt;
* Requires good weight initialization and choice of activation functions&lt;br /&gt;
&lt;br /&gt;
=== 📚 Summary Table ===&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Concept&lt;br /&gt;
! Meaning&lt;br /&gt;
|-&lt;br /&gt;
| Backpropagation&lt;br /&gt;
| Algorithm for updating weights based on error&lt;br /&gt;
|-&lt;br /&gt;
| Gradient&lt;br /&gt;
| Direction and size of weight adjustment&lt;br /&gt;
|-&lt;br /&gt;
| Chain Rule&lt;br /&gt;
| Math rule used to calculate gradients in multi-layer networks&lt;br /&gt;
|-&lt;br /&gt;
| Loss Function&lt;br /&gt;
| Measures how wrong the prediction is&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== 📎 See Also ===&lt;br /&gt;
&lt;br /&gt;
* [[Gradient Descent]]&lt;br /&gt;
* [[Loss Function]]&lt;br /&gt;
* [[Activation Functions]]&lt;br /&gt;
* [[Vanishing Gradient Problem]]&lt;br /&gt;
* [[Exploding Gradient Problem]]&lt;br /&gt;
* [[Neural Networks]]&lt;/div&gt;</summary>
		<author><name>Thakshashila</name></author>
	</entry>
</feed>