Skip to content

Commit 0035ea0

Browse files
commit
1 parent 97f2d67 commit 0035ea0

File tree

3 files changed

+72
-1
lines changed

3 files changed

+72
-1
lines changed

README.md

Lines changed: 72 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ relu_md.ipynb: Analysis and implementation of the ReLU activation function.
1313
********************************************************************************************************************************
1414

1515
## 1. XOR Problem with Backpropagation
16-
[Check the code in Jupyter Notebook]("C:\veccinternship_projects\XOR_Backpropagation_md.ipynb")
16+
[Check the code in Jupyter Notebook](XOR_Backpropagation_md.ipynb)
1717

1818
### XOR Problem and Neural Network Implementation
1919

@@ -46,4 +46,75 @@ The neural network used to solve the XOR problem typically has:
4646

4747
This neural network training process helps the model learn the XOR pattern through iterative updates, gradually reducing the error in predictions.
4848

49+
## XOR deeplearning
50+
[Check the code in Jupyter Notebook](XORdeeplearning_md.ipynb)
4951

52+
**Introduction**: Delves into deep learning techniques for solving the XOR problem.
53+
54+
**Deep Neural Network**:
55+
- A "deep" learning approach uses multiple layers in the neural network (hidden layers) to learn more complex patterns. Implements a deeper neural network, possibly with more hidden layers, compared to the previous notebook.
56+
- Uses advanced neural network architectures to improve performance.
57+
58+
**Training and Optimization**:
59+
- Explains the training process for deep neural networks, including optimization techniques like gradient descent.
60+
- May include regularization techniques to prevent overfitting.
61+
62+
**Architecture**:
63+
Input layer → Hidden layer(s) → Output layer.
64+
65+
**Activation Functions**:
66+
Used in neurons to introduce non-linearity, helping the model solve complex problems like XOR.
67+
68+
**Key Concepts**:
69+
* Deep Learning: Involves neural networks with multiple hidden layers.
70+
* Optimization: Techniques to adjust weights for minimizing the loss function.
71+
* Regularization: Methods to reduce overfitting and improve generalization.
72+
73+
## Sigmoid Neuron Model
74+
[Check the code in Jupyter Notebook](sigmoid_neuron.ipynb)
75+
76+
**What is a Sigmoid Neuron?**
77+
- A basic unit of neural networks where the activation function is sigmoid:
78+
<img src="images\sigmoid.png" alt="Sigmoid formula" title="Sigmoid formula pic">
79+
It squashes the input into a range between 0 and 1.
80+
81+
**Sigmoid Activation Function** : <img src="images\sigmoid.png" alt="Sigmoid formula" title="Sigmoid formula pic">
82+
- Explains how the sigmoid function outputs values between 0 and 1, making it suitable for binary classification.
83+
84+
**Implementation**:
85+
- Code implementation of a sigmoid neuron, including forward and backward propagation steps.
86+
- Calculates the gradient of the sigmoid function for use in backpropagation.
87+
88+
Why Use It?
89+
Useful for binary classification and probabilistic outputs.
90+
91+
**Limitations**:
92+
Vanishing gradient problem: Gradients become too small during backpropagation for large networks, slowing learning.
93+
94+
**Key Concepts**:
95+
Sigmoid Function: A smooth, differentiable function that outputs values between 0 and 1.
96+
97+
**Gradient**: The derivative of the sigmoid function, used for updating weights during training.
98+
99+
## ReLU Activation Function
100+
[Check the code in Jupyter Notebook](relu_md.ipynb)
101+
102+
**What is ReLU**?
103+
- Rectified Linear Unit (ReLU) is a popular activation function: f(x)=max(0,x)
104+
- RelU Activation function described the RelU function, explains how RelU introduces non-linearity into the model, allowing the network to learn complex patterns.
105+
106+
**Why Use ReLU**?
107+
- Computationally efficient.
108+
- Helps with the vanishing gradient problem by not saturating for positive values.
109+
110+
**Variants**:
111+
Leaky ReLU, Parametric ReLU, etc., are used to address issues like "dead neurons."
112+
113+
**Implementation**:
114+
- Code implementation of a ReLU neuron, including forward and backward propagation steps.
115+
- Discusses the advantages and drawbacks of ReLU, such as the issue of dying ReLUs.
116+
117+
**Key Concepts**:
118+
- ReLU Function: Outputs the input if positive, otherwise zero.
119+
- Non-Linearity: Introduced by ReLU, enabling the learning of complex patterns.
120+
- Dying ReLUs: A problem where neurons stop activating, often addressed with variants like Leaky ReLU.

images/relu.png

3.89 KB
Loading

images/sigmoid.png

2.99 KB
Loading

0 commit comments

Comments
 (0)