Skip to content

Commit 6b77991

Browse files
updated readme
1 parent 592079f commit 6b77991

File tree

1 file changed

+35
-1
lines changed

1 file changed

+35
-1
lines changed

README.md

Lines changed: 35 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
# Neural-Networks
2-
### **Implemented Convolutional Neural Network, Neural Network From Scratch.**
2+
### **Implemented Convolutional Neural Network, LSTM Neural Network and Neural Network From Scratch.**
33

44
## **Convolutional Neural Network**
55

@@ -52,6 +52,40 @@ They have applications in image and video recognition, recommender systems and n
5252

5353
* Learned about the implementation details of forward propagation and backward propagation algorithms used in CNN's and how different activation functions affect the training accuracy.
5454

55+
## **LSTM Neural Network**:
56+
57+
Long short-term memory (LSTM) units (or blocks) are a building unit for layers of a recurrent neural network (RNN). A RNN composed of LSTM units is often called an LSTM network. A common LSTM unit is composed of a cell, an input gate, an output gate and a forget gate. The cell is responsible for "remembering" values over arbitrary time intervals; hence the word "memory" in LSTM. Each of the three gates can be thought of as a "conventional" artificial neuron, as in a multi-layer (or feedforward) neural network: that is, they compute an activation (using an activation function) of a weighted sum. Intuitively, they can be thought as regulators of the flow of values that goes through the connections of the LSTM; hence the denotation "gate". There are connections between these gates and the cell.
58+
59+
The expression long short-term refers to the fact that LSTM is a model for the short-term memory which can last for a long period of time. An LSTM is well-suited to classify, process and predict time series given time lags of unknown size and duration between important events. LSTMs were developed to deal with the exploding and vanishing gradient problem when training traditional RNNs.
60+
61+
--> **Implemented LSTM Neural Network from SCRATCH!**
62+
* **Dataset** : US Baby's First Names
63+
* **Network Architecture** :
64+
![](https://github.com/navjindervirdee/neural-networks/blob/master/Recurrent%20Neural%20Network/LSTMForward.JPG?raw=true)
65+
66+
* **Actvation Function** :
67+
1. Sigmoid
68+
2. Softmax
69+
3. Tanh
70+
71+
* **Functions Implemented using Numpy** :
72+
1. Initialize Parameters
73+
2. Forward Propagation
74+
3. Cal Loss And Accuracy and Perplexity
75+
4. Backward Propagation
76+
5. Update Parameters
77+
6. Update Embeddings
78+
79+
* **Libraries Used** :
80+
1. Numpy
81+
2. Pandas
82+
3. Matplot
83+
84+
* **Results** :
85+
1. Perplexity : 1.05
86+
2. Accuracy : 85%
87+
88+
* Learned about the implementation details of forward propagation and backward propagation through time algorithms used in RNN's and how different activation functions affect the training accuracy.
5589

5690
## **Neural Network**:
5791

0 commit comments

Comments
 (0)