mirror of
https://github.com/TheAlgorithms/Python.git
synced 2024-11-23 21:11:08 +00:00
Add 2-hidden layer neural network with back propagation using sigmoid activation function (#4037)
* added neural network with 2 hidden layers
* Revert "added neural network with 2 hidden layers"
This reverts commit fa4e2ac86e
.
* added neural network with 2 hidden layers
* passing pre-commit requirements
* doctest completed
* added return hints
* added example
* example added
* completed doctest's
* changes made as per the review
* changes made
* changes after review
* changes
* spacing
* changed return type
* changed dtype
This commit is contained in:
parent
207ac957ef
commit
64d8504170
295
neural_network/2_hidden_layers_neural_network.py
Normal file
295
neural_network/2_hidden_layers_neural_network.py
Normal file
|
@ -0,0 +1,295 @@
|
|||
"""
|
||||
References:
|
||||
- http://neuralnetworksanddeeplearning.com/chap2.html (Backpropagation)
|
||||
- https://en.wikipedia.org/wiki/Sigmoid_function (Sigmoid activation function)
|
||||
- https://en.wikipedia.org/wiki/Feedforward_neural_network (Feedforward)
|
||||
"""
|
||||
|
||||
import numpy
|
||||
|
||||
|
||||
class TwoHiddenLayerNeuralNetwork:
|
||||
def __init__(self, input_array: numpy.ndarray, output_array: numpy.ndarray) -> None:
|
||||
"""
|
||||
This function initializes the TwoHiddenLayerNeuralNetwork class with random
|
||||
weights for every layer and initializes predicted output with zeroes.
|
||||
|
||||
input_array : input values for training the neural network (i.e training data) .
|
||||
output_array : expected output values of the given inputs.
|
||||
"""
|
||||
|
||||
# Input values provided for training the model.
|
||||
self.input_array = input_array
|
||||
|
||||
# Random initial weights are assigned where first argument is the
|
||||
# number of nodes in previous layer and second argument is the
|
||||
# number of nodes in the next layer.
|
||||
|
||||
# Random initial weights are assigned.
|
||||
# self.input_array.shape[1] is used to represent number of nodes in input layer.
|
||||
# First hidden layer consists of 4 nodes.
|
||||
self.input_layer_and_first_hidden_layer_weights = numpy.random.rand(
|
||||
self.input_array.shape[1], 4
|
||||
)
|
||||
|
||||
# Random initial values for the first hidden layer.
|
||||
# First hidden layer has 4 nodes.
|
||||
# Second hidden layer has 3 nodes.
|
||||
self.first_hidden_layer_and_second_hidden_layer_weights = numpy.random.rand(
|
||||
4, 3
|
||||
)
|
||||
|
||||
# Random initial values for the second hidden layer.
|
||||
# Second hidden layer has 3 nodes.
|
||||
# Output layer has 1 node.
|
||||
self.second_hidden_layer_and_output_layer_weights = numpy.random.rand(3, 1)
|
||||
|
||||
# Real output values provided.
|
||||
self.output_array = output_array
|
||||
|
||||
# Predicted output values by the neural network.
|
||||
# Predicted_output array initially consists of zeroes.
|
||||
self.predicted_output = numpy.zeros(output_array.shape)
|
||||
|
||||
def feedforward(self) -> numpy.ndarray:
|
||||
"""
|
||||
The information moves in only one direction i.e. forward from the input nodes,
|
||||
through the two hidden nodes and to the output nodes.
|
||||
There are no cycles or loops in the network.
|
||||
|
||||
Return layer_between_second_hidden_layer_and_output
|
||||
(i.e the last layer of the neural network).
|
||||
|
||||
>>> input_val = numpy.array(([0, 0, 0], [0, 0, 0], [0, 0, 0]), dtype=float)
|
||||
>>> output_val = numpy.array(([0], [0], [0]), dtype=float)
|
||||
>>> nn = TwoHiddenLayerNeuralNetwork(input_val, output_val)
|
||||
>>> res = nn.feedforward()
|
||||
>>> array_sum = numpy.sum(res)
|
||||
>>> numpy.isnan(array_sum)
|
||||
False
|
||||
"""
|
||||
# Layer_between_input_and_first_hidden_layer is the layer connecting the
|
||||
# input nodes with the first hidden layer nodes.
|
||||
self.layer_between_input_and_first_hidden_layer = sigmoid(
|
||||
numpy.dot(self.input_array, self.input_layer_and_first_hidden_layer_weights)
|
||||
)
|
||||
|
||||
# layer_between_first_hidden_layer_and_second_hidden_layer is the layer
|
||||
# connecting the first hidden set of nodes with the second hidden set of nodes.
|
||||
self.layer_between_first_hidden_layer_and_second_hidden_layer = sigmoid(
|
||||
numpy.dot(
|
||||
self.layer_between_input_and_first_hidden_layer,
|
||||
self.first_hidden_layer_and_second_hidden_layer_weights,
|
||||
)
|
||||
)
|
||||
|
||||
# layer_between_second_hidden_layer_and_output is the layer connecting
|
||||
# second hidden layer with the output node.
|
||||
self.layer_between_second_hidden_layer_and_output = sigmoid(
|
||||
numpy.dot(
|
||||
self.layer_between_first_hidden_layer_and_second_hidden_layer,
|
||||
self.second_hidden_layer_and_output_layer_weights,
|
||||
)
|
||||
)
|
||||
|
||||
return self.layer_between_second_hidden_layer_and_output
|
||||
|
||||
def back_propagation(self) -> None:
|
||||
"""
|
||||
Function for fine-tuning the weights of the neural net based on the
|
||||
error rate obtained in the previous epoch (i.e., iteration).
|
||||
Updation is done using derivative of sogmoid activation function.
|
||||
|
||||
>>> input_val = numpy.array(([0, 0, 0], [0, 0, 0], [0, 0, 0]), dtype=float)
|
||||
>>> output_val = numpy.array(([0], [0], [0]), dtype=float)
|
||||
>>> nn = TwoHiddenLayerNeuralNetwork(input_val, output_val)
|
||||
>>> res = nn.feedforward()
|
||||
>>> nn.back_propagation()
|
||||
>>> updated_weights = nn.second_hidden_layer_and_output_layer_weights
|
||||
>>> (res == updated_weights).all()
|
||||
False
|
||||
"""
|
||||
|
||||
updated_second_hidden_layer_and_output_layer_weights = numpy.dot(
|
||||
self.layer_between_first_hidden_layer_and_second_hidden_layer.T,
|
||||
2
|
||||
* (self.output_array - self.predicted_output)
|
||||
* sigmoid_derivative(self.predicted_output),
|
||||
)
|
||||
updated_first_hidden_layer_and_second_hidden_layer_weights = numpy.dot(
|
||||
self.layer_between_input_and_first_hidden_layer.T,
|
||||
numpy.dot(
|
||||
2
|
||||
* (self.output_array - self.predicted_output)
|
||||
* sigmoid_derivative(self.predicted_output),
|
||||
self.second_hidden_layer_and_output_layer_weights.T,
|
||||
)
|
||||
* sigmoid_derivative(
|
||||
self.layer_between_first_hidden_layer_and_second_hidden_layer
|
||||
),
|
||||
)
|
||||
updated_input_layer_and_first_hidden_layer_weights = numpy.dot(
|
||||
self.input_array.T,
|
||||
numpy.dot(
|
||||
numpy.dot(
|
||||
2
|
||||
* (self.output_array - self.predicted_output)
|
||||
* sigmoid_derivative(self.predicted_output),
|
||||
self.second_hidden_layer_and_output_layer_weights.T,
|
||||
)
|
||||
* sigmoid_derivative(
|
||||
self.layer_between_first_hidden_layer_and_second_hidden_layer
|
||||
),
|
||||
self.first_hidden_layer_and_second_hidden_layer_weights.T,
|
||||
)
|
||||
* sigmoid_derivative(self.layer_between_input_and_first_hidden_layer),
|
||||
)
|
||||
|
||||
self.input_layer_and_first_hidden_layer_weights += (
|
||||
updated_input_layer_and_first_hidden_layer_weights
|
||||
)
|
||||
self.first_hidden_layer_and_second_hidden_layer_weights += (
|
||||
updated_first_hidden_layer_and_second_hidden_layer_weights
|
||||
)
|
||||
self.second_hidden_layer_and_output_layer_weights += (
|
||||
updated_second_hidden_layer_and_output_layer_weights
|
||||
)
|
||||
|
||||
def train(self, output: numpy.ndarray, iterations: int, give_loss: bool) -> None:
|
||||
"""
|
||||
Performs the feedforwarding and back propagation process for the
|
||||
given number of iterations.
|
||||
Every iteration will update the weights of neural network.
|
||||
|
||||
output : real output values,required for calculating loss.
|
||||
iterations : number of times the weights are to be updated.
|
||||
give_loss : boolean value, If True then prints loss for each iteration,
|
||||
If False then nothing is printed
|
||||
|
||||
>>> input_val = numpy.array(([0, 0, 0], [0, 1, 0], [0, 0, 1]), dtype=float)
|
||||
>>> output_val = numpy.array(([0], [1], [1]), dtype=float)
|
||||
>>> nn = TwoHiddenLayerNeuralNetwork(input_val, output_val)
|
||||
>>> first_iteration_weights = nn.feedforward()
|
||||
>>> nn.back_propagation()
|
||||
>>> updated_weights = nn.second_hidden_layer_and_output_layer_weights
|
||||
>>> (first_iteration_weights == updated_weights).all()
|
||||
False
|
||||
"""
|
||||
for iteration in range(1, iterations + 1):
|
||||
self.output = self.feedforward()
|
||||
self.back_propagation()
|
||||
if give_loss:
|
||||
loss = numpy.mean(numpy.square(output - self.feedforward()))
|
||||
print(f"Iteration {iteration} Loss: {loss}")
|
||||
|
||||
def predict(self, input: numpy.ndarray) -> int:
|
||||
"""
|
||||
Predict's the output for the given input values using
|
||||
the trained neural network.
|
||||
|
||||
The output value given by the model ranges in-between 0 and 1.
|
||||
The predict function returns 1 if the model value is greater
|
||||
than the threshold value else returns 0,
|
||||
as the real output values are in binary.
|
||||
|
||||
>>> input_val = numpy.array(([0, 0, 0], [0, 1, 0], [0, 0, 1]), dtype=float)
|
||||
>>> output_val = numpy.array(([0], [1], [1]), dtype=float)
|
||||
>>> nn = TwoHiddenLayerNeuralNetwork(input_val, output_val)
|
||||
>>> nn.train(output_val, 1000, False)
|
||||
>>> nn.predict([0,1,0])
|
||||
1
|
||||
"""
|
||||
|
||||
# Input values for which the predictions are to be made.
|
||||
self.array = input
|
||||
|
||||
self.layer_between_input_and_first_hidden_layer = sigmoid(
|
||||
numpy.dot(self.array, self.input_layer_and_first_hidden_layer_weights)
|
||||
)
|
||||
|
||||
self.layer_between_first_hidden_layer_and_second_hidden_layer = sigmoid(
|
||||
numpy.dot(
|
||||
self.layer_between_input_and_first_hidden_layer,
|
||||
self.first_hidden_layer_and_second_hidden_layer_weights,
|
||||
)
|
||||
)
|
||||
|
||||
self.layer_between_second_hidden_layer_and_output = sigmoid(
|
||||
numpy.dot(
|
||||
self.layer_between_first_hidden_layer_and_second_hidden_layer,
|
||||
self.second_hidden_layer_and_output_layer_weights,
|
||||
)
|
||||
)
|
||||
|
||||
return int(self.layer_between_second_hidden_layer_and_output > 0.6)
|
||||
|
||||
|
||||
def sigmoid(value: numpy.ndarray) -> numpy.ndarray:
|
||||
"""
|
||||
Applies sigmoid activation function.
|
||||
|
||||
return normalized values
|
||||
|
||||
>>> sigmoid(numpy.array(([1, 0, 2], [1, 0, 0]), dtype=numpy.float64))
|
||||
array([[0.73105858, 0.5 , 0.88079708],
|
||||
[0.73105858, 0.5 , 0.5 ]])
|
||||
"""
|
||||
return 1 / (1 + numpy.exp(-value))
|
||||
|
||||
|
||||
def sigmoid_derivative(value: numpy.ndarray) -> numpy.ndarray:
|
||||
"""
|
||||
Provides the derivative value of the sigmoid function.
|
||||
|
||||
returns derivative of the sigmoid value
|
||||
|
||||
>>> sigmoid_derivative(numpy.array(([1, 0, 2], [1, 0, 0]), dtype=numpy.float64))
|
||||
array([[ 0., 0., -2.],
|
||||
[ 0., 0., 0.]])
|
||||
"""
|
||||
return (value) * (1 - (value))
|
||||
|
||||
|
||||
def example() -> int:
|
||||
"""
|
||||
Example for "how to use the neural network class and use the
|
||||
respected methods for the desired output".
|
||||
Calls the TwoHiddenLayerNeuralNetwork class and
|
||||
provides the fixed input output values to the model.
|
||||
Model is trained for a fixed amount of iterations then the predict method is called.
|
||||
In this example the output is divided into 2 classes i.e. binary classification,
|
||||
the two classes are represented by '0' and '1'.
|
||||
|
||||
>>> example()
|
||||
1
|
||||
"""
|
||||
# Input values.
|
||||
input = numpy.array(
|
||||
(
|
||||
[0, 0, 0],
|
||||
[0, 0, 1],
|
||||
[0, 1, 0],
|
||||
[0, 1, 1],
|
||||
[1, 0, 0],
|
||||
[1, 0, 1],
|
||||
[1, 1, 0],
|
||||
[1, 1, 1],
|
||||
),
|
||||
dtype=numpy.float64,
|
||||
)
|
||||
|
||||
# True output values for the given input values.
|
||||
output = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]), dtype=numpy.float64)
|
||||
|
||||
# Calling neural network class.
|
||||
neural_network = TwoHiddenLayerNeuralNetwork(input_array=input, output_array=output)
|
||||
|
||||
# Calling training function.
|
||||
# Set give_loss to True if you want to see loss in every iteration.
|
||||
neural_network.train(output=output, iterations=10, give_loss=False)
|
||||
|
||||
return neural_network.predict(numpy.array(([1, 1, 1]), dtype=numpy.float64))
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
example()
|
Loading…
Reference in New Issue
Block a user