Artificial Intelligence
Understanding AI via neural networks
A frustration I had when looking at explanations of neural networks (NNs) was there is typically no explanation of what an artificial neuron does, but not what one actually is. How is a neuron instantiated?
Well, the logic and mathematical machinery of NNs applies to many and various instantiations. A neuron could be a piece of card bluetak'd to a whiteboard, and links to other neurons could be shown with pieces of string. We could write on each card the identities of the neurons that send data to it, and we could write on the weights for reach of those inputs. We could then manually update the inputs, maybe by writing on the whiteboard on the left, and move on to updating the weighted sums in the various layers, also using a whiteboard marker.
Another instantiation, maybe a more common one, is that the NN itself is an object, an occurrence of an NN class. The various neurons don't need to have a very defined existence, they can be represented simply by tables of weights, weighted sums and activations ( the outputs that are allowed to be only between 0 and 1).
I have piece of python code that makes an NN exactly like this.
import numpy as np
class NeuralNetwork:
def __init__(self, input_size, hidden_size, output_size):
# Initialize weights and biases randomly
hidden_layer = [0][0]
hidden_layer_activation = [0][0]
self.weights1 = np.random.randn(input_size, hidden_size)
self.bias1 = np.zeros(hidden_size)
self.weights2 = np.random.randn(hidden_size, output_size)
self.bias2 = np.zeros(output_size)
def sigmoid(self, x):
return 1 / (1 + np.exp(-x))
def sigmoid_derivative(self, x):
return self.sigmoid(x) * (1 - self.sigmoid(x))
def forward_propagation(self, X):
# Calculate hidden layer activations
hidden_layer = np.dot(X, self.weights1) + self.bias1
hidden_layer_activation = self.sigmoid(hidden_layer)
# Calculate output layer activations
output_layer = np.dot(hidden_layer_activation, self.weights2) + self.bias2
output_layer_activation = self.sigmoid(output_layer)
return output_layer_activation
def backward_propagation(self, X, y, output):
# Calculate output layer error
hidden_layer = [0][0]
output_error = output - y
output_delta = output_error * self.sigmoid_derivative(output)
# Calculate hidden layer error
hidden_layer_error = np.dot(output_delta, self.weights2.T)
hidden_layer_delta = hidden_layer_error * self.sigmoid_derivative(hidden_layer)
hidden_layer_activation = self.sigmoid(hidden_layer)
# Update weights and biases
self.weights2 -= np.dot(hidden_layer_activation.T, output_delta) * learning_rate
self.bias2 -= np.sum(output_delta, axis=0) * learning_rate
self.weights1 -= np.dot(X.T, hidden_layer_delta) * learning_rate
self.bias1 -= np.sum(hidden_layer_delta, axis=0) * learning_rate
def train(self, X, y, learning_rate, epochs):
for epoch in range(epochs):
output = self.forward_propagation(X)
self.backward_propagation(X, y,output)
error = np.mean(np.square(output - y))
print(f"Epoch {epoch+1}: Error = {error}")
# Example usage
input_size = 2
hidden_size = 4
output_size = 1
learning_rate = 0.1
epochs = 1000
# Create a neural network instance
nn = NeuralNetwork(input_size, hidden_size, output_size)
# Sample data
X = np.array([[0, 0], [0, 1], [1, 0], [1, 1]])
y = np.array([[0], [1], [1], [0]])
# Train the neural network
nn.train(X, y, learning_rate, epochs)
# Make predictions
new_input = np.array([[0.5, 0.7]])
prediction = nn.forward_propagation(new_input)
print("Prediction:", prediction)