This repository has been archived on 2026-04-20. You can view files and clone it, but you cannot make any changes to it's state, such as pushing and creating new issues, pull requests or comments.
c-net/neuronal_network.c

351 lines
No EOL
12 KiB
C

#include <stdlib.h>
#include "neuronal_network.h"
#include <stdio.h>
#include <math.h>
double sigmoid(double input);
Matrix* predict(Neural_Network* network, Matrix* image_data);
double square(double input);
Matrix* sigmoid_derivative(Matrix* matrix);
Matrix* calculate_weights_delta(Matrix* previous_layer_output, Matrix* delta_matrix);
void apply_weights(Neural_Network* network, Matrix* delta_weights_matrix, int index);
Matrix* calculate_delta_hidden(Matrix* next_layer_delta, Matrix* weights, Matrix* current_layer_output);
Neural_Network* new_network(int input_size, int hidden_size, int hidden_amount, int output_size, double learning_rate){
Neural_Network* network = malloc(sizeof(Neural_Network));
network->input_size = input_size;
network->hidden_size = hidden_size;
network->hidden_amount = hidden_amount;
network->output_size = output_size;
network->learning_rate = learning_rate;
Matrix** weights = malloc(sizeof(Matrix*) * (hidden_amount + 1));
network->weights = weights;
network->weights[0] = matrix_create(hidden_size, input_size + 1);
for(int i=1;i<hidden_amount;i++){
network->weights[i] = matrix_create(hidden_size, hidden_size + 1);
}
network->weights[hidden_amount] = matrix_create(output_size, hidden_size + 1);
return network;
}
void randomize_network(Neural_Network* network, int scope){
for (int i = 0; i < network->hidden_amount + 1; i++) {
matrix_randomize(network->weights[i], scope);
}
}
void free_network(Neural_Network* network){
for (int i = 0; i < network->hidden_amount + 1; i++) {
matrix_free(network->weights[i]);
}
free(network->weights);
free(network);
}
void save_network(Neural_Network* network) {
char* file_name = "../networks/newest_network.txt";
// create file
FILE* save_file = fopen(file_name, "w");
// check if file is successfully opened
if(save_file == NULL) {
printf("ERROR: Something went wrong in file creation! (save_network)");
exit(1);
}
// save network size to first line of the file
fprintf(save_file, "%d\n", network->input_size);
fprintf(save_file, "%d\n", network->hidden_size);
fprintf(save_file, "%d\n", network->hidden_amount);
fprintf(save_file, "%d\n", network->output_size);
// close the file
fclose(save_file);
for (int i = 0; i < network->hidden_amount + 1; ++i) {
matrix_save(network->weights[i], file_name);
}
printf("Network Saved!");
}
Neural_Network* load_network(char* file) {
// create file pointer and open file
FILE* save_file = fopen(file, "r");
// check if file could be opened
if(save_file == NULL) {
printf("ERROR: File could not be opened/found! (load_network)");
exit(1);
}
// read & store the information on the size of the network from the save file
char buffer[MAX_BYTES];
fgets(buffer, MAX_BYTES, save_file);
int input_size = (int) strtol(buffer, NULL, 10);
fgets(buffer, MAX_BYTES, save_file);
int hidden_size = (int) strtol(buffer, NULL, 10);
fgets(buffer, MAX_BYTES, save_file);
int hidden_amount = (int) strtol(buffer, NULL, 10);
fgets(buffer, MAX_BYTES, save_file);
int output_size = (int) strtol(buffer, NULL, 10);
// create a new network to fill with the saved data
Neural_Network* saved_network = new_network(input_size, hidden_size, hidden_amount, output_size, 0);
for (int i = 0; i < saved_network->hidden_amount + 1; ++i) {
saved_network->weights[i] = load_next_matrix(save_file);
}
// return saved network
fclose(save_file);
return saved_network;
}
void print_network(Neural_Network* network) {
for (int i = 0; i < network->hidden_amount; ++i) {
matrix_print(network->weights[i]);
}
}
double measure_network_accuracy(Neural_Network* network, Image** images, int amount) {
int num_correct = 0;
for (int i = 0; i < amount; i++) {
Matrix* prediction = predict_image(network, images[i]);
if (matrix_argmax(prediction) == images[i]->label) {
num_correct++;
}
matrix_free(prediction);
}
return ((double) num_correct) / amount;
}
Matrix* predict_image(Neural_Network* network, Image* image){
Matrix* image_data = matrix_flatten(image->pixel_values, 0);
Matrix* res = predict(network, image_data);
matrix_free(image_data);
return res;
}
Matrix* predict(Neural_Network* network, Matrix* image_data) {
Matrix* input = matrix_add_bias(image_data);
Matrix* output[network->hidden_amount + 1];
for (int i = 0; i < network->hidden_amount + 1; ++i) {
Matrix* neuron_input = dot(network->weights[i], input);
Matrix* neuron_activation = apply(sigmoid, neuron_input);
output[i] = neuron_activation;
matrix_free(neuron_input);
matrix_free(input);
input = matrix_add_bias(neuron_activation);
}
for (int i = 0; i < network->hidden_amount; ++i) {
matrix_free(output[i]);
}
matrix_free(input);
return output[network->hidden_amount];
}
void batch_train(Neural_Network* network, Image** images, int amount, int batch_size) {
for (int i = 0; i < amount; ++i) {
Matrix* batch_weights[network->hidden_amount + 1];
for (int j = 0; j < batch_size; ++j) {
Matrix** delta_weights = train_network(network, images[i], images[i]->label);
for (int k = 0; k < network->hidden_amount + 1; k++) {
if(j == 0) {
batch_weights[k] = delta_weights[k];
continue;
}
Matrix* temp_result = add(batch_weights[k], delta_weights[k]);
matrix_free(batch_weights[k]);
matrix_free(delta_weights[k]);
batch_weights[k] = temp_result;
}
free(delta_weights);
}
for (int j = 0; j < network->hidden_amount + 1; ++j) {
Matrix* average_delta_weight = scale(batch_weights[j], (1.0 / batch_size));
apply_weights(network, average_delta_weight, j);
matrix_free(average_delta_weight);
matrix_free(batch_weights[j]);
}
}
}
Matrix ** train_network(Neural_Network* network, Image *image, int label) {
Matrix* image_data = matrix_flatten(image->pixel_values, 0);
Matrix* input = matrix_add_bias(image_data);
Matrix* output[network->hidden_amount + 1];
for (int i = 0; i < network->hidden_amount + 1; ++i) {
Matrix* neuron_input = dot(network->weights[i], input);
Matrix* neuron_activation = apply(sigmoid, neuron_input);
output[i] = neuron_activation;
matrix_free(neuron_input);
matrix_free(input);
input = matrix_add_bias(neuron_activation);
}
// back propagation
//list to store the new weights
Matrix** delta_weights = malloc(sizeof(Matrix*) * (network->hidden_amount + 1));
// calculate the derivative of the sigmoid function of the input of the result layer
Matrix* sigmoid_prime = sigmoid_derivative(output[network->hidden_amount]);
// create wanted out-put matrix, calculate the difference and delta values (output layer only)
Matrix* wanted_output = matrix_create(output[network->hidden_amount]->rows, output[network->hidden_amount]->columns);
matrix_fill(wanted_output, 0);
wanted_output->numbers[label][0] = 1;
Matrix* error = subtract(wanted_output, output[network->hidden_amount]);
Matrix* delta = multiply(sigmoid_prime, error);
//calculate and apply the delta for all weights in out-put layer
delta_weights[network->hidden_amount] = calculate_weights_delta(output[network->hidden_amount - 1], delta);
//hidden layers
Matrix* previous_delta = delta;
for (int i = network->hidden_amount; i > 1; i--) {
delta = calculate_delta_hidden(previous_delta, network->weights[i], output[i - 1]);
delta_weights[i - 1] = calculate_weights_delta(output[i - 2], delta);
matrix_free(previous_delta);
previous_delta = delta;
}
// Input Layer
delta = calculate_delta_hidden(previous_delta, network->weights[1], output[0]);
delta_weights[0] = calculate_weights_delta(image_data, delta);
for (int i = 0; i < network->hidden_amount + 1; ++i) {
apply_weights(network, delta_weights[i], i);
}
// De-allocate stuff
matrix_free(image_data);
matrix_free(input);
for (int i = 0; i < network->hidden_amount + 1; ++i) {
matrix_free(output[i]);
}
for (int i = 0; i < network->hidden_amount + 1; ++i) {
matrix_free(delta_weights[i]);
}
matrix_free(sigmoid_prime);
matrix_free(wanted_output);
matrix_free(error);
matrix_free(delta);
matrix_free(previous_delta);
return delta_weights;
}
Matrix* calculate_delta_hidden(Matrix* next_layer_delta, Matrix* weights, Matrix* current_layer_output) {
// remove bias weights from weights
Matrix* weights_without_biases = matrix_create(weights->rows, weights->columns - 1);
for (int i = 0; i < weights->rows; ++i) {
for (int j = 0; j < weights->columns - 1; ++j) {
weights_without_biases->numbers[i][j] = weights->numbers[i][j + 1];
}
}
// transpose the new weights and multiply with deltas
Matrix* transposed_weight_without_biases = transpose(weights_without_biases);
Matrix* sum_delta_weights = dot(transposed_weight_without_biases, next_layer_delta);
//multiply with derivative of current layer output
Matrix* sigmoid_prime = sigmoid_derivative(current_layer_output);
// multiply to find deltas for current layer
Matrix* new_deltas = multiply(sigmoid_prime, sum_delta_weights);
matrix_free(weights_without_biases);
matrix_free(transposed_weight_without_biases);
matrix_free(sum_delta_weights);
matrix_free(sigmoid_prime);
return new_deltas;
}
void apply_weights(Neural_Network* network, Matrix* delta_weights_matrix, int index) {
if(index > network->hidden_amount || index < 0) {
printf("ERROR: Index out of range! (apply_weights)");
exit(1);
}
if(delta_weights_matrix->rows != network->weights[index]->rows ||
delta_weights_matrix->columns != network->weights[index]->columns) {
printf("ERROR: Size of weight matrices do not match! (apply_weights)");
exit(1);
}
for (int i = 0; i < delta_weights_matrix->rows; i++) {
for (int j = 0; j < delta_weights_matrix->columns; j++) {
network->weights[index]->numbers[i][j] += delta_weights_matrix->numbers[i][j]; // multiply delta_weights_matrix with learning rate AND - instead of + because soll-ist
}
}
}
Matrix* calculate_weights_delta(Matrix* previous_layer_output, Matrix* delta_matrix) {
Matrix* previous_out_with_one = matrix_add_bias(previous_layer_output);
Matrix* transposed_previous_out_with_bias = transpose(previous_out_with_one);
Matrix* weights_delta_matrix = dot(delta_matrix, transposed_previous_out_with_bias);
matrix_free(previous_out_with_one);
matrix_free(transposed_previous_out_with_bias);
return weights_delta_matrix;
}
Matrix* sigmoid_derivative(Matrix* matrix) {
Matrix* ones = matrix_create(matrix->rows, matrix->columns);
matrix_fill(ones, 1);
Matrix* ones_minus_out = subtract(ones, matrix);
Matrix* sigmoid_derivative = multiply(matrix, ones_minus_out);
matrix_free(ones);
matrix_free(ones_minus_out);
return sigmoid_derivative;
}
double sigmoid(double input) {
return 1.0 / (1 + exp(-1 * input));
}
double square(double input) {
return input * input;
}