mirror of https://github.com/glouw/tinn
203 lines
5.4 KiB
C
203 lines
5.4 KiB
C
// gcc test.c Tinn.c -lm
|
|
//
|
|
// Tinn does not include functionality for loading
|
|
// and parsing data sets as all data sets are different.
|
|
//
|
|
// This example shows how to open an example data file from the machine learning archives.
|
|
// The training data consists of hand written digits and can be found at:
|
|
//
|
|
// http://archive.ics.uci.edu/ml/machine-learning-databases/semeion/semeion.data
|
|
//
|
|
// Each line is one digit. A digit consists of 256 pixels (16 x 16 display).
|
|
// Each line finishes with 10 digits indicating the digit:
|
|
//
|
|
// 0: 1 0 0 0 0 0 0 0 0 0
|
|
// 1: 0 1 0 0 0 0 0 0 0 0
|
|
// 2: 0 0 1 0 0 0 0 0 0 0
|
|
// 3: 0 0 0 1 0 0 0 0 0 0
|
|
// 4: 0 0 0 0 1 0 0 0 0 0
|
|
// ...
|
|
// 9: 0 0 0 0 0 0 0 0 0 1
|
|
//
|
|
// This gives 256 inputs and 10 outputs to the neural network.
|
|
|
|
#include "Tinn.h"
|
|
#include <stdio.h>
|
|
#include <time.h>
|
|
#include <string.h>
|
|
#include <stdlib.h>
|
|
|
|
typedef struct
|
|
{
|
|
float** in;
|
|
float** tg;
|
|
int nips;
|
|
int nops;
|
|
int rows;
|
|
}
|
|
Data;
|
|
|
|
static int lns(FILE* const file)
|
|
{
|
|
int ch = EOF;
|
|
int lines = 0;
|
|
int pc = '\n';
|
|
while((ch = getc(file)) != EOF)
|
|
{
|
|
if(ch == '\n')
|
|
lines++;
|
|
pc = ch;
|
|
}
|
|
if(pc != '\n')
|
|
lines++;
|
|
rewind(file);
|
|
return lines;
|
|
}
|
|
|
|
static char* readln(FILE* const file)
|
|
{
|
|
int ch = EOF;
|
|
int reads = 0;
|
|
int size = 128;
|
|
char* line = (char*) malloc((size) * sizeof(char));
|
|
while((ch = getc(file)) != '\n' && ch != EOF)
|
|
{
|
|
line[reads++] = ch;
|
|
if(reads + 1 == size)
|
|
line = (char*) realloc((line), (size *= 2) * sizeof(char));
|
|
}
|
|
line[reads] = '\0';
|
|
return line;
|
|
}
|
|
|
|
static float** new2d(const int rows, const int cols)
|
|
{
|
|
float** row = (float**) malloc((rows) * sizeof(float*));
|
|
for(int r = 0; r < rows; r++)
|
|
row[r] = (float*) malloc((cols) * sizeof(float));
|
|
return row;
|
|
}
|
|
|
|
static Data ndata(const int nips, const int nops, const int rows)
|
|
{
|
|
const Data data = {
|
|
new2d(rows, nips), new2d(rows, nops), nips, nops, rows
|
|
};
|
|
return data;
|
|
}
|
|
|
|
static void parse(const Data data, char* line, const int row)
|
|
{
|
|
const int cols = data.nips + data.nops;
|
|
for(int col = 0; col < cols; col++)
|
|
{
|
|
const float val = atof(strtok(col == 0 ? line : NULL, " "));
|
|
if(col < data.nips)
|
|
data.in[row][col] = val;
|
|
else
|
|
data.tg[row][col - data.nips] = val;
|
|
}
|
|
}
|
|
|
|
static void dfree(const Data d)
|
|
{
|
|
for(int row = 0; row < d.rows; row++)
|
|
{
|
|
free(d.in[row]);
|
|
free(d.tg[row]);
|
|
}
|
|
free(d.in);
|
|
free(d.tg);
|
|
}
|
|
|
|
static void shuffle(const Data d)
|
|
{
|
|
for(int a = 0; a < d.rows; a++)
|
|
{
|
|
const int b = rand() % d.rows;
|
|
float* ot = d.tg[a];
|
|
float* it = d.in[a];
|
|
// Swap output.
|
|
d.tg[a] = d.tg[b];
|
|
d.tg[b] = ot;
|
|
// Swap input.
|
|
d.in[a] = d.in[b];
|
|
d.in[b] = it;
|
|
}
|
|
}
|
|
|
|
static Data build(const char* path, const int nips, const int nops)
|
|
{
|
|
FILE* file = fopen(path, "r");
|
|
if(file == NULL)
|
|
{
|
|
printf("Could not open %s\n", path);
|
|
printf("Get it from the machine learning database: ");
|
|
printf("wget http://archive.ics.uci.edu/ml/machine-learning-databases/semeion/semeion.data\n");
|
|
exit(1);
|
|
}
|
|
const int rows = lns(file);
|
|
Data data = ndata(nips, nops, rows);
|
|
for(int row = 0; row < rows; row++)
|
|
{
|
|
char* line = readln(file);
|
|
parse(data, line, row);
|
|
free(line);
|
|
}
|
|
fclose(file);
|
|
return data;
|
|
}
|
|
|
|
int main()
|
|
{
|
|
// Tinn does not seed the random number generator.
|
|
srand(time(0));
|
|
// Input and output size is harded coded here as machine learning
|
|
// repositories usually don't include the input and output size in the data itself.
|
|
const int nips = 256;
|
|
const int nops = 10;
|
|
// Hyper Parameters.
|
|
// Learning rate is annealed and thus not constant.
|
|
// It can be fine tuned along with the number of hidden layers.
|
|
// Feel free to modify the anneal rate as well.
|
|
const int nhid = 28;
|
|
float rate = 1.0f;
|
|
const float anneal = 0.99f;
|
|
// Load the training set.
|
|
const Data data = build("semeion.data", nips, nops);
|
|
// Train, baby, train.
|
|
const Tinn tinn = xtbuild(nips, nhid, nops);
|
|
for(int i = 0; i < 100; i++)
|
|
{
|
|
shuffle(data);
|
|
float error = 0.0f;
|
|
for(int j = 0; j < data.rows; j++)
|
|
{
|
|
const float* const in = data.in[j];
|
|
const float* const tg = data.tg[j];
|
|
error += xttrain(tinn, in, tg, rate);
|
|
}
|
|
printf("error %.12f :: learning rate %f\n",
|
|
(double) error / data.rows,
|
|
(double) rate);
|
|
rate *= anneal;
|
|
}
|
|
// This is how you save the neural network to disk.
|
|
xtsave(tinn, "saved.tinn");
|
|
xtfree(tinn);
|
|
// This is how you load the neural network from disk.
|
|
const Tinn loaded = xtload("saved.tinn");
|
|
// Now we do a prediction with the neural network we loaded from disk.
|
|
// Ideally, we would also load a testing set to make the prediction with,
|
|
// but for the sake of brevity here we just reuse the training set from earlier.
|
|
const float* const in = data.in[0];
|
|
const float* const tg = data.tg[0];
|
|
const float* const pd = xtpredict(loaded, in);
|
|
for(int i = 0; i < data.nops; i++) { printf("%f ", (double) tg[i]); } printf("\n");
|
|
for(int i = 0; i < data.nops; i++) { printf("%f ", (double) pd[i]); } printf("\n");
|
|
// All done. Let's clean up.
|
|
xtfree(loaded);
|
|
dfree(data);
|
|
return 0;
|
|
}
|