mirror of
https://github.com/Relintai/pmlpp.git
synced 2025-04-03 02:12:44 +02:00
Use real_t instead of doubles.
This commit is contained in:
parent
7a291b451a
commit
4deb34c852
@ -30,7 +30,7 @@ g++ main.cpp /usr/local/lib/MLPP.so --std=c++17
|
|||||||
```
|
```
|
||||||
|
|
||||||
## Usage
|
## Usage
|
||||||
Please note that ML++ uses the ```std::vector<double>``` data type for emulating vectors, and the ```std::vector<std::vector<double>>``` data type for emulating matrices.
|
Please note that ML++ uses the ```std::vector<real_t>``` data type for emulating vectors, and the ```std::vector<std::vector<real_t>>``` data type for emulating matrices.
|
||||||
|
|
||||||
Begin by including the respective header file of your choice.
|
Begin by including the respective header file of your choice.
|
||||||
```cpp
|
```cpp
|
||||||
|
162
main.cpp
162
main.cpp
@ -52,19 +52,19 @@
|
|||||||
#include "MLPP/Transforms/Transforms.hpp"
|
#include "MLPP/Transforms/Transforms.hpp"
|
||||||
|
|
||||||
|
|
||||||
// double f(double x){
|
// real_t f(real_t x){
|
||||||
// return x*x*x + 2*x - 2;
|
// return x*x*x + 2*x - 2;
|
||||||
// }
|
// }
|
||||||
|
|
||||||
double f(double x){
|
real_t f(real_t x){
|
||||||
return sin(x);
|
return sin(x);
|
||||||
}
|
}
|
||||||
|
|
||||||
double f_prime(double x){
|
real_t f_prime(real_t x){
|
||||||
return 2 * x;
|
return 2 * x;
|
||||||
}
|
}
|
||||||
|
|
||||||
double f_prime_2var(std::vector<double> x){
|
real_t f_prime_2var(std::vector<real_t> x){
|
||||||
return 2 * x[0] + x[1];
|
return 2 * x[0] + x[1];
|
||||||
}
|
}
|
||||||
/*
|
/*
|
||||||
@ -74,7 +74,7 @@ double f_prime_2var(std::vector<double> x){
|
|||||||
y''(2) = 12
|
y''(2) = 12
|
||||||
*/
|
*/
|
||||||
|
|
||||||
// double f_mv(std::vector<double> x){
|
// real_t f_mv(std::vector<real_t> x){
|
||||||
// return x[0] * x[0] + x[0] * x[1] * x[1] + x[1] + 5;
|
// return x[0] * x[0] + x[0] * x[1] * x[1] + x[1] + 5;
|
||||||
// }
|
// }
|
||||||
|
|
||||||
@ -85,7 +85,7 @@ double f_prime_2var(std::vector<double> x){
|
|||||||
∂^2f/∂x∂y = 2
|
∂^2f/∂x∂y = 2
|
||||||
*/
|
*/
|
||||||
|
|
||||||
double f_mv(std::vector<double> x){
|
real_t f_mv(std::vector<real_t> x){
|
||||||
return x[0] * x[0] * x[0] + x[0] + x[1] * x[1] * x[1] * x[0] + x[2] * x[2] * x[1];
|
return x[0] * x[0] * x[0] + x[0] + x[1] * x[1] * x[1] * x[0] + x[2] * x[2] * x[1];
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -128,30 +128,30 @@ int main() {
|
|||||||
MLPPConvolutions conv;
|
MLPPConvolutions conv;
|
||||||
|
|
||||||
// DATA SETS
|
// DATA SETS
|
||||||
// std::vector<std::vector<double>> inputSet = {{1,2,3,4,5,6,7,8,9,10}, {3,5,9,12,15,18,21,24,27,30}};
|
// std::vector<std::vector<real_t>> inputSet = {{1,2,3,4,5,6,7,8,9,10}, {3,5,9,12,15,18,21,24,27,30}};
|
||||||
// std::vector<double> outputSet = {2,4,6,8,10,12,14,16,18,20};
|
// std::vector<real_t> outputSet = {2,4,6,8,10,12,14,16,18,20};
|
||||||
|
|
||||||
// std::vector<std::vector<double>> inputSet = {{1,2,3,4,5,6,7,8}, {0,0,0,0,1,1,1,1}};
|
// std::vector<std::vector<real_t>> inputSet = {{1,2,3,4,5,6,7,8}, {0,0,0,0,1,1,1,1}};
|
||||||
// std::vector<double> outputSet = {0,0,0,0,1,1,1,1};
|
// std::vector<real_t> outputSet = {0,0,0,0,1,1,1,1};
|
||||||
|
|
||||||
// std::vector<std::vector<double>> inputSet = {{4,3,0,-3,-4}, {0,0,0,1,1}};
|
// std::vector<std::vector<real_t>> inputSet = {{4,3,0,-3,-4}, {0,0,0,1,1}};
|
||||||
// std::vector<double> outputSet = {1,1,0,-1,-1};
|
// std::vector<real_t> outputSet = {1,1,0,-1,-1};
|
||||||
|
|
||||||
// std::vector<std::vector<double>> inputSet = {{0,1,2,3,4}};
|
// std::vector<std::vector<real_t>> inputSet = {{0,1,2,3,4}};
|
||||||
// std::vector<double> outputSet = {1,2,4,8,16};
|
// std::vector<real_t> outputSet = {1,2,4,8,16};
|
||||||
|
|
||||||
//std::vector<std::vector<double>> inputSet = {{32, 0, 7}, {2, 28, 17}, {0, 9, 23}};
|
//std::vector<std::vector<real_t>> inputSet = {{32, 0, 7}, {2, 28, 17}, {0, 9, 23}};
|
||||||
|
|
||||||
// std::vector<std::vector<double>> inputSet = {{1,1,0,0,1}, {0,0,1,1,1}, {0,1,1,0,1}};
|
// std::vector<std::vector<real_t>> inputSet = {{1,1,0,0,1}, {0,0,1,1,1}, {0,1,1,0,1}};
|
||||||
// std::vector<double> outputSet = {0,1,0,1,1};
|
// std::vector<real_t> outputSet = {0,1,0,1,1};
|
||||||
|
|
||||||
// std::vector<std::vector<double>> inputSet = {{0,0,1,1}, {0,1,0,1}};
|
// std::vector<std::vector<real_t>> inputSet = {{0,0,1,1}, {0,1,0,1}};
|
||||||
// std::vector<double> outputSet = {0,1,1,0};
|
// std::vector<real_t> outputSet = {0,1,1,0};
|
||||||
|
|
||||||
// // STATISTICS
|
// // STATISTICS
|
||||||
// std::vector<double> x = {1,2,3,4,5,6,7,8,9,10};
|
// std::vector<real_t> x = {1,2,3,4,5,6,7,8,9,10};
|
||||||
// std::vector<double> y = {10,9,8,7,6,5,4,3,2,1};
|
// std::vector<real_t> y = {10,9,8,7,6,5,4,3,2,1};
|
||||||
// std::vector<double> w = {0.1, 0.1, 0.1, 0.1, 0.1, 0.1, 0.1, 0.1, 0.1, 0.1};
|
// std::vector<real_t> w = {0.1, 0.1, 0.1, 0.1, 0.1, 0.1, 0.1, 0.1, 0.1, 0.1};
|
||||||
|
|
||||||
// std::cout << "Arithmetic Mean: " << stat.mean(x) << std::endl;
|
// std::cout << "Arithmetic Mean: " << stat.mean(x) << std::endl;
|
||||||
// std::cout << "Median: " << stat.median(x) << std::endl;
|
// std::cout << "Median: " << stat.median(x) << std::endl;
|
||||||
@ -184,16 +184,16 @@ int main() {
|
|||||||
// std::cout << "Absolute Average Deviation: " << stat.absAvgDeviation(x) << std::endl;
|
// std::cout << "Absolute Average Deviation: " << stat.absAvgDeviation(x) << std::endl;
|
||||||
|
|
||||||
// LINEAR ALGEBRA
|
// LINEAR ALGEBRA
|
||||||
// std::vector<std::vector<double>> square = {{1, 1}, {-1, 1}, {1, -1}, {-1, -1}};
|
// std::vector<std::vector<real_t>> square = {{1, 1}, {-1, 1}, {1, -1}, {-1, -1}};
|
||||||
|
|
||||||
// alg.printMatrix(alg.rotate(square, M_PI/4));
|
// alg.printMatrix(alg.rotate(square, M_PI/4));
|
||||||
|
|
||||||
// std::vector<std::vector<double>> A = {
|
// std::vector<std::vector<real_t>> A = {
|
||||||
// {1, 2, 3, 4, 5, 6, 7, 8, 9, 10},
|
// {1, 2, 3, 4, 5, 6, 7, 8, 9, 10},
|
||||||
// {1, 2, 3, 4, 5, 6, 7, 8, 9, 10},
|
// {1, 2, 3, 4, 5, 6, 7, 8, 9, 10},
|
||||||
// };
|
// };
|
||||||
// std::vector<double> a = {4, 3, 1, 3};
|
// std::vector<real_t> a = {4, 3, 1, 3};
|
||||||
// std::vector<double> b = {3, 5, 6, 1};
|
// std::vector<real_t> b = {3, 5, 6, 1};
|
||||||
|
|
||||||
// alg.printMatrix(alg.matmult(alg.transpose(A), A));
|
// alg.printMatrix(alg.matmult(alg.transpose(A), A));
|
||||||
// std::cout << std::endl;
|
// std::cout << std::endl;
|
||||||
@ -226,8 +226,8 @@ int main() {
|
|||||||
|
|
||||||
// const int TRIAL_NUM = 1000;
|
// const int TRIAL_NUM = 1000;
|
||||||
|
|
||||||
// double scoreSGD = 0;
|
// real_t scoreSGD = 0;
|
||||||
// double scoreADAM = 0;
|
// real_t scoreADAM = 0;
|
||||||
// for(int i = 0; i < TRIAL_NUM; i++){
|
// for(int i = 0; i < TRIAL_NUM; i++){
|
||||||
// LinReg model(alg.transpose(inputSet), outputSet);
|
// LinReg model(alg.transpose(inputSet), outputSet);
|
||||||
// model.MBGD(0.001, 5, 1, 0);
|
// model.MBGD(0.001, 5, 1, 0);
|
||||||
@ -269,8 +269,8 @@ int main() {
|
|||||||
// std::cout << "ACCURACY: " << 100 * model.score() << "%" << std::endl;
|
// std::cout << "ACCURACY: " << 100 * model.score() << "%" << std::endl;
|
||||||
|
|
||||||
// // PROBIT REGRESSION
|
// // PROBIT REGRESSION
|
||||||
// std::vector<std::vector<double>> inputSet;
|
// std::vector<std::vector<real_t>> inputSet;
|
||||||
// std::vector<double> outputSet;
|
// std::vector<real_t> outputSet;
|
||||||
// data.setData(30, "/Users/marcmelikyan/Desktop/Data/BreastCancer.csv", inputSet, outputSet);
|
// data.setData(30, "/Users/marcmelikyan/Desktop/Data/BreastCancer.csv", inputSet, outputSet);
|
||||||
// ProbitReg model(inputSet, outputSet);
|
// ProbitReg model(inputSet, outputSet);
|
||||||
// model.SGD(0.001, 10000, 1);
|
// model.SGD(0.001, 10000, 1);
|
||||||
@ -278,24 +278,24 @@ int main() {
|
|||||||
// std::cout << "ACCURACY: " << 100 * model.score() << "%" << std::endl;
|
// std::cout << "ACCURACY: " << 100 * model.score() << "%" << std::endl;
|
||||||
|
|
||||||
// // CLOGLOG REGRESSION
|
// // CLOGLOG REGRESSION
|
||||||
// std::vector<std::vector<double>> inputSet = {{1,2,3,4,5,6,7,8}, {0,0,0,0,1,1,1,1}};
|
// std::vector<std::vector<real_t>> inputSet = {{1,2,3,4,5,6,7,8}, {0,0,0,0,1,1,1,1}};
|
||||||
// std::vector<double> outputSet = {0,0,0,0,1,1,1,1};
|
// std::vector<real_t> outputSet = {0,0,0,0,1,1,1,1};
|
||||||
// CLogLogReg model(alg.transpose(inputSet), outputSet);
|
// CLogLogReg model(alg.transpose(inputSet), outputSet);
|
||||||
// model.SGD(0.1, 10000, 0);
|
// model.SGD(0.1, 10000, 0);
|
||||||
// alg.printVector(model.modelSetTest(alg.transpose(inputSet)));
|
// alg.printVector(model.modelSetTest(alg.transpose(inputSet)));
|
||||||
// std::cout << "ACCURACY: " << 100 * model.score() << "%" << std::endl;
|
// std::cout << "ACCURACY: " << 100 * model.score() << "%" << std::endl;
|
||||||
|
|
||||||
// // EXPREG REGRESSION
|
// // EXPREG REGRESSION
|
||||||
// std::vector<std::vector<double>> inputSet = {{0,1,2,3,4}};
|
// std::vector<std::vector<real_t>> inputSet = {{0,1,2,3,4}};
|
||||||
// std::vector<double> outputSet = {1,2,4,8,16};
|
// std::vector<real_t> outputSet = {1,2,4,8,16};
|
||||||
// ExpReg model(alg.transpose(inputSet), outputSet);
|
// ExpReg model(alg.transpose(inputSet), outputSet);
|
||||||
// model.SGD(0.001, 10000, 0);
|
// model.SGD(0.001, 10000, 0);
|
||||||
// alg.printVector(model.modelSetTest(alg.transpose(inputSet)));
|
// alg.printVector(model.modelSetTest(alg.transpose(inputSet)));
|
||||||
// std::cout << "ACCURACY: " << 100 * model.score() << "%" << std::endl;
|
// std::cout << "ACCURACY: " << 100 * model.score() << "%" << std::endl;
|
||||||
|
|
||||||
// // TANH REGRESSION
|
// // TANH REGRESSION
|
||||||
// std::vector<std::vector<double>> inputSet = {{4,3,0,-3,-4}, {0,0,0,1,1}};
|
// std::vector<std::vector<real_t>> inputSet = {{4,3,0,-3,-4}, {0,0,0,1,1}};
|
||||||
// std::vector<double> outputSet = {1,1,0,-1,-1};
|
// std::vector<real_t> outputSet = {1,1,0,-1,-1};
|
||||||
// TanhReg model(alg.transpose(inputSet), outputSet);
|
// TanhReg model(alg.transpose(inputSet), outputSet);
|
||||||
// model.SGD(0.1, 10000, 0);
|
// model.SGD(0.1, 10000, 0);
|
||||||
// alg.printVector(model.modelSetTest(alg.transpose(inputSet)));
|
// alg.printVector(model.modelSetTest(alg.transpose(inputSet)));
|
||||||
@ -320,9 +320,9 @@ int main() {
|
|||||||
// alg.printMatrix(model.modelSetTest(inputSet));
|
// alg.printMatrix(model.modelSetTest(inputSet));
|
||||||
|
|
||||||
// // MLP
|
// // MLP
|
||||||
// std::vector<std::vector<double>> inputSet = {{0,0,1,1}, {0,1,0,1}};
|
// std::vector<std::vector<real_t>> inputSet = {{0,0,1,1}, {0,1,0,1}};
|
||||||
// inputSet = alg.transpose(inputSet);
|
// inputSet = alg.transpose(inputSet);
|
||||||
// std::vector<double> outputSet = {0,1,1,0};
|
// std::vector<real_t> outputSet = {0,1,1,0};
|
||||||
|
|
||||||
// MLP model(inputSet, outputSet, 2);
|
// MLP model(inputSet, outputSet, 2);
|
||||||
// model.gradientDescent(0.1, 10000, 0);
|
// model.gradientDescent(0.1, 10000, 0);
|
||||||
@ -337,7 +337,7 @@ int main() {
|
|||||||
// std::cout << "ACCURACY: " << 100 * model.score() << "%" << std::endl;
|
// std::cout << "ACCURACY: " << 100 * model.score() << "%" << std::endl;
|
||||||
|
|
||||||
// // AUTOENCODER
|
// // AUTOENCODER
|
||||||
// std::vector<std::vector<double>> inputSet = {{1,2,3,4,5,6,7,8,9,10}, {3,5,9,12,15,18,21,24,27,30}};
|
// std::vector<std::vector<real_t>> inputSet = {{1,2,3,4,5,6,7,8,9,10}, {3,5,9,12,15,18,21,24,27,30}};
|
||||||
// AutoEncoder model(alg.transpose(inputSet), 5);
|
// AutoEncoder model(alg.transpose(inputSet), 5);
|
||||||
// model.SGD(0.001, 300000, 0);
|
// model.SGD(0.001, 300000, 0);
|
||||||
// alg.printMatrix(model.modelSetTest(alg.transpose(inputSet)));
|
// alg.printMatrix(model.modelSetTest(alg.transpose(inputSet)));
|
||||||
@ -347,8 +347,8 @@ int main() {
|
|||||||
// Possible Weight Init Methods: Default, Uniform, HeNormal, HeUniform, XavierNormal, XavierUniform
|
// Possible Weight Init Methods: Default, Uniform, HeNormal, HeUniform, XavierNormal, XavierUniform
|
||||||
// Possible Activations: Linear, Sigmoid, Swish, Softplus, Softsign, CLogLog, Ar{Sinh, Cosh, Tanh, Csch, Sech, Coth}, GaussianCDF, GELU, UnitStep
|
// Possible Activations: Linear, Sigmoid, Swish, Softplus, Softsign, CLogLog, Ar{Sinh, Cosh, Tanh, Csch, Sech, Coth}, GaussianCDF, GELU, UnitStep
|
||||||
// Possible Loss Functions: MSE, RMSE, MBE, LogLoss, CrossEntropy, HingeLoss
|
// Possible Loss Functions: MSE, RMSE, MBE, LogLoss, CrossEntropy, HingeLoss
|
||||||
// std::vector<std::vector<double>> inputSet = {{0,0,1,1}, {0,1,0,1}};
|
// std::vector<std::vector<real_t>> inputSet = {{0,0,1,1}, {0,1,0,1}};
|
||||||
// std::vector<double> outputSet = {0,1,1,0};
|
// std::vector<real_t> outputSet = {0,1,1,0};
|
||||||
// ANN ann(alg.transpose(inputSet), outputSet);
|
// ANN ann(alg.transpose(inputSet), outputSet);
|
||||||
// ann.addLayer(2, "Cosh");
|
// ann.addLayer(2, "Cosh");
|
||||||
// ann.addOutputLayer("Sigmoid", "LogLoss");
|
// ann.addOutputLayer("Sigmoid", "LogLoss");
|
||||||
@ -363,7 +363,7 @@ int main() {
|
|||||||
// alg.printVector(ann.modelSetTest(alg.transpose(inputSet)));
|
// alg.printVector(ann.modelSetTest(alg.transpose(inputSet)));
|
||||||
// std::cout << "ACCURACY: " << 100 * ann.score() << "%" << std::endl;
|
// std::cout << "ACCURACY: " << 100 * ann.score() << "%" << std::endl;
|
||||||
|
|
||||||
std::vector<std::vector<double>> outputSet = {{1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20},
|
std::vector<std::vector<real_t>> outputSet = {{1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20},
|
||||||
{2,4,6,8,10,12,14,16,18,20,22,24,26,28,30,32,34,36,38,40}};
|
{2,4,6,8,10,12,14,16,18,20,22,24,26,28,30,32,34,36,38,40}};
|
||||||
|
|
||||||
WGAN gan(2, alg.transpose(outputSet)); // our gan is a wasserstein gan (wgan)
|
WGAN gan(2, alg.transpose(outputSet)); // our gan is a wasserstein gan (wgan)
|
||||||
@ -376,8 +376,8 @@ int main() {
|
|||||||
alg.printMatrix(gan.generateExample(100));
|
alg.printMatrix(gan.generateExample(100));
|
||||||
|
|
||||||
|
|
||||||
// typedef std::vector<std::vector<double>> Matrix;
|
// typedef std::vector<std::vector<real_t>> Matrix;
|
||||||
// typedef std::vector<double> Vector;
|
// typedef std::vector<real_t> Vector;
|
||||||
|
|
||||||
// Matrix inputSet = {{0,0}, {0,1}, {1,0}, {1,1}}; // XOR
|
// Matrix inputSet = {{0,0}, {0,1}, {1,0}, {1,1}}; // XOR
|
||||||
// Vector outputSet = {0,1,1,0};
|
// Vector outputSet = {0,1,1,0};
|
||||||
@ -393,8 +393,8 @@ int main() {
|
|||||||
// std::cout << "ACCURACY: " << 100 * ann.score() << "%" << std::endl; // Accuracy.
|
// std::cout << "ACCURACY: " << 100 * ann.score() << "%" << std::endl; // Accuracy.
|
||||||
|
|
||||||
// // DYNAMICALLY SIZED MANN (Multidimensional Output ANN)
|
// // DYNAMICALLY SIZED MANN (Multidimensional Output ANN)
|
||||||
// std::vector<std::vector<double>> inputSet = {{1,2,3},{2,4,6},{3,6,9},{4,8,12}};
|
// std::vector<std::vector<real_t>> inputSet = {{1,2,3},{2,4,6},{3,6,9},{4,8,12}};
|
||||||
// std::vector<std::vector<double>> outputSet = {{1,5}, {2,10}, {3,15}, {4,20}};
|
// std::vector<std::vector<real_t>> outputSet = {{1,5}, {2,10}, {3,15}, {4,20}};
|
||||||
|
|
||||||
// MANN mann(inputSet, outputSet);
|
// MANN mann(inputSet, outputSet);
|
||||||
// mann.addOutputLayer("Linear", "MSE");
|
// mann.addOutputLayer("Linear", "MSE");
|
||||||
@ -402,14 +402,14 @@ int main() {
|
|||||||
// alg.printMatrix(mann.modelSetTest(inputSet));
|
// alg.printMatrix(mann.modelSetTest(inputSet));
|
||||||
// std::cout << "ACCURACY: " << 100 * mann.score() << "%" << std::endl;
|
// std::cout << "ACCURACY: " << 100 * mann.score() << "%" << std::endl;
|
||||||
|
|
||||||
// std::vector<std::vector<double>> inputSet;
|
// std::vector<std::vector<real_t>> inputSet;
|
||||||
// std::vector<double> tempOutputSet;
|
// std::vector<real_t> tempOutputSet;
|
||||||
// data.setData(4, "/Users/marcmelikyan/Desktop/Data/Iris.csv", inputSet, tempOutputSet);
|
// data.setData(4, "/Users/marcmelikyan/Desktop/Data/Iris.csv", inputSet, tempOutputSet);
|
||||||
// std::vector<std::vector<double>> outputSet = data.oneHotRep(tempOutputSet, 3);
|
// std::vector<std::vector<real_t>> outputSet = data.oneHotRep(tempOutputSet, 3);
|
||||||
|
|
||||||
// TRAIN TEST SPLIT CHECK
|
// TRAIN TEST SPLIT CHECK
|
||||||
// std::vector<std::vector<double>> inputSet1 = {{1,2,3,4,5,6,7,8,9,10}, {3,5,9,12,15,18,21,24,27,30}};
|
// std::vector<std::vector<real_t>> inputSet1 = {{1,2,3,4,5,6,7,8,9,10}, {3,5,9,12,15,18,21,24,27,30}};
|
||||||
// std::vector<std::vector<double>> outputSet1 = {{2,4,6,8,10,12,14,16,18,20}};
|
// std::vector<std::vector<real_t>> outputSet1 = {{2,4,6,8,10,12,14,16,18,20}};
|
||||||
// auto [inputSet, outputSet, inputTestSet, outputTestSet] = data.trainTestSplit(alg.transpose(inputSet1), alg.transpose(outputSet1), 0.2);
|
// auto [inputSet, outputSet, inputTestSet, outputTestSet] = data.trainTestSplit(alg.transpose(inputSet1), alg.transpose(outputSet1), 0.2);
|
||||||
// alg.printMatrix(inputSet);
|
// alg.printMatrix(inputSet);
|
||||||
// alg.printMatrix(outputSet);
|
// alg.printMatrix(outputSet);
|
||||||
@ -428,8 +428,8 @@ int main() {
|
|||||||
// std::cout << "ACCURACY: " << 100 * mann.score() << "%" << std::endl;
|
// std::cout << "ACCURACY: " << 100 * mann.score() << "%" << std::endl;
|
||||||
|
|
||||||
// // NAIVE BAYES
|
// // NAIVE BAYES
|
||||||
// std::vector<std::vector<double>> inputSet = {{1,1,1,1,1}, {0,0,1,1,1}, {0,0,1,0,1}};
|
// std::vector<std::vector<real_t>> inputSet = {{1,1,1,1,1}, {0,0,1,1,1}, {0,0,1,0,1}};
|
||||||
// std::vector<double> outputSet = {0,1,0,1,1};
|
// std::vector<real_t> outputSet = {0,1,0,1,1};
|
||||||
|
|
||||||
// MultinomialNB MNB(alg.transpose(inputSet), outputSet, 2);
|
// MultinomialNB MNB(alg.transpose(inputSet), outputSet, 2);
|
||||||
// alg.printVector(MNB.modelSetTest(alg.transpose(inputSet)));
|
// alg.printVector(MNB.modelSetTest(alg.transpose(inputSet)));
|
||||||
@ -441,7 +441,7 @@ int main() {
|
|||||||
// alg.printVector(GNB.modelSetTest(alg.transpose(inputSet)));
|
// alg.printVector(GNB.modelSetTest(alg.transpose(inputSet)));
|
||||||
|
|
||||||
// // KMeans
|
// // KMeans
|
||||||
// std::vector<std::vector<double>> inputSet = {{32, 0, 7}, {2, 28, 17}, {0, 9, 23}};
|
// std::vector<std::vector<real_t>> inputSet = {{32, 0, 7}, {2, 28, 17}, {0, 9, 23}};
|
||||||
// KMeans kmeans(inputSet, 3, "KMeans++");
|
// KMeans kmeans(inputSet, 3, "KMeans++");
|
||||||
// kmeans.train(3, 1);
|
// kmeans.train(3, 1);
|
||||||
// std::cout << std::endl;
|
// std::cout << std::endl;
|
||||||
@ -450,26 +450,26 @@ int main() {
|
|||||||
// alg.printVector(kmeans.silhouette_scores());
|
// alg.printVector(kmeans.silhouette_scores());
|
||||||
|
|
||||||
// // kNN
|
// // kNN
|
||||||
// std::vector<std::vector<double>> inputSet = {{1,2,3,4,5,6,7,8}, {0,0,0,0,1,1,1,1}};
|
// std::vector<std::vector<real_t>> inputSet = {{1,2,3,4,5,6,7,8}, {0,0,0,0,1,1,1,1}};
|
||||||
// std::vector<double> outputSet = {0,0,0,0,1,1,1,1};
|
// std::vector<real_t> outputSet = {0,0,0,0,1,1,1,1};
|
||||||
// kNN knn(alg.transpose(inputSet), outputSet, 8);
|
// kNN knn(alg.transpose(inputSet), outputSet, 8);
|
||||||
// alg.printVector(knn.modelSetTest(alg.transpose(inputSet)));
|
// alg.printVector(knn.modelSetTest(alg.transpose(inputSet)));
|
||||||
// std::cout << "ACCURACY: " << 100 * knn.score() << "%" << std::endl;
|
// std::cout << "ACCURACY: " << 100 * knn.score() << "%" << std::endl;
|
||||||
|
|
||||||
|
|
||||||
// // CONVOLUTION, POOLING, ETC..
|
// // CONVOLUTION, POOLING, ETC..
|
||||||
// std::vector<std::vector<double>> input = {
|
// std::vector<std::vector<real_t>> input = {
|
||||||
// {1},
|
// {1},
|
||||||
// };
|
// };
|
||||||
|
|
||||||
// std::vector<std::vector<std::vector<double>>> tensorSet;
|
// std::vector<std::vector<std::vector<real_t>>> tensorSet;
|
||||||
// tensorSet.push_back(input);
|
// tensorSet.push_back(input);
|
||||||
// tensorSet.push_back(input);
|
// tensorSet.push_back(input);
|
||||||
// tensorSet.push_back(input);
|
// tensorSet.push_back(input);
|
||||||
|
|
||||||
// alg.printTensor(data.rgb2xyz(tensorSet));
|
// alg.printTensor(data.rgb2xyz(tensorSet));
|
||||||
|
|
||||||
// std::vector<std::vector<double>> input = {
|
// std::vector<std::vector<real_t>> input = {
|
||||||
// {62,55,55,54,49,48,47,55},
|
// {62,55,55,54,49,48,47,55},
|
||||||
// {62,57,54,52,48,47,48,53},
|
// {62,57,54,52,48,47,48,53},
|
||||||
// {61,60,52,49,48,47,49,54},
|
// {61,60,52,49,48,47,49,54},
|
||||||
@ -487,17 +487,17 @@ int main() {
|
|||||||
// alg.printMatrix(conv.convolve(input, conv.getPrewittVertical(), 1)); // Can use padding
|
// alg.printMatrix(conv.convolve(input, conv.getPrewittVertical(), 1)); // Can use padding
|
||||||
// alg.printMatrix(conv.pool(input, 4, 4, "Max")); // Can use Max, Min, or Average pooling.
|
// alg.printMatrix(conv.pool(input, 4, 4, "Max")); // Can use Max, Min, or Average pooling.
|
||||||
|
|
||||||
// std::vector<std::vector<std::vector<double>>> tensorSet;
|
// std::vector<std::vector<std::vector<real_t>>> tensorSet;
|
||||||
// tensorSet.push_back(input);
|
// tensorSet.push_back(input);
|
||||||
// tensorSet.push_back(input);
|
// tensorSet.push_back(input);
|
||||||
// alg.printVector(conv.globalPool(tensorSet, "Average")); // Can use Max, Min, or Average global pooling.
|
// alg.printVector(conv.globalPool(tensorSet, "Average")); // Can use Max, Min, or Average global pooling.
|
||||||
|
|
||||||
// std::vector<std::vector<double>> laplacian = {{1, 1, 1}, {1, -4, 1}, {1, 1, 1}};
|
// std::vector<std::vector<real_t>> laplacian = {{1, 1, 1}, {1, -4, 1}, {1, 1, 1}};
|
||||||
// alg.printMatrix(conv.convolve(conv.gaussianFilter2D(5, 1), laplacian, 1));
|
// alg.printMatrix(conv.convolve(conv.gaussianFilter2D(5, 1), laplacian, 1));
|
||||||
|
|
||||||
|
|
||||||
// // PCA, SVD, eigenvalues & eigenvectors
|
// // PCA, SVD, eigenvalues & eigenvectors
|
||||||
// std::vector<std::vector<double>> inputSet = {{1,1}, {1,1}};
|
// std::vector<std::vector<real_t>> inputSet = {{1,1}, {1,1}};
|
||||||
// auto [Eigenvectors, Eigenvalues] = alg.eig(inputSet);
|
// auto [Eigenvectors, Eigenvalues] = alg.eig(inputSet);
|
||||||
// std::cout << "Eigenvectors:" << std::endl;
|
// std::cout << "Eigenvectors:" << std::endl;
|
||||||
// alg.printMatrix(Eigenvectors);
|
// alg.printMatrix(Eigenvectors);
|
||||||
@ -547,7 +547,7 @@ int main() {
|
|||||||
// std::cout << std::endl;
|
// std::cout << std::endl;
|
||||||
|
|
||||||
|
|
||||||
// std::vector<std::vector<double>> inputSet = {{1,2},{2,3},{3,4},{4,5},{5,6}};
|
// std::vector<std::vector<real_t>> inputSet = {{1,2},{2,3},{3,4},{4,5},{5,6}};
|
||||||
// std::cout << "Feature Scaling Example:" << std::endl;
|
// std::cout << "Feature Scaling Example:" << std::endl;
|
||||||
// alg.printMatrix(data.featureScaling(inputSet));
|
// alg.printMatrix(data.featureScaling(inputSet));
|
||||||
// std::cout << std::endl;
|
// std::cout << std::endl;
|
||||||
@ -561,20 +561,20 @@ int main() {
|
|||||||
// std::cout << std::endl;
|
// std::cout << std::endl;
|
||||||
|
|
||||||
// // Outlier Finder
|
// // Outlier Finder
|
||||||
// std::vector<double> inputSet = {1,2,3,4,5,6,7,8,9,23554332523523};
|
// std::vector<real_t> inputSet = {1,2,3,4,5,6,7,8,9,23554332523523};
|
||||||
// OutlierFinder outlierFinder(2); // Any datapoint outside of 2 stds from the mean is marked as an outlier.
|
// OutlierFinder outlierFinder(2); // Any datapoint outside of 2 stds from the mean is marked as an outlier.
|
||||||
// alg.printVector(outlierFinder.modelTest(inputSet));
|
// alg.printVector(outlierFinder.modelTest(inputSet));
|
||||||
|
|
||||||
// // Testing new Functions
|
// // Testing new Functions
|
||||||
// double z_s = 0.001;
|
// real_t z_s = 0.001;
|
||||||
// std::cout << avn.logit(z_s) << std::endl;
|
// std::cout << avn.logit(z_s) << std::endl;
|
||||||
// std::cout << avn.logit(z_s, 1) << std::endl;
|
// std::cout << avn.logit(z_s, 1) << std::endl;
|
||||||
|
|
||||||
// std::vector<double> z_v = {0.001};
|
// std::vector<real_t> z_v = {0.001};
|
||||||
// alg.printVector(avn.logit(z_v));
|
// alg.printVector(avn.logit(z_v));
|
||||||
// alg.printVector(avn.logit(z_v, 1));
|
// alg.printVector(avn.logit(z_v, 1));
|
||||||
|
|
||||||
// std::vector<std::vector<double>> Z_m = {{0.001}};
|
// std::vector<std::vector<real_t>> Z_m = {{0.001}};
|
||||||
// alg.printMatrix(avn.logit(Z_m));
|
// alg.printMatrix(avn.logit(Z_m));
|
||||||
// alg.printMatrix(avn.logit(Z_m, 1));
|
// alg.printMatrix(avn.logit(Z_m, 1));
|
||||||
|
|
||||||
@ -585,18 +585,18 @@ int main() {
|
|||||||
// alg.printMatrix(alg.matrixPower({{5,5},{5,5}}, 2));
|
// alg.printMatrix(alg.matrixPower({{5,5},{5,5}}, 2));
|
||||||
// alg.printVector(alg.solve({{1,1}, {1.5, 4.0}}, {2200, 5050}));
|
// alg.printVector(alg.solve({{1,1}, {1.5, 4.0}}, {2200, 5050}));
|
||||||
|
|
||||||
// std::vector<std::vector<double>> matrixOfCubes = {{1,2,64,27}};
|
// std::vector<std::vector<real_t>> matrixOfCubes = {{1,2,64,27}};
|
||||||
// std::vector<double> vectorOfCubes = {1,2,64,27};
|
// std::vector<real_t> vectorOfCubes = {1,2,64,27};
|
||||||
// alg.printMatrix(alg.cbrt(matrixOfCubes));
|
// alg.printMatrix(alg.cbrt(matrixOfCubes));
|
||||||
// alg.printVector(alg.cbrt(vectorOfCubes));
|
// alg.printVector(alg.cbrt(vectorOfCubes));
|
||||||
// std::cout << alg.max({{1,2,3,4,5}, {6,5,3,4,1}, {9,9,9,9,9}}) << std::endl;
|
// std::cout << alg.max({{1,2,3,4,5}, {6,5,3,4,1}, {9,9,9,9,9}}) << std::endl;
|
||||||
// std::cout << alg.min({{1,2,3,4,5}, {6,5,3,4,1}, {9,9,9,9,9}}) << std::endl;
|
// std::cout << alg.min({{1,2,3,4,5}, {6,5,3,4,1}, {9,9,9,9,9}}) << std::endl;
|
||||||
|
|
||||||
// std::vector<double> chicken;
|
// std::vector<real_t> chicken;
|
||||||
// data.getImage("../../Data/apple.jpeg", chicken);
|
// data.getImage("../../Data/apple.jpeg", chicken);
|
||||||
// alg.printVector(chicken);
|
// alg.printVector(chicken);
|
||||||
|
|
||||||
// std::vector<std::vector<double>> P = {{12, -51, 4}, {6, 167, -68}, {-4, 24, -41}};
|
// std::vector<std::vector<real_t>> P = {{12, -51, 4}, {6, 167, -68}, {-4, 24, -41}};
|
||||||
// alg.printMatrix(P);
|
// alg.printMatrix(P);
|
||||||
|
|
||||||
// alg.printMatrix(alg.gramSchmidtProcess(P));
|
// alg.printMatrix(alg.gramSchmidtProcess(P));
|
||||||
@ -608,7 +608,7 @@ int main() {
|
|||||||
// alg.printMatrix(R);
|
// alg.printMatrix(R);
|
||||||
|
|
||||||
// // Checking positive-definiteness checker. For Cholesky Decomp.
|
// // Checking positive-definiteness checker. For Cholesky Decomp.
|
||||||
// std::vector<std::vector<double>> A =
|
// std::vector<std::vector<real_t>> A =
|
||||||
// {
|
// {
|
||||||
// {1,-1,-1,-1},
|
// {1,-1,-1,-1},
|
||||||
// {-1,2,2,2},
|
// {-1,2,2,2},
|
||||||
@ -653,7 +653,7 @@ int main() {
|
|||||||
|
|
||||||
// std::cout << numAn.laplacian(f_mv, {1,1,1}) << std::endl;
|
// std::cout << numAn.laplacian(f_mv, {1,1,1}) << std::endl;
|
||||||
|
|
||||||
// std::vector<std::vector<std::vector<double>>> tensor;
|
// std::vector<std::vector<std::vector<real_t>>> tensor;
|
||||||
// tensor.push_back({{1,2}, {1,2}, {1,2}});
|
// tensor.push_back({{1,2}, {1,2}, {1,2}});
|
||||||
// tensor.push_back({{1,2}, {1,2}, {1,2}});
|
// tensor.push_back({{1,2}, {1,2}, {1,2}});
|
||||||
|
|
||||||
@ -672,7 +672,7 @@ int main() {
|
|||||||
|
|
||||||
// alg.printMatrix(conv.gradOrientation(A));
|
// alg.printMatrix(conv.gradOrientation(A));
|
||||||
|
|
||||||
// std::vector<std::vector<double>> A =
|
// std::vector<std::vector<real_t>> A =
|
||||||
// {
|
// {
|
||||||
// {1,0,0,0},
|
// {1,0,0,0},
|
||||||
// {0,0,0,0},
|
// {0,0,0,0},
|
||||||
@ -689,23 +689,23 @@ int main() {
|
|||||||
// std::cout << std::endl;
|
// std::cout << std::endl;
|
||||||
// } // Harris detector works. Life is good!
|
// } // Harris detector works. Life is good!
|
||||||
|
|
||||||
// std::vector<double> a = {3,4,4};
|
// std::vector<real_t> a = {3,4,4};
|
||||||
// std::vector<double> b = {4,4,4};
|
// std::vector<real_t> b = {4,4,4};
|
||||||
// alg.printVector(alg.cross(a,b));
|
// alg.printVector(alg.cross(a,b));
|
||||||
|
|
||||||
//SUPPORT VECTOR CLASSIFICATION (kernel method)
|
//SUPPORT VECTOR CLASSIFICATION (kernel method)
|
||||||
// std::vector<std::vector<double>> inputSet;
|
// std::vector<std::vector<real_t>> inputSet;
|
||||||
// std::vector<double> outputSet;
|
// std::vector<real_t> outputSet;
|
||||||
// data.setData(30, "/Users/marcmelikyan/Desktop/Data/BreastCancerSVM.csv", inputSet, outputSet);
|
// data.setData(30, "/Users/marcmelikyan/Desktop/Data/BreastCancerSVM.csv", inputSet, outputSet);
|
||||||
|
|
||||||
// std::vector<std::vector<double>> inputSet;
|
// std::vector<std::vector<real_t>> inputSet;
|
||||||
// std::vector<double> outputSet;
|
// std::vector<real_t> outputSet;
|
||||||
// data.setData(4, "/Users/marcmelikyan/Desktop/Data/IrisSVM.csv", inputSet, outputSet);
|
// data.setData(4, "/Users/marcmelikyan/Desktop/Data/IrisSVM.csv", inputSet, outputSet);
|
||||||
|
|
||||||
// DualSVC kernelSVM(inputSet, outputSet, 1000);
|
// DualSVC kernelSVM(inputSet, outputSet, 1000);
|
||||||
// kernelSVM.gradientDescent(0.0001, 20, 1);
|
// kernelSVM.gradientDescent(0.0001, 20, 1);
|
||||||
|
|
||||||
// std::vector<std::vector<double>> linearlyIndependentMat =
|
// std::vector<std::vector<real_t>> linearlyIndependentMat =
|
||||||
|
|
||||||
// {
|
// {
|
||||||
// {1,2,3,4},
|
// {1,2,3,4},
|
||||||
|
@ -10,14 +10,14 @@
|
|||||||
#include <cmath>
|
#include <cmath>
|
||||||
#include <iostream>
|
#include <iostream>
|
||||||
|
|
||||||
double MLPPActivation::linear(double z, bool deriv) {
|
real_t MLPPActivation::linear(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
return z;
|
return z;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::linear(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::linear(std::vector<real_t> z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.onevec(z.size());
|
return alg.onevec(z.size());
|
||||||
@ -25,7 +25,7 @@ std::vector<double> MLPPActivation::linear(std::vector<double> z, bool deriv) {
|
|||||||
return z;
|
return z;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::linear(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::linear(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.onemat(z.size(), z[0].size());
|
return alg.onemat(z.size(), z[0].size());
|
||||||
@ -33,14 +33,14 @@ std::vector<std::vector<double>> MLPPActivation::linear(std::vector<std::vector<
|
|||||||
return z;
|
return z;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::sigmoid(double z, bool deriv) {
|
real_t MLPPActivation::sigmoid(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return sigmoid(z) * (1 - sigmoid(z));
|
return sigmoid(z) * (1 - sigmoid(z));
|
||||||
}
|
}
|
||||||
return 1 / (1 + exp(-z));
|
return 1 / (1 + exp(-z));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::sigmoid(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::sigmoid(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.subtraction(sigmoid(z), alg.hadamard_product(sigmoid(z), sigmoid(z)));
|
return alg.subtraction(sigmoid(z), alg.hadamard_product(sigmoid(z), sigmoid(z)));
|
||||||
@ -48,7 +48,7 @@ std::vector<double> MLPPActivation::sigmoid(std::vector<double> z, bool deriv) {
|
|||||||
return alg.elementWiseDivision(alg.onevec(z.size()), alg.addition(alg.onevec(z.size()), alg.exp(alg.scalarMultiply(-1, z))));
|
return alg.elementWiseDivision(alg.onevec(z.size()), alg.addition(alg.onevec(z.size()), alg.exp(alg.scalarMultiply(-1, z))));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::sigmoid(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::sigmoid(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.subtraction(sigmoid(z), alg.hadamard_product(sigmoid(z), sigmoid(z)));
|
return alg.subtraction(sigmoid(z), alg.hadamard_product(sigmoid(z), sigmoid(z)));
|
||||||
@ -56,12 +56,12 @@ std::vector<std::vector<double>> MLPPActivation::sigmoid(std::vector<std::vector
|
|||||||
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.addition(alg.onemat(z.size(), z[0].size()), alg.exp(alg.scalarMultiply(-1, z))));
|
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.addition(alg.onemat(z.size(), z[0].size()), alg.exp(alg.scalarMultiply(-1, z))));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::softmax(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::softmax(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<double> a;
|
std::vector<real_t> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
std::vector<double> expZ = alg.exp(z);
|
std::vector<real_t> expZ = alg.exp(z);
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
|
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
sum += expZ[i];
|
sum += expZ[i];
|
||||||
@ -72,9 +72,9 @@ std::vector<double> MLPPActivation::softmax(std::vector<double> z, bool deriv) {
|
|||||||
return a;
|
return a;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::softmax(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::softmax(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<double>> a;
|
std::vector<std::vector<real_t>> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
|
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
@ -83,18 +83,18 @@ std::vector<std::vector<double>> MLPPActivation::softmax(std::vector<std::vector
|
|||||||
return a;
|
return a;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::adjSoftmax(std::vector<double> z) {
|
std::vector<real_t> MLPPActivation::adjSoftmax(std::vector<real_t> z) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<double> a;
|
std::vector<real_t> a;
|
||||||
double C = -*std::max_element(z.begin(), z.end());
|
real_t C = -*std::max_element(z.begin(), z.end());
|
||||||
z = alg.scalarAdd(C, z);
|
z = alg.scalarAdd(C, z);
|
||||||
|
|
||||||
return softmax(z);
|
return softmax(z);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::adjSoftmax(std::vector<std::vector<double>> z) {
|
std::vector<std::vector<real_t>> MLPPActivation::adjSoftmax(std::vector<std::vector<real_t>> z) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<double>> a;
|
std::vector<std::vector<real_t>> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
|
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
@ -103,10 +103,10 @@ std::vector<std::vector<double>> MLPPActivation::adjSoftmax(std::vector<std::vec
|
|||||||
return a;
|
return a;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::softmaxDeriv(std::vector<double> z) {
|
std::vector<std::vector<real_t>> MLPPActivation::softmaxDeriv(std::vector<real_t> z) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<double>> deriv;
|
std::vector<std::vector<real_t>> deriv;
|
||||||
std::vector<double> a = softmax(z);
|
std::vector<real_t> a = softmax(z);
|
||||||
deriv.resize(a.size());
|
deriv.resize(a.size());
|
||||||
for (int i = 0; i < deriv.size(); i++) {
|
for (int i = 0; i < deriv.size(); i++) {
|
||||||
deriv[i].resize(a.size());
|
deriv[i].resize(a.size());
|
||||||
@ -123,10 +123,10 @@ std::vector<std::vector<double>> MLPPActivation::softmaxDeriv(std::vector<double
|
|||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> MLPPActivation::softmaxDeriv(std::vector<std::vector<double>> z) {
|
std::vector<std::vector<std::vector<real_t>>> MLPPActivation::softmaxDeriv(std::vector<std::vector<real_t>> z) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<std::vector<double>>> deriv;
|
std::vector<std::vector<std::vector<real_t>>> deriv;
|
||||||
std::vector<std::vector<double>> a = softmax(z);
|
std::vector<std::vector<real_t>> a = softmax(z);
|
||||||
|
|
||||||
deriv.resize(a.size());
|
deriv.resize(a.size());
|
||||||
for (int i = 0; i < deriv.size(); i++) {
|
for (int i = 0; i < deriv.size(); i++) {
|
||||||
@ -144,14 +144,14 @@ std::vector<std::vector<std::vector<double>>> MLPPActivation::softmaxDeriv(std::
|
|||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::softplus(double z, bool deriv) {
|
real_t MLPPActivation::softplus(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return sigmoid(z);
|
return sigmoid(z);
|
||||||
}
|
}
|
||||||
return std::log(1 + exp(z));
|
return std::log(1 + exp(z));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::softplus(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::softplus(std::vector<real_t> z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return sigmoid(z);
|
return sigmoid(z);
|
||||||
}
|
}
|
||||||
@ -159,7 +159,7 @@ std::vector<double> MLPPActivation::softplus(std::vector<double> z, bool deriv)
|
|||||||
return alg.log(alg.addition(alg.onevec(z.size()), alg.exp(z)));
|
return alg.log(alg.addition(alg.onevec(z.size()), alg.exp(z)));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::softplus(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::softplus(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return sigmoid(z);
|
return sigmoid(z);
|
||||||
}
|
}
|
||||||
@ -167,14 +167,14 @@ std::vector<std::vector<double>> MLPPActivation::softplus(std::vector<std::vecto
|
|||||||
return alg.log(alg.addition(alg.onemat(z.size(), z[0].size()), alg.exp(z)));
|
return alg.log(alg.addition(alg.onemat(z.size(), z[0].size()), alg.exp(z)));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::softsign(double z, bool deriv) {
|
real_t MLPPActivation::softsign(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return 1 / ((1 + abs(z)) * (1 + abs(z)));
|
return 1 / ((1 + abs(z)) * (1 + abs(z)));
|
||||||
}
|
}
|
||||||
return z / (1 + abs(z));
|
return z / (1 + abs(z));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::softsign(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::softsign(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.elementWiseDivision(alg.onevec(z.size()), alg.exponentiate(alg.addition(alg.onevec(z.size()), alg.abs(z)), 2));
|
return alg.elementWiseDivision(alg.onevec(z.size()), alg.exponentiate(alg.addition(alg.onevec(z.size()), alg.abs(z)), 2));
|
||||||
@ -182,7 +182,7 @@ std::vector<double> MLPPActivation::softsign(std::vector<double> z, bool deriv)
|
|||||||
return alg.elementWiseDivision(z, alg.addition(alg.onevec(z.size()), alg.abs(z)));
|
return alg.elementWiseDivision(z, alg.addition(alg.onevec(z.size()), alg.abs(z)));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::softsign(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::softsign(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.exponentiate(alg.addition(alg.onemat(z.size(), z[0].size()), alg.abs(z)), 2));
|
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.exponentiate(alg.addition(alg.onemat(z.size(), z[0].size()), alg.abs(z)), 2));
|
||||||
@ -190,14 +190,14 @@ std::vector<std::vector<double>> MLPPActivation::softsign(std::vector<std::vecto
|
|||||||
return alg.elementWiseDivision(z, alg.addition(alg.onemat(z.size(), z[0].size()), alg.abs(z)));
|
return alg.elementWiseDivision(z, alg.addition(alg.onemat(z.size(), z[0].size()), alg.abs(z)));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::gaussianCDF(double z, bool deriv) {
|
real_t MLPPActivation::gaussianCDF(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return (1 / sqrt(2 * M_PI)) * exp(-z * z / 2);
|
return (1 / sqrt(2 * M_PI)) * exp(-z * z / 2);
|
||||||
}
|
}
|
||||||
return 0.5 * (1 + erf(z / sqrt(2)));
|
return 0.5 * (1 + erf(z / sqrt(2)));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::gaussianCDF(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::gaussianCDF(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.scalarMultiply(1 / sqrt(2 * M_PI), alg.exp(alg.scalarMultiply(-1 / 2, alg.hadamard_product(z, z))));
|
return alg.scalarMultiply(1 / sqrt(2 * M_PI), alg.exp(alg.scalarMultiply(-1 / 2, alg.hadamard_product(z, z))));
|
||||||
@ -205,7 +205,7 @@ std::vector<double> MLPPActivation::gaussianCDF(std::vector<double> z, bool deri
|
|||||||
return alg.scalarMultiply(0.5, alg.addition(alg.onevec(z.size()), alg.erf(alg.scalarMultiply(1 / sqrt(2), z))));
|
return alg.scalarMultiply(0.5, alg.addition(alg.onevec(z.size()), alg.erf(alg.scalarMultiply(1 / sqrt(2), z))));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::gaussianCDF(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::gaussianCDF(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.scalarMultiply(1 / sqrt(2 * M_PI), alg.exp(alg.scalarMultiply(-1 / 2, alg.hadamard_product(z, z))));
|
return alg.scalarMultiply(1 / sqrt(2 * M_PI), alg.exp(alg.scalarMultiply(-1 / 2, alg.hadamard_product(z, z))));
|
||||||
@ -213,14 +213,14 @@ std::vector<std::vector<double>> MLPPActivation::gaussianCDF(std::vector<std::ve
|
|||||||
return alg.scalarMultiply(0.5, alg.addition(alg.onemat(z.size(), z[0].size()), alg.erf(alg.scalarMultiply(1 / sqrt(2), z))));
|
return alg.scalarMultiply(0.5, alg.addition(alg.onemat(z.size(), z[0].size()), alg.erf(alg.scalarMultiply(1 / sqrt(2), z))));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::cloglog(double z, bool deriv) {
|
real_t MLPPActivation::cloglog(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return exp(z - exp(z));
|
return exp(z - exp(z));
|
||||||
}
|
}
|
||||||
return 1 - exp(-exp(z));
|
return 1 - exp(-exp(z));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::cloglog(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::cloglog(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.exp(alg.scalarMultiply(-1, alg.exp(z)));
|
return alg.exp(alg.scalarMultiply(-1, alg.exp(z)));
|
||||||
@ -228,7 +228,7 @@ std::vector<double> MLPPActivation::cloglog(std::vector<double> z, bool deriv) {
|
|||||||
return alg.scalarMultiply(-1, alg.scalarAdd(-1, alg.exp(alg.scalarMultiply(-1, alg.exp(z)))));
|
return alg.scalarMultiply(-1, alg.scalarAdd(-1, alg.exp(alg.scalarMultiply(-1, alg.exp(z)))));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::cloglog(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::cloglog(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.exp(alg.scalarMultiply(-1, alg.exp(z)));
|
return alg.exp(alg.scalarMultiply(-1, alg.exp(z)));
|
||||||
@ -236,14 +236,14 @@ std::vector<std::vector<double>> MLPPActivation::cloglog(std::vector<std::vector
|
|||||||
return alg.scalarMultiply(-1, alg.scalarAdd(-1, alg.exp(alg.scalarMultiply(-1, alg.exp(z)))));
|
return alg.scalarMultiply(-1, alg.scalarAdd(-1, alg.exp(alg.scalarMultiply(-1, alg.exp(z)))));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::logit(double z, bool deriv) {
|
real_t MLPPActivation::logit(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return 1 / z - 1 / (z - 1);
|
return 1 / z - 1 / (z - 1);
|
||||||
}
|
}
|
||||||
return std::log(z / (1 - z));
|
return std::log(z / (1 - z));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::logit(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::logit(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.subtraction(alg.elementWiseDivision(alg.onevec(z.size()), z), alg.elementWiseDivision(alg.onevec(z.size()), alg.subtraction(z, alg.onevec(z.size()))));
|
return alg.subtraction(alg.elementWiseDivision(alg.onevec(z.size()), z), alg.elementWiseDivision(alg.onevec(z.size()), alg.subtraction(z, alg.onevec(z.size()))));
|
||||||
@ -251,7 +251,7 @@ std::vector<double> MLPPActivation::logit(std::vector<double> z, bool deriv) {
|
|||||||
return alg.log(alg.elementWiseDivision(z, alg.subtraction(alg.onevec(z.size()), z)));
|
return alg.log(alg.elementWiseDivision(z, alg.subtraction(alg.onevec(z.size()), z)));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::logit(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::logit(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.subtraction(alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), z), alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.subtraction(z, alg.onemat(z.size(), z[0].size()))));
|
return alg.subtraction(alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), z), alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.subtraction(z, alg.onemat(z.size(), z[0].size()))));
|
||||||
@ -259,23 +259,23 @@ std::vector<std::vector<double>> MLPPActivation::logit(std::vector<std::vector<d
|
|||||||
return alg.log(alg.elementWiseDivision(z, alg.subtraction(alg.onemat(z.size(), z[0].size()), z)));
|
return alg.log(alg.elementWiseDivision(z, alg.subtraction(alg.onemat(z.size(), z[0].size()), z)));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::unitStep(double z, bool deriv) {
|
real_t MLPPActivation::unitStep(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
return z < 0 ? 0 : 1;
|
return z < 0 ? 0 : 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::unitStep(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::unitStep(std::vector<real_t> z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
std::vector<double> deriv;
|
std::vector<real_t> deriv;
|
||||||
deriv.resize(z.size());
|
deriv.resize(z.size());
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
deriv[i] = unitStep(z[i], 1);
|
deriv[i] = unitStep(z[i], 1);
|
||||||
}
|
}
|
||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
std::vector<double> a;
|
std::vector<real_t> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
|
|
||||||
for (int i = 0; i < a.size(); i++) {
|
for (int i = 0; i < a.size(); i++) {
|
||||||
@ -284,16 +284,16 @@ std::vector<double> MLPPActivation::unitStep(std::vector<double> z, bool deriv)
|
|||||||
return a;
|
return a;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::unitStep(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::unitStep(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
std::vector<std::vector<double>> deriv;
|
std::vector<std::vector<real_t>> deriv;
|
||||||
deriv.resize(z.size());
|
deriv.resize(z.size());
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
deriv[i] = unitStep(z[i], 1);
|
deriv[i] = unitStep(z[i], 1);
|
||||||
}
|
}
|
||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
std::vector<std::vector<double>> a;
|
std::vector<std::vector<real_t>> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
|
|
||||||
for (int i = 0; i < a.size(); i++) {
|
for (int i = 0; i < a.size(); i++) {
|
||||||
@ -302,14 +302,14 @@ std::vector<std::vector<double>> MLPPActivation::unitStep(std::vector<std::vecto
|
|||||||
return a;
|
return a;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::swish(double z, bool deriv) {
|
real_t MLPPActivation::swish(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return swish(z) + sigmoid(z) * (1 - swish(z));
|
return swish(z) + sigmoid(z) * (1 - swish(z));
|
||||||
}
|
}
|
||||||
return z * sigmoid(z);
|
return z * sigmoid(z);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::swish(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::swish(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
alg.addition(swish(z), alg.subtraction(sigmoid(z), alg.hadamard_product(sigmoid(z), swish(z))));
|
alg.addition(swish(z), alg.subtraction(sigmoid(z), alg.hadamard_product(sigmoid(z), swish(z))));
|
||||||
@ -317,7 +317,7 @@ std::vector<double> MLPPActivation::swish(std::vector<double> z, bool deriv) {
|
|||||||
return alg.hadamard_product(z, sigmoid(z));
|
return alg.hadamard_product(z, sigmoid(z));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::swish(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::swish(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
alg.addition(swish(z), alg.subtraction(sigmoid(z), alg.hadamard_product(sigmoid(z), swish(z))));
|
alg.addition(swish(z), alg.subtraction(sigmoid(z), alg.hadamard_product(sigmoid(z), swish(z))));
|
||||||
@ -325,14 +325,14 @@ std::vector<std::vector<double>> MLPPActivation::swish(std::vector<std::vector<d
|
|||||||
return alg.hadamard_product(z, sigmoid(z));
|
return alg.hadamard_product(z, sigmoid(z));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::mish(double z, bool deriv) {
|
real_t MLPPActivation::mish(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return sech(softplus(z)) * sech(softplus(z)) * z * sigmoid(z) + mish(z) / z;
|
return sech(softplus(z)) * sech(softplus(z)) * z * sigmoid(z) + mish(z) / z;
|
||||||
}
|
}
|
||||||
return z * tanh(softplus(z));
|
return z * tanh(softplus(z));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::mish(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::mish(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.addition(alg.hadamard_product(alg.hadamard_product(alg.hadamard_product(sech(softplus(z)), sech(softplus(z))), z), sigmoid(z)), alg.elementWiseDivision(mish(z), z));
|
return alg.addition(alg.hadamard_product(alg.hadamard_product(alg.hadamard_product(sech(softplus(z)), sech(softplus(z))), z), sigmoid(z)), alg.elementWiseDivision(mish(z), z));
|
||||||
@ -340,7 +340,7 @@ std::vector<double> MLPPActivation::mish(std::vector<double> z, bool deriv) {
|
|||||||
return alg.hadamard_product(z, tanh(softplus(z)));
|
return alg.hadamard_product(z, tanh(softplus(z)));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::mish(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::mish(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.addition(alg.hadamard_product(alg.hadamard_product(alg.hadamard_product(sech(softplus(z)), sech(softplus(z))), z), sigmoid(z)), alg.elementWiseDivision(mish(z), z));
|
return alg.addition(alg.hadamard_product(alg.hadamard_product(alg.hadamard_product(sech(softplus(z)), sech(softplus(z))), z), sigmoid(z)), alg.elementWiseDivision(mish(z), z));
|
||||||
@ -348,14 +348,14 @@ std::vector<std::vector<double>> MLPPActivation::mish(std::vector<std::vector<do
|
|||||||
return alg.hadamard_product(z, tanh(softplus(z)));
|
return alg.hadamard_product(z, tanh(softplus(z)));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::sinc(double z, bool deriv) {
|
real_t MLPPActivation::sinc(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return (z * std::cos(z) - std::sin(z)) / (z * z);
|
return (z * std::cos(z) - std::sin(z)) / (z * z);
|
||||||
}
|
}
|
||||||
return std::sin(z) / z;
|
return std::sin(z) / z;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::sinc(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::sinc(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.elementWiseDivision(alg.subtraction(alg.hadamard_product(z, alg.cos(z)), alg.sin(z)), alg.hadamard_product(z, z));
|
return alg.elementWiseDivision(alg.subtraction(alg.hadamard_product(z, alg.cos(z)), alg.sin(z)), alg.hadamard_product(z, z));
|
||||||
@ -363,7 +363,7 @@ std::vector<double> MLPPActivation::sinc(std::vector<double> z, bool deriv) {
|
|||||||
return alg.elementWiseDivision(alg.sin(z), z);
|
return alg.elementWiseDivision(alg.sin(z), z);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::sinc(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::sinc(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.elementWiseDivision(alg.subtraction(alg.hadamard_product(z, alg.cos(z)), alg.sin(z)), alg.hadamard_product(z, z));
|
return alg.elementWiseDivision(alg.subtraction(alg.hadamard_product(z, alg.cos(z)), alg.sin(z)), alg.hadamard_product(z, z));
|
||||||
@ -371,7 +371,7 @@ std::vector<std::vector<double>> MLPPActivation::sinc(std::vector<std::vector<do
|
|||||||
return alg.elementWiseDivision(alg.sin(z), z);
|
return alg.elementWiseDivision(alg.sin(z), z);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::RELU(double z, bool deriv) {
|
real_t MLPPActivation::RELU(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
if (z <= 0) {
|
if (z <= 0) {
|
||||||
return 0;
|
return 0;
|
||||||
@ -382,16 +382,16 @@ double MLPPActivation::RELU(double z, bool deriv) {
|
|||||||
return fmax(0, z);
|
return fmax(0, z);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::RELU(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::RELU(std::vector<real_t> z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
std::vector<double> deriv;
|
std::vector<real_t> deriv;
|
||||||
deriv.resize(z.size());
|
deriv.resize(z.size());
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
deriv[i] = RELU(z[i], 1);
|
deriv[i] = RELU(z[i], 1);
|
||||||
}
|
}
|
||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
std::vector<double> a;
|
std::vector<real_t> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
|
|
||||||
for (int i = 0; i < a.size(); i++) {
|
for (int i = 0; i < a.size(); i++) {
|
||||||
@ -400,16 +400,16 @@ std::vector<double> MLPPActivation::RELU(std::vector<double> z, bool deriv) {
|
|||||||
return a;
|
return a;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::RELU(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::RELU(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
std::vector<std::vector<double>> deriv;
|
std::vector<std::vector<real_t>> deriv;
|
||||||
deriv.resize(z.size());
|
deriv.resize(z.size());
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
deriv[i] = RELU(z[i], 1);
|
deriv[i] = RELU(z[i], 1);
|
||||||
}
|
}
|
||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
std::vector<std::vector<double>> a;
|
std::vector<std::vector<real_t>> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
|
|
||||||
for (int i = 0; i < a.size(); i++) {
|
for (int i = 0; i < a.size(); i++) {
|
||||||
@ -418,7 +418,7 @@ std::vector<std::vector<double>> MLPPActivation::RELU(std::vector<std::vector<do
|
|||||||
return a;
|
return a;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::leakyReLU(double z, double c, bool deriv) {
|
real_t MLPPActivation::leakyReLU(real_t z, real_t c, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
if (z <= 0) {
|
if (z <= 0) {
|
||||||
return c;
|
return c;
|
||||||
@ -429,16 +429,16 @@ double MLPPActivation::leakyReLU(double z, double c, bool deriv) {
|
|||||||
return fmax(c * z, z);
|
return fmax(c * z, z);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::leakyReLU(std::vector<double> z, double c, bool deriv) {
|
std::vector<real_t> MLPPActivation::leakyReLU(std::vector<real_t> z, real_t c, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
std::vector<double> deriv;
|
std::vector<real_t> deriv;
|
||||||
deriv.resize(z.size());
|
deriv.resize(z.size());
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
deriv[i] = leakyReLU(z[i], c, 1);
|
deriv[i] = leakyReLU(z[i], c, 1);
|
||||||
}
|
}
|
||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
std::vector<double> a;
|
std::vector<real_t> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
|
|
||||||
for (int i = 0; i < a.size(); i++) {
|
for (int i = 0; i < a.size(); i++) {
|
||||||
@ -447,16 +447,16 @@ std::vector<double> MLPPActivation::leakyReLU(std::vector<double> z, double c, b
|
|||||||
return a;
|
return a;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::leakyReLU(std::vector<std::vector<double>> z, double c, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::leakyReLU(std::vector<std::vector<real_t>> z, real_t c, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
std::vector<std::vector<double>> deriv;
|
std::vector<std::vector<real_t>> deriv;
|
||||||
deriv.resize(z.size());
|
deriv.resize(z.size());
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
deriv[i] = leakyReLU(z[i], c, 1);
|
deriv[i] = leakyReLU(z[i], c, 1);
|
||||||
}
|
}
|
||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
std::vector<std::vector<double>> a;
|
std::vector<std::vector<real_t>> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
|
|
||||||
for (int i = 0; i < a.size(); i++) {
|
for (int i = 0; i < a.size(); i++) {
|
||||||
@ -465,7 +465,7 @@ std::vector<std::vector<double>> MLPPActivation::leakyReLU(std::vector<std::vect
|
|||||||
return a;
|
return a;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::ELU(double z, double c, bool deriv) {
|
real_t MLPPActivation::ELU(real_t z, real_t c, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
if (z <= 0) {
|
if (z <= 0) {
|
||||||
return c * exp(z);
|
return c * exp(z);
|
||||||
@ -480,16 +480,16 @@ double MLPPActivation::ELU(double z, double c, bool deriv) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::ELU(std::vector<double> z, double c, bool deriv) {
|
std::vector<real_t> MLPPActivation::ELU(std::vector<real_t> z, real_t c, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
std::vector<double> deriv;
|
std::vector<real_t> deriv;
|
||||||
deriv.resize(z.size());
|
deriv.resize(z.size());
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
deriv[i] = ELU(z[i], c, 1);
|
deriv[i] = ELU(z[i], c, 1);
|
||||||
}
|
}
|
||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
std::vector<double> a;
|
std::vector<real_t> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
|
|
||||||
for (int i = 0; i < a.size(); i++) {
|
for (int i = 0; i < a.size(); i++) {
|
||||||
@ -498,16 +498,16 @@ std::vector<double> MLPPActivation::ELU(std::vector<double> z, double c, bool de
|
|||||||
return a;
|
return a;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::ELU(std::vector<std::vector<double>> z, double c, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::ELU(std::vector<std::vector<real_t>> z, real_t c, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
std::vector<std::vector<double>> deriv;
|
std::vector<std::vector<real_t>> deriv;
|
||||||
deriv.resize(z.size());
|
deriv.resize(z.size());
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
deriv[i] = ELU(z[i], c, 1);
|
deriv[i] = ELU(z[i], c, 1);
|
||||||
}
|
}
|
||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
std::vector<std::vector<double>> a;
|
std::vector<std::vector<real_t>> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
|
|
||||||
for (int i = 0; i < a.size(); i++) {
|
for (int i = 0; i < a.size(); i++) {
|
||||||
@ -516,23 +516,23 @@ std::vector<std::vector<double>> MLPPActivation::ELU(std::vector<std::vector<dou
|
|||||||
return a;
|
return a;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::SELU(double z, double lambda, double c, bool deriv) {
|
real_t MLPPActivation::SELU(real_t z, real_t lambda, real_t c, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return ELU(z, c, 1);
|
return ELU(z, c, 1);
|
||||||
}
|
}
|
||||||
return lambda * ELU(z, c);
|
return lambda * ELU(z, c);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::SELU(std::vector<double> z, double lambda, double c, bool deriv) {
|
std::vector<real_t> MLPPActivation::SELU(std::vector<real_t> z, real_t lambda, real_t c, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
std::vector<double> deriv;
|
std::vector<real_t> deriv;
|
||||||
deriv.resize(z.size());
|
deriv.resize(z.size());
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
deriv[i] = SELU(z[i], lambda, c, 1);
|
deriv[i] = SELU(z[i], lambda, c, 1);
|
||||||
}
|
}
|
||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
std::vector<double> a;
|
std::vector<real_t> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
|
|
||||||
for (int i = 0; i < a.size(); i++) {
|
for (int i = 0; i < a.size(); i++) {
|
||||||
@ -541,16 +541,16 @@ std::vector<double> MLPPActivation::SELU(std::vector<double> z, double lambda, d
|
|||||||
return a;
|
return a;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::SELU(std::vector<std::vector<double>> z, double lambda, double c, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::SELU(std::vector<std::vector<real_t>> z, real_t lambda, real_t c, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
std::vector<std::vector<double>> deriv;
|
std::vector<std::vector<real_t>> deriv;
|
||||||
deriv.resize(z.size());
|
deriv.resize(z.size());
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
deriv[i] = SELU(z[i], lambda, c, 1);
|
deriv[i] = SELU(z[i], lambda, c, 1);
|
||||||
}
|
}
|
||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
std::vector<std::vector<double>> a;
|
std::vector<std::vector<real_t>> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
|
|
||||||
for (int i = 0; i < a.size(); i++) {
|
for (int i = 0; i < a.size(); i++) {
|
||||||
@ -559,23 +559,23 @@ std::vector<std::vector<double>> MLPPActivation::SELU(std::vector<std::vector<do
|
|||||||
return a;
|
return a;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::GELU(double z, bool deriv) {
|
real_t MLPPActivation::GELU(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return 0.5 * tanh(0.0356774 * std::pow(z, 3) + 0.797885 * z) + (0.0535161 * std::pow(z, 3) + 0.398942 * z) * std::pow(sech(0.0356774 * std::pow(z, 3) + 0.797885 * z), 2) + 0.5;
|
return 0.5 * tanh(0.0356774 * std::pow(z, 3) + 0.797885 * z) + (0.0535161 * std::pow(z, 3) + 0.398942 * z) * std::pow(sech(0.0356774 * std::pow(z, 3) + 0.797885 * z), 2) + 0.5;
|
||||||
}
|
}
|
||||||
return 0.5 * z * (1 + tanh(sqrt(2 / M_PI) * (z + 0.044715 * std::pow(z, 3))));
|
return 0.5 * z * (1 + tanh(sqrt(2 / M_PI) * (z + 0.044715 * std::pow(z, 3))));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::GELU(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::GELU(std::vector<real_t> z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
std::vector<double> deriv;
|
std::vector<real_t> deriv;
|
||||||
deriv.resize(z.size());
|
deriv.resize(z.size());
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
deriv[i] = GELU(z[i], 1);
|
deriv[i] = GELU(z[i], 1);
|
||||||
}
|
}
|
||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
std::vector<double> a;
|
std::vector<real_t> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
|
|
||||||
for (int i = 0; i < a.size(); i++) {
|
for (int i = 0; i < a.size(); i++) {
|
||||||
@ -584,16 +584,16 @@ std::vector<double> MLPPActivation::GELU(std::vector<double> z, bool deriv) {
|
|||||||
return a;
|
return a;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::GELU(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::GELU(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
std::vector<std::vector<double>> deriv;
|
std::vector<std::vector<real_t>> deriv;
|
||||||
deriv.resize(z.size());
|
deriv.resize(z.size());
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
deriv[i] = GELU(z[i], 1);
|
deriv[i] = GELU(z[i], 1);
|
||||||
}
|
}
|
||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
std::vector<std::vector<double>> a;
|
std::vector<std::vector<real_t>> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
|
|
||||||
for (int i = 0; i < a.size(); i++) {
|
for (int i = 0; i < a.size(); i++) {
|
||||||
@ -602,7 +602,7 @@ std::vector<std::vector<double>> MLPPActivation::GELU(std::vector<std::vector<do
|
|||||||
return a;
|
return a;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::sign(double z, bool deriv) {
|
real_t MLPPActivation::sign(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
@ -615,16 +615,16 @@ double MLPPActivation::sign(double z, bool deriv) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::sign(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::sign(std::vector<real_t> z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
std::vector<double> deriv;
|
std::vector<real_t> deriv;
|
||||||
deriv.resize(z.size());
|
deriv.resize(z.size());
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
deriv[i] = sign(z[i], 1);
|
deriv[i] = sign(z[i], 1);
|
||||||
}
|
}
|
||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
std::vector<double> a;
|
std::vector<real_t> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
|
|
||||||
for (int i = 0; i < a.size(); i++) {
|
for (int i = 0; i < a.size(); i++) {
|
||||||
@ -633,16 +633,16 @@ std::vector<double> MLPPActivation::sign(std::vector<double> z, bool deriv) {
|
|||||||
return a;
|
return a;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::sign(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::sign(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
std::vector<std::vector<double>> deriv;
|
std::vector<std::vector<real_t>> deriv;
|
||||||
deriv.resize(z.size());
|
deriv.resize(z.size());
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
deriv[i] = sign(z[i], 1);
|
deriv[i] = sign(z[i], 1);
|
||||||
}
|
}
|
||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
std::vector<std::vector<double>> a;
|
std::vector<std::vector<real_t>> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
|
|
||||||
for (int i = 0; i < a.size(); i++) {
|
for (int i = 0; i < a.size(); i++) {
|
||||||
@ -651,14 +651,14 @@ std::vector<std::vector<double>> MLPPActivation::sign(std::vector<std::vector<do
|
|||||||
return a;
|
return a;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::sinh(double z, bool deriv) {
|
real_t MLPPActivation::sinh(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return cosh(z);
|
return cosh(z);
|
||||||
}
|
}
|
||||||
return 0.5 * (exp(z) - exp(-z));
|
return 0.5 * (exp(z) - exp(-z));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::sinh(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::sinh(std::vector<real_t> z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return cosh(z);
|
return cosh(z);
|
||||||
}
|
}
|
||||||
@ -666,7 +666,7 @@ std::vector<double> MLPPActivation::sinh(std::vector<double> z, bool deriv) {
|
|||||||
return alg.scalarMultiply(0.5, alg.subtraction(alg.exp(z), alg.exp(alg.scalarMultiply(-1, z))));
|
return alg.scalarMultiply(0.5, alg.subtraction(alg.exp(z), alg.exp(alg.scalarMultiply(-1, z))));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::sinh(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::sinh(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return cosh(z);
|
return cosh(z);
|
||||||
}
|
}
|
||||||
@ -674,14 +674,14 @@ std::vector<std::vector<double>> MLPPActivation::sinh(std::vector<std::vector<do
|
|||||||
return alg.scalarMultiply(0.5, alg.subtraction(alg.exp(z), alg.exp(alg.scalarMultiply(-1, z))));
|
return alg.scalarMultiply(0.5, alg.subtraction(alg.exp(z), alg.exp(alg.scalarMultiply(-1, z))));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::cosh(double z, bool deriv) {
|
real_t MLPPActivation::cosh(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return sinh(z);
|
return sinh(z);
|
||||||
}
|
}
|
||||||
return 0.5 * (exp(z) + exp(-z));
|
return 0.5 * (exp(z) + exp(-z));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::cosh(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::cosh(std::vector<real_t> z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return sinh(z);
|
return sinh(z);
|
||||||
}
|
}
|
||||||
@ -689,7 +689,7 @@ std::vector<double> MLPPActivation::cosh(std::vector<double> z, bool deriv) {
|
|||||||
return alg.scalarMultiply(0.5, alg.addition(alg.exp(z), alg.exp(alg.scalarMultiply(-1, z))));
|
return alg.scalarMultiply(0.5, alg.addition(alg.exp(z), alg.exp(alg.scalarMultiply(-1, z))));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::cosh(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::cosh(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return sinh(z);
|
return sinh(z);
|
||||||
}
|
}
|
||||||
@ -697,14 +697,14 @@ std::vector<std::vector<double>> MLPPActivation::cosh(std::vector<std::vector<do
|
|||||||
return alg.scalarMultiply(0.5, alg.addition(alg.exp(z), alg.exp(alg.scalarMultiply(-1, z))));
|
return alg.scalarMultiply(0.5, alg.addition(alg.exp(z), alg.exp(alg.scalarMultiply(-1, z))));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::tanh(double z, bool deriv) {
|
real_t MLPPActivation::tanh(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return 1 - tanh(z) * tanh(z);
|
return 1 - tanh(z) * tanh(z);
|
||||||
}
|
}
|
||||||
return (exp(z) - exp(-z)) / (exp(z) + exp(-z));
|
return (exp(z) - exp(-z)) / (exp(z) + exp(-z));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::tanh(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::tanh(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.scalarMultiply(-1, alg.scalarAdd(-1, alg.hadamard_product(tanh(z), tanh(z))));
|
return alg.scalarMultiply(-1, alg.scalarAdd(-1, alg.hadamard_product(tanh(z), tanh(z))));
|
||||||
@ -712,7 +712,7 @@ std::vector<double> MLPPActivation::tanh(std::vector<double> z, bool deriv) {
|
|||||||
return alg.elementWiseDivision(alg.subtraction(alg.exp(z), alg.exp(alg.scalarMultiply(-1, z))), alg.addition(alg.exp(z), alg.exp(alg.scalarMultiply(-1, z))));
|
return alg.elementWiseDivision(alg.subtraction(alg.exp(z), alg.exp(alg.scalarMultiply(-1, z))), alg.addition(alg.exp(z), alg.exp(alg.scalarMultiply(-1, z))));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::tanh(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::tanh(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.scalarMultiply(-1, alg.scalarAdd(-1, alg.hadamard_product(tanh(z), tanh(z))));
|
return alg.scalarMultiply(-1, alg.scalarAdd(-1, alg.hadamard_product(tanh(z), tanh(z))));
|
||||||
@ -721,14 +721,14 @@ std::vector<std::vector<double>> MLPPActivation::tanh(std::vector<std::vector<do
|
|||||||
return alg.elementWiseDivision(alg.subtraction(alg.exp(z), alg.exp(alg.scalarMultiply(-1, z))), alg.addition(alg.exp(z), alg.exp(alg.scalarMultiply(-1, z))));
|
return alg.elementWiseDivision(alg.subtraction(alg.exp(z), alg.exp(alg.scalarMultiply(-1, z))), alg.addition(alg.exp(z), alg.exp(alg.scalarMultiply(-1, z))));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::csch(double z, bool deriv) {
|
real_t MLPPActivation::csch(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return -csch(z) * coth(z);
|
return -csch(z) * coth(z);
|
||||||
}
|
}
|
||||||
return 1 / sinh(z);
|
return 1 / sinh(z);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::csch(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::csch(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.hadamard_product(alg.scalarMultiply(-1, csch(z)), coth(z));
|
return alg.hadamard_product(alg.scalarMultiply(-1, csch(z)), coth(z));
|
||||||
@ -736,7 +736,7 @@ std::vector<double> MLPPActivation::csch(std::vector<double> z, bool deriv) {
|
|||||||
return alg.elementWiseDivision(alg.onevec(z.size()), sinh(z));
|
return alg.elementWiseDivision(alg.onevec(z.size()), sinh(z));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::csch(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::csch(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.hadamard_product(alg.scalarMultiply(-1, csch(z)), coth(z));
|
return alg.hadamard_product(alg.scalarMultiply(-1, csch(z)), coth(z));
|
||||||
@ -744,41 +744,41 @@ std::vector<std::vector<double>> MLPPActivation::csch(std::vector<std::vector<do
|
|||||||
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), sinh(z));
|
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), sinh(z));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::sech(double z, bool deriv) {
|
real_t MLPPActivation::sech(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return -sech(z) * tanh(z);
|
return -sech(z) * tanh(z);
|
||||||
}
|
}
|
||||||
return 1 / cosh(z);
|
return 1 / cosh(z);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::sech(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::sech(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.hadamard_product(alg.scalarMultiply(-1, sech(z)), tanh(z));
|
return alg.hadamard_product(alg.scalarMultiply(-1, sech(z)), tanh(z));
|
||||||
}
|
}
|
||||||
return alg.elementWiseDivision(alg.onevec(z.size()), cosh(z));
|
return alg.elementWiseDivision(alg.onevec(z.size()), cosh(z));
|
||||||
|
|
||||||
// return activation(z, deriv, static_cast<void (*)(double, bool)>(&sech));
|
// return activation(z, deriv, static_cast<void (*)(real_t, bool)>(&sech));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::sech(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::sech(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.hadamard_product(alg.scalarMultiply(-1, sech(z)), tanh(z));
|
return alg.hadamard_product(alg.scalarMultiply(-1, sech(z)), tanh(z));
|
||||||
}
|
}
|
||||||
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), cosh(z));
|
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), cosh(z));
|
||||||
|
|
||||||
// return activation(z, deriv, static_cast<void (*)(double, bool)>(&sech));
|
// return activation(z, deriv, static_cast<void (*)(real_t, bool)>(&sech));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::coth(double z, bool deriv) {
|
real_t MLPPActivation::coth(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return -csch(z) * csch(z);
|
return -csch(z) * csch(z);
|
||||||
}
|
}
|
||||||
return 1 / tanh(z);
|
return 1 / tanh(z);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::coth(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::coth(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.hadamard_product(alg.scalarMultiply(-1, csch(z)), csch(z));
|
return alg.hadamard_product(alg.scalarMultiply(-1, csch(z)), csch(z));
|
||||||
@ -786,7 +786,7 @@ std::vector<double> MLPPActivation::coth(std::vector<double> z, bool deriv) {
|
|||||||
return alg.elementWiseDivision(alg.onevec(z.size()), tanh(z));
|
return alg.elementWiseDivision(alg.onevec(z.size()), tanh(z));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::coth(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::coth(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.hadamard_product(alg.scalarMultiply(-1, csch(z)), csch(z));
|
return alg.hadamard_product(alg.scalarMultiply(-1, csch(z)), csch(z));
|
||||||
@ -794,14 +794,14 @@ std::vector<std::vector<double>> MLPPActivation::coth(std::vector<std::vector<do
|
|||||||
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), tanh(z));
|
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), tanh(z));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::arsinh(double z, bool deriv) {
|
real_t MLPPActivation::arsinh(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return 1 / sqrt(z * z + 1);
|
return 1 / sqrt(z * z + 1);
|
||||||
}
|
}
|
||||||
return std::log(z + sqrt(z * z + 1));
|
return std::log(z + sqrt(z * z + 1));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::arsinh(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::arsinh(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.elementWiseDivision(alg.onevec(z.size()), alg.sqrt(alg.addition(alg.hadamard_product(z, z), alg.onevec(z.size()))));
|
return alg.elementWiseDivision(alg.onevec(z.size()), alg.sqrt(alg.addition(alg.hadamard_product(z, z), alg.onevec(z.size()))));
|
||||||
@ -809,7 +809,7 @@ std::vector<double> MLPPActivation::arsinh(std::vector<double> z, bool deriv) {
|
|||||||
return alg.log(alg.addition(z, alg.sqrt(alg.addition(alg.hadamard_product(z, z), alg.onevec(z.size())))));
|
return alg.log(alg.addition(z, alg.sqrt(alg.addition(alg.hadamard_product(z, z), alg.onevec(z.size())))));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::arsinh(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::arsinh(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.sqrt(alg.addition(alg.hadamard_product(z, z), alg.onemat(z.size(), z[0].size()))));
|
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.sqrt(alg.addition(alg.hadamard_product(z, z), alg.onemat(z.size(), z[0].size()))));
|
||||||
@ -817,14 +817,14 @@ std::vector<std::vector<double>> MLPPActivation::arsinh(std::vector<std::vector<
|
|||||||
return alg.log(alg.addition(z, alg.sqrt(alg.addition(alg.hadamard_product(z, z), alg.onemat(z.size(), z[0].size())))));
|
return alg.log(alg.addition(z, alg.sqrt(alg.addition(alg.hadamard_product(z, z), alg.onemat(z.size(), z[0].size())))));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::arcosh(double z, bool deriv) {
|
real_t MLPPActivation::arcosh(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return 1 / sqrt(z * z - 1);
|
return 1 / sqrt(z * z - 1);
|
||||||
}
|
}
|
||||||
return std::log(z + sqrt(z * z - 1));
|
return std::log(z + sqrt(z * z - 1));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::arcosh(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::arcosh(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.elementWiseDivision(alg.onevec(z.size()), alg.sqrt(alg.subtraction(alg.hadamard_product(z, z), alg.onevec(z.size()))));
|
return alg.elementWiseDivision(alg.onevec(z.size()), alg.sqrt(alg.subtraction(alg.hadamard_product(z, z), alg.onevec(z.size()))));
|
||||||
@ -832,7 +832,7 @@ std::vector<double> MLPPActivation::arcosh(std::vector<double> z, bool deriv) {
|
|||||||
return alg.log(alg.addition(z, alg.sqrt(alg.subtraction(alg.hadamard_product(z, z), alg.onevec(z.size())))));
|
return alg.log(alg.addition(z, alg.sqrt(alg.subtraction(alg.hadamard_product(z, z), alg.onevec(z.size())))));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::arcosh(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::arcosh(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.sqrt(alg.subtraction(alg.hadamard_product(z, z), alg.onemat(z.size(), z[0].size()))));
|
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.sqrt(alg.subtraction(alg.hadamard_product(z, z), alg.onemat(z.size(), z[0].size()))));
|
||||||
@ -840,14 +840,14 @@ std::vector<std::vector<double>> MLPPActivation::arcosh(std::vector<std::vector<
|
|||||||
return alg.log(alg.addition(z, alg.sqrt(alg.subtraction(alg.hadamard_product(z, z), alg.onemat(z.size(), z[0].size())))));
|
return alg.log(alg.addition(z, alg.sqrt(alg.subtraction(alg.hadamard_product(z, z), alg.onemat(z.size(), z[0].size())))));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::artanh(double z, bool deriv) {
|
real_t MLPPActivation::artanh(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return 1 / (1 - z * z);
|
return 1 / (1 - z * z);
|
||||||
}
|
}
|
||||||
return 0.5 * std::log((1 + z) / (1 - z));
|
return 0.5 * std::log((1 + z) / (1 - z));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::artanh(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::artanh(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.elementWiseDivision(alg.onevec(z.size()), alg.subtraction(alg.onevec(z.size()), alg.hadamard_product(z, z)));
|
return alg.elementWiseDivision(alg.onevec(z.size()), alg.subtraction(alg.onevec(z.size()), alg.hadamard_product(z, z)));
|
||||||
@ -855,7 +855,7 @@ std::vector<double> MLPPActivation::artanh(std::vector<double> z, bool deriv) {
|
|||||||
return alg.scalarMultiply(0.5, alg.log(alg.elementWiseDivision(alg.addition(alg.onevec(z.size()), z), alg.subtraction(alg.onevec(z.size()), z))));
|
return alg.scalarMultiply(0.5, alg.log(alg.elementWiseDivision(alg.addition(alg.onevec(z.size()), z), alg.subtraction(alg.onevec(z.size()), z))));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::artanh(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::artanh(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.subtraction(alg.onemat(z.size(), z[0].size()), alg.hadamard_product(z, z)));
|
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.subtraction(alg.onemat(z.size(), z[0].size()), alg.hadamard_product(z, z)));
|
||||||
@ -863,14 +863,14 @@ std::vector<std::vector<double>> MLPPActivation::artanh(std::vector<std::vector<
|
|||||||
return alg.scalarMultiply(0.5, alg.log(alg.elementWiseDivision(alg.addition(alg.onemat(z.size(), z[0].size()), z), alg.subtraction(alg.onemat(z.size(), z[0].size()), z))));
|
return alg.scalarMultiply(0.5, alg.log(alg.elementWiseDivision(alg.addition(alg.onemat(z.size(), z[0].size()), z), alg.subtraction(alg.onemat(z.size(), z[0].size()), z))));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::arcsch(double z, bool deriv) {
|
real_t MLPPActivation::arcsch(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return -1 / ((z * z) * sqrt(1 + (1 / (z * z))));
|
return -1 / ((z * z) * sqrt(1 + (1 / (z * z))));
|
||||||
}
|
}
|
||||||
return std::log(sqrt(1 + (1 / (z * z))) + (1 / z));
|
return std::log(sqrt(1 + (1 / (z * z))) + (1 / z));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::arcsch(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::arcsch(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.elementWiseDivision(alg.full(z.size(), -1), alg.hadamard_product(alg.hadamard_product(z, z), alg.sqrt(alg.addition(alg.onevec(z.size()), alg.elementWiseDivision(alg.onevec(z.size()), alg.hadamard_product(z, z))))));
|
return alg.elementWiseDivision(alg.full(z.size(), -1), alg.hadamard_product(alg.hadamard_product(z, z), alg.sqrt(alg.addition(alg.onevec(z.size()), alg.elementWiseDivision(alg.onevec(z.size()), alg.hadamard_product(z, z))))));
|
||||||
@ -878,7 +878,7 @@ std::vector<double> MLPPActivation::arcsch(std::vector<double> z, bool deriv) {
|
|||||||
return alg.log(alg.addition(alg.sqrt(alg.addition(alg.onevec(z.size()), alg.elementWiseDivision(alg.onevec(z.size()), alg.hadamard_product(z, z)))), alg.elementWiseDivision(alg.onevec(z.size()), z)));
|
return alg.log(alg.addition(alg.sqrt(alg.addition(alg.onevec(z.size()), alg.elementWiseDivision(alg.onevec(z.size()), alg.hadamard_product(z, z)))), alg.elementWiseDivision(alg.onevec(z.size()), z)));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::arcsch(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::arcsch(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.elementWiseDivision(alg.full(z.size(), z[0].size(), -1), alg.hadamard_product(alg.hadamard_product(z, z), alg.sqrt(alg.addition(alg.onemat(z.size(), z[0].size()), alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.hadamard_product(z, z))))));
|
return alg.elementWiseDivision(alg.full(z.size(), z[0].size(), -1), alg.hadamard_product(alg.hadamard_product(z, z), alg.sqrt(alg.addition(alg.onemat(z.size(), z[0].size()), alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.hadamard_product(z, z))))));
|
||||||
@ -886,14 +886,14 @@ std::vector<std::vector<double>> MLPPActivation::arcsch(std::vector<std::vector<
|
|||||||
return alg.log(alg.addition(alg.sqrt(alg.addition(alg.onemat(z.size(), z[0].size()), alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.hadamard_product(z, z)))), alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), z)));
|
return alg.log(alg.addition(alg.sqrt(alg.addition(alg.onemat(z.size(), z[0].size()), alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.hadamard_product(z, z)))), alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), z)));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::arsech(double z, bool deriv) {
|
real_t MLPPActivation::arsech(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return -1 / (z * sqrt(1 - z * z));
|
return -1 / (z * sqrt(1 - z * z));
|
||||||
}
|
}
|
||||||
return std::log((1 / z) + ((1 / z) + 1) * ((1 / z) - 1));
|
return std::log((1 / z) + ((1 / z) + 1) * ((1 / z) - 1));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::arsech(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::arsech(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.elementWiseDivision(alg.full(z.size(), -1), alg.hadamard_product(z, alg.sqrt(alg.subtraction(alg.onevec(z.size()), alg.hadamard_product(z, z)))));
|
return alg.elementWiseDivision(alg.full(z.size(), -1), alg.hadamard_product(z, alg.sqrt(alg.subtraction(alg.onevec(z.size()), alg.hadamard_product(z, z)))));
|
||||||
@ -901,7 +901,7 @@ std::vector<double> MLPPActivation::arsech(std::vector<double> z, bool deriv) {
|
|||||||
return alg.log(alg.addition(alg.elementWiseDivision(alg.onevec(z.size()), z), alg.hadamard_product(alg.addition(alg.elementWiseDivision(alg.onevec(z.size()), z), alg.onevec(z.size())), alg.subtraction(alg.elementWiseDivision(alg.onevec(z.size()), z), alg.onevec(z.size())))));
|
return alg.log(alg.addition(alg.elementWiseDivision(alg.onevec(z.size()), z), alg.hadamard_product(alg.addition(alg.elementWiseDivision(alg.onevec(z.size()), z), alg.onevec(z.size())), alg.subtraction(alg.elementWiseDivision(alg.onevec(z.size()), z), alg.onevec(z.size())))));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::arsech(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::arsech(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.elementWiseDivision(alg.full(z.size(), z[0].size(), -1), alg.hadamard_product(z, alg.sqrt(alg.subtraction(alg.onemat(z.size(), z[0].size()), alg.hadamard_product(z, z)))));
|
return alg.elementWiseDivision(alg.full(z.size(), z[0].size(), -1), alg.hadamard_product(z, alg.sqrt(alg.subtraction(alg.onemat(z.size(), z[0].size()), alg.hadamard_product(z, z)))));
|
||||||
@ -909,14 +909,14 @@ std::vector<std::vector<double>> MLPPActivation::arsech(std::vector<std::vector<
|
|||||||
return alg.log(alg.addition(alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), z), alg.hadamard_product(alg.addition(alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), z), alg.onemat(z.size(), z[0].size())), alg.subtraction(alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), z), alg.onemat(z.size(), z[0].size())))));
|
return alg.log(alg.addition(alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), z), alg.hadamard_product(alg.addition(alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), z), alg.onemat(z.size(), z[0].size())), alg.subtraction(alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), z), alg.onemat(z.size(), z[0].size())))));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPActivation::arcoth(double z, bool deriv) {
|
real_t MLPPActivation::arcoth(real_t z, bool deriv) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return 1 / (1 - z * z);
|
return 1 / (1 - z * z);
|
||||||
}
|
}
|
||||||
return 0.5 * std::log((1 + z) / (z - 1));
|
return 0.5 * std::log((1 + z) / (z - 1));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPActivation::arcoth(std::vector<double> z, bool deriv) {
|
std::vector<real_t> MLPPActivation::arcoth(std::vector<real_t> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.elementWiseDivision(alg.onevec(z.size()), alg.subtraction(alg.onevec(z.size()), alg.hadamard_product(z, z)));
|
return alg.elementWiseDivision(alg.onevec(z.size()), alg.subtraction(alg.onevec(z.size()), alg.hadamard_product(z, z)));
|
||||||
@ -924,7 +924,7 @@ std::vector<double> MLPPActivation::arcoth(std::vector<double> z, bool deriv) {
|
|||||||
return alg.scalarMultiply(0.5, alg.log(alg.elementWiseDivision(alg.addition(alg.onevec(z.size()), z), alg.subtraction(z, alg.onevec(z.size())))));
|
return alg.scalarMultiply(0.5, alg.log(alg.elementWiseDivision(alg.addition(alg.onevec(z.size()), z), alg.subtraction(z, alg.onevec(z.size())))));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPActivation::arcoth(std::vector<std::vector<double>> z, bool deriv) {
|
std::vector<std::vector<real_t>> MLPPActivation::arcoth(std::vector<std::vector<real_t>> z, bool deriv) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.subtraction(alg.onemat(z.size(), z[0].size()), alg.hadamard_product(z, z)));
|
return alg.elementWiseDivision(alg.onemat(z.size(), z[0].size()), alg.subtraction(alg.onemat(z.size(), z[0].size()), alg.hadamard_product(z, z)));
|
||||||
@ -933,16 +933,16 @@ std::vector<std::vector<double>> MLPPActivation::arcoth(std::vector<std::vector<
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TO DO: Implement this template activation
|
// TO DO: Implement this template activation
|
||||||
std::vector<double> MLPPActivation::activation(std::vector<double> z, bool deriv, double (*function)(double, bool)) {
|
std::vector<real_t> MLPPActivation::activation(std::vector<real_t> z, bool deriv, real_t (*function)(real_t, bool)) {
|
||||||
if (deriv) {
|
if (deriv) {
|
||||||
std::vector<double> deriv;
|
std::vector<real_t> deriv;
|
||||||
deriv.resize(z.size());
|
deriv.resize(z.size());
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
deriv[i] = function(z[i], 1);
|
deriv[i] = function(z[i], 1);
|
||||||
}
|
}
|
||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
std::vector<double> a;
|
std::vector<real_t> a;
|
||||||
a.resize(z.size());
|
a.resize(z.size());
|
||||||
for (int i = 0; i < z.size(); i++) {
|
for (int i = 0; i < z.size(); i++) {
|
||||||
a[i] = function(z[i], deriv);
|
a[i] = function(z[i], deriv);
|
||||||
|
@ -8,136 +8,138 @@
|
|||||||
// Created by Marc Melikyan on 1/16/21.
|
// Created by Marc Melikyan on 1/16/21.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
class MLPPActivation {
|
class MLPPActivation {
|
||||||
public:
|
public:
|
||||||
double linear(double z, bool deriv = 0);
|
real_t linear(real_t z, bool deriv = 0);
|
||||||
std::vector<double> linear(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> linear(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> linear(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> linear(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double sigmoid(double z, bool deriv = 0);
|
real_t sigmoid(real_t z, bool deriv = 0);
|
||||||
std::vector<double> sigmoid(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> sigmoid(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> sigmoid(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> sigmoid(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
std::vector<double> softmax(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> softmax(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> softmax(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> softmax(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
std::vector<double> adjSoftmax(std::vector<double> z);
|
std::vector<real_t> adjSoftmax(std::vector<real_t> z);
|
||||||
std::vector<std::vector<double>> adjSoftmax(std::vector<std::vector<double>> z);
|
std::vector<std::vector<real_t>> adjSoftmax(std::vector<std::vector<real_t>> z);
|
||||||
|
|
||||||
std::vector<std::vector<double>> softmaxDeriv(std::vector<double> z);
|
std::vector<std::vector<real_t>> softmaxDeriv(std::vector<real_t> z);
|
||||||
std::vector<std::vector<std::vector<double>>> softmaxDeriv(std::vector<std::vector<double>> z);
|
std::vector<std::vector<std::vector<real_t>>> softmaxDeriv(std::vector<std::vector<real_t>> z);
|
||||||
|
|
||||||
double softplus(double z, bool deriv = 0);
|
real_t softplus(real_t z, bool deriv = 0);
|
||||||
std::vector<double> softplus(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> softplus(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> softplus(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> softplus(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double softsign(double z, bool deriv = 0);
|
real_t softsign(real_t z, bool deriv = 0);
|
||||||
std::vector<double> softsign(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> softsign(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> softsign(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> softsign(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double gaussianCDF(double z, bool deriv = 0);
|
real_t gaussianCDF(real_t z, bool deriv = 0);
|
||||||
std::vector<double> gaussianCDF(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> gaussianCDF(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> gaussianCDF(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> gaussianCDF(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double cloglog(double z, bool deriv = 0);
|
real_t cloglog(real_t z, bool deriv = 0);
|
||||||
std::vector<double> cloglog(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> cloglog(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> cloglog(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> cloglog(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double logit(double z, bool deriv = 0);
|
real_t logit(real_t z, bool deriv = 0);
|
||||||
std::vector<double> logit(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> logit(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> logit(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> logit(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double unitStep(double z, bool deriv = 0);
|
real_t unitStep(real_t z, bool deriv = 0);
|
||||||
std::vector<double> unitStep(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> unitStep(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> unitStep(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> unitStep(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double swish(double z, bool deriv = 0);
|
real_t swish(real_t z, bool deriv = 0);
|
||||||
std::vector<double> swish(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> swish(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> swish(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> swish(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double mish(double z, bool deriv = 0);
|
real_t mish(real_t z, bool deriv = 0);
|
||||||
std::vector<double> mish(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> mish(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> mish(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> mish(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double sinc(double z, bool deriv = 0);
|
real_t sinc(real_t z, bool deriv = 0);
|
||||||
std::vector<double> sinc(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> sinc(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> sinc(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> sinc(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double RELU(double z, bool deriv = 0);
|
real_t RELU(real_t z, bool deriv = 0);
|
||||||
std::vector<double> RELU(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> RELU(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> RELU(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> RELU(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double leakyReLU(double z, double c, bool deriv = 0);
|
real_t leakyReLU(real_t z, real_t c, bool deriv = 0);
|
||||||
std::vector<double> leakyReLU(std::vector<double> z, double c, bool deriv = 0);
|
std::vector<real_t> leakyReLU(std::vector<real_t> z, real_t c, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> leakyReLU(std::vector<std::vector<double>> z, double c, bool deriv = 0);
|
std::vector<std::vector<real_t>> leakyReLU(std::vector<std::vector<real_t>> z, real_t c, bool deriv = 0);
|
||||||
|
|
||||||
double ELU(double z, double c, bool deriv = 0);
|
real_t ELU(real_t z, real_t c, bool deriv = 0);
|
||||||
std::vector<double> ELU(std::vector<double> z, double c, bool deriv = 0);
|
std::vector<real_t> ELU(std::vector<real_t> z, real_t c, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> ELU(std::vector<std::vector<double>> z, double c, bool deriv = 0);
|
std::vector<std::vector<real_t>> ELU(std::vector<std::vector<real_t>> z, real_t c, bool deriv = 0);
|
||||||
|
|
||||||
double SELU(double z, double lambda, double c, bool deriv = 0);
|
real_t SELU(real_t z, real_t lambda, real_t c, bool deriv = 0);
|
||||||
std::vector<double> SELU(std::vector<double> z, double lambda, double c, bool deriv = 0);
|
std::vector<real_t> SELU(std::vector<real_t> z, real_t lambda, real_t c, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> SELU(std::vector<std::vector<double>>, double lambda, double c, bool deriv = 0);
|
std::vector<std::vector<real_t>> SELU(std::vector<std::vector<real_t>>, real_t lambda, real_t c, bool deriv = 0);
|
||||||
|
|
||||||
double GELU(double z, bool deriv = 0);
|
real_t GELU(real_t z, bool deriv = 0);
|
||||||
std::vector<double> GELU(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> GELU(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> GELU(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> GELU(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double sign(double z, bool deriv = 0);
|
real_t sign(real_t z, bool deriv = 0);
|
||||||
std::vector<double> sign(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> sign(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> sign(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> sign(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double sinh(double z, bool deriv = 0);
|
real_t sinh(real_t z, bool deriv = 0);
|
||||||
std::vector<double> sinh(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> sinh(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> sinh(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> sinh(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double cosh(double z, bool deriv = 0);
|
real_t cosh(real_t z, bool deriv = 0);
|
||||||
std::vector<double> cosh(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> cosh(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> cosh(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> cosh(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double tanh(double z, bool deriv = 0);
|
real_t tanh(real_t z, bool deriv = 0);
|
||||||
std::vector<double> tanh(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> tanh(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> tanh(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> tanh(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double csch(double z, bool deriv = 0);
|
real_t csch(real_t z, bool deriv = 0);
|
||||||
std::vector<double> csch(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> csch(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> csch(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> csch(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double sech(double z, bool deriv = 0);
|
real_t sech(real_t z, bool deriv = 0);
|
||||||
std::vector<double> sech(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> sech(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> sech(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> sech(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double coth(double z, bool deriv = 0);
|
real_t coth(real_t z, bool deriv = 0);
|
||||||
std::vector<double> coth(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> coth(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> coth(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> coth(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double arsinh(double z, bool deriv = 0);
|
real_t arsinh(real_t z, bool deriv = 0);
|
||||||
std::vector<double> arsinh(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> arsinh(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> arsinh(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> arsinh(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double arcosh(double z, bool deriv = 0);
|
real_t arcosh(real_t z, bool deriv = 0);
|
||||||
std::vector<double> arcosh(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> arcosh(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> arcosh(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> arcosh(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double artanh(double z, bool deriv = 0);
|
real_t artanh(real_t z, bool deriv = 0);
|
||||||
std::vector<double> artanh(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> artanh(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> artanh(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> artanh(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double arcsch(double z, bool deriv = 0);
|
real_t arcsch(real_t z, bool deriv = 0);
|
||||||
std::vector<double> arcsch(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> arcsch(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> arcsch(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> arcsch(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double arsech(double z, bool deriv = 0);
|
real_t arsech(real_t z, bool deriv = 0);
|
||||||
std::vector<double> arsech(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> arsech(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> arsech(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> arsech(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
double arcoth(double z, bool deriv = 0);
|
real_t arcoth(real_t z, bool deriv = 0);
|
||||||
std::vector<double> arcoth(std::vector<double> z, bool deriv = 0);
|
std::vector<real_t> arcoth(std::vector<real_t> z, bool deriv = 0);
|
||||||
std::vector<std::vector<double>> arcoth(std::vector<std::vector<double>> z, bool deriv = 0);
|
std::vector<std::vector<real_t>> arcoth(std::vector<std::vector<real_t>> z, bool deriv = 0);
|
||||||
|
|
||||||
std::vector<double> activation(std::vector<double> z, bool deriv, double (*function)(double, bool));
|
std::vector<real_t> activation(std::vector<real_t> z, bool deriv, real_t (*function)(real_t, bool));
|
||||||
|
|
||||||
private:
|
private:
|
||||||
};
|
};
|
||||||
|
212
mlpp/ann/ann.cpp
212
mlpp/ann/ann.cpp
@ -15,7 +15,7 @@
|
|||||||
#include <iostream>
|
#include <iostream>
|
||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
MLPPANN::MLPPANN(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet) :
|
MLPPANN::MLPPANN(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet) :
|
||||||
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), lrScheduler("None"), decayConstant(0), dropRate(0) {
|
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), lrScheduler("None"), decayConstant(0), dropRate(0) {
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -23,7 +23,7 @@ MLPPANN::~MLPPANN() {
|
|||||||
delete outputLayer;
|
delete outputLayer;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPANN::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPANN::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
network[0].input = X;
|
network[0].input = X;
|
||||||
network[0].forwardPass();
|
network[0].forwardPass();
|
||||||
@ -40,7 +40,7 @@ std::vector<double> MLPPANN::modelSetTest(std::vector<std::vector<double>> X) {
|
|||||||
return outputLayer->a;
|
return outputLayer->a;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPANN::modelTest(std::vector<double> x) {
|
real_t MLPPANN::modelTest(std::vector<real_t> x) {
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
network[0].Test(x);
|
network[0].Test(x);
|
||||||
for (int i = 1; i < network.size(); i++) {
|
for (int i = 1; i < network.size(); i++) {
|
||||||
@ -53,13 +53,13 @@ double MLPPANN::modelTest(std::vector<double> x) {
|
|||||||
return outputLayer->a_test;
|
return outputLayer->a_test;
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPANN::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
void MLPPANN::gradientDescent(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
double initial_learning_rate = learning_rate;
|
real_t initial_learning_rate = learning_rate;
|
||||||
|
|
||||||
alg.printMatrix(network[network.size() - 1].weights);
|
alg.printMatrix(network[network.size() - 1].weights);
|
||||||
while (true) {
|
while (true) {
|
||||||
@ -87,13 +87,13 @@ void MLPPANN::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPANN::SGD(double learning_rate, int max_epoch, bool UI) {
|
void MLPPANN::SGD(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
double initial_learning_rate = learning_rate;
|
real_t initial_learning_rate = learning_rate;
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
||||||
@ -103,7 +103,7 @@ void MLPPANN::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
||||||
int outputIndex = distribution(generator);
|
int outputIndex = distribution(generator);
|
||||||
|
|
||||||
std::vector<double> y_hat = modelSetTest({ inputSet[outputIndex] });
|
std::vector<real_t> y_hat = modelSetTest({ inputSet[outputIndex] });
|
||||||
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
||||||
|
|
||||||
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, { outputSet[outputIndex] });
|
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, { outputSet[outputIndex] });
|
||||||
@ -125,13 +125,13 @@ void MLPPANN::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPANN::MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
void MLPPANN::MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
double initial_learning_rate = learning_rate;
|
real_t initial_learning_rate = learning_rate;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
int n_mini_batch = n / mini_batch_size;
|
int n_mini_batch = n / mini_batch_size;
|
||||||
@ -141,7 +141,7 @@ void MLPPANN::MBGD(double learning_rate, int max_epoch, int mini_batch_size, boo
|
|||||||
while (true) {
|
while (true) {
|
||||||
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = modelSetTest(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = modelSetTest(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, outputMiniBatches[i]);
|
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, outputMiniBatches[i]);
|
||||||
@ -163,13 +163,13 @@ void MLPPANN::MBGD(double learning_rate, int max_epoch, int mini_batch_size, boo
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPANN::Momentum(double learning_rate, int max_epoch, int mini_batch_size, double gamma, bool NAG, bool UI) {
|
void MLPPANN::Momentum(real_t learning_rate, int max_epoch, int mini_batch_size, real_t gamma, bool NAG, bool UI) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
double initial_learning_rate = learning_rate;
|
real_t initial_learning_rate = learning_rate;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
int n_mini_batch = n / mini_batch_size;
|
int n_mini_batch = n / mini_batch_size;
|
||||||
@ -178,13 +178,13 @@ void MLPPANN::Momentum(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
||||||
|
|
||||||
// Initializing necessary components for Adam.
|
// Initializing necessary components for Adam.
|
||||||
std::vector<std::vector<std::vector<double>>> v_hidden;
|
std::vector<std::vector<std::vector<real_t>>> v_hidden;
|
||||||
|
|
||||||
std::vector<double> v_output;
|
std::vector<real_t> v_output;
|
||||||
while (true) {
|
while (true) {
|
||||||
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = modelSetTest(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = modelSetTest(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, outputMiniBatches[i]);
|
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, outputMiniBatches[i]);
|
||||||
@ -220,13 +220,13 @@ void MLPPANN::Momentum(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPANN::Adagrad(double learning_rate, int max_epoch, int mini_batch_size, double e, bool UI) {
|
void MLPPANN::Adagrad(real_t learning_rate, int max_epoch, int mini_batch_size, real_t e, bool UI) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
double initial_learning_rate = learning_rate;
|
real_t initial_learning_rate = learning_rate;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
int n_mini_batch = n / mini_batch_size;
|
int n_mini_batch = n / mini_batch_size;
|
||||||
@ -235,13 +235,13 @@ void MLPPANN::Adagrad(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
||||||
|
|
||||||
// Initializing necessary components for Adam.
|
// Initializing necessary components for Adam.
|
||||||
std::vector<std::vector<std::vector<double>>> v_hidden;
|
std::vector<std::vector<std::vector<real_t>>> v_hidden;
|
||||||
|
|
||||||
std::vector<double> v_output;
|
std::vector<real_t> v_output;
|
||||||
while (true) {
|
while (true) {
|
||||||
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = modelSetTest(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = modelSetTest(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, outputMiniBatches[i]);
|
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, outputMiniBatches[i]);
|
||||||
@ -258,8 +258,8 @@ void MLPPANN::Adagrad(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
|
|
||||||
v_output = alg.addition(v_output, alg.exponentiate(outputWGrad, 2));
|
v_output = alg.addition(v_output, alg.exponentiate(outputWGrad, 2));
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> hiddenLayerUpdations = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(cumulativeHiddenLayerWGrad, alg.scalarAdd(e, alg.sqrt(v_hidden))));
|
std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(cumulativeHiddenLayerWGrad, alg.scalarAdd(e, alg.sqrt(v_hidden))));
|
||||||
std::vector<double> outputLayerUpdation = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(outputWGrad, alg.scalarAdd(e, alg.sqrt(v_output))));
|
std::vector<real_t> outputLayerUpdation = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(outputWGrad, alg.scalarAdd(e, alg.sqrt(v_output))));
|
||||||
|
|
||||||
updateParameters(hiddenLayerUpdations, outputLayerUpdation, learning_rate); // subject to change. may want bias to have this matrix too.
|
updateParameters(hiddenLayerUpdations, outputLayerUpdation, learning_rate); // subject to change. may want bias to have this matrix too.
|
||||||
y_hat = modelSetTest(inputMiniBatches[i]);
|
y_hat = modelSetTest(inputMiniBatches[i]);
|
||||||
@ -276,13 +276,13 @@ void MLPPANN::Adagrad(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPANN::Adadelta(double learning_rate, int max_epoch, int mini_batch_size, double b1, double e, bool UI) {
|
void MLPPANN::Adadelta(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t e, bool UI) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
double initial_learning_rate = learning_rate;
|
real_t initial_learning_rate = learning_rate;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
int n_mini_batch = n / mini_batch_size;
|
int n_mini_batch = n / mini_batch_size;
|
||||||
@ -291,13 +291,13 @@ void MLPPANN::Adadelta(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
||||||
|
|
||||||
// Initializing necessary components for Adam.
|
// Initializing necessary components for Adam.
|
||||||
std::vector<std::vector<std::vector<double>>> v_hidden;
|
std::vector<std::vector<std::vector<real_t>>> v_hidden;
|
||||||
|
|
||||||
std::vector<double> v_output;
|
std::vector<real_t> v_output;
|
||||||
while (true) {
|
while (true) {
|
||||||
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = modelSetTest(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = modelSetTest(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, outputMiniBatches[i]);
|
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, outputMiniBatches[i]);
|
||||||
@ -314,8 +314,8 @@ void MLPPANN::Adadelta(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
|
|
||||||
v_output = alg.addition(v_output, alg.exponentiate(outputWGrad, 2));
|
v_output = alg.addition(v_output, alg.exponentiate(outputWGrad, 2));
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> hiddenLayerUpdations = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(cumulativeHiddenLayerWGrad, alg.scalarAdd(e, alg.sqrt(v_hidden))));
|
std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(cumulativeHiddenLayerWGrad, alg.scalarAdd(e, alg.sqrt(v_hidden))));
|
||||||
std::vector<double> outputLayerUpdation = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(outputWGrad, alg.scalarAdd(e, alg.sqrt(v_output))));
|
std::vector<real_t> outputLayerUpdation = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(outputWGrad, alg.scalarAdd(e, alg.sqrt(v_output))));
|
||||||
|
|
||||||
updateParameters(hiddenLayerUpdations, outputLayerUpdation, learning_rate); // subject to change. may want bias to have this matrix too.
|
updateParameters(hiddenLayerUpdations, outputLayerUpdation, learning_rate); // subject to change. may want bias to have this matrix too.
|
||||||
y_hat = modelSetTest(inputMiniBatches[i]);
|
y_hat = modelSetTest(inputMiniBatches[i]);
|
||||||
@ -332,13 +332,13 @@ void MLPPANN::Adadelta(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPANN::Adam(double learning_rate, int max_epoch, int mini_batch_size, double b1, double b2, double e, bool UI) {
|
void MLPPANN::Adam(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t b2, real_t e, bool UI) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
double initial_learning_rate = learning_rate;
|
real_t initial_learning_rate = learning_rate;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
int n_mini_batch = n / mini_batch_size;
|
int n_mini_batch = n / mini_batch_size;
|
||||||
@ -347,15 +347,15 @@ void MLPPANN::Adam(double learning_rate, int max_epoch, int mini_batch_size, dou
|
|||||||
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
||||||
|
|
||||||
// Initializing necessary components for Adam.
|
// Initializing necessary components for Adam.
|
||||||
std::vector<std::vector<std::vector<double>>> m_hidden;
|
std::vector<std::vector<std::vector<real_t>>> m_hidden;
|
||||||
std::vector<std::vector<std::vector<double>>> v_hidden;
|
std::vector<std::vector<std::vector<real_t>>> v_hidden;
|
||||||
|
|
||||||
std::vector<double> m_output;
|
std::vector<real_t> m_output;
|
||||||
std::vector<double> v_output;
|
std::vector<real_t> v_output;
|
||||||
while (true) {
|
while (true) {
|
||||||
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = modelSetTest(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = modelSetTest(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, outputMiniBatches[i]);
|
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, outputMiniBatches[i]);
|
||||||
@ -375,14 +375,14 @@ void MLPPANN::Adam(double learning_rate, int max_epoch, int mini_batch_size, dou
|
|||||||
m_output = alg.addition(alg.scalarMultiply(b1, m_output), alg.scalarMultiply(1 - b1, outputWGrad));
|
m_output = alg.addition(alg.scalarMultiply(b1, m_output), alg.scalarMultiply(1 - b1, outputWGrad));
|
||||||
v_output = alg.addition(alg.scalarMultiply(b2, v_output), alg.scalarMultiply(1 - b2, alg.exponentiate(outputWGrad, 2)));
|
v_output = alg.addition(alg.scalarMultiply(b2, v_output), alg.scalarMultiply(1 - b2, alg.exponentiate(outputWGrad, 2)));
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> m_hidden_hat = alg.scalarMultiply(1 / (1 - std::pow(b1, epoch)), m_hidden);
|
std::vector<std::vector<std::vector<real_t>>> m_hidden_hat = alg.scalarMultiply(1 / (1 - std::pow(b1, epoch)), m_hidden);
|
||||||
std::vector<std::vector<std::vector<double>>> v_hidden_hat = alg.scalarMultiply(1 / (1 - std::pow(b2, epoch)), v_hidden);
|
std::vector<std::vector<std::vector<real_t>>> v_hidden_hat = alg.scalarMultiply(1 / (1 - std::pow(b2, epoch)), v_hidden);
|
||||||
|
|
||||||
std::vector<double> m_output_hat = alg.scalarMultiply(1 / (1 - std::pow(b1, epoch)), m_output);
|
std::vector<real_t> m_output_hat = alg.scalarMultiply(1 / (1 - std::pow(b1, epoch)), m_output);
|
||||||
std::vector<double> v_output_hat = alg.scalarMultiply(1 / (1 - std::pow(b2, epoch)), v_output);
|
std::vector<real_t> v_output_hat = alg.scalarMultiply(1 / (1 - std::pow(b2, epoch)), v_output);
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> hiddenLayerUpdations = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(m_hidden_hat, alg.scalarAdd(e, alg.sqrt(v_hidden_hat))));
|
std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(m_hidden_hat, alg.scalarAdd(e, alg.sqrt(v_hidden_hat))));
|
||||||
std::vector<double> outputLayerUpdation = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(m_output_hat, alg.scalarAdd(e, alg.sqrt(v_output_hat))));
|
std::vector<real_t> outputLayerUpdation = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(m_output_hat, alg.scalarAdd(e, alg.sqrt(v_output_hat))));
|
||||||
|
|
||||||
updateParameters(hiddenLayerUpdations, outputLayerUpdation, learning_rate); // subject to change. may want bias to have this matrix too.
|
updateParameters(hiddenLayerUpdations, outputLayerUpdation, learning_rate); // subject to change. may want bias to have this matrix too.
|
||||||
y_hat = modelSetTest(inputMiniBatches[i]);
|
y_hat = modelSetTest(inputMiniBatches[i]);
|
||||||
@ -399,13 +399,13 @@ void MLPPANN::Adam(double learning_rate, int max_epoch, int mini_batch_size, dou
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPANN::Adamax(double learning_rate, int max_epoch, int mini_batch_size, double b1, double b2, double e, bool UI) {
|
void MLPPANN::Adamax(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t b2, real_t e, bool UI) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
double initial_learning_rate = learning_rate;
|
real_t initial_learning_rate = learning_rate;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
int n_mini_batch = n / mini_batch_size;
|
int n_mini_batch = n / mini_batch_size;
|
||||||
@ -414,15 +414,15 @@ void MLPPANN::Adamax(double learning_rate, int max_epoch, int mini_batch_size, d
|
|||||||
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
||||||
|
|
||||||
// Initializing necessary components for Adam.
|
// Initializing necessary components for Adam.
|
||||||
std::vector<std::vector<std::vector<double>>> m_hidden;
|
std::vector<std::vector<std::vector<real_t>>> m_hidden;
|
||||||
std::vector<std::vector<std::vector<double>>> u_hidden;
|
std::vector<std::vector<std::vector<real_t>>> u_hidden;
|
||||||
|
|
||||||
std::vector<double> m_output;
|
std::vector<real_t> m_output;
|
||||||
std::vector<double> u_output;
|
std::vector<real_t> u_output;
|
||||||
while (true) {
|
while (true) {
|
||||||
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = modelSetTest(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = modelSetTest(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, outputMiniBatches[i]);
|
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, outputMiniBatches[i]);
|
||||||
@ -442,12 +442,12 @@ void MLPPANN::Adamax(double learning_rate, int max_epoch, int mini_batch_size, d
|
|||||||
m_output = alg.addition(alg.scalarMultiply(b1, m_output), alg.scalarMultiply(1 - b1, outputWGrad));
|
m_output = alg.addition(alg.scalarMultiply(b1, m_output), alg.scalarMultiply(1 - b1, outputWGrad));
|
||||||
u_output = alg.max(alg.scalarMultiply(b2, u_output), alg.abs(outputWGrad));
|
u_output = alg.max(alg.scalarMultiply(b2, u_output), alg.abs(outputWGrad));
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> m_hidden_hat = alg.scalarMultiply(1 / (1 - std::pow(b1, epoch)), m_hidden);
|
std::vector<std::vector<std::vector<real_t>>> m_hidden_hat = alg.scalarMultiply(1 / (1 - std::pow(b1, epoch)), m_hidden);
|
||||||
|
|
||||||
std::vector<double> m_output_hat = alg.scalarMultiply(1 / (1 - std::pow(b1, epoch)), m_output);
|
std::vector<real_t> m_output_hat = alg.scalarMultiply(1 / (1 - std::pow(b1, epoch)), m_output);
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> hiddenLayerUpdations = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(m_hidden_hat, alg.scalarAdd(e, u_hidden)));
|
std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(m_hidden_hat, alg.scalarAdd(e, u_hidden)));
|
||||||
std::vector<double> outputLayerUpdation = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(m_output_hat, alg.scalarAdd(e, u_output)));
|
std::vector<real_t> outputLayerUpdation = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(m_output_hat, alg.scalarAdd(e, u_output)));
|
||||||
|
|
||||||
updateParameters(hiddenLayerUpdations, outputLayerUpdation, learning_rate); // subject to change. may want bias to have this matrix too.
|
updateParameters(hiddenLayerUpdations, outputLayerUpdation, learning_rate); // subject to change. may want bias to have this matrix too.
|
||||||
y_hat = modelSetTest(inputMiniBatches[i]);
|
y_hat = modelSetTest(inputMiniBatches[i]);
|
||||||
@ -464,13 +464,13 @@ void MLPPANN::Adamax(double learning_rate, int max_epoch, int mini_batch_size, d
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPANN::Nadam(double learning_rate, int max_epoch, int mini_batch_size, double b1, double b2, double e, bool UI) {
|
void MLPPANN::Nadam(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t b2, real_t e, bool UI) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
double initial_learning_rate = learning_rate;
|
real_t initial_learning_rate = learning_rate;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
int n_mini_batch = n / mini_batch_size;
|
int n_mini_batch = n / mini_batch_size;
|
||||||
@ -479,16 +479,16 @@ void MLPPANN::Nadam(double learning_rate, int max_epoch, int mini_batch_size, do
|
|||||||
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
||||||
|
|
||||||
// Initializing necessary components for Adam.
|
// Initializing necessary components for Adam.
|
||||||
std::vector<std::vector<std::vector<double>>> m_hidden;
|
std::vector<std::vector<std::vector<real_t>>> m_hidden;
|
||||||
std::vector<std::vector<std::vector<double>>> v_hidden;
|
std::vector<std::vector<std::vector<real_t>>> v_hidden;
|
||||||
std::vector<std::vector<std::vector<double>>> m_hidden_final;
|
std::vector<std::vector<std::vector<real_t>>> m_hidden_final;
|
||||||
|
|
||||||
std::vector<double> m_output;
|
std::vector<real_t> m_output;
|
||||||
std::vector<double> v_output;
|
std::vector<real_t> v_output;
|
||||||
while (true) {
|
while (true) {
|
||||||
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = modelSetTest(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = modelSetTest(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, outputMiniBatches[i]);
|
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, outputMiniBatches[i]);
|
||||||
@ -508,16 +508,16 @@ void MLPPANN::Nadam(double learning_rate, int max_epoch, int mini_batch_size, do
|
|||||||
m_output = alg.addition(alg.scalarMultiply(b1, m_output), alg.scalarMultiply(1 - b1, outputWGrad));
|
m_output = alg.addition(alg.scalarMultiply(b1, m_output), alg.scalarMultiply(1 - b1, outputWGrad));
|
||||||
v_output = alg.addition(alg.scalarMultiply(b2, v_output), alg.scalarMultiply(1 - b2, alg.exponentiate(outputWGrad, 2)));
|
v_output = alg.addition(alg.scalarMultiply(b2, v_output), alg.scalarMultiply(1 - b2, alg.exponentiate(outputWGrad, 2)));
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> m_hidden_hat = alg.scalarMultiply(1 / (1 - std::pow(b1, epoch)), m_hidden);
|
std::vector<std::vector<std::vector<real_t>>> m_hidden_hat = alg.scalarMultiply(1 / (1 - std::pow(b1, epoch)), m_hidden);
|
||||||
std::vector<std::vector<std::vector<double>>> v_hidden_hat = alg.scalarMultiply(1 / (1 - std::pow(b2, epoch)), v_hidden);
|
std::vector<std::vector<std::vector<real_t>>> v_hidden_hat = alg.scalarMultiply(1 / (1 - std::pow(b2, epoch)), v_hidden);
|
||||||
std::vector<std::vector<std::vector<double>>> m_hidden_final = alg.addition(alg.scalarMultiply(b1, m_hidden_hat), alg.scalarMultiply((1 - b1) / (1 - std::pow(b1, epoch)), cumulativeHiddenLayerWGrad));
|
std::vector<std::vector<std::vector<real_t>>> m_hidden_final = alg.addition(alg.scalarMultiply(b1, m_hidden_hat), alg.scalarMultiply((1 - b1) / (1 - std::pow(b1, epoch)), cumulativeHiddenLayerWGrad));
|
||||||
|
|
||||||
std::vector<double> m_output_hat = alg.scalarMultiply(1 / (1 - std::pow(b1, epoch)), m_output);
|
std::vector<real_t> m_output_hat = alg.scalarMultiply(1 / (1 - std::pow(b1, epoch)), m_output);
|
||||||
std::vector<double> v_output_hat = alg.scalarMultiply(1 / (1 - std::pow(b2, epoch)), v_output);
|
std::vector<real_t> v_output_hat = alg.scalarMultiply(1 / (1 - std::pow(b2, epoch)), v_output);
|
||||||
std::vector<double> m_output_final = alg.addition(alg.scalarMultiply(b1, m_output_hat), alg.scalarMultiply((1 - b1) / (1 - std::pow(b1, epoch)), outputWGrad));
|
std::vector<real_t> m_output_final = alg.addition(alg.scalarMultiply(b1, m_output_hat), alg.scalarMultiply((1 - b1) / (1 - std::pow(b1, epoch)), outputWGrad));
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> hiddenLayerUpdations = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(m_hidden_final, alg.scalarAdd(e, alg.sqrt(v_hidden_hat))));
|
std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(m_hidden_final, alg.scalarAdd(e, alg.sqrt(v_hidden_hat))));
|
||||||
std::vector<double> outputLayerUpdation = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(m_output_final, alg.scalarAdd(e, alg.sqrt(v_output_hat))));
|
std::vector<real_t> outputLayerUpdation = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(m_output_final, alg.scalarAdd(e, alg.sqrt(v_output_hat))));
|
||||||
|
|
||||||
updateParameters(hiddenLayerUpdations, outputLayerUpdation, learning_rate); // subject to change. may want bias to have this matrix too.
|
updateParameters(hiddenLayerUpdations, outputLayerUpdation, learning_rate); // subject to change. may want bias to have this matrix too.
|
||||||
y_hat = modelSetTest(inputMiniBatches[i]);
|
y_hat = modelSetTest(inputMiniBatches[i]);
|
||||||
@ -534,13 +534,13 @@ void MLPPANN::Nadam(double learning_rate, int max_epoch, int mini_batch_size, do
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPANN::AMSGrad(double learning_rate, int max_epoch, int mini_batch_size, double b1, double b2, double e, bool UI) {
|
void MLPPANN::AMSGrad(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t b2, real_t e, bool UI) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
double initial_learning_rate = learning_rate;
|
real_t initial_learning_rate = learning_rate;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
int n_mini_batch = n / mini_batch_size;
|
int n_mini_batch = n / mini_batch_size;
|
||||||
@ -549,19 +549,19 @@ void MLPPANN::AMSGrad(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
||||||
|
|
||||||
// Initializing necessary components for Adam.
|
// Initializing necessary components for Adam.
|
||||||
std::vector<std::vector<std::vector<double>>> m_hidden;
|
std::vector<std::vector<std::vector<real_t>>> m_hidden;
|
||||||
std::vector<std::vector<std::vector<double>>> v_hidden;
|
std::vector<std::vector<std::vector<real_t>>> v_hidden;
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> v_hidden_hat;
|
std::vector<std::vector<std::vector<real_t>>> v_hidden_hat;
|
||||||
|
|
||||||
std::vector<double> m_output;
|
std::vector<real_t> m_output;
|
||||||
std::vector<double> v_output;
|
std::vector<real_t> v_output;
|
||||||
|
|
||||||
std::vector<double> v_output_hat;
|
std::vector<real_t> v_output_hat;
|
||||||
while (true) {
|
while (true) {
|
||||||
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
learning_rate = applyLearningRateScheduler(initial_learning_rate, decayConstant, epoch, dropRate);
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = modelSetTest(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = modelSetTest(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, outputMiniBatches[i]);
|
auto [cumulativeHiddenLayerWGrad, outputWGrad] = computeGradients(y_hat, outputMiniBatches[i]);
|
||||||
@ -587,8 +587,8 @@ void MLPPANN::AMSGrad(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
|
|
||||||
v_output_hat = alg.max(v_output_hat, v_output);
|
v_output_hat = alg.max(v_output_hat, v_output);
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> hiddenLayerUpdations = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(m_hidden, alg.scalarAdd(e, alg.sqrt(v_hidden_hat))));
|
std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(m_hidden, alg.scalarAdd(e, alg.sqrt(v_hidden_hat))));
|
||||||
std::vector<double> outputLayerUpdation = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(m_output, alg.scalarAdd(e, alg.sqrt(v_output_hat))));
|
std::vector<real_t> outputLayerUpdation = alg.scalarMultiply(learning_rate / n, alg.elementWiseDivision(m_output, alg.scalarAdd(e, alg.sqrt(v_output_hat))));
|
||||||
|
|
||||||
updateParameters(hiddenLayerUpdations, outputLayerUpdation, learning_rate); // subject to change. may want bias to have this matrix too.
|
updateParameters(hiddenLayerUpdations, outputLayerUpdation, learning_rate); // subject to change. may want bias to have this matrix too.
|
||||||
y_hat = modelSetTest(inputMiniBatches[i]);
|
y_hat = modelSetTest(inputMiniBatches[i]);
|
||||||
@ -605,7 +605,7 @@ void MLPPANN::AMSGrad(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPANN::score() {
|
real_t MLPPANN::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
return util.performance(y_hat, outputSet);
|
return util.performance(y_hat, outputSet);
|
||||||
@ -624,12 +624,12 @@ void MLPPANN::save(std::string fileName) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPANN::setLearningRateScheduler(std::string type, double decayConstant) {
|
void MLPPANN::setLearningRateScheduler(std::string type, real_t decayConstant) {
|
||||||
lrScheduler = type;
|
lrScheduler = type;
|
||||||
MLPPANN::decayConstant = decayConstant;
|
MLPPANN::decayConstant = decayConstant;
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPANN::setLearningRateScheduler(std::string type, double decayConstant, double dropRate) {
|
void MLPPANN::setLearningRateScheduler(std::string type, real_t decayConstant, real_t dropRate) {
|
||||||
lrScheduler = type;
|
lrScheduler = type;
|
||||||
MLPPANN::decayConstant = decayConstant;
|
MLPPANN::decayConstant = decayConstant;
|
||||||
MLPPANN::dropRate = dropRate;
|
MLPPANN::dropRate = dropRate;
|
||||||
@ -637,7 +637,7 @@ void MLPPANN::setLearningRateScheduler(std::string type, double decayConstant, d
|
|||||||
|
|
||||||
// https://en.wikipedia.org/wiki/Learning_rate
|
// https://en.wikipedia.org/wiki/Learning_rate
|
||||||
// Learning Rate Decay (C2W2L09) - Andrew Ng - Deep Learning Specialization
|
// Learning Rate Decay (C2W2L09) - Andrew Ng - Deep Learning Specialization
|
||||||
double MLPPANN::applyLearningRateScheduler(double learningRate, double decayConstant, double epoch, double dropRate) {
|
real_t MLPPANN::applyLearningRateScheduler(real_t learningRate, real_t decayConstant, real_t epoch, real_t dropRate) {
|
||||||
if (lrScheduler == "Time") {
|
if (lrScheduler == "Time") {
|
||||||
return learningRate / (1 + decayConstant * epoch);
|
return learningRate / (1 + decayConstant * epoch);
|
||||||
} else if (lrScheduler == "Epoch") {
|
} else if (lrScheduler == "Epoch") {
|
||||||
@ -650,7 +650,7 @@ double MLPPANN::applyLearningRateScheduler(double learningRate, double decayCons
|
|||||||
return learningRate;
|
return learningRate;
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPANN::addLayer(int n_hidden, std::string activation, std::string weightInit, std::string reg, double lambda, double alpha) {
|
void MLPPANN::addLayer(int n_hidden, std::string activation, std::string weightInit, std::string reg, real_t lambda, real_t alpha) {
|
||||||
if (network.empty()) {
|
if (network.empty()) {
|
||||||
network.push_back(MLPPHiddenLayer(n_hidden, activation, inputSet, weightInit, reg, lambda, alpha));
|
network.push_back(MLPPHiddenLayer(n_hidden, activation, inputSet, weightInit, reg, lambda, alpha));
|
||||||
network[0].forwardPass();
|
network[0].forwardPass();
|
||||||
@ -660,7 +660,7 @@ void MLPPANN::addLayer(int n_hidden, std::string activation, std::string weightI
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPANN::addOutputLayer(std::string activation, std::string loss, std::string weightInit, std::string reg, double lambda, double alpha) {
|
void MLPPANN::addOutputLayer(std::string activation, std::string loss, std::string weightInit, std::string reg, real_t lambda, real_t alpha) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
outputLayer = new MLPPOutputLayer(network[network.size() - 1].n_hidden, activation, loss, network[network.size() - 1].a, weightInit, reg, lambda, alpha);
|
outputLayer = new MLPPOutputLayer(network[network.size() - 1].n_hidden, activation, loss, network[network.size() - 1].a, weightInit, reg, lambda, alpha);
|
||||||
@ -669,10 +669,10 @@ void MLPPANN::addOutputLayer(std::string activation, std::string loss, std::stri
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPANN::Cost(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPANN::Cost(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
double totalRegTerm = 0;
|
real_t totalRegTerm = 0;
|
||||||
|
|
||||||
auto cost_function = outputLayer->cost_map[outputLayer->cost];
|
auto cost_function = outputLayer->cost_map[outputLayer->cost];
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
@ -700,7 +700,7 @@ void MLPPANN::forwardPass() {
|
|||||||
y_hat = outputLayer->a;
|
y_hat = outputLayer->a;
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPANN::updateParameters(std::vector<std::vector<std::vector<double>>> hiddenLayerUpdations, std::vector<double> outputLayerUpdation, double learning_rate) {
|
void MLPPANN::updateParameters(std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations, std::vector<real_t> outputLayerUpdation, real_t learning_rate) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
outputLayer->weights = alg.subtraction(outputLayer->weights, outputLayerUpdation);
|
outputLayer->weights = alg.subtraction(outputLayer->weights, outputLayerUpdation);
|
||||||
@ -717,39 +717,39 @@ void MLPPANN::updateParameters(std::vector<std::vector<std::vector<double>>> hid
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<double>> MLPPANN::computeGradients(std::vector<double> y_hat, std::vector<double> outputSet) {
|
std::tuple<std::vector<std::vector<std::vector<real_t>>>, std::vector<real_t>> MLPPANN::computeGradients(std::vector<real_t> y_hat, std::vector<real_t> outputSet) {
|
||||||
// std::cout << "BEGIN" << std::endl;
|
// std::cout << "BEGIN" << std::endl;
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> cumulativeHiddenLayerWGrad; // Tensor containing ALL hidden grads.
|
std::vector<std::vector<std::vector<real_t>>> cumulativeHiddenLayerWGrad; // Tensor containing ALL hidden grads.
|
||||||
|
|
||||||
auto costDeriv = outputLayer->costDeriv_map[outputLayer->cost];
|
auto costDeriv = outputLayer->costDeriv_map[outputLayer->cost];
|
||||||
auto outputAvn = outputLayer->activation_map[outputLayer->activation];
|
auto outputAvn = outputLayer->activation_map[outputLayer->activation];
|
||||||
outputLayer->delta = alg.hadamard_product((cost.*costDeriv)(y_hat, outputSet), (avn.*outputAvn)(outputLayer->z, 1));
|
outputLayer->delta = alg.hadamard_product((cost.*costDeriv)(y_hat, outputSet), (avn.*outputAvn)(outputLayer->z, 1));
|
||||||
std::vector<double> outputWGrad = alg.mat_vec_mult(alg.transpose(outputLayer->input), outputLayer->delta);
|
std::vector<real_t> outputWGrad = alg.mat_vec_mult(alg.transpose(outputLayer->input), outputLayer->delta);
|
||||||
outputWGrad = alg.addition(outputWGrad, regularization.regDerivTerm(outputLayer->weights, outputLayer->lambda, outputLayer->alpha, outputLayer->reg));
|
outputWGrad = alg.addition(outputWGrad, regularization.regDerivTerm(outputLayer->weights, outputLayer->lambda, outputLayer->alpha, outputLayer->reg));
|
||||||
|
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
auto hiddenLayerAvn = network[network.size() - 1].activation_map[network[network.size() - 1].activation];
|
auto hiddenLayerAvn = network[network.size() - 1].activation_map[network[network.size() - 1].activation];
|
||||||
network[network.size() - 1].delta = alg.hadamard_product(alg.outerProduct(outputLayer->delta, outputLayer->weights), (avn.*hiddenLayerAvn)(network[network.size() - 1].z, 1));
|
network[network.size() - 1].delta = alg.hadamard_product(alg.outerProduct(outputLayer->delta, outputLayer->weights), (avn.*hiddenLayerAvn)(network[network.size() - 1].z, 1));
|
||||||
std::vector<std::vector<double>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[network.size() - 1].input), network[network.size() - 1].delta);
|
std::vector<std::vector<real_t>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[network.size() - 1].input), network[network.size() - 1].delta);
|
||||||
|
|
||||||
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[network.size() - 1].weights, network[network.size() - 1].lambda, network[network.size() - 1].alpha, network[network.size() - 1].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[network.size() - 1].weights, network[network.size() - 1].lambda, network[network.size() - 1].alpha, network[network.size() - 1].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
||||||
|
|
||||||
for (int i = network.size() - 2; i >= 0; i--) {
|
for (int i = network.size() - 2; i >= 0; i--) {
|
||||||
auto hiddenLayerAvn = network[i].activation_map[network[i].activation];
|
auto hiddenLayerAvn = network[i].activation_map[network[i].activation];
|
||||||
network[i].delta = alg.hadamard_product(alg.matmult(network[i + 1].delta, alg.transpose(network[i + 1].weights)), (avn.*hiddenLayerAvn)(network[i].z, 1));
|
network[i].delta = alg.hadamard_product(alg.matmult(network[i + 1].delta, alg.transpose(network[i + 1].weights)), (avn.*hiddenLayerAvn)(network[i].z, 1));
|
||||||
std::vector<std::vector<double>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[i].input), network[i].delta);
|
std::vector<std::vector<real_t>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[i].input), network[i].delta);
|
||||||
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[i].weights, network[i].lambda, network[i].alpha, network[i].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[i].weights, network[i].lambda, network[i].alpha, network[i].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return { cumulativeHiddenLayerWGrad, outputWGrad };
|
return { cumulativeHiddenLayerWGrad, outputWGrad };
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPANN::UI(int epoch, double cost_prev, std::vector<double> y_hat, std::vector<double> outputSet) {
|
void MLPPANN::UI(int epoch, real_t cost_prev, std::vector<real_t> y_hat, std::vector<real_t> outputSet) {
|
||||||
MLPPUtilities::CostInfo(epoch, cost_prev, Cost(y_hat, outputSet));
|
MLPPUtilities::CostInfo(epoch, cost_prev, Cost(y_hat, outputSet));
|
||||||
std::cout << "Layer " << network.size() + 1 << ": " << std::endl;
|
std::cout << "Layer " << network.size() + 1 << ": " << std::endl;
|
||||||
MLPPUtilities::UI(outputLayer->weights, outputLayer->bias);
|
MLPPUtilities::UI(outputLayer->weights, outputLayer->bias);
|
||||||
|
@ -7,6 +7,8 @@
|
|||||||
// Created by Marc Melikyan on 11/4/20.
|
// Created by Marc Melikyan on 11/4/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include "../hidden_layer/hidden_layer.h"
|
#include "../hidden_layer/hidden_layer.h"
|
||||||
#include "../output_layer/output_layer.h"
|
#include "../output_layer/output_layer.h"
|
||||||
|
|
||||||
@ -16,43 +18,43 @@
|
|||||||
|
|
||||||
class MLPPANN {
|
class MLPPANN {
|
||||||
public:
|
public:
|
||||||
MLPPANN(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet);
|
MLPPANN(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet);
|
||||||
~MLPPANN();
|
~MLPPANN();
|
||||||
std::vector<double> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<real_t> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
double modelTest(std::vector<double> x);
|
real_t modelTest(std::vector<real_t> x);
|
||||||
void gradientDescent(double learning_rate, int max_epoch, bool UI = 1);
|
void gradientDescent(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void SGD(double learning_rate, int max_epoch, bool UI = 1);
|
void SGD(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
void MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
||||||
void Momentum(double learning_rate, int max_epoch, int mini_batch_size, double gamma, bool NAG, bool UI = 1);
|
void Momentum(real_t learning_rate, int max_epoch, int mini_batch_size, real_t gamma, bool NAG, bool UI = 1);
|
||||||
void Adagrad(double learning_rate, int max_epoch, int mini_batch_size, double e, bool UI = 1);
|
void Adagrad(real_t learning_rate, int max_epoch, int mini_batch_size, real_t e, bool UI = 1);
|
||||||
void Adadelta(double learning_rate, int max_epoch, int mini_batch_size, double b1, double e, bool UI = 1);
|
void Adadelta(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t e, bool UI = 1);
|
||||||
void Adam(double learning_rate, int max_epoch, int mini_batch_size, double b1, double b2, double e, bool UI = 1);
|
void Adam(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t b2, real_t e, bool UI = 1);
|
||||||
void Adamax(double learning_rate, int max_epoch, int mini_batch_size, double b1, double b2, double e, bool UI = 1);
|
void Adamax(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t b2, real_t e, bool UI = 1);
|
||||||
void Nadam(double learning_rate, int max_epoch, int mini_batch_size, double b1, double b2, double e, bool UI = 1);
|
void Nadam(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t b2, real_t e, bool UI = 1);
|
||||||
void AMSGrad(double learning_rate, int max_epoch, int mini_batch_size, double b1, double b2, double e, bool UI = 1);
|
void AMSGrad(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t b2, real_t e, bool UI = 1);
|
||||||
double score();
|
real_t score();
|
||||||
void save(std::string fileName);
|
void save(std::string fileName);
|
||||||
|
|
||||||
void setLearningRateScheduler(std::string type, double decayConstant);
|
void setLearningRateScheduler(std::string type, real_t decayConstant);
|
||||||
void setLearningRateScheduler(std::string type, double decayConstant, double dropRate);
|
void setLearningRateScheduler(std::string type, real_t decayConstant, real_t dropRate);
|
||||||
|
|
||||||
void addLayer(int n_hidden, std::string activation, std::string weightInit = "Default", std::string reg = "None", double lambda = 0.5, double alpha = 0.5);
|
void addLayer(int n_hidden, std::string activation, std::string weightInit = "Default", std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
||||||
void addOutputLayer(std::string activation, std::string loss, std::string weightInit = "Default", std::string reg = "None", double lambda = 0.5, double alpha = 0.5);
|
void addOutputLayer(std::string activation, std::string loss, std::string weightInit = "Default", std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
double applyLearningRateScheduler(double learningRate, double decayConstant, double epoch, double dropRate);
|
real_t applyLearningRateScheduler(real_t learningRate, real_t decayConstant, real_t epoch, real_t dropRate);
|
||||||
|
|
||||||
double Cost(std::vector<double> y_hat, std::vector<double> y);
|
real_t Cost(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
|
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
void updateParameters(std::vector<std::vector<std::vector<double>>> hiddenLayerUpdations, std::vector<double> outputLayerUpdation, double learning_rate);
|
void updateParameters(std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations, std::vector<real_t> outputLayerUpdation, real_t learning_rate);
|
||||||
std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<double>> computeGradients(std::vector<double> y_hat, std::vector<double> outputSet);
|
std::tuple<std::vector<std::vector<std::vector<real_t>>>, std::vector<real_t>> computeGradients(std::vector<real_t> y_hat, std::vector<real_t> outputSet);
|
||||||
|
|
||||||
void UI(int epoch, double cost_prev, std::vector<double> y_hat, std::vector<double> outputSet);
|
void UI(int epoch, real_t cost_prev, std::vector<real_t> y_hat, std::vector<real_t> outputSet);
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
|
|
||||||
std::vector<MLPPHiddenLayer> network;
|
std::vector<MLPPHiddenLayer> network;
|
||||||
MLPPOutputLayer *outputLayer;
|
MLPPOutputLayer *outputLayer;
|
||||||
@ -61,8 +63,8 @@ private:
|
|||||||
int k;
|
int k;
|
||||||
|
|
||||||
std::string lrScheduler;
|
std::string lrScheduler;
|
||||||
double decayConstant;
|
real_t decayConstant;
|
||||||
double dropRate;
|
real_t dropRate;
|
||||||
};
|
};
|
||||||
|
|
||||||
#endif /* ANN_hpp */
|
#endif /* ANN_hpp */
|
@ -13,7 +13,7 @@
|
|||||||
#include <iostream>
|
#include <iostream>
|
||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
MLPPAutoEncoder::MLPPAutoEncoder(std::vector<std::vector<double>> inputSet, int n_hidden) :
|
MLPPAutoEncoder::MLPPAutoEncoder(std::vector<std::vector<real_t>> inputSet, int n_hidden) :
|
||||||
inputSet(inputSet), n_hidden(n_hidden), n(inputSet.size()), k(inputSet[0].size()) {
|
inputSet(inputSet), n_hidden(n_hidden), n(inputSet.size()), k(inputSet[0].size()) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
y_hat.resize(inputSet.size());
|
y_hat.resize(inputSet.size());
|
||||||
@ -24,18 +24,18 @@ MLPPAutoEncoder::MLPPAutoEncoder(std::vector<std::vector<double>> inputSet, int
|
|||||||
bias2 = MLPPUtilities::biasInitialization(k);
|
bias2 = MLPPUtilities::biasInitialization(k);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPAutoEncoder::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<std::vector<real_t>> MLPPAutoEncoder::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
return Evaluate(X);
|
return Evaluate(X);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPAutoEncoder::modelTest(std::vector<double> x) {
|
std::vector<real_t> MLPPAutoEncoder::modelTest(std::vector<real_t> x) {
|
||||||
return Evaluate(x);
|
return Evaluate(x);
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPAutoEncoder::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
void MLPPAutoEncoder::gradientDescent(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
@ -43,10 +43,10 @@ void MLPPAutoEncoder::gradientDescent(double learning_rate, int max_epoch, bool
|
|||||||
cost_prev = Cost(y_hat, inputSet);
|
cost_prev = Cost(y_hat, inputSet);
|
||||||
|
|
||||||
// Calculating the errors
|
// Calculating the errors
|
||||||
std::vector<std::vector<double>> error = alg.subtraction(y_hat, inputSet);
|
std::vector<std::vector<real_t>> error = alg.subtraction(y_hat, inputSet);
|
||||||
|
|
||||||
// Calculating the weight/bias gradients for layer 2
|
// Calculating the weight/bias gradients for layer 2
|
||||||
std::vector<std::vector<double>> D2_1 = alg.matmult(alg.transpose(a2), error);
|
std::vector<std::vector<real_t>> D2_1 = alg.matmult(alg.transpose(a2), error);
|
||||||
|
|
||||||
// weights and bias updation for layer 2
|
// weights and bias updation for layer 2
|
||||||
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate / n, D2_1));
|
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate / n, D2_1));
|
||||||
@ -56,11 +56,11 @@ void MLPPAutoEncoder::gradientDescent(double learning_rate, int max_epoch, bool
|
|||||||
|
|
||||||
//Calculating the weight/bias for layer 1
|
//Calculating the weight/bias for layer 1
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_1 = alg.matmult(error, alg.transpose(weights2));
|
std::vector<std::vector<real_t>> D1_1 = alg.matmult(error, alg.transpose(weights2));
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
std::vector<std::vector<real_t>> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_3 = alg.matmult(alg.transpose(inputSet), D1_2);
|
std::vector<std::vector<real_t>> D1_3 = alg.matmult(alg.transpose(inputSet), D1_2);
|
||||||
|
|
||||||
// weight an bias updation for layer 1
|
// weight an bias updation for layer 1
|
||||||
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate / n, D1_3));
|
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate / n, D1_3));
|
||||||
@ -85,10 +85,10 @@ void MLPPAutoEncoder::gradientDescent(double learning_rate, int max_epoch, bool
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPAutoEncoder::SGD(double learning_rate, int max_epoch, bool UI) {
|
void MLPPAutoEncoder::SGD(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
@ -97,22 +97,22 @@ void MLPPAutoEncoder::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
||||||
int outputIndex = distribution(generator);
|
int outputIndex = distribution(generator);
|
||||||
|
|
||||||
std::vector<double> y_hat = Evaluate(inputSet[outputIndex]);
|
std::vector<real_t> y_hat = Evaluate(inputSet[outputIndex]);
|
||||||
auto [z2, a2] = propagate(inputSet[outputIndex]);
|
auto [z2, a2] = propagate(inputSet[outputIndex]);
|
||||||
cost_prev = Cost({ y_hat }, { inputSet[outputIndex] });
|
cost_prev = Cost({ y_hat }, { inputSet[outputIndex] });
|
||||||
std::vector<double> error = alg.subtraction(y_hat, inputSet[outputIndex]);
|
std::vector<real_t> error = alg.subtraction(y_hat, inputSet[outputIndex]);
|
||||||
|
|
||||||
// Weight updation for layer 2
|
// Weight updation for layer 2
|
||||||
std::vector<std::vector<double>> D2_1 = alg.outerProduct(error, a2);
|
std::vector<std::vector<real_t>> D2_1 = alg.outerProduct(error, a2);
|
||||||
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate, alg.transpose(D2_1)));
|
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate, alg.transpose(D2_1)));
|
||||||
|
|
||||||
// Bias updation for layer 2
|
// Bias updation for layer 2
|
||||||
bias2 = alg.subtraction(bias2, alg.scalarMultiply(learning_rate, error));
|
bias2 = alg.subtraction(bias2, alg.scalarMultiply(learning_rate, error));
|
||||||
|
|
||||||
// Weight updation for layer 1
|
// Weight updation for layer 1
|
||||||
std::vector<double> D1_1 = alg.mat_vec_mult(weights2, error);
|
std::vector<real_t> D1_1 = alg.mat_vec_mult(weights2, error);
|
||||||
std::vector<double> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
std::vector<real_t> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
||||||
std::vector<std::vector<double>> D1_3 = alg.outerProduct(inputSet[outputIndex], D1_2);
|
std::vector<std::vector<real_t>> D1_3 = alg.outerProduct(inputSet[outputIndex], D1_2);
|
||||||
|
|
||||||
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate, D1_3));
|
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate, D1_3));
|
||||||
// Bias updation for layer 1
|
// Bias updation for layer 1
|
||||||
@ -136,28 +136,28 @@ void MLPPAutoEncoder::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPAutoEncoder::MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
void MLPPAutoEncoder::MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
int n_mini_batch = n / mini_batch_size;
|
int n_mini_batch = n / mini_batch_size;
|
||||||
std::vector<std::vector<std::vector<double>>> inputMiniBatches = MLPPUtilities::createMiniBatches(inputSet, n_mini_batch);
|
std::vector<std::vector<std::vector<real_t>>> inputMiniBatches = MLPPUtilities::createMiniBatches(inputSet, n_mini_batch);
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<std::vector<double>> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<std::vector<real_t>> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
auto [z2, a2] = propagate(inputMiniBatches[i]);
|
auto [z2, a2] = propagate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, inputMiniBatches[i]);
|
cost_prev = Cost(y_hat, inputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the errors
|
// Calculating the errors
|
||||||
std::vector<std::vector<double>> error = alg.subtraction(y_hat, inputMiniBatches[i]);
|
std::vector<std::vector<real_t>> error = alg.subtraction(y_hat, inputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the weight/bias gradients for layer 2
|
// Calculating the weight/bias gradients for layer 2
|
||||||
|
|
||||||
std::vector<std::vector<double>> D2_1 = alg.matmult(alg.transpose(a2), error);
|
std::vector<std::vector<real_t>> D2_1 = alg.matmult(alg.transpose(a2), error);
|
||||||
|
|
||||||
// weights and bias updation for layer 2
|
// weights and bias updation for layer 2
|
||||||
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate / inputMiniBatches[i].size(), D2_1));
|
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate / inputMiniBatches[i].size(), D2_1));
|
||||||
@ -167,11 +167,11 @@ void MLPPAutoEncoder::MBGD(double learning_rate, int max_epoch, int mini_batch_s
|
|||||||
|
|
||||||
//Calculating the weight/bias for layer 1
|
//Calculating the weight/bias for layer 1
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_1 = alg.matmult(error, alg.transpose(weights2));
|
std::vector<std::vector<real_t>> D1_1 = alg.matmult(error, alg.transpose(weights2));
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
std::vector<std::vector<real_t>> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_3 = alg.matmult(alg.transpose(inputMiniBatches[i]), D1_2);
|
std::vector<std::vector<real_t>> D1_3 = alg.matmult(alg.transpose(inputMiniBatches[i]), D1_2);
|
||||||
|
|
||||||
// weight an bias updation for layer 1
|
// weight an bias updation for layer 1
|
||||||
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate / inputMiniBatches[i].size(), D1_3));
|
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate / inputMiniBatches[i].size(), D1_3));
|
||||||
@ -196,7 +196,7 @@ void MLPPAutoEncoder::MBGD(double learning_rate, int max_epoch, int mini_batch_s
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPAutoEncoder::score() {
|
real_t MLPPAutoEncoder::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
return util.performance(y_hat, inputSet);
|
return util.performance(y_hat, inputSet);
|
||||||
}
|
}
|
||||||
@ -207,40 +207,40 @@ void MLPPAutoEncoder::save(std::string fileName) {
|
|||||||
util.saveParameters(fileName, weights2, bias2, 1, 2);
|
util.saveParameters(fileName, weights2, bias2, 1, 2);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPAutoEncoder::Cost(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
real_t MLPPAutoEncoder::Cost(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
return cost.MSE(y_hat, inputSet);
|
return cost.MSE(y_hat, inputSet);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPAutoEncoder::Evaluate(std::vector<std::vector<double>> X) {
|
std::vector<std::vector<real_t>> MLPPAutoEncoder::Evaluate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
std::vector<std::vector<double>> z2 = alg.mat_vec_add(alg.matmult(X, weights1), bias1);
|
std::vector<std::vector<real_t>> z2 = alg.mat_vec_add(alg.matmult(X, weights1), bias1);
|
||||||
std::vector<std::vector<double>> a2 = avn.sigmoid(z2);
|
std::vector<std::vector<real_t>> a2 = avn.sigmoid(z2);
|
||||||
return alg.mat_vec_add(alg.matmult(a2, weights2), bias2);
|
return alg.mat_vec_add(alg.matmult(a2, weights2), bias2);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>> MLPPAutoEncoder::propagate(std::vector<std::vector<double>> X) {
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> MLPPAutoEncoder::propagate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
std::vector<std::vector<double>> z2 = alg.mat_vec_add(alg.matmult(X, weights1), bias1);
|
std::vector<std::vector<real_t>> z2 = alg.mat_vec_add(alg.matmult(X, weights1), bias1);
|
||||||
std::vector<std::vector<double>> a2 = avn.sigmoid(z2);
|
std::vector<std::vector<real_t>> a2 = avn.sigmoid(z2);
|
||||||
return { z2, a2 };
|
return { z2, a2 };
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPAutoEncoder::Evaluate(std::vector<double> x) {
|
std::vector<real_t> MLPPAutoEncoder::Evaluate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
std::vector<double> z2 = alg.addition(alg.mat_vec_mult(alg.transpose(weights1), x), bias1);
|
std::vector<real_t> z2 = alg.addition(alg.mat_vec_mult(alg.transpose(weights1), x), bias1);
|
||||||
std::vector<double> a2 = avn.sigmoid(z2);
|
std::vector<real_t> a2 = avn.sigmoid(z2);
|
||||||
return alg.addition(alg.mat_vec_mult(alg.transpose(weights2), a2), bias2);
|
return alg.addition(alg.mat_vec_mult(alg.transpose(weights2), a2), bias2);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<double>, std::vector<double>> MLPPAutoEncoder::propagate(std::vector<double> x) {
|
std::tuple<std::vector<real_t>, std::vector<real_t>> MLPPAutoEncoder::propagate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
std::vector<double> z2 = alg.addition(alg.mat_vec_mult(alg.transpose(weights1), x), bias1);
|
std::vector<real_t> z2 = alg.addition(alg.mat_vec_mult(alg.transpose(weights1), x), bias1);
|
||||||
std::vector<double> a2 = avn.sigmoid(z2);
|
std::vector<real_t> a2 = avn.sigmoid(z2);
|
||||||
return { z2, a2 };
|
return { z2, a2 };
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -8,41 +8,43 @@
|
|||||||
// Created by Marc Melikyan on 11/4/20.
|
// Created by Marc Melikyan on 11/4/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <tuple>
|
#include <tuple>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
class MLPPAutoEncoder {
|
class MLPPAutoEncoder {
|
||||||
public:
|
public:
|
||||||
MLPPAutoEncoder(std::vector<std::vector<double>> inputSet, int n_hidden);
|
MLPPAutoEncoder(std::vector<std::vector<real_t>> inputSet, int n_hidden);
|
||||||
std::vector<std::vector<double>> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<std::vector<real_t>> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
std::vector<double> modelTest(std::vector<double> x);
|
std::vector<real_t> modelTest(std::vector<real_t> x);
|
||||||
void gradientDescent(double learning_rate, int max_epoch, bool UI = 1);
|
void gradientDescent(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void SGD(double learning_rate, int max_epoch, bool UI = 1);
|
void SGD(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
void MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
||||||
double score();
|
real_t score();
|
||||||
void save(std::string fileName);
|
void save(std::string fileName);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
double Cost(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
real_t Cost(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
std::vector<std::vector<double>> Evaluate(std::vector<std::vector<double>> X);
|
std::vector<std::vector<real_t>> Evaluate(std::vector<std::vector<real_t>> X);
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>> propagate(std::vector<std::vector<double>> X);
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> propagate(std::vector<std::vector<real_t>> X);
|
||||||
std::vector<double> Evaluate(std::vector<double> x);
|
std::vector<real_t> Evaluate(std::vector<real_t> x);
|
||||||
std::tuple<std::vector<double>, std::vector<double>> propagate(std::vector<double> x);
|
std::tuple<std::vector<real_t>, std::vector<real_t>> propagate(std::vector<real_t> x);
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<std::vector<double>> y_hat;
|
std::vector<std::vector<real_t>> y_hat;
|
||||||
|
|
||||||
std::vector<std::vector<double>> weights1;
|
std::vector<std::vector<real_t>> weights1;
|
||||||
std::vector<std::vector<double>> weights2;
|
std::vector<std::vector<real_t>> weights2;
|
||||||
|
|
||||||
std::vector<double> bias1;
|
std::vector<real_t> bias1;
|
||||||
std::vector<double> bias2;
|
std::vector<real_t> bias2;
|
||||||
|
|
||||||
std::vector<std::vector<double>> z2;
|
std::vector<std::vector<real_t>> z2;
|
||||||
std::vector<std::vector<double>> a2;
|
std::vector<std::vector<real_t>> a2;
|
||||||
|
|
||||||
int n;
|
int n;
|
||||||
int k;
|
int k;
|
||||||
|
@ -12,23 +12,23 @@
|
|||||||
#include <iostream>
|
#include <iostream>
|
||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
MLPPBernoulliNB::MLPPBernoulliNB(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet) :
|
MLPPBernoulliNB::MLPPBernoulliNB(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet) :
|
||||||
inputSet(inputSet), outputSet(outputSet), class_num(2) {
|
inputSet(inputSet), outputSet(outputSet), class_num(2) {
|
||||||
y_hat.resize(outputSet.size());
|
y_hat.resize(outputSet.size());
|
||||||
Evaluate();
|
Evaluate();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPBernoulliNB::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPBernoulliNB::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
for (int i = 0; i < X.size(); i++) {
|
for (int i = 0; i < X.size(); i++) {
|
||||||
y_hat.push_back(modelTest(X[i]));
|
y_hat.push_back(modelTest(X[i]));
|
||||||
}
|
}
|
||||||
return y_hat;
|
return y_hat;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPBernoulliNB::modelTest(std::vector<double> x) {
|
real_t MLPPBernoulliNB::modelTest(std::vector<real_t> x) {
|
||||||
double score_0 = 1;
|
real_t score_0 = 1;
|
||||||
double score_1 = 1;
|
real_t score_1 = 1;
|
||||||
|
|
||||||
std::vector<int> foundIndices;
|
std::vector<int> foundIndices;
|
||||||
|
|
||||||
@ -68,7 +68,7 @@ double MLPPBernoulliNB::modelTest(std::vector<double> x) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPBernoulliNB::score() {
|
real_t MLPPBernoulliNB::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
return util.performance(y_hat, outputSet);
|
return util.performance(y_hat, outputSet);
|
||||||
}
|
}
|
||||||
@ -76,7 +76,7 @@ double MLPPBernoulliNB::score() {
|
|||||||
void MLPPBernoulliNB::computeVocab() {
|
void MLPPBernoulliNB::computeVocab() {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPData data;
|
MLPPData data;
|
||||||
vocab = data.vecToSet<double>(alg.flatten(inputSet));
|
vocab = data.vecToSet<real_t>(alg.flatten(inputSet));
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPBernoulliNB::computeTheta() {
|
void MLPPBernoulliNB::computeTheta() {
|
||||||
@ -110,10 +110,10 @@ void MLPPBernoulliNB::computeTheta() {
|
|||||||
void MLPPBernoulliNB::Evaluate() {
|
void MLPPBernoulliNB::Evaluate() {
|
||||||
for (int i = 0; i < outputSet.size(); i++) {
|
for (int i = 0; i < outputSet.size(); i++) {
|
||||||
// Pr(B | A) * Pr(A)
|
// Pr(B | A) * Pr(A)
|
||||||
double score_0 = 1;
|
real_t score_0 = 1;
|
||||||
double score_1 = 1;
|
real_t score_1 = 1;
|
||||||
|
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < outputSet.size(); i++) {
|
for (int i = 0; i < outputSet.size(); i++) {
|
||||||
if (outputSet[i] == 1) {
|
if (outputSet[i] == 1) {
|
||||||
sum += outputSet[i];
|
sum += outputSet[i];
|
||||||
|
@ -8,15 +8,17 @@
|
|||||||
// Created by Marc Melikyan on 1/17/21.
|
// Created by Marc Melikyan on 1/17/21.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <map>
|
#include <map>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
class MLPPBernoulliNB {
|
class MLPPBernoulliNB {
|
||||||
public:
|
public:
|
||||||
MLPPBernoulliNB(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet);
|
MLPPBernoulliNB(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet);
|
||||||
std::vector<double> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<real_t> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
double modelTest(std::vector<double> x);
|
real_t modelTest(std::vector<real_t> x);
|
||||||
double score();
|
real_t score();
|
||||||
|
|
||||||
private:
|
private:
|
||||||
void computeVocab();
|
void computeVocab();
|
||||||
@ -24,17 +26,17 @@ private:
|
|||||||
void Evaluate();
|
void Evaluate();
|
||||||
|
|
||||||
// Model Params
|
// Model Params
|
||||||
double prior_1 = 0;
|
real_t prior_1 = 0;
|
||||||
double prior_0 = 0;
|
real_t prior_0 = 0;
|
||||||
|
|
||||||
std::vector<std::map<double, int>> theta;
|
std::vector<std::map<real_t, int>> theta;
|
||||||
std::vector<double> vocab;
|
std::vector<real_t> vocab;
|
||||||
int class_num;
|
int class_num;
|
||||||
|
|
||||||
// Datasets
|
// Datasets
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
};
|
};
|
||||||
|
|
||||||
#endif /* BernoulliNB_hpp */
|
#endif /* BernoulliNB_hpp */
|
@ -14,33 +14,33 @@
|
|||||||
#include <iostream>
|
#include <iostream>
|
||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
MLPPCLogLogReg::MLPPCLogLogReg(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, std::string reg, double lambda, double alpha) :
|
MLPPCLogLogReg::MLPPCLogLogReg(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, std::string reg, real_t lambda, real_t alpha) :
|
||||||
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
||||||
y_hat.resize(n);
|
y_hat.resize(n);
|
||||||
weights = MLPPUtilities::weightInitialization(k);
|
weights = MLPPUtilities::weightInitialization(k);
|
||||||
bias = MLPPUtilities::biasInitialization();
|
bias = MLPPUtilities::biasInitialization();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPCLogLogReg::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPCLogLogReg::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
return Evaluate(X);
|
return Evaluate(X);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCLogLogReg::modelTest(std::vector<double> x) {
|
real_t MLPPCLogLogReg::modelTest(std::vector<real_t> x) {
|
||||||
return Evaluate(x);
|
return Evaluate(x);
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPCLogLogReg::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
void MLPPCLogLogReg::gradientDescent(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
cost_prev = Cost(y_hat, outputSet);
|
cost_prev = Cost(y_hat, outputSet);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputSet);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputSet);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputSet), alg.hadamard_product(error, avn.cloglog(z, 1)))));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputSet), alg.hadamard_product(error, avn.cloglog(z, 1)))));
|
||||||
@ -63,18 +63,18 @@ void MLPPCLogLogReg::gradientDescent(double learning_rate, int max_epoch, bool U
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPCLogLogReg::MLE(double learning_rate, int max_epoch, bool UI) {
|
void MLPPCLogLogReg::MLE(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
cost_prev = Cost(y_hat, outputSet);
|
cost_prev = Cost(y_hat, outputSet);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputSet);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputSet);
|
||||||
|
|
||||||
weights = alg.addition(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputSet), alg.hadamard_product(error, avn.cloglog(z, 1)))));
|
weights = alg.addition(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputSet), alg.hadamard_product(error, avn.cloglog(z, 1)))));
|
||||||
weights = regularization.regWeights(weights, lambda, alpha, reg);
|
weights = regularization.regWeights(weights, lambda, alpha, reg);
|
||||||
@ -95,10 +95,10 @@ void MLPPCLogLogReg::MLE(double learning_rate, int max_epoch, bool UI) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPCLogLogReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
void MLPPCLogLogReg::SGD(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
@ -108,11 +108,11 @@ void MLPPCLogLogReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
||||||
int outputIndex = distribution(generator);
|
int outputIndex = distribution(generator);
|
||||||
|
|
||||||
double y_hat = Evaluate(inputSet[outputIndex]);
|
real_t y_hat = Evaluate(inputSet[outputIndex]);
|
||||||
double z = propagate(inputSet[outputIndex]);
|
real_t z = propagate(inputSet[outputIndex]);
|
||||||
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
||||||
|
|
||||||
double error = y_hat - outputSet[outputIndex];
|
real_t error = y_hat - outputSet[outputIndex];
|
||||||
|
|
||||||
// Weight Updation
|
// Weight Updation
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate * error * exp(z - exp(z)), inputSet[outputIndex]));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate * error * exp(z - exp(z)), inputSet[outputIndex]));
|
||||||
@ -136,11 +136,11 @@ void MLPPCLogLogReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPCLogLogReg::MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
void MLPPCLogLogReg::MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
@ -149,11 +149,11 @@ void MLPPCLogLogReg::MBGD(double learning_rate, int max_epoch, int mini_batch_si
|
|||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
std::vector<double> z = propagate(inputMiniBatches[i]);
|
std::vector<real_t> z = propagate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), alg.hadamard_product(error, avn.cloglog(z, 1)))));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), alg.hadamard_product(error, avn.cloglog(z, 1)))));
|
||||||
@ -179,35 +179,35 @@ void MLPPCLogLogReg::MBGD(double learning_rate, int max_epoch, int mini_batch_si
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCLogLogReg::score() {
|
real_t MLPPCLogLogReg::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
return util.performance(y_hat, outputSet);
|
return util.performance(y_hat, outputSet);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCLogLogReg::Cost(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPCLogLogReg::Cost(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
return cost.MSE(y_hat, y) + regularization.regTerm(weights, lambda, alpha, reg);
|
return cost.MSE(y_hat, y) + regularization.regTerm(weights, lambda, alpha, reg);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPCLogLogReg::Evaluate(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPCLogLogReg::Evaluate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
return avn.cloglog(alg.scalarAdd(bias, alg.mat_vec_mult(X, weights)));
|
return avn.cloglog(alg.scalarAdd(bias, alg.mat_vec_mult(X, weights)));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPCLogLogReg::propagate(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPCLogLogReg::propagate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.scalarAdd(bias, alg.mat_vec_mult(X, weights));
|
return alg.scalarAdd(bias, alg.mat_vec_mult(X, weights));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCLogLogReg::Evaluate(std::vector<double> x) {
|
real_t MLPPCLogLogReg::Evaluate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
return avn.cloglog(alg.dot(weights, x) + bias);
|
return avn.cloglog(alg.dot(weights, x) + bias);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCLogLogReg::propagate(std::vector<double> x) {
|
real_t MLPPCLogLogReg::propagate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.dot(weights, x) + bias;
|
return alg.dot(weights, x) + bias;
|
||||||
}
|
}
|
||||||
|
@ -8,45 +8,47 @@
|
|||||||
// Created by Marc Melikyan on 10/2/20.
|
// Created by Marc Melikyan on 10/2/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
class MLPPCLogLogReg {
|
class MLPPCLogLogReg {
|
||||||
public:
|
public:
|
||||||
MLPPCLogLogReg(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, std::string reg = "None", double lambda = 0.5, double alpha = 0.5);
|
MLPPCLogLogReg(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
||||||
std::vector<double> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<real_t> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
double modelTest(std::vector<double> x);
|
real_t modelTest(std::vector<real_t> x);
|
||||||
void gradientDescent(double learning_rate, int max_epoch, bool UI = 1);
|
void gradientDescent(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void MLE(double learning_rate, int max_epoch, bool UI = 1);
|
void MLE(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void SGD(double learning_rate, int max_epoch, bool UI = 1);
|
void SGD(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
void MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
||||||
double score();
|
real_t score();
|
||||||
|
|
||||||
private:
|
private:
|
||||||
void weightInitialization(int k);
|
void weightInitialization(int k);
|
||||||
void biasInitialization();
|
void biasInitialization();
|
||||||
double Cost(std::vector<double> y_hat, std::vector<double> y);
|
real_t Cost(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
|
|
||||||
std::vector<double> Evaluate(std::vector<std::vector<double>> X);
|
std::vector<real_t> Evaluate(std::vector<std::vector<real_t>> X);
|
||||||
std::vector<double> propagate(std::vector<std::vector<double>> X);
|
std::vector<real_t> propagate(std::vector<std::vector<real_t>> X);
|
||||||
double Evaluate(std::vector<double> x);
|
real_t Evaluate(std::vector<real_t> x);
|
||||||
double propagate(std::vector<double> x);
|
real_t propagate(std::vector<real_t> x);
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
std::vector<double> z;
|
std::vector<real_t> z;
|
||||||
std::vector<double> weights;
|
std::vector<real_t> weights;
|
||||||
double bias;
|
real_t bias;
|
||||||
|
|
||||||
int n;
|
int n;
|
||||||
int k;
|
int k;
|
||||||
|
|
||||||
// Regularization Params
|
// Regularization Params
|
||||||
std::string reg;
|
std::string reg;
|
||||||
double lambda;
|
real_t lambda;
|
||||||
double alpha; /* This is the controlling param for Elastic Net*/
|
real_t alpha; /* This is the controlling param for Elastic Net*/
|
||||||
};
|
};
|
||||||
|
|
||||||
#endif /* CLogLogReg_hpp */
|
#endif /* CLogLogReg_hpp */
|
||||||
|
@ -14,15 +14,15 @@ MLPPConvolutions::MLPPConvolutions() :
|
|||||||
prewittHorizontal({ { 1, 1, 1 }, { 0, 0, 0 }, { -1, -1, -1 } }), prewittVertical({ { 1, 0, -1 }, { 1, 0, -1 }, { 1, 0, -1 } }), sobelHorizontal({ { 1, 2, 1 }, { 0, 0, 0 }, { -1, -2, -1 } }), sobelVertical({ { -1, 0, 1 }, { -2, 0, 2 }, { -1, 0, 1 } }), scharrHorizontal({ { 3, 10, 3 }, { 0, 0, 0 }, { -3, -10, -3 } }), scharrVertical({ { 3, 0, -3 }, { 10, 0, -10 }, { 3, 0, -3 } }), robertsHorizontal({ { 0, 1 }, { -1, 0 } }), robertsVertical({ { 1, 0 }, { 0, -1 } }) {
|
prewittHorizontal({ { 1, 1, 1 }, { 0, 0, 0 }, { -1, -1, -1 } }), prewittVertical({ { 1, 0, -1 }, { 1, 0, -1 }, { 1, 0, -1 } }), sobelHorizontal({ { 1, 2, 1 }, { 0, 0, 0 }, { -1, -2, -1 } }), sobelVertical({ { -1, 0, 1 }, { -2, 0, 2 }, { -1, 0, 1 } }), scharrHorizontal({ { 3, 10, 3 }, { 0, 0, 0 }, { -3, -10, -3 } }), scharrVertical({ { 3, 0, -3 }, { 10, 0, -10 }, { 3, 0, -3 } }), robertsHorizontal({ { 0, 1 }, { -1, 0 } }), robertsVertical({ { 1, 0 }, { 0, -1 } }) {
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPConvolutions::convolve(std::vector<std::vector<double>> input, std::vector<std::vector<double>> filter, int S, int P) {
|
std::vector<std::vector<real_t>> MLPPConvolutions::convolve(std::vector<std::vector<real_t>> input, std::vector<std::vector<real_t>> filter, int S, int P) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<double>> featureMap;
|
std::vector<std::vector<real_t>> featureMap;
|
||||||
int N = input.size();
|
int N = input.size();
|
||||||
int F = filter.size();
|
int F = filter.size();
|
||||||
int mapSize = (N - F + 2 * P) / S + 1; // This is computed as ⌊mapSize⌋ by def- thanks C++!
|
int mapSize = (N - F + 2 * P) / S + 1; // This is computed as ⌊mapSize⌋ by def- thanks C++!
|
||||||
|
|
||||||
if (P != 0) {
|
if (P != 0) {
|
||||||
std::vector<std::vector<double>> paddedInput;
|
std::vector<std::vector<real_t>> paddedInput;
|
||||||
paddedInput.resize(N + 2 * P);
|
paddedInput.resize(N + 2 * P);
|
||||||
for (int i = 0; i < paddedInput.size(); i++) {
|
for (int i = 0; i < paddedInput.size(); i++) {
|
||||||
paddedInput[i].resize(N + 2 * P);
|
paddedInput[i].resize(N + 2 * P);
|
||||||
@ -50,7 +50,7 @@ std::vector<std::vector<double>> MLPPConvolutions::convolve(std::vector<std::vec
|
|||||||
|
|
||||||
for (int i = 0; i < mapSize; i++) {
|
for (int i = 0; i < mapSize; i++) {
|
||||||
for (int j = 0; j < mapSize; j++) {
|
for (int j = 0; j < mapSize; j++) {
|
||||||
std::vector<double> convolvingInput;
|
std::vector<real_t> convolvingInput;
|
||||||
for (int k = 0; k < F; k++) {
|
for (int k = 0; k < F; k++) {
|
||||||
for (int p = 0; p < F; p++) {
|
for (int p = 0; p < F; p++) {
|
||||||
if (i == 0 && j == 0) {
|
if (i == 0 && j == 0) {
|
||||||
@ -70,9 +70,9 @@ std::vector<std::vector<double>> MLPPConvolutions::convolve(std::vector<std::vec
|
|||||||
return featureMap;
|
return featureMap;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> MLPPConvolutions::convolve(std::vector<std::vector<std::vector<double>>> input, std::vector<std::vector<std::vector<double>>> filter, int S, int P) {
|
std::vector<std::vector<std::vector<real_t>>> MLPPConvolutions::convolve(std::vector<std::vector<std::vector<real_t>>> input, std::vector<std::vector<std::vector<real_t>>> filter, int S, int P) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<std::vector<double>>> featureMap;
|
std::vector<std::vector<std::vector<real_t>>> featureMap;
|
||||||
int N = input[0].size();
|
int N = input[0].size();
|
||||||
int F = filter[0].size();
|
int F = filter[0].size();
|
||||||
int C = filter.size() / input.size();
|
int C = filter.size() / input.size();
|
||||||
@ -80,7 +80,7 @@ std::vector<std::vector<std::vector<double>>> MLPPConvolutions::convolve(std::ve
|
|||||||
|
|
||||||
if (P != 0) {
|
if (P != 0) {
|
||||||
for (int c = 0; c < input.size(); c++) {
|
for (int c = 0; c < input.size(); c++) {
|
||||||
std::vector<std::vector<double>> paddedInput;
|
std::vector<std::vector<real_t>> paddedInput;
|
||||||
paddedInput.resize(N + 2 * P);
|
paddedInput.resize(N + 2 * P);
|
||||||
for (int i = 0; i < paddedInput.size(); i++) {
|
for (int i = 0; i < paddedInput.size(); i++) {
|
||||||
paddedInput[i].resize(N + 2 * P);
|
paddedInput[i].resize(N + 2 * P);
|
||||||
@ -113,7 +113,7 @@ std::vector<std::vector<std::vector<double>>> MLPPConvolutions::convolve(std::ve
|
|||||||
for (int c = 0; c < C; c++) {
|
for (int c = 0; c < C; c++) {
|
||||||
for (int i = 0; i < mapSize; i++) {
|
for (int i = 0; i < mapSize; i++) {
|
||||||
for (int j = 0; j < mapSize; j++) {
|
for (int j = 0; j < mapSize; j++) {
|
||||||
std::vector<double> convolvingInput;
|
std::vector<real_t> convolvingInput;
|
||||||
for (int t = 0; t < input.size(); t++) {
|
for (int t = 0; t < input.size(); t++) {
|
||||||
for (int k = 0; k < F; k++) {
|
for (int k = 0; k < F; k++) {
|
||||||
for (int p = 0; p < F; p++) {
|
for (int p = 0; p < F; p++) {
|
||||||
@ -136,9 +136,9 @@ std::vector<std::vector<std::vector<double>>> MLPPConvolutions::convolve(std::ve
|
|||||||
return featureMap;
|
return featureMap;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPConvolutions::pool(std::vector<std::vector<double>> input, int F, int S, std::string type) {
|
std::vector<std::vector<real_t>> MLPPConvolutions::pool(std::vector<std::vector<real_t>> input, int F, int S, std::string type) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<double>> pooledMap;
|
std::vector<std::vector<real_t>> pooledMap;
|
||||||
int N = input.size();
|
int N = input.size();
|
||||||
int mapSize = floor((N - F) / S + 1);
|
int mapSize = floor((N - F) / S + 1);
|
||||||
|
|
||||||
@ -149,7 +149,7 @@ std::vector<std::vector<double>> MLPPConvolutions::pool(std::vector<std::vector<
|
|||||||
|
|
||||||
for (int i = 0; i < mapSize; i++) {
|
for (int i = 0; i < mapSize; i++) {
|
||||||
for (int j = 0; j < mapSize; j++) {
|
for (int j = 0; j < mapSize; j++) {
|
||||||
std::vector<double> poolingInput;
|
std::vector<real_t> poolingInput;
|
||||||
for (int k = 0; k < F; k++) {
|
for (int k = 0; k < F; k++) {
|
||||||
for (int p = 0; p < F; p++) {
|
for (int p = 0; p < F; p++) {
|
||||||
if (i == 0 && j == 0) {
|
if (i == 0 && j == 0) {
|
||||||
@ -176,15 +176,15 @@ std::vector<std::vector<double>> MLPPConvolutions::pool(std::vector<std::vector<
|
|||||||
return pooledMap;
|
return pooledMap;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> MLPPConvolutions::pool(std::vector<std::vector<std::vector<double>>> input, int F, int S, std::string type) {
|
std::vector<std::vector<std::vector<real_t>>> MLPPConvolutions::pool(std::vector<std::vector<std::vector<real_t>>> input, int F, int S, std::string type) {
|
||||||
std::vector<std::vector<std::vector<double>>> pooledMap;
|
std::vector<std::vector<std::vector<real_t>>> pooledMap;
|
||||||
for (int i = 0; i < input.size(); i++) {
|
for (int i = 0; i < input.size(); i++) {
|
||||||
pooledMap.push_back(pool(input[i], F, S, type));
|
pooledMap.push_back(pool(input[i], F, S, type));
|
||||||
}
|
}
|
||||||
return pooledMap;
|
return pooledMap;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPConvolutions::globalPool(std::vector<std::vector<double>> input, std::string type) {
|
real_t MLPPConvolutions::globalPool(std::vector<std::vector<real_t>> input, std::string type) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (type == "Average") {
|
if (type == "Average") {
|
||||||
MLPPStat stat;
|
MLPPStat stat;
|
||||||
@ -196,21 +196,21 @@ double MLPPConvolutions::globalPool(std::vector<std::vector<double>> input, std:
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPConvolutions::globalPool(std::vector<std::vector<std::vector<double>>> input, std::string type) {
|
std::vector<real_t> MLPPConvolutions::globalPool(std::vector<std::vector<std::vector<real_t>>> input, std::string type) {
|
||||||
std::vector<double> pooledMap;
|
std::vector<real_t> pooledMap;
|
||||||
for (int i = 0; i < input.size(); i++) {
|
for (int i = 0; i < input.size(); i++) {
|
||||||
pooledMap.push_back(globalPool(input[i], type));
|
pooledMap.push_back(globalPool(input[i], type));
|
||||||
}
|
}
|
||||||
return pooledMap;
|
return pooledMap;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPConvolutions::gaussian2D(double x, double y, double std) {
|
real_t MLPPConvolutions::gaussian2D(real_t x, real_t y, real_t std) {
|
||||||
double std_sq = std * std;
|
real_t std_sq = std * std;
|
||||||
return 1 / (2 * M_PI * std_sq) * std::exp(-(x * x + y * y) / 2 * std_sq);
|
return 1 / (2 * M_PI * std_sq) * std::exp(-(x * x + y * y) / 2 * std_sq);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPConvolutions::gaussianFilter2D(int size, double std) {
|
std::vector<std::vector<real_t>> MLPPConvolutions::gaussianFilter2D(int size, real_t std) {
|
||||||
std::vector<std::vector<double>> filter;
|
std::vector<std::vector<real_t>> filter;
|
||||||
filter.resize(size);
|
filter.resize(size);
|
||||||
for (int i = 0; i < filter.size(); i++) {
|
for (int i = 0; i < filter.size(); i++) {
|
||||||
filter[i].resize(size);
|
filter[i].resize(size);
|
||||||
@ -229,8 +229,8 @@ been easier to carry out the calculation explicitly, mainly because it is more i
|
|||||||
and also because my convolution algorithm is only built for filters with equally sized
|
and also because my convolution algorithm is only built for filters with equally sized
|
||||||
heights and widths.
|
heights and widths.
|
||||||
*/
|
*/
|
||||||
std::vector<std::vector<double>> MLPPConvolutions::dx(std::vector<std::vector<double>> input) {
|
std::vector<std::vector<real_t>> MLPPConvolutions::dx(std::vector<std::vector<real_t>> input) {
|
||||||
std::vector<std::vector<double>> deriv; // We assume a gray scale image.
|
std::vector<std::vector<real_t>> deriv; // We assume a gray scale image.
|
||||||
deriv.resize(input.size());
|
deriv.resize(input.size());
|
||||||
for (int i = 0; i < deriv.size(); i++) {
|
for (int i = 0; i < deriv.size(); i++) {
|
||||||
deriv[i].resize(input[i].size());
|
deriv[i].resize(input[i].size());
|
||||||
@ -250,8 +250,8 @@ std::vector<std::vector<double>> MLPPConvolutions::dx(std::vector<std::vector<do
|
|||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPConvolutions::dy(std::vector<std::vector<double>> input) {
|
std::vector<std::vector<real_t>> MLPPConvolutions::dy(std::vector<std::vector<real_t>> input) {
|
||||||
std::vector<std::vector<double>> deriv;
|
std::vector<std::vector<real_t>> deriv;
|
||||||
deriv.resize(input.size());
|
deriv.resize(input.size());
|
||||||
for (int i = 0; i < deriv.size(); i++) {
|
for (int i = 0; i < deriv.size(); i++) {
|
||||||
deriv[i].resize(input[i].size());
|
deriv[i].resize(input[i].size());
|
||||||
@ -271,22 +271,22 @@ std::vector<std::vector<double>> MLPPConvolutions::dy(std::vector<std::vector<do
|
|||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPConvolutions::gradMagnitude(std::vector<std::vector<double>> input) {
|
std::vector<std::vector<real_t>> MLPPConvolutions::gradMagnitude(std::vector<std::vector<real_t>> input) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<double>> xDeriv_2 = alg.hadamard_product(dx(input), dx(input));
|
std::vector<std::vector<real_t>> xDeriv_2 = alg.hadamard_product(dx(input), dx(input));
|
||||||
std::vector<std::vector<double>> yDeriv_2 = alg.hadamard_product(dy(input), dy(input));
|
std::vector<std::vector<real_t>> yDeriv_2 = alg.hadamard_product(dy(input), dy(input));
|
||||||
return alg.sqrt(alg.addition(xDeriv_2, yDeriv_2));
|
return alg.sqrt(alg.addition(xDeriv_2, yDeriv_2));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPConvolutions::gradOrientation(std::vector<std::vector<double>> input) {
|
std::vector<std::vector<real_t>> MLPPConvolutions::gradOrientation(std::vector<std::vector<real_t>> input) {
|
||||||
std::vector<std::vector<double>> deriv;
|
std::vector<std::vector<real_t>> deriv;
|
||||||
deriv.resize(input.size());
|
deriv.resize(input.size());
|
||||||
for (int i = 0; i < deriv.size(); i++) {
|
for (int i = 0; i < deriv.size(); i++) {
|
||||||
deriv[i].resize(input[i].size());
|
deriv[i].resize(input[i].size());
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> xDeriv = dx(input);
|
std::vector<std::vector<real_t>> xDeriv = dx(input);
|
||||||
std::vector<std::vector<double>> yDeriv = dy(input);
|
std::vector<std::vector<real_t>> yDeriv = dy(input);
|
||||||
for (int i = 0; i < deriv.size(); i++) {
|
for (int i = 0; i < deriv.size(); i++) {
|
||||||
for (int j = 0; j < deriv[i].size(); j++) {
|
for (int j = 0; j < deriv[i].size(); j++) {
|
||||||
deriv[i][j] = std::atan2(yDeriv[i][j], xDeriv[i][j]);
|
deriv[i][j] = std::atan2(yDeriv[i][j], xDeriv[i][j]);
|
||||||
@ -295,33 +295,33 @@ std::vector<std::vector<double>> MLPPConvolutions::gradOrientation(std::vector<s
|
|||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> MLPPConvolutions::computeM(std::vector<std::vector<double>> input) {
|
std::vector<std::vector<std::vector<real_t>>> MLPPConvolutions::computeM(std::vector<std::vector<real_t>> input) {
|
||||||
double const SIGMA = 1;
|
real_t const SIGMA = 1;
|
||||||
double const GAUSSIAN_SIZE = 3;
|
real_t const GAUSSIAN_SIZE = 3;
|
||||||
|
|
||||||
double const GAUSSIAN_PADDING = ((input.size() - 1) + GAUSSIAN_SIZE - input.size()) / 2; // Convs must be same.
|
real_t const GAUSSIAN_PADDING = ((input.size() - 1) + GAUSSIAN_SIZE - input.size()) / 2; // Convs must be same.
|
||||||
std::cout << GAUSSIAN_PADDING << std::endl;
|
std::cout << GAUSSIAN_PADDING << std::endl;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<double>> xDeriv = dx(input);
|
std::vector<std::vector<real_t>> xDeriv = dx(input);
|
||||||
std::vector<std::vector<double>> yDeriv = dy(input);
|
std::vector<std::vector<real_t>> yDeriv = dy(input);
|
||||||
|
|
||||||
std::vector<std::vector<double>> gaussianFilter = gaussianFilter2D(GAUSSIAN_SIZE, SIGMA); // Sigma of 1, size of 3.
|
std::vector<std::vector<real_t>> gaussianFilter = gaussianFilter2D(GAUSSIAN_SIZE, SIGMA); // Sigma of 1, size of 3.
|
||||||
std::vector<std::vector<double>> xxDeriv = convolve(alg.hadamard_product(xDeriv, xDeriv), gaussianFilter, 1, GAUSSIAN_PADDING);
|
std::vector<std::vector<real_t>> xxDeriv = convolve(alg.hadamard_product(xDeriv, xDeriv), gaussianFilter, 1, GAUSSIAN_PADDING);
|
||||||
std::vector<std::vector<double>> yyDeriv = convolve(alg.hadamard_product(yDeriv, yDeriv), gaussianFilter, 1, GAUSSIAN_PADDING);
|
std::vector<std::vector<real_t>> yyDeriv = convolve(alg.hadamard_product(yDeriv, yDeriv), gaussianFilter, 1, GAUSSIAN_PADDING);
|
||||||
std::vector<std::vector<double>> xyDeriv = convolve(alg.hadamard_product(xDeriv, yDeriv), gaussianFilter, 1, GAUSSIAN_PADDING);
|
std::vector<std::vector<real_t>> xyDeriv = convolve(alg.hadamard_product(xDeriv, yDeriv), gaussianFilter, 1, GAUSSIAN_PADDING);
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> M = { xxDeriv, yyDeriv, xyDeriv };
|
std::vector<std::vector<std::vector<real_t>>> M = { xxDeriv, yyDeriv, xyDeriv };
|
||||||
return M;
|
return M;
|
||||||
}
|
}
|
||||||
std::vector<std::vector<std::string>> MLPPConvolutions::harrisCornerDetection(std::vector<std::vector<double>> input) {
|
std::vector<std::vector<std::string>> MLPPConvolutions::harrisCornerDetection(std::vector<std::vector<real_t>> input) {
|
||||||
double const k = 0.05; // Empirically determined wherein k -> [0.04, 0.06], though conventionally 0.05 is typically used as well.
|
real_t const k = 0.05; // Empirically determined wherein k -> [0.04, 0.06], though conventionally 0.05 is typically used as well.
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<std::vector<double>>> M = computeM(input);
|
std::vector<std::vector<std::vector<real_t>>> M = computeM(input);
|
||||||
std::vector<std::vector<double>> det = alg.subtraction(alg.hadamard_product(M[0], M[1]), alg.hadamard_product(M[2], M[2]));
|
std::vector<std::vector<real_t>> det = alg.subtraction(alg.hadamard_product(M[0], M[1]), alg.hadamard_product(M[2], M[2]));
|
||||||
std::vector<std::vector<double>> trace = alg.addition(M[0], M[1]);
|
std::vector<std::vector<real_t>> trace = alg.addition(M[0], M[1]);
|
||||||
|
|
||||||
// The reason this is not a scalar is because xxDeriv, xyDeriv, yxDeriv, and yyDeriv are not scalars.
|
// The reason this is not a scalar is because xxDeriv, xyDeriv, yxDeriv, and yyDeriv are not scalars.
|
||||||
std::vector<std::vector<double>> r = alg.subtraction(det, alg.scalarMultiply(k, alg.hadamard_product(trace, trace)));
|
std::vector<std::vector<real_t>> r = alg.subtraction(det, alg.scalarMultiply(k, alg.hadamard_product(trace, trace)));
|
||||||
std::vector<std::vector<std::string>> imageTypes;
|
std::vector<std::vector<std::string>> imageTypes;
|
||||||
imageTypes.resize(r.size());
|
imageTypes.resize(r.size());
|
||||||
alg.printMatrix(r);
|
alg.printMatrix(r);
|
||||||
@ -340,34 +340,34 @@ std::vector<std::vector<std::string>> MLPPConvolutions::harrisCornerDetection(st
|
|||||||
return imageTypes;
|
return imageTypes;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPConvolutions::getPrewittHorizontal() {
|
std::vector<std::vector<real_t>> MLPPConvolutions::getPrewittHorizontal() {
|
||||||
return prewittHorizontal;
|
return prewittHorizontal;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPConvolutions::getPrewittVertical() {
|
std::vector<std::vector<real_t>> MLPPConvolutions::getPrewittVertical() {
|
||||||
return prewittVertical;
|
return prewittVertical;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPConvolutions::getSobelHorizontal() {
|
std::vector<std::vector<real_t>> MLPPConvolutions::getSobelHorizontal() {
|
||||||
return sobelHorizontal;
|
return sobelHorizontal;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPConvolutions::getSobelVertical() {
|
std::vector<std::vector<real_t>> MLPPConvolutions::getSobelVertical() {
|
||||||
return sobelVertical;
|
return sobelVertical;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPConvolutions::getScharrHorizontal() {
|
std::vector<std::vector<real_t>> MLPPConvolutions::getScharrHorizontal() {
|
||||||
return scharrHorizontal;
|
return scharrHorizontal;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPConvolutions::getScharrVertical() {
|
std::vector<std::vector<real_t>> MLPPConvolutions::getScharrVertical() {
|
||||||
return scharrVertical;
|
return scharrVertical;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPConvolutions::getRobertsHorizontal() {
|
std::vector<std::vector<real_t>> MLPPConvolutions::getRobertsHorizontal() {
|
||||||
return robertsHorizontal;
|
return robertsHorizontal;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPConvolutions::getRobertsVertical() {
|
std::vector<std::vector<real_t>> MLPPConvolutions::getRobertsVertical() {
|
||||||
return robertsVertical;
|
return robertsVertical;
|
||||||
}
|
}
|
||||||
|
@ -5,46 +5,48 @@
|
|||||||
#include <vector>
|
#include <vector>
|
||||||
#include <string>
|
#include <string>
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
class MLPPConvolutions {
|
class MLPPConvolutions {
|
||||||
public:
|
public:
|
||||||
MLPPConvolutions();
|
MLPPConvolutions();
|
||||||
std::vector<std::vector<double>> convolve(std::vector<std::vector<double>> input, std::vector<std::vector<double>> filter, int S, int P = 0);
|
std::vector<std::vector<real_t>> convolve(std::vector<std::vector<real_t>> input, std::vector<std::vector<real_t>> filter, int S, int P = 0);
|
||||||
std::vector<std::vector<std::vector<double>>> convolve(std::vector<std::vector<std::vector<double>>> input, std::vector<std::vector<std::vector<double>>> filter, int S, int P = 0);
|
std::vector<std::vector<std::vector<real_t>>> convolve(std::vector<std::vector<std::vector<real_t>>> input, std::vector<std::vector<std::vector<real_t>>> filter, int S, int P = 0);
|
||||||
std::vector<std::vector<double>> pool(std::vector<std::vector<double>> input, int F, int S, std::string type);
|
std::vector<std::vector<real_t>> pool(std::vector<std::vector<real_t>> input, int F, int S, std::string type);
|
||||||
std::vector<std::vector<std::vector<double>>> pool(std::vector<std::vector<std::vector<double>>> input, int F, int S, std::string type);
|
std::vector<std::vector<std::vector<real_t>>> pool(std::vector<std::vector<std::vector<real_t>>> input, int F, int S, std::string type);
|
||||||
double globalPool(std::vector<std::vector<double>> input, std::string type);
|
real_t globalPool(std::vector<std::vector<real_t>> input, std::string type);
|
||||||
std::vector<double> globalPool(std::vector<std::vector<std::vector<double>>> input, std::string type);
|
std::vector<real_t> globalPool(std::vector<std::vector<std::vector<real_t>>> input, std::string type);
|
||||||
|
|
||||||
double gaussian2D(double x, double y, double std);
|
real_t gaussian2D(real_t x, real_t y, real_t std);
|
||||||
std::vector<std::vector<double>> gaussianFilter2D(int size, double std);
|
std::vector<std::vector<real_t>> gaussianFilter2D(int size, real_t std);
|
||||||
|
|
||||||
std::vector<std::vector<double>> dx(std::vector<std::vector<double>> input);
|
std::vector<std::vector<real_t>> dx(std::vector<std::vector<real_t>> input);
|
||||||
std::vector<std::vector<double>> dy(std::vector<std::vector<double>> input);
|
std::vector<std::vector<real_t>> dy(std::vector<std::vector<real_t>> input);
|
||||||
|
|
||||||
std::vector<std::vector<double>> gradMagnitude(std::vector<std::vector<double>> input);
|
std::vector<std::vector<real_t>> gradMagnitude(std::vector<std::vector<real_t>> input);
|
||||||
std::vector<std::vector<double>> gradOrientation(std::vector<std::vector<double>> input);
|
std::vector<std::vector<real_t>> gradOrientation(std::vector<std::vector<real_t>> input);
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> computeM(std::vector<std::vector<double>> input);
|
std::vector<std::vector<std::vector<real_t>>> computeM(std::vector<std::vector<real_t>> input);
|
||||||
std::vector<std::vector<std::string>> harrisCornerDetection(std::vector<std::vector<double>> input);
|
std::vector<std::vector<std::string>> harrisCornerDetection(std::vector<std::vector<real_t>> input);
|
||||||
|
|
||||||
std::vector<std::vector<double>> getPrewittHorizontal();
|
std::vector<std::vector<real_t>> getPrewittHorizontal();
|
||||||
std::vector<std::vector<double>> getPrewittVertical();
|
std::vector<std::vector<real_t>> getPrewittVertical();
|
||||||
std::vector<std::vector<double>> getSobelHorizontal();
|
std::vector<std::vector<real_t>> getSobelHorizontal();
|
||||||
std::vector<std::vector<double>> getSobelVertical();
|
std::vector<std::vector<real_t>> getSobelVertical();
|
||||||
std::vector<std::vector<double>> getScharrHorizontal();
|
std::vector<std::vector<real_t>> getScharrHorizontal();
|
||||||
std::vector<std::vector<double>> getScharrVertical();
|
std::vector<std::vector<real_t>> getScharrVertical();
|
||||||
std::vector<std::vector<double>> getRobertsHorizontal();
|
std::vector<std::vector<real_t>> getRobertsHorizontal();
|
||||||
std::vector<std::vector<double>> getRobertsVertical();
|
std::vector<std::vector<real_t>> getRobertsVertical();
|
||||||
|
|
||||||
private:
|
private:
|
||||||
std::vector<std::vector<double>> prewittHorizontal;
|
std::vector<std::vector<real_t>> prewittHorizontal;
|
||||||
std::vector<std::vector<double>> prewittVertical;
|
std::vector<std::vector<real_t>> prewittVertical;
|
||||||
std::vector<std::vector<double>> sobelHorizontal;
|
std::vector<std::vector<real_t>> sobelHorizontal;
|
||||||
std::vector<std::vector<double>> sobelVertical;
|
std::vector<std::vector<real_t>> sobelVertical;
|
||||||
std::vector<std::vector<double>> scharrHorizontal;
|
std::vector<std::vector<real_t>> scharrHorizontal;
|
||||||
std::vector<std::vector<double>> scharrVertical;
|
std::vector<std::vector<real_t>> scharrVertical;
|
||||||
std::vector<std::vector<double>> robertsHorizontal;
|
std::vector<std::vector<real_t>> robertsHorizontal;
|
||||||
std::vector<std::vector<double>> robertsVertical;
|
std::vector<std::vector<real_t>> robertsVertical;
|
||||||
};
|
};
|
||||||
|
|
||||||
#endif // Convolutions_hpp
|
#endif // Convolutions_hpp
|
@ -11,16 +11,16 @@
|
|||||||
#include <iostream>
|
#include <iostream>
|
||||||
|
|
||||||
|
|
||||||
double MLPPCost::MSE(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPCost::MSE(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
sum += (y_hat[i] - y[i]) * (y_hat[i] - y[i]);
|
sum += (y_hat[i] - y[i]) * (y_hat[i] - y[i]);
|
||||||
}
|
}
|
||||||
return sum / 2 * y_hat.size();
|
return sum / 2 * y_hat.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::MSE(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
real_t MLPPCost::MSE(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
for (int j = 0; j < y_hat[i].size(); j++) {
|
for (int j = 0; j < y_hat[i].size(); j++) {
|
||||||
sum += (y_hat[i][j] - y[i][j]) * (y_hat[i][j] - y[i][j]);
|
sum += (y_hat[i][j] - y[i][j]) * (y_hat[i][j] - y[i][j]);
|
||||||
@ -29,26 +29,26 @@ double MLPPCost::MSE(std::vector<std::vector<double>> y_hat, std::vector<std::ve
|
|||||||
return sum / 2 * y_hat.size();
|
return sum / 2 * y_hat.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPCost::MSEDeriv(std::vector<double> y_hat, std::vector<double> y) {
|
std::vector<real_t> MLPPCost::MSEDeriv(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.subtraction(y_hat, y);
|
return alg.subtraction(y_hat, y);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPCost::MSEDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
std::vector<std::vector<real_t>> MLPPCost::MSEDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.subtraction(y_hat, y);
|
return alg.subtraction(y_hat, y);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::RMSE(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPCost::RMSE(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
sum += (y_hat[i] - y[i]) * (y_hat[i] - y[i]);
|
sum += (y_hat[i] - y[i]) * (y_hat[i] - y[i]);
|
||||||
}
|
}
|
||||||
return sqrt(sum / y_hat.size());
|
return sqrt(sum / y_hat.size());
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::RMSE(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
real_t MLPPCost::RMSE(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
for (int j = 0; j < y_hat[i].size(); j++) {
|
for (int j = 0; j < y_hat[i].size(); j++) {
|
||||||
sum += (y_hat[i][j] - y[i][j]) * (y_hat[i][j] - y[i][j]);
|
sum += (y_hat[i][j] - y[i][j]) * (y_hat[i][j] - y[i][j]);
|
||||||
@ -57,26 +57,26 @@ double MLPPCost::RMSE(std::vector<std::vector<double>> y_hat, std::vector<std::v
|
|||||||
return sqrt(sum / y_hat.size());
|
return sqrt(sum / y_hat.size());
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPCost::RMSEDeriv(std::vector<double> y_hat, std::vector<double> y) {
|
std::vector<real_t> MLPPCost::RMSEDeriv(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.scalarMultiply(1 / (2 * sqrt(MSE(y_hat, y))), MSEDeriv(y_hat, y));
|
return alg.scalarMultiply(1 / (2 * sqrt(MSE(y_hat, y))), MSEDeriv(y_hat, y));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPCost::RMSEDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
std::vector<std::vector<real_t>> MLPPCost::RMSEDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.scalarMultiply(1 / (2 / sqrt(MSE(y_hat, y))), MSEDeriv(y_hat, y));
|
return alg.scalarMultiply(1 / (2 / sqrt(MSE(y_hat, y))), MSEDeriv(y_hat, y));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::MAE(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPCost::MAE(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
sum += abs((y_hat[i] - y[i]));
|
sum += abs((y_hat[i] - y[i]));
|
||||||
}
|
}
|
||||||
return sum / y_hat.size();
|
return sum / y_hat.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::MAE(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
real_t MLPPCost::MAE(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
for (int j = 0; j < y_hat[i].size(); j++) {
|
for (int j = 0; j < y_hat[i].size(); j++) {
|
||||||
sum += abs((y_hat[i][j] - y[i][j]));
|
sum += abs((y_hat[i][j] - y[i][j]));
|
||||||
@ -85,8 +85,8 @@ double MLPPCost::MAE(std::vector<std::vector<double>> y_hat, std::vector<std::ve
|
|||||||
return sum / y_hat.size();
|
return sum / y_hat.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPCost::MAEDeriv(std::vector<double> y_hat, std::vector<double> y) {
|
std::vector<real_t> MLPPCost::MAEDeriv(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
std::vector<double> deriv;
|
std::vector<real_t> deriv;
|
||||||
deriv.resize(y_hat.size());
|
deriv.resize(y_hat.size());
|
||||||
for (int i = 0; i < deriv.size(); i++) {
|
for (int i = 0; i < deriv.size(); i++) {
|
||||||
if (y_hat[i] < 0) {
|
if (y_hat[i] < 0) {
|
||||||
@ -100,8 +100,8 @@ std::vector<double> MLPPCost::MAEDeriv(std::vector<double> y_hat, std::vector<do
|
|||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPCost::MAEDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
std::vector<std::vector<real_t>> MLPPCost::MAEDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
std::vector<std::vector<double>> deriv;
|
std::vector<std::vector<real_t>> deriv;
|
||||||
deriv.resize(y_hat.size());
|
deriv.resize(y_hat.size());
|
||||||
for (int i = 0; i < deriv.size(); i++) {
|
for (int i = 0; i < deriv.size(); i++) {
|
||||||
deriv.resize(y_hat[i].size());
|
deriv.resize(y_hat[i].size());
|
||||||
@ -120,16 +120,16 @@ std::vector<std::vector<double>> MLPPCost::MAEDeriv(std::vector<std::vector<doub
|
|||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::MBE(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPCost::MBE(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
sum += (y_hat[i] - y[i]);
|
sum += (y_hat[i] - y[i]);
|
||||||
}
|
}
|
||||||
return sum / y_hat.size();
|
return sum / y_hat.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::MBE(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
real_t MLPPCost::MBE(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
for (int j = 0; j < y_hat[i].size(); j++) {
|
for (int j = 0; j < y_hat[i].size(); j++) {
|
||||||
sum += (y_hat[i][j] - y[i][j]);
|
sum += (y_hat[i][j] - y[i][j]);
|
||||||
@ -138,19 +138,19 @@ double MLPPCost::MBE(std::vector<std::vector<double>> y_hat, std::vector<std::ve
|
|||||||
return sum / y_hat.size();
|
return sum / y_hat.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPCost::MBEDeriv(std::vector<double> y_hat, std::vector<double> y) {
|
std::vector<real_t> MLPPCost::MBEDeriv(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.onevec(y_hat.size());
|
return alg.onevec(y_hat.size());
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPCost::MBEDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
std::vector<std::vector<real_t>> MLPPCost::MBEDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.onemat(y_hat.size(), y_hat[0].size());
|
return alg.onemat(y_hat.size(), y_hat[0].size());
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::LogLoss(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPCost::LogLoss(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
double eps = 1e-8;
|
real_t eps = 1e-8;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
sum += -(y[i] * std::log(y_hat[i] + eps) + (1 - y[i]) * std::log(1 - y_hat[i] + eps));
|
sum += -(y[i] * std::log(y_hat[i] + eps) + (1 - y[i]) * std::log(1 - y_hat[i] + eps));
|
||||||
}
|
}
|
||||||
@ -158,9 +158,9 @@ double MLPPCost::LogLoss(std::vector<double> y_hat, std::vector<double> y) {
|
|||||||
return sum / y_hat.size();
|
return sum / y_hat.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::LogLoss(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
real_t MLPPCost::LogLoss(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
double eps = 1e-8;
|
real_t eps = 1e-8;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
for (int j = 0; j < y_hat[i].size(); j++) {
|
for (int j = 0; j < y_hat[i].size(); j++) {
|
||||||
sum += -(y[i][j] * std::log(y_hat[i][j] + eps) + (1 - y[i][j]) * std::log(1 - y_hat[i][j] + eps));
|
sum += -(y[i][j] * std::log(y_hat[i][j] + eps) + (1 - y[i][j]) * std::log(1 - y_hat[i][j] + eps));
|
||||||
@ -170,18 +170,18 @@ double MLPPCost::LogLoss(std::vector<std::vector<double>> y_hat, std::vector<std
|
|||||||
return sum / y_hat.size();
|
return sum / y_hat.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPCost::LogLossDeriv(std::vector<double> y_hat, std::vector<double> y) {
|
std::vector<real_t> MLPPCost::LogLossDeriv(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.addition(alg.scalarMultiply(-1, alg.elementWiseDivision(y, y_hat)), alg.elementWiseDivision(alg.scalarMultiply(-1, alg.scalarAdd(-1, y)), alg.scalarMultiply(-1, alg.scalarAdd(-1, y_hat))));
|
return alg.addition(alg.scalarMultiply(-1, alg.elementWiseDivision(y, y_hat)), alg.elementWiseDivision(alg.scalarMultiply(-1, alg.scalarAdd(-1, y)), alg.scalarMultiply(-1, alg.scalarAdd(-1, y_hat))));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPCost::LogLossDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
std::vector<std::vector<real_t>> MLPPCost::LogLossDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.addition(alg.scalarMultiply(-1, alg.elementWiseDivision(y, y_hat)), alg.elementWiseDivision(alg.scalarMultiply(-1, alg.scalarAdd(-1, y)), alg.scalarMultiply(-1, alg.scalarAdd(-1, y_hat))));
|
return alg.addition(alg.scalarMultiply(-1, alg.elementWiseDivision(y, y_hat)), alg.elementWiseDivision(alg.scalarMultiply(-1, alg.scalarAdd(-1, y)), alg.scalarMultiply(-1, alg.scalarAdd(-1, y_hat))));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::CrossEntropy(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPCost::CrossEntropy(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
sum += y[i] * std::log(y_hat[i]);
|
sum += y[i] * std::log(y_hat[i]);
|
||||||
}
|
}
|
||||||
@ -189,8 +189,8 @@ double MLPPCost::CrossEntropy(std::vector<double> y_hat, std::vector<double> y)
|
|||||||
return -1 * sum;
|
return -1 * sum;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::CrossEntropy(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
real_t MLPPCost::CrossEntropy(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
for (int j = 0; j < y_hat[i].size(); j++) {
|
for (int j = 0; j < y_hat[i].size(); j++) {
|
||||||
sum += y[i][j] * std::log(y_hat[i][j]);
|
sum += y[i][j] * std::log(y_hat[i][j]);
|
||||||
@ -200,19 +200,19 @@ double MLPPCost::CrossEntropy(std::vector<std::vector<double>> y_hat, std::vecto
|
|||||||
return -1 * sum;
|
return -1 * sum;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPCost::CrossEntropyDeriv(std::vector<double> y_hat, std::vector<double> y) {
|
std::vector<real_t> MLPPCost::CrossEntropyDeriv(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.scalarMultiply(-1, alg.elementWiseDivision(y, y_hat));
|
return alg.scalarMultiply(-1, alg.elementWiseDivision(y, y_hat));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPCost::CrossEntropyDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
std::vector<std::vector<real_t>> MLPPCost::CrossEntropyDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.scalarMultiply(-1, alg.elementWiseDivision(y, y_hat));
|
return alg.scalarMultiply(-1, alg.elementWiseDivision(y, y_hat));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::HuberLoss(std::vector<double> y_hat, std::vector<double> y, double delta) {
|
real_t MLPPCost::HuberLoss(std::vector<real_t> y_hat, std::vector<real_t> y, real_t delta) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
if (abs(y[i] - y_hat[i]) <= delta) {
|
if (abs(y[i] - y_hat[i]) <= delta) {
|
||||||
sum += (y[i] - y_hat[i]) * (y[i] - y_hat[i]);
|
sum += (y[i] - y_hat[i]) * (y[i] - y_hat[i]);
|
||||||
@ -223,9 +223,9 @@ double MLPPCost::HuberLoss(std::vector<double> y_hat, std::vector<double> y, dou
|
|||||||
return sum;
|
return sum;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::HuberLoss(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y, double delta) {
|
real_t MLPPCost::HuberLoss(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y, real_t delta) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
for (int j = 0; j < y_hat[i].size(); j++) {
|
for (int j = 0; j < y_hat[i].size(); j++) {
|
||||||
if (abs(y[i][j] - y_hat[i][j]) <= delta) {
|
if (abs(y[i][j] - y_hat[i][j]) <= delta) {
|
||||||
@ -238,10 +238,10 @@ double MLPPCost::HuberLoss(std::vector<std::vector<double>> y_hat, std::vector<s
|
|||||||
return sum;
|
return sum;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPCost::HuberLossDeriv(std::vector<double> y_hat, std::vector<double> y, double delta) {
|
std::vector<real_t> MLPPCost::HuberLossDeriv(std::vector<real_t> y_hat, std::vector<real_t> y, real_t delta) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
std::vector<double> deriv;
|
std::vector<real_t> deriv;
|
||||||
deriv.resize(y_hat.size());
|
deriv.resize(y_hat.size());
|
||||||
|
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
@ -258,10 +258,10 @@ std::vector<double> MLPPCost::HuberLossDeriv(std::vector<double> y_hat, std::vec
|
|||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPCost::HuberLossDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y, double delta) {
|
std::vector<std::vector<real_t>> MLPPCost::HuberLossDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y, real_t delta) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
std::vector<std::vector<double>> deriv;
|
std::vector<std::vector<real_t>> deriv;
|
||||||
deriv.resize(y_hat.size());
|
deriv.resize(y_hat.size());
|
||||||
for (int i = 0; i < deriv.size(); i++) {
|
for (int i = 0; i < deriv.size(); i++) {
|
||||||
deriv[i].resize(y_hat[i].size());
|
deriv[i].resize(y_hat[i].size());
|
||||||
@ -283,8 +283,8 @@ std::vector<std::vector<double>> MLPPCost::HuberLossDeriv(std::vector<std::vecto
|
|||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::HingeLoss(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPCost::HingeLoss(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
sum += fmax(0, 1 - y[i] * y_hat[i]);
|
sum += fmax(0, 1 - y[i] * y_hat[i]);
|
||||||
}
|
}
|
||||||
@ -292,8 +292,8 @@ double MLPPCost::HingeLoss(std::vector<double> y_hat, std::vector<double> y) {
|
|||||||
return sum / y_hat.size();
|
return sum / y_hat.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::HingeLoss(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
real_t MLPPCost::HingeLoss(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
for (int j = 0; j < y_hat[i].size(); j++) {
|
for (int j = 0; j < y_hat[i].size(); j++) {
|
||||||
sum += fmax(0, 1 - y[i][j] * y_hat[i][j]);
|
sum += fmax(0, 1 - y[i][j] * y_hat[i][j]);
|
||||||
@ -303,8 +303,8 @@ double MLPPCost::HingeLoss(std::vector<std::vector<double>> y_hat, std::vector<s
|
|||||||
return sum / y_hat.size();
|
return sum / y_hat.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPCost::HingeLossDeriv(std::vector<double> y_hat, std::vector<double> y) {
|
std::vector<real_t> MLPPCost::HingeLossDeriv(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
std::vector<double> deriv;
|
std::vector<real_t> deriv;
|
||||||
deriv.resize(y_hat.size());
|
deriv.resize(y_hat.size());
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
if (1 - y[i] * y_hat[i] > 0) {
|
if (1 - y[i] * y_hat[i] > 0) {
|
||||||
@ -316,8 +316,8 @@ std::vector<double> MLPPCost::HingeLossDeriv(std::vector<double> y_hat, std::vec
|
|||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPCost::HingeLossDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
std::vector<std::vector<real_t>> MLPPCost::HingeLossDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
std::vector<std::vector<double>> deriv;
|
std::vector<std::vector<real_t>> deriv;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
for (int j = 0; j < y_hat[i].size(); j++) {
|
for (int j = 0; j < y_hat[i].size(); j++) {
|
||||||
if (1 - y[i][j] * y_hat[i][j] > 0) {
|
if (1 - y[i][j] * y_hat[i][j] > 0) {
|
||||||
@ -330,16 +330,16 @@ std::vector<std::vector<double>> MLPPCost::HingeLossDeriv(std::vector<std::vecto
|
|||||||
return deriv;
|
return deriv;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::WassersteinLoss(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPCost::WassersteinLoss(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
sum += y_hat[i] * y[i];
|
sum += y_hat[i] * y[i];
|
||||||
}
|
}
|
||||||
return -sum / y_hat.size();
|
return -sum / y_hat.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::WassersteinLoss(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
real_t MLPPCost::WassersteinLoss(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
for (int j = 0; j < y_hat[i].size(); j++) {
|
for (int j = 0; j < y_hat[i].size(); j++) {
|
||||||
sum += y_hat[i][j] * y[i][j];
|
sum += y_hat[i][j] * y[i][j];
|
||||||
@ -348,59 +348,59 @@ double MLPPCost::WassersteinLoss(std::vector<std::vector<double>> y_hat, std::ve
|
|||||||
return -sum / y_hat.size();
|
return -sum / y_hat.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPCost::WassersteinLossDeriv(std::vector<double> y_hat, std::vector<double> y) {
|
std::vector<real_t> MLPPCost::WassersteinLossDeriv(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.scalarMultiply(-1, y); // Simple.
|
return alg.scalarMultiply(-1, y); // Simple.
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPCost::WassersteinLossDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
std::vector<std::vector<real_t>> MLPPCost::WassersteinLossDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.scalarMultiply(-1, y); // Simple.
|
return alg.scalarMultiply(-1, y); // Simple.
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::HingeLoss(std::vector<double> y_hat, std::vector<double> y, std::vector<double> weights, double C) {
|
real_t MLPPCost::HingeLoss(std::vector<real_t> y_hat, std::vector<real_t> y, std::vector<real_t> weights, real_t C) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
return C * HingeLoss(y_hat, y) + regularization.regTerm(weights, 1, 0, "Ridge");
|
return C * HingeLoss(y_hat, y) + regularization.regTerm(weights, 1, 0, "Ridge");
|
||||||
}
|
}
|
||||||
double MLPPCost::HingeLoss(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y, std::vector<std::vector<double>> weights, double C) {
|
real_t MLPPCost::HingeLoss(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y, std::vector<std::vector<real_t>> weights, real_t C) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
return C * HingeLoss(y_hat, y) + regularization.regTerm(weights, 1, 0, "Ridge");
|
return C * HingeLoss(y_hat, y) + regularization.regTerm(weights, 1, 0, "Ridge");
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPCost::HingeLossDeriv(std::vector<double> y_hat, std::vector<double> y, double C) {
|
std::vector<real_t> MLPPCost::HingeLossDeriv(std::vector<real_t> y_hat, std::vector<real_t> y, real_t C) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
return alg.scalarMultiply(C, HingeLossDeriv(y_hat, y));
|
return alg.scalarMultiply(C, HingeLossDeriv(y_hat, y));
|
||||||
}
|
}
|
||||||
std::vector<std::vector<double>> MLPPCost::HingeLossDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y, double C) {
|
std::vector<std::vector<real_t>> MLPPCost::HingeLossDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y, real_t C) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
return alg.scalarMultiply(C, HingeLossDeriv(y_hat, y));
|
return alg.scalarMultiply(C, HingeLossDeriv(y_hat, y));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPCost::dualFormSVM(std::vector<double> alpha, std::vector<std::vector<double>> X, std::vector<double> y) {
|
real_t MLPPCost::dualFormSVM(std::vector<real_t> alpha, std::vector<std::vector<real_t>> X, std::vector<real_t> y) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<double>> Y = alg.diag(y); // Y is a diagnoal matrix. Y[i][j] = y[i] if i = i, else Y[i][j] = 0. Yt = Y.
|
std::vector<std::vector<real_t>> Y = alg.diag(y); // Y is a diagnoal matrix. Y[i][j] = y[i] if i = i, else Y[i][j] = 0. Yt = Y.
|
||||||
std::vector<std::vector<double>> K = alg.matmult(X, alg.transpose(X)); // TO DO: DON'T forget to add non-linear kernelizations.
|
std::vector<std::vector<real_t>> K = alg.matmult(X, alg.transpose(X)); // TO DO: DON'T forget to add non-linear kernelizations.
|
||||||
std::vector<std::vector<double>> Q = alg.matmult(alg.matmult(alg.transpose(Y), K), Y);
|
std::vector<std::vector<real_t>> Q = alg.matmult(alg.matmult(alg.transpose(Y), K), Y);
|
||||||
double alphaQ = alg.matmult(alg.matmult({ alpha }, Q), alg.transpose({ alpha }))[0][0];
|
real_t alphaQ = alg.matmult(alg.matmult({ alpha }, Q), alg.transpose({ alpha }))[0][0];
|
||||||
std::vector<double> one = alg.onevec(alpha.size());
|
std::vector<real_t> one = alg.onevec(alpha.size());
|
||||||
|
|
||||||
return -alg.dot(one, alpha) + 0.5 * alphaQ;
|
return -alg.dot(one, alpha) + 0.5 * alphaQ;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPCost::dualFormSVMDeriv(std::vector<double> alpha, std::vector<std::vector<double>> X, std::vector<double> y) {
|
std::vector<real_t> MLPPCost::dualFormSVMDeriv(std::vector<real_t> alpha, std::vector<std::vector<real_t>> X, std::vector<real_t> y) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<double>> Y = alg.zeromat(y.size(), y.size());
|
std::vector<std::vector<real_t>> Y = alg.zeromat(y.size(), y.size());
|
||||||
for (int i = 0; i < y.size(); i++) {
|
for (int i = 0; i < y.size(); i++) {
|
||||||
Y[i][i] = y[i]; // Y is a diagnoal matrix. Y[i][j] = y[i] if i = i, else Y[i][j] = 0. Yt = Y.
|
Y[i][i] = y[i]; // Y is a diagnoal matrix. Y[i][j] = y[i] if i = i, else Y[i][j] = 0. Yt = Y.
|
||||||
}
|
}
|
||||||
std::vector<std::vector<double>> K = alg.matmult(X, alg.transpose(X)); // TO DO: DON'T forget to add non-linear kernelizations.
|
std::vector<std::vector<real_t>> K = alg.matmult(X, alg.transpose(X)); // TO DO: DON'T forget to add non-linear kernelizations.
|
||||||
std::vector<std::vector<double>> Q = alg.matmult(alg.matmult(alg.transpose(Y), K), Y);
|
std::vector<std::vector<real_t>> Q = alg.matmult(alg.matmult(alg.transpose(Y), K), Y);
|
||||||
std::vector<double> alphaQDeriv = alg.mat_vec_mult(Q, alpha);
|
std::vector<real_t> alphaQDeriv = alg.mat_vec_mult(Q, alpha);
|
||||||
std::vector<double> one = alg.onevec(alpha.size());
|
std::vector<real_t> one = alg.onevec(alpha.size());
|
||||||
|
|
||||||
return alg.subtraction(alphaQDeriv, one);
|
return alg.subtraction(alphaQDeriv, one);
|
||||||
}
|
}
|
||||||
|
@ -8,76 +8,78 @@
|
|||||||
// Created by Marc Melikyan on 1/16/21.
|
// Created by Marc Melikyan on 1/16/21.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
|
|
||||||
class MLPPCost {
|
class MLPPCost {
|
||||||
public:
|
public:
|
||||||
// Regression Costs
|
// Regression Costs
|
||||||
double MSE(std::vector<double> y_hat, std::vector<double> y);
|
real_t MSE(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
double MSE(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
real_t MSE(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
std::vector<double> MSEDeriv(std::vector<double> y_hat, std::vector<double> y);
|
std::vector<real_t> MSEDeriv(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
std::vector<std::vector<double>> MSEDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
std::vector<std::vector<real_t>> MSEDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
double RMSE(std::vector<double> y_hat, std::vector<double> y);
|
real_t RMSE(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
double RMSE(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
real_t RMSE(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
std::vector<double> RMSEDeriv(std::vector<double> y_hat, std::vector<double> y);
|
std::vector<real_t> RMSEDeriv(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
std::vector<std::vector<double>> RMSEDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
std::vector<std::vector<real_t>> RMSEDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
double MAE(std::vector<double> y_hat, std::vector<double> y);
|
real_t MAE(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
double MAE(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
real_t MAE(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
std::vector<double> MAEDeriv(std::vector<double> y_hat, std::vector<double> y);
|
std::vector<real_t> MAEDeriv(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
std::vector<std::vector<double>> MAEDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
std::vector<std::vector<real_t>> MAEDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
double MBE(std::vector<double> y_hat, std::vector<double> y);
|
real_t MBE(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
double MBE(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
real_t MBE(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
std::vector<double> MBEDeriv(std::vector<double> y_hat, std::vector<double> y);
|
std::vector<real_t> MBEDeriv(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
std::vector<std::vector<double>> MBEDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
std::vector<std::vector<real_t>> MBEDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
// Classification Costs
|
// Classification Costs
|
||||||
double LogLoss(std::vector<double> y_hat, std::vector<double> y);
|
real_t LogLoss(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
double LogLoss(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
real_t LogLoss(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
std::vector<double> LogLossDeriv(std::vector<double> y_hat, std::vector<double> y);
|
std::vector<real_t> LogLossDeriv(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
std::vector<std::vector<double>> LogLossDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
std::vector<std::vector<real_t>> LogLossDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
double CrossEntropy(std::vector<double> y_hat, std::vector<double> y);
|
real_t CrossEntropy(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
double CrossEntropy(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
real_t CrossEntropy(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
std::vector<double> CrossEntropyDeriv(std::vector<double> y_hat, std::vector<double> y);
|
std::vector<real_t> CrossEntropyDeriv(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
std::vector<std::vector<double>> CrossEntropyDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
std::vector<std::vector<real_t>> CrossEntropyDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
double HuberLoss(std::vector<double> y_hat, std::vector<double> y, double delta);
|
real_t HuberLoss(std::vector<real_t> y_hat, std::vector<real_t> y, real_t delta);
|
||||||
double HuberLoss(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y, double delta);
|
real_t HuberLoss(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y, real_t delta);
|
||||||
|
|
||||||
std::vector<double> HuberLossDeriv(std::vector<double> y_hat, std::vector<double> y, double delta);
|
std::vector<real_t> HuberLossDeriv(std::vector<real_t> y_hat, std::vector<real_t> y, real_t delta);
|
||||||
std::vector<std::vector<double>> HuberLossDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y, double delta);
|
std::vector<std::vector<real_t>> HuberLossDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y, real_t delta);
|
||||||
|
|
||||||
double HingeLoss(std::vector<double> y_hat, std::vector<double> y);
|
real_t HingeLoss(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
double HingeLoss(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
real_t HingeLoss(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
std::vector<double> HingeLossDeriv(std::vector<double> y_hat, std::vector<double> y);
|
std::vector<real_t> HingeLossDeriv(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
std::vector<std::vector<double>> HingeLossDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
std::vector<std::vector<real_t>> HingeLossDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
double HingeLoss(std::vector<double> y_hat, std::vector<double> y, std::vector<double> weights, double C);
|
real_t HingeLoss(std::vector<real_t> y_hat, std::vector<real_t> y, std::vector<real_t> weights, real_t C);
|
||||||
double HingeLoss(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y, std::vector<std::vector<double>> weights, double C);
|
real_t HingeLoss(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y, std::vector<std::vector<real_t>> weights, real_t C);
|
||||||
|
|
||||||
std::vector<double> HingeLossDeriv(std::vector<double> y_hat, std::vector<double> y, double C);
|
std::vector<real_t> HingeLossDeriv(std::vector<real_t> y_hat, std::vector<real_t> y, real_t C);
|
||||||
std::vector<std::vector<double>> HingeLossDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y, double C);
|
std::vector<std::vector<real_t>> HingeLossDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y, real_t C);
|
||||||
|
|
||||||
double WassersteinLoss(std::vector<double> y_hat, std::vector<double> y);
|
real_t WassersteinLoss(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
double WassersteinLoss(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
real_t WassersteinLoss(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
std::vector<double> WassersteinLossDeriv(std::vector<double> y_hat, std::vector<double> y);
|
std::vector<real_t> WassersteinLossDeriv(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
std::vector<std::vector<double>> WassersteinLossDeriv(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
std::vector<std::vector<real_t>> WassersteinLossDeriv(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
double dualFormSVM(std::vector<double> alpha, std::vector<std::vector<double>> X, std::vector<double> y); // TO DO: DON'T forget to add non-linear kernelizations.
|
real_t dualFormSVM(std::vector<real_t> alpha, std::vector<std::vector<real_t>> X, std::vector<real_t> y); // TO DO: DON'T forget to add non-linear kernelizations.
|
||||||
|
|
||||||
std::vector<double> dualFormSVMDeriv(std::vector<double> alpha, std::vector<std::vector<double>> X, std::vector<double> y);
|
std::vector<real_t> dualFormSVMDeriv(std::vector<real_t> alpha, std::vector<std::vector<real_t>> X, std::vector<real_t> y);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
};
|
};
|
||||||
|
@ -56,7 +56,7 @@ Ref<MLPPDataComplex> MLPPData::load_iris(const String &path) {
|
|||||||
const int IRIS_SIZE = 4;
|
const int IRIS_SIZE = 4;
|
||||||
const int ONE_HOT_NUM = 3;
|
const int ONE_HOT_NUM = 3;
|
||||||
|
|
||||||
std::vector<double> tempOutputSet;
|
std::vector<real_t> tempOutputSet;
|
||||||
|
|
||||||
Ref<MLPPDataComplex> data;
|
Ref<MLPPDataComplex> data;
|
||||||
data.instance();
|
data.instance();
|
||||||
@ -71,7 +71,7 @@ Ref<MLPPDataComplex> MLPPData::load_wine(const String &path) {
|
|||||||
const int WINE_SIZE = 4;
|
const int WINE_SIZE = 4;
|
||||||
const int ONE_HOT_NUM = 3;
|
const int ONE_HOT_NUM = 3;
|
||||||
|
|
||||||
std::vector<double> tempOutputSet;
|
std::vector<real_t> tempOutputSet;
|
||||||
|
|
||||||
Ref<MLPPDataComplex> data;
|
Ref<MLPPDataComplex> data;
|
||||||
data.instance();
|
data.instance();
|
||||||
@ -86,8 +86,8 @@ Ref<MLPPDataComplex> MLPPData::load_mnist_train(const String &path) {
|
|||||||
const int MNIST_SIZE = 784;
|
const int MNIST_SIZE = 784;
|
||||||
const int ONE_HOT_NUM = 10;
|
const int ONE_HOT_NUM = 10;
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> tempOutputSet;
|
std::vector<real_t> tempOutputSet;
|
||||||
|
|
||||||
Ref<MLPPDataComplex> data;
|
Ref<MLPPDataComplex> data;
|
||||||
data.instance();
|
data.instance();
|
||||||
@ -101,8 +101,8 @@ Ref<MLPPDataComplex> MLPPData::load_mnist_train(const String &path) {
|
|||||||
Ref<MLPPDataComplex> MLPPData::load_mnist_test(const String &path) {
|
Ref<MLPPDataComplex> MLPPData::load_mnist_test(const String &path) {
|
||||||
const int MNIST_SIZE = 784;
|
const int MNIST_SIZE = 784;
|
||||||
const int ONE_HOT_NUM = 10;
|
const int ONE_HOT_NUM = 10;
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> tempOutputSet;
|
std::vector<real_t> tempOutputSet;
|
||||||
|
|
||||||
Ref<MLPPDataComplex> data;
|
Ref<MLPPDataComplex> data;
|
||||||
data.instance();
|
data.instance();
|
||||||
@ -137,7 +137,7 @@ Ref<MLPPDataESimple> MLPPData::load_fires_and_crime(const String &path) {
|
|||||||
|
|
||||||
// MULTIVARIATE SUPERVISED
|
// MULTIVARIATE SUPERVISED
|
||||||
|
|
||||||
void MLPPData::set_data_supervised(int k, const String &file_name, std::vector<std::vector<double>> &inputSet, std::vector<double> &outputSet) {
|
void MLPPData::set_data_supervised(int k, const String &file_name, std::vector<std::vector<real_t>> &inputSet, std::vector<real_t> &outputSet) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
inputSet.resize(k);
|
inputSet.resize(k);
|
||||||
@ -150,10 +150,10 @@ void MLPPData::set_data_supervised(int k, const String &file_name, std::vector<s
|
|||||||
Vector<String> ll = file->get_csv_line();
|
Vector<String> ll = file->get_csv_line();
|
||||||
|
|
||||||
for (int i = 0; i < k; ++i) {
|
for (int i = 0; i < k; ++i) {
|
||||||
inputSet[i].push_back(ll[i].to_double());
|
inputSet[i].push_back(static_cast<real_t>(ll[i].to_double()));
|
||||||
}
|
}
|
||||||
|
|
||||||
outputSet.push_back(ll[k].to_double());
|
outputSet.push_back(static_cast<real_t>(ll[k].to_double()));
|
||||||
}
|
}
|
||||||
|
|
||||||
inputSet = alg.transpose(inputSet);
|
inputSet = alg.transpose(inputSet);
|
||||||
@ -161,7 +161,7 @@ void MLPPData::set_data_supervised(int k, const String &file_name, std::vector<s
|
|||||||
memdelete(file);
|
memdelete(file);
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPData::set_data_unsupervised(int k, const String &file_name, std::vector<std::vector<double>> &inputSet) {
|
void MLPPData::set_data_unsupervised(int k, const String &file_name, std::vector<std::vector<real_t>> &inputSet) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
inputSet.resize(k);
|
inputSet.resize(k);
|
||||||
@ -174,7 +174,7 @@ void MLPPData::set_data_unsupervised(int k, const String &file_name, std::vector
|
|||||||
Vector<String> ll = file->get_csv_line();
|
Vector<String> ll = file->get_csv_line();
|
||||||
|
|
||||||
for (int i = 0; i < k; ++i) {
|
for (int i = 0; i < k; ++i) {
|
||||||
inputSet[i].push_back(ll[i].to_double());
|
inputSet[i].push_back(static_cast<real_t>(ll[i].to_double()));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -183,7 +183,7 @@ void MLPPData::set_data_unsupervised(int k, const String &file_name, std::vector
|
|||||||
memdelete(file);
|
memdelete(file);
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPData::set_data_simple(const String &file_name, std::vector<double> &inputSet, std::vector<double> &outputSet) {
|
void MLPPData::set_data_simple(const String &file_name, std::vector<real_t> &inputSet, std::vector<real_t> &outputSet) {
|
||||||
FileAccess *file = FileAccess::open(file_name, FileAccess::READ);
|
FileAccess *file = FileAccess::open(file_name, FileAccess::READ);
|
||||||
|
|
||||||
ERR_FAIL_COND(!file);
|
ERR_FAIL_COND(!file);
|
||||||
@ -192,15 +192,15 @@ void MLPPData::set_data_simple(const String &file_name, std::vector<double> &inp
|
|||||||
Vector<String> ll = file->get_csv_line();
|
Vector<String> ll = file->get_csv_line();
|
||||||
|
|
||||||
for (int i = 0; i < ll.size(); i += 2) {
|
for (int i = 0; i < ll.size(); i += 2) {
|
||||||
inputSet.push_back(ll[i].to_double());
|
inputSet.push_back(static_cast<real_t>(ll[i].to_double()));
|
||||||
outputSet.push_back(ll[i + 1].to_double());
|
outputSet.push_back(static_cast<real_t>(ll[i + 1].to_double()));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
memdelete(file);
|
memdelete(file);
|
||||||
}
|
}
|
||||||
|
|
||||||
MLPPData::SplitComplexData MLPPData::train_test_split(const Ref<MLPPDataComplex> &data, double test_size) {
|
MLPPData::SplitComplexData MLPPData::train_test_split(const Ref<MLPPDataComplex> &data, real_t test_size) {
|
||||||
SplitComplexData res;
|
SplitComplexData res;
|
||||||
|
|
||||||
res.train.instance();
|
res.train.instance();
|
||||||
@ -237,7 +237,7 @@ MLPPData::SplitComplexData MLPPData::train_test_split(const Ref<MLPPDataComplex>
|
|||||||
|
|
||||||
return res;
|
return res;
|
||||||
}
|
}
|
||||||
Array MLPPData::train_test_split_bind(const Ref<MLPPDataComplex> &data, double test_size) {
|
Array MLPPData::train_test_split_bind(const Ref<MLPPDataComplex> &data, real_t test_size) {
|
||||||
SplitComplexData res = train_test_split(data, test_size);
|
SplitComplexData res = train_test_split(data, test_size);
|
||||||
|
|
||||||
Array arr;
|
Array arr;
|
||||||
@ -248,80 +248,80 @@ Array MLPPData::train_test_split_bind(const Ref<MLPPDataComplex> &data, double t
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Loading Datasets
|
// Loading Datasets
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<double>> MLPPData::loadBreastCancer() {
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<real_t>> MLPPData::loadBreastCancer() {
|
||||||
const int BREAST_CANCER_SIZE = 30; // k = 30
|
const int BREAST_CANCER_SIZE = 30; // k = 30
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
|
|
||||||
setData(BREAST_CANCER_SIZE, "MLPP/Data/Datasets/BreastCancer.csv", inputSet, outputSet);
|
setData(BREAST_CANCER_SIZE, "MLPP/Data/Datasets/BreastCancer.csv", inputSet, outputSet);
|
||||||
return { inputSet, outputSet };
|
return { inputSet, outputSet };
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<double>> MLPPData::loadBreastCancerSVC() {
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<real_t>> MLPPData::loadBreastCancerSVC() {
|
||||||
const int BREAST_CANCER_SIZE = 30; // k = 30
|
const int BREAST_CANCER_SIZE = 30; // k = 30
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
|
|
||||||
setData(BREAST_CANCER_SIZE, "MLPP/Data/Datasets/BreastCancerSVM.csv", inputSet, outputSet);
|
setData(BREAST_CANCER_SIZE, "MLPP/Data/Datasets/BreastCancerSVM.csv", inputSet, outputSet);
|
||||||
return { inputSet, outputSet };
|
return { inputSet, outputSet };
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>> MLPPData::loadIris() {
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> MLPPData::loadIris() {
|
||||||
const int IRIS_SIZE = 4;
|
const int IRIS_SIZE = 4;
|
||||||
const int ONE_HOT_NUM = 3;
|
const int ONE_HOT_NUM = 3;
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> tempOutputSet;
|
std::vector<real_t> tempOutputSet;
|
||||||
|
|
||||||
setData(IRIS_SIZE, "/Users/marcmelikyan/Desktop/Data/Iris.csv", inputSet, tempOutputSet);
|
setData(IRIS_SIZE, "/Users/marcmelikyan/Desktop/Data/Iris.csv", inputSet, tempOutputSet);
|
||||||
std::vector<std::vector<double>> outputSet = oneHotRep(tempOutputSet, ONE_HOT_NUM);
|
std::vector<std::vector<real_t>> outputSet = oneHotRep(tempOutputSet, ONE_HOT_NUM);
|
||||||
return { inputSet, outputSet };
|
return { inputSet, outputSet };
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>> MLPPData::loadWine() {
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> MLPPData::loadWine() {
|
||||||
const int WINE_SIZE = 4;
|
const int WINE_SIZE = 4;
|
||||||
const int ONE_HOT_NUM = 3;
|
const int ONE_HOT_NUM = 3;
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> tempOutputSet;
|
std::vector<real_t> tempOutputSet;
|
||||||
|
|
||||||
setData(WINE_SIZE, "MLPP/Data/Datasets/Iris.csv", inputSet, tempOutputSet);
|
setData(WINE_SIZE, "MLPP/Data/Datasets/Iris.csv", inputSet, tempOutputSet);
|
||||||
std::vector<std::vector<double>> outputSet = oneHotRep(tempOutputSet, ONE_HOT_NUM);
|
std::vector<std::vector<real_t>> outputSet = oneHotRep(tempOutputSet, ONE_HOT_NUM);
|
||||||
return { inputSet, outputSet };
|
return { inputSet, outputSet };
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>> MLPPData::loadMnistTrain() {
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> MLPPData::loadMnistTrain() {
|
||||||
const int MNIST_SIZE = 784;
|
const int MNIST_SIZE = 784;
|
||||||
const int ONE_HOT_NUM = 10;
|
const int ONE_HOT_NUM = 10;
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> tempOutputSet;
|
std::vector<real_t> tempOutputSet;
|
||||||
|
|
||||||
setData(MNIST_SIZE, "MLPP/Data/Datasets/MnistTrain.csv", inputSet, tempOutputSet);
|
setData(MNIST_SIZE, "MLPP/Data/Datasets/MnistTrain.csv", inputSet, tempOutputSet);
|
||||||
std::vector<std::vector<double>> outputSet = oneHotRep(tempOutputSet, ONE_HOT_NUM);
|
std::vector<std::vector<real_t>> outputSet = oneHotRep(tempOutputSet, ONE_HOT_NUM);
|
||||||
return { inputSet, outputSet };
|
return { inputSet, outputSet };
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>> MLPPData::loadMnistTest() {
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> MLPPData::loadMnistTest() {
|
||||||
const int MNIST_SIZE = 784;
|
const int MNIST_SIZE = 784;
|
||||||
const int ONE_HOT_NUM = 10;
|
const int ONE_HOT_NUM = 10;
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> tempOutputSet;
|
std::vector<real_t> tempOutputSet;
|
||||||
|
|
||||||
setData(MNIST_SIZE, "MLPP/Data/Datasets/MnistTest.csv", inputSet, tempOutputSet);
|
setData(MNIST_SIZE, "MLPP/Data/Datasets/MnistTest.csv", inputSet, tempOutputSet);
|
||||||
std::vector<std::vector<double>> outputSet = oneHotRep(tempOutputSet, ONE_HOT_NUM);
|
std::vector<std::vector<real_t>> outputSet = oneHotRep(tempOutputSet, ONE_HOT_NUM);
|
||||||
return { inputSet, outputSet };
|
return { inputSet, outputSet };
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<double>> MLPPData::loadCaliforniaHousing() {
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<real_t>> MLPPData::loadCaliforniaHousing() {
|
||||||
const int CALIFORNIA_HOUSING_SIZE = 13; // k = 30
|
const int CALIFORNIA_HOUSING_SIZE = 13; // k = 30
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
|
|
||||||
setData(CALIFORNIA_HOUSING_SIZE, "MLPP/Data/Datasets/CaliforniaHousing.csv", inputSet, outputSet);
|
setData(CALIFORNIA_HOUSING_SIZE, "MLPP/Data/Datasets/CaliforniaHousing.csv", inputSet, outputSet);
|
||||||
return { inputSet, outputSet };
|
return { inputSet, outputSet };
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<double>, std::vector<double>> MLPPData::loadFiresAndCrime() {
|
std::tuple<std::vector<real_t>, std::vector<real_t>> MLPPData::loadFiresAndCrime() {
|
||||||
std::vector<double> inputSet; // k is implicitly 1.
|
std::vector<real_t> inputSet; // k is implicitly 1.
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
|
|
||||||
setData("MLPP/Data/Datasets/FiresAndCrime.csv", inputSet, outputSet);
|
setData("MLPP/Data/Datasets/FiresAndCrime.csv", inputSet, outputSet);
|
||||||
return { inputSet, outputSet };
|
return { inputSet, outputSet };
|
||||||
@ -330,15 +330,15 @@ std::tuple<std::vector<double>, std::vector<double>> MLPPData::loadFiresAndCrime
|
|||||||
// Note that inputs and outputs should be pairs (technically), but this
|
// Note that inputs and outputs should be pairs (technically), but this
|
||||||
// implementation will separate them. (My implementation keeps them tied together.)
|
// implementation will separate them. (My implementation keeps them tied together.)
|
||||||
// Not yet sure whether this is intentional or not (or it's something like a compiler specific difference)
|
// Not yet sure whether this is intentional or not (or it's something like a compiler specific difference)
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>, std::vector<std::vector<double>>, std::vector<std::vector<double>>> MLPPData::trainTestSplit(std::vector<std::vector<double>> inputSet, std::vector<std::vector<double>> outputSet, double testSize) {
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> MLPPData::trainTestSplit(std::vector<std::vector<real_t>> inputSet, std::vector<std::vector<real_t>> outputSet, real_t testSize) {
|
||||||
std::random_device rd;
|
std::random_device rd;
|
||||||
std::default_random_engine generator(rd());
|
std::default_random_engine generator(rd());
|
||||||
|
|
||||||
std::shuffle(inputSet.begin(), inputSet.end(), generator); // inputSet random shuffle
|
std::shuffle(inputSet.begin(), inputSet.end(), generator); // inputSet random shuffle
|
||||||
std::shuffle(outputSet.begin(), outputSet.end(), generator); // outputSet random shuffle)
|
std::shuffle(outputSet.begin(), outputSet.end(), generator); // outputSet random shuffle)
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputTestSet;
|
std::vector<std::vector<real_t>> inputTestSet;
|
||||||
std::vector<std::vector<double>> outputTestSet;
|
std::vector<std::vector<real_t>> outputTestSet;
|
||||||
|
|
||||||
int testInputNumber = testSize * inputSet.size(); // implicit usage of floor
|
int testInputNumber = testSize * inputSet.size(); // implicit usage of floor
|
||||||
int testOutputNumber = testSize * outputSet.size(); // implicit usage of floor
|
int testOutputNumber = testSize * outputSet.size(); // implicit usage of floor
|
||||||
@ -358,7 +358,7 @@ std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>, s
|
|||||||
|
|
||||||
// MULTIVARIATE SUPERVISED
|
// MULTIVARIATE SUPERVISED
|
||||||
|
|
||||||
void MLPPData::setData(int k, std::string fileName, std::vector<std::vector<double>> &inputSet, std::vector<double> &outputSet) {
|
void MLPPData::setData(int k, std::string fileName, std::vector<std::vector<real_t>> &inputSet, std::vector<real_t> &outputSet) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::string inputTemp;
|
std::string inputTemp;
|
||||||
std::string outputTemp;
|
std::string outputTemp;
|
||||||
@ -386,7 +386,7 @@ void MLPPData::setData(int k, std::string fileName, std::vector<std::vector<doub
|
|||||||
dataFile.close();
|
dataFile.close();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPData::printData(std::vector<std::string> inputName, std::string outputName, std::vector<std::vector<double>> inputSet, std::vector<double> outputSet) {
|
void MLPPData::printData(std::vector<std::string> inputName, std::string outputName, std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
inputSet = alg.transpose(inputSet);
|
inputSet = alg.transpose(inputSet);
|
||||||
for (int i = 0; i < inputSet.size(); i++) {
|
for (int i = 0; i < inputSet.size(); i++) {
|
||||||
@ -404,7 +404,7 @@ void MLPPData::printData(std::vector<std::string> inputName, std::string outputN
|
|||||||
|
|
||||||
// UNSUPERVISED
|
// UNSUPERVISED
|
||||||
|
|
||||||
void MLPPData::setData(int k, std::string fileName, std::vector<std::vector<double>> &inputSet) {
|
void MLPPData::setData(int k, std::string fileName, std::vector<std::vector<real_t>> &inputSet) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::string inputTemp;
|
std::string inputTemp;
|
||||||
|
|
||||||
@ -428,7 +428,7 @@ void MLPPData::setData(int k, std::string fileName, std::vector<std::vector<doub
|
|||||||
dataFile.close();
|
dataFile.close();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPData::printData(std::vector<std::string> inputName, std::vector<std::vector<double>> inputSet) {
|
void MLPPData::printData(std::vector<std::string> inputName, std::vector<std::vector<real_t>> inputSet) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
inputSet = alg.transpose(inputSet);
|
inputSet = alg.transpose(inputSet);
|
||||||
for (int i = 0; i < inputSet.size(); i++) {
|
for (int i = 0; i < inputSet.size(); i++) {
|
||||||
@ -441,7 +441,7 @@ void MLPPData::printData(std::vector<std::string> inputName, std::vector<std::ve
|
|||||||
|
|
||||||
// SIMPLE
|
// SIMPLE
|
||||||
|
|
||||||
void MLPPData::setData(std::string fileName, std::vector<double> &inputSet, std::vector<double> &outputSet) {
|
void MLPPData::setData(std::string fileName, std::vector<real_t> &inputSet, std::vector<real_t> &outputSet) {
|
||||||
std::string inputTemp, outputTemp;
|
std::string inputTemp, outputTemp;
|
||||||
|
|
||||||
std::ifstream dataFile(fileName);
|
std::ifstream dataFile(fileName);
|
||||||
@ -464,7 +464,7 @@ void MLPPData::setData(std::string fileName, std::vector<double> &inputSet, std:
|
|||||||
dataFile.close();
|
dataFile.close();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPData::printData(std::string &inputName, std::string &outputName, std::vector<double> &inputSet, std::vector<double> &outputSet) {
|
void MLPPData::printData(std::string &inputName, std::string &outputName, std::vector<real_t> &inputSet, std::vector<real_t> &outputSet) {
|
||||||
std::cout << inputName << std::endl;
|
std::cout << inputName << std::endl;
|
||||||
for (int i = 0; i < inputSet.size(); i++) {
|
for (int i = 0; i < inputSet.size(); i++) {
|
||||||
std::cout << inputSet[i] << std::endl;
|
std::cout << inputSet[i] << std::endl;
|
||||||
@ -477,8 +477,8 @@ void MLPPData::printData(std::string &inputName, std::string &outputName, std::v
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Images
|
// Images
|
||||||
std::vector<std::vector<double>> MLPPData::rgb2gray(std::vector<std::vector<std::vector<double>>> input) {
|
std::vector<std::vector<real_t>> MLPPData::rgb2gray(std::vector<std::vector<std::vector<real_t>>> input) {
|
||||||
std::vector<std::vector<double>> grayScale;
|
std::vector<std::vector<real_t>> grayScale;
|
||||||
grayScale.resize(input[0].size());
|
grayScale.resize(input[0].size());
|
||||||
for (int i = 0; i < grayScale.size(); i++) {
|
for (int i = 0; i < grayScale.size(); i++) {
|
||||||
grayScale[i].resize(input[0][i].size());
|
grayScale[i].resize(input[0][i].size());
|
||||||
@ -491,9 +491,9 @@ std::vector<std::vector<double>> MLPPData::rgb2gray(std::vector<std::vector<std:
|
|||||||
return grayScale;
|
return grayScale;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> MLPPData::rgb2ycbcr(std::vector<std::vector<std::vector<double>>> input) {
|
std::vector<std::vector<std::vector<real_t>>> MLPPData::rgb2ycbcr(std::vector<std::vector<std::vector<real_t>>> input) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<std::vector<double>>> YCbCr;
|
std::vector<std::vector<std::vector<real_t>>> YCbCr;
|
||||||
YCbCr = alg.resize(YCbCr, input);
|
YCbCr = alg.resize(YCbCr, input);
|
||||||
for (int i = 0; i < YCbCr[0].size(); i++) {
|
for (int i = 0; i < YCbCr[0].size(); i++) {
|
||||||
for (int j = 0; j < YCbCr[0][i].size(); j++) {
|
for (int j = 0; j < YCbCr[0][i].size(); j++) {
|
||||||
@ -507,19 +507,19 @@ std::vector<std::vector<std::vector<double>>> MLPPData::rgb2ycbcr(std::vector<st
|
|||||||
|
|
||||||
// Conversion formulas available here:
|
// Conversion formulas available here:
|
||||||
// https://www.rapidtables.com/convert/color/rgb-to-hsv.html
|
// https://www.rapidtables.com/convert/color/rgb-to-hsv.html
|
||||||
std::vector<std::vector<std::vector<double>>> MLPPData::rgb2hsv(std::vector<std::vector<std::vector<double>>> input) {
|
std::vector<std::vector<std::vector<real_t>>> MLPPData::rgb2hsv(std::vector<std::vector<std::vector<real_t>>> input) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<std::vector<double>>> HSV;
|
std::vector<std::vector<std::vector<real_t>>> HSV;
|
||||||
HSV = alg.resize(HSV, input);
|
HSV = alg.resize(HSV, input);
|
||||||
for (int i = 0; i < HSV[0].size(); i++) {
|
for (int i = 0; i < HSV[0].size(); i++) {
|
||||||
for (int j = 0; j < HSV[0][i].size(); j++) {
|
for (int j = 0; j < HSV[0][i].size(); j++) {
|
||||||
double rPrime = input[0][i][j] / 255;
|
real_t rPrime = input[0][i][j] / 255;
|
||||||
double gPrime = input[1][i][j] / 255;
|
real_t gPrime = input[1][i][j] / 255;
|
||||||
double bPrime = input[2][i][j] / 255;
|
real_t bPrime = input[2][i][j] / 255;
|
||||||
|
|
||||||
double cMax = alg.max({ rPrime, gPrime, bPrime });
|
real_t cMax = alg.max({ rPrime, gPrime, bPrime });
|
||||||
double cMin = alg.min({ rPrime, gPrime, bPrime });
|
real_t cMin = alg.min({ rPrime, gPrime, bPrime });
|
||||||
double delta = cMax - cMin;
|
real_t delta = cMax - cMin;
|
||||||
|
|
||||||
// H calculation.
|
// H calculation.
|
||||||
if (delta == 0) {
|
if (delta == 0) {
|
||||||
@ -549,19 +549,19 @@ std::vector<std::vector<std::vector<double>>> MLPPData::rgb2hsv(std::vector<std:
|
|||||||
}
|
}
|
||||||
|
|
||||||
// http://machinethatsees.blogspot.com/2013/07/how-to-convert-rgb-to-xyz-or-vice-versa.html
|
// http://machinethatsees.blogspot.com/2013/07/how-to-convert-rgb-to-xyz-or-vice-versa.html
|
||||||
std::vector<std::vector<std::vector<double>>> MLPPData::rgb2xyz(std::vector<std::vector<std::vector<double>>> input) {
|
std::vector<std::vector<std::vector<real_t>>> MLPPData::rgb2xyz(std::vector<std::vector<std::vector<real_t>>> input) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<std::vector<double>>> XYZ;
|
std::vector<std::vector<std::vector<real_t>>> XYZ;
|
||||||
XYZ = alg.resize(XYZ, input);
|
XYZ = alg.resize(XYZ, input);
|
||||||
std::vector<std::vector<double>> RGB2XYZ = { { 0.4124564, 0.3575761, 0.1804375 }, { 0.2126726, 0.7151522, 0.0721750 }, { 0.0193339, 0.1191920, 0.9503041 } };
|
std::vector<std::vector<real_t>> RGB2XYZ = { { 0.4124564, 0.3575761, 0.1804375 }, { 0.2126726, 0.7151522, 0.0721750 }, { 0.0193339, 0.1191920, 0.9503041 } };
|
||||||
return alg.vector_wise_tensor_product(input, RGB2XYZ);
|
return alg.vector_wise_tensor_product(input, RGB2XYZ);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> MLPPData::xyz2rgb(std::vector<std::vector<std::vector<double>>> input) {
|
std::vector<std::vector<std::vector<real_t>>> MLPPData::xyz2rgb(std::vector<std::vector<std::vector<real_t>>> input) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<std::vector<double>>> XYZ;
|
std::vector<std::vector<std::vector<real_t>>> XYZ;
|
||||||
XYZ = alg.resize(XYZ, input);
|
XYZ = alg.resize(XYZ, input);
|
||||||
std::vector<std::vector<double>> RGB2XYZ = alg.inverse({ { 0.4124564, 0.3575761, 0.1804375 }, { 0.2126726, 0.7151522, 0.0721750 }, { 0.0193339, 0.1191920, 0.9503041 } });
|
std::vector<std::vector<real_t>> RGB2XYZ = alg.inverse({ { 0.4124564, 0.3575761, 0.1804375 }, { 0.2126726, 0.7151522, 0.0721750 }, { 0.0193339, 0.1191920, 0.9503041 } });
|
||||||
return alg.vector_wise_tensor_product(input, RGB2XYZ);
|
return alg.vector_wise_tensor_product(input, RGB2XYZ);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -640,11 +640,11 @@ std::vector<std::string> MLPPData::segment(std::string text) {
|
|||||||
return segmented_data;
|
return segmented_data;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPData::tokenize(std::string text) {
|
std::vector<real_t> MLPPData::tokenize(std::string text) {
|
||||||
int max_num = 0;
|
int max_num = 0;
|
||||||
bool new_num = true;
|
bool new_num = true;
|
||||||
std::vector<std::string> segmented_data = segment(text);
|
std::vector<std::string> segmented_data = segment(text);
|
||||||
std::vector<double> tokenized_data;
|
std::vector<real_t> tokenized_data;
|
||||||
tokenized_data.resize(segmented_data.size());
|
tokenized_data.resize(segmented_data.size());
|
||||||
for (int i = 0; i < segmented_data.size(); i++) {
|
for (int i = 0; i < segmented_data.size(); i++) {
|
||||||
for (int j = i - 1; j >= 0; j--) {
|
for (int j = i - 1; j >= 0; j--) {
|
||||||
@ -710,7 +710,7 @@ std::string MLPPData::stemming(std::string text) {
|
|||||||
return text;
|
return text;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPData::BOW(std::vector<std::string> sentences, std::string type) {
|
std::vector<std::vector<real_t>> MLPPData::BOW(std::vector<std::string> sentences, std::string type) {
|
||||||
/*
|
/*
|
||||||
STEPS OF BOW:
|
STEPS OF BOW:
|
||||||
1) To lowercase (done by removeStopWords function by def)
|
1) To lowercase (done by removeStopWords function by def)
|
||||||
@ -729,7 +729,7 @@ std::vector<std::vector<double>> MLPPData::BOW(std::vector<std::string> sentence
|
|||||||
segmented_sentences[i] = removeStopWords(sentences[i]);
|
segmented_sentences[i] = removeStopWords(sentences[i]);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> bow;
|
std::vector<std::vector<real_t>> bow;
|
||||||
|
|
||||||
bow.resize(sentences.size());
|
bow.resize(sentences.size());
|
||||||
for (int i = 0; i < bow.size(); i++) {
|
for (int i = 0; i < bow.size(); i++) {
|
||||||
@ -752,7 +752,7 @@ std::vector<std::vector<double>> MLPPData::BOW(std::vector<std::string> sentence
|
|||||||
return bow;
|
return bow;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPData::TFIDF(std::vector<std::string> sentences) {
|
std::vector<std::vector<real_t>> MLPPData::TFIDF(std::vector<std::string> sentences) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::string> wordList = removeNullByte(removeStopWords(createWordList(sentences)));
|
std::vector<std::string> wordList = removeNullByte(removeStopWords(createWordList(sentences)));
|
||||||
|
|
||||||
@ -763,7 +763,7 @@ std::vector<std::vector<double>> MLPPData::TFIDF(std::vector<std::string> senten
|
|||||||
segmented_sentences[i] = removeStopWords(sentences[i]);
|
segmented_sentences[i] = removeStopWords(sentences[i]);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> TF;
|
std::vector<std::vector<real_t>> TF;
|
||||||
std::vector<int> frequency;
|
std::vector<int> frequency;
|
||||||
frequency.resize(wordList.size());
|
frequency.resize(wordList.size());
|
||||||
TF.resize(segmented_sentences.size());
|
TF.resize(segmented_sentences.size());
|
||||||
@ -783,17 +783,17 @@ std::vector<std::vector<double>> MLPPData::TFIDF(std::vector<std::string> senten
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
TF[i] = alg.scalarMultiply(double(1) / double(segmented_sentences[i].size()), TF[i]);
|
TF[i] = alg.scalarMultiply(real_t(1) / real_t(segmented_sentences[i].size()), TF[i]);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> IDF;
|
std::vector<real_t> IDF;
|
||||||
IDF.resize(frequency.size());
|
IDF.resize(frequency.size());
|
||||||
|
|
||||||
for (int i = 0; i < IDF.size(); i++) {
|
for (int i = 0; i < IDF.size(); i++) {
|
||||||
IDF[i] = std::log((double)segmented_sentences.size() / (double)frequency[i]);
|
IDF[i] = std::log((real_t)segmented_sentences.size() / (real_t)frequency[i]);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> TFIDF;
|
std::vector<std::vector<real_t>> TFIDF;
|
||||||
TFIDF.resize(segmented_sentences.size());
|
TFIDF.resize(segmented_sentences.size());
|
||||||
for (int i = 0; i < TFIDF.size(); i++) {
|
for (int i = 0; i < TFIDF.size(); i++) {
|
||||||
TFIDF[i].resize(wordList.size());
|
TFIDF[i].resize(wordList.size());
|
||||||
@ -808,7 +808,7 @@ std::vector<std::vector<double>> MLPPData::TFIDF(std::vector<std::string> senten
|
|||||||
return TFIDF;
|
return TFIDF;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::string>> MLPPData::word2Vec(std::vector<std::string> sentences, std::string type, int windowSize, int dimension, double learning_rate, int max_epoch) {
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::string>> MLPPData::word2Vec(std::vector<std::string> sentences, std::string type, int windowSize, int dimension, real_t learning_rate, int max_epoch) {
|
||||||
std::vector<std::string> wordList = removeNullByte(removeStopWords(createWordList(sentences)));
|
std::vector<std::string> wordList = removeNullByte(removeStopWords(createWordList(sentences)));
|
||||||
|
|
||||||
std::vector<std::vector<std::string>> segmented_sentences;
|
std::vector<std::vector<std::string>> segmented_sentences;
|
||||||
@ -841,10 +841,10 @@ std::tuple<std::vector<std::vector<double>>, std::vector<std::string>> MLPPData:
|
|||||||
|
|
||||||
inputStrings.insert(inputStrings.end(), outputStrings.begin(), outputStrings.end());
|
inputStrings.insert(inputStrings.end(), outputStrings.begin(), outputStrings.end());
|
||||||
|
|
||||||
std::vector<std::vector<double>> BOW = MLPPData::BOW(inputStrings, "Binary");
|
std::vector<std::vector<real_t>> BOW = MLPPData::BOW(inputStrings, "Binary");
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<std::vector<double>> outputSet;
|
std::vector<std::vector<real_t>> outputSet;
|
||||||
|
|
||||||
for (int i = 0; i < inputSize; i++) {
|
for (int i = 0; i < inputSize; i++) {
|
||||||
inputSet.push_back(BOW[i]);
|
inputSet.push_back(BOW[i]);
|
||||||
@ -862,17 +862,17 @@ std::tuple<std::vector<std::vector<double>>, std::vector<std::string>> MLPPData:
|
|||||||
}
|
}
|
||||||
model->gradientDescent(learning_rate, max_epoch, 1);
|
model->gradientDescent(learning_rate, max_epoch, 1);
|
||||||
|
|
||||||
std::vector<std::vector<double>> wordEmbeddings = model->getEmbeddings();
|
std::vector<std::vector<real_t>> wordEmbeddings = model->getEmbeddings();
|
||||||
delete model;
|
delete model;
|
||||||
return { wordEmbeddings, wordList };
|
return { wordEmbeddings, wordList };
|
||||||
}
|
}
|
||||||
|
|
||||||
struct WordsToVecResult {
|
struct WordsToVecResult {
|
||||||
std::vector<std::vector<double>> word_embeddings;
|
std::vector<std::vector<real_t>> word_embeddings;
|
||||||
std::vector<std::string> word_list;
|
std::vector<std::string> word_list;
|
||||||
};
|
};
|
||||||
|
|
||||||
MLPPData::WordsToVecResult MLPPData::word_to_vec(std::vector<std::string> sentences, std::string type, int windowSize, int dimension, double learning_rate, int max_epoch) {
|
MLPPData::WordsToVecResult MLPPData::word_to_vec(std::vector<std::string> sentences, std::string type, int windowSize, int dimension, real_t learning_rate, int max_epoch) {
|
||||||
WordsToVecResult res;
|
WordsToVecResult res;
|
||||||
|
|
||||||
res.word_list = removeNullByte(removeStopWords(createWordList(sentences)));
|
res.word_list = removeNullByte(removeStopWords(createWordList(sentences)));
|
||||||
@ -907,10 +907,10 @@ MLPPData::WordsToVecResult MLPPData::word_to_vec(std::vector<std::string> senten
|
|||||||
|
|
||||||
inputStrings.insert(inputStrings.end(), outputStrings.begin(), outputStrings.end());
|
inputStrings.insert(inputStrings.end(), outputStrings.begin(), outputStrings.end());
|
||||||
|
|
||||||
std::vector<std::vector<double>> BOW = MLPPData::BOW(inputStrings, "Binary");
|
std::vector<std::vector<real_t>> BOW = MLPPData::BOW(inputStrings, "Binary");
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<std::vector<double>> outputSet;
|
std::vector<std::vector<real_t>> outputSet;
|
||||||
|
|
||||||
for (int i = 0; i < inputSize; i++) {
|
for (int i = 0; i < inputSize; i++) {
|
||||||
inputSet.push_back(BOW[i]);
|
inputSet.push_back(BOW[i]);
|
||||||
@ -934,19 +934,19 @@ MLPPData::WordsToVecResult MLPPData::word_to_vec(std::vector<std::string> senten
|
|||||||
return res;
|
return res;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPData::LSA(std::vector<std::string> sentences, int dim) {
|
std::vector<std::vector<real_t>> MLPPData::LSA(std::vector<std::string> sentences, int dim) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<double>> docWordData = BOW(sentences, "Binary");
|
std::vector<std::vector<real_t>> docWordData = BOW(sentences, "Binary");
|
||||||
|
|
||||||
auto [U, S, Vt] = alg.SVD(docWordData);
|
auto [U, S, Vt] = alg.SVD(docWordData);
|
||||||
std::vector<std::vector<double>> S_trunc = alg.zeromat(dim, dim);
|
std::vector<std::vector<real_t>> S_trunc = alg.zeromat(dim, dim);
|
||||||
std::vector<std::vector<double>> Vt_trunc;
|
std::vector<std::vector<real_t>> Vt_trunc;
|
||||||
for (int i = 0; i < dim; i++) {
|
for (int i = 0; i < dim; i++) {
|
||||||
S_trunc[i][i] = S[i][i];
|
S_trunc[i][i] = S[i][i];
|
||||||
Vt_trunc.push_back(Vt[i]);
|
Vt_trunc.push_back(Vt[i]);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> embeddings = alg.matmult(S_trunc, Vt_trunc);
|
std::vector<std::vector<real_t>> embeddings = alg.matmult(S_trunc, Vt_trunc);
|
||||||
return embeddings;
|
return embeddings;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -977,10 +977,10 @@ void MLPPData::setInputNames(std::string fileName, std::vector<std::string> &inp
|
|||||||
dataFile.close();
|
dataFile.close();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPData::featureScaling(std::vector<std::vector<double>> X) {
|
std::vector<std::vector<real_t>> MLPPData::featureScaling(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
X = alg.transpose(X);
|
X = alg.transpose(X);
|
||||||
std::vector<double> max_elements, min_elements;
|
std::vector<real_t> max_elements, min_elements;
|
||||||
max_elements.resize(X.size());
|
max_elements.resize(X.size());
|
||||||
min_elements.resize(X.size());
|
min_elements.resize(X.size());
|
||||||
|
|
||||||
@ -997,7 +997,7 @@ std::vector<std::vector<double>> MLPPData::featureScaling(std::vector<std::vecto
|
|||||||
return alg.transpose(X);
|
return alg.transpose(X);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPData::meanNormalization(std::vector<std::vector<double>> X) {
|
std::vector<std::vector<real_t>> MLPPData::meanNormalization(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPStat stat;
|
MLPPStat stat;
|
||||||
// (X_j - mu_j) / std_j, for every j
|
// (X_j - mu_j) / std_j, for every j
|
||||||
@ -1009,11 +1009,11 @@ std::vector<std::vector<double>> MLPPData::meanNormalization(std::vector<std::ve
|
|||||||
return X;
|
return X;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPData::meanCentering(std::vector<std::vector<double>> X) {
|
std::vector<std::vector<real_t>> MLPPData::meanCentering(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPStat stat;
|
MLPPStat stat;
|
||||||
for (int i = 0; i < X.size(); i++) {
|
for (int i = 0; i < X.size(); i++) {
|
||||||
double mean_i = stat.mean(X[i]);
|
real_t mean_i = stat.mean(X[i]);
|
||||||
for (int j = 0; j < X[i].size(); j++) {
|
for (int j = 0; j < X[i].size(); j++) {
|
||||||
X[i][j] -= mean_i;
|
X[i][j] -= mean_i;
|
||||||
}
|
}
|
||||||
@ -1021,8 +1021,8 @@ std::vector<std::vector<double>> MLPPData::meanCentering(std::vector<std::vector
|
|||||||
return X;
|
return X;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPData::oneHotRep(std::vector<double> tempOutputSet, int n_class) {
|
std::vector<std::vector<real_t>> MLPPData::oneHotRep(std::vector<real_t> tempOutputSet, int n_class) {
|
||||||
std::vector<std::vector<double>> outputSet;
|
std::vector<std::vector<real_t>> outputSet;
|
||||||
outputSet.resize(tempOutputSet.size());
|
outputSet.resize(tempOutputSet.size());
|
||||||
for (int i = 0; i < tempOutputSet.size(); i++) {
|
for (int i = 0; i < tempOutputSet.size(); i++) {
|
||||||
for (int j = 0; j <= n_class - 1; j++) {
|
for (int j = 0; j <= n_class - 1; j++) {
|
||||||
@ -1036,8 +1036,8 @@ std::vector<std::vector<double>> MLPPData::oneHotRep(std::vector<double> tempOut
|
|||||||
return outputSet;
|
return outputSet;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPData::reverseOneHot(std::vector<std::vector<double>> tempOutputSet) {
|
std::vector<real_t> MLPPData::reverseOneHot(std::vector<std::vector<real_t>> tempOutputSet) {
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
int n_class = tempOutputSet[0].size();
|
int n_class = tempOutputSet[0].size();
|
||||||
for (int i = 0; i < tempOutputSet.size(); i++) {
|
for (int i = 0; i < tempOutputSet.size(); i++) {
|
||||||
int current_class = 1;
|
int current_class = 1;
|
||||||
|
@ -9,6 +9,8 @@
|
|||||||
// Created by Marc Melikyan on 11/4/20.
|
// Created by Marc Melikyan on 11/4/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include "core/string/ustring.h"
|
#include "core/string/ustring.h"
|
||||||
#include "core/variant/array.h"
|
#include "core/variant/array.h"
|
||||||
|
|
||||||
@ -22,8 +24,8 @@ class MLPPDataESimple : public Reference {
|
|||||||
GDCLASS(MLPPDataESimple, Reference);
|
GDCLASS(MLPPDataESimple, Reference);
|
||||||
|
|
||||||
public:
|
public:
|
||||||
std::vector<double> input;
|
std::vector<real_t> input;
|
||||||
std::vector<double> output;
|
std::vector<real_t> output;
|
||||||
|
|
||||||
protected:
|
protected:
|
||||||
static void _bind_methods();
|
static void _bind_methods();
|
||||||
@ -33,8 +35,8 @@ class MLPPDataSimple : public Reference {
|
|||||||
GDCLASS(MLPPDataSimple, Reference);
|
GDCLASS(MLPPDataSimple, Reference);
|
||||||
|
|
||||||
public:
|
public:
|
||||||
std::vector<std::vector<double>> input;
|
std::vector<std::vector<real_t>> input;
|
||||||
std::vector<double> output;
|
std::vector<real_t> output;
|
||||||
|
|
||||||
protected:
|
protected:
|
||||||
static void _bind_methods();
|
static void _bind_methods();
|
||||||
@ -44,8 +46,8 @@ class MLPPDataComplex : public Reference {
|
|||||||
GDCLASS(MLPPDataComplex, Reference);
|
GDCLASS(MLPPDataComplex, Reference);
|
||||||
|
|
||||||
public:
|
public:
|
||||||
std::vector<std::vector<double>> input;
|
std::vector<std::vector<real_t>> input;
|
||||||
std::vector<std::vector<double>> output;
|
std::vector<std::vector<real_t>> output;
|
||||||
|
|
||||||
protected:
|
protected:
|
||||||
static void _bind_methods();
|
static void _bind_methods();
|
||||||
@ -65,48 +67,48 @@ public:
|
|||||||
Ref<MLPPDataSimple> load_california_housing(const String &path);
|
Ref<MLPPDataSimple> load_california_housing(const String &path);
|
||||||
Ref<MLPPDataESimple> load_fires_and_crime(const String &path);
|
Ref<MLPPDataESimple> load_fires_and_crime(const String &path);
|
||||||
|
|
||||||
void set_data_supervised(int k, const String &file_name, std::vector<std::vector<double>> &inputSet, std::vector<double> &outputSet);
|
void set_data_supervised(int k, const String &file_name, std::vector<std::vector<real_t>> &inputSet, std::vector<real_t> &outputSet);
|
||||||
void set_data_unsupervised(int k, const String &file_name, std::vector<std::vector<double>> &inputSet);
|
void set_data_unsupervised(int k, const String &file_name, std::vector<std::vector<real_t>> &inputSet);
|
||||||
void set_data_simple(const String &file_name, std::vector<double> &inputSet, std::vector<double> &outputSet);
|
void set_data_simple(const String &file_name, std::vector<real_t> &inputSet, std::vector<real_t> &outputSet);
|
||||||
|
|
||||||
struct SplitComplexData {
|
struct SplitComplexData {
|
||||||
Ref<MLPPDataComplex> train;
|
Ref<MLPPDataComplex> train;
|
||||||
Ref<MLPPDataComplex> test;
|
Ref<MLPPDataComplex> test;
|
||||||
};
|
};
|
||||||
|
|
||||||
SplitComplexData train_test_split(const Ref<MLPPDataComplex> &data, double test_size);
|
SplitComplexData train_test_split(const Ref<MLPPDataComplex> &data, real_t test_size);
|
||||||
Array train_test_split_bind(const Ref<MLPPDataComplex> &data, double test_size);
|
Array train_test_split_bind(const Ref<MLPPDataComplex> &data, real_t test_size);
|
||||||
|
|
||||||
// Load Datasets
|
// Load Datasets
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<double>> loadBreastCancer();
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<real_t>> loadBreastCancer();
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<double>> loadBreastCancerSVC();
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<real_t>> loadBreastCancerSVC();
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>> loadIris();
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> loadIris();
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>> loadWine();
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> loadWine();
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>> loadMnistTrain();
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> loadMnistTrain();
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>> loadMnistTest();
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> loadMnistTest();
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<double>> loadCaliforniaHousing();
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<real_t>> loadCaliforniaHousing();
|
||||||
std::tuple<std::vector<double>, std::vector<double>> loadFiresAndCrime();
|
std::tuple<std::vector<real_t>, std::vector<real_t>> loadFiresAndCrime();
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>, std::vector<std::vector<double>>, std::vector<std::vector<double>>> trainTestSplit(std::vector<std::vector<double>> inputSet, std::vector<std::vector<double>> outputSet, double testSize);
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> trainTestSplit(std::vector<std::vector<real_t>> inputSet, std::vector<std::vector<real_t>> outputSet, real_t testSize);
|
||||||
|
|
||||||
// Supervised
|
// Supervised
|
||||||
void setData(int k, std::string fileName, std::vector<std::vector<double>> &inputSet, std::vector<double> &outputSet);
|
void setData(int k, std::string fileName, std::vector<std::vector<real_t>> &inputSet, std::vector<real_t> &outputSet);
|
||||||
void printData(std::vector<std::string> inputName, std::string outputName, std::vector<std::vector<double>> inputSet, std::vector<double> outputSet);
|
void printData(std::vector<std::string> inputName, std::string outputName, std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet);
|
||||||
|
|
||||||
// Unsupervised
|
// Unsupervised
|
||||||
void setData(int k, std::string fileName, std::vector<std::vector<double>> &inputSet);
|
void setData(int k, std::string fileName, std::vector<std::vector<real_t>> &inputSet);
|
||||||
void printData(std::vector<std::string> inputName, std::vector<std::vector<double>> inputSet);
|
void printData(std::vector<std::string> inputName, std::vector<std::vector<real_t>> inputSet);
|
||||||
|
|
||||||
// Simple
|
// Simple
|
||||||
void setData(std::string fileName, std::vector<double> &inputSet, std::vector<double> &outputSet);
|
void setData(std::string fileName, std::vector<real_t> &inputSet, std::vector<real_t> &outputSet);
|
||||||
void printData(std::string &inputName, std::string &outputName, std::vector<double> &inputSet, std::vector<double> &outputSet);
|
void printData(std::string &inputName, std::string &outputName, std::vector<real_t> &inputSet, std::vector<real_t> &outputSet);
|
||||||
|
|
||||||
// Images
|
// Images
|
||||||
std::vector<std::vector<double>> rgb2gray(std::vector<std::vector<std::vector<double>>> input);
|
std::vector<std::vector<real_t>> rgb2gray(std::vector<std::vector<std::vector<real_t>>> input);
|
||||||
std::vector<std::vector<std::vector<double>>> rgb2ycbcr(std::vector<std::vector<std::vector<double>>> input);
|
std::vector<std::vector<std::vector<real_t>>> rgb2ycbcr(std::vector<std::vector<std::vector<real_t>>> input);
|
||||||
std::vector<std::vector<std::vector<double>>> rgb2hsv(std::vector<std::vector<std::vector<double>>> input);
|
std::vector<std::vector<std::vector<real_t>>> rgb2hsv(std::vector<std::vector<std::vector<real_t>>> input);
|
||||||
std::vector<std::vector<std::vector<double>>> rgb2xyz(std::vector<std::vector<std::vector<double>>> input);
|
std::vector<std::vector<std::vector<real_t>>> rgb2xyz(std::vector<std::vector<std::vector<real_t>>> input);
|
||||||
std::vector<std::vector<std::vector<double>>> xyz2rgb(std::vector<std::vector<std::vector<double>>> input);
|
std::vector<std::vector<std::vector<real_t>>> xyz2rgb(std::vector<std::vector<std::vector<real_t>>> input);
|
||||||
|
|
||||||
// Text-Based & NLP
|
// Text-Based & NLP
|
||||||
std::string toLower(std::string text);
|
std::string toLower(std::string text);
|
||||||
@ -115,35 +117,35 @@ public:
|
|||||||
std::vector<std::string> removeSpaces(std::vector<std::string> data);
|
std::vector<std::string> removeSpaces(std::vector<std::string> data);
|
||||||
std::vector<std::string> removeNullByte(std::vector<std::string> data);
|
std::vector<std::string> removeNullByte(std::vector<std::string> data);
|
||||||
std::vector<std::string> segment(std::string text);
|
std::vector<std::string> segment(std::string text);
|
||||||
std::vector<double> tokenize(std::string text);
|
std::vector<real_t> tokenize(std::string text);
|
||||||
std::vector<std::string> removeStopWords(std::string text);
|
std::vector<std::string> removeStopWords(std::string text);
|
||||||
std::vector<std::string> removeStopWords(std::vector<std::string> segmented_data);
|
std::vector<std::string> removeStopWords(std::vector<std::string> segmented_data);
|
||||||
|
|
||||||
std::string stemming(std::string text);
|
std::string stemming(std::string text);
|
||||||
|
|
||||||
std::vector<std::vector<double>> BOW(std::vector<std::string> sentences, std::string = "Default");
|
std::vector<std::vector<real_t>> BOW(std::vector<std::string> sentences, std::string = "Default");
|
||||||
std::vector<std::vector<double>> TFIDF(std::vector<std::string> sentences);
|
std::vector<std::vector<real_t>> TFIDF(std::vector<std::string> sentences);
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::string>> word2Vec(std::vector<std::string> sentences, std::string type, int windowSize, int dimension, double learning_rate, int max_epoch);
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::string>> word2Vec(std::vector<std::string> sentences, std::string type, int windowSize, int dimension, real_t learning_rate, int max_epoch);
|
||||||
|
|
||||||
struct WordsToVecResult {
|
struct WordsToVecResult {
|
||||||
std::vector<std::vector<double>> word_embeddings;
|
std::vector<std::vector<real_t>> word_embeddings;
|
||||||
std::vector<std::string> word_list;
|
std::vector<std::string> word_list;
|
||||||
};
|
};
|
||||||
|
|
||||||
WordsToVecResult word_to_vec(std::vector<std::string> sentences, std::string type, int windowSize, int dimension, double learning_rate, int max_epoch);
|
WordsToVecResult word_to_vec(std::vector<std::string> sentences, std::string type, int windowSize, int dimension, real_t learning_rate, int max_epoch);
|
||||||
|
|
||||||
std::vector<std::vector<double>> LSA(std::vector<std::string> sentences, int dim);
|
std::vector<std::vector<real_t>> LSA(std::vector<std::string> sentences, int dim);
|
||||||
|
|
||||||
std::vector<std::string> createWordList(std::vector<std::string> sentences);
|
std::vector<std::string> createWordList(std::vector<std::string> sentences);
|
||||||
|
|
||||||
// Extra
|
// Extra
|
||||||
void setInputNames(std::string fileName, std::vector<std::string> &inputNames);
|
void setInputNames(std::string fileName, std::vector<std::string> &inputNames);
|
||||||
std::vector<std::vector<double>> featureScaling(std::vector<std::vector<double>> X);
|
std::vector<std::vector<real_t>> featureScaling(std::vector<std::vector<real_t>> X);
|
||||||
std::vector<std::vector<double>> meanNormalization(std::vector<std::vector<double>> X);
|
std::vector<std::vector<real_t>> meanNormalization(std::vector<std::vector<real_t>> X);
|
||||||
std::vector<std::vector<double>> meanCentering(std::vector<std::vector<double>> X);
|
std::vector<std::vector<real_t>> meanCentering(std::vector<std::vector<real_t>> X);
|
||||||
std::vector<std::vector<double>> oneHotRep(std::vector<double> tempOutputSet, int n_class);
|
std::vector<std::vector<real_t>> oneHotRep(std::vector<real_t> tempOutputSet, int n_class);
|
||||||
std::vector<double> reverseOneHot(std::vector<std::vector<double>> tempOutputSet);
|
std::vector<real_t> reverseOneHot(std::vector<std::vector<real_t>> tempOutputSet);
|
||||||
|
|
||||||
template <class T>
|
template <class T>
|
||||||
std::vector<T> vecToSet(std::vector<T> inputSet) {
|
std::vector<T> vecToSet(std::vector<T> inputSet) {
|
||||||
|
@ -15,7 +15,7 @@
|
|||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
|
|
||||||
MLPPDualSVC::MLPPDualSVC(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, double C, std::string kernel) :
|
MLPPDualSVC::MLPPDualSVC(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, real_t C, std::string kernel) :
|
||||||
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), C(C), kernel(kernel) {
|
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), C(C), kernel(kernel) {
|
||||||
y_hat.resize(n);
|
y_hat.resize(n);
|
||||||
bias = MLPPUtilities::biasInitialization();
|
bias = MLPPUtilities::biasInitialization();
|
||||||
@ -23,20 +23,20 @@ MLPPDualSVC::MLPPDualSVC(std::vector<std::vector<double>> inputSet, std::vector<
|
|||||||
K = kernelFunction(inputSet, inputSet, kernel); // For now this is unused. When non-linear kernels are added, the K will be manipulated.
|
K = kernelFunction(inputSet, inputSet, kernel); // For now this is unused. When non-linear kernels are added, the K will be manipulated.
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPDualSVC::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPDualSVC::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
return Evaluate(X);
|
return Evaluate(X);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPDualSVC::modelTest(std::vector<double> x) {
|
real_t MLPPDualSVC::modelTest(std::vector<real_t> x) {
|
||||||
return Evaluate(x);
|
return Evaluate(x);
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPDualSVC::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
void MLPPDualSVC::gradientDescent(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
@ -48,9 +48,9 @@ void MLPPDualSVC::gradientDescent(double learning_rate, int max_epoch, bool UI)
|
|||||||
alphaProjection();
|
alphaProjection();
|
||||||
|
|
||||||
// Calculating the bias
|
// Calculating the bias
|
||||||
double biasGradient = 0;
|
real_t biasGradient = 0;
|
||||||
for (int i = 0; i < alpha.size(); i++) {
|
for (int i = 0; i < alpha.size(); i++) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
if (alpha[i] < C && alpha[i] > 0) {
|
if (alpha[i] < C && alpha[i] > 0) {
|
||||||
for (int j = 0; j < alpha.size(); j++) {
|
for (int j = 0; j < alpha.size(); j++) {
|
||||||
if (alpha[j] > 0) {
|
if (alpha[j] > 0) {
|
||||||
@ -79,13 +79,13 @@ void MLPPDualSVC::gradientDescent(double learning_rate, int max_epoch, bool UI)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// void MLPPDualSVC::SGD(double learning_rate, int max_epoch, bool UI){
|
// void MLPPDualSVC::SGD(real_t learning_rate, int max_epoch, bool UI){
|
||||||
// class MLPPCost cost;
|
// class MLPPCost cost;
|
||||||
// MLPPActivation avn;
|
// MLPPActivation avn;
|
||||||
// MLPPLinAlg alg;
|
// MLPPLinAlg alg;
|
||||||
// MLPPReg regularization;
|
// MLPPReg regularization;
|
||||||
|
|
||||||
// double cost_prev = 0;
|
// real_t cost_prev = 0;
|
||||||
// int epoch = 1;
|
// int epoch = 1;
|
||||||
|
|
||||||
// while(true){
|
// while(true){
|
||||||
@ -112,12 +112,12 @@ void MLPPDualSVC::gradientDescent(double learning_rate, int max_epoch, bool UI)
|
|||||||
// forwardPass();
|
// forwardPass();
|
||||||
// }
|
// }
|
||||||
|
|
||||||
// void MLPPDualSVC::MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI){
|
// void MLPPDualSVC::MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI){
|
||||||
// class MLPPCost cost;
|
// class MLPPCost cost;
|
||||||
// MLPPActivation avn;
|
// MLPPActivation avn;
|
||||||
// MLPPLinAlg alg;
|
// MLPPLinAlg alg;
|
||||||
// MLPPReg regularization;
|
// MLPPReg regularization;
|
||||||
// double cost_prev = 0;
|
// real_t cost_prev = 0;
|
||||||
// int epoch = 1;
|
// int epoch = 1;
|
||||||
|
|
||||||
// // Creating the mini-batches
|
// // Creating the mini-batches
|
||||||
@ -126,8 +126,8 @@ void MLPPDualSVC::gradientDescent(double learning_rate, int max_epoch, bool UI)
|
|||||||
|
|
||||||
// while(true){
|
// while(true){
|
||||||
// for(int i = 0; i < n_mini_batch; i++){
|
// for(int i = 0; i < n_mini_batch; i++){
|
||||||
// std::vector<double> y_hat = Evaluate(inputMiniBatches[i]);
|
// std::vector<real_t> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
// std::vector<double> z = propagate(inputMiniBatches[i]);
|
// std::vector<real_t> z = propagate(inputMiniBatches[i]);
|
||||||
// cost_prev = Cost(z, outputMiniBatches[i], weights, C);
|
// cost_prev = Cost(z, outputMiniBatches[i], weights, C);
|
||||||
|
|
||||||
// // Calculating the weight gradients
|
// // Calculating the weight gradients
|
||||||
@ -152,7 +152,7 @@ void MLPPDualSVC::gradientDescent(double learning_rate, int max_epoch, bool UI)
|
|||||||
// forwardPass();
|
// forwardPass();
|
||||||
// }
|
// }
|
||||||
|
|
||||||
double MLPPDualSVC::score() {
|
real_t MLPPDualSVC::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
return util.performance(y_hat, outputSet);
|
return util.performance(y_hat, outputSet);
|
||||||
}
|
}
|
||||||
@ -162,21 +162,21 @@ void MLPPDualSVC::save(std::string fileName) {
|
|||||||
util.saveParameters(fileName, alpha, bias);
|
util.saveParameters(fileName, alpha, bias);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPDualSVC::Cost(std::vector<double> alpha, std::vector<std::vector<double>> X, std::vector<double> y) {
|
real_t MLPPDualSVC::Cost(std::vector<real_t> alpha, std::vector<std::vector<real_t>> X, std::vector<real_t> y) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
return cost.dualFormSVM(alpha, X, y);
|
return cost.dualFormSVM(alpha, X, y);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPDualSVC::Evaluate(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPDualSVC::Evaluate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
return avn.sign(propagate(X));
|
return avn.sign(propagate(X));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPDualSVC::propagate(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPDualSVC::propagate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<double> z;
|
std::vector<real_t> z;
|
||||||
for (int i = 0; i < X.size(); i++) {
|
for (int i = 0; i < X.size(); i++) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int j = 0; j < alpha.size(); j++) {
|
for (int j = 0; j < alpha.size(); j++) {
|
||||||
if (alpha[j] != 0) {
|
if (alpha[j] != 0) {
|
||||||
sum += alpha[j] * outputSet[j] * alg.dot(inputSet[j], X[i]); // TO DO: DON'T forget to add non-linear kernelizations.
|
sum += alpha[j] * outputSet[j] * alg.dot(inputSet[j], X[i]); // TO DO: DON'T forget to add non-linear kernelizations.
|
||||||
@ -188,14 +188,14 @@ std::vector<double> MLPPDualSVC::propagate(std::vector<std::vector<double>> X) {
|
|||||||
return z;
|
return z;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPDualSVC::Evaluate(std::vector<double> x) {
|
real_t MLPPDualSVC::Evaluate(std::vector<real_t> x) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
return avn.sign(propagate(x));
|
return avn.sign(propagate(x));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPDualSVC::propagate(std::vector<double> x) {
|
real_t MLPPDualSVC::propagate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
double z = 0;
|
real_t z = 0;
|
||||||
for (int j = 0; j < alpha.size(); j++) {
|
for (int j = 0; j < alpha.size(); j++) {
|
||||||
if (alpha[j] != 0) {
|
if (alpha[j] != 0) {
|
||||||
z += alpha[j] * outputSet[j] * alg.dot(inputSet[j], x); // TO DO: DON'T forget to add non-linear kernelizations.
|
z += alpha[j] * outputSet[j] * alg.dot(inputSet[j], x); // TO DO: DON'T forget to add non-linear kernelizations.
|
||||||
@ -223,14 +223,14 @@ void MLPPDualSVC::alphaProjection() {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPDualSVC::kernelFunction(std::vector<double> u, std::vector<double> v, std::string kernel) {
|
real_t MLPPDualSVC::kernelFunction(std::vector<real_t> u, std::vector<real_t> v, std::string kernel) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (kernel == "Linear") {
|
if (kernel == "Linear") {
|
||||||
return alg.dot(u, v);
|
return alg.dot(u, v);
|
||||||
} // warning: non-void function does not return a value in all control paths [-Wreturn-type]
|
} // warning: non-void function does not return a value in all control paths [-Wreturn-type]
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPDualSVC::kernelFunction(std::vector<std::vector<double>> A, std::vector<std::vector<double>> B, std::string kernel) {
|
std::vector<std::vector<real_t>> MLPPDualSVC::kernelFunction(std::vector<std::vector<real_t>> A, std::vector<std::vector<real_t>> B, std::string kernel) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (kernel == "Linear") {
|
if (kernel == "Linear") {
|
||||||
return alg.matmult(inputSet, alg.transpose(inputSet));
|
return alg.matmult(inputSet, alg.transpose(inputSet));
|
||||||
|
@ -11,6 +11,8 @@
|
|||||||
// http://ciml.info/dl/v0_99/ciml-v0_99-ch11.pdf
|
// http://ciml.info/dl/v0_99/ciml-v0_99-ch11.pdf
|
||||||
// Were excellent for the practical intution behind the dual formulation.
|
// Were excellent for the practical intution behind the dual formulation.
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
@ -18,52 +20,52 @@
|
|||||||
|
|
||||||
class MLPPDualSVC {
|
class MLPPDualSVC {
|
||||||
public:
|
public:
|
||||||
MLPPDualSVC(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, double C, std::string kernel = "Linear");
|
MLPPDualSVC(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, real_t C, std::string kernel = "Linear");
|
||||||
MLPPDualSVC(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, double C, std::string kernel, double p, double c);
|
MLPPDualSVC(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, real_t C, std::string kernel, real_t p, real_t c);
|
||||||
|
|
||||||
std::vector<double> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<real_t> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
double modelTest(std::vector<double> x);
|
real_t modelTest(std::vector<real_t> x);
|
||||||
void gradientDescent(double learning_rate, int max_epoch, bool UI = 1);
|
void gradientDescent(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void SGD(double learning_rate, int max_epoch, bool UI = 1);
|
void SGD(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
void MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
||||||
double score();
|
real_t score();
|
||||||
void save(std::string fileName);
|
void save(std::string fileName);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
void init();
|
void init();
|
||||||
|
|
||||||
double Cost(std::vector<double> alpha, std::vector<std::vector<double>> X, std::vector<double> y);
|
real_t Cost(std::vector<real_t> alpha, std::vector<std::vector<real_t>> X, std::vector<real_t> y);
|
||||||
|
|
||||||
std::vector<double> Evaluate(std::vector<std::vector<double>> X);
|
std::vector<real_t> Evaluate(std::vector<std::vector<real_t>> X);
|
||||||
std::vector<double> propagate(std::vector<std::vector<double>> X);
|
std::vector<real_t> propagate(std::vector<std::vector<real_t>> X);
|
||||||
double Evaluate(std::vector<double> x);
|
real_t Evaluate(std::vector<real_t> x);
|
||||||
double propagate(std::vector<double> x);
|
real_t propagate(std::vector<real_t> x);
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
|
|
||||||
void alphaProjection();
|
void alphaProjection();
|
||||||
|
|
||||||
double kernelFunction(std::vector<double> v, std::vector<double> u, std::string kernel);
|
real_t kernelFunction(std::vector<real_t> v, std::vector<real_t> u, std::string kernel);
|
||||||
std::vector<std::vector<double>> kernelFunction(std::vector<std::vector<double>> U, std::vector<std::vector<double>> V, std::string kernel);
|
std::vector<std::vector<real_t>> kernelFunction(std::vector<std::vector<real_t>> U, std::vector<std::vector<real_t>> V, std::string kernel);
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
std::vector<double> z;
|
std::vector<real_t> z;
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
double bias;
|
real_t bias;
|
||||||
|
|
||||||
std::vector<double> alpha;
|
std::vector<real_t> alpha;
|
||||||
std::vector<std::vector<double>> K;
|
std::vector<std::vector<real_t>> K;
|
||||||
|
|
||||||
double C;
|
real_t C;
|
||||||
int n;
|
int n;
|
||||||
int k;
|
int k;
|
||||||
|
|
||||||
std::string kernel;
|
std::string kernel;
|
||||||
double p; // Poly
|
real_t p; // Poly
|
||||||
double c; // Poly
|
real_t c; // Poly
|
||||||
|
|
||||||
// UI Portion
|
// UI Portion
|
||||||
void UI(int epoch, double cost_prev);
|
void UI(int epoch, real_t cost_prev);
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -15,7 +15,7 @@
|
|||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
|
|
||||||
MLPPExpReg::MLPPExpReg(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, std::string reg, double lambda, double alpha) :
|
MLPPExpReg::MLPPExpReg(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, std::string reg, real_t lambda, real_t alpha) :
|
||||||
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
||||||
y_hat.resize(n);
|
y_hat.resize(n);
|
||||||
weights = MLPPUtilities::weightInitialization(k);
|
weights = MLPPUtilities::weightInitialization(k);
|
||||||
@ -23,41 +23,41 @@ MLPPExpReg::MLPPExpReg(std::vector<std::vector<double>> inputSet, std::vector<do
|
|||||||
bias = MLPPUtilities::biasInitialization();
|
bias = MLPPUtilities::biasInitialization();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPExpReg::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPExpReg::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
return Evaluate(X);
|
return Evaluate(X);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPExpReg::modelTest(std::vector<double> x) {
|
real_t MLPPExpReg::modelTest(std::vector<real_t> x) {
|
||||||
return Evaluate(x);
|
return Evaluate(x);
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPExpReg::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
void MLPPExpReg::gradientDescent(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
cost_prev = Cost(y_hat, outputSet);
|
cost_prev = Cost(y_hat, outputSet);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputSet);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputSet);
|
||||||
|
|
||||||
for (int i = 0; i < k; i++) {
|
for (int i = 0; i < k; i++) {
|
||||||
// Calculating the weight gradient
|
// Calculating the weight gradient
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int j = 0; j < n; j++) {
|
for (int j = 0; j < n; j++) {
|
||||||
sum += error[j] * inputSet[j][i] * std::pow(weights[i], inputSet[j][i] - 1);
|
sum += error[j] * inputSet[j][i] * std::pow(weights[i], inputSet[j][i] - 1);
|
||||||
}
|
}
|
||||||
double w_gradient = sum / n;
|
real_t w_gradient = sum / n;
|
||||||
|
|
||||||
// Calculating the initial gradient
|
// Calculating the initial gradient
|
||||||
double sum2 = 0;
|
real_t sum2 = 0;
|
||||||
for (int j = 0; j < n; j++) {
|
for (int j = 0; j < n; j++) {
|
||||||
sum2 += error[j] * std::pow(weights[i], inputSet[j][i]);
|
sum2 += error[j] * std::pow(weights[i], inputSet[j][i]);
|
||||||
}
|
}
|
||||||
|
|
||||||
double i_gradient = sum2 / n;
|
real_t i_gradient = sum2 / n;
|
||||||
|
|
||||||
// Weight/initial updation
|
// Weight/initial updation
|
||||||
weights[i] -= learning_rate * w_gradient;
|
weights[i] -= learning_rate * w_gradient;
|
||||||
@ -66,11 +66,11 @@ void MLPPExpReg::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
weights = regularization.regWeights(weights, lambda, alpha, reg);
|
weights = regularization.regWeights(weights, lambda, alpha, reg);
|
||||||
|
|
||||||
// Calculating the bias gradient
|
// Calculating the bias gradient
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int j = 0; j < n; j++) {
|
for (int j = 0; j < n; j++) {
|
||||||
sum += (y_hat[j] - outputSet[j]);
|
sum += (y_hat[j] - outputSet[j]);
|
||||||
}
|
}
|
||||||
double b_gradient = sum / n;
|
real_t b_gradient = sum / n;
|
||||||
|
|
||||||
// bias updation
|
// bias updation
|
||||||
bias -= learning_rate * b_gradient;
|
bias -= learning_rate * b_gradient;
|
||||||
@ -88,9 +88,9 @@ void MLPPExpReg::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPExpReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
void MLPPExpReg::SGD(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
@ -99,14 +99,14 @@ void MLPPExpReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
||||||
int outputIndex = distribution(generator);
|
int outputIndex = distribution(generator);
|
||||||
|
|
||||||
double y_hat = Evaluate(inputSet[outputIndex]);
|
real_t y_hat = Evaluate(inputSet[outputIndex]);
|
||||||
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
||||||
|
|
||||||
for (int i = 0; i < k; i++) {
|
for (int i = 0; i < k; i++) {
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
|
|
||||||
double w_gradient = (y_hat - outputSet[outputIndex]) * inputSet[outputIndex][i] * std::pow(weights[i], inputSet[outputIndex][i] - 1);
|
real_t w_gradient = (y_hat - outputSet[outputIndex]) * inputSet[outputIndex][i] * std::pow(weights[i], inputSet[outputIndex][i] - 1);
|
||||||
double i_gradient = (y_hat - outputSet[outputIndex]) * std::pow(weights[i], inputSet[outputIndex][i]);
|
real_t i_gradient = (y_hat - outputSet[outputIndex]) * std::pow(weights[i], inputSet[outputIndex][i]);
|
||||||
|
|
||||||
// Weight/initial updation
|
// Weight/initial updation
|
||||||
weights[i] -= learning_rate * w_gradient;
|
weights[i] -= learning_rate * w_gradient;
|
||||||
@ -115,7 +115,7 @@ void MLPPExpReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
weights = regularization.regWeights(weights, lambda, alpha, reg);
|
weights = regularization.regWeights(weights, lambda, alpha, reg);
|
||||||
|
|
||||||
// Calculating the bias gradients
|
// Calculating the bias gradients
|
||||||
double b_gradient = (y_hat - outputSet[outputIndex]);
|
real_t b_gradient = (y_hat - outputSet[outputIndex]);
|
||||||
|
|
||||||
// Bias updation
|
// Bias updation
|
||||||
bias -= learning_rate * b_gradient;
|
bias -= learning_rate * b_gradient;
|
||||||
@ -134,10 +134,10 @@ void MLPPExpReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPExpReg::MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
void MLPPExpReg::MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
@ -146,25 +146,25 @@ void MLPPExpReg::MBGD(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
for (int j = 0; j < k; j++) {
|
for (int j = 0; j < k; j++) {
|
||||||
// Calculating the weight gradient
|
// Calculating the weight gradient
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int k = 0; k < outputMiniBatches[i].size(); k++) {
|
for (int k = 0; k < outputMiniBatches[i].size(); k++) {
|
||||||
sum += error[k] * inputMiniBatches[i][k][j] * std::pow(weights[j], inputMiniBatches[i][k][j] - 1);
|
sum += error[k] * inputMiniBatches[i][k][j] * std::pow(weights[j], inputMiniBatches[i][k][j] - 1);
|
||||||
}
|
}
|
||||||
double w_gradient = sum / outputMiniBatches[i].size();
|
real_t w_gradient = sum / outputMiniBatches[i].size();
|
||||||
|
|
||||||
// Calculating the initial gradient
|
// Calculating the initial gradient
|
||||||
double sum2 = 0;
|
real_t sum2 = 0;
|
||||||
for (int k = 0; k < outputMiniBatches[i].size(); k++) {
|
for (int k = 0; k < outputMiniBatches[i].size(); k++) {
|
||||||
sum2 += error[k] * std::pow(weights[j], inputMiniBatches[i][k][j]);
|
sum2 += error[k] * std::pow(weights[j], inputMiniBatches[i][k][j]);
|
||||||
}
|
}
|
||||||
|
|
||||||
double i_gradient = sum2 / outputMiniBatches[i].size();
|
real_t i_gradient = sum2 / outputMiniBatches[i].size();
|
||||||
|
|
||||||
// Weight/initial updation
|
// Weight/initial updation
|
||||||
weights[j] -= learning_rate * w_gradient;
|
weights[j] -= learning_rate * w_gradient;
|
||||||
@ -173,11 +173,11 @@ void MLPPExpReg::MBGD(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
weights = regularization.regWeights(weights, lambda, alpha, reg);
|
weights = regularization.regWeights(weights, lambda, alpha, reg);
|
||||||
|
|
||||||
// Calculating the bias gradient
|
// Calculating the bias gradient
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int j = 0; j < outputMiniBatches[i].size(); j++) {
|
for (int j = 0; j < outputMiniBatches[i].size(); j++) {
|
||||||
sum += (y_hat[j] - outputMiniBatches[i][j]);
|
sum += (y_hat[j] - outputMiniBatches[i][j]);
|
||||||
}
|
}
|
||||||
double b_gradient = sum / outputMiniBatches[i].size();
|
real_t b_gradient = sum / outputMiniBatches[i].size();
|
||||||
y_hat = Evaluate(inputMiniBatches[i]);
|
y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
|
|
||||||
if (UI) {
|
if (UI) {
|
||||||
@ -193,7 +193,7 @@ void MLPPExpReg::MBGD(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPExpReg::score() {
|
real_t MLPPExpReg::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
return util.performance(y_hat, outputSet);
|
return util.performance(y_hat, outputSet);
|
||||||
}
|
}
|
||||||
@ -203,14 +203,14 @@ void MLPPExpReg::save(std::string fileName) {
|
|||||||
util.saveParameters(fileName, weights, initial, bias);
|
util.saveParameters(fileName, weights, initial, bias);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPExpReg::Cost(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPExpReg::Cost(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
return cost.MSE(y_hat, y) + regularization.regTerm(weights, lambda, alpha, reg);
|
return cost.MSE(y_hat, y) + regularization.regTerm(weights, lambda, alpha, reg);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPExpReg::Evaluate(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPExpReg::Evaluate(std::vector<std::vector<real_t>> X) {
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
y_hat.resize(X.size());
|
y_hat.resize(X.size());
|
||||||
for (int i = 0; i < X.size(); i++) {
|
for (int i = 0; i < X.size(); i++) {
|
||||||
y_hat[i] = 0;
|
y_hat[i] = 0;
|
||||||
@ -222,8 +222,8 @@ std::vector<double> MLPPExpReg::Evaluate(std::vector<std::vector<double>> X) {
|
|||||||
return y_hat;
|
return y_hat;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPExpReg::Evaluate(std::vector<double> x) {
|
real_t MLPPExpReg::Evaluate(std::vector<real_t> x) {
|
||||||
double y_hat = 0;
|
real_t y_hat = 0;
|
||||||
for (int i = 0; i < x.size(); i++) {
|
for (int i = 0; i < x.size(); i++) {
|
||||||
y_hat += initial[i] * std::pow(weights[i], x[i]);
|
y_hat += initial[i] * std::pow(weights[i], x[i]);
|
||||||
}
|
}
|
||||||
|
@ -8,42 +8,44 @@
|
|||||||
// Created by Marc Melikyan on 10/2/20.
|
// Created by Marc Melikyan on 10/2/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
|
|
||||||
class MLPPExpReg {
|
class MLPPExpReg {
|
||||||
public:
|
public:
|
||||||
MLPPExpReg(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, std::string reg = "None", double lambda = 0.5, double alpha = 0.5);
|
MLPPExpReg(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
||||||
std::vector<double> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<real_t> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
double modelTest(std::vector<double> x);
|
real_t modelTest(std::vector<real_t> x);
|
||||||
void gradientDescent(double learning_rate, int max_epoch, bool UI = 1);
|
void gradientDescent(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void SGD(double learning_rate, int max_epoch, bool UI = 1);
|
void SGD(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
void MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
||||||
double score();
|
real_t score();
|
||||||
void save(std::string fileName);
|
void save(std::string fileName);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
double Cost(std::vector<double> y_hat, std::vector<double> y);
|
real_t Cost(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
|
|
||||||
std::vector<double> Evaluate(std::vector<std::vector<double>> X);
|
std::vector<real_t> Evaluate(std::vector<std::vector<real_t>> X);
|
||||||
double Evaluate(std::vector<double> x);
|
real_t Evaluate(std::vector<real_t> x);
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
std::vector<double> weights;
|
std::vector<real_t> weights;
|
||||||
std::vector<double> initial;
|
std::vector<real_t> initial;
|
||||||
double bias;
|
real_t bias;
|
||||||
|
|
||||||
int n;
|
int n;
|
||||||
int k;
|
int k;
|
||||||
|
|
||||||
// Regularization Params
|
// Regularization Params
|
||||||
std::string reg;
|
std::string reg;
|
||||||
double lambda;
|
real_t lambda;
|
||||||
double alpha; /* This is the controlling param for Elastic Net*/
|
real_t alpha; /* This is the controlling param for Elastic Net*/
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -15,7 +15,7 @@
|
|||||||
#include <iostream>
|
#include <iostream>
|
||||||
|
|
||||||
|
|
||||||
MLPPGAN::MLPPGAN(double k, std::vector<std::vector<double>> outputSet) :
|
MLPPGAN::MLPPGAN(real_t k, std::vector<std::vector<real_t>> outputSet) :
|
||||||
outputSet(outputSet), n(outputSet.size()), k(k) {
|
outputSet(outputSet), n(outputSet.size()), k(k) {
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -23,15 +23,15 @@ MLPPGAN::~MLPPGAN() {
|
|||||||
delete outputLayer;
|
delete outputLayer;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPGAN::generateExample(int n) {
|
std::vector<std::vector<real_t>> MLPPGAN::generateExample(int n) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return modelSetTestGenerator(alg.gaussianNoise(n, k));
|
return modelSetTestGenerator(alg.gaussianNoise(n, k));
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPGAN::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
void MLPPGAN::gradientDescent(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
@ -40,13 +40,13 @@ void MLPPGAN::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
|
|
||||||
// Training of the discriminator.
|
// Training of the discriminator.
|
||||||
|
|
||||||
std::vector<std::vector<double>> generatorInputSet = alg.gaussianNoise(n, k);
|
std::vector<std::vector<real_t>> generatorInputSet = alg.gaussianNoise(n, k);
|
||||||
std::vector<std::vector<double>> discriminatorInputSet = modelSetTestGenerator(generatorInputSet);
|
std::vector<std::vector<real_t>> discriminatorInputSet = modelSetTestGenerator(generatorInputSet);
|
||||||
discriminatorInputSet.insert(discriminatorInputSet.end(), outputSet.begin(), outputSet.end()); // Fake + real inputs.
|
discriminatorInputSet.insert(discriminatorInputSet.end(), outputSet.begin(), outputSet.end()); // Fake + real inputs.
|
||||||
|
|
||||||
std::vector<double> y_hat = modelSetTestDiscriminator(discriminatorInputSet);
|
std::vector<real_t> y_hat = modelSetTestDiscriminator(discriminatorInputSet);
|
||||||
std::vector<double> outputSet = alg.zerovec(n);
|
std::vector<real_t> outputSet = alg.zerovec(n);
|
||||||
std::vector<double> outputSetReal = alg.onevec(n);
|
std::vector<real_t> outputSetReal = alg.onevec(n);
|
||||||
outputSet.insert(outputSet.end(), outputSetReal.begin(), outputSetReal.end()); // Fake + real output scores.
|
outputSet.insert(outputSet.end(), outputSetReal.begin(), outputSetReal.end()); // Fake + real output scores.
|
||||||
|
|
||||||
auto [cumulativeDiscriminatorHiddenLayerWGrad, outputDiscriminatorWGrad] = computeDiscriminatorGradients(y_hat, outputSet);
|
auto [cumulativeDiscriminatorHiddenLayerWGrad, outputDiscriminatorWGrad] = computeDiscriminatorGradients(y_hat, outputSet);
|
||||||
@ -60,7 +60,7 @@ void MLPPGAN::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
y_hat = modelSetTestDiscriminator(discriminatorInputSet);
|
y_hat = modelSetTestDiscriminator(discriminatorInputSet);
|
||||||
outputSet = alg.onevec(n);
|
outputSet = alg.onevec(n);
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> cumulativeGeneratorHiddenLayerWGrad = computeGeneratorGradients(y_hat, outputSet);
|
std::vector<std::vector<std::vector<real_t>>> cumulativeGeneratorHiddenLayerWGrad = computeGeneratorGradients(y_hat, outputSet);
|
||||||
cumulativeGeneratorHiddenLayerWGrad = alg.scalarMultiply(learning_rate / n, cumulativeGeneratorHiddenLayerWGrad);
|
cumulativeGeneratorHiddenLayerWGrad = alg.scalarMultiply(learning_rate / n, cumulativeGeneratorHiddenLayerWGrad);
|
||||||
updateGeneratorParameters(cumulativeGeneratorHiddenLayerWGrad, learning_rate);
|
updateGeneratorParameters(cumulativeGeneratorHiddenLayerWGrad, learning_rate);
|
||||||
|
|
||||||
@ -76,7 +76,7 @@ void MLPPGAN::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPGAN::score() {
|
real_t MLPPGAN::score() {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
@ -96,7 +96,7 @@ void MLPPGAN::save(std::string fileName) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPGAN::addLayer(int n_hidden, std::string activation, std::string weightInit, std::string reg, double lambda, double alpha) {
|
void MLPPGAN::addLayer(int n_hidden, std::string activation, std::string weightInit, std::string reg, real_t lambda, real_t alpha) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (network.empty()) {
|
if (network.empty()) {
|
||||||
network.push_back(MLPPHiddenLayer(n_hidden, activation, alg.gaussianNoise(n, k), weightInit, reg, lambda, alpha));
|
network.push_back(MLPPHiddenLayer(n_hidden, activation, alg.gaussianNoise(n, k), weightInit, reg, lambda, alpha));
|
||||||
@ -107,7 +107,7 @@ void MLPPGAN::addLayer(int n_hidden, std::string activation, std::string weightI
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPGAN::addOutputLayer(std::string weightInit, std::string reg, double lambda, double alpha) {
|
void MLPPGAN::addOutputLayer(std::string weightInit, std::string reg, real_t lambda, real_t alpha) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
outputLayer = new MLPPOutputLayer(network[network.size() - 1].n_hidden, "Sigmoid", "LogLoss", network[network.size() - 1].a, weightInit, reg, lambda, alpha);
|
outputLayer = new MLPPOutputLayer(network[network.size() - 1].n_hidden, "Sigmoid", "LogLoss", network[network.size() - 1].a, weightInit, reg, lambda, alpha);
|
||||||
@ -116,7 +116,7 @@ void MLPPGAN::addOutputLayer(std::string weightInit, std::string reg, double lam
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPGAN::modelSetTestGenerator(std::vector<std::vector<double>> X) {
|
std::vector<std::vector<real_t>> MLPPGAN::modelSetTestGenerator(std::vector<std::vector<real_t>> X) {
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
network[0].input = X;
|
network[0].input = X;
|
||||||
network[0].forwardPass();
|
network[0].forwardPass();
|
||||||
@ -129,7 +129,7 @@ std::vector<std::vector<double>> MLPPGAN::modelSetTestGenerator(std::vector<std:
|
|||||||
return network[network.size() / 2].a;
|
return network[network.size() / 2].a;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPGAN::modelSetTestDiscriminator(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPGAN::modelSetTestDiscriminator(std::vector<std::vector<real_t>> X) {
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
for (int i = network.size() / 2 + 1; i < network.size(); i++) {
|
for (int i = network.size() / 2 + 1; i < network.size(); i++) {
|
||||||
if (i == network.size() / 2 + 1) {
|
if (i == network.size() / 2 + 1) {
|
||||||
@ -145,10 +145,10 @@ std::vector<double> MLPPGAN::modelSetTestDiscriminator(std::vector<std::vector<d
|
|||||||
return outputLayer->a;
|
return outputLayer->a;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPGAN::Cost(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPGAN::Cost(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
double totalRegTerm = 0;
|
real_t totalRegTerm = 0;
|
||||||
|
|
||||||
auto cost_function = outputLayer->cost_map[outputLayer->cost];
|
auto cost_function = outputLayer->cost_map[outputLayer->cost];
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
@ -177,7 +177,7 @@ void MLPPGAN::forwardPass() {
|
|||||||
y_hat = outputLayer->a;
|
y_hat = outputLayer->a;
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPGAN::updateDiscriminatorParameters(std::vector<std::vector<std::vector<double>>> hiddenLayerUpdations, std::vector<double> outputLayerUpdation, double learning_rate) {
|
void MLPPGAN::updateDiscriminatorParameters(std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations, std::vector<real_t> outputLayerUpdation, real_t learning_rate) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
outputLayer->weights = alg.subtraction(outputLayer->weights, outputLayerUpdation);
|
outputLayer->weights = alg.subtraction(outputLayer->weights, outputLayerUpdation);
|
||||||
@ -194,7 +194,7 @@ void MLPPGAN::updateDiscriminatorParameters(std::vector<std::vector<std::vector<
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPGAN::updateGeneratorParameters(std::vector<std::vector<std::vector<double>>> hiddenLayerUpdations, double learning_rate) {
|
void MLPPGAN::updateGeneratorParameters(std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations, real_t learning_rate) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
@ -207,25 +207,25 @@ void MLPPGAN::updateGeneratorParameters(std::vector<std::vector<std::vector<doub
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<double>> MLPPGAN::computeDiscriminatorGradients(std::vector<double> y_hat, std::vector<double> outputSet) {
|
std::tuple<std::vector<std::vector<std::vector<real_t>>>, std::vector<real_t>> MLPPGAN::computeDiscriminatorGradients(std::vector<real_t> y_hat, std::vector<real_t> outputSet) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> cumulativeHiddenLayerWGrad; // Tensor containing ALL hidden grads.
|
std::vector<std::vector<std::vector<real_t>>> cumulativeHiddenLayerWGrad; // Tensor containing ALL hidden grads.
|
||||||
|
|
||||||
auto costDeriv = outputLayer->costDeriv_map[outputLayer->cost];
|
auto costDeriv = outputLayer->costDeriv_map[outputLayer->cost];
|
||||||
auto outputAvn = outputLayer->activation_map[outputLayer->activation];
|
auto outputAvn = outputLayer->activation_map[outputLayer->activation];
|
||||||
outputLayer->delta = alg.hadamard_product((cost.*costDeriv)(y_hat, outputSet), (avn.*outputAvn)(outputLayer->z, 1));
|
outputLayer->delta = alg.hadamard_product((cost.*costDeriv)(y_hat, outputSet), (avn.*outputAvn)(outputLayer->z, 1));
|
||||||
std::vector<double> outputWGrad = alg.mat_vec_mult(alg.transpose(outputLayer->input), outputLayer->delta);
|
std::vector<real_t> outputWGrad = alg.mat_vec_mult(alg.transpose(outputLayer->input), outputLayer->delta);
|
||||||
outputWGrad = alg.addition(outputWGrad, regularization.regDerivTerm(outputLayer->weights, outputLayer->lambda, outputLayer->alpha, outputLayer->reg));
|
outputWGrad = alg.addition(outputWGrad, regularization.regDerivTerm(outputLayer->weights, outputLayer->lambda, outputLayer->alpha, outputLayer->reg));
|
||||||
|
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
auto hiddenLayerAvn = network[network.size() - 1].activation_map[network[network.size() - 1].activation];
|
auto hiddenLayerAvn = network[network.size() - 1].activation_map[network[network.size() - 1].activation];
|
||||||
|
|
||||||
network[network.size() - 1].delta = alg.hadamard_product(alg.outerProduct(outputLayer->delta, outputLayer->weights), (avn.*hiddenLayerAvn)(network[network.size() - 1].z, 1));
|
network[network.size() - 1].delta = alg.hadamard_product(alg.outerProduct(outputLayer->delta, outputLayer->weights), (avn.*hiddenLayerAvn)(network[network.size() - 1].z, 1));
|
||||||
std::vector<std::vector<double>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[network.size() - 1].input), network[network.size() - 1].delta);
|
std::vector<std::vector<real_t>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[network.size() - 1].input), network[network.size() - 1].delta);
|
||||||
|
|
||||||
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[network.size() - 1].weights, network[network.size() - 1].lambda, network[network.size() - 1].alpha, network[network.size() - 1].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[network.size() - 1].weights, network[network.size() - 1].lambda, network[network.size() - 1].alpha, network[network.size() - 1].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
||||||
|
|
||||||
@ -235,7 +235,7 @@ std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<double>> M
|
|||||||
for (int i = network.size() - 2; i > network.size() / 2; i--) {
|
for (int i = network.size() - 2; i > network.size() / 2; i--) {
|
||||||
auto hiddenLayerAvn = network[i].activation_map[network[i].activation];
|
auto hiddenLayerAvn = network[i].activation_map[network[i].activation];
|
||||||
network[i].delta = alg.hadamard_product(alg.matmult(network[i + 1].delta, alg.transpose(network[i + 1].weights)), (avn.*hiddenLayerAvn)(network[i].z, 1));
|
network[i].delta = alg.hadamard_product(alg.matmult(network[i + 1].delta, alg.transpose(network[i + 1].weights)), (avn.*hiddenLayerAvn)(network[i].z, 1));
|
||||||
std::vector<std::vector<double>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[i].input), network[i].delta);
|
std::vector<std::vector<real_t>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[i].input), network[i].delta);
|
||||||
|
|
||||||
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[i].weights, network[i].lambda, network[i].alpha, network[i].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[i].weights, network[i].lambda, network[i].alpha, network[i].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
||||||
}
|
}
|
||||||
@ -243,36 +243,36 @@ std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<double>> M
|
|||||||
return { cumulativeHiddenLayerWGrad, outputWGrad };
|
return { cumulativeHiddenLayerWGrad, outputWGrad };
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> MLPPGAN::computeGeneratorGradients(std::vector<double> y_hat, std::vector<double> outputSet) {
|
std::vector<std::vector<std::vector<real_t>>> MLPPGAN::computeGeneratorGradients(std::vector<real_t> y_hat, std::vector<real_t> outputSet) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> cumulativeHiddenLayerWGrad; // Tensor containing ALL hidden grads.
|
std::vector<std::vector<std::vector<real_t>>> cumulativeHiddenLayerWGrad; // Tensor containing ALL hidden grads.
|
||||||
|
|
||||||
auto costDeriv = outputLayer->costDeriv_map[outputLayer->cost];
|
auto costDeriv = outputLayer->costDeriv_map[outputLayer->cost];
|
||||||
auto outputAvn = outputLayer->activation_map[outputLayer->activation];
|
auto outputAvn = outputLayer->activation_map[outputLayer->activation];
|
||||||
outputLayer->delta = alg.hadamard_product((cost.*costDeriv)(y_hat, outputSet), (avn.*outputAvn)(outputLayer->z, 1));
|
outputLayer->delta = alg.hadamard_product((cost.*costDeriv)(y_hat, outputSet), (avn.*outputAvn)(outputLayer->z, 1));
|
||||||
std::vector<double> outputWGrad = alg.mat_vec_mult(alg.transpose(outputLayer->input), outputLayer->delta);
|
std::vector<real_t> outputWGrad = alg.mat_vec_mult(alg.transpose(outputLayer->input), outputLayer->delta);
|
||||||
outputWGrad = alg.addition(outputWGrad, regularization.regDerivTerm(outputLayer->weights, outputLayer->lambda, outputLayer->alpha, outputLayer->reg));
|
outputWGrad = alg.addition(outputWGrad, regularization.regDerivTerm(outputLayer->weights, outputLayer->lambda, outputLayer->alpha, outputLayer->reg));
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
auto hiddenLayerAvn = network[network.size() - 1].activation_map[network[network.size() - 1].activation];
|
auto hiddenLayerAvn = network[network.size() - 1].activation_map[network[network.size() - 1].activation];
|
||||||
network[network.size() - 1].delta = alg.hadamard_product(alg.outerProduct(outputLayer->delta, outputLayer->weights), (avn.*hiddenLayerAvn)(network[network.size() - 1].z, 1));
|
network[network.size() - 1].delta = alg.hadamard_product(alg.outerProduct(outputLayer->delta, outputLayer->weights), (avn.*hiddenLayerAvn)(network[network.size() - 1].z, 1));
|
||||||
std::vector<std::vector<double>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[network.size() - 1].input), network[network.size() - 1].delta);
|
std::vector<std::vector<real_t>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[network.size() - 1].input), network[network.size() - 1].delta);
|
||||||
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[network.size() - 1].weights, network[network.size() - 1].lambda, network[network.size() - 1].alpha, network[network.size() - 1].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[network.size() - 1].weights, network[network.size() - 1].lambda, network[network.size() - 1].alpha, network[network.size() - 1].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
||||||
|
|
||||||
for (int i = network.size() - 2; i >= 0; i--) {
|
for (int i = network.size() - 2; i >= 0; i--) {
|
||||||
auto hiddenLayerAvn = network[i].activation_map[network[i].activation];
|
auto hiddenLayerAvn = network[i].activation_map[network[i].activation];
|
||||||
network[i].delta = alg.hadamard_product(alg.matmult(network[i + 1].delta, alg.transpose(network[i + 1].weights)), (avn.*hiddenLayerAvn)(network[i].z, 1));
|
network[i].delta = alg.hadamard_product(alg.matmult(network[i + 1].delta, alg.transpose(network[i + 1].weights)), (avn.*hiddenLayerAvn)(network[i].z, 1));
|
||||||
std::vector<std::vector<double>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[i].input), network[i].delta);
|
std::vector<std::vector<real_t>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[i].input), network[i].delta);
|
||||||
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[i].weights, network[i].lambda, network[i].alpha, network[i].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[i].weights, network[i].lambda, network[i].alpha, network[i].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return cumulativeHiddenLayerWGrad;
|
return cumulativeHiddenLayerWGrad;
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPGAN::UI(int epoch, double cost_prev, std::vector<double> y_hat, std::vector<double> outputSet) {
|
void MLPPGAN::UI(int epoch, real_t cost_prev, std::vector<real_t> y_hat, std::vector<real_t> outputSet) {
|
||||||
MLPPUtilities::CostInfo(epoch, cost_prev, Cost(y_hat, outputSet));
|
MLPPUtilities::CostInfo(epoch, cost_prev, Cost(y_hat, outputSet));
|
||||||
std::cout << "Layer " << network.size() + 1 << ": " << std::endl;
|
std::cout << "Layer " << network.size() + 1 << ": " << std::endl;
|
||||||
MLPPUtilities::UI(outputLayer->weights, outputLayer->bias);
|
MLPPUtilities::UI(outputLayer->weights, outputLayer->bias);
|
||||||
|
@ -8,6 +8,8 @@
|
|||||||
// Created by Marc Melikyan on 11/4/20.
|
// Created by Marc Melikyan on 11/4/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include "../hidden_layer/hidden_layer.h"
|
#include "../hidden_layer/hidden_layer.h"
|
||||||
#include "../output_layer/output_layer.h"
|
#include "../output_layer/output_layer.h"
|
||||||
|
|
||||||
@ -19,32 +21,32 @@
|
|||||||
|
|
||||||
class MLPPGAN {
|
class MLPPGAN {
|
||||||
public:
|
public:
|
||||||
MLPPGAN(double k, std::vector<std::vector<double>> outputSet);
|
MLPPGAN(real_t k, std::vector<std::vector<real_t>> outputSet);
|
||||||
~MLPPGAN();
|
~MLPPGAN();
|
||||||
std::vector<std::vector<double>> generateExample(int n);
|
std::vector<std::vector<real_t>> generateExample(int n);
|
||||||
void gradientDescent(double learning_rate, int max_epoch, bool UI = 1);
|
void gradientDescent(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
double score();
|
real_t score();
|
||||||
void save(std::string fileName);
|
void save(std::string fileName);
|
||||||
|
|
||||||
void addLayer(int n_hidden, std::string activation, std::string weightInit = "Default", std::string reg = "None", double lambda = 0.5, double alpha = 0.5);
|
void addLayer(int n_hidden, std::string activation, std::string weightInit = "Default", std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
||||||
void addOutputLayer(std::string weightInit = "Default", std::string reg = "None", double lambda = 0.5, double alpha = 0.5);
|
void addOutputLayer(std::string weightInit = "Default", std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
std::vector<std::vector<double>> modelSetTestGenerator(std::vector<std::vector<double>> X); // Evaluator for the generator of the gan.
|
std::vector<std::vector<real_t>> modelSetTestGenerator(std::vector<std::vector<real_t>> X); // Evaluator for the generator of the gan.
|
||||||
std::vector<double> modelSetTestDiscriminator(std::vector<std::vector<double>> X); // Evaluator for the discriminator of the gan.
|
std::vector<real_t> modelSetTestDiscriminator(std::vector<std::vector<real_t>> X); // Evaluator for the discriminator of the gan.
|
||||||
|
|
||||||
double Cost(std::vector<double> y_hat, std::vector<double> y);
|
real_t Cost(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
|
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
void updateDiscriminatorParameters(std::vector<std::vector<std::vector<double>>> hiddenLayerUpdations, std::vector<double> outputLayerUpdation, double learning_rate);
|
void updateDiscriminatorParameters(std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations, std::vector<real_t> outputLayerUpdation, real_t learning_rate);
|
||||||
void updateGeneratorParameters(std::vector<std::vector<std::vector<double>>> hiddenLayerUpdations, double learning_rate);
|
void updateGeneratorParameters(std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations, real_t learning_rate);
|
||||||
std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<double>> computeDiscriminatorGradients(std::vector<double> y_hat, std::vector<double> outputSet);
|
std::tuple<std::vector<std::vector<std::vector<real_t>>>, std::vector<real_t>> computeDiscriminatorGradients(std::vector<real_t> y_hat, std::vector<real_t> outputSet);
|
||||||
std::vector<std::vector<std::vector<double>>> computeGeneratorGradients(std::vector<double> y_hat, std::vector<double> outputSet);
|
std::vector<std::vector<std::vector<real_t>>> computeGeneratorGradients(std::vector<real_t> y_hat, std::vector<real_t> outputSet);
|
||||||
|
|
||||||
void UI(int epoch, double cost_prev, std::vector<double> y_hat, std::vector<double> outputSet);
|
void UI(int epoch, real_t cost_prev, std::vector<real_t> y_hat, std::vector<real_t> outputSet);
|
||||||
|
|
||||||
std::vector<std::vector<double>> outputSet;
|
std::vector<std::vector<real_t>> outputSet;
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
|
|
||||||
std::vector<MLPPHiddenLayer> network;
|
std::vector<MLPPHiddenLayer> network;
|
||||||
MLPPOutputLayer *outputLayer;
|
MLPPOutputLayer *outputLayer;
|
||||||
|
@ -9,7 +9,7 @@
|
|||||||
#include <iostream>
|
#include <iostream>
|
||||||
|
|
||||||
|
|
||||||
void MLPPGaussMarkovChecker::checkGMConditions(std::vector<double> eps) {
|
void MLPPGaussMarkovChecker::checkGMConditions(std::vector<real_t> eps) {
|
||||||
bool condition1 = arithmeticMean(eps);
|
bool condition1 = arithmeticMean(eps);
|
||||||
bool condition2 = homoscedasticity(eps);
|
bool condition2 = homoscedasticity(eps);
|
||||||
bool condition3 = exogeneity(eps);
|
bool condition3 = exogeneity(eps);
|
||||||
@ -21,7 +21,7 @@ void MLPPGaussMarkovChecker::checkGMConditions(std::vector<double> eps) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
bool MLPPGaussMarkovChecker::arithmeticMean(std::vector<double> eps) {
|
bool MLPPGaussMarkovChecker::arithmeticMean(std::vector<real_t> eps) {
|
||||||
MLPPStat stat;
|
MLPPStat stat;
|
||||||
if (stat.mean(eps) == 0) {
|
if (stat.mean(eps) == 0) {
|
||||||
return 1;
|
return 1;
|
||||||
@ -30,9 +30,9 @@ bool MLPPGaussMarkovChecker::arithmeticMean(std::vector<double> eps) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
bool MLPPGaussMarkovChecker::homoscedasticity(std::vector<double> eps) {
|
bool MLPPGaussMarkovChecker::homoscedasticity(std::vector<real_t> eps) {
|
||||||
MLPPStat stat;
|
MLPPStat stat;
|
||||||
double currentVar = (eps[0] - stat.mean(eps)) * (eps[0] - stat.mean(eps)) / eps.size();
|
real_t currentVar = (eps[0] - stat.mean(eps)) * (eps[0] - stat.mean(eps)) / eps.size();
|
||||||
for (int i = 0; i < eps.size(); i++) {
|
for (int i = 0; i < eps.size(); i++) {
|
||||||
if (currentVar != (eps[i] - stat.mean(eps)) * (eps[i] - stat.mean(eps)) / eps.size()) {
|
if (currentVar != (eps[i] - stat.mean(eps)) * (eps[i] - stat.mean(eps)) / eps.size()) {
|
||||||
return 0;
|
return 0;
|
||||||
@ -41,7 +41,7 @@ bool MLPPGaussMarkovChecker::homoscedasticity(std::vector<double> eps) {
|
|||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
bool MLPPGaussMarkovChecker::exogeneity(std::vector<double> eps) {
|
bool MLPPGaussMarkovChecker::exogeneity(std::vector<real_t> eps) {
|
||||||
MLPPStat stat;
|
MLPPStat stat;
|
||||||
for (int i = 0; i < eps.size(); i++) {
|
for (int i = 0; i < eps.size(); i++) {
|
||||||
for (int j = 0; j < eps.size(); j++) {
|
for (int j = 0; j < eps.size(); j++) {
|
||||||
|
@ -8,18 +8,20 @@
|
|||||||
// Created by Marc Melikyan on 11/13/20.
|
// Created by Marc Melikyan on 11/13/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
|
|
||||||
class MLPPGaussMarkovChecker {
|
class MLPPGaussMarkovChecker {
|
||||||
public:
|
public:
|
||||||
void checkGMConditions(std::vector<double> eps);
|
void checkGMConditions(std::vector<real_t> eps);
|
||||||
|
|
||||||
// Independent, 3 Gauss-Markov Conditions
|
// Independent, 3 Gauss-Markov Conditions
|
||||||
bool arithmeticMean(std::vector<double> eps); // 1) Arithmetic Mean of 0.
|
bool arithmeticMean(std::vector<real_t> eps); // 1) Arithmetic Mean of 0.
|
||||||
bool homoscedasticity(std::vector<double> eps); // 2) Homoscedasticity
|
bool homoscedasticity(std::vector<real_t> eps); // 2) Homoscedasticity
|
||||||
bool exogeneity(std::vector<double> eps); // 3) Cov of any 2 non-equal eps values = 0.
|
bool exogeneity(std::vector<real_t> eps); // 3) Cov of any 2 non-equal eps values = 0.
|
||||||
private:
|
private:
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -14,35 +14,35 @@
|
|||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
|
|
||||||
MLPPGaussianNB::MLPPGaussianNB(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, int class_num) :
|
MLPPGaussianNB::MLPPGaussianNB(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, int class_num) :
|
||||||
inputSet(inputSet), outputSet(outputSet), class_num(class_num) {
|
inputSet(inputSet), outputSet(outputSet), class_num(class_num) {
|
||||||
y_hat.resize(outputSet.size());
|
y_hat.resize(outputSet.size());
|
||||||
Evaluate();
|
Evaluate();
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPGaussianNB::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPGaussianNB::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
for (int i = 0; i < X.size(); i++) {
|
for (int i = 0; i < X.size(); i++) {
|
||||||
y_hat.push_back(modelTest(X[i]));
|
y_hat.push_back(modelTest(X[i]));
|
||||||
}
|
}
|
||||||
return y_hat;
|
return y_hat;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPGaussianNB::modelTest(std::vector<double> x) {
|
real_t MLPPGaussianNB::modelTest(std::vector<real_t> x) {
|
||||||
MLPPStat stat;
|
MLPPStat stat;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
double score[class_num];
|
real_t score[class_num];
|
||||||
double y_hat_i = 1;
|
real_t y_hat_i = 1;
|
||||||
for (int i = class_num - 1; i >= 0; i--) {
|
for (int i = class_num - 1; i >= 0; i--) {
|
||||||
y_hat_i += std::log(priors[i] * (1 / sqrt(2 * M_PI * sigma[i] * sigma[i])) * exp(-(x[i] * mu[i]) * (x[i] * mu[i]) / (2 * sigma[i] * sigma[i])));
|
y_hat_i += std::log(priors[i] * (1 / sqrt(2 * M_PI * sigma[i] * sigma[i])) * exp(-(x[i] * mu[i]) * (x[i] * mu[i]) / (2 * sigma[i] * sigma[i])));
|
||||||
score[i] = exp(y_hat_i);
|
score[i] = exp(y_hat_i);
|
||||||
}
|
}
|
||||||
return std::distance(score, std::max_element(score, score + sizeof(score) / sizeof(double)));
|
return std::distance(score, std::max_element(score, score + sizeof(score) / sizeof(real_t)));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPGaussianNB::score() {
|
real_t MLPPGaussianNB::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
return util.performance(y_hat, outputSet);
|
return util.performance(y_hat, outputSet);
|
||||||
}
|
}
|
||||||
@ -55,7 +55,7 @@ void MLPPGaussianNB::Evaluate() {
|
|||||||
mu.resize(class_num);
|
mu.resize(class_num);
|
||||||
sigma.resize(class_num);
|
sigma.resize(class_num);
|
||||||
for (int i = class_num - 1; i >= 0; i--) {
|
for (int i = class_num - 1; i >= 0; i--) {
|
||||||
std::vector<double> set;
|
std::vector<real_t> set;
|
||||||
for (int j = 0; j < inputSet.size(); j++) {
|
for (int j = 0; j < inputSet.size(); j++) {
|
||||||
for (int k = 0; k < inputSet[j].size(); k++) {
|
for (int k = 0; k < inputSet[j].size(); k++) {
|
||||||
if (outputSet[j] == i) {
|
if (outputSet[j] == i) {
|
||||||
@ -72,11 +72,11 @@ void MLPPGaussianNB::Evaluate() {
|
|||||||
for (int i = 0; i < outputSet.size(); i++) {
|
for (int i = 0; i < outputSet.size(); i++) {
|
||||||
priors[int(outputSet[i])]++;
|
priors[int(outputSet[i])]++;
|
||||||
}
|
}
|
||||||
priors = alg.scalarMultiply(double(1) / double(outputSet.size()), priors);
|
priors = alg.scalarMultiply(real_t(1) / real_t(outputSet.size()), priors);
|
||||||
|
|
||||||
for (int i = 0; i < outputSet.size(); i++) {
|
for (int i = 0; i < outputSet.size(); i++) {
|
||||||
double score[class_num];
|
real_t score[class_num];
|
||||||
double y_hat_i = 1;
|
real_t y_hat_i = 1;
|
||||||
for (int j = class_num - 1; j >= 0; j--) {
|
for (int j = class_num - 1; j >= 0; j--) {
|
||||||
for (int k = 0; k < inputSet[i].size(); k++) {
|
for (int k = 0; k < inputSet[i].size(); k++) {
|
||||||
y_hat_i += std::log(priors[j] * (1 / sqrt(2 * M_PI * sigma[j] * sigma[j])) * exp(-(inputSet[i][k] * mu[j]) * (inputSet[i][k] * mu[j]) / (2 * sigma[j] * sigma[j])));
|
y_hat_i += std::log(priors[j] * (1 / sqrt(2 * M_PI * sigma[j] * sigma[j])) * exp(-(inputSet[i][k] * mu[j]) * (inputSet[i][k] * mu[j]) / (2 * sigma[j] * sigma[j])));
|
||||||
@ -84,7 +84,7 @@ void MLPPGaussianNB::Evaluate() {
|
|||||||
score[j] = exp(y_hat_i);
|
score[j] = exp(y_hat_i);
|
||||||
std::cout << score[j] << std::endl;
|
std::cout << score[j] << std::endl;
|
||||||
}
|
}
|
||||||
y_hat[i] = std::distance(score, std::max_element(score, score + sizeof(score) / sizeof(double)));
|
y_hat[i] = std::distance(score, std::max_element(score, score + sizeof(score) / sizeof(real_t)));
|
||||||
std::cout << std::distance(score, std::max_element(score, score + sizeof(score) / sizeof(double))) << std::endl;
|
std::cout << std::distance(score, std::max_element(score, score + sizeof(score) / sizeof(real_t))) << std::endl;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -8,29 +8,31 @@
|
|||||||
// Created by Marc Melikyan on 1/17/21.
|
// Created by Marc Melikyan on 1/17/21.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
|
|
||||||
class MLPPGaussianNB {
|
class MLPPGaussianNB {
|
||||||
public:
|
public:
|
||||||
MLPPGaussianNB(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, int class_num);
|
MLPPGaussianNB(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, int class_num);
|
||||||
std::vector<double> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<real_t> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
double modelTest(std::vector<double> x);
|
real_t modelTest(std::vector<real_t> x);
|
||||||
double score();
|
real_t score();
|
||||||
|
|
||||||
private:
|
private:
|
||||||
void Evaluate();
|
void Evaluate();
|
||||||
|
|
||||||
int class_num;
|
int class_num;
|
||||||
|
|
||||||
std::vector<double> priors;
|
std::vector<real_t> priors;
|
||||||
std::vector<double> mu;
|
std::vector<real_t> mu;
|
||||||
std::vector<double> sigma;
|
std::vector<real_t> sigma;
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
|
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
};
|
};
|
||||||
|
|
||||||
#endif /* GaussianNB_hpp */
|
#endif /* GaussianNB_hpp */
|
||||||
|
@ -13,7 +13,7 @@
|
|||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
|
|
||||||
MLPPHiddenLayer::MLPPHiddenLayer(int n_hidden, std::string activation, std::vector<std::vector<double>> input, std::string weightInit, std::string reg, double lambda, double alpha) :
|
MLPPHiddenLayer::MLPPHiddenLayer(int n_hidden, std::string activation, std::vector<std::vector<real_t>> input, std::string weightInit, std::string reg, real_t lambda, real_t alpha) :
|
||||||
n_hidden(n_hidden), activation(activation), input(input), weightInit(weightInit), reg(reg), lambda(lambda), alpha(alpha) {
|
n_hidden(n_hidden), activation(activation), input(input), weightInit(weightInit), reg(reg), lambda(lambda), alpha(alpha) {
|
||||||
weights = MLPPUtilities::weightInitialization(input[0].size(), n_hidden, weightInit);
|
weights = MLPPUtilities::weightInitialization(input[0].size(), n_hidden, weightInit);
|
||||||
bias = MLPPUtilities::biasInitialization(n_hidden);
|
bias = MLPPUtilities::biasInitialization(n_hidden);
|
||||||
@ -104,7 +104,7 @@ void MLPPHiddenLayer::forwardPass() {
|
|||||||
a = (avn.*activation_map[activation])(z, 0);
|
a = (avn.*activation_map[activation])(z, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPHiddenLayer::Test(std::vector<double> x) {
|
void MLPPHiddenLayer::Test(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
z_test = alg.addition(alg.mat_vec_mult(alg.transpose(weights), x), bias);
|
z_test = alg.addition(alg.mat_vec_mult(alg.transpose(weights), x), bias);
|
||||||
|
@ -8,6 +8,8 @@
|
|||||||
// Created by Marc Melikyan on 11/4/20.
|
// Created by Marc Melikyan on 11/4/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include "../activation/activation.h"
|
#include "../activation/activation.h"
|
||||||
|
|
||||||
#include <map>
|
#include <map>
|
||||||
@ -17,36 +19,36 @@
|
|||||||
|
|
||||||
class MLPPHiddenLayer {
|
class MLPPHiddenLayer {
|
||||||
public:
|
public:
|
||||||
MLPPHiddenLayer(int n_hidden, std::string activation, std::vector<std::vector<double>> input, std::string weightInit, std::string reg, double lambda, double alpha);
|
MLPPHiddenLayer(int n_hidden, std::string activation, std::vector<std::vector<real_t>> input, std::string weightInit, std::string reg, real_t lambda, real_t alpha);
|
||||||
|
|
||||||
int n_hidden;
|
int n_hidden;
|
||||||
std::string activation;
|
std::string activation;
|
||||||
|
|
||||||
std::vector<std::vector<double>> input;
|
std::vector<std::vector<real_t>> input;
|
||||||
|
|
||||||
std::vector<std::vector<double>> weights;
|
std::vector<std::vector<real_t>> weights;
|
||||||
std::vector<double> bias;
|
std::vector<real_t> bias;
|
||||||
|
|
||||||
std::vector<std::vector<double>> z;
|
std::vector<std::vector<real_t>> z;
|
||||||
std::vector<std::vector<double>> a;
|
std::vector<std::vector<real_t>> a;
|
||||||
|
|
||||||
std::map<std::string, std::vector<std::vector<double>> (MLPPActivation::*)(std::vector<std::vector<double>>, bool)> activation_map;
|
std::map<std::string, std::vector<std::vector<real_t>> (MLPPActivation::*)(std::vector<std::vector<real_t>>, bool)> activation_map;
|
||||||
std::map<std::string, std::vector<double> (MLPPActivation::*)(std::vector<double>, bool)> activationTest_map;
|
std::map<std::string, std::vector<real_t> (MLPPActivation::*)(std::vector<real_t>, bool)> activationTest_map;
|
||||||
|
|
||||||
std::vector<double> z_test;
|
std::vector<real_t> z_test;
|
||||||
std::vector<double> a_test;
|
std::vector<real_t> a_test;
|
||||||
|
|
||||||
std::vector<std::vector<double>> delta;
|
std::vector<std::vector<real_t>> delta;
|
||||||
|
|
||||||
// Regularization Params
|
// Regularization Params
|
||||||
std::string reg;
|
std::string reg;
|
||||||
double lambda; /* Regularization Parameter */
|
real_t lambda; /* Regularization Parameter */
|
||||||
double alpha; /* This is the controlling param for Elastic Net*/
|
real_t alpha; /* This is the controlling param for Elastic Net*/
|
||||||
|
|
||||||
std::string weightInit;
|
std::string weightInit;
|
||||||
|
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
void Test(std::vector<double> x);
|
void Test(std::vector<real_t> x);
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -8,9 +8,9 @@
|
|||||||
|
|
||||||
|
|
||||||
|
|
||||||
std::tuple<bool, double> MLPPHypothesisTesting::chiSquareTest(std::vector<double> observed, std::vector<double> expected) {
|
std::tuple<bool, real_t> MLPPHypothesisTesting::chiSquareTest(std::vector<real_t> observed, std::vector<real_t> expected) {
|
||||||
double df = observed.size() - 1; // These are our degrees of freedom
|
real_t df = observed.size() - 1; // These are our degrees of freedom
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < observed.size(); i++) {
|
for (int i = 0; i < observed.size(); i++) {
|
||||||
sum += (observed[i] - expected[i]) * (observed[i] - expected[i]) / expected[i];
|
sum += (observed[i] - expected[i]) * (observed[i] - expected[i]) / expected[i];
|
||||||
}
|
}
|
||||||
|
@ -8,13 +8,15 @@
|
|||||||
// Created by Marc Melikyan on 3/10/21.
|
// Created by Marc Melikyan on 3/10/21.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <tuple>
|
#include <tuple>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
|
|
||||||
class MLPPHypothesisTesting {
|
class MLPPHypothesisTesting {
|
||||||
public:
|
public:
|
||||||
std::tuple<bool, double> chiSquareTest(std::vector<double> observed, std::vector<double> expected);
|
std::tuple<bool, real_t> chiSquareTest(std::vector<real_t> observed, std::vector<real_t> expected);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
};
|
};
|
||||||
|
@ -13,7 +13,7 @@
|
|||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
|
|
||||||
MLPPKMeans::MLPPKMeans(std::vector<std::vector<double>> inputSet, int k, std::string init_type) :
|
MLPPKMeans::MLPPKMeans(std::vector<std::vector<real_t>> inputSet, int k, std::string init_type) :
|
||||||
inputSet(inputSet), k(k), init_type(init_type) {
|
inputSet(inputSet), k(k), init_type(init_type) {
|
||||||
if (init_type == "KMeans++") {
|
if (init_type == "KMeans++") {
|
||||||
kmeansppInitialization(k);
|
kmeansppInitialization(k);
|
||||||
@ -22,11 +22,11 @@ MLPPKMeans::MLPPKMeans(std::vector<std::vector<double>> inputSet, int k, std::st
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPKMeans::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<std::vector<real_t>> MLPPKMeans::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<double>> closestCentroids;
|
std::vector<std::vector<real_t>> closestCentroids;
|
||||||
for (int i = 0; i < inputSet.size(); i++) {
|
for (int i = 0; i < inputSet.size(); i++) {
|
||||||
std::vector<double> closestCentroid = mu[0];
|
std::vector<real_t> closestCentroid = mu[0];
|
||||||
for (int j = 0; j < r[0].size(); j++) {
|
for (int j = 0; j < r[0].size(); j++) {
|
||||||
bool isCentroidCloser = alg.euclideanDistance(X[i], mu[j]) < alg.euclideanDistance(X[i], closestCentroid);
|
bool isCentroidCloser = alg.euclideanDistance(X[i], mu[j]) < alg.euclideanDistance(X[i], closestCentroid);
|
||||||
if (isCentroidCloser) {
|
if (isCentroidCloser) {
|
||||||
@ -38,9 +38,9 @@ std::vector<std::vector<double>> MLPPKMeans::modelSetTest(std::vector<std::vecto
|
|||||||
return closestCentroids;
|
return closestCentroids;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPKMeans::modelTest(std::vector<double> x) {
|
std::vector<real_t> MLPPKMeans::modelTest(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<double> closestCentroid = mu[0];
|
std::vector<real_t> closestCentroid = mu[0];
|
||||||
for (int j = 0; j < mu.size(); j++) {
|
for (int j = 0; j < mu.size(); j++) {
|
||||||
if (alg.euclideanDistance(x, mu[j]) < alg.euclideanDistance(x, closestCentroid)) {
|
if (alg.euclideanDistance(x, mu[j]) < alg.euclideanDistance(x, closestCentroid)) {
|
||||||
closestCentroid = mu[j];
|
closestCentroid = mu[j];
|
||||||
@ -50,7 +50,7 @@ std::vector<double> MLPPKMeans::modelTest(std::vector<double> x) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
void MLPPKMeans::train(int epoch_num, bool UI) {
|
void MLPPKMeans::train(int epoch_num, bool UI) {
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
Evaluate();
|
Evaluate();
|
||||||
@ -80,17 +80,17 @@ void MLPPKMeans::train(int epoch_num, bool UI) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPKMeans::score() {
|
real_t MLPPKMeans::score() {
|
||||||
return Cost();
|
return Cost();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPKMeans::silhouette_scores() {
|
std::vector<real_t> MLPPKMeans::silhouette_scores() {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<double>> closestCentroids = modelSetTest(inputSet);
|
std::vector<std::vector<real_t>> closestCentroids = modelSetTest(inputSet);
|
||||||
std::vector<double> silhouette_scores;
|
std::vector<real_t> silhouette_scores;
|
||||||
for (int i = 0; i < inputSet.size(); i++) {
|
for (int i = 0; i < inputSet.size(); i++) {
|
||||||
// COMPUTING a[i]
|
// COMPUTING a[i]
|
||||||
double a = 0;
|
real_t a = 0;
|
||||||
for (int j = 0; j < inputSet.size(); j++) {
|
for (int j = 0; j < inputSet.size(); j++) {
|
||||||
if (i != j && r[i] == r[j]) {
|
if (i != j && r[i] == r[j]) {
|
||||||
a += alg.euclideanDistance(inputSet[i], inputSet[j]);
|
a += alg.euclideanDistance(inputSet[i], inputSet[j]);
|
||||||
@ -100,15 +100,15 @@ std::vector<double> MLPPKMeans::silhouette_scores() {
|
|||||||
a /= closestCentroids[i].size() - 1;
|
a /= closestCentroids[i].size() - 1;
|
||||||
|
|
||||||
// COMPUTING b[i]
|
// COMPUTING b[i]
|
||||||
double b = INT_MAX;
|
real_t b = INT_MAX;
|
||||||
for (int j = 0; j < mu.size(); j++) {
|
for (int j = 0; j < mu.size(); j++) {
|
||||||
if (closestCentroids[i] != mu[j]) {
|
if (closestCentroids[i] != mu[j]) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int k = 0; k < inputSet.size(); k++) {
|
for (int k = 0; k < inputSet.size(); k++) {
|
||||||
sum += alg.euclideanDistance(inputSet[i], inputSet[k]);
|
sum += alg.euclideanDistance(inputSet[i], inputSet[k]);
|
||||||
}
|
}
|
||||||
// NORMALIZE b[i]
|
// NORMALIZE b[i]
|
||||||
double k_clusterSize = 0;
|
real_t k_clusterSize = 0;
|
||||||
for (int k = 0; k < closestCentroids.size(); k++) {
|
for (int k = 0; k < closestCentroids.size(); k++) {
|
||||||
if (closestCentroids[k] == mu[j]) {
|
if (closestCentroids[k] == mu[j]) {
|
||||||
k_clusterSize++;
|
k_clusterSize++;
|
||||||
@ -144,7 +144,7 @@ void MLPPKMeans::Evaluate() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
for (int i = 0; i < r.size(); i++) {
|
for (int i = 0; i < r.size(); i++) {
|
||||||
std::vector<double> closestCentroid = mu[0];
|
std::vector<real_t> closestCentroid = mu[0];
|
||||||
for (int j = 0; j < r[0].size(); j++) {
|
for (int j = 0; j < r[0].size(); j++) {
|
||||||
bool isCentroidCloser = alg.euclideanDistance(inputSet[i], mu[j]) < alg.euclideanDistance(inputSet[i], closestCentroid);
|
bool isCentroidCloser = alg.euclideanDistance(inputSet[i], mu[j]) < alg.euclideanDistance(inputSet[i], closestCentroid);
|
||||||
if (isCentroidCloser) {
|
if (isCentroidCloser) {
|
||||||
@ -165,21 +165,21 @@ void MLPPKMeans::Evaluate() {
|
|||||||
void MLPPKMeans::computeMu() {
|
void MLPPKMeans::computeMu() {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
for (int i = 0; i < mu.size(); i++) {
|
for (int i = 0; i < mu.size(); i++) {
|
||||||
std::vector<double> num;
|
std::vector<real_t> num;
|
||||||
num.resize(r.size());
|
num.resize(r.size());
|
||||||
|
|
||||||
for (int i = 0; i < num.size(); i++) {
|
for (int i = 0; i < num.size(); i++) {
|
||||||
num[i] = 0;
|
num[i] = 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
double den = 0;
|
real_t den = 0;
|
||||||
for (int j = 0; j < r.size(); j++) {
|
for (int j = 0; j < r.size(); j++) {
|
||||||
num = alg.addition(num, alg.scalarMultiply(r[j][i], inputSet[j]));
|
num = alg.addition(num, alg.scalarMultiply(r[j][i], inputSet[j]));
|
||||||
}
|
}
|
||||||
for (int j = 0; j < r.size(); j++) {
|
for (int j = 0; j < r.size(); j++) {
|
||||||
den += r[j][i];
|
den += r[j][i];
|
||||||
}
|
}
|
||||||
mu[i] = alg.scalarMultiply(double(1) / double(den), num);
|
mu[i] = alg.scalarMultiply(real_t(1) / real_t(den), num);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -204,12 +204,12 @@ void MLPPKMeans::kmeansppInitialization(int k) {
|
|||||||
mu.push_back(inputSet[distribution(generator)]);
|
mu.push_back(inputSet[distribution(generator)]);
|
||||||
|
|
||||||
for (int i = 0; i < k - 1; i++) {
|
for (int i = 0; i < k - 1; i++) {
|
||||||
std::vector<double> farthestCentroid;
|
std::vector<real_t> farthestCentroid;
|
||||||
for (int j = 0; j < inputSet.size(); j++) {
|
for (int j = 0; j < inputSet.size(); j++) {
|
||||||
double max_dist = 0;
|
real_t max_dist = 0;
|
||||||
/* SUM ALL THE SQUARED DISTANCES, CHOOSE THE ONE THAT'S FARTHEST
|
/* SUM ALL THE SQUARED DISTANCES, CHOOSE THE ONE THAT'S FARTHEST
|
||||||
AS TO SPREAD OUT THE CLUSTER CENTROIDS. */
|
AS TO SPREAD OUT THE CLUSTER CENTROIDS. */
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int k = 0; k < mu.size(); k++) {
|
for (int k = 0; k < mu.size(); k++) {
|
||||||
sum += alg.euclideanDistance(inputSet[j], mu[k]);
|
sum += alg.euclideanDistance(inputSet[j], mu[k]);
|
||||||
}
|
}
|
||||||
@ -222,9 +222,9 @@ void MLPPKMeans::kmeansppInitialization(int k) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPKMeans::Cost() {
|
real_t MLPPKMeans::Cost() {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < r.size(); i++) {
|
for (int i = 0; i < r.size(); i++) {
|
||||||
for (int j = 0; j < r[0].size(); j++) {
|
for (int j = 0; j < r[0].size(); j++) {
|
||||||
sum += r[i][j] * alg.norm_sq(alg.subtraction(inputSet[i], mu[j]));
|
sum += r[i][j] * alg.norm_sq(alg.subtraction(inputSet[i], mu[j]));
|
||||||
|
@ -8,18 +8,20 @@
|
|||||||
// Created by Marc Melikyan on 10/2/20.
|
// Created by Marc Melikyan on 10/2/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
|
|
||||||
class MLPPKMeans {
|
class MLPPKMeans {
|
||||||
public:
|
public:
|
||||||
MLPPKMeans(std::vector<std::vector<double>> inputSet, int k, std::string init_type = "Default");
|
MLPPKMeans(std::vector<std::vector<real_t>> inputSet, int k, std::string init_type = "Default");
|
||||||
std::vector<std::vector<double>> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<std::vector<real_t>> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
std::vector<double> modelTest(std::vector<double> x);
|
std::vector<real_t> modelTest(std::vector<real_t> x);
|
||||||
void train(int epoch_num, bool UI = 1);
|
void train(int epoch_num, bool UI = 1);
|
||||||
double score();
|
real_t score();
|
||||||
std::vector<double> silhouette_scores();
|
std::vector<real_t> silhouette_scores();
|
||||||
|
|
||||||
private:
|
private:
|
||||||
void Evaluate();
|
void Evaluate();
|
||||||
@ -27,15 +29,15 @@ private:
|
|||||||
|
|
||||||
void centroidInitialization(int k);
|
void centroidInitialization(int k);
|
||||||
void kmeansppInitialization(int k);
|
void kmeansppInitialization(int k);
|
||||||
double Cost();
|
real_t Cost();
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<std::vector<double>> mu;
|
std::vector<std::vector<real_t>> mu;
|
||||||
std::vector<std::vector<double>> r;
|
std::vector<std::vector<real_t>> r;
|
||||||
|
|
||||||
double euclideanDistance(std::vector<double> A, std::vector<double> B);
|
real_t euclideanDistance(std::vector<real_t> A, std::vector<real_t> B);
|
||||||
|
|
||||||
double accuracy_threshold;
|
real_t accuracy_threshold;
|
||||||
int k;
|
int k;
|
||||||
|
|
||||||
std::string init_type;
|
std::string init_type;
|
||||||
|
@ -13,28 +13,28 @@
|
|||||||
#include <map>
|
#include <map>
|
||||||
|
|
||||||
|
|
||||||
MLPPKNN::MLPPKNN(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, int k) :
|
MLPPKNN::MLPPKNN(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, int k) :
|
||||||
inputSet(inputSet), outputSet(outputSet), k(k) {
|
inputSet(inputSet), outputSet(outputSet), k(k) {
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPKNN::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPKNN::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
for (int i = 0; i < X.size(); i++) {
|
for (int i = 0; i < X.size(); i++) {
|
||||||
y_hat.push_back(modelTest(X[i]));
|
y_hat.push_back(modelTest(X[i]));
|
||||||
}
|
}
|
||||||
return y_hat;
|
return y_hat;
|
||||||
}
|
}
|
||||||
|
|
||||||
int MLPPKNN::modelTest(std::vector<double> x) {
|
int MLPPKNN::modelTest(std::vector<real_t> x) {
|
||||||
return determineClass(nearestNeighbors(x));
|
return determineClass(nearestNeighbors(x));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPKNN::score() {
|
real_t MLPPKNN::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
return util.performance(modelSetTest(inputSet), outputSet);
|
return util.performance(modelSetTest(inputSet), outputSet);
|
||||||
}
|
}
|
||||||
|
|
||||||
int MLPPKNN::determineClass(std::vector<double> knn) {
|
int MLPPKNN::determineClass(std::vector<real_t> knn) {
|
||||||
std::map<int, int> class_nums;
|
std::map<int, int> class_nums;
|
||||||
for (int i = 0; i < outputSet.size(); i++) {
|
for (int i = 0; i < outputSet.size(); i++) {
|
||||||
class_nums[outputSet[i]] = 0;
|
class_nums[outputSet[i]] = 0;
|
||||||
@ -62,12 +62,12 @@ int MLPPKNN::determineClass(std::vector<double> knn) {
|
|||||||
return final_class;
|
return final_class;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPKNN::nearestNeighbors(std::vector<double> x) {
|
std::vector<real_t> MLPPKNN::nearestNeighbors(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
// The nearest neighbors
|
// The nearest neighbors
|
||||||
std::vector<double> knn;
|
std::vector<real_t> knn;
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputUseSet = inputSet;
|
std::vector<std::vector<real_t>> inputUseSet = inputSet;
|
||||||
//Perfom this loop unless and until all k nearest neighbors are found, appended, and returned
|
//Perfom this loop unless and until all k nearest neighbors are found, appended, and returned
|
||||||
for (int i = 0; i < k; i++) {
|
for (int i = 0; i < k; i++) {
|
||||||
int neighbor = 0;
|
int neighbor = 0;
|
||||||
|
@ -8,24 +8,26 @@
|
|||||||
// Created by Marc Melikyan on 10/2/20.
|
// Created by Marc Melikyan on 10/2/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
|
|
||||||
class MLPPKNN {
|
class MLPPKNN {
|
||||||
public:
|
public:
|
||||||
MLPPKNN(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, int k);
|
MLPPKNN(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, int k);
|
||||||
std::vector<double> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<real_t> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
int modelTest(std::vector<double> x);
|
int modelTest(std::vector<real_t> x);
|
||||||
double score();
|
real_t score();
|
||||||
|
|
||||||
private:
|
private:
|
||||||
// Private Model Functions
|
// Private Model Functions
|
||||||
std::vector<double> nearestNeighbors(std::vector<double> x);
|
std::vector<real_t> nearestNeighbors(std::vector<real_t> x);
|
||||||
int determineClass(std::vector<double> knn);
|
int determineClass(std::vector<real_t> knn);
|
||||||
|
|
||||||
// Model Inputs and Parameters
|
// Model Inputs and Parameters
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
int k;
|
int k;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
File diff suppressed because it is too large
Load Diff
@ -8,6 +8,8 @@
|
|||||||
// Created by Marc Melikyan on 1/8/21.
|
// Created by Marc Melikyan on 1/8/21.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <tuple>
|
#include <tuple>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
@ -15,246 +17,246 @@ class MLPPLinAlg {
|
|||||||
public:
|
public:
|
||||||
// MATRIX FUNCTIONS
|
// MATRIX FUNCTIONS
|
||||||
|
|
||||||
std::vector<std::vector<double>> gramMatrix(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> gramMatrix(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
bool linearIndependenceChecker(std::vector<std::vector<double>> A);
|
bool linearIndependenceChecker(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> gaussianNoise(int n, int m);
|
std::vector<std::vector<real_t>> gaussianNoise(int n, int m);
|
||||||
|
|
||||||
std::vector<std::vector<double>> addition(std::vector<std::vector<double>> A, std::vector<std::vector<double>> B);
|
std::vector<std::vector<real_t>> addition(std::vector<std::vector<real_t>> A, std::vector<std::vector<real_t>> B);
|
||||||
|
|
||||||
std::vector<std::vector<double>> subtraction(std::vector<std::vector<double>> A, std::vector<std::vector<double>> B);
|
std::vector<std::vector<real_t>> subtraction(std::vector<std::vector<real_t>> A, std::vector<std::vector<real_t>> B);
|
||||||
|
|
||||||
std::vector<std::vector<double>> matmult(std::vector<std::vector<double>> A, std::vector<std::vector<double>> B);
|
std::vector<std::vector<real_t>> matmult(std::vector<std::vector<real_t>> A, std::vector<std::vector<real_t>> B);
|
||||||
|
|
||||||
std::vector<std::vector<double>> hadamard_product(std::vector<std::vector<double>> A, std::vector<std::vector<double>> B);
|
std::vector<std::vector<real_t>> hadamard_product(std::vector<std::vector<real_t>> A, std::vector<std::vector<real_t>> B);
|
||||||
|
|
||||||
std::vector<std::vector<double>> kronecker_product(std::vector<std::vector<double>> A, std::vector<std::vector<double>> B);
|
std::vector<std::vector<real_t>> kronecker_product(std::vector<std::vector<real_t>> A, std::vector<std::vector<real_t>> B);
|
||||||
|
|
||||||
std::vector<std::vector<double>> elementWiseDivision(std::vector<std::vector<double>> A, std::vector<std::vector<double>> B);
|
std::vector<std::vector<real_t>> elementWiseDivision(std::vector<std::vector<real_t>> A, std::vector<std::vector<real_t>> B);
|
||||||
|
|
||||||
std::vector<std::vector<double>> transpose(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> transpose(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> scalarMultiply(double scalar, std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> scalarMultiply(real_t scalar, std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> scalarAdd(double scalar, std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> scalarAdd(real_t scalar, std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> log(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> log(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> log10(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> log10(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> exp(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> exp(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> erf(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> erf(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> exponentiate(std::vector<std::vector<double>> A, double p);
|
std::vector<std::vector<real_t>> exponentiate(std::vector<std::vector<real_t>> A, real_t p);
|
||||||
|
|
||||||
std::vector<std::vector<double>> sqrt(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> sqrt(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> cbrt(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> cbrt(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> matrixPower(std::vector<std::vector<double>> A, int n);
|
std::vector<std::vector<real_t>> matrixPower(std::vector<std::vector<real_t>> A, int n);
|
||||||
|
|
||||||
std::vector<std::vector<double>> abs(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> abs(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
double det(std::vector<std::vector<double>> A, int d);
|
real_t det(std::vector<std::vector<real_t>> A, int d);
|
||||||
|
|
||||||
double trace(std::vector<std::vector<double>> A);
|
real_t trace(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> cofactor(std::vector<std::vector<double>> A, int n, int i, int j);
|
std::vector<std::vector<real_t>> cofactor(std::vector<std::vector<real_t>> A, int n, int i, int j);
|
||||||
|
|
||||||
std::vector<std::vector<double>> adjoint(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> adjoint(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> inverse(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> inverse(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> pinverse(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> pinverse(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> zeromat(int n, int m);
|
std::vector<std::vector<real_t>> zeromat(int n, int m);
|
||||||
|
|
||||||
std::vector<std::vector<double>> onemat(int n, int m);
|
std::vector<std::vector<real_t>> onemat(int n, int m);
|
||||||
|
|
||||||
std::vector<std::vector<double>> full(int n, int m, int k);
|
std::vector<std::vector<real_t>> full(int n, int m, int k);
|
||||||
|
|
||||||
std::vector<std::vector<double>> sin(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> sin(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> cos(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> cos(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> rotate(std::vector<std::vector<double>> A, double theta, int axis = -1);
|
std::vector<std::vector<real_t>> rotate(std::vector<std::vector<real_t>> A, real_t theta, int axis = -1);
|
||||||
|
|
||||||
std::vector<std::vector<double>> max(std::vector<std::vector<double>> A, std::vector<std::vector<double>> B);
|
std::vector<std::vector<real_t>> max(std::vector<std::vector<real_t>> A, std::vector<std::vector<real_t>> B);
|
||||||
|
|
||||||
double max(std::vector<std::vector<double>> A);
|
real_t max(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
double min(std::vector<std::vector<double>> A);
|
real_t min(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> round(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> round(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
double norm_2(std::vector<std::vector<double>> A);
|
real_t norm_2(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<std::vector<double>> identity(double d);
|
std::vector<std::vector<real_t>> identity(real_t d);
|
||||||
|
|
||||||
std::vector<std::vector<double>> cov(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> cov(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>> eig(std::vector<std::vector<double>> A);
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> eig(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
struct EigenResult {
|
struct EigenResult {
|
||||||
std::vector<std::vector<double>> eigen_vectors;
|
std::vector<std::vector<real_t>> eigen_vectors;
|
||||||
std::vector<std::vector<double>> eigen_values;
|
std::vector<std::vector<real_t>> eigen_values;
|
||||||
};
|
};
|
||||||
|
|
||||||
EigenResult eigen(std::vector<std::vector<double>> A);
|
EigenResult eigen(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>, std::vector<std::vector<double>>> SVD(std::vector<std::vector<double>> A);
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> SVD(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
struct SDVResult {
|
struct SDVResult {
|
||||||
std::vector<std::vector<double>> U;
|
std::vector<std::vector<real_t>> U;
|
||||||
std::vector<std::vector<double>> S;
|
std::vector<std::vector<real_t>> S;
|
||||||
std::vector<std::vector<double>> Vt;
|
std::vector<std::vector<real_t>> Vt;
|
||||||
};
|
};
|
||||||
|
|
||||||
SDVResult svd(std::vector<std::vector<double>> A);
|
SDVResult svd(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<double> vectorProjection(std::vector<double> a, std::vector<double> b);
|
std::vector<real_t> vectorProjection(std::vector<real_t> a, std::vector<real_t> b);
|
||||||
|
|
||||||
std::vector<std::vector<double>> gramSchmidtProcess(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> gramSchmidtProcess(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>> QRD(std::vector<std::vector<double>> A);
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> QRD(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
struct QRDResult {
|
struct QRDResult {
|
||||||
std::vector<std::vector<double>> Q;
|
std::vector<std::vector<real_t>> Q;
|
||||||
std::vector<std::vector<double>> R;
|
std::vector<std::vector<real_t>> R;
|
||||||
};
|
};
|
||||||
|
|
||||||
QRDResult qrd(std::vector<std::vector<double>> A);
|
QRDResult qrd(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>> chol(std::vector<std::vector<double>> A);
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> chol(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
struct CholeskyResult {
|
struct CholeskyResult {
|
||||||
std::vector<std::vector<double>> L;
|
std::vector<std::vector<real_t>> L;
|
||||||
std::vector<std::vector<double>> Lt;
|
std::vector<std::vector<real_t>> Lt;
|
||||||
};
|
};
|
||||||
|
|
||||||
CholeskyResult cholesky(std::vector<std::vector<double>> A);
|
CholeskyResult cholesky(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
double sum_elements(std::vector<std::vector<double>> A);
|
real_t sum_elements(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<double> flatten(std::vector<std::vector<double>> A);
|
std::vector<real_t> flatten(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
std::vector<double> solve(std::vector<std::vector<double>> A, std::vector<double> b);
|
std::vector<real_t> solve(std::vector<std::vector<real_t>> A, std::vector<real_t> b);
|
||||||
|
|
||||||
bool positiveDefiniteChecker(std::vector<std::vector<double>> A);
|
bool positiveDefiniteChecker(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
bool negativeDefiniteChecker(std::vector<std::vector<double>> A);
|
bool negativeDefiniteChecker(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
bool zeroEigenvalue(std::vector<std::vector<double>> A);
|
bool zeroEigenvalue(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
void printMatrix(std::vector<std::vector<double>> A);
|
void printMatrix(std::vector<std::vector<real_t>> A);
|
||||||
|
|
||||||
// VECTOR FUNCTIONS
|
// VECTOR FUNCTIONS
|
||||||
|
|
||||||
std::vector<std::vector<double>> outerProduct(std::vector<double> a, std::vector<double> b); // This multiplies a, bT
|
std::vector<std::vector<real_t>> outerProduct(std::vector<real_t> a, std::vector<real_t> b); // This multiplies a, bT
|
||||||
|
|
||||||
std::vector<double> hadamard_product(std::vector<double> a, std::vector<double> b);
|
std::vector<real_t> hadamard_product(std::vector<real_t> a, std::vector<real_t> b);
|
||||||
|
|
||||||
std::vector<double> elementWiseDivision(std::vector<double> a, std::vector<double> b);
|
std::vector<real_t> elementWiseDivision(std::vector<real_t> a, std::vector<real_t> b);
|
||||||
|
|
||||||
std::vector<double> scalarMultiply(double scalar, std::vector<double> a);
|
std::vector<real_t> scalarMultiply(real_t scalar, std::vector<real_t> a);
|
||||||
|
|
||||||
std::vector<double> scalarAdd(double scalar, std::vector<double> a);
|
std::vector<real_t> scalarAdd(real_t scalar, std::vector<real_t> a);
|
||||||
|
|
||||||
std::vector<double> addition(std::vector<double> a, std::vector<double> b);
|
std::vector<real_t> addition(std::vector<real_t> a, std::vector<real_t> b);
|
||||||
|
|
||||||
std::vector<double> subtraction(std::vector<double> a, std::vector<double> b);
|
std::vector<real_t> subtraction(std::vector<real_t> a, std::vector<real_t> b);
|
||||||
|
|
||||||
std::vector<double> subtractMatrixRows(std::vector<double> a, std::vector<std::vector<double>> B);
|
std::vector<real_t> subtractMatrixRows(std::vector<real_t> a, std::vector<std::vector<real_t>> B);
|
||||||
|
|
||||||
std::vector<double> log(std::vector<double> a);
|
std::vector<real_t> log(std::vector<real_t> a);
|
||||||
|
|
||||||
std::vector<double> log10(std::vector<double> a);
|
std::vector<real_t> log10(std::vector<real_t> a);
|
||||||
|
|
||||||
std::vector<double> exp(std::vector<double> a);
|
std::vector<real_t> exp(std::vector<real_t> a);
|
||||||
|
|
||||||
std::vector<double> erf(std::vector<double> a);
|
std::vector<real_t> erf(std::vector<real_t> a);
|
||||||
|
|
||||||
std::vector<double> exponentiate(std::vector<double> a, double p);
|
std::vector<real_t> exponentiate(std::vector<real_t> a, real_t p);
|
||||||
|
|
||||||
std::vector<double> sqrt(std::vector<double> a);
|
std::vector<real_t> sqrt(std::vector<real_t> a);
|
||||||
|
|
||||||
std::vector<double> cbrt(std::vector<double> a);
|
std::vector<real_t> cbrt(std::vector<real_t> a);
|
||||||
|
|
||||||
double dot(std::vector<double> a, std::vector<double> b);
|
real_t dot(std::vector<real_t> a, std::vector<real_t> b);
|
||||||
|
|
||||||
std::vector<double> cross(std::vector<double> a, std::vector<double> b);
|
std::vector<real_t> cross(std::vector<real_t> a, std::vector<real_t> b);
|
||||||
|
|
||||||
std::vector<double> abs(std::vector<double> a);
|
std::vector<real_t> abs(std::vector<real_t> a);
|
||||||
|
|
||||||
std::vector<double> zerovec(int n);
|
std::vector<real_t> zerovec(int n);
|
||||||
|
|
||||||
std::vector<double> onevec(int n);
|
std::vector<real_t> onevec(int n);
|
||||||
|
|
||||||
std::vector<std::vector<double>> diag(std::vector<double> a);
|
std::vector<std::vector<real_t>> diag(std::vector<real_t> a);
|
||||||
|
|
||||||
std::vector<double> full(int n, int k);
|
std::vector<real_t> full(int n, int k);
|
||||||
|
|
||||||
std::vector<double> sin(std::vector<double> a);
|
std::vector<real_t> sin(std::vector<real_t> a);
|
||||||
|
|
||||||
std::vector<double> cos(std::vector<double> a);
|
std::vector<real_t> cos(std::vector<real_t> a);
|
||||||
|
|
||||||
std::vector<double> max(std::vector<double> a, std::vector<double> b);
|
std::vector<real_t> max(std::vector<real_t> a, std::vector<real_t> b);
|
||||||
|
|
||||||
double max(std::vector<double> a);
|
real_t max(std::vector<real_t> a);
|
||||||
|
|
||||||
double min(std::vector<double> a);
|
real_t min(std::vector<real_t> a);
|
||||||
|
|
||||||
std::vector<double> round(std::vector<double> a);
|
std::vector<real_t> round(std::vector<real_t> a);
|
||||||
|
|
||||||
double euclideanDistance(std::vector<double> a, std::vector<double> b);
|
real_t euclideanDistance(std::vector<real_t> a, std::vector<real_t> b);
|
||||||
|
|
||||||
double norm_2(std::vector<double> a);
|
real_t norm_2(std::vector<real_t> a);
|
||||||
|
|
||||||
double norm_sq(std::vector<double> a);
|
real_t norm_sq(std::vector<real_t> a);
|
||||||
|
|
||||||
double sum_elements(std::vector<double> a);
|
real_t sum_elements(std::vector<real_t> a);
|
||||||
|
|
||||||
double cosineSimilarity(std::vector<double> a, std::vector<double> b);
|
real_t cosineSimilarity(std::vector<real_t> a, std::vector<real_t> b);
|
||||||
|
|
||||||
void printVector(std::vector<double> a);
|
void printVector(std::vector<real_t> a);
|
||||||
|
|
||||||
// MATRIX-VECTOR FUNCTIONS
|
// MATRIX-VECTOR FUNCTIONS
|
||||||
std::vector<std::vector<double>> mat_vec_add(std::vector<std::vector<double>> A, std::vector<double> b);
|
std::vector<std::vector<real_t>> mat_vec_add(std::vector<std::vector<real_t>> A, std::vector<real_t> b);
|
||||||
|
|
||||||
std::vector<double> mat_vec_mult(std::vector<std::vector<double>> A, std::vector<double> b);
|
std::vector<real_t> mat_vec_mult(std::vector<std::vector<real_t>> A, std::vector<real_t> b);
|
||||||
|
|
||||||
// TENSOR FUNCTIONS
|
// TENSOR FUNCTIONS
|
||||||
std::vector<std::vector<std::vector<double>>> addition(std::vector<std::vector<std::vector<double>>> A, std::vector<std::vector<std::vector<double>>> B);
|
std::vector<std::vector<std::vector<real_t>>> addition(std::vector<std::vector<std::vector<real_t>>> A, std::vector<std::vector<std::vector<real_t>>> B);
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> elementWiseDivision(std::vector<std::vector<std::vector<double>>> A, std::vector<std::vector<std::vector<double>>> B);
|
std::vector<std::vector<std::vector<real_t>>> elementWiseDivision(std::vector<std::vector<std::vector<real_t>>> A, std::vector<std::vector<std::vector<real_t>>> B);
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> sqrt(std::vector<std::vector<std::vector<double>>> A);
|
std::vector<std::vector<std::vector<real_t>>> sqrt(std::vector<std::vector<std::vector<real_t>>> A);
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> exponentiate(std::vector<std::vector<std::vector<double>>> A, double p);
|
std::vector<std::vector<std::vector<real_t>>> exponentiate(std::vector<std::vector<std::vector<real_t>>> A, real_t p);
|
||||||
|
|
||||||
std::vector<std::vector<double>> tensor_vec_mult(std::vector<std::vector<std::vector<double>>> A, std::vector<double> b);
|
std::vector<std::vector<real_t>> tensor_vec_mult(std::vector<std::vector<std::vector<real_t>>> A, std::vector<real_t> b);
|
||||||
|
|
||||||
std::vector<double> flatten(std::vector<std::vector<std::vector<double>>> A);
|
std::vector<real_t> flatten(std::vector<std::vector<std::vector<real_t>>> A);
|
||||||
|
|
||||||
void printTensor(std::vector<std::vector<std::vector<double>>> A);
|
void printTensor(std::vector<std::vector<std::vector<real_t>>> A);
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> scalarMultiply(double scalar, std::vector<std::vector<std::vector<double>>> A);
|
std::vector<std::vector<std::vector<real_t>>> scalarMultiply(real_t scalar, std::vector<std::vector<std::vector<real_t>>> A);
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> scalarAdd(double scalar, std::vector<std::vector<std::vector<double>>> A);
|
std::vector<std::vector<std::vector<real_t>>> scalarAdd(real_t scalar, std::vector<std::vector<std::vector<real_t>>> A);
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> resize(std::vector<std::vector<std::vector<double>>> A, std::vector<std::vector<std::vector<double>>> B);
|
std::vector<std::vector<std::vector<real_t>>> resize(std::vector<std::vector<std::vector<real_t>>> A, std::vector<std::vector<std::vector<real_t>>> B);
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> hadamard_product(std::vector<std::vector<std::vector<double>>> A, std::vector<std::vector<std::vector<double>>> B);
|
std::vector<std::vector<std::vector<real_t>>> hadamard_product(std::vector<std::vector<std::vector<real_t>>> A, std::vector<std::vector<std::vector<real_t>>> B);
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> max(std::vector<std::vector<std::vector<double>>> A, std::vector<std::vector<std::vector<double>>> B);
|
std::vector<std::vector<std::vector<real_t>>> max(std::vector<std::vector<std::vector<real_t>>> A, std::vector<std::vector<std::vector<real_t>>> B);
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> abs(std::vector<std::vector<std::vector<double>>> A);
|
std::vector<std::vector<std::vector<real_t>>> abs(std::vector<std::vector<std::vector<real_t>>> A);
|
||||||
|
|
||||||
double norm_2(std::vector<std::vector<std::vector<double>>> A);
|
real_t norm_2(std::vector<std::vector<std::vector<real_t>>> A);
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> vector_wise_tensor_product(std::vector<std::vector<std::vector<double>>> A, std::vector<std::vector<double>> B);
|
std::vector<std::vector<std::vector<real_t>>> vector_wise_tensor_product(std::vector<std::vector<std::vector<real_t>>> A, std::vector<std::vector<real_t>> B);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
};
|
};
|
||||||
|
@ -1,6 +1,8 @@
|
|||||||
#ifndef MLPP_MATRIX_H
|
#ifndef MLPP_MATRIX_H
|
||||||
#define MLPP_MATRIX_H
|
#define MLPP_MATRIX_H
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include "core/containers/pool_vector.h"
|
#include "core/containers/pool_vector.h"
|
||||||
#include "core/containers/sort_array.h"
|
#include "core/containers/sort_array.h"
|
||||||
#include "core/containers/vector.h"
|
#include "core/containers/vector.h"
|
||||||
@ -19,15 +21,15 @@ class MLPPMatrix : public Reference {
|
|||||||
GDCLASS(MLPPMatrix, Reference);
|
GDCLASS(MLPPMatrix, Reference);
|
||||||
|
|
||||||
public:
|
public:
|
||||||
double *ptr() {
|
real_t *ptr() {
|
||||||
return _data;
|
return _data;
|
||||||
}
|
}
|
||||||
|
|
||||||
const double *ptr() const {
|
const real_t *ptr() const {
|
||||||
return _data;
|
return _data;
|
||||||
}
|
}
|
||||||
|
|
||||||
_FORCE_INLINE_ void add_row(const Vector<double> &p_row) {
|
_FORCE_INLINE_ void add_row(const Vector<real_t> &p_row) {
|
||||||
if (_size.x == 0) {
|
if (_size.x == 0) {
|
||||||
_size.x = p_row.size();
|
_size.x = p_row.size();
|
||||||
}
|
}
|
||||||
@ -38,10 +40,10 @@ public:
|
|||||||
|
|
||||||
++_size.y;
|
++_size.y;
|
||||||
|
|
||||||
_data = (double *)memrealloc(_data, data_size() * sizeof(double));
|
_data = (real_t *)memrealloc(_data, data_size() * sizeof(real_t));
|
||||||
CRASH_COND_MSG(!_data, "Out of memory");
|
CRASH_COND_MSG(!_data, "Out of memory");
|
||||||
|
|
||||||
const double *row_arr = p_row.ptr();
|
const real_t *row_arr = p_row.ptr();
|
||||||
|
|
||||||
for (int i = 0; i < p_row.size(); ++i) {
|
for (int i = 0; i < p_row.size(); ++i) {
|
||||||
_data[ci + i] = row_arr[i];
|
_data[ci + i] = row_arr[i];
|
||||||
@ -59,7 +61,7 @@ public:
|
|||||||
|
|
||||||
++_size.y;
|
++_size.y;
|
||||||
|
|
||||||
_data = (double *)memrealloc(_data, data_size() * sizeof(double));
|
_data = (real_t *)memrealloc(_data, data_size() * sizeof(real_t));
|
||||||
CRASH_COND_MSG(!_data, "Out of memory");
|
CRASH_COND_MSG(!_data, "Out of memory");
|
||||||
|
|
||||||
PoolRealArray::Read rread = p_row.read();
|
PoolRealArray::Read rread = p_row.read();
|
||||||
@ -70,7 +72,7 @@ public:
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void remove_row(double p_index) {
|
void remove_row(real_t p_index) {
|
||||||
ERR_FAIL_INDEX(p_index, _size.y);
|
ERR_FAIL_INDEX(p_index, _size.y);
|
||||||
|
|
||||||
--_size.y;
|
--_size.y;
|
||||||
@ -87,7 +89,7 @@ public:
|
|||||||
_data[i] = _data[i + _size.x];
|
_data[i] = _data[i + _size.x];
|
||||||
}
|
}
|
||||||
|
|
||||||
_data = (double *)memrealloc(_data, data_size() * sizeof(double));
|
_data = (real_t *)memrealloc(_data, data_size() * sizeof(real_t));
|
||||||
CRASH_COND_MSG(!_data, "Out of memory");
|
CRASH_COND_MSG(!_data, "Out of memory");
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -113,7 +115,7 @@ public:
|
|||||||
_data[i] = _data[ds + i];
|
_data[i] = _data[ds + i];
|
||||||
}
|
}
|
||||||
|
|
||||||
_data = (double *)memrealloc(_data, data_size() * sizeof(double));
|
_data = (real_t *)memrealloc(_data, data_size() * sizeof(real_t));
|
||||||
CRASH_COND_MSG(!_data, "Out of memory");
|
CRASH_COND_MSG(!_data, "Out of memory");
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -154,26 +156,26 @@ public:
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
_data = (double *)memrealloc(_data, ds * sizeof(double));
|
_data = (real_t *)memrealloc(_data, ds * sizeof(real_t));
|
||||||
CRASH_COND_MSG(!_data, "Out of memory");
|
CRASH_COND_MSG(!_data, "Out of memory");
|
||||||
}
|
}
|
||||||
|
|
||||||
_FORCE_INLINE_ const double &operator[](int p_index) const {
|
_FORCE_INLINE_ const real_t &operator[](int p_index) const {
|
||||||
CRASH_BAD_INDEX(p_index, data_size());
|
CRASH_BAD_INDEX(p_index, data_size());
|
||||||
return _data[p_index];
|
return _data[p_index];
|
||||||
}
|
}
|
||||||
_FORCE_INLINE_ double &operator[](int p_index) {
|
_FORCE_INLINE_ real_t &operator[](int p_index) {
|
||||||
CRASH_BAD_INDEX(p_index, data_size());
|
CRASH_BAD_INDEX(p_index, data_size());
|
||||||
return _data[p_index];
|
return _data[p_index];
|
||||||
}
|
}
|
||||||
|
|
||||||
_FORCE_INLINE_ double get_element(int p_index_x, int p_index_y) const {
|
_FORCE_INLINE_ real_t get_element(int p_index_x, int p_index_y) const {
|
||||||
ERR_FAIL_INDEX_V(p_index_x, _size.x, 0);
|
ERR_FAIL_INDEX_V(p_index_x, _size.x, 0);
|
||||||
ERR_FAIL_INDEX_V(p_index_y, _size.y, 0);
|
ERR_FAIL_INDEX_V(p_index_y, _size.y, 0);
|
||||||
|
|
||||||
return _data[p_index_x * p_index_y];
|
return _data[p_index_x * p_index_y];
|
||||||
}
|
}
|
||||||
_FORCE_INLINE_ double get_element(int p_index_x, int p_index_y) {
|
_FORCE_INLINE_ real_t get_element(int p_index_x, int p_index_y) {
|
||||||
ERR_FAIL_INDEX_V(p_index_x, _size.x, 0);
|
ERR_FAIL_INDEX_V(p_index_x, _size.x, 0);
|
||||||
ERR_FAIL_INDEX_V(p_index_y, _size.y, 0);
|
ERR_FAIL_INDEX_V(p_index_y, _size.y, 0);
|
||||||
|
|
||||||
@ -187,7 +189,7 @@ public:
|
|||||||
return static_cast<real_t>(_data[p_index_x * p_index_y]);
|
return static_cast<real_t>(_data[p_index_x * p_index_y]);
|
||||||
}
|
}
|
||||||
|
|
||||||
_FORCE_INLINE_ void set_element(int p_index_x, int p_index_y, double p_val) {
|
_FORCE_INLINE_ void set_element(int p_index_x, int p_index_y, real_t p_val) {
|
||||||
ERR_FAIL_INDEX(p_index_x, _size.x);
|
ERR_FAIL_INDEX(p_index_x, _size.x);
|
||||||
ERR_FAIL_INDEX(p_index_y, _size.y);
|
ERR_FAIL_INDEX(p_index_y, _size.y);
|
||||||
|
|
||||||
@ -201,13 +203,13 @@ public:
|
|||||||
_data[p_index_x * p_index_y] = p_val;
|
_data[p_index_x * p_index_y] = p_val;
|
||||||
}
|
}
|
||||||
|
|
||||||
_FORCE_INLINE_ void set_row_vector(int p_index_y, const Vector<double> &p_row) {
|
_FORCE_INLINE_ void set_row_vector(int p_index_y, const Vector<real_t> &p_row) {
|
||||||
ERR_FAIL_COND(p_row.size() != _size.x);
|
ERR_FAIL_COND(p_row.size() != _size.x);
|
||||||
ERR_FAIL_INDEX(p_index_y, _size.y);
|
ERR_FAIL_INDEX(p_index_y, _size.y);
|
||||||
|
|
||||||
int ind_start = p_index_y * _size.x;
|
int ind_start = p_index_y * _size.x;
|
||||||
|
|
||||||
const double *row_ptr = p_row.ptr();
|
const real_t *row_ptr = p_row.ptr();
|
||||||
|
|
||||||
for (int i = 0; i < _size.x; ++i) {
|
for (int i = 0; i < _size.x; ++i) {
|
||||||
_data[ind_start + i] = row_ptr[i];
|
_data[ind_start + i] = row_ptr[i];
|
||||||
@ -228,18 +230,18 @@ public:
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void fill(double p_val) {
|
void fill(real_t p_val) {
|
||||||
int ds = data_size();
|
int ds = data_size();
|
||||||
for (int i = 0; i < ds; i++) {
|
for (int i = 0; i < ds; i++) {
|
||||||
_data[i] = p_val;
|
_data[i] = p_val;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
Vector<double> to_flat_vector() const {
|
Vector<real_t> to_flat_vector() const {
|
||||||
Vector<double> ret;
|
Vector<real_t> ret;
|
||||||
ret.resize(data_size());
|
ret.resize(data_size());
|
||||||
double *w = ret.ptrw();
|
real_t *w = ret.ptrw();
|
||||||
memcpy(w, _data, sizeof(double) * data_size());
|
memcpy(w, _data, sizeof(real_t) * data_size());
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -259,9 +261,9 @@ public:
|
|||||||
|
|
||||||
Vector<uint8_t> to_flat_byte_array() const {
|
Vector<uint8_t> to_flat_byte_array() const {
|
||||||
Vector<uint8_t> ret;
|
Vector<uint8_t> ret;
|
||||||
ret.resize(data_size() * sizeof(double));
|
ret.resize(data_size() * sizeof(real_t));
|
||||||
uint8_t *w = ret.ptrw();
|
uint8_t *w = ret.ptrw();
|
||||||
memcpy(w, _data, sizeof(double) * data_size());
|
memcpy(w, _data, sizeof(real_t) * data_size());
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -307,7 +309,7 @@ public:
|
|||||||
|
|
||||||
int start_index = i * _size.x;
|
int start_index = i * _size.x;
|
||||||
|
|
||||||
const double *from_ptr = r->ptr();
|
const real_t *from_ptr = r->ptr();
|
||||||
for (int j = 0; j < _size.x; j++) {
|
for (int j = 0; j < _size.x; j++) {
|
||||||
_data[start_index + j] = from_ptr[j];
|
_data[start_index + j] = from_ptr[j];
|
||||||
}
|
}
|
||||||
@ -342,14 +344,14 @@ public:
|
|||||||
|
|
||||||
int start_index = i * _size.x;
|
int start_index = i * _size.x;
|
||||||
|
|
||||||
const double *from_ptr = r->ptr();
|
const real_t *from_ptr = r->ptr();
|
||||||
for (int j = 0; j < _size.x; j++) {
|
for (int j = 0; j < _size.x; j++) {
|
||||||
_data[start_index + j] = from_ptr[j];
|
_data[start_index + j] = from_ptr[j];
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
_FORCE_INLINE_ void set_from_vectors(const Vector<Vector<double>> &p_from) {
|
_FORCE_INLINE_ void set_from_vectors(const Vector<Vector<real_t>> &p_from) {
|
||||||
if (p_from.size() == 0) {
|
if (p_from.size() == 0) {
|
||||||
reset();
|
reset();
|
||||||
return;
|
return;
|
||||||
@ -363,13 +365,13 @@ public:
|
|||||||
}
|
}
|
||||||
|
|
||||||
for (int i = 0; i < p_from.size(); ++i) {
|
for (int i = 0; i < p_from.size(); ++i) {
|
||||||
const Vector<double> &r = p_from[i];
|
const Vector<real_t> &r = p_from[i];
|
||||||
|
|
||||||
ERR_CONTINUE(r.size() != _size.x);
|
ERR_CONTINUE(r.size() != _size.x);
|
||||||
|
|
||||||
int start_index = i * _size.x;
|
int start_index = i * _size.x;
|
||||||
|
|
||||||
const double *from_ptr = r.ptr();
|
const real_t *from_ptr = r.ptr();
|
||||||
for (int j = 0; j < _size.x; j++) {
|
for (int j = 0; j < _size.x; j++) {
|
||||||
_data[start_index + j] = from_ptr[j];
|
_data[start_index + j] = from_ptr[j];
|
||||||
}
|
}
|
||||||
@ -420,7 +422,7 @@ public:
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
MLPPMatrix(const Vector<Vector<double>> &p_from) {
|
MLPPMatrix(const Vector<Vector<real_t>> &p_from) {
|
||||||
_data = NULL;
|
_data = NULL;
|
||||||
|
|
||||||
set_from_vectors(p_from);
|
set_from_vectors(p_from);
|
||||||
@ -439,15 +441,15 @@ public:
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TODO: These are temporary
|
// TODO: These are temporary
|
||||||
std::vector<double> to_flat_std_vector() const {
|
std::vector<real_t> to_flat_std_vector() const {
|
||||||
std::vector<double> ret;
|
std::vector<real_t> ret;
|
||||||
ret.resize(data_size());
|
ret.resize(data_size());
|
||||||
double *w = &ret[0];
|
real_t *w = &ret[0];
|
||||||
memcpy(w, _data, sizeof(double) * data_size());
|
memcpy(w, _data, sizeof(real_t) * data_size());
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
_FORCE_INLINE_ void set_from_std_vectors(const std::vector<std::vector<double>> &p_from) {
|
_FORCE_INLINE_ void set_from_std_vectors(const std::vector<std::vector<real_t>> &p_from) {
|
||||||
if (p_from.size() == 0) {
|
if (p_from.size() == 0) {
|
||||||
reset();
|
reset();
|
||||||
return;
|
return;
|
||||||
@ -461,33 +463,33 @@ public:
|
|||||||
}
|
}
|
||||||
|
|
||||||
for (uint32_t i = 0; i < p_from.size(); ++i) {
|
for (uint32_t i = 0; i < p_from.size(); ++i) {
|
||||||
const std::vector<double> &r = p_from[i];
|
const std::vector<real_t> &r = p_from[i];
|
||||||
|
|
||||||
ERR_CONTINUE(r.size() != static_cast<uint32_t>(_size.x));
|
ERR_CONTINUE(r.size() != static_cast<uint32_t>(_size.x));
|
||||||
|
|
||||||
int start_index = i * _size.x;
|
int start_index = i * _size.x;
|
||||||
|
|
||||||
const double *from_ptr = &r[0];
|
const real_t *from_ptr = &r[0];
|
||||||
for (int j = 0; j < _size.x; j++) {
|
for (int j = 0; j < _size.x; j++) {
|
||||||
_data[start_index + j] = from_ptr[j];
|
_data[start_index + j] = from_ptr[j];
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
_FORCE_INLINE_ void set_row_std_vector(int p_index_y, const std::vector<double> &p_row) {
|
_FORCE_INLINE_ void set_row_std_vector(int p_index_y, const std::vector<real_t> &p_row) {
|
||||||
ERR_FAIL_COND(p_row.size() != static_cast<uint32_t>(_size.x));
|
ERR_FAIL_COND(p_row.size() != static_cast<uint32_t>(_size.x));
|
||||||
ERR_FAIL_INDEX(p_index_y, _size.y);
|
ERR_FAIL_INDEX(p_index_y, _size.y);
|
||||||
|
|
||||||
int ind_start = p_index_y * _size.x;
|
int ind_start = p_index_y * _size.x;
|
||||||
|
|
||||||
const double *row_ptr = &p_row[0];
|
const real_t *row_ptr = &p_row[0];
|
||||||
|
|
||||||
for (int i = 0; i < _size.x; ++i) {
|
for (int i = 0; i < _size.x; ++i) {
|
||||||
_data[ind_start + i] = row_ptr[i];
|
_data[ind_start + i] = row_ptr[i];
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
MLPPMatrix(const std::vector<std::vector<double>> &p_from) {
|
MLPPMatrix(const std::vector<std::vector<real_t>> &p_from) {
|
||||||
_data = NULL;
|
_data = NULL;
|
||||||
|
|
||||||
set_from_std_vectors(p_from);
|
set_from_std_vectors(p_from);
|
||||||
@ -498,7 +500,7 @@ protected:
|
|||||||
|
|
||||||
protected:
|
protected:
|
||||||
Size2i _size;
|
Size2i _size;
|
||||||
double *_data;
|
real_t *_data;
|
||||||
};
|
};
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
@ -1,6 +1,8 @@
|
|||||||
#ifndef MLPP_VECTOR_H
|
#ifndef MLPP_VECTOR_H
|
||||||
#define MLPP_VECTOR_H
|
#define MLPP_VECTOR_H
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include "core/containers/pool_vector.h"
|
#include "core/containers/pool_vector.h"
|
||||||
#include "core/containers/sort_array.h"
|
#include "core/containers/sort_array.h"
|
||||||
#include "core/containers/vector.h"
|
#include "core/containers/vector.h"
|
||||||
@ -16,24 +18,24 @@ class MLPPVector : public Reference {
|
|||||||
GDCLASS(MLPPVector, Reference);
|
GDCLASS(MLPPVector, Reference);
|
||||||
|
|
||||||
public:
|
public:
|
||||||
double *ptr() {
|
real_t *ptr() {
|
||||||
return _data;
|
return _data;
|
||||||
}
|
}
|
||||||
|
|
||||||
const double *ptr() const {
|
const real_t *ptr() const {
|
||||||
return _data;
|
return _data;
|
||||||
}
|
}
|
||||||
|
|
||||||
_FORCE_INLINE_ void push_back(double p_elem) {
|
_FORCE_INLINE_ void push_back(real_t p_elem) {
|
||||||
++_size;
|
++_size;
|
||||||
|
|
||||||
_data = (double *)memrealloc(_data, _size * sizeof(double));
|
_data = (real_t *)memrealloc(_data, _size * sizeof(real_t));
|
||||||
CRASH_COND_MSG(!_data, "Out of memory");
|
CRASH_COND_MSG(!_data, "Out of memory");
|
||||||
|
|
||||||
_data[_size - 1] = p_elem;
|
_data[_size - 1] = p_elem;
|
||||||
}
|
}
|
||||||
|
|
||||||
void remove(double p_index) {
|
void remove(real_t p_index) {
|
||||||
ERR_FAIL_INDEX(p_index, _size);
|
ERR_FAIL_INDEX(p_index, _size);
|
||||||
|
|
||||||
--_size;
|
--_size;
|
||||||
@ -42,7 +44,7 @@ public:
|
|||||||
_data[i] = _data[i + 1];
|
_data[i] = _data[i + 1];
|
||||||
}
|
}
|
||||||
|
|
||||||
_data = (double *)memrealloc(_data, _size * sizeof(double));
|
_data = (real_t *)memrealloc(_data, _size * sizeof(real_t));
|
||||||
CRASH_COND_MSG(!_data, "Out of memory");
|
CRASH_COND_MSG(!_data, "Out of memory");
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -56,18 +58,18 @@ public:
|
|||||||
_data[p_index] = _data[_size];
|
_data[p_index] = _data[_size];
|
||||||
}
|
}
|
||||||
|
|
||||||
_data = (double *)memrealloc(_data, _size * sizeof(double));
|
_data = (real_t *)memrealloc(_data, _size * sizeof(real_t));
|
||||||
CRASH_COND_MSG(!_data, "Out of memory");
|
CRASH_COND_MSG(!_data, "Out of memory");
|
||||||
}
|
}
|
||||||
|
|
||||||
void erase(const double &p_val) {
|
void erase(const real_t &p_val) {
|
||||||
int idx = find(p_val);
|
int idx = find(p_val);
|
||||||
if (idx >= 0) {
|
if (idx >= 0) {
|
||||||
remove(idx);
|
remove(idx);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
int erase_multiple_unordered(const double &p_val) {
|
int erase_multiple_unordered(const real_t &p_val) {
|
||||||
int from = 0;
|
int from = 0;
|
||||||
int count = 0;
|
int count = 0;
|
||||||
while (true) {
|
while (true) {
|
||||||
@ -111,24 +113,24 @@ public:
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
_data = (double *)memrealloc(_data, _size * sizeof(double));
|
_data = (real_t *)memrealloc(_data, _size * sizeof(real_t));
|
||||||
CRASH_COND_MSG(!_data, "Out of memory");
|
CRASH_COND_MSG(!_data, "Out of memory");
|
||||||
}
|
}
|
||||||
|
|
||||||
_FORCE_INLINE_ const double &operator[](int p_index) const {
|
_FORCE_INLINE_ const real_t &operator[](int p_index) const {
|
||||||
CRASH_BAD_INDEX(p_index, _size);
|
CRASH_BAD_INDEX(p_index, _size);
|
||||||
return _data[p_index];
|
return _data[p_index];
|
||||||
}
|
}
|
||||||
_FORCE_INLINE_ double &operator[](int p_index) {
|
_FORCE_INLINE_ real_t &operator[](int p_index) {
|
||||||
CRASH_BAD_INDEX(p_index, _size);
|
CRASH_BAD_INDEX(p_index, _size);
|
||||||
return _data[p_index];
|
return _data[p_index];
|
||||||
}
|
}
|
||||||
|
|
||||||
_FORCE_INLINE_ double get_element(int p_index) const {
|
_FORCE_INLINE_ real_t get_element(int p_index) const {
|
||||||
ERR_FAIL_INDEX_V(p_index, _size, 0);
|
ERR_FAIL_INDEX_V(p_index, _size, 0);
|
||||||
return _data[p_index];
|
return _data[p_index];
|
||||||
}
|
}
|
||||||
_FORCE_INLINE_ double get_element(int p_index) {
|
_FORCE_INLINE_ real_t get_element(int p_index) {
|
||||||
ERR_FAIL_INDEX_V(p_index, _size, 0);
|
ERR_FAIL_INDEX_V(p_index, _size, 0);
|
||||||
return _data[p_index];
|
return _data[p_index];
|
||||||
}
|
}
|
||||||
@ -138,7 +140,7 @@ public:
|
|||||||
return static_cast<real_t>(_data[p_index]);
|
return static_cast<real_t>(_data[p_index]);
|
||||||
}
|
}
|
||||||
|
|
||||||
_FORCE_INLINE_ void set_element(int p_index, double p_val) {
|
_FORCE_INLINE_ void set_element(int p_index, real_t p_val) {
|
||||||
ERR_FAIL_INDEX(p_index, _size);
|
ERR_FAIL_INDEX(p_index, _size);
|
||||||
_data[p_index] = p_val;
|
_data[p_index] = p_val;
|
||||||
}
|
}
|
||||||
@ -148,13 +150,13 @@ public:
|
|||||||
_data[p_index] = p_val;
|
_data[p_index] = p_val;
|
||||||
}
|
}
|
||||||
|
|
||||||
void fill(double p_val) {
|
void fill(real_t p_val) {
|
||||||
for (int i = 0; i < _size; i++) {
|
for (int i = 0; i < _size; i++) {
|
||||||
_data[i] = p_val;
|
_data[i] = p_val;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void insert(int p_pos, double p_val) {
|
void insert(int p_pos, real_t p_val) {
|
||||||
ERR_FAIL_INDEX(p_pos, _size + 1);
|
ERR_FAIL_INDEX(p_pos, _size + 1);
|
||||||
if (p_pos == _size) {
|
if (p_pos == _size) {
|
||||||
push_back(p_val);
|
push_back(p_val);
|
||||||
@ -167,7 +169,7 @@ public:
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
int find(const double &p_val, int p_from = 0) const {
|
int find(const real_t &p_val, int p_from = 0) const {
|
||||||
for (int i = p_from; i < _size; i++) {
|
for (int i = p_from; i < _size; i++) {
|
||||||
if (_data[i] == p_val) {
|
if (_data[i] == p_val) {
|
||||||
return i;
|
return i;
|
||||||
@ -183,15 +185,15 @@ public:
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
SortArray<double, C> sorter;
|
SortArray<real_t, C> sorter;
|
||||||
sorter.sort(_data, len);
|
sorter.sort(_data, len);
|
||||||
}
|
}
|
||||||
|
|
||||||
void sort() {
|
void sort() {
|
||||||
sort_custom<_DefaultComparator<double>>();
|
sort_custom<_DefaultComparator<real_t>>();
|
||||||
}
|
}
|
||||||
|
|
||||||
void ordered_insert(double p_val) {
|
void ordered_insert(real_t p_val) {
|
||||||
int i;
|
int i;
|
||||||
for (i = 0; i < _size; i++) {
|
for (i = 0; i < _size; i++) {
|
||||||
if (p_val < _data[i]) {
|
if (p_val < _data[i]) {
|
||||||
@ -201,11 +203,11 @@ public:
|
|||||||
insert(i, p_val);
|
insert(i, p_val);
|
||||||
}
|
}
|
||||||
|
|
||||||
Vector<double> to_vector() const {
|
Vector<real_t> to_vector() const {
|
||||||
Vector<double> ret;
|
Vector<real_t> ret;
|
||||||
ret.resize(size());
|
ret.resize(size());
|
||||||
double *w = ret.ptrw();
|
real_t *w = ret.ptrw();
|
||||||
memcpy(w, _data, sizeof(double) * _size);
|
memcpy(w, _data, sizeof(real_t) * _size);
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -225,9 +227,9 @@ public:
|
|||||||
|
|
||||||
Vector<uint8_t> to_byte_array() const {
|
Vector<uint8_t> to_byte_array() const {
|
||||||
Vector<uint8_t> ret;
|
Vector<uint8_t> ret;
|
||||||
ret.resize(_size * sizeof(double));
|
ret.resize(_size * sizeof(real_t));
|
||||||
uint8_t *w = ret.ptrw();
|
uint8_t *w = ret.ptrw();
|
||||||
memcpy(w, _data, sizeof(double) * _size);
|
memcpy(w, _data, sizeof(real_t) * _size);
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -255,7 +257,7 @@ public:
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
_FORCE_INLINE_ void set_from_vector(const Vector<double> &p_from) {
|
_FORCE_INLINE_ void set_from_vector(const Vector<real_t> &p_from) {
|
||||||
resize(p_from.size());
|
resize(p_from.size());
|
||||||
for (int i = 0; i < _size; i++) {
|
for (int i = 0; i < _size; i++) {
|
||||||
_data[i] = p_from[i];
|
_data[i] = p_from[i];
|
||||||
@ -286,7 +288,7 @@ public:
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
MLPPVector(const Vector<double> &p_from) {
|
MLPPVector(const Vector<real_t> &p_from) {
|
||||||
_size = 0;
|
_size = 0;
|
||||||
_data = NULL;
|
_data = NULL;
|
||||||
|
|
||||||
@ -314,22 +316,22 @@ public:
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TODO: These are temporary
|
// TODO: These are temporary
|
||||||
std::vector<double> to_std_vector() const {
|
std::vector<real_t> to_std_vector() const {
|
||||||
std::vector<double> ret;
|
std::vector<real_t> ret;
|
||||||
ret.resize(size());
|
ret.resize(size());
|
||||||
double *w = &ret[0];
|
real_t *w = &ret[0];
|
||||||
memcpy(w, _data, sizeof(double) * _size);
|
memcpy(w, _data, sizeof(real_t) * _size);
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
_FORCE_INLINE_ void set_from_std_vector(const std::vector<double> &p_from) {
|
_FORCE_INLINE_ void set_from_std_vector(const std::vector<real_t> &p_from) {
|
||||||
resize(p_from.size());
|
resize(p_from.size());
|
||||||
for (int i = 0; i < _size; i++) {
|
for (int i = 0; i < _size; i++) {
|
||||||
_data[i] = p_from[i];
|
_data[i] = p_from[i];
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
MLPPVector(const std::vector<double> &p_from) {
|
MLPPVector(const std::vector<real_t> &p_from) {
|
||||||
_size = 0;
|
_size = 0;
|
||||||
_data = NULL;
|
_data = NULL;
|
||||||
|
|
||||||
@ -344,7 +346,7 @@ protected:
|
|||||||
|
|
||||||
protected:
|
protected:
|
||||||
int _size;
|
int _size;
|
||||||
double *_data;
|
real_t *_data;
|
||||||
};
|
};
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
@ -15,7 +15,7 @@
|
|||||||
#include <iostream>
|
#include <iostream>
|
||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
MLPPLinReg::MLPPLinReg(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, std::string reg, double lambda, double alpha) :
|
MLPPLinReg::MLPPLinReg(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, std::string reg, real_t lambda, real_t alpha) :
|
||||||
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
||||||
y_hat.resize(n);
|
y_hat.resize(n);
|
||||||
|
|
||||||
@ -23,28 +23,28 @@ MLPPLinReg::MLPPLinReg(std::vector<std::vector<double>> inputSet, std::vector<do
|
|||||||
bias = MLPPUtilities::biasInitialization();
|
bias = MLPPUtilities::biasInitialization();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPLinReg::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPLinReg::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
return Evaluate(X);
|
return Evaluate(X);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPLinReg::modelTest(std::vector<double> x) {
|
real_t MLPPLinReg::modelTest(std::vector<real_t> x) {
|
||||||
return Evaluate(x);
|
return Evaluate(x);
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPLinReg::NewtonRaphson(double learning_rate, int max_epoch, bool UI) {
|
void MLPPLinReg::NewtonRaphson(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
while (true) {
|
while (true) {
|
||||||
cost_prev = Cost(y_hat, outputSet);
|
cost_prev = Cost(y_hat, outputSet);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputSet);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputSet);
|
||||||
|
|
||||||
// Calculating the weight gradients (2nd derivative)
|
// Calculating the weight gradients (2nd derivative)
|
||||||
std::vector<double> first_derivative = alg.mat_vec_mult(alg.transpose(inputSet), error);
|
std::vector<real_t> first_derivative = alg.mat_vec_mult(alg.transpose(inputSet), error);
|
||||||
std::vector<std::vector<double>> second_derivative = alg.matmult(alg.transpose(inputSet), inputSet);
|
std::vector<std::vector<real_t>> second_derivative = alg.matmult(alg.transpose(inputSet), inputSet);
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(alg.inverse(second_derivative)), first_derivative)));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(alg.inverse(second_derivative)), first_derivative)));
|
||||||
weights = regularization.regWeights(weights, lambda, alpha, reg);
|
weights = regularization.regWeights(weights, lambda, alpha, reg);
|
||||||
|
|
||||||
@ -63,17 +63,17 @@ void MLPPLinReg::NewtonRaphson(double learning_rate, int max_epoch, bool UI) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPLinReg::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
void MLPPLinReg::gradientDescent(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
cost_prev = Cost(y_hat, outputSet);
|
cost_prev = Cost(y_hat, outputSet);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputSet);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputSet);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputSet), error)));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputSet), error)));
|
||||||
@ -94,10 +94,10 @@ void MLPPLinReg::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPLinReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
void MLPPLinReg::SGD(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
@ -106,10 +106,10 @@ void MLPPLinReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
||||||
int outputIndex = distribution(generator);
|
int outputIndex = distribution(generator);
|
||||||
|
|
||||||
double y_hat = Evaluate(inputSet[outputIndex]);
|
real_t y_hat = Evaluate(inputSet[outputIndex]);
|
||||||
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
||||||
|
|
||||||
double error = y_hat - outputSet[outputIndex];
|
real_t error = y_hat - outputSet[outputIndex];
|
||||||
|
|
||||||
// Weight updation
|
// Weight updation
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate * error, inputSet[outputIndex]));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate * error, inputSet[outputIndex]));
|
||||||
@ -133,10 +133,10 @@ void MLPPLinReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPLinReg::MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
void MLPPLinReg::MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
@ -145,10 +145,10 @@ void MLPPLinReg::MBGD(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error)));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error)));
|
||||||
@ -171,10 +171,10 @@ void MLPPLinReg::MBGD(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPLinReg::Momentum(double learning_rate, int max_epoch, int mini_batch_size, double gamma, bool UI) {
|
void MLPPLinReg::Momentum(real_t learning_rate, int max_epoch, int mini_batch_size, real_t gamma, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
@ -182,18 +182,18 @@ void MLPPLinReg::Momentum(double learning_rate, int max_epoch, int mini_batch_si
|
|||||||
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
||||||
|
|
||||||
// Initializing necessary components for Momentum.
|
// Initializing necessary components for Momentum.
|
||||||
std::vector<double> v = alg.zerovec(weights.size());
|
std::vector<real_t> v = alg.zerovec(weights.size());
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
std::vector<double> gradient = alg.scalarMultiply(1 / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error));
|
std::vector<real_t> gradient = alg.scalarMultiply(1 / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error));
|
||||||
std::vector<double> RegDerivTerm = regularization.regDerivTerm(weights, lambda, alpha, reg);
|
std::vector<real_t> RegDerivTerm = regularization.regDerivTerm(weights, lambda, alpha, reg);
|
||||||
std::vector<double> weight_grad = alg.addition(gradient, RegDerivTerm); // Weight_grad_final
|
std::vector<real_t> weight_grad = alg.addition(gradient, RegDerivTerm); // Weight_grad_final
|
||||||
|
|
||||||
v = alg.addition(alg.scalarMultiply(gamma, v), alg.scalarMultiply(learning_rate, weight_grad));
|
v = alg.addition(alg.scalarMultiply(gamma, v), alg.scalarMultiply(learning_rate, weight_grad));
|
||||||
|
|
||||||
@ -216,10 +216,10 @@ void MLPPLinReg::Momentum(double learning_rate, int max_epoch, int mini_batch_si
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPLinReg::NAG(double learning_rate, int max_epoch, int mini_batch_size, double gamma, bool UI) {
|
void MLPPLinReg::NAG(real_t learning_rate, int max_epoch, int mini_batch_size, real_t gamma, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
@ -227,20 +227,20 @@ void MLPPLinReg::NAG(double learning_rate, int max_epoch, int mini_batch_size, d
|
|||||||
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
||||||
|
|
||||||
// Initializing necessary components for Momentum.
|
// Initializing necessary components for Momentum.
|
||||||
std::vector<double> v = alg.zerovec(weights.size());
|
std::vector<real_t> v = alg.zerovec(weights.size());
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(gamma, v)); // "Aposterori" calculation
|
weights = alg.subtraction(weights, alg.scalarMultiply(gamma, v)); // "Aposterori" calculation
|
||||||
|
|
||||||
std::vector<double> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
std::vector<double> gradient = alg.scalarMultiply(1 / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error));
|
std::vector<real_t> gradient = alg.scalarMultiply(1 / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error));
|
||||||
std::vector<double> RegDerivTerm = regularization.regDerivTerm(weights, lambda, alpha, reg);
|
std::vector<real_t> RegDerivTerm = regularization.regDerivTerm(weights, lambda, alpha, reg);
|
||||||
std::vector<double> weight_grad = alg.addition(gradient, RegDerivTerm); // Weight_grad_final
|
std::vector<real_t> weight_grad = alg.addition(gradient, RegDerivTerm); // Weight_grad_final
|
||||||
|
|
||||||
v = alg.addition(alg.scalarMultiply(gamma, v), alg.scalarMultiply(learning_rate, weight_grad));
|
v = alg.addition(alg.scalarMultiply(gamma, v), alg.scalarMultiply(learning_rate, weight_grad));
|
||||||
|
|
||||||
@ -263,10 +263,10 @@ void MLPPLinReg::NAG(double learning_rate, int max_epoch, int mini_batch_size, d
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPLinReg::Adagrad(double learning_rate, int max_epoch, int mini_batch_size, double e, bool UI) {
|
void MLPPLinReg::Adagrad(real_t learning_rate, int max_epoch, int mini_batch_size, real_t e, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
@ -274,18 +274,18 @@ void MLPPLinReg::Adagrad(double learning_rate, int max_epoch, int mini_batch_siz
|
|||||||
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
||||||
|
|
||||||
// Initializing necessary components for Adagrad.
|
// Initializing necessary components for Adagrad.
|
||||||
std::vector<double> v = alg.zerovec(weights.size());
|
std::vector<real_t> v = alg.zerovec(weights.size());
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
std::vector<double> gradient = alg.scalarMultiply(1 / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error));
|
std::vector<real_t> gradient = alg.scalarMultiply(1 / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error));
|
||||||
std::vector<double> RegDerivTerm = regularization.regDerivTerm(weights, lambda, alpha, reg);
|
std::vector<real_t> RegDerivTerm = regularization.regDerivTerm(weights, lambda, alpha, reg);
|
||||||
std::vector<double> weight_grad = alg.addition(gradient, RegDerivTerm); // Weight_grad_final
|
std::vector<real_t> weight_grad = alg.addition(gradient, RegDerivTerm); // Weight_grad_final
|
||||||
|
|
||||||
v = alg.hadamard_product(weight_grad, weight_grad);
|
v = alg.hadamard_product(weight_grad, weight_grad);
|
||||||
|
|
||||||
@ -308,11 +308,11 @@ void MLPPLinReg::Adagrad(double learning_rate, int max_epoch, int mini_batch_siz
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPLinReg::Adadelta(double learning_rate, int max_epoch, int mini_batch_size, double b1, double e, bool UI) {
|
void MLPPLinReg::Adadelta(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t e, bool UI) {
|
||||||
// Adagrad upgrade. Momentum is applied.
|
// Adagrad upgrade. Momentum is applied.
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
@ -320,18 +320,18 @@ void MLPPLinReg::Adadelta(double learning_rate, int max_epoch, int mini_batch_si
|
|||||||
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
||||||
|
|
||||||
// Initializing necessary components for Adagrad.
|
// Initializing necessary components for Adagrad.
|
||||||
std::vector<double> v = alg.zerovec(weights.size());
|
std::vector<real_t> v = alg.zerovec(weights.size());
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
std::vector<double> gradient = alg.scalarMultiply(1 / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error));
|
std::vector<real_t> gradient = alg.scalarMultiply(1 / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error));
|
||||||
std::vector<double> RegDerivTerm = regularization.regDerivTerm(weights, lambda, alpha, reg);
|
std::vector<real_t> RegDerivTerm = regularization.regDerivTerm(weights, lambda, alpha, reg);
|
||||||
std::vector<double> weight_grad = alg.addition(gradient, RegDerivTerm); // Weight_grad_final
|
std::vector<real_t> weight_grad = alg.addition(gradient, RegDerivTerm); // Weight_grad_final
|
||||||
|
|
||||||
v = alg.addition(alg.scalarMultiply(b1, v), alg.scalarMultiply(1 - b1, alg.hadamard_product(weight_grad, weight_grad)));
|
v = alg.addition(alg.scalarMultiply(b1, v), alg.scalarMultiply(1 - b1, alg.hadamard_product(weight_grad, weight_grad)));
|
||||||
|
|
||||||
@ -354,10 +354,10 @@ void MLPPLinReg::Adadelta(double learning_rate, int max_epoch, int mini_batch_si
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPLinReg::Adam(double learning_rate, int max_epoch, int mini_batch_size, double b1, double b2, double e, bool UI) {
|
void MLPPLinReg::Adam(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t b2, real_t e, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
@ -365,26 +365,26 @@ void MLPPLinReg::Adam(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
||||||
|
|
||||||
// Initializing necessary components for Adam.
|
// Initializing necessary components for Adam.
|
||||||
std::vector<double> m = alg.zerovec(weights.size());
|
std::vector<real_t> m = alg.zerovec(weights.size());
|
||||||
|
|
||||||
std::vector<double> v = alg.zerovec(weights.size());
|
std::vector<real_t> v = alg.zerovec(weights.size());
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
std::vector<double> gradient = alg.scalarMultiply(1 / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error));
|
std::vector<real_t> gradient = alg.scalarMultiply(1 / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error));
|
||||||
std::vector<double> RegDerivTerm = regularization.regDerivTerm(weights, lambda, alpha, reg);
|
std::vector<real_t> RegDerivTerm = regularization.regDerivTerm(weights, lambda, alpha, reg);
|
||||||
std::vector<double> weight_grad = alg.addition(gradient, RegDerivTerm); // Weight_grad_final
|
std::vector<real_t> weight_grad = alg.addition(gradient, RegDerivTerm); // Weight_grad_final
|
||||||
|
|
||||||
m = alg.addition(alg.scalarMultiply(b1, m), alg.scalarMultiply(1 - b1, weight_grad));
|
m = alg.addition(alg.scalarMultiply(b1, m), alg.scalarMultiply(1 - b1, weight_grad));
|
||||||
v = alg.addition(alg.scalarMultiply(b2, v), alg.scalarMultiply(1 - b2, alg.exponentiate(weight_grad, 2)));
|
v = alg.addition(alg.scalarMultiply(b2, v), alg.scalarMultiply(1 - b2, alg.exponentiate(weight_grad, 2)));
|
||||||
|
|
||||||
std::vector<double> m_hat = alg.scalarMultiply(1 / (1 - pow(b1, epoch)), m);
|
std::vector<real_t> m_hat = alg.scalarMultiply(1 / (1 - pow(b1, epoch)), m);
|
||||||
std::vector<double> v_hat = alg.scalarMultiply(1 / (1 - pow(b2, epoch)), v);
|
std::vector<real_t> v_hat = alg.scalarMultiply(1 / (1 - pow(b2, epoch)), v);
|
||||||
|
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate, alg.elementWiseDivision(m_hat, alg.scalarAdd(e, alg.sqrt(v_hat)))));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate, alg.elementWiseDivision(m_hat, alg.scalarAdd(e, alg.sqrt(v_hat)))));
|
||||||
|
|
||||||
@ -405,35 +405,35 @@ void MLPPLinReg::Adam(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPLinReg::Adamax(double learning_rate, int max_epoch, int mini_batch_size, double b1, double b2, double e, bool UI) {
|
void MLPPLinReg::Adamax(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t b2, real_t e, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
int n_mini_batch = n / mini_batch_size;
|
int n_mini_batch = n / mini_batch_size;
|
||||||
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
||||||
|
|
||||||
std::vector<double> m = alg.zerovec(weights.size());
|
std::vector<real_t> m = alg.zerovec(weights.size());
|
||||||
|
|
||||||
std::vector<double> u = alg.zerovec(weights.size());
|
std::vector<real_t> u = alg.zerovec(weights.size());
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
std::vector<double> gradient = alg.scalarMultiply(1 / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error));
|
std::vector<real_t> gradient = alg.scalarMultiply(1 / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error));
|
||||||
std::vector<double> RegDerivTerm = regularization.regDerivTerm(weights, lambda, alpha, reg);
|
std::vector<real_t> RegDerivTerm = regularization.regDerivTerm(weights, lambda, alpha, reg);
|
||||||
std::vector<double> weight_grad = alg.addition(gradient, RegDerivTerm); // Weight_grad_final
|
std::vector<real_t> weight_grad = alg.addition(gradient, RegDerivTerm); // Weight_grad_final
|
||||||
|
|
||||||
m = alg.addition(alg.scalarMultiply(b1, m), alg.scalarMultiply(1 - b1, weight_grad));
|
m = alg.addition(alg.scalarMultiply(b1, m), alg.scalarMultiply(1 - b1, weight_grad));
|
||||||
u = alg.max(alg.scalarMultiply(b2, u), alg.abs(weight_grad));
|
u = alg.max(alg.scalarMultiply(b2, u), alg.abs(weight_grad));
|
||||||
|
|
||||||
std::vector<double> m_hat = alg.scalarMultiply(1 / (1 - pow(b1, epoch)), m);
|
std::vector<real_t> m_hat = alg.scalarMultiply(1 / (1 - pow(b1, epoch)), m);
|
||||||
|
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate, alg.elementWiseDivision(m_hat, u)));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate, alg.elementWiseDivision(m_hat, u)));
|
||||||
|
|
||||||
@ -454,10 +454,10 @@ void MLPPLinReg::Adamax(double learning_rate, int max_epoch, int mini_batch_size
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPLinReg::Nadam(double learning_rate, int max_epoch, int mini_batch_size, double b1, double b2, double e, bool UI) {
|
void MLPPLinReg::Nadam(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t b2, real_t e, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
@ -465,27 +465,27 @@ void MLPPLinReg::Nadam(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
auto [inputMiniBatches, outputMiniBatches] = MLPPUtilities::createMiniBatches(inputSet, outputSet, n_mini_batch);
|
||||||
|
|
||||||
// Initializing necessary components for Adam.
|
// Initializing necessary components for Adam.
|
||||||
std::vector<double> m = alg.zerovec(weights.size());
|
std::vector<real_t> m = alg.zerovec(weights.size());
|
||||||
std::vector<double> v = alg.zerovec(weights.size());
|
std::vector<real_t> v = alg.zerovec(weights.size());
|
||||||
std::vector<double> m_final = alg.zerovec(weights.size());
|
std::vector<real_t> m_final = alg.zerovec(weights.size());
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
std::vector<double> gradient = alg.scalarMultiply(1 / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error));
|
std::vector<real_t> gradient = alg.scalarMultiply(1 / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error));
|
||||||
std::vector<double> RegDerivTerm = regularization.regDerivTerm(weights, lambda, alpha, reg);
|
std::vector<real_t> RegDerivTerm = regularization.regDerivTerm(weights, lambda, alpha, reg);
|
||||||
std::vector<double> weight_grad = alg.addition(gradient, RegDerivTerm); // Weight_grad_final
|
std::vector<real_t> weight_grad = alg.addition(gradient, RegDerivTerm); // Weight_grad_final
|
||||||
|
|
||||||
m = alg.addition(alg.scalarMultiply(b1, m), alg.scalarMultiply(1 - b1, weight_grad));
|
m = alg.addition(alg.scalarMultiply(b1, m), alg.scalarMultiply(1 - b1, weight_grad));
|
||||||
v = alg.addition(alg.scalarMultiply(b2, v), alg.scalarMultiply(1 - b2, alg.exponentiate(weight_grad, 2)));
|
v = alg.addition(alg.scalarMultiply(b2, v), alg.scalarMultiply(1 - b2, alg.exponentiate(weight_grad, 2)));
|
||||||
m_final = alg.addition(alg.scalarMultiply(b1, m), alg.scalarMultiply((1 - b1) / (1 - pow(b1, epoch)), weight_grad));
|
m_final = alg.addition(alg.scalarMultiply(b1, m), alg.scalarMultiply((1 - b1) / (1 - pow(b1, epoch)), weight_grad));
|
||||||
|
|
||||||
std::vector<double> m_hat = alg.scalarMultiply(1 / (1 - pow(b1, epoch)), m);
|
std::vector<real_t> m_hat = alg.scalarMultiply(1 / (1 - pow(b1, epoch)), m);
|
||||||
std::vector<double> v_hat = alg.scalarMultiply(1 / (1 - pow(b2, epoch)), v);
|
std::vector<real_t> v_hat = alg.scalarMultiply(1 / (1 - pow(b2, epoch)), v);
|
||||||
|
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate, alg.elementWiseDivision(m_final, alg.scalarAdd(e, alg.sqrt(v_hat)))));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate, alg.elementWiseDivision(m_final, alg.scalarAdd(e, alg.sqrt(v_hat)))));
|
||||||
|
|
||||||
@ -509,8 +509,8 @@ void MLPPLinReg::Nadam(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
void MLPPLinReg::normalEquation() {
|
void MLPPLinReg::normalEquation() {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPStat stat;
|
MLPPStat stat;
|
||||||
std::vector<double> x_means;
|
std::vector<real_t> x_means;
|
||||||
std::vector<std::vector<double>> inputSetT = alg.transpose(inputSet);
|
std::vector<std::vector<real_t>> inputSetT = alg.transpose(inputSet);
|
||||||
|
|
||||||
x_means.resize(inputSetT.size());
|
x_means.resize(inputSetT.size());
|
||||||
for (int i = 0; i < inputSetT.size(); i++) {
|
for (int i = 0; i < inputSetT.size(); i++) {
|
||||||
@ -518,7 +518,7 @@ void MLPPLinReg::normalEquation() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
//try {
|
//try {
|
||||||
std::vector<double> temp;
|
std::vector<real_t> temp;
|
||||||
temp.resize(k);
|
temp.resize(k);
|
||||||
temp = alg.mat_vec_mult(alg.inverse(alg.matmult(alg.transpose(inputSet), inputSet)), alg.mat_vec_mult(alg.transpose(inputSet), outputSet));
|
temp = alg.mat_vec_mult(alg.inverse(alg.matmult(alg.transpose(inputSet), inputSet)), alg.mat_vec_mult(alg.transpose(inputSet), outputSet));
|
||||||
if (std::isnan(temp[0])) {
|
if (std::isnan(temp[0])) {
|
||||||
@ -542,7 +542,7 @@ void MLPPLinReg::normalEquation() {
|
|||||||
//}
|
//}
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPLinReg::score() {
|
real_t MLPPLinReg::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
return util.performance(y_hat, outputSet);
|
return util.performance(y_hat, outputSet);
|
||||||
}
|
}
|
||||||
@ -552,18 +552,18 @@ void MLPPLinReg::save(std::string fileName) {
|
|||||||
util.saveParameters(fileName, weights, bias);
|
util.saveParameters(fileName, weights, bias);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPLinReg::Cost(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPLinReg::Cost(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
return cost.MSE(y_hat, y) + regularization.regTerm(weights, lambda, alpha, reg);
|
return cost.MSE(y_hat, y) + regularization.regTerm(weights, lambda, alpha, reg);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPLinReg::Evaluate(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPLinReg::Evaluate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.scalarAdd(bias, alg.mat_vec_mult(X, weights));
|
return alg.scalarAdd(bias, alg.mat_vec_mult(X, weights));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPLinReg::Evaluate(std::vector<double> x) {
|
real_t MLPPLinReg::Evaluate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.dot(weights, x) + bias;
|
return alg.dot(weights, x) + bias;
|
||||||
}
|
}
|
||||||
|
@ -8,43 +8,45 @@
|
|||||||
// Created by Marc Melikyan on 10/2/20.
|
// Created by Marc Melikyan on 10/2/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
class MLPPLinReg {
|
class MLPPLinReg {
|
||||||
public:
|
public:
|
||||||
MLPPLinReg(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, std::string reg = "None", double lambda = 0.5, double alpha = 0.5);
|
MLPPLinReg(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
||||||
std::vector<double> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<real_t> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
double modelTest(std::vector<double> x);
|
real_t modelTest(std::vector<real_t> x);
|
||||||
void NewtonRaphson(double learning_rate, int max_epoch, bool UI);
|
void NewtonRaphson(real_t learning_rate, int max_epoch, bool UI);
|
||||||
void gradientDescent(double learning_rate, int max_epoch, bool UI = 1);
|
void gradientDescent(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void SGD(double learning_rate, int max_epoch, bool UI = 1);
|
void SGD(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
|
|
||||||
void Momentum(double learning_rate, int max_epoch, int mini_batch_size, double gamma, bool UI = 1);
|
void Momentum(real_t learning_rate, int max_epoch, int mini_batch_size, real_t gamma, bool UI = 1);
|
||||||
void NAG(double learning_rate, int max_epoch, int mini_batch_size, double gamma, bool UI = 1);
|
void NAG(real_t learning_rate, int max_epoch, int mini_batch_size, real_t gamma, bool UI = 1);
|
||||||
void Adagrad(double learning_rate, int max_epoch, int mini_batch_size, double e, bool UI = 1);
|
void Adagrad(real_t learning_rate, int max_epoch, int mini_batch_size, real_t e, bool UI = 1);
|
||||||
void Adadelta(double learning_rate, int max_epoch, int mini_batch_size, double b1, double e, bool UI = 1);
|
void Adadelta(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t e, bool UI = 1);
|
||||||
void Adam(double learning_rate, int max_epoch, int mini_batch_size, double b1, double b2, double e, bool UI = 1);
|
void Adam(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t b2, real_t e, bool UI = 1);
|
||||||
void Adamax(double learning_rate, int max_epoch, int mini_batch_size, double b1, double b2, double e, bool UI = 1);
|
void Adamax(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t b2, real_t e, bool UI = 1);
|
||||||
void Nadam(double learning_rate, int max_epoch, int mini_batch_size, double b1, double b2, double e, bool UI = 1);
|
void Nadam(real_t learning_rate, int max_epoch, int mini_batch_size, real_t b1, real_t b2, real_t e, bool UI = 1);
|
||||||
|
|
||||||
void MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
void MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
||||||
void normalEquation();
|
void normalEquation();
|
||||||
double score();
|
real_t score();
|
||||||
void save(std::string fileName);
|
void save(std::string fileName);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
double Cost(std::vector<double> y_hat, std::vector<double> y);
|
real_t Cost(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
|
|
||||||
std::vector<double> Evaluate(std::vector<std::vector<double>> X);
|
std::vector<real_t> Evaluate(std::vector<std::vector<real_t>> X);
|
||||||
double Evaluate(std::vector<double> x);
|
real_t Evaluate(std::vector<real_t> x);
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
std::vector<double> weights;
|
std::vector<real_t> weights;
|
||||||
double bias;
|
real_t bias;
|
||||||
|
|
||||||
int n;
|
int n;
|
||||||
int k;
|
int k;
|
||||||
|
@ -15,32 +15,32 @@
|
|||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
|
|
||||||
MLPPLogReg::MLPPLogReg(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, std::string reg, double lambda, double alpha) :
|
MLPPLogReg::MLPPLogReg(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, std::string reg, real_t lambda, real_t alpha) :
|
||||||
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
||||||
y_hat.resize(n);
|
y_hat.resize(n);
|
||||||
weights = MLPPUtilities::weightInitialization(k);
|
weights = MLPPUtilities::weightInitialization(k);
|
||||||
bias = MLPPUtilities::biasInitialization();
|
bias = MLPPUtilities::biasInitialization();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPLogReg::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPLogReg::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
return Evaluate(X);
|
return Evaluate(X);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPLogReg::modelTest(std::vector<double> x) {
|
real_t MLPPLogReg::modelTest(std::vector<real_t> x) {
|
||||||
return Evaluate(x);
|
return Evaluate(x);
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPLogReg::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
void MLPPLogReg::gradientDescent(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
cost_prev = Cost(y_hat, outputSet);
|
cost_prev = Cost(y_hat, outputSet);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputSet);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputSet);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputSet), error)));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputSet), error)));
|
||||||
@ -62,17 +62,17 @@ void MLPPLogReg::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPLogReg::MLE(double learning_rate, int max_epoch, bool UI) {
|
void MLPPLogReg::MLE(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
cost_prev = Cost(y_hat, outputSet);
|
cost_prev = Cost(y_hat, outputSet);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(outputSet, y_hat);
|
std::vector<real_t> error = alg.subtraction(outputSet, y_hat);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
weights = alg.addition(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputSet), error)));
|
weights = alg.addition(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputSet), error)));
|
||||||
@ -93,10 +93,10 @@ void MLPPLogReg::MLE(double learning_rate, int max_epoch, bool UI) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPLogReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
void MLPPLogReg::SGD(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
@ -105,10 +105,10 @@ void MLPPLogReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
||||||
int outputIndex = distribution(generator);
|
int outputIndex = distribution(generator);
|
||||||
|
|
||||||
double y_hat = Evaluate(inputSet[outputIndex]);
|
real_t y_hat = Evaluate(inputSet[outputIndex]);
|
||||||
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
||||||
|
|
||||||
double error = y_hat - outputSet[outputIndex];
|
real_t error = y_hat - outputSet[outputIndex];
|
||||||
|
|
||||||
// Weight updation
|
// Weight updation
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate * error, inputSet[outputIndex]));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate * error, inputSet[outputIndex]));
|
||||||
@ -132,10 +132,10 @@ void MLPPLogReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPLogReg::MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
void MLPPLogReg::MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
@ -144,10 +144,10 @@ void MLPPLogReg::MBGD(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error)));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / outputMiniBatches[i].size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), error)));
|
||||||
@ -170,7 +170,7 @@ void MLPPLogReg::MBGD(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPLogReg::score() {
|
real_t MLPPLogReg::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
return util.performance(y_hat, outputSet);
|
return util.performance(y_hat, outputSet);
|
||||||
}
|
}
|
||||||
@ -180,19 +180,19 @@ void MLPPLogReg::save(std::string fileName) {
|
|||||||
util.saveParameters(fileName, weights, bias);
|
util.saveParameters(fileName, weights, bias);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPLogReg::Cost(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPLogReg::Cost(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
return cost.LogLoss(y_hat, y) + regularization.regTerm(weights, lambda, alpha, reg);
|
return cost.LogLoss(y_hat, y) + regularization.regTerm(weights, lambda, alpha, reg);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPLogReg::Evaluate(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPLogReg::Evaluate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
return avn.sigmoid(alg.scalarAdd(bias, alg.mat_vec_mult(X, weights)));
|
return avn.sigmoid(alg.scalarAdd(bias, alg.mat_vec_mult(X, weights)));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPLogReg::Evaluate(std::vector<double> x) {
|
real_t MLPPLogReg::Evaluate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
return avn.sigmoid(alg.dot(weights, x) + bias);
|
return avn.sigmoid(alg.dot(weights, x) + bias);
|
||||||
|
@ -8,6 +8,8 @@
|
|||||||
// Created by Marc Melikyan on 10/2/20.
|
// Created by Marc Melikyan on 10/2/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
@ -15,37 +17,37 @@
|
|||||||
|
|
||||||
class MLPPLogReg {
|
class MLPPLogReg {
|
||||||
public:
|
public:
|
||||||
MLPPLogReg(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, std::string reg = "None", double lambda = 0.5, double alpha = 0.5);
|
MLPPLogReg(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
||||||
std::vector<double> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<real_t> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
double modelTest(std::vector<double> x);
|
real_t modelTest(std::vector<real_t> x);
|
||||||
void gradientDescent(double learning_rate, int max_epoch, bool UI = 1);
|
void gradientDescent(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void MLE(double learning_rate, int max_epoch, bool UI = 1);
|
void MLE(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void SGD(double learning_rate, int max_epoch, bool UI = 1);
|
void SGD(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
void MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
||||||
double score();
|
real_t score();
|
||||||
void save(std::string fileName);
|
void save(std::string fileName);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
double Cost(std::vector<double> y_hat, std::vector<double> y);
|
real_t Cost(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
|
|
||||||
std::vector<double> Evaluate(std::vector<std::vector<double>> X);
|
std::vector<real_t> Evaluate(std::vector<std::vector<real_t>> X);
|
||||||
double Evaluate(std::vector<double> x);
|
real_t Evaluate(std::vector<real_t> x);
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
std::vector<double> weights;
|
std::vector<real_t> weights;
|
||||||
double bias;
|
real_t bias;
|
||||||
|
|
||||||
int n;
|
int n;
|
||||||
int k;
|
int k;
|
||||||
double learning_rate;
|
real_t learning_rate;
|
||||||
|
|
||||||
// Regularization Params
|
// Regularization Params
|
||||||
std::string reg;
|
std::string reg;
|
||||||
double lambda; /* Regularization Parameter */
|
real_t lambda; /* Regularization Parameter */
|
||||||
double alpha; /* This is the controlling param for Elastic Net*/
|
real_t alpha; /* This is the controlling param for Elastic Net*/
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -14,7 +14,7 @@
|
|||||||
#include <iostream>
|
#include <iostream>
|
||||||
|
|
||||||
|
|
||||||
MLPPMANN::MLPPMANN(std::vector<std::vector<double>> inputSet, std::vector<std::vector<double>> outputSet) :
|
MLPPMANN::MLPPMANN(std::vector<std::vector<real_t>> inputSet, std::vector<std::vector<real_t>> outputSet) :
|
||||||
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), n_output(outputSet[0].size()) {
|
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), n_output(outputSet[0].size()) {
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -22,7 +22,7 @@ MLPPMANN::~MLPPMANN() {
|
|||||||
delete outputLayer;
|
delete outputLayer;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPMANN::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<std::vector<real_t>> MLPPMANN::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
network[0].input = X;
|
network[0].input = X;
|
||||||
network[0].forwardPass();
|
network[0].forwardPass();
|
||||||
@ -39,7 +39,7 @@ std::vector<std::vector<double>> MLPPMANN::modelSetTest(std::vector<std::vector<
|
|||||||
return outputLayer->a;
|
return outputLayer->a;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPMANN::modelTest(std::vector<double> x) {
|
std::vector<real_t> MLPPMANN::modelTest(std::vector<real_t> x) {
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
network[0].Test(x);
|
network[0].Test(x);
|
||||||
for (int i = 1; i < network.size(); i++) {
|
for (int i = 1; i < network.size(); i++) {
|
||||||
@ -52,13 +52,13 @@ std::vector<double> MLPPMANN::modelTest(std::vector<double> x) {
|
|||||||
return outputLayer->a_test;
|
return outputLayer->a_test;
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPMANN::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
void MLPPMANN::gradientDescent(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
|
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
@ -73,7 +73,7 @@ void MLPPMANN::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
outputLayer->delta = alg.hadamard_product((cost.*costDeriv)(y_hat, outputSet), (avn.*outputAvn)(outputLayer->z, 1));
|
outputLayer->delta = alg.hadamard_product((cost.*costDeriv)(y_hat, outputSet), (avn.*outputAvn)(outputLayer->z, 1));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> outputWGrad = alg.matmult(alg.transpose(outputLayer->input), outputLayer->delta);
|
std::vector<std::vector<real_t>> outputWGrad = alg.matmult(alg.transpose(outputLayer->input), outputLayer->delta);
|
||||||
|
|
||||||
outputLayer->weights = alg.subtraction(outputLayer->weights, alg.scalarMultiply(learning_rate / n, outputWGrad));
|
outputLayer->weights = alg.subtraction(outputLayer->weights, alg.scalarMultiply(learning_rate / n, outputWGrad));
|
||||||
outputLayer->weights = regularization.regWeights(outputLayer->weights, outputLayer->lambda, outputLayer->alpha, outputLayer->reg);
|
outputLayer->weights = regularization.regWeights(outputLayer->weights, outputLayer->lambda, outputLayer->alpha, outputLayer->reg);
|
||||||
@ -82,7 +82,7 @@ void MLPPMANN::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
auto hiddenLayerAvn = network[network.size() - 1].activation_map[network[network.size() - 1].activation];
|
auto hiddenLayerAvn = network[network.size() - 1].activation_map[network[network.size() - 1].activation];
|
||||||
network[network.size() - 1].delta = alg.hadamard_product(alg.matmult(outputLayer->delta, alg.transpose(outputLayer->weights)), (avn.*hiddenLayerAvn)(network[network.size() - 1].z, 1));
|
network[network.size() - 1].delta = alg.hadamard_product(alg.matmult(outputLayer->delta, alg.transpose(outputLayer->weights)), (avn.*hiddenLayerAvn)(network[network.size() - 1].z, 1));
|
||||||
std::vector<std::vector<double>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[network.size() - 1].input), network[network.size() - 1].delta);
|
std::vector<std::vector<real_t>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[network.size() - 1].input), network[network.size() - 1].delta);
|
||||||
|
|
||||||
network[network.size() - 1].weights = alg.subtraction(network[network.size() - 1].weights, alg.scalarMultiply(learning_rate / n, hiddenLayerWGrad));
|
network[network.size() - 1].weights = alg.subtraction(network[network.size() - 1].weights, alg.scalarMultiply(learning_rate / n, hiddenLayerWGrad));
|
||||||
network[network.size() - 1].weights = regularization.regWeights(network[network.size() - 1].weights, network[network.size() - 1].lambda, network[network.size() - 1].alpha, network[network.size() - 1].reg);
|
network[network.size() - 1].weights = regularization.regWeights(network[network.size() - 1].weights, network[network.size() - 1].lambda, network[network.size() - 1].alpha, network[network.size() - 1].reg);
|
||||||
@ -91,7 +91,7 @@ void MLPPMANN::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
for (int i = network.size() - 2; i >= 0; i--) {
|
for (int i = network.size() - 2; i >= 0; i--) {
|
||||||
auto hiddenLayerAvn = network[i].activation_map[network[i].activation];
|
auto hiddenLayerAvn = network[i].activation_map[network[i].activation];
|
||||||
network[i].delta = alg.hadamard_product(alg.matmult(network[i + 1].delta, network[i + 1].weights), (avn.*hiddenLayerAvn)(network[i].z, 1));
|
network[i].delta = alg.hadamard_product(alg.matmult(network[i + 1].delta, network[i + 1].weights), (avn.*hiddenLayerAvn)(network[i].z, 1));
|
||||||
std::vector<std::vector<double>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[i].input), network[i].delta);
|
std::vector<std::vector<real_t>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[i].input), network[i].delta);
|
||||||
network[i].weights = alg.subtraction(network[i].weights, alg.scalarMultiply(learning_rate / n, hiddenLayerWGrad));
|
network[i].weights = alg.subtraction(network[i].weights, alg.scalarMultiply(learning_rate / n, hiddenLayerWGrad));
|
||||||
network[i].weights = regularization.regWeights(network[i].weights, network[i].lambda, network[i].alpha, network[i].reg);
|
network[i].weights = regularization.regWeights(network[i].weights, network[i].lambda, network[i].alpha, network[i].reg);
|
||||||
network[i].bias = alg.subtractMatrixRows(network[i].bias, alg.scalarMultiply(learning_rate / n, network[i].delta));
|
network[i].bias = alg.subtractMatrixRows(network[i].bias, alg.scalarMultiply(learning_rate / n, network[i].delta));
|
||||||
@ -120,7 +120,7 @@ void MLPPMANN::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPMANN::score() {
|
real_t MLPPMANN::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
return util.performance(y_hat, outputSet);
|
return util.performance(y_hat, outputSet);
|
||||||
@ -139,7 +139,7 @@ void MLPPMANN::save(std::string fileName) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPMANN::addLayer(int n_hidden, std::string activation, std::string weightInit, std::string reg, double lambda, double alpha) {
|
void MLPPMANN::addLayer(int n_hidden, std::string activation, std::string weightInit, std::string reg, real_t lambda, real_t alpha) {
|
||||||
if (network.empty()) {
|
if (network.empty()) {
|
||||||
network.push_back(MLPPHiddenLayer(n_hidden, activation, inputSet, weightInit, reg, lambda, alpha));
|
network.push_back(MLPPHiddenLayer(n_hidden, activation, inputSet, weightInit, reg, lambda, alpha));
|
||||||
network[0].forwardPass();
|
network[0].forwardPass();
|
||||||
@ -149,7 +149,7 @@ void MLPPMANN::addLayer(int n_hidden, std::string activation, std::string weight
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPMANN::addOutputLayer(std::string activation, std::string loss, std::string weightInit, std::string reg, double lambda, double alpha) {
|
void MLPPMANN::addOutputLayer(std::string activation, std::string loss, std::string weightInit, std::string reg, real_t lambda, real_t alpha) {
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
outputLayer = new MLPPMultiOutputLayer(n_output, network[0].n_hidden, activation, loss, network[network.size() - 1].a, weightInit, reg, lambda, alpha);
|
outputLayer = new MLPPMultiOutputLayer(n_output, network[0].n_hidden, activation, loss, network[network.size() - 1].a, weightInit, reg, lambda, alpha);
|
||||||
} else {
|
} else {
|
||||||
@ -157,10 +157,10 @@ void MLPPMANN::addOutputLayer(std::string activation, std::string loss, std::str
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPMANN::Cost(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
real_t MLPPMANN::Cost(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
double totalRegTerm = 0;
|
real_t totalRegTerm = 0;
|
||||||
|
|
||||||
auto cost_function = outputLayer->cost_map[outputLayer->cost];
|
auto cost_function = outputLayer->cost_map[outputLayer->cost];
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
|
@ -8,6 +8,8 @@
|
|||||||
// Created by Marc Melikyan on 11/4/20.
|
// Created by Marc Melikyan on 11/4/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include "../hidden_layer/hidden_layer.h"
|
#include "../hidden_layer/hidden_layer.h"
|
||||||
#include "../multi_output_layer/multi_output_layer.h"
|
#include "../multi_output_layer/multi_output_layer.h"
|
||||||
|
|
||||||
@ -18,24 +20,24 @@
|
|||||||
|
|
||||||
class MLPPMANN {
|
class MLPPMANN {
|
||||||
public:
|
public:
|
||||||
MLPPMANN(std::vector<std::vector<double>> inputSet, std::vector<std::vector<double>> outputSet);
|
MLPPMANN(std::vector<std::vector<real_t>> inputSet, std::vector<std::vector<real_t>> outputSet);
|
||||||
~MLPPMANN();
|
~MLPPMANN();
|
||||||
std::vector<std::vector<double>> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<std::vector<real_t>> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
std::vector<double> modelTest(std::vector<double> x);
|
std::vector<real_t> modelTest(std::vector<real_t> x);
|
||||||
void gradientDescent(double learning_rate, int max_epoch, bool UI = 1);
|
void gradientDescent(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
double score();
|
real_t score();
|
||||||
void save(std::string fileName);
|
void save(std::string fileName);
|
||||||
|
|
||||||
void addLayer(int n_hidden, std::string activation, std::string weightInit = "Default", std::string reg = "None", double lambda = 0.5, double alpha = 0.5);
|
void addLayer(int n_hidden, std::string activation, std::string weightInit = "Default", std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
||||||
void addOutputLayer(std::string activation, std::string loss, std::string weightInit = "Default", std::string reg = "None", double lambda = 0.5, double alpha = 0.5);
|
void addOutputLayer(std::string activation, std::string loss, std::string weightInit = "Default", std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
double Cost(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
real_t Cost(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<std::vector<double>> outputSet;
|
std::vector<std::vector<real_t>> outputSet;
|
||||||
std::vector<std::vector<double>> y_hat;
|
std::vector<std::vector<real_t>> y_hat;
|
||||||
|
|
||||||
std::vector<MLPPHiddenLayer> network;
|
std::vector<MLPPHiddenLayer> network;
|
||||||
MLPPMultiOutputLayer *outputLayer;
|
MLPPMultiOutputLayer *outputLayer;
|
||||||
|
@ -16,7 +16,7 @@
|
|||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
|
|
||||||
MLPPMLP::MLPPMLP(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, int n_hidden, std::string reg, double lambda, double alpha) :
|
MLPPMLP::MLPPMLP(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, int n_hidden, std::string reg, real_t lambda, real_t alpha) :
|
||||||
inputSet(inputSet), outputSet(outputSet), n_hidden(n_hidden), n(inputSet.size()), k(inputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
inputSet(inputSet), outputSet(outputSet), n_hidden(n_hidden), n(inputSet.size()), k(inputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
y_hat.resize(n);
|
y_hat.resize(n);
|
||||||
@ -27,19 +27,19 @@ MLPPMLP::MLPPMLP(std::vector<std::vector<double>> inputSet, std::vector<double>
|
|||||||
bias2 = MLPPUtilities::biasInitialization();
|
bias2 = MLPPUtilities::biasInitialization();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPMLP::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPMLP::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
return Evaluate(X);
|
return Evaluate(X);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPMLP::modelTest(std::vector<double> x) {
|
real_t MLPPMLP::modelTest(std::vector<real_t> x) {
|
||||||
return Evaluate(x);
|
return Evaluate(x);
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPMLP::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
void MLPPMLP::gradientDescent(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
@ -47,11 +47,11 @@ void MLPPMLP::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
cost_prev = Cost(y_hat, outputSet);
|
cost_prev = Cost(y_hat, outputSet);
|
||||||
|
|
||||||
// Calculating the errors
|
// Calculating the errors
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputSet);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputSet);
|
||||||
|
|
||||||
// Calculating the weight/bias gradients for layer 2
|
// Calculating the weight/bias gradients for layer 2
|
||||||
|
|
||||||
std::vector<double> D2_1 = alg.mat_vec_mult(alg.transpose(a2), error);
|
std::vector<real_t> D2_1 = alg.mat_vec_mult(alg.transpose(a2), error);
|
||||||
|
|
||||||
// weights and bias updation for layer 2
|
// weights and bias updation for layer 2
|
||||||
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate / n, D2_1));
|
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate / n, D2_1));
|
||||||
@ -61,14 +61,14 @@ void MLPPMLP::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
|
|
||||||
// Calculating the weight/bias for layer 1
|
// Calculating the weight/bias for layer 1
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_1;
|
std::vector<std::vector<real_t>> D1_1;
|
||||||
D1_1.resize(n);
|
D1_1.resize(n);
|
||||||
|
|
||||||
D1_1 = alg.outerProduct(error, weights2);
|
D1_1 = alg.outerProduct(error, weights2);
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
std::vector<std::vector<real_t>> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_3 = alg.matmult(alg.transpose(inputSet), D1_2);
|
std::vector<std::vector<real_t>> D1_3 = alg.matmult(alg.transpose(inputSet), D1_2);
|
||||||
|
|
||||||
// weight an bias updation for layer 1
|
// weight an bias updation for layer 1
|
||||||
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate / n, D1_3));
|
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate / n, D1_3));
|
||||||
@ -94,11 +94,11 @@ void MLPPMLP::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPMLP::SGD(double learning_rate, int max_epoch, bool UI) {
|
void MLPPMLP::SGD(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
@ -107,13 +107,13 @@ void MLPPMLP::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
||||||
int outputIndex = distribution(generator);
|
int outputIndex = distribution(generator);
|
||||||
|
|
||||||
double y_hat = Evaluate(inputSet[outputIndex]);
|
real_t y_hat = Evaluate(inputSet[outputIndex]);
|
||||||
auto [z2, a2] = propagate(inputSet[outputIndex]);
|
auto [z2, a2] = propagate(inputSet[outputIndex]);
|
||||||
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
||||||
double error = y_hat - outputSet[outputIndex];
|
real_t error = y_hat - outputSet[outputIndex];
|
||||||
|
|
||||||
// Weight updation for layer 2
|
// Weight updation for layer 2
|
||||||
std::vector<double> D2_1 = alg.scalarMultiply(error, a2);
|
std::vector<real_t> D2_1 = alg.scalarMultiply(error, a2);
|
||||||
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate, D2_1));
|
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate, D2_1));
|
||||||
weights2 = regularization.regWeights(weights2, lambda, alpha, reg);
|
weights2 = regularization.regWeights(weights2, lambda, alpha, reg);
|
||||||
|
|
||||||
@ -121,9 +121,9 @@ void MLPPMLP::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
bias2 -= learning_rate * error;
|
bias2 -= learning_rate * error;
|
||||||
|
|
||||||
// Weight updation for layer 1
|
// Weight updation for layer 1
|
||||||
std::vector<double> D1_1 = alg.scalarMultiply(error, weights2);
|
std::vector<real_t> D1_1 = alg.scalarMultiply(error, weights2);
|
||||||
std::vector<double> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
std::vector<real_t> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
||||||
std::vector<std::vector<double>> D1_3 = alg.outerProduct(inputSet[outputIndex], D1_2);
|
std::vector<std::vector<real_t>> D1_3 = alg.outerProduct(inputSet[outputIndex], D1_2);
|
||||||
|
|
||||||
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate, D1_3));
|
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate, D1_3));
|
||||||
weights1 = regularization.regWeights(weights1, lambda, alpha, reg);
|
weights1 = regularization.regWeights(weights1, lambda, alpha, reg);
|
||||||
@ -148,11 +148,11 @@ void MLPPMLP::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPMLP::MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
void MLPPMLP::MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
@ -161,34 +161,34 @@ void MLPPMLP::MBGD(double learning_rate, int max_epoch, int mini_batch_size, boo
|
|||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
auto [z2, a2] = propagate(inputMiniBatches[i]);
|
auto [z2, a2] = propagate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the errors
|
// Calculating the errors
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the weight/bias gradients for layer 2
|
// Calculating the weight/bias gradients for layer 2
|
||||||
|
|
||||||
std::vector<double> D2_1 = alg.mat_vec_mult(alg.transpose(a2), error);
|
std::vector<real_t> D2_1 = alg.mat_vec_mult(alg.transpose(a2), error);
|
||||||
|
|
||||||
// weights and bias updation for layser 2
|
// weights and bias updation for layser 2
|
||||||
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate / outputMiniBatches[i].size(), D2_1));
|
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate / outputMiniBatches[i].size(), D2_1));
|
||||||
weights2 = regularization.regWeights(weights2, lambda, alpha, reg);
|
weights2 = regularization.regWeights(weights2, lambda, alpha, reg);
|
||||||
|
|
||||||
// Calculating the bias gradients for layer 2
|
// Calculating the bias gradients for layer 2
|
||||||
double b_gradient = alg.sum_elements(error);
|
real_t b_gradient = alg.sum_elements(error);
|
||||||
|
|
||||||
// Bias Updation for layer 2
|
// Bias Updation for layer 2
|
||||||
bias2 -= learning_rate * alg.sum_elements(error) / outputMiniBatches[i].size();
|
bias2 -= learning_rate * alg.sum_elements(error) / outputMiniBatches[i].size();
|
||||||
|
|
||||||
//Calculating the weight/bias for layer 1
|
//Calculating the weight/bias for layer 1
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_1 = alg.outerProduct(error, weights2);
|
std::vector<std::vector<real_t>> D1_1 = alg.outerProduct(error, weights2);
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
std::vector<std::vector<real_t>> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_3 = alg.matmult(alg.transpose(inputMiniBatches[i]), D1_2);
|
std::vector<std::vector<real_t>> D1_3 = alg.matmult(alg.transpose(inputMiniBatches[i]), D1_2);
|
||||||
|
|
||||||
// weight an bias updation for layer 1
|
// weight an bias updation for layer 1
|
||||||
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate / outputMiniBatches[i].size(), D1_3));
|
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate / outputMiniBatches[i].size(), D1_3));
|
||||||
@ -214,7 +214,7 @@ void MLPPMLP::MBGD(double learning_rate, int max_epoch, int mini_batch_size, boo
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPMLP::score() {
|
real_t MLPPMLP::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
return util.performance(y_hat, outputSet);
|
return util.performance(y_hat, outputSet);
|
||||||
}
|
}
|
||||||
@ -225,41 +225,41 @@ void MLPPMLP::save(std::string fileName) {
|
|||||||
util.saveParameters(fileName, weights2, bias2, 1, 2);
|
util.saveParameters(fileName, weights2, bias2, 1, 2);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPMLP::Cost(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPMLP::Cost(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
return cost.LogLoss(y_hat, y) + regularization.regTerm(weights2, lambda, alpha, reg) + regularization.regTerm(weights1, lambda, alpha, reg);
|
return cost.LogLoss(y_hat, y) + regularization.regTerm(weights2, lambda, alpha, reg) + regularization.regTerm(weights1, lambda, alpha, reg);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPMLP::Evaluate(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPMLP::Evaluate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
std::vector<std::vector<double>> z2 = alg.mat_vec_add(alg.matmult(X, weights1), bias1);
|
std::vector<std::vector<real_t>> z2 = alg.mat_vec_add(alg.matmult(X, weights1), bias1);
|
||||||
std::vector<std::vector<double>> a2 = avn.sigmoid(z2);
|
std::vector<std::vector<real_t>> a2 = avn.sigmoid(z2);
|
||||||
return avn.sigmoid(alg.scalarAdd(bias2, alg.mat_vec_mult(a2, weights2)));
|
return avn.sigmoid(alg.scalarAdd(bias2, alg.mat_vec_mult(a2, weights2)));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>> MLPPMLP::propagate(std::vector<std::vector<double>> X) {
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> MLPPMLP::propagate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
std::vector<std::vector<double>> z2 = alg.mat_vec_add(alg.matmult(X, weights1), bias1);
|
std::vector<std::vector<real_t>> z2 = alg.mat_vec_add(alg.matmult(X, weights1), bias1);
|
||||||
std::vector<std::vector<double>> a2 = avn.sigmoid(z2);
|
std::vector<std::vector<real_t>> a2 = avn.sigmoid(z2);
|
||||||
return { z2, a2 };
|
return { z2, a2 };
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPMLP::Evaluate(std::vector<double> x) {
|
real_t MLPPMLP::Evaluate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
std::vector<double> z2 = alg.addition(alg.mat_vec_mult(alg.transpose(weights1), x), bias1);
|
std::vector<real_t> z2 = alg.addition(alg.mat_vec_mult(alg.transpose(weights1), x), bias1);
|
||||||
std::vector<double> a2 = avn.sigmoid(z2);
|
std::vector<real_t> a2 = avn.sigmoid(z2);
|
||||||
return avn.sigmoid(alg.dot(weights2, a2) + bias2);
|
return avn.sigmoid(alg.dot(weights2, a2) + bias2);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<double>, std::vector<double>> MLPPMLP::propagate(std::vector<double> x) {
|
std::tuple<std::vector<real_t>, std::vector<real_t>> MLPPMLP::propagate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
std::vector<double> z2 = alg.addition(alg.mat_vec_mult(alg.transpose(weights1), x), bias1);
|
std::vector<real_t> z2 = alg.addition(alg.mat_vec_mult(alg.transpose(weights1), x), bias1);
|
||||||
std::vector<double> a2 = avn.sigmoid(z2);
|
std::vector<real_t> a2 = avn.sigmoid(z2);
|
||||||
return { z2, a2 };
|
return { z2, a2 };
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -8,6 +8,8 @@
|
|||||||
// Created by Marc Melikyan on 11/4/20.
|
// Created by Marc Melikyan on 11/4/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <map>
|
#include <map>
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
@ -16,36 +18,36 @@
|
|||||||
|
|
||||||
class MLPPMLP {
|
class MLPPMLP {
|
||||||
public:
|
public:
|
||||||
MLPPMLP(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, int n_hidden, std::string reg = "None", double lambda = 0.5, double alpha = 0.5);
|
MLPPMLP(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, int n_hidden, std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
||||||
std::vector<double> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<real_t> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
double modelTest(std::vector<double> x);
|
real_t modelTest(std::vector<real_t> x);
|
||||||
void gradientDescent(double learning_rate, int max_epoch, bool UI = 1);
|
void gradientDescent(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void SGD(double learning_rate, int max_epoch, bool UI = 1);
|
void SGD(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
void MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
||||||
double score();
|
real_t score();
|
||||||
void save(std::string fileName);
|
void save(std::string fileName);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
double Cost(std::vector<double> y_hat, std::vector<double> y);
|
real_t Cost(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
|
|
||||||
std::vector<double> Evaluate(std::vector<std::vector<double>> X);
|
std::vector<real_t> Evaluate(std::vector<std::vector<real_t>> X);
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>> propagate(std::vector<std::vector<double>> X);
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> propagate(std::vector<std::vector<real_t>> X);
|
||||||
double Evaluate(std::vector<double> x);
|
real_t Evaluate(std::vector<real_t> x);
|
||||||
std::tuple<std::vector<double>, std::vector<double>> propagate(std::vector<double> x);
|
std::tuple<std::vector<real_t>, std::vector<real_t>> propagate(std::vector<real_t> x);
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
|
|
||||||
std::vector<std::vector<double>> weights1;
|
std::vector<std::vector<real_t>> weights1;
|
||||||
std::vector<double> weights2;
|
std::vector<real_t> weights2;
|
||||||
|
|
||||||
std::vector<double> bias1;
|
std::vector<real_t> bias1;
|
||||||
double bias2;
|
real_t bias2;
|
||||||
|
|
||||||
std::vector<std::vector<double>> z2;
|
std::vector<std::vector<real_t>> z2;
|
||||||
std::vector<std::vector<double>> a2;
|
std::vector<std::vector<real_t>> a2;
|
||||||
|
|
||||||
int n;
|
int n;
|
||||||
int k;
|
int k;
|
||||||
@ -53,8 +55,8 @@ private:
|
|||||||
|
|
||||||
// Regularization Params
|
// Regularization Params
|
||||||
std::string reg;
|
std::string reg;
|
||||||
double lambda; /* Regularization Parameter */
|
real_t lambda; /* Regularization Parameter */
|
||||||
double alpha; /* This is the controlling param for Elastic Net*/
|
real_t alpha; /* This is the controlling param for Elastic Net*/
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -12,7 +12,7 @@
|
|||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
|
|
||||||
MLPPMultiOutputLayer::MLPPMultiOutputLayer(int n_output, int n_hidden, std::string activation, std::string cost, std::vector<std::vector<double>> input, std::string weightInit, std::string reg, double lambda, double alpha) :
|
MLPPMultiOutputLayer::MLPPMultiOutputLayer(int n_output, int n_hidden, std::string activation, std::string cost, std::vector<std::vector<real_t>> input, std::string weightInit, std::string reg, real_t lambda, real_t alpha) :
|
||||||
n_output(n_output), n_hidden(n_hidden), activation(activation), cost(cost), input(input), weightInit(weightInit), reg(reg), lambda(lambda), alpha(alpha) {
|
n_output(n_output), n_hidden(n_hidden), activation(activation), cost(cost), input(input), weightInit(weightInit), reg(reg), lambda(lambda), alpha(alpha) {
|
||||||
weights = MLPPUtilities::weightInitialization(n_hidden, n_output, weightInit);
|
weights = MLPPUtilities::weightInitialization(n_hidden, n_output, weightInit);
|
||||||
bias = MLPPUtilities::biasInitialization(n_output);
|
bias = MLPPUtilities::biasInitialization(n_output);
|
||||||
@ -123,7 +123,7 @@ void MLPPMultiOutputLayer::forwardPass() {
|
|||||||
a = (avn.*activation_map[activation])(z, 0);
|
a = (avn.*activation_map[activation])(z, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPMultiOutputLayer::Test(std::vector<double> x) {
|
void MLPPMultiOutputLayer::Test(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
z_test = alg.addition(alg.mat_vec_mult(alg.transpose(weights), x), bias);
|
z_test = alg.addition(alg.mat_vec_mult(alg.transpose(weights), x), bias);
|
||||||
|
@ -8,6 +8,8 @@
|
|||||||
// Created by Marc Melikyan on 11/4/20.
|
// Created by Marc Melikyan on 11/4/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include "../activation/activation.h"
|
#include "../activation/activation.h"
|
||||||
#include "../cost/cost.h"
|
#include "../cost/cost.h"
|
||||||
|
|
||||||
@ -18,40 +20,40 @@
|
|||||||
|
|
||||||
class MLPPMultiOutputLayer {
|
class MLPPMultiOutputLayer {
|
||||||
public:
|
public:
|
||||||
MLPPMultiOutputLayer(int n_output, int n_hidden, std::string activation, std::string cost, std::vector<std::vector<double>> input, std::string weightInit, std::string reg, double lambda, double alpha);
|
MLPPMultiOutputLayer(int n_output, int n_hidden, std::string activation, std::string cost, std::vector<std::vector<real_t>> input, std::string weightInit, std::string reg, real_t lambda, real_t alpha);
|
||||||
|
|
||||||
int n_output;
|
int n_output;
|
||||||
int n_hidden;
|
int n_hidden;
|
||||||
std::string activation;
|
std::string activation;
|
||||||
std::string cost;
|
std::string cost;
|
||||||
|
|
||||||
std::vector<std::vector<double>> input;
|
std::vector<std::vector<real_t>> input;
|
||||||
|
|
||||||
std::vector<std::vector<double>> weights;
|
std::vector<std::vector<real_t>> weights;
|
||||||
std::vector<double> bias;
|
std::vector<real_t> bias;
|
||||||
|
|
||||||
std::vector<std::vector<double>> z;
|
std::vector<std::vector<real_t>> z;
|
||||||
std::vector<std::vector<double>> a;
|
std::vector<std::vector<real_t>> a;
|
||||||
|
|
||||||
std::map<std::string, std::vector<std::vector<double>> (MLPPActivation::*)(std::vector<std::vector<double>>, bool)> activation_map;
|
std::map<std::string, std::vector<std::vector<real_t>> (MLPPActivation::*)(std::vector<std::vector<real_t>>, bool)> activation_map;
|
||||||
std::map<std::string, std::vector<double> (MLPPActivation::*)(std::vector<double>, bool)> activationTest_map;
|
std::map<std::string, std::vector<real_t> (MLPPActivation::*)(std::vector<real_t>, bool)> activationTest_map;
|
||||||
std::map<std::string, double (MLPPCost::*)(std::vector<std::vector<double>>, std::vector<std::vector<double>>)> cost_map;
|
std::map<std::string, real_t (MLPPCost::*)(std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>)> cost_map;
|
||||||
std::map<std::string, std::vector<std::vector<double>> (MLPPCost::*)(std::vector<std::vector<double>>, std::vector<std::vector<double>>)> costDeriv_map;
|
std::map<std::string, std::vector<std::vector<real_t>> (MLPPCost::*)(std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>)> costDeriv_map;
|
||||||
|
|
||||||
std::vector<double> z_test;
|
std::vector<real_t> z_test;
|
||||||
std::vector<double> a_test;
|
std::vector<real_t> a_test;
|
||||||
|
|
||||||
std::vector<std::vector<double>> delta;
|
std::vector<std::vector<real_t>> delta;
|
||||||
|
|
||||||
// Regularization Params
|
// Regularization Params
|
||||||
std::string reg;
|
std::string reg;
|
||||||
double lambda; /* Regularization Parameter */
|
real_t lambda; /* Regularization Parameter */
|
||||||
double alpha; /* This is the controlling param for Elastic Net*/
|
real_t alpha; /* This is the controlling param for Elastic Net*/
|
||||||
|
|
||||||
std::string weightInit;
|
std::string weightInit;
|
||||||
|
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
void Test(std::vector<double> x);
|
void Test(std::vector<real_t> x);
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -13,22 +13,22 @@
|
|||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
|
|
||||||
MLPPMultinomialNB::MLPPMultinomialNB(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, int class_num) :
|
MLPPMultinomialNB::MLPPMultinomialNB(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, int class_num) :
|
||||||
inputSet(inputSet), outputSet(outputSet), class_num(class_num) {
|
inputSet(inputSet), outputSet(outputSet), class_num(class_num) {
|
||||||
y_hat.resize(outputSet.size());
|
y_hat.resize(outputSet.size());
|
||||||
Evaluate();
|
Evaluate();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPMultinomialNB::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPMultinomialNB::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
for (int i = 0; i < X.size(); i++) {
|
for (int i = 0; i < X.size(); i++) {
|
||||||
y_hat.push_back(modelTest(X[i]));
|
y_hat.push_back(modelTest(X[i]));
|
||||||
}
|
}
|
||||||
return y_hat;
|
return y_hat;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPMultinomialNB::modelTest(std::vector<double> x) {
|
real_t MLPPMultinomialNB::modelTest(std::vector<real_t> x) {
|
||||||
double score[class_num];
|
real_t score[class_num];
|
||||||
computeTheta();
|
computeTheta();
|
||||||
|
|
||||||
for (int j = 0; j < x.size(); j++) {
|
for (int j = 0; j < x.size(); j++) {
|
||||||
@ -45,10 +45,10 @@ double MLPPMultinomialNB::modelTest(std::vector<double> x) {
|
|||||||
score[i] += std::log(priors[i]);
|
score[i] += std::log(priors[i]);
|
||||||
}
|
}
|
||||||
|
|
||||||
return std::distance(score, std::max_element(score, score + sizeof(score) / sizeof(double)));
|
return std::distance(score, std::max_element(score, score + sizeof(score) / sizeof(real_t)));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPMultinomialNB::score() {
|
real_t MLPPMultinomialNB::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
return util.performance(y_hat, outputSet);
|
return util.performance(y_hat, outputSet);
|
||||||
}
|
}
|
||||||
@ -81,14 +81,14 @@ void MLPPMultinomialNB::Evaluate() {
|
|||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
for (int i = 0; i < outputSet.size(); i++) {
|
for (int i = 0; i < outputSet.size(); i++) {
|
||||||
// Pr(B | A) * Pr(A)
|
// Pr(B | A) * Pr(A)
|
||||||
double score[class_num];
|
real_t score[class_num];
|
||||||
|
|
||||||
// Easy computation of priors, i.e. Pr(C_k)
|
// Easy computation of priors, i.e. Pr(C_k)
|
||||||
priors.resize(class_num);
|
priors.resize(class_num);
|
||||||
for (int i = 0; i < outputSet.size(); i++) {
|
for (int i = 0; i < outputSet.size(); i++) {
|
||||||
priors[int(outputSet[i])]++;
|
priors[int(outputSet[i])]++;
|
||||||
}
|
}
|
||||||
priors = alg.scalarMultiply(double(1) / double(outputSet.size()), priors);
|
priors = alg.scalarMultiply(real_t(1) / real_t(outputSet.size()), priors);
|
||||||
|
|
||||||
// Evaluating Theta...
|
// Evaluating Theta...
|
||||||
computeTheta();
|
computeTheta();
|
||||||
@ -113,6 +113,6 @@ void MLPPMultinomialNB::Evaluate() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Assigning the traning example's y_hat to a class
|
// Assigning the traning example's y_hat to a class
|
||||||
y_hat[i] = std::distance(score, std::max_element(score, score + sizeof(score) / sizeof(double)));
|
y_hat[i] = std::distance(score, std::max_element(score, score + sizeof(score) / sizeof(real_t)));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -8,32 +8,34 @@
|
|||||||
// Created by Marc Melikyan on 1/17/21.
|
// Created by Marc Melikyan on 1/17/21.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <map>
|
#include <map>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
|
|
||||||
class MLPPMultinomialNB {
|
class MLPPMultinomialNB {
|
||||||
public:
|
public:
|
||||||
MLPPMultinomialNB(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, int class_num);
|
MLPPMultinomialNB(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, int class_num);
|
||||||
std::vector<double> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<real_t> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
double modelTest(std::vector<double> x);
|
real_t modelTest(std::vector<real_t> x);
|
||||||
double score();
|
real_t score();
|
||||||
|
|
||||||
private:
|
private:
|
||||||
void computeTheta();
|
void computeTheta();
|
||||||
void Evaluate();
|
void Evaluate();
|
||||||
|
|
||||||
// Model Params
|
// Model Params
|
||||||
std::vector<double> priors;
|
std::vector<real_t> priors;
|
||||||
|
|
||||||
std::vector<std::map<double, int>> theta;
|
std::vector<std::map<real_t, int>> theta;
|
||||||
std::vector<double> vocab;
|
std::vector<real_t> vocab;
|
||||||
int class_num;
|
int class_num;
|
||||||
|
|
||||||
// Datasets
|
// Datasets
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
};
|
};
|
||||||
|
|
||||||
#endif /* MultinomialNB_hpp */
|
#endif /* MultinomialNB_hpp */
|
||||||
|
@ -14,127 +14,127 @@
|
|||||||
|
|
||||||
|
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::numDiff(double (*function)(double), double x) {
|
real_t MLPPNumericalAnalysis::numDiff(real_t (*function)(real_t), real_t x) {
|
||||||
double eps = 1e-10;
|
real_t eps = 1e-10;
|
||||||
return (function(x + eps) - function(x)) / eps; // This is just the formal def. of the derivative.
|
return (function(x + eps) - function(x)) / eps; // This is just the formal def. of the derivative.
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::numDiff_2(double (*function)(double), double x) {
|
real_t MLPPNumericalAnalysis::numDiff_2(real_t (*function)(real_t), real_t x) {
|
||||||
double eps = 1e-5;
|
real_t eps = 1e-5;
|
||||||
return (function(x + 2 * eps) - 2 * function(x + eps) + function(x)) / (eps * eps);
|
return (function(x + 2 * eps) - 2 * function(x + eps) + function(x)) / (eps * eps);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::numDiff_3(double (*function)(double), double x) {
|
real_t MLPPNumericalAnalysis::numDiff_3(real_t (*function)(real_t), real_t x) {
|
||||||
double eps = 1e-5;
|
real_t eps = 1e-5;
|
||||||
double t1 = function(x + 3 * eps) - 2 * function(x + 2 * eps) + function(x + eps);
|
real_t t1 = function(x + 3 * eps) - 2 * function(x + 2 * eps) + function(x + eps);
|
||||||
double t2 = function(x + 2 * eps) - 2 * function(x + eps) + function(x);
|
real_t t2 = function(x + 2 * eps) - 2 * function(x + eps) + function(x);
|
||||||
return (t1 - t2) / (eps * eps * eps);
|
return (t1 - t2) / (eps * eps * eps);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::constantApproximation(double (*function)(double), double c) {
|
real_t MLPPNumericalAnalysis::constantApproximation(real_t (*function)(real_t), real_t c) {
|
||||||
return function(c);
|
return function(c);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::linearApproximation(double (*function)(double), double c, double x) {
|
real_t MLPPNumericalAnalysis::linearApproximation(real_t (*function)(real_t), real_t c, real_t x) {
|
||||||
return constantApproximation(function, c) + numDiff(function, c) * (x - c);
|
return constantApproximation(function, c) + numDiff(function, c) * (x - c);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::quadraticApproximation(double (*function)(double), double c, double x) {
|
real_t MLPPNumericalAnalysis::quadraticApproximation(real_t (*function)(real_t), real_t c, real_t x) {
|
||||||
return linearApproximation(function, c, x) + 0.5 * numDiff_2(function, c) * (x - c) * (x - c);
|
return linearApproximation(function, c, x) + 0.5 * numDiff_2(function, c) * (x - c) * (x - c);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::cubicApproximation(double (*function)(double), double c, double x) {
|
real_t MLPPNumericalAnalysis::cubicApproximation(real_t (*function)(real_t), real_t c, real_t x) {
|
||||||
return quadraticApproximation(function, c, x) + (1 / 6) * numDiff_3(function, c) * (x - c) * (x - c) * (x - c);
|
return quadraticApproximation(function, c, x) + (1 / 6) * numDiff_3(function, c) * (x - c) * (x - c) * (x - c);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::numDiff(double (*function)(std::vector<double>), std::vector<double> x, int axis) {
|
real_t MLPPNumericalAnalysis::numDiff(real_t (*function)(std::vector<real_t>), std::vector<real_t> x, int axis) {
|
||||||
// For multivariable function analysis.
|
// For multivariable function analysis.
|
||||||
// This will be used for calculating Jacobian vectors.
|
// This will be used for calculating Jacobian vectors.
|
||||||
// Diffrentiate with respect to indicated axis. (0, 1, 2 ...)
|
// Diffrentiate with respect to indicated axis. (0, 1, 2 ...)
|
||||||
double eps = 1e-10;
|
real_t eps = 1e-10;
|
||||||
std::vector<double> x_eps = x;
|
std::vector<real_t> x_eps = x;
|
||||||
x_eps[axis] += eps;
|
x_eps[axis] += eps;
|
||||||
|
|
||||||
return (function(x_eps) - function(x)) / eps;
|
return (function(x_eps) - function(x)) / eps;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::numDiff_2(double (*function)(std::vector<double>), std::vector<double> x, int axis1, int axis2) {
|
real_t MLPPNumericalAnalysis::numDiff_2(real_t (*function)(std::vector<real_t>), std::vector<real_t> x, int axis1, int axis2) {
|
||||||
//For Hessians.
|
//For Hessians.
|
||||||
double eps = 1e-5;
|
real_t eps = 1e-5;
|
||||||
|
|
||||||
std::vector<double> x_pp = x;
|
std::vector<real_t> x_pp = x;
|
||||||
x_pp[axis1] += eps;
|
x_pp[axis1] += eps;
|
||||||
x_pp[axis2] += eps;
|
x_pp[axis2] += eps;
|
||||||
|
|
||||||
std::vector<double> x_np = x;
|
std::vector<real_t> x_np = x;
|
||||||
x_np[axis2] += eps;
|
x_np[axis2] += eps;
|
||||||
|
|
||||||
std::vector<double> x_pn = x;
|
std::vector<real_t> x_pn = x;
|
||||||
x_pn[axis1] += eps;
|
x_pn[axis1] += eps;
|
||||||
|
|
||||||
return (function(x_pp) - function(x_np) - function(x_pn) + function(x)) / (eps * eps);
|
return (function(x_pp) - function(x_np) - function(x_pn) + function(x)) / (eps * eps);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::numDiff_3(double (*function)(std::vector<double>), std::vector<double> x, int axis1, int axis2, int axis3) {
|
real_t MLPPNumericalAnalysis::numDiff_3(real_t (*function)(std::vector<real_t>), std::vector<real_t> x, int axis1, int axis2, int axis3) {
|
||||||
// For third order derivative tensors.
|
// For third order derivative tensors.
|
||||||
// NOTE: Approximations do not appear to be accurate for sinusodial functions...
|
// NOTE: Approximations do not appear to be accurate for sinusodial functions...
|
||||||
// Should revisit this later.
|
// Should revisit this later.
|
||||||
double eps = INT_MAX;
|
real_t eps = INT_MAX;
|
||||||
|
|
||||||
std::vector<double> x_ppp = x;
|
std::vector<real_t> x_ppp = x;
|
||||||
x_ppp[axis1] += eps;
|
x_ppp[axis1] += eps;
|
||||||
x_ppp[axis2] += eps;
|
x_ppp[axis2] += eps;
|
||||||
x_ppp[axis3] += eps;
|
x_ppp[axis3] += eps;
|
||||||
|
|
||||||
std::vector<double> x_npp = x;
|
std::vector<real_t> x_npp = x;
|
||||||
x_npp[axis2] += eps;
|
x_npp[axis2] += eps;
|
||||||
x_npp[axis3] += eps;
|
x_npp[axis3] += eps;
|
||||||
|
|
||||||
std::vector<double> x_pnp = x;
|
std::vector<real_t> x_pnp = x;
|
||||||
x_pnp[axis1] += eps;
|
x_pnp[axis1] += eps;
|
||||||
x_pnp[axis3] += eps;
|
x_pnp[axis3] += eps;
|
||||||
|
|
||||||
std::vector<double> x_nnp = x;
|
std::vector<real_t> x_nnp = x;
|
||||||
x_nnp[axis3] += eps;
|
x_nnp[axis3] += eps;
|
||||||
|
|
||||||
std::vector<double> x_ppn = x;
|
std::vector<real_t> x_ppn = x;
|
||||||
x_ppn[axis1] += eps;
|
x_ppn[axis1] += eps;
|
||||||
x_ppn[axis2] += eps;
|
x_ppn[axis2] += eps;
|
||||||
|
|
||||||
std::vector<double> x_npn = x;
|
std::vector<real_t> x_npn = x;
|
||||||
x_npn[axis2] += eps;
|
x_npn[axis2] += eps;
|
||||||
|
|
||||||
std::vector<double> x_pnn = x;
|
std::vector<real_t> x_pnn = x;
|
||||||
x_pnn[axis1] += eps;
|
x_pnn[axis1] += eps;
|
||||||
|
|
||||||
double thirdAxis = function(x_ppp) - function(x_npp) - function(x_pnp) + function(x_nnp);
|
real_t thirdAxis = function(x_ppp) - function(x_npp) - function(x_pnp) + function(x_nnp);
|
||||||
double noThirdAxis = function(x_ppn) - function(x_npn) - function(x_pnn) + function(x);
|
real_t noThirdAxis = function(x_ppn) - function(x_npn) - function(x_pnn) + function(x);
|
||||||
return (thirdAxis - noThirdAxis) / (eps * eps * eps);
|
return (thirdAxis - noThirdAxis) / (eps * eps * eps);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::newtonRaphsonMethod(double (*function)(double), double x_0, double epoch_num) {
|
real_t MLPPNumericalAnalysis::newtonRaphsonMethod(real_t (*function)(real_t), real_t x_0, real_t epoch_num) {
|
||||||
double x = x_0;
|
real_t x = x_0;
|
||||||
for (int i = 0; i < epoch_num; i++) {
|
for (int i = 0; i < epoch_num; i++) {
|
||||||
x -= function(x) / numDiff(function, x);
|
x -= function(x) / numDiff(function, x);
|
||||||
}
|
}
|
||||||
return x;
|
return x;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::halleyMethod(double (*function)(double), double x_0, double epoch_num) {
|
real_t MLPPNumericalAnalysis::halleyMethod(real_t (*function)(real_t), real_t x_0, real_t epoch_num) {
|
||||||
double x = x_0;
|
real_t x = x_0;
|
||||||
for (int i = 0; i < epoch_num; i++) {
|
for (int i = 0; i < epoch_num; i++) {
|
||||||
x -= ((2 * function(x) * numDiff(function, x)) / (2 * numDiff(function, x) * numDiff(function, x) - function(x) * numDiff_2(function, x)));
|
x -= ((2 * function(x) * numDiff(function, x)) / (2 * numDiff(function, x) * numDiff(function, x) - function(x) * numDiff_2(function, x)));
|
||||||
}
|
}
|
||||||
return x;
|
return x;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::invQuadraticInterpolation(double (*function)(double), std::vector<double> x_0, double epoch_num) {
|
real_t MLPPNumericalAnalysis::invQuadraticInterpolation(real_t (*function)(real_t), std::vector<real_t> x_0, real_t epoch_num) {
|
||||||
double x = 0;
|
real_t x = 0;
|
||||||
std::vector<double> currentThree = x_0;
|
std::vector<real_t> currentThree = x_0;
|
||||||
for (int i = 0; i < epoch_num; i++) {
|
for (int i = 0; i < epoch_num; i++) {
|
||||||
double t1 = ((function(currentThree[1]) * function(currentThree[2])) / ((function(currentThree[0]) - function(currentThree[1])) * (function(currentThree[0]) - function(currentThree[2])))) * currentThree[0];
|
real_t t1 = ((function(currentThree[1]) * function(currentThree[2])) / ((function(currentThree[0]) - function(currentThree[1])) * (function(currentThree[0]) - function(currentThree[2])))) * currentThree[0];
|
||||||
double t2 = ((function(currentThree[0]) * function(currentThree[2])) / ((function(currentThree[1]) - function(currentThree[0])) * (function(currentThree[1]) - function(currentThree[2])))) * currentThree[1];
|
real_t t2 = ((function(currentThree[0]) * function(currentThree[2])) / ((function(currentThree[1]) - function(currentThree[0])) * (function(currentThree[1]) - function(currentThree[2])))) * currentThree[1];
|
||||||
double t3 = ((function(currentThree[0]) * function(currentThree[1])) / ((function(currentThree[2]) - function(currentThree[0])) * (function(currentThree[2]) - function(currentThree[1])))) * currentThree[2];
|
real_t t3 = ((function(currentThree[0]) * function(currentThree[1])) / ((function(currentThree[2]) - function(currentThree[0])) * (function(currentThree[2]) - function(currentThree[1])))) * currentThree[2];
|
||||||
x = t1 + t2 + t3;
|
x = t1 + t2 + t3;
|
||||||
|
|
||||||
currentThree.erase(currentThree.begin());
|
currentThree.erase(currentThree.begin());
|
||||||
@ -143,10 +143,10 @@ double MLPPNumericalAnalysis::invQuadraticInterpolation(double (*function)(doubl
|
|||||||
return x;
|
return x;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::eulerianMethod(double (*derivative)(double), std::vector<double> q_0, double p, double h) {
|
real_t MLPPNumericalAnalysis::eulerianMethod(real_t (*derivative)(real_t), std::vector<real_t> q_0, real_t p, real_t h) {
|
||||||
double max_epoch = (p - q_0[0]) / h;
|
real_t max_epoch = (p - q_0[0]) / h;
|
||||||
double x = q_0[0];
|
real_t x = q_0[0];
|
||||||
double y = q_0[1];
|
real_t y = q_0[1];
|
||||||
for (int i = 0; i < max_epoch; i++) {
|
for (int i = 0; i < max_epoch; i++) {
|
||||||
y = y + h * derivative(x);
|
y = y + h * derivative(x);
|
||||||
x += h;
|
x += h;
|
||||||
@ -154,10 +154,10 @@ double MLPPNumericalAnalysis::eulerianMethod(double (*derivative)(double), std::
|
|||||||
return y;
|
return y;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::eulerianMethod(double (*derivative)(std::vector<double>), std::vector<double> q_0, double p, double h) {
|
real_t MLPPNumericalAnalysis::eulerianMethod(real_t (*derivative)(std::vector<real_t>), std::vector<real_t> q_0, real_t p, real_t h) {
|
||||||
double max_epoch = (p - q_0[0]) / h;
|
real_t max_epoch = (p - q_0[0]) / h;
|
||||||
double x = q_0[0];
|
real_t x = q_0[0];
|
||||||
double y = q_0[1];
|
real_t y = q_0[1];
|
||||||
for (int i = 0; i < max_epoch; i++) {
|
for (int i = 0; i < max_epoch; i++) {
|
||||||
y = y + h * derivative({ x, y });
|
y = y + h * derivative({ x, y });
|
||||||
x += h;
|
x += h;
|
||||||
@ -165,7 +165,7 @@ double MLPPNumericalAnalysis::eulerianMethod(double (*derivative)(std::vector<do
|
|||||||
return y;
|
return y;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::growthMethod(double C, double k, double t) {
|
real_t MLPPNumericalAnalysis::growthMethod(real_t C, real_t k, real_t t) {
|
||||||
/*
|
/*
|
||||||
dP/dt = kP
|
dP/dt = kP
|
||||||
dP/P = kdt
|
dP/P = kdt
|
||||||
@ -177,20 +177,20 @@ double MLPPNumericalAnalysis::growthMethod(double C, double k, double t) {
|
|||||||
P = C * e^(kt)
|
P = C * e^(kt)
|
||||||
*/
|
*/
|
||||||
|
|
||||||
// auto growthFunction = [&C, &k](double t) { return C * exp(k * t); };
|
// auto growthFunction = [&C, &k](real_t t) { return C * exp(k * t); };
|
||||||
return C * std::exp(k * t);
|
return C * std::exp(k * t);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPNumericalAnalysis::jacobian(double (*function)(std::vector<double>), std::vector<double> x) {
|
std::vector<real_t> MLPPNumericalAnalysis::jacobian(real_t (*function)(std::vector<real_t>), std::vector<real_t> x) {
|
||||||
std::vector<double> jacobian;
|
std::vector<real_t> jacobian;
|
||||||
jacobian.resize(x.size());
|
jacobian.resize(x.size());
|
||||||
for (int i = 0; i < jacobian.size(); i++) {
|
for (int i = 0; i < jacobian.size(); i++) {
|
||||||
jacobian[i] = numDiff(function, x, i); // Derivative w.r.t axis i evaluated at x. For all x_i.
|
jacobian[i] = numDiff(function, x, i); // Derivative w.r.t axis i evaluated at x. For all x_i.
|
||||||
}
|
}
|
||||||
return jacobian;
|
return jacobian;
|
||||||
}
|
}
|
||||||
std::vector<std::vector<double>> MLPPNumericalAnalysis::hessian(double (*function)(std::vector<double>), std::vector<double> x) {
|
std::vector<std::vector<real_t>> MLPPNumericalAnalysis::hessian(real_t (*function)(std::vector<real_t>), std::vector<real_t> x) {
|
||||||
std::vector<std::vector<double>> hessian;
|
std::vector<std::vector<real_t>> hessian;
|
||||||
hessian.resize(x.size());
|
hessian.resize(x.size());
|
||||||
for (int i = 0; i < hessian.size(); i++) {
|
for (int i = 0; i < hessian.size(); i++) {
|
||||||
hessian[i].resize(x.size());
|
hessian[i].resize(x.size());
|
||||||
@ -203,8 +203,8 @@ std::vector<std::vector<double>> MLPPNumericalAnalysis::hessian(double (*functio
|
|||||||
return hessian;
|
return hessian;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> MLPPNumericalAnalysis::thirdOrderTensor(double (*function)(std::vector<double>), std::vector<double> x) {
|
std::vector<std::vector<std::vector<real_t>>> MLPPNumericalAnalysis::thirdOrderTensor(real_t (*function)(std::vector<real_t>), std::vector<real_t> x) {
|
||||||
std::vector<std::vector<std::vector<double>>> tensor;
|
std::vector<std::vector<std::vector<real_t>>> tensor;
|
||||||
tensor.resize(x.size());
|
tensor.resize(x.size());
|
||||||
for (int i = 0; i < tensor.size(); i++) {
|
for (int i = 0; i < tensor.size(); i++) {
|
||||||
tensor[i].resize(x.size());
|
tensor[i].resize(x.size());
|
||||||
@ -221,21 +221,21 @@ std::vector<std::vector<std::vector<double>>> MLPPNumericalAnalysis::thirdOrderT
|
|||||||
return tensor;
|
return tensor;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::constantApproximation(double (*function)(std::vector<double>), std::vector<double> c) {
|
real_t MLPPNumericalAnalysis::constantApproximation(real_t (*function)(std::vector<real_t>), std::vector<real_t> c) {
|
||||||
return function(c);
|
return function(c);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::linearApproximation(double (*function)(std::vector<double>), std::vector<double> c, std::vector<double> x) {
|
real_t MLPPNumericalAnalysis::linearApproximation(real_t (*function)(std::vector<real_t>), std::vector<real_t> c, std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return constantApproximation(function, c) + alg.matmult(alg.transpose({ jacobian(function, c) }), { alg.subtraction(x, c) })[0][0];
|
return constantApproximation(function, c) + alg.matmult(alg.transpose({ jacobian(function, c) }), { alg.subtraction(x, c) })[0][0];
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::quadraticApproximation(double (*function)(std::vector<double>), std::vector<double> c, std::vector<double> x) {
|
real_t MLPPNumericalAnalysis::quadraticApproximation(real_t (*function)(std::vector<real_t>), std::vector<real_t> c, std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return linearApproximation(function, c, x) + 0.5 * alg.matmult({ (alg.subtraction(x, c)) }, alg.matmult(hessian(function, c), alg.transpose({ alg.subtraction(x, c) })))[0][0];
|
return linearApproximation(function, c, x) + 0.5 * alg.matmult({ (alg.subtraction(x, c)) }, alg.matmult(hessian(function, c), alg.transpose({ alg.subtraction(x, c) })))[0][0];
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::cubicApproximation(double (*function)(std::vector<double>), std::vector<double> c, std::vector<double> x) {
|
real_t MLPPNumericalAnalysis::cubicApproximation(real_t (*function)(std::vector<real_t>), std::vector<real_t> c, std::vector<real_t> x) {
|
||||||
/*
|
/*
|
||||||
Not completely sure as the literature seldom discusses the third order taylor approximation,
|
Not completely sure as the literature seldom discusses the third order taylor approximation,
|
||||||
in particular for multivariate cases, but ostensibly, the matrix/tensor/vector multiplies
|
in particular for multivariate cases, but ostensibly, the matrix/tensor/vector multiplies
|
||||||
@ -246,32 +246,32 @@ double MLPPNumericalAnalysis::cubicApproximation(double (*function)(std::vector<
|
|||||||
Result is a scalar.
|
Result is a scalar.
|
||||||
*/
|
*/
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<double>> resultMat = alg.tensor_vec_mult(thirdOrderTensor(function, c), alg.subtraction(x, c));
|
std::vector<std::vector<real_t>> resultMat = alg.tensor_vec_mult(thirdOrderTensor(function, c), alg.subtraction(x, c));
|
||||||
double resultScalar = alg.matmult({ (alg.subtraction(x, c)) }, alg.matmult(resultMat, alg.transpose({ alg.subtraction(x, c) })))[0][0];
|
real_t resultScalar = alg.matmult({ (alg.subtraction(x, c)) }, alg.matmult(resultMat, alg.transpose({ alg.subtraction(x, c) })))[0][0];
|
||||||
|
|
||||||
return quadraticApproximation(function, c, x) + (1 / 6) * resultScalar;
|
return quadraticApproximation(function, c, x) + (1 / 6) * resultScalar;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPNumericalAnalysis::laplacian(double (*function)(std::vector<double>), std::vector<double> x) {
|
real_t MLPPNumericalAnalysis::laplacian(real_t (*function)(std::vector<real_t>), std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<double>> hessian_matrix = hessian(function, x);
|
std::vector<std::vector<real_t>> hessian_matrix = hessian(function, x);
|
||||||
double laplacian = 0;
|
real_t laplacian = 0;
|
||||||
for (int i = 0; i < hessian_matrix.size(); i++) {
|
for (int i = 0; i < hessian_matrix.size(); i++) {
|
||||||
laplacian += hessian_matrix[i][i]; // homogenous 2nd derivs w.r.t i, then i
|
laplacian += hessian_matrix[i][i]; // homogenous 2nd derivs w.r.t i, then i
|
||||||
}
|
}
|
||||||
return laplacian;
|
return laplacian;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::string MLPPNumericalAnalysis::secondPartialDerivativeTest(double (*function)(std::vector<double>), std::vector<double> x) {
|
std::string MLPPNumericalAnalysis::secondPartialDerivativeTest(real_t (*function)(std::vector<real_t>), std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<double>> hessianMatrix = hessian(function, x);
|
std::vector<std::vector<real_t>> hessianMatrix = hessian(function, x);
|
||||||
/*
|
/*
|
||||||
The reason we do this is because the 2nd partial derivative test is less conclusive for functions of variables greater than
|
The reason we do this is because the 2nd partial derivative test is less conclusive for functions of variables greater than
|
||||||
2, and the calculations specific to the bivariate case are less computationally intensive.
|
2, and the calculations specific to the bivariate case are less computationally intensive.
|
||||||
*/
|
*/
|
||||||
if (x.size() == 2) {
|
if (x.size() == 2) {
|
||||||
double det = alg.det(hessianMatrix, hessianMatrix.size());
|
real_t det = alg.det(hessianMatrix, hessianMatrix.size());
|
||||||
double secondDerivative = numDiff_2(function, x, 0, 0);
|
real_t secondDerivative = numDiff_2(function, x, 0, 0);
|
||||||
if (secondDerivative > 0 && det > 0) {
|
if (secondDerivative > 0 && det > 0) {
|
||||||
return "min";
|
return "min";
|
||||||
} else if (secondDerivative < 0 && det > 0) {
|
} else if (secondDerivative < 0 && det > 0) {
|
||||||
|
@ -7,6 +7,8 @@
|
|||||||
//
|
//
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
@ -17,40 +19,40 @@ public:
|
|||||||
as an analytical method for calculating derivatives will most likely be used in
|
as an analytical method for calculating derivatives will most likely be used in
|
||||||
the future.
|
the future.
|
||||||
*/
|
*/
|
||||||
double numDiff(double (*function)(double), double x);
|
real_t numDiff(real_t (*function)(real_t), real_t x);
|
||||||
double numDiff_2(double (*function)(double), double x);
|
real_t numDiff_2(real_t (*function)(real_t), real_t x);
|
||||||
double numDiff_3(double (*function)(double), double x);
|
real_t numDiff_3(real_t (*function)(real_t), real_t x);
|
||||||
|
|
||||||
double constantApproximation(double (*function)(double), double c);
|
real_t constantApproximation(real_t (*function)(real_t), real_t c);
|
||||||
double linearApproximation(double (*function)(double), double c, double x);
|
real_t linearApproximation(real_t (*function)(real_t), real_t c, real_t x);
|
||||||
double quadraticApproximation(double (*function)(double), double c, double x);
|
real_t quadraticApproximation(real_t (*function)(real_t), real_t c, real_t x);
|
||||||
double cubicApproximation(double (*function)(double), double c, double x);
|
real_t cubicApproximation(real_t (*function)(real_t), real_t c, real_t x);
|
||||||
|
|
||||||
double numDiff(double (*function)(std::vector<double>), std::vector<double> x, int axis);
|
real_t numDiff(real_t (*function)(std::vector<real_t>), std::vector<real_t> x, int axis);
|
||||||
double numDiff_2(double (*function)(std::vector<double>), std::vector<double> x, int axis1, int axis2);
|
real_t numDiff_2(real_t (*function)(std::vector<real_t>), std::vector<real_t> x, int axis1, int axis2);
|
||||||
double numDiff_3(double (*function)(std::vector<double>), std::vector<double> x, int axis1, int axis2, int axis3);
|
real_t numDiff_3(real_t (*function)(std::vector<real_t>), std::vector<real_t> x, int axis1, int axis2, int axis3);
|
||||||
|
|
||||||
double newtonRaphsonMethod(double (*function)(double), double x_0, double epoch_num);
|
real_t newtonRaphsonMethod(real_t (*function)(real_t), real_t x_0, real_t epoch_num);
|
||||||
double halleyMethod(double (*function)(double), double x_0, double epoch_num);
|
real_t halleyMethod(real_t (*function)(real_t), real_t x_0, real_t epoch_num);
|
||||||
double invQuadraticInterpolation(double (*function)(double), std::vector<double> x_0, double epoch_num);
|
real_t invQuadraticInterpolation(real_t (*function)(real_t), std::vector<real_t> x_0, real_t epoch_num);
|
||||||
|
|
||||||
double eulerianMethod(double (*derivative)(double), std::vector<double> q_0, double p, double h); // Euler's method for solving diffrential equations.
|
real_t eulerianMethod(real_t (*derivative)(real_t), std::vector<real_t> q_0, real_t p, real_t h); // Euler's method for solving diffrential equations.
|
||||||
double eulerianMethod(double (*derivative)(std::vector<double>), std::vector<double> q_0, double p, double h); // Euler's method for solving diffrential equations.
|
real_t eulerianMethod(real_t (*derivative)(std::vector<real_t>), std::vector<real_t> q_0, real_t p, real_t h); // Euler's method for solving diffrential equations.
|
||||||
|
|
||||||
double growthMethod(double C, double k, double t); // General growth-based diffrential equations can be solved by seperation of variables.
|
real_t growthMethod(real_t C, real_t k, real_t t); // General growth-based diffrential equations can be solved by seperation of variables.
|
||||||
|
|
||||||
std::vector<double> jacobian(double (*function)(std::vector<double>), std::vector<double> x); // Indeed, for functions with scalar outputs the Jacobians will be vectors.
|
std::vector<real_t> jacobian(real_t (*function)(std::vector<real_t>), std::vector<real_t> x); // Indeed, for functions with scalar outputs the Jacobians will be vectors.
|
||||||
std::vector<std::vector<double>> hessian(double (*function)(std::vector<double>), std::vector<double> x);
|
std::vector<std::vector<real_t>> hessian(real_t (*function)(std::vector<real_t>), std::vector<real_t> x);
|
||||||
std::vector<std::vector<std::vector<double>>> thirdOrderTensor(double (*function)(std::vector<double>), std::vector<double> x);
|
std::vector<std::vector<std::vector<real_t>>> thirdOrderTensor(real_t (*function)(std::vector<real_t>), std::vector<real_t> x);
|
||||||
|
|
||||||
double constantApproximation(double (*function)(std::vector<double>), std::vector<double> c);
|
real_t constantApproximation(real_t (*function)(std::vector<real_t>), std::vector<real_t> c);
|
||||||
double linearApproximation(double (*function)(std::vector<double>), std::vector<double> c, std::vector<double> x);
|
real_t linearApproximation(real_t (*function)(std::vector<real_t>), std::vector<real_t> c, std::vector<real_t> x);
|
||||||
double quadraticApproximation(double (*function)(std::vector<double>), std::vector<double> c, std::vector<double> x);
|
real_t quadraticApproximation(real_t (*function)(std::vector<real_t>), std::vector<real_t> c, std::vector<real_t> x);
|
||||||
double cubicApproximation(double (*function)(std::vector<double>), std::vector<double> c, std::vector<double> x);
|
real_t cubicApproximation(real_t (*function)(std::vector<real_t>), std::vector<real_t> c, std::vector<real_t> x);
|
||||||
|
|
||||||
double laplacian(double (*function)(std::vector<double>), std::vector<double> x); // laplacian
|
real_t laplacian(real_t (*function)(std::vector<real_t>), std::vector<real_t> x); // laplacian
|
||||||
|
|
||||||
std::string secondPartialDerivativeTest(double (*function)(std::vector<double>), std::vector<double> x);
|
std::string secondPartialDerivativeTest(real_t (*function)(std::vector<real_t>), std::vector<real_t> x);
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -13,13 +13,13 @@ MLPPOutlierFinder::MLPPOutlierFinder(int threshold) :
|
|||||||
threshold(threshold) {
|
threshold(threshold) {
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPOutlierFinder::modelSetTest(std::vector<std::vector<double>> inputSet) {
|
std::vector<std::vector<real_t>> MLPPOutlierFinder::modelSetTest(std::vector<std::vector<real_t>> inputSet) {
|
||||||
MLPPStat stat;
|
MLPPStat stat;
|
||||||
std::vector<std::vector<double>> outliers;
|
std::vector<std::vector<real_t>> outliers;
|
||||||
outliers.resize(inputSet.size());
|
outliers.resize(inputSet.size());
|
||||||
for (int i = 0; i < inputSet.size(); i++) {
|
for (int i = 0; i < inputSet.size(); i++) {
|
||||||
for (int j = 0; j < inputSet[i].size(); j++) {
|
for (int j = 0; j < inputSet[i].size(); j++) {
|
||||||
double z = (inputSet[i][j] - stat.mean(inputSet[i])) / stat.standardDeviation(inputSet[i]);
|
real_t z = (inputSet[i][j] - stat.mean(inputSet[i])) / stat.standardDeviation(inputSet[i]);
|
||||||
if (abs(z) > threshold) {
|
if (abs(z) > threshold) {
|
||||||
outliers[i].push_back(inputSet[i][j]);
|
outliers[i].push_back(inputSet[i][j]);
|
||||||
}
|
}
|
||||||
@ -28,11 +28,11 @@ std::vector<std::vector<double>> MLPPOutlierFinder::modelSetTest(std::vector<std
|
|||||||
return outliers;
|
return outliers;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPOutlierFinder::modelTest(std::vector<double> inputSet) {
|
std::vector<real_t> MLPPOutlierFinder::modelTest(std::vector<real_t> inputSet) {
|
||||||
MLPPStat stat;
|
MLPPStat stat;
|
||||||
std::vector<double> outliers;
|
std::vector<real_t> outliers;
|
||||||
for (int i = 0; i < inputSet.size(); i++) {
|
for (int i = 0; i < inputSet.size(); i++) {
|
||||||
double z = (inputSet[i] - stat.mean(inputSet)) / stat.standardDeviation(inputSet);
|
real_t z = (inputSet[i] - stat.mean(inputSet)) / stat.standardDeviation(inputSet);
|
||||||
if (abs(z) > threshold) {
|
if (abs(z) > threshold) {
|
||||||
outliers.push_back(inputSet[i]);
|
outliers.push_back(inputSet[i]);
|
||||||
}
|
}
|
||||||
|
@ -8,6 +8,8 @@
|
|||||||
// Created by Marc Melikyan on 11/13/20.
|
// Created by Marc Melikyan on 11/13/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
|
|
||||||
@ -16,8 +18,8 @@ public:
|
|||||||
// Cnstr
|
// Cnstr
|
||||||
MLPPOutlierFinder(int threshold);
|
MLPPOutlierFinder(int threshold);
|
||||||
|
|
||||||
std::vector<std::vector<double>> modelSetTest(std::vector<std::vector<double>> inputSet);
|
std::vector<std::vector<real_t>> modelSetTest(std::vector<std::vector<real_t>> inputSet);
|
||||||
std::vector<double> modelTest(std::vector<double> inputSet);
|
std::vector<real_t> modelTest(std::vector<real_t> inputSet);
|
||||||
|
|
||||||
// Variables required
|
// Variables required
|
||||||
int threshold;
|
int threshold;
|
||||||
|
@ -12,7 +12,7 @@
|
|||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
|
|
||||||
MLPPOutputLayer::MLPPOutputLayer(int n_hidden, std::string activation, std::string cost, std::vector<std::vector<double>> input, std::string weightInit, std::string reg, double lambda, double alpha) :
|
MLPPOutputLayer::MLPPOutputLayer(int n_hidden, std::string activation, std::string cost, std::vector<std::vector<real_t>> input, std::string weightInit, std::string reg, real_t lambda, real_t alpha) :
|
||||||
n_hidden(n_hidden), activation(activation), cost(cost), input(input), weightInit(weightInit), reg(reg), lambda(lambda), alpha(alpha) {
|
n_hidden(n_hidden), activation(activation), cost(cost), input(input), weightInit(weightInit), reg(reg), lambda(lambda), alpha(alpha) {
|
||||||
weights = MLPPUtilities::weightInitialization(n_hidden, weightInit);
|
weights = MLPPUtilities::weightInitialization(n_hidden, weightInit);
|
||||||
bias = MLPPUtilities::biasInitialization();
|
bias = MLPPUtilities::biasInitialization();
|
||||||
@ -120,7 +120,7 @@ void MLPPOutputLayer::forwardPass() {
|
|||||||
a = (avn.*activation_map[activation])(z, 0);
|
a = (avn.*activation_map[activation])(z, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPOutputLayer::Test(std::vector<double> x) {
|
void MLPPOutputLayer::Test(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
z_test = alg.dot(weights, x) + bias;
|
z_test = alg.dot(weights, x) + bias;
|
||||||
|
@ -8,6 +8,8 @@
|
|||||||
// Created by Marc Melikyan on 11/4/20.
|
// Created by Marc Melikyan on 11/4/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include "../activation/activation.h"
|
#include "../activation/activation.h"
|
||||||
#include "../cost/cost.h"
|
#include "../cost/cost.h"
|
||||||
|
|
||||||
@ -18,39 +20,39 @@
|
|||||||
|
|
||||||
class MLPPOutputLayer {
|
class MLPPOutputLayer {
|
||||||
public:
|
public:
|
||||||
MLPPOutputLayer(int n_hidden, std::string activation, std::string cost, std::vector<std::vector<double>> input, std::string weightInit, std::string reg, double lambda, double alpha);
|
MLPPOutputLayer(int n_hidden, std::string activation, std::string cost, std::vector<std::vector<real_t>> input, std::string weightInit, std::string reg, real_t lambda, real_t alpha);
|
||||||
|
|
||||||
int n_hidden;
|
int n_hidden;
|
||||||
std::string activation;
|
std::string activation;
|
||||||
std::string cost;
|
std::string cost;
|
||||||
|
|
||||||
std::vector<std::vector<double>> input;
|
std::vector<std::vector<real_t>> input;
|
||||||
|
|
||||||
std::vector<double> weights;
|
std::vector<real_t> weights;
|
||||||
double bias;
|
real_t bias;
|
||||||
|
|
||||||
std::vector<double> z;
|
std::vector<real_t> z;
|
||||||
std::vector<double> a;
|
std::vector<real_t> a;
|
||||||
|
|
||||||
std::map<std::string, std::vector<double> (MLPPActivation::*)(std::vector<double>, bool)> activation_map;
|
std::map<std::string, std::vector<real_t> (MLPPActivation::*)(std::vector<real_t>, bool)> activation_map;
|
||||||
std::map<std::string, double (MLPPActivation::*)(double, bool)> activationTest_map;
|
std::map<std::string, real_t (MLPPActivation::*)(real_t, bool)> activationTest_map;
|
||||||
std::map<std::string, double (MLPPCost::*)(std::vector<double>, std::vector<double>)> cost_map;
|
std::map<std::string, real_t (MLPPCost::*)(std::vector<real_t>, std::vector<real_t>)> cost_map;
|
||||||
std::map<std::string, std::vector<double> (MLPPCost::*)(std::vector<double>, std::vector<double>)> costDeriv_map;
|
std::map<std::string, std::vector<real_t> (MLPPCost::*)(std::vector<real_t>, std::vector<real_t>)> costDeriv_map;
|
||||||
|
|
||||||
double z_test;
|
real_t z_test;
|
||||||
double a_test;
|
real_t a_test;
|
||||||
|
|
||||||
std::vector<double> delta;
|
std::vector<real_t> delta;
|
||||||
|
|
||||||
// Regularization Params
|
// Regularization Params
|
||||||
std::string reg;
|
std::string reg;
|
||||||
double lambda; /* Regularization Parameter */
|
real_t lambda; /* Regularization Parameter */
|
||||||
double alpha; /* This is the controlling param for Elastic Net*/
|
real_t alpha; /* This is the controlling param for Elastic Net*/
|
||||||
|
|
||||||
std::string weightInit;
|
std::string weightInit;
|
||||||
|
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
void Test(std::vector<double> x);
|
void Test(std::vector<real_t> x);
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -13,11 +13,11 @@
|
|||||||
|
|
||||||
|
|
||||||
|
|
||||||
MLPPPCA::MLPPPCA(std::vector<std::vector<double>> inputSet, int k) :
|
MLPPPCA::MLPPPCA(std::vector<std::vector<real_t>> inputSet, int k) :
|
||||||
inputSet(inputSet), k(k) {
|
inputSet(inputSet), k(k) {
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPPCA::principalComponents() {
|
std::vector<std::vector<real_t>> MLPPPCA::principalComponents() {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPData data;
|
MLPPData data;
|
||||||
|
|
||||||
@ -33,10 +33,10 @@ std::vector<std::vector<double>> MLPPPCA::principalComponents() {
|
|||||||
return Z;
|
return Z;
|
||||||
}
|
}
|
||||||
// Simply tells us the percentage of variance maintained.
|
// Simply tells us the percentage of variance maintained.
|
||||||
double MLPPPCA::score() {
|
real_t MLPPPCA::score() {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
std::vector<std::vector<double>> X_approx = alg.matmult(U_reduce, Z);
|
std::vector<std::vector<real_t>> X_approx = alg.matmult(U_reduce, Z);
|
||||||
double num, den = 0;
|
real_t num, den = 0;
|
||||||
for (int i = 0; i < X_normalized.size(); i++) {
|
for (int i = 0; i < X_normalized.size(); i++) {
|
||||||
num += alg.norm_sq(alg.subtraction(X_normalized[i], X_approx[i]));
|
num += alg.norm_sq(alg.subtraction(X_normalized[i], X_approx[i]));
|
||||||
}
|
}
|
||||||
|
@ -8,20 +8,22 @@
|
|||||||
// Created by Marc Melikyan on 10/2/20.
|
// Created by Marc Melikyan on 10/2/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
|
|
||||||
class MLPPPCA {
|
class MLPPPCA {
|
||||||
public:
|
public:
|
||||||
MLPPPCA(std::vector<std::vector<double>> inputSet, int k);
|
MLPPPCA(std::vector<std::vector<real_t>> inputSet, int k);
|
||||||
std::vector<std::vector<double>> principalComponents();
|
std::vector<std::vector<real_t>> principalComponents();
|
||||||
double score();
|
real_t score();
|
||||||
|
|
||||||
private:
|
private:
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<std::vector<double>> X_normalized;
|
std::vector<std::vector<real_t>> X_normalized;
|
||||||
std::vector<std::vector<double>> U_reduce;
|
std::vector<std::vector<real_t>> U_reduce;
|
||||||
std::vector<std::vector<double>> Z;
|
std::vector<std::vector<real_t>> Z;
|
||||||
int k;
|
int k;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -15,33 +15,33 @@
|
|||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
|
|
||||||
MLPPProbitReg::MLPPProbitReg(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, std::string reg, double lambda, double alpha) :
|
MLPPProbitReg::MLPPProbitReg(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, std::string reg, real_t lambda, real_t alpha) :
|
||||||
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
||||||
y_hat.resize(n);
|
y_hat.resize(n);
|
||||||
weights = MLPPUtilities::weightInitialization(k);
|
weights = MLPPUtilities::weightInitialization(k);
|
||||||
bias = MLPPUtilities::biasInitialization();
|
bias = MLPPUtilities::biasInitialization();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPProbitReg::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPProbitReg::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
return Evaluate(X);
|
return Evaluate(X);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPProbitReg::modelTest(std::vector<double> x) {
|
real_t MLPPProbitReg::modelTest(std::vector<real_t> x) {
|
||||||
return Evaluate(x);
|
return Evaluate(x);
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPProbitReg::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
void MLPPProbitReg::gradientDescent(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
cost_prev = Cost(y_hat, outputSet);
|
cost_prev = Cost(y_hat, outputSet);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputSet);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputSet);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputSet), alg.hadamard_product(error, avn.gaussianCDF(z, 1)))));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputSet), alg.hadamard_product(error, avn.gaussianCDF(z, 1)))));
|
||||||
@ -63,18 +63,18 @@ void MLPPProbitReg::gradientDescent(double learning_rate, int max_epoch, bool UI
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPProbitReg::MLE(double learning_rate, int max_epoch, bool UI) {
|
void MLPPProbitReg::MLE(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
cost_prev = Cost(y_hat, outputSet);
|
cost_prev = Cost(y_hat, outputSet);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(outputSet, y_hat);
|
std::vector<real_t> error = alg.subtraction(outputSet, y_hat);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
weights = alg.addition(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputSet), alg.hadamard_product(error, avn.gaussianCDF(z, 1)))));
|
weights = alg.addition(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputSet), alg.hadamard_product(error, avn.gaussianCDF(z, 1)))));
|
||||||
@ -96,12 +96,12 @@ void MLPPProbitReg::MLE(double learning_rate, int max_epoch, bool UI) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPProbitReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
void MLPPProbitReg::SGD(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
// NOTE: ∂y_hat/∂z is sparse
|
// NOTE: ∂y_hat/∂z is sparse
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
@ -110,11 +110,11 @@ void MLPPProbitReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
||||||
int outputIndex = distribution(generator);
|
int outputIndex = distribution(generator);
|
||||||
|
|
||||||
double y_hat = Evaluate(inputSet[outputIndex]);
|
real_t y_hat = Evaluate(inputSet[outputIndex]);
|
||||||
double z = propagate(inputSet[outputIndex]);
|
real_t z = propagate(inputSet[outputIndex]);
|
||||||
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
||||||
|
|
||||||
double error = y_hat - outputSet[outputIndex];
|
real_t error = y_hat - outputSet[outputIndex];
|
||||||
|
|
||||||
// Weight Updation
|
// Weight Updation
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate * error * ((1 / sqrt(2 * M_PI)) * exp(-z * z / 2)), inputSet[outputIndex]));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate * error * ((1 / sqrt(2 * M_PI)) * exp(-z * z / 2)), inputSet[outputIndex]));
|
||||||
@ -138,11 +138,11 @@ void MLPPProbitReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPProbitReg::MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
void MLPPProbitReg::MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
@ -151,8 +151,8 @@ void MLPPProbitReg::MBGD(double learning_rate, int max_epoch, int mini_batch_siz
|
|||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<std::vector<double>> currentInputSet;
|
std::vector<std::vector<real_t>> currentInputSet;
|
||||||
std::vector<double> currentOutputSet;
|
std::vector<real_t> currentOutputSet;
|
||||||
for (int j = 0; j < n / n_mini_batch; j++) {
|
for (int j = 0; j < n / n_mini_batch; j++) {
|
||||||
currentInputSet.push_back(inputSet[n / n_mini_batch * i + j]);
|
currentInputSet.push_back(inputSet[n / n_mini_batch * i + j]);
|
||||||
currentOutputSet.push_back(outputSet[n / n_mini_batch * i + j]);
|
currentOutputSet.push_back(outputSet[n / n_mini_batch * i + j]);
|
||||||
@ -161,7 +161,7 @@ void MLPPProbitReg::MBGD(double learning_rate, int max_epoch, int mini_batch_siz
|
|||||||
outputMiniBatches.push_back(currentOutputSet);
|
outputMiniBatches.push_back(currentOutputSet);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (double(n) / double(n_mini_batch) - int(n / n_mini_batch) != 0) {
|
if (real_t(n) / real_t(n_mini_batch) - int(n / n_mini_batch) != 0) {
|
||||||
for (int i = 0; i < n - n / n_mini_batch * n_mini_batch; i++) {
|
for (int i = 0; i < n - n / n_mini_batch * n_mini_batch; i++) {
|
||||||
inputMiniBatches[n_mini_batch - 1].push_back(inputSet[n / n_mini_batch * n_mini_batch + i]);
|
inputMiniBatches[n_mini_batch - 1].push_back(inputSet[n / n_mini_batch * n_mini_batch + i]);
|
||||||
outputMiniBatches[n_mini_batch - 1].push_back(outputSet[n / n_mini_batch * n_mini_batch + i]);
|
outputMiniBatches[n_mini_batch - 1].push_back(outputSet[n / n_mini_batch * n_mini_batch + i]);
|
||||||
@ -170,11 +170,11 @@ void MLPPProbitReg::MBGD(double learning_rate, int max_epoch, int mini_batch_siz
|
|||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
std::vector<double> z = propagate(inputMiniBatches[i]);
|
std::vector<real_t> z = propagate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / outputMiniBatches.size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), alg.hadamard_product(error, avn.gaussianCDF(z, 1)))));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / outputMiniBatches.size(), alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), alg.hadamard_product(error, avn.gaussianCDF(z, 1)))));
|
||||||
@ -197,7 +197,7 @@ void MLPPProbitReg::MBGD(double learning_rate, int max_epoch, int mini_batch_siz
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPProbitReg::score() {
|
real_t MLPPProbitReg::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
return util.performance(y_hat, outputSet);
|
return util.performance(y_hat, outputSet);
|
||||||
}
|
}
|
||||||
@ -207,30 +207,30 @@ void MLPPProbitReg::save(std::string fileName) {
|
|||||||
util.saveParameters(fileName, weights, bias);
|
util.saveParameters(fileName, weights, bias);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPProbitReg::Cost(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPProbitReg::Cost(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
return cost.MSE(y_hat, y) + regularization.regTerm(weights, lambda, alpha, reg);
|
return cost.MSE(y_hat, y) + regularization.regTerm(weights, lambda, alpha, reg);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPProbitReg::Evaluate(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPProbitReg::Evaluate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
return avn.gaussianCDF(alg.scalarAdd(bias, alg.mat_vec_mult(X, weights)));
|
return avn.gaussianCDF(alg.scalarAdd(bias, alg.mat_vec_mult(X, weights)));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPProbitReg::propagate(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPProbitReg::propagate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.scalarAdd(bias, alg.mat_vec_mult(X, weights));
|
return alg.scalarAdd(bias, alg.mat_vec_mult(X, weights));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPProbitReg::Evaluate(std::vector<double> x) {
|
real_t MLPPProbitReg::Evaluate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
return avn.gaussianCDF(alg.dot(weights, x) + bias);
|
return avn.gaussianCDF(alg.dot(weights, x) + bias);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPProbitReg::propagate(std::vector<double> x) {
|
real_t MLPPProbitReg::propagate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.dot(weights, x) + bias;
|
return alg.dot(weights, x) + bias;
|
||||||
}
|
}
|
||||||
|
@ -8,6 +8,8 @@
|
|||||||
// Created by Marc Melikyan on 10/2/20.
|
// Created by Marc Melikyan on 10/2/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
@ -15,39 +17,39 @@
|
|||||||
|
|
||||||
class MLPPProbitReg {
|
class MLPPProbitReg {
|
||||||
public:
|
public:
|
||||||
MLPPProbitReg(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, std::string reg = "None", double lambda = 0.5, double alpha = 0.5);
|
MLPPProbitReg(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
||||||
std::vector<double> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<real_t> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
double modelTest(std::vector<double> x);
|
real_t modelTest(std::vector<real_t> x);
|
||||||
void gradientDescent(double learning_rate, int max_epoch = 0, bool UI = 1);
|
void gradientDescent(real_t learning_rate, int max_epoch = 0, bool UI = 1);
|
||||||
void MLE(double learning_rate, int max_epoch = 0, bool UI = 1);
|
void MLE(real_t learning_rate, int max_epoch = 0, bool UI = 1);
|
||||||
void SGD(double learning_rate, int max_epoch = 0, bool UI = 1);
|
void SGD(real_t learning_rate, int max_epoch = 0, bool UI = 1);
|
||||||
void MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
void MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
||||||
double score();
|
real_t score();
|
||||||
void save(std::string fileName);
|
void save(std::string fileName);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
double Cost(std::vector<double> y_hat, std::vector<double> y);
|
real_t Cost(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
|
|
||||||
std::vector<double> Evaluate(std::vector<std::vector<double>> X);
|
std::vector<real_t> Evaluate(std::vector<std::vector<real_t>> X);
|
||||||
std::vector<double> propagate(std::vector<std::vector<double>> X);
|
std::vector<real_t> propagate(std::vector<std::vector<real_t>> X);
|
||||||
double Evaluate(std::vector<double> x);
|
real_t Evaluate(std::vector<real_t> x);
|
||||||
double propagate(std::vector<double> x);
|
real_t propagate(std::vector<real_t> x);
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
std::vector<double> z;
|
std::vector<real_t> z;
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
std::vector<double> weights;
|
std::vector<real_t> weights;
|
||||||
double bias;
|
real_t bias;
|
||||||
|
|
||||||
int n;
|
int n;
|
||||||
int k;
|
int k;
|
||||||
|
|
||||||
// Regularization Params
|
// Regularization Params
|
||||||
std::string reg;
|
std::string reg;
|
||||||
double lambda;
|
real_t lambda;
|
||||||
double alpha; /* This is the controlling param for Elastic Net*/
|
real_t alpha; /* This is the controlling param for Elastic Net*/
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -12,21 +12,21 @@
|
|||||||
|
|
||||||
|
|
||||||
|
|
||||||
double MLPPReg::regTerm(std::vector<double> weights, double lambda, double alpha, std::string reg) {
|
real_t MLPPReg::regTerm(std::vector<real_t> weights, real_t lambda, real_t alpha, std::string reg) {
|
||||||
if (reg == "Ridge") {
|
if (reg == "Ridge") {
|
||||||
double reg = 0;
|
real_t reg = 0;
|
||||||
for (int i = 0; i < weights.size(); i++) {
|
for (int i = 0; i < weights.size(); i++) {
|
||||||
reg += weights[i] * weights[i];
|
reg += weights[i] * weights[i];
|
||||||
}
|
}
|
||||||
return reg * lambda / 2;
|
return reg * lambda / 2;
|
||||||
} else if (reg == "Lasso") {
|
} else if (reg == "Lasso") {
|
||||||
double reg = 0;
|
real_t reg = 0;
|
||||||
for (int i = 0; i < weights.size(); i++) {
|
for (int i = 0; i < weights.size(); i++) {
|
||||||
reg += abs(weights[i]);
|
reg += abs(weights[i]);
|
||||||
}
|
}
|
||||||
return reg * lambda;
|
return reg * lambda;
|
||||||
} else if (reg == "ElasticNet") {
|
} else if (reg == "ElasticNet") {
|
||||||
double reg = 0;
|
real_t reg = 0;
|
||||||
for (int i = 0; i < weights.size(); i++) {
|
for (int i = 0; i < weights.size(); i++) {
|
||||||
reg += alpha * abs(weights[i]); // Lasso Reg
|
reg += alpha * abs(weights[i]); // Lasso Reg
|
||||||
reg += ((1 - alpha) / 2) * weights[i] * weights[i]; // Ridge Reg
|
reg += ((1 - alpha) / 2) * weights[i] * weights[i]; // Ridge Reg
|
||||||
@ -36,9 +36,9 @@ double MLPPReg::regTerm(std::vector<double> weights, double lambda, double alpha
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPReg::regTerm(std::vector<std::vector<double>> weights, double lambda, double alpha, std::string reg) {
|
real_t MLPPReg::regTerm(std::vector<std::vector<real_t>> weights, real_t lambda, real_t alpha, std::string reg) {
|
||||||
if (reg == "Ridge") {
|
if (reg == "Ridge") {
|
||||||
double reg = 0;
|
real_t reg = 0;
|
||||||
for (int i = 0; i < weights.size(); i++) {
|
for (int i = 0; i < weights.size(); i++) {
|
||||||
for (int j = 0; j < weights[i].size(); j++) {
|
for (int j = 0; j < weights[i].size(); j++) {
|
||||||
reg += weights[i][j] * weights[i][j];
|
reg += weights[i][j] * weights[i][j];
|
||||||
@ -46,7 +46,7 @@ double MLPPReg::regTerm(std::vector<std::vector<double>> weights, double lambda,
|
|||||||
}
|
}
|
||||||
return reg * lambda / 2;
|
return reg * lambda / 2;
|
||||||
} else if (reg == "Lasso") {
|
} else if (reg == "Lasso") {
|
||||||
double reg = 0;
|
real_t reg = 0;
|
||||||
for (int i = 0; i < weights.size(); i++) {
|
for (int i = 0; i < weights.size(); i++) {
|
||||||
for (int j = 0; j < weights[i].size(); j++) {
|
for (int j = 0; j < weights[i].size(); j++) {
|
||||||
reg += abs(weights[i][j]);
|
reg += abs(weights[i][j]);
|
||||||
@ -54,7 +54,7 @@ double MLPPReg::regTerm(std::vector<std::vector<double>> weights, double lambda,
|
|||||||
}
|
}
|
||||||
return reg * lambda;
|
return reg * lambda;
|
||||||
} else if (reg == "ElasticNet") {
|
} else if (reg == "ElasticNet") {
|
||||||
double reg = 0;
|
real_t reg = 0;
|
||||||
for (int i = 0; i < weights.size(); i++) {
|
for (int i = 0; i < weights.size(); i++) {
|
||||||
for (int j = 0; j < weights[i].size(); j++) {
|
for (int j = 0; j < weights[i].size(); j++) {
|
||||||
reg += alpha * abs(weights[i][j]); // Lasso Reg
|
reg += alpha * abs(weights[i][j]); // Lasso Reg
|
||||||
@ -66,7 +66,7 @@ double MLPPReg::regTerm(std::vector<std::vector<double>> weights, double lambda,
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPReg::regWeights(std::vector<double> weights, double lambda, double alpha, std::string reg) {
|
std::vector<real_t> MLPPReg::regWeights(std::vector<real_t> weights, real_t lambda, real_t alpha, std::string reg) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (reg == "WeightClipping") {
|
if (reg == "WeightClipping") {
|
||||||
return regDerivTerm(weights, lambda, alpha, reg);
|
return regDerivTerm(weights, lambda, alpha, reg);
|
||||||
@ -78,7 +78,7 @@ std::vector<double> MLPPReg::regWeights(std::vector<double> weights, double lamb
|
|||||||
// return weights;
|
// return weights;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPReg::regWeights(std::vector<std::vector<double>> weights, double lambda, double alpha, std::string reg) {
|
std::vector<std::vector<real_t>> MLPPReg::regWeights(std::vector<std::vector<real_t>> weights, real_t lambda, real_t alpha, std::string reg) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (reg == "WeightClipping") {
|
if (reg == "WeightClipping") {
|
||||||
return regDerivTerm(weights, lambda, alpha, reg);
|
return regDerivTerm(weights, lambda, alpha, reg);
|
||||||
@ -92,8 +92,8 @@ std::vector<std::vector<double>> MLPPReg::regWeights(std::vector<std::vector<dou
|
|||||||
// return weights;
|
// return weights;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPReg::regDerivTerm(std::vector<double> weights, double lambda, double alpha, std::string reg) {
|
std::vector<real_t> MLPPReg::regDerivTerm(std::vector<real_t> weights, real_t lambda, real_t alpha, std::string reg) {
|
||||||
std::vector<double> regDeriv;
|
std::vector<real_t> regDeriv;
|
||||||
regDeriv.resize(weights.size());
|
regDeriv.resize(weights.size());
|
||||||
|
|
||||||
for (int i = 0; i < regDeriv.size(); i++) {
|
for (int i = 0; i < regDeriv.size(); i++) {
|
||||||
@ -102,8 +102,8 @@ std::vector<double> MLPPReg::regDerivTerm(std::vector<double> weights, double la
|
|||||||
return regDeriv;
|
return regDeriv;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPReg::regDerivTerm(std::vector<std::vector<double>> weights, double lambda, double alpha, std::string reg) {
|
std::vector<std::vector<real_t>> MLPPReg::regDerivTerm(std::vector<std::vector<real_t>> weights, real_t lambda, real_t alpha, std::string reg) {
|
||||||
std::vector<std::vector<double>> regDeriv;
|
std::vector<std::vector<real_t>> regDeriv;
|
||||||
regDeriv.resize(weights.size());
|
regDeriv.resize(weights.size());
|
||||||
for (int i = 0; i < regDeriv.size(); i++) {
|
for (int i = 0; i < regDeriv.size(); i++) {
|
||||||
regDeriv[i].resize(weights[0].size());
|
regDeriv[i].resize(weights[0].size());
|
||||||
@ -117,7 +117,7 @@ std::vector<std::vector<double>> MLPPReg::regDerivTerm(std::vector<std::vector<d
|
|||||||
return regDeriv;
|
return regDeriv;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPReg::regDerivTerm(std::vector<double> weights, double lambda, double alpha, std::string reg, int j) {
|
real_t MLPPReg::regDerivTerm(std::vector<real_t> weights, real_t lambda, real_t alpha, std::string reg, int j) {
|
||||||
MLPPActivation act;
|
MLPPActivation act;
|
||||||
if (reg == "Ridge") {
|
if (reg == "Ridge") {
|
||||||
return lambda * weights[j];
|
return lambda * weights[j];
|
||||||
@ -140,7 +140,7 @@ double MLPPReg::regDerivTerm(std::vector<double> weights, double lambda, double
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPReg::regDerivTerm(std::vector<std::vector<double>> weights, double lambda, double alpha, std::string reg, int i, int j) {
|
real_t MLPPReg::regDerivTerm(std::vector<std::vector<real_t>> weights, real_t lambda, real_t alpha, std::string reg, int i, int j) {
|
||||||
MLPPActivation act;
|
MLPPActivation act;
|
||||||
if (reg == "Ridge") {
|
if (reg == "Ridge") {
|
||||||
return lambda * weights[i][j];
|
return lambda * weights[i][j];
|
||||||
|
@ -9,24 +9,26 @@
|
|||||||
// Created by Marc Melikyan on 1/16/21.
|
// Created by Marc Melikyan on 1/16/21.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <vector>
|
#include <vector>
|
||||||
#include <string>
|
#include <string>
|
||||||
|
|
||||||
|
|
||||||
class MLPPReg {
|
class MLPPReg {
|
||||||
public:
|
public:
|
||||||
double regTerm(std::vector<double> weights, double lambda, double alpha, std::string reg);
|
real_t regTerm(std::vector<real_t> weights, real_t lambda, real_t alpha, std::string reg);
|
||||||
double regTerm(std::vector<std::vector<double>> weights, double lambda, double alpha, std::string reg);
|
real_t regTerm(std::vector<std::vector<real_t>> weights, real_t lambda, real_t alpha, std::string reg);
|
||||||
|
|
||||||
std::vector<double> regWeights(std::vector<double> weights, double lambda, double alpha, std::string reg);
|
std::vector<real_t> regWeights(std::vector<real_t> weights, real_t lambda, real_t alpha, std::string reg);
|
||||||
std::vector<std::vector<double>> regWeights(std::vector<std::vector<double>> weights, double lambda, double alpha, std::string reg);
|
std::vector<std::vector<real_t>> regWeights(std::vector<std::vector<real_t>> weights, real_t lambda, real_t alpha, std::string reg);
|
||||||
|
|
||||||
std::vector<double> regDerivTerm(std::vector<double> weights, double lambda, double alpha, std::string reg);
|
std::vector<real_t> regDerivTerm(std::vector<real_t> weights, real_t lambda, real_t alpha, std::string reg);
|
||||||
std::vector<std::vector<double>> regDerivTerm(std::vector<std::vector<double>>, double lambda, double alpha, std::string reg);
|
std::vector<std::vector<real_t>> regDerivTerm(std::vector<std::vector<real_t>>, real_t lambda, real_t alpha, std::string reg);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
double regDerivTerm(std::vector<double> weights, double lambda, double alpha, std::string reg, int j);
|
real_t regDerivTerm(std::vector<real_t> weights, real_t lambda, real_t alpha, std::string reg, int j);
|
||||||
double regDerivTerm(std::vector<std::vector<double>> weights, double lambda, double alpha, std::string reg, int i, int j);
|
real_t regDerivTerm(std::vector<std::vector<real_t>> weights, real_t lambda, real_t alpha, std::string reg, int i, int j);
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -16,7 +16,7 @@
|
|||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
|
|
||||||
MLPPSoftmaxNet::MLPPSoftmaxNet(std::vector<std::vector<double>> inputSet, std::vector<std::vector<double>> outputSet, int n_hidden, std::string reg, double lambda, double alpha) :
|
MLPPSoftmaxNet::MLPPSoftmaxNet(std::vector<std::vector<real_t>> inputSet, std::vector<std::vector<real_t>> outputSet, int n_hidden, std::string reg, real_t lambda, real_t alpha) :
|
||||||
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), n_hidden(n_hidden), n_class(outputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), n_hidden(n_hidden), n_class(outputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
||||||
y_hat.resize(n);
|
y_hat.resize(n);
|
||||||
|
|
||||||
@ -26,19 +26,19 @@ MLPPSoftmaxNet::MLPPSoftmaxNet(std::vector<std::vector<double>> inputSet, std::v
|
|||||||
bias2 = MLPPUtilities::biasInitialization(n_class);
|
bias2 = MLPPUtilities::biasInitialization(n_class);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPSoftmaxNet::modelTest(std::vector<double> x) {
|
std::vector<real_t> MLPPSoftmaxNet::modelTest(std::vector<real_t> x) {
|
||||||
return Evaluate(x);
|
return Evaluate(x);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPSoftmaxNet::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<std::vector<real_t>> MLPPSoftmaxNet::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
return Evaluate(X);
|
return Evaluate(X);
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPSoftmaxNet::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
void MLPPSoftmaxNet::gradientDescent(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
@ -46,11 +46,11 @@ void MLPPSoftmaxNet::gradientDescent(double learning_rate, int max_epoch, bool U
|
|||||||
cost_prev = Cost(y_hat, outputSet);
|
cost_prev = Cost(y_hat, outputSet);
|
||||||
|
|
||||||
// Calculating the errors
|
// Calculating the errors
|
||||||
std::vector<std::vector<double>> error = alg.subtraction(y_hat, outputSet);
|
std::vector<std::vector<real_t>> error = alg.subtraction(y_hat, outputSet);
|
||||||
|
|
||||||
// Calculating the weight/bias gradients for layer 2
|
// Calculating the weight/bias gradients for layer 2
|
||||||
|
|
||||||
std::vector<std::vector<double>> D2_1 = alg.matmult(alg.transpose(a2), error);
|
std::vector<std::vector<real_t>> D2_1 = alg.matmult(alg.transpose(a2), error);
|
||||||
|
|
||||||
// weights and bias updation for layer 2
|
// weights and bias updation for layer 2
|
||||||
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate, D2_1));
|
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate, D2_1));
|
||||||
@ -60,11 +60,11 @@ void MLPPSoftmaxNet::gradientDescent(double learning_rate, int max_epoch, bool U
|
|||||||
|
|
||||||
//Calculating the weight/bias for layer 1
|
//Calculating the weight/bias for layer 1
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_1 = alg.matmult(error, alg.transpose(weights2));
|
std::vector<std::vector<real_t>> D1_1 = alg.matmult(error, alg.transpose(weights2));
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
std::vector<std::vector<real_t>> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_3 = alg.matmult(alg.transpose(inputSet), D1_2);
|
std::vector<std::vector<real_t>> D1_3 = alg.matmult(alg.transpose(inputSet), D1_2);
|
||||||
|
|
||||||
// weight an bias updation for layer 1
|
// weight an bias updation for layer 1
|
||||||
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate, D1_3));
|
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate, D1_3));
|
||||||
@ -90,11 +90,11 @@ void MLPPSoftmaxNet::gradientDescent(double learning_rate, int max_epoch, bool U
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPSoftmaxNet::SGD(double learning_rate, int max_epoch, bool UI) {
|
void MLPPSoftmaxNet::SGD(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
@ -103,13 +103,13 @@ void MLPPSoftmaxNet::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
||||||
int outputIndex = distribution(generator);
|
int outputIndex = distribution(generator);
|
||||||
|
|
||||||
std::vector<double> y_hat = Evaluate(inputSet[outputIndex]);
|
std::vector<real_t> y_hat = Evaluate(inputSet[outputIndex]);
|
||||||
auto [z2, a2] = propagate(inputSet[outputIndex]);
|
auto [z2, a2] = propagate(inputSet[outputIndex]);
|
||||||
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputSet[outputIndex]);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputSet[outputIndex]);
|
||||||
|
|
||||||
// Weight updation for layer 2
|
// Weight updation for layer 2
|
||||||
std::vector<std::vector<double>> D2_1 = alg.outerProduct(error, a2);
|
std::vector<std::vector<real_t>> D2_1 = alg.outerProduct(error, a2);
|
||||||
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate, alg.transpose(D2_1)));
|
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate, alg.transpose(D2_1)));
|
||||||
weights2 = regularization.regWeights(weights2, lambda, alpha, reg);
|
weights2 = regularization.regWeights(weights2, lambda, alpha, reg);
|
||||||
|
|
||||||
@ -117,9 +117,9 @@ void MLPPSoftmaxNet::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
bias2 = alg.subtraction(bias2, alg.scalarMultiply(learning_rate, error));
|
bias2 = alg.subtraction(bias2, alg.scalarMultiply(learning_rate, error));
|
||||||
|
|
||||||
// Weight updation for layer 1
|
// Weight updation for layer 1
|
||||||
std::vector<double> D1_1 = alg.mat_vec_mult(weights2, error);
|
std::vector<real_t> D1_1 = alg.mat_vec_mult(weights2, error);
|
||||||
std::vector<double> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
std::vector<real_t> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
||||||
std::vector<std::vector<double>> D1_3 = alg.outerProduct(inputSet[outputIndex], D1_2);
|
std::vector<std::vector<real_t>> D1_3 = alg.outerProduct(inputSet[outputIndex], D1_2);
|
||||||
|
|
||||||
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate, D1_3));
|
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate, D1_3));
|
||||||
weights1 = regularization.regWeights(weights1, lambda, alpha, reg);
|
weights1 = regularization.regWeights(weights1, lambda, alpha, reg);
|
||||||
@ -144,11 +144,11 @@ void MLPPSoftmaxNet::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPSoftmaxNet::MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
void MLPPSoftmaxNet::MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
@ -157,8 +157,8 @@ void MLPPSoftmaxNet::MBGD(double learning_rate, int max_epoch, int mini_batch_si
|
|||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<std::vector<double>> currentInputSet;
|
std::vector<std::vector<real_t>> currentInputSet;
|
||||||
std::vector<std::vector<double>> currentOutputSet;
|
std::vector<std::vector<real_t>> currentOutputSet;
|
||||||
for (int j = 0; j < n / n_mini_batch; j++) {
|
for (int j = 0; j < n / n_mini_batch; j++) {
|
||||||
currentInputSet.push_back(inputSet[n / n_mini_batch * i + j]);
|
currentInputSet.push_back(inputSet[n / n_mini_batch * i + j]);
|
||||||
currentOutputSet.push_back(outputSet[n / n_mini_batch * i + j]);
|
currentOutputSet.push_back(outputSet[n / n_mini_batch * i + j]);
|
||||||
@ -167,7 +167,7 @@ void MLPPSoftmaxNet::MBGD(double learning_rate, int max_epoch, int mini_batch_si
|
|||||||
outputMiniBatches.push_back(currentOutputSet);
|
outputMiniBatches.push_back(currentOutputSet);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (double(n) / double(n_mini_batch) - int(n / n_mini_batch) != 0) {
|
if (real_t(n) / real_t(n_mini_batch) - int(n / n_mini_batch) != 0) {
|
||||||
for (int i = 0; i < n - n / n_mini_batch * n_mini_batch; i++) {
|
for (int i = 0; i < n - n / n_mini_batch * n_mini_batch; i++) {
|
||||||
inputMiniBatches[n_mini_batch - 1].push_back(inputSet[n / n_mini_batch * n_mini_batch + i]);
|
inputMiniBatches[n_mini_batch - 1].push_back(inputSet[n / n_mini_batch * n_mini_batch + i]);
|
||||||
outputMiniBatches[n_mini_batch - 1].push_back(outputSet[n / n_mini_batch * n_mini_batch + i]);
|
outputMiniBatches[n_mini_batch - 1].push_back(outputSet[n / n_mini_batch * n_mini_batch + i]);
|
||||||
@ -176,16 +176,16 @@ void MLPPSoftmaxNet::MBGD(double learning_rate, int max_epoch, int mini_batch_si
|
|||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<std::vector<double>> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<std::vector<real_t>> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
auto [z2, a2] = propagate(inputMiniBatches[i]);
|
auto [z2, a2] = propagate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the errors
|
// Calculating the errors
|
||||||
std::vector<std::vector<double>> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
std::vector<std::vector<real_t>> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the weight/bias gradients for layer 2
|
// Calculating the weight/bias gradients for layer 2
|
||||||
|
|
||||||
std::vector<std::vector<double>> D2_1 = alg.matmult(alg.transpose(a2), error);
|
std::vector<std::vector<real_t>> D2_1 = alg.matmult(alg.transpose(a2), error);
|
||||||
|
|
||||||
// weights and bias updation for layser 2
|
// weights and bias updation for layser 2
|
||||||
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate, D2_1));
|
weights2 = alg.subtraction(weights2, alg.scalarMultiply(learning_rate, D2_1));
|
||||||
@ -196,11 +196,11 @@ void MLPPSoftmaxNet::MBGD(double learning_rate, int max_epoch, int mini_batch_si
|
|||||||
|
|
||||||
//Calculating the weight/bias for layer 1
|
//Calculating the weight/bias for layer 1
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_1 = alg.matmult(error, alg.transpose(weights2));
|
std::vector<std::vector<real_t>> D1_1 = alg.matmult(error, alg.transpose(weights2));
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
std::vector<std::vector<real_t>> D1_2 = alg.hadamard_product(D1_1, avn.sigmoid(z2, 1));
|
||||||
|
|
||||||
std::vector<std::vector<double>> D1_3 = alg.matmult(alg.transpose(inputMiniBatches[i]), D1_2);
|
std::vector<std::vector<real_t>> D1_3 = alg.matmult(alg.transpose(inputMiniBatches[i]), D1_2);
|
||||||
|
|
||||||
// weight an bias updation for layer 1
|
// weight an bias updation for layer 1
|
||||||
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate, D1_3));
|
weights1 = alg.subtraction(weights1, alg.scalarMultiply(learning_rate, D1_3));
|
||||||
@ -226,7 +226,7 @@ void MLPPSoftmaxNet::MBGD(double learning_rate, int max_epoch, int mini_batch_si
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPSoftmaxNet::score() {
|
real_t MLPPSoftmaxNet::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
return util.performance(y_hat, outputSet);
|
return util.performance(y_hat, outputSet);
|
||||||
}
|
}
|
||||||
@ -239,46 +239,46 @@ void MLPPSoftmaxNet::save(std::string fileName) {
|
|||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPSoftmaxNet::getEmbeddings() {
|
std::vector<std::vector<real_t>> MLPPSoftmaxNet::getEmbeddings() {
|
||||||
return weights1;
|
return weights1;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPSoftmaxNet::Cost(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
real_t MLPPSoftmaxNet::Cost(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
MLPPData data;
|
MLPPData data;
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
return cost.CrossEntropy(y_hat, y) + regularization.regTerm(weights1, lambda, alpha, reg) + regularization.regTerm(weights2, lambda, alpha, reg);
|
return cost.CrossEntropy(y_hat, y) + regularization.regTerm(weights1, lambda, alpha, reg) + regularization.regTerm(weights2, lambda, alpha, reg);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPSoftmaxNet::Evaluate(std::vector<std::vector<double>> X) {
|
std::vector<std::vector<real_t>> MLPPSoftmaxNet::Evaluate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
std::vector<std::vector<double>> z2 = alg.mat_vec_add(alg.matmult(X, weights1), bias1);
|
std::vector<std::vector<real_t>> z2 = alg.mat_vec_add(alg.matmult(X, weights1), bias1);
|
||||||
std::vector<std::vector<double>> a2 = avn.sigmoid(z2);
|
std::vector<std::vector<real_t>> a2 = avn.sigmoid(z2);
|
||||||
return avn.adjSoftmax(alg.mat_vec_add(alg.matmult(a2, weights2), bias2));
|
return avn.adjSoftmax(alg.mat_vec_add(alg.matmult(a2, weights2), bias2));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>> MLPPSoftmaxNet::propagate(std::vector<std::vector<double>> X) {
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> MLPPSoftmaxNet::propagate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
std::vector<std::vector<double>> z2 = alg.mat_vec_add(alg.matmult(X, weights1), bias1);
|
std::vector<std::vector<real_t>> z2 = alg.mat_vec_add(alg.matmult(X, weights1), bias1);
|
||||||
std::vector<std::vector<double>> a2 = avn.sigmoid(z2);
|
std::vector<std::vector<real_t>> a2 = avn.sigmoid(z2);
|
||||||
return { z2, a2 };
|
return { z2, a2 };
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPSoftmaxNet::Evaluate(std::vector<double> x) {
|
std::vector<real_t> MLPPSoftmaxNet::Evaluate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
std::vector<double> z2 = alg.addition(alg.mat_vec_mult(alg.transpose(weights1), x), bias1);
|
std::vector<real_t> z2 = alg.addition(alg.mat_vec_mult(alg.transpose(weights1), x), bias1);
|
||||||
std::vector<double> a2 = avn.sigmoid(z2);
|
std::vector<real_t> a2 = avn.sigmoid(z2);
|
||||||
return avn.adjSoftmax(alg.addition(alg.mat_vec_mult(alg.transpose(weights2), a2), bias2));
|
return avn.adjSoftmax(alg.addition(alg.mat_vec_mult(alg.transpose(weights2), a2), bias2));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<double>, std::vector<double>> MLPPSoftmaxNet::propagate(std::vector<double> x) {
|
std::tuple<std::vector<real_t>, std::vector<real_t>> MLPPSoftmaxNet::propagate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
std::vector<double> z2 = alg.addition(alg.mat_vec_mult(alg.transpose(weights1), x), bias1);
|
std::vector<real_t> z2 = alg.addition(alg.mat_vec_mult(alg.transpose(weights1), x), bias1);
|
||||||
std::vector<double> a2 = avn.sigmoid(z2);
|
std::vector<real_t> a2 = avn.sigmoid(z2);
|
||||||
return { z2, a2 };
|
return { z2, a2 };
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -8,6 +8,8 @@
|
|||||||
// Created by Marc Melikyan on 10/2/20.
|
// Created by Marc Melikyan on 10/2/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
@ -15,37 +17,37 @@
|
|||||||
|
|
||||||
class MLPPSoftmaxNet {
|
class MLPPSoftmaxNet {
|
||||||
public:
|
public:
|
||||||
MLPPSoftmaxNet(std::vector<std::vector<double>> inputSet, std::vector<std::vector<double>> outputSet, int n_hidden, std::string reg = "None", double lambda = 0.5, double alpha = 0.5);
|
MLPPSoftmaxNet(std::vector<std::vector<real_t>> inputSet, std::vector<std::vector<real_t>> outputSet, int n_hidden, std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
||||||
std::vector<double> modelTest(std::vector<double> x);
|
std::vector<real_t> modelTest(std::vector<real_t> x);
|
||||||
std::vector<std::vector<double>> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<std::vector<real_t>> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
void gradientDescent(double learning_rate, int max_epoch, bool UI = 1);
|
void gradientDescent(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void SGD(double learning_rate, int max_epoch, bool UI = 1);
|
void SGD(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
void MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
||||||
double score();
|
real_t score();
|
||||||
void save(std::string fileName);
|
void save(std::string fileName);
|
||||||
|
|
||||||
std::vector<std::vector<double>> getEmbeddings(); // This class is used (mostly) for word2Vec. This function returns our embeddings.
|
std::vector<std::vector<real_t>> getEmbeddings(); // This class is used (mostly) for word2Vec. This function returns our embeddings.
|
||||||
private:
|
private:
|
||||||
double Cost(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
real_t Cost(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
std::vector<std::vector<double>> Evaluate(std::vector<std::vector<double>> X);
|
std::vector<std::vector<real_t>> Evaluate(std::vector<std::vector<real_t>> X);
|
||||||
std::tuple<std::vector<std::vector<double>>, std::vector<std::vector<double>>> propagate(std::vector<std::vector<double>> X);
|
std::tuple<std::vector<std::vector<real_t>>, std::vector<std::vector<real_t>>> propagate(std::vector<std::vector<real_t>> X);
|
||||||
std::vector<double> Evaluate(std::vector<double> x);
|
std::vector<real_t> Evaluate(std::vector<real_t> x);
|
||||||
std::tuple<std::vector<double>, std::vector<double>> propagate(std::vector<double> x);
|
std::tuple<std::vector<real_t>, std::vector<real_t>> propagate(std::vector<real_t> x);
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<std::vector<double>> outputSet;
|
std::vector<std::vector<real_t>> outputSet;
|
||||||
std::vector<std::vector<double>> y_hat;
|
std::vector<std::vector<real_t>> y_hat;
|
||||||
|
|
||||||
std::vector<std::vector<double>> weights1;
|
std::vector<std::vector<real_t>> weights1;
|
||||||
std::vector<std::vector<double>> weights2;
|
std::vector<std::vector<real_t>> weights2;
|
||||||
|
|
||||||
std::vector<double> bias1;
|
std::vector<real_t> bias1;
|
||||||
std::vector<double> bias2;
|
std::vector<real_t> bias2;
|
||||||
|
|
||||||
std::vector<std::vector<double>> z2;
|
std::vector<std::vector<real_t>> z2;
|
||||||
std::vector<std::vector<double>> a2;
|
std::vector<std::vector<real_t>> a2;
|
||||||
|
|
||||||
int n;
|
int n;
|
||||||
int k;
|
int k;
|
||||||
@ -54,8 +56,8 @@ private:
|
|||||||
|
|
||||||
// Regularization Params
|
// Regularization Params
|
||||||
std::string reg;
|
std::string reg;
|
||||||
double lambda;
|
real_t lambda;
|
||||||
double alpha; /* This is the controlling param for Elastic Net*/
|
real_t alpha; /* This is the controlling param for Elastic Net*/
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -15,41 +15,41 @@
|
|||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
|
|
||||||
MLPPSoftmaxReg::MLPPSoftmaxReg(std::vector<std::vector<double>> inputSet, std::vector<std::vector<double>> outputSet, std::string reg, double lambda, double alpha) :
|
MLPPSoftmaxReg::MLPPSoftmaxReg(std::vector<std::vector<real_t>> inputSet, std::vector<std::vector<real_t>> outputSet, std::string reg, real_t lambda, real_t alpha) :
|
||||||
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), n_class(outputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), n_class(outputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
||||||
y_hat.resize(n);
|
y_hat.resize(n);
|
||||||
weights = MLPPUtilities::weightInitialization(k, n_class);
|
weights = MLPPUtilities::weightInitialization(k, n_class);
|
||||||
bias = MLPPUtilities::biasInitialization(n_class);
|
bias = MLPPUtilities::biasInitialization(n_class);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPSoftmaxReg::modelTest(std::vector<double> x) {
|
std::vector<real_t> MLPPSoftmaxReg::modelTest(std::vector<real_t> x) {
|
||||||
return Evaluate(x);
|
return Evaluate(x);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPSoftmaxReg::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<std::vector<real_t>> MLPPSoftmaxReg::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
return Evaluate(X);
|
return Evaluate(X);
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPSoftmaxReg::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
void MLPPSoftmaxReg::gradientDescent(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
cost_prev = Cost(y_hat, outputSet);
|
cost_prev = Cost(y_hat, outputSet);
|
||||||
std::vector<std::vector<double>> error = alg.subtraction(y_hat, outputSet);
|
std::vector<std::vector<real_t>> error = alg.subtraction(y_hat, outputSet);
|
||||||
|
|
||||||
//Calculating the weight gradients
|
//Calculating the weight gradients
|
||||||
std::vector<std::vector<double>> w_gradient = alg.matmult(alg.transpose(inputSet), error);
|
std::vector<std::vector<real_t>> w_gradient = alg.matmult(alg.transpose(inputSet), error);
|
||||||
|
|
||||||
//Weight updation
|
//Weight updation
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate, w_gradient));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate, w_gradient));
|
||||||
weights = regularization.regWeights(weights, lambda, alpha, reg);
|
weights = regularization.regWeights(weights, lambda, alpha, reg);
|
||||||
|
|
||||||
// Calculating the bias gradients
|
// Calculating the bias gradients
|
||||||
//double b_gradient = alg.sum_elements(error);
|
//real_t b_gradient = alg.sum_elements(error);
|
||||||
|
|
||||||
// Bias Updation
|
// Bias Updation
|
||||||
bias = alg.subtractMatrixRows(bias, alg.scalarMultiply(learning_rate, error));
|
bias = alg.subtractMatrixRows(bias, alg.scalarMultiply(learning_rate, error));
|
||||||
@ -69,30 +69,30 @@ void MLPPSoftmaxReg::gradientDescent(double learning_rate, int max_epoch, bool U
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPSoftmaxReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
void MLPPSoftmaxReg::SGD(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
std::random_device rd;
|
std::random_device rd;
|
||||||
std::default_random_engine generator(rd());
|
std::default_random_engine generator(rd());
|
||||||
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
||||||
double outputIndex = distribution(generator);
|
real_t outputIndex = distribution(generator);
|
||||||
|
|
||||||
std::vector<double> y_hat = Evaluate(inputSet[outputIndex]);
|
std::vector<real_t> y_hat = Evaluate(inputSet[outputIndex]);
|
||||||
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
std::vector<std::vector<double>> w_gradient = alg.outerProduct(inputSet[outputIndex], alg.subtraction(y_hat, outputSet[outputIndex]));
|
std::vector<std::vector<real_t>> w_gradient = alg.outerProduct(inputSet[outputIndex], alg.subtraction(y_hat, outputSet[outputIndex]));
|
||||||
|
|
||||||
// Weight Updation
|
// Weight Updation
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate, w_gradient));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate, w_gradient));
|
||||||
weights = regularization.regWeights(weights, lambda, alpha, reg);
|
weights = regularization.regWeights(weights, lambda, alpha, reg);
|
||||||
|
|
||||||
// Calculating the bias gradients
|
// Calculating the bias gradients
|
||||||
std::vector<double> b_gradient = alg.subtraction(y_hat, outputSet[outputIndex]);
|
std::vector<real_t> b_gradient = alg.subtraction(y_hat, outputSet[outputIndex]);
|
||||||
|
|
||||||
// Bias updation
|
// Bias updation
|
||||||
bias = alg.subtraction(bias, alg.scalarMultiply(learning_rate, b_gradient));
|
bias = alg.subtraction(bias, alg.scalarMultiply(learning_rate, b_gradient));
|
||||||
@ -112,10 +112,10 @@ void MLPPSoftmaxReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPSoftmaxReg::MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
void MLPPSoftmaxReg::MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
@ -124,13 +124,13 @@ void MLPPSoftmaxReg::MBGD(double learning_rate, int max_epoch, int mini_batch_si
|
|||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<std::vector<double>> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<std::vector<real_t>> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
std::vector<std::vector<double>> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
std::vector<std::vector<real_t>> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
std::vector<std::vector<double>> w_gradient = alg.matmult(alg.transpose(inputMiniBatches[i]), error);
|
std::vector<std::vector<real_t>> w_gradient = alg.matmult(alg.transpose(inputMiniBatches[i]), error);
|
||||||
|
|
||||||
//Weight updation
|
//Weight updation
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate, w_gradient));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate, w_gradient));
|
||||||
@ -153,7 +153,7 @@ void MLPPSoftmaxReg::MBGD(double learning_rate, int max_epoch, int mini_batch_si
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPSoftmaxReg::score() {
|
real_t MLPPSoftmaxReg::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
return util.performance(y_hat, outputSet);
|
return util.performance(y_hat, outputSet);
|
||||||
}
|
}
|
||||||
@ -163,19 +163,19 @@ void MLPPSoftmaxReg::save(std::string fileName) {
|
|||||||
util.saveParameters(fileName, weights, bias);
|
util.saveParameters(fileName, weights, bias);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPSoftmaxReg::Cost(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
real_t MLPPSoftmaxReg::Cost(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
return cost.CrossEntropy(y_hat, y) + regularization.regTerm(weights, lambda, alpha, reg);
|
return cost.CrossEntropy(y_hat, y) + regularization.regTerm(weights, lambda, alpha, reg);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPSoftmaxReg::Evaluate(std::vector<double> x) {
|
std::vector<real_t> MLPPSoftmaxReg::Evaluate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
return avn.softmax(alg.addition(bias, alg.mat_vec_mult(alg.transpose(weights), x)));
|
return avn.softmax(alg.addition(bias, alg.mat_vec_mult(alg.transpose(weights), x)));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPSoftmaxReg::Evaluate(std::vector<std::vector<double>> X) {
|
std::vector<std::vector<real_t>> MLPPSoftmaxReg::Evaluate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
|
|
||||||
|
@ -8,6 +8,8 @@
|
|||||||
// Created by Marc Melikyan on 10/2/20.
|
// Created by Marc Melikyan on 10/2/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
@ -15,27 +17,27 @@
|
|||||||
|
|
||||||
class MLPPSoftmaxReg {
|
class MLPPSoftmaxReg {
|
||||||
public:
|
public:
|
||||||
MLPPSoftmaxReg(std::vector<std::vector<double>> inputSet, std::vector<std::vector<double>> outputSet, std::string reg = "None", double lambda = 0.5, double alpha = 0.5);
|
MLPPSoftmaxReg(std::vector<std::vector<real_t>> inputSet, std::vector<std::vector<real_t>> outputSet, std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
||||||
std::vector<double> modelTest(std::vector<double> x);
|
std::vector<real_t> modelTest(std::vector<real_t> x);
|
||||||
std::vector<std::vector<double>> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<std::vector<real_t>> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
void gradientDescent(double learning_rate, int max_epoch, bool UI = 1);
|
void gradientDescent(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void SGD(double learning_rate, int max_epoch, bool UI = 1);
|
void SGD(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
void MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
||||||
double score();
|
real_t score();
|
||||||
void save(std::string fileName);
|
void save(std::string fileName);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
double Cost(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
real_t Cost(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
std::vector<std::vector<double>> Evaluate(std::vector<std::vector<double>> X);
|
std::vector<std::vector<real_t>> Evaluate(std::vector<std::vector<real_t>> X);
|
||||||
std::vector<double> Evaluate(std::vector<double> x);
|
std::vector<real_t> Evaluate(std::vector<real_t> x);
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<std::vector<double>> outputSet;
|
std::vector<std::vector<real_t>> outputSet;
|
||||||
std::vector<std::vector<double>> y_hat;
|
std::vector<std::vector<real_t>> y_hat;
|
||||||
std::vector<std::vector<double>> weights;
|
std::vector<std::vector<real_t>> weights;
|
||||||
std::vector<double> bias;
|
std::vector<real_t> bias;
|
||||||
|
|
||||||
int n;
|
int n;
|
||||||
int k;
|
int k;
|
||||||
@ -43,8 +45,8 @@ private:
|
|||||||
|
|
||||||
// Regularization Params
|
// Regularization Params
|
||||||
std::string reg;
|
std::string reg;
|
||||||
double lambda;
|
real_t lambda;
|
||||||
double alpha; /* This is the controlling param for Elastic Net*/
|
real_t alpha; /* This is the controlling param for Elastic Net*/
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -15,24 +15,24 @@
|
|||||||
#include <iostream>
|
#include <iostream>
|
||||||
|
|
||||||
|
|
||||||
double MLPPStat::b0Estimation(const std::vector<double> &x, const std::vector<double> &y) {
|
real_t MLPPStat::b0Estimation(const std::vector<real_t> &x, const std::vector<real_t> &y) {
|
||||||
return mean(y) - b1Estimation(x, y) * mean(x);
|
return mean(y) - b1Estimation(x, y) * mean(x);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::b1Estimation(const std::vector<double> &x, const std::vector<double> &y) {
|
real_t MLPPStat::b1Estimation(const std::vector<real_t> &x, const std::vector<real_t> &y) {
|
||||||
return covariance(x, y) / variance(x);
|
return covariance(x, y) / variance(x);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::mean(const std::vector<double> &x) {
|
real_t MLPPStat::mean(const std::vector<real_t> &x) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < x.size(); i++) {
|
for (int i = 0; i < x.size(); i++) {
|
||||||
sum += x[i];
|
sum += x[i];
|
||||||
}
|
}
|
||||||
return sum / x.size();
|
return sum / x.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::median(std::vector<double> x) {
|
real_t MLPPStat::median(std::vector<real_t> x) {
|
||||||
double center = double(x.size()) / double(2);
|
real_t center = real_t(x.size()) / real_t(2);
|
||||||
sort(x.begin(), x.end());
|
sort(x.begin(), x.end());
|
||||||
if (x.size() % 2 == 0) {
|
if (x.size() % 2 == 0) {
|
||||||
return mean({ x[center - 1], x[center] });
|
return mean({ x[center - 1], x[center] });
|
||||||
@ -41,18 +41,18 @@ double MLPPStat::median(std::vector<double> x) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPStat::mode(const std::vector<double> &x) {
|
std::vector<real_t> MLPPStat::mode(const std::vector<real_t> &x) {
|
||||||
MLPPData data;
|
MLPPData data;
|
||||||
std::vector<double> x_set = data.vecToSet(x);
|
std::vector<real_t> x_set = data.vecToSet(x);
|
||||||
std::map<double, int> element_num;
|
std::map<real_t, int> element_num;
|
||||||
for (int i = 0; i < x_set.size(); i++) {
|
for (int i = 0; i < x_set.size(); i++) {
|
||||||
element_num[x[i]] = 0;
|
element_num[x[i]] = 0;
|
||||||
}
|
}
|
||||||
for (int i = 0; i < x.size(); i++) {
|
for (int i = 0; i < x.size(); i++) {
|
||||||
element_num[x[i]]++;
|
element_num[x[i]]++;
|
||||||
}
|
}
|
||||||
std::vector<double> modes;
|
std::vector<real_t> modes;
|
||||||
double max_num = element_num[x_set[0]];
|
real_t max_num = element_num[x_set[0]];
|
||||||
for (int i = 0; i < x_set.size(); i++) {
|
for (int i = 0; i < x_set.size(); i++) {
|
||||||
if (element_num[x_set[i]] > max_num) {
|
if (element_num[x_set[i]] > max_num) {
|
||||||
max_num = element_num[x_set[i]];
|
max_num = element_num[x_set[i]];
|
||||||
@ -65,59 +65,59 @@ std::vector<double> MLPPStat::mode(const std::vector<double> &x) {
|
|||||||
return modes;
|
return modes;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::range(const std::vector<double> &x) {
|
real_t MLPPStat::range(const std::vector<real_t> &x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.max(x) - alg.min(x);
|
return alg.max(x) - alg.min(x);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::midrange(const std::vector<double> &x) {
|
real_t MLPPStat::midrange(const std::vector<real_t> &x) {
|
||||||
return range(x) / 2;
|
return range(x) / 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::absAvgDeviation(const std::vector<double> &x) {
|
real_t MLPPStat::absAvgDeviation(const std::vector<real_t> &x) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < x.size(); i++) {
|
for (int i = 0; i < x.size(); i++) {
|
||||||
sum += std::abs(x[i] - mean(x));
|
sum += std::abs(x[i] - mean(x));
|
||||||
}
|
}
|
||||||
return sum / x.size();
|
return sum / x.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::standardDeviation(const std::vector<double> &x) {
|
real_t MLPPStat::standardDeviation(const std::vector<real_t> &x) {
|
||||||
return std::sqrt(variance(x));
|
return std::sqrt(variance(x));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::variance(const std::vector<double> &x) {
|
real_t MLPPStat::variance(const std::vector<real_t> &x) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < x.size(); i++) {
|
for (int i = 0; i < x.size(); i++) {
|
||||||
sum += (x[i] - mean(x)) * (x[i] - mean(x));
|
sum += (x[i] - mean(x)) * (x[i] - mean(x));
|
||||||
}
|
}
|
||||||
return sum / (x.size() - 1);
|
return sum / (x.size() - 1);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::covariance(const std::vector<double> &x, const std::vector<double> &y) {
|
real_t MLPPStat::covariance(const std::vector<real_t> &x, const std::vector<real_t> &y) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < x.size(); i++) {
|
for (int i = 0; i < x.size(); i++) {
|
||||||
sum += (x[i] - mean(x)) * (y[i] - mean(y));
|
sum += (x[i] - mean(x)) * (y[i] - mean(y));
|
||||||
}
|
}
|
||||||
return sum / (x.size() - 1);
|
return sum / (x.size() - 1);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::correlation(const std::vector<double> &x, const std::vector<double> &y) {
|
real_t MLPPStat::correlation(const std::vector<real_t> &x, const std::vector<real_t> &y) {
|
||||||
return covariance(x, y) / (standardDeviation(x) * standardDeviation(y));
|
return covariance(x, y) / (standardDeviation(x) * standardDeviation(y));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::R2(const std::vector<double> &x, const std::vector<double> &y) {
|
real_t MLPPStat::R2(const std::vector<real_t> &x, const std::vector<real_t> &y) {
|
||||||
return correlation(x, y) * correlation(x, y);
|
return correlation(x, y) * correlation(x, y);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::chebyshevIneq(const double k) {
|
real_t MLPPStat::chebyshevIneq(const real_t k) {
|
||||||
// X may or may not belong to a Gaussian Distribution
|
// X may or may not belong to a Gaussian Distribution
|
||||||
return 1 - 1 / (k * k);
|
return 1 - 1 / (k * k);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::weightedMean(const std::vector<double> &x, const std::vector<double> &weights) {
|
real_t MLPPStat::weightedMean(const std::vector<real_t> &x, const std::vector<real_t> &weights) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
double weights_sum = 0;
|
real_t weights_sum = 0;
|
||||||
for (int i = 0; i < x.size(); i++) {
|
for (int i = 0; i < x.size(); i++) {
|
||||||
sum += x[i] * weights[i];
|
sum += x[i] * weights[i];
|
||||||
weights_sum += weights[i];
|
weights_sum += weights[i];
|
||||||
@ -125,41 +125,41 @@ double MLPPStat::weightedMean(const std::vector<double> &x, const std::vector<do
|
|||||||
return sum / weights_sum;
|
return sum / weights_sum;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::geometricMean(const std::vector<double> &x) {
|
real_t MLPPStat::geometricMean(const std::vector<real_t> &x) {
|
||||||
double product = 1;
|
real_t product = 1;
|
||||||
for (int i = 0; i < x.size(); i++) {
|
for (int i = 0; i < x.size(); i++) {
|
||||||
product *= x[i];
|
product *= x[i];
|
||||||
}
|
}
|
||||||
return std::pow(product, 1.0 / x.size());
|
return std::pow(product, 1.0 / x.size());
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::harmonicMean(const std::vector<double> &x) {
|
real_t MLPPStat::harmonicMean(const std::vector<real_t> &x) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < x.size(); i++) {
|
for (int i = 0; i < x.size(); i++) {
|
||||||
sum += 1 / x[i];
|
sum += 1 / x[i];
|
||||||
}
|
}
|
||||||
return x.size() / sum;
|
return x.size() / sum;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::RMS(const std::vector<double> &x) {
|
real_t MLPPStat::RMS(const std::vector<real_t> &x) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < x.size(); i++) {
|
for (int i = 0; i < x.size(); i++) {
|
||||||
sum += x[i] * x[i];
|
sum += x[i] * x[i];
|
||||||
}
|
}
|
||||||
return sqrt(sum / x.size());
|
return sqrt(sum / x.size());
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::powerMean(const std::vector<double> &x, const double p) {
|
real_t MLPPStat::powerMean(const std::vector<real_t> &x, const real_t p) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
for (int i = 0; i < x.size(); i++) {
|
for (int i = 0; i < x.size(); i++) {
|
||||||
sum += std::pow(x[i], p);
|
sum += std::pow(x[i], p);
|
||||||
}
|
}
|
||||||
return std::pow(sum / x.size(), 1 / p);
|
return std::pow(sum / x.size(), 1 / p);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::lehmerMean(const std::vector<double> &x, const double p) {
|
real_t MLPPStat::lehmerMean(const std::vector<real_t> &x, const real_t p) {
|
||||||
double num = 0;
|
real_t num = 0;
|
||||||
double den = 0;
|
real_t den = 0;
|
||||||
for (int i = 0; i < x.size(); i++) {
|
for (int i = 0; i < x.size(); i++) {
|
||||||
num += std::pow(x[i], p);
|
num += std::pow(x[i], p);
|
||||||
den += std::pow(x[i], p - 1);
|
den += std::pow(x[i], p - 1);
|
||||||
@ -167,9 +167,9 @@ double MLPPStat::lehmerMean(const std::vector<double> &x, const double p) {
|
|||||||
return num / den;
|
return num / den;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::weightedLehmerMean(const std::vector<double> &x, const std::vector<double> &weights, const double p) {
|
real_t MLPPStat::weightedLehmerMean(const std::vector<real_t> &x, const std::vector<real_t> &weights, const real_t p) {
|
||||||
double num = 0;
|
real_t num = 0;
|
||||||
double den = 0;
|
real_t den = 0;
|
||||||
for (int i = 0; i < x.size(); i++) {
|
for (int i = 0; i < x.size(); i++) {
|
||||||
num += weights[i] * std::pow(x[i], p);
|
num += weights[i] * std::pow(x[i], p);
|
||||||
den += weights[i] * std::pow(x[i], p - 1);
|
den += weights[i] * std::pow(x[i], p - 1);
|
||||||
@ -177,38 +177,38 @@ double MLPPStat::weightedLehmerMean(const std::vector<double> &x, const std::vec
|
|||||||
return num / den;
|
return num / den;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::heronianMean(const double A, const double B) {
|
real_t MLPPStat::heronianMean(const real_t A, const real_t B) {
|
||||||
return (A + sqrt(A * B) + B) / 3;
|
return (A + sqrt(A * B) + B) / 3;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::contraHarmonicMean(const std::vector<double> &x) {
|
real_t MLPPStat::contraHarmonicMean(const std::vector<real_t> &x) {
|
||||||
return lehmerMean(x, 2);
|
return lehmerMean(x, 2);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::heinzMean(const double A, const double B, const double x) {
|
real_t MLPPStat::heinzMean(const real_t A, const real_t B, const real_t x) {
|
||||||
return (std::pow(A, x) * std::pow(B, 1 - x) + std::pow(A, 1 - x) * std::pow(B, x)) / 2;
|
return (std::pow(A, x) * std::pow(B, 1 - x) + std::pow(A, 1 - x) * std::pow(B, x)) / 2;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::neumanSandorMean(const double a, const double b) {
|
real_t MLPPStat::neumanSandorMean(const real_t a, const real_t b) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
return (a - b) / 2 * avn.arsinh((a - b) / (a + b));
|
return (a - b) / 2 * avn.arsinh((a - b) / (a + b));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::stolarskyMean(const double x, const double y, const double p) {
|
real_t MLPPStat::stolarskyMean(const real_t x, const real_t y, const real_t p) {
|
||||||
if (x == y) {
|
if (x == y) {
|
||||||
return x;
|
return x;
|
||||||
}
|
}
|
||||||
return std::pow((std::pow(x, p) - std::pow(y, p)) / (p * (x - y)), 1 / (p - 1));
|
return std::pow((std::pow(x, p) - std::pow(y, p)) / (p * (x - y)), 1 / (p - 1));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::identricMean(const double x, const double y) {
|
real_t MLPPStat::identricMean(const real_t x, const real_t y) {
|
||||||
if (x == y) {
|
if (x == y) {
|
||||||
return x;
|
return x;
|
||||||
}
|
}
|
||||||
return (1 / M_E) * std::pow(std::pow(x, x) / std::pow(y, y), 1 / (x - y));
|
return (1 / M_E) * std::pow(std::pow(x, x) / std::pow(y, y), 1 / (x - y));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPStat::logMean(const double x, const double y) {
|
real_t MLPPStat::logMean(const real_t x, const real_t y) {
|
||||||
if (x == y) {
|
if (x == y) {
|
||||||
return x;
|
return x;
|
||||||
}
|
}
|
||||||
|
@ -8,44 +8,46 @@
|
|||||||
// Created by Marc Melikyan on 9/29/20.
|
// Created by Marc Melikyan on 9/29/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
|
|
||||||
class MLPPStat {
|
class MLPPStat {
|
||||||
public:
|
public:
|
||||||
// These functions are for univariate lin reg module- not for users.
|
// These functions are for univariate lin reg module- not for users.
|
||||||
double b0Estimation(const std::vector<double> &x, const std::vector<double> &y);
|
real_t b0Estimation(const std::vector<real_t> &x, const std::vector<real_t> &y);
|
||||||
double b1Estimation(const std::vector<double> &x, const std::vector<double> &y);
|
real_t b1Estimation(const std::vector<real_t> &x, const std::vector<real_t> &y);
|
||||||
|
|
||||||
// Statistical Functions
|
// Statistical Functions
|
||||||
double mean(const std::vector<double> &x);
|
real_t mean(const std::vector<real_t> &x);
|
||||||
double median(std::vector<double> x);
|
real_t median(std::vector<real_t> x);
|
||||||
std::vector<double> mode(const std::vector<double> &x);
|
std::vector<real_t> mode(const std::vector<real_t> &x);
|
||||||
double range(const std::vector<double> &x);
|
real_t range(const std::vector<real_t> &x);
|
||||||
double midrange(const std::vector<double> &x);
|
real_t midrange(const std::vector<real_t> &x);
|
||||||
double absAvgDeviation(const std::vector<double> &x);
|
real_t absAvgDeviation(const std::vector<real_t> &x);
|
||||||
double standardDeviation(const std::vector<double> &x);
|
real_t standardDeviation(const std::vector<real_t> &x);
|
||||||
double variance(const std::vector<double> &x);
|
real_t variance(const std::vector<real_t> &x);
|
||||||
double covariance(const std::vector<double> &x, const std::vector<double> &y);
|
real_t covariance(const std::vector<real_t> &x, const std::vector<real_t> &y);
|
||||||
double correlation(const std::vector<double> &x, const std::vector<double> &y);
|
real_t correlation(const std::vector<real_t> &x, const std::vector<real_t> &y);
|
||||||
double R2(const std::vector<double> &x, const std::vector<double> &y);
|
real_t R2(const std::vector<real_t> &x, const std::vector<real_t> &y);
|
||||||
double chebyshevIneq(const double k);
|
real_t chebyshevIneq(const real_t k);
|
||||||
|
|
||||||
// Extras
|
// Extras
|
||||||
double weightedMean(const std::vector<double> &x, const std::vector<double> &weights);
|
real_t weightedMean(const std::vector<real_t> &x, const std::vector<real_t> &weights);
|
||||||
double geometricMean(const std::vector<double> &x);
|
real_t geometricMean(const std::vector<real_t> &x);
|
||||||
double harmonicMean(const std::vector<double> &x);
|
real_t harmonicMean(const std::vector<real_t> &x);
|
||||||
double RMS(const std::vector<double> &x);
|
real_t RMS(const std::vector<real_t> &x);
|
||||||
double powerMean(const std::vector<double> &x, const double p);
|
real_t powerMean(const std::vector<real_t> &x, const real_t p);
|
||||||
double lehmerMean(const std::vector<double> &x, const double p);
|
real_t lehmerMean(const std::vector<real_t> &x, const real_t p);
|
||||||
double weightedLehmerMean(const std::vector<double> &x, const std::vector<double> &weights, const double p);
|
real_t weightedLehmerMean(const std::vector<real_t> &x, const std::vector<real_t> &weights, const real_t p);
|
||||||
double contraHarmonicMean(const std::vector<double> &x);
|
real_t contraHarmonicMean(const std::vector<real_t> &x);
|
||||||
double heronianMean(const double A, const double B);
|
real_t heronianMean(const real_t A, const real_t B);
|
||||||
double heinzMean(const double A, const double B, const double x);
|
real_t heinzMean(const real_t A, const real_t B, const real_t x);
|
||||||
double neumanSandorMean(const double a, const double b);
|
real_t neumanSandorMean(const real_t a, const real_t b);
|
||||||
double stolarskyMean(const double x, const double y, const double p);
|
real_t stolarskyMean(const real_t x, const real_t y, const real_t p);
|
||||||
double identricMean(const double x, const double y);
|
real_t identricMean(const real_t x, const real_t y);
|
||||||
double logMean(const double x, const double y);
|
real_t logMean(const real_t x, const real_t y);
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -15,27 +15,27 @@
|
|||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
|
|
||||||
MLPPSVC::MLPPSVC(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, double C) :
|
MLPPSVC::MLPPSVC(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, real_t C) :
|
||||||
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), C(C) {
|
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), C(C) {
|
||||||
y_hat.resize(n);
|
y_hat.resize(n);
|
||||||
weights = MLPPUtilities::weightInitialization(k);
|
weights = MLPPUtilities::weightInitialization(k);
|
||||||
bias = MLPPUtilities::biasInitialization();
|
bias = MLPPUtilities::biasInitialization();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPSVC::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPSVC::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
return Evaluate(X);
|
return Evaluate(X);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPSVC::modelTest(std::vector<double> x) {
|
real_t MLPPSVC::modelTest(std::vector<real_t> x) {
|
||||||
return Evaluate(x);
|
return Evaluate(x);
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPSVC::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
void MLPPSVC::gradientDescent(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
@ -63,13 +63,13 @@ void MLPPSVC::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPSVC::SGD(double learning_rate, int max_epoch, bool UI) {
|
void MLPPSVC::SGD(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
|
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
@ -78,11 +78,11 @@ void MLPPSVC::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
||||||
int outputIndex = distribution(generator);
|
int outputIndex = distribution(generator);
|
||||||
|
|
||||||
double y_hat = Evaluate(inputSet[outputIndex]);
|
real_t y_hat = Evaluate(inputSet[outputIndex]);
|
||||||
double z = propagate(inputSet[outputIndex]);
|
real_t z = propagate(inputSet[outputIndex]);
|
||||||
cost_prev = Cost({ z }, { outputSet[outputIndex] }, weights, C);
|
cost_prev = Cost({ z }, { outputSet[outputIndex] }, weights, C);
|
||||||
|
|
||||||
double costDeriv = cost.HingeLossDeriv(std::vector<double>({ z }), std::vector<double>({ outputSet[outputIndex] }), C)[0]; // Explicit conversion to avoid ambiguity with overloaded function. Error occured on Ubuntu.
|
real_t costDeriv = cost.HingeLossDeriv(std::vector<real_t>({ z }), std::vector<real_t>({ outputSet[outputIndex] }), C)[0]; // Explicit conversion to avoid ambiguity with overloaded function. Error occured on Ubuntu.
|
||||||
|
|
||||||
// Weight Updation
|
// Weight Updation
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate * costDeriv, inputSet[outputIndex]));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate * costDeriv, inputSet[outputIndex]));
|
||||||
@ -106,12 +106,12 @@ void MLPPSVC::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPSVC::MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
void MLPPSVC::MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
@ -120,8 +120,8 @@ void MLPPSVC::MBGD(double learning_rate, int max_epoch, int mini_batch_size, boo
|
|||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
std::vector<double> z = propagate(inputMiniBatches[i]);
|
std::vector<real_t> z = propagate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(z, outputMiniBatches[i], weights, C);
|
cost_prev = Cost(z, outputMiniBatches[i], weights, C);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
@ -148,7 +148,7 @@ void MLPPSVC::MBGD(double learning_rate, int max_epoch, int mini_batch_size, boo
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPSVC::score() {
|
real_t MLPPSVC::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
return util.performance(y_hat, outputSet);
|
return util.performance(y_hat, outputSet);
|
||||||
}
|
}
|
||||||
@ -158,30 +158,30 @@ void MLPPSVC::save(std::string fileName) {
|
|||||||
util.saveParameters(fileName, weights, bias);
|
util.saveParameters(fileName, weights, bias);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPSVC::Cost(std::vector<double> z, std::vector<double> y, std::vector<double> weights, double C) {
|
real_t MLPPSVC::Cost(std::vector<real_t> z, std::vector<real_t> y, std::vector<real_t> weights, real_t C) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
return cost.HingeLoss(z, y, weights, C);
|
return cost.HingeLoss(z, y, weights, C);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPSVC::Evaluate(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPSVC::Evaluate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
return avn.sign(alg.scalarAdd(bias, alg.mat_vec_mult(X, weights)));
|
return avn.sign(alg.scalarAdd(bias, alg.mat_vec_mult(X, weights)));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPSVC::propagate(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPSVC::propagate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
return alg.scalarAdd(bias, alg.mat_vec_mult(X, weights));
|
return alg.scalarAdd(bias, alg.mat_vec_mult(X, weights));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPSVC::Evaluate(std::vector<double> x) {
|
real_t MLPPSVC::Evaluate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
return avn.sign(alg.dot(weights, x) + bias);
|
return avn.sign(alg.dot(weights, x) + bias);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPSVC::propagate(std::vector<double> x) {
|
real_t MLPPSVC::propagate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
return alg.dot(weights, x) + bias;
|
return alg.dot(weights, x) + bias;
|
||||||
|
@ -11,6 +11,8 @@
|
|||||||
// https://towardsdatascience.com/svm-implementation-from-scratch-python-2db2fc52e5c2
|
// https://towardsdatascience.com/svm-implementation-from-scratch-python-2db2fc52e5c2
|
||||||
// Illustratd a practical definition of the Hinge Loss function and its gradient when optimizing with SGD.
|
// Illustratd a practical definition of the Hinge Loss function and its gradient when optimizing with SGD.
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
@ -18,37 +20,37 @@
|
|||||||
|
|
||||||
class MLPPSVC {
|
class MLPPSVC {
|
||||||
public:
|
public:
|
||||||
MLPPSVC(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, double C);
|
MLPPSVC(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, real_t C);
|
||||||
std::vector<double> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<real_t> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
double modelTest(std::vector<double> x);
|
real_t modelTest(std::vector<real_t> x);
|
||||||
void gradientDescent(double learning_rate, int max_epoch, bool UI = 1);
|
void gradientDescent(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void SGD(double learning_rate, int max_epoch, bool UI = 1);
|
void SGD(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
void MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
||||||
double score();
|
real_t score();
|
||||||
void save(std::string fileName);
|
void save(std::string fileName);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
double Cost(std::vector<double> y_hat, std::vector<double> y, std::vector<double> weights, double C);
|
real_t Cost(std::vector<real_t> y_hat, std::vector<real_t> y, std::vector<real_t> weights, real_t C);
|
||||||
|
|
||||||
std::vector<double> Evaluate(std::vector<std::vector<double>> X);
|
std::vector<real_t> Evaluate(std::vector<std::vector<real_t>> X);
|
||||||
std::vector<double> propagate(std::vector<std::vector<double>> X);
|
std::vector<real_t> propagate(std::vector<std::vector<real_t>> X);
|
||||||
double Evaluate(std::vector<double> x);
|
real_t Evaluate(std::vector<real_t> x);
|
||||||
double propagate(std::vector<double> x);
|
real_t propagate(std::vector<real_t> x);
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
std::vector<double> z;
|
std::vector<real_t> z;
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
std::vector<double> weights;
|
std::vector<real_t> weights;
|
||||||
double bias;
|
real_t bias;
|
||||||
|
|
||||||
double C;
|
real_t C;
|
||||||
int n;
|
int n;
|
||||||
int k;
|
int k;
|
||||||
|
|
||||||
// UI Portion
|
// UI Portion
|
||||||
void UI(int epoch, double cost_prev);
|
void UI(int epoch, real_t cost_prev);
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -15,33 +15,33 @@
|
|||||||
#include <random>
|
#include <random>
|
||||||
|
|
||||||
|
|
||||||
MLPPTanhReg::MLPPTanhReg(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, std::string reg, double lambda, double alpha) :
|
MLPPTanhReg::MLPPTanhReg(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, std::string reg, real_t lambda, real_t alpha) :
|
||||||
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
inputSet(inputSet), outputSet(outputSet), n(inputSet.size()), k(inputSet[0].size()), reg(reg), lambda(lambda), alpha(alpha) {
|
||||||
y_hat.resize(n);
|
y_hat.resize(n);
|
||||||
weights = MLPPUtilities::weightInitialization(k);
|
weights = MLPPUtilities::weightInitialization(k);
|
||||||
bias = MLPPUtilities::biasInitialization();
|
bias = MLPPUtilities::biasInitialization();
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPTanhReg::modelSetTest(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPTanhReg::modelSetTest(std::vector<std::vector<real_t>> X) {
|
||||||
return Evaluate(X);
|
return Evaluate(X);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPTanhReg::modelTest(std::vector<double> x) {
|
real_t MLPPTanhReg::modelTest(std::vector<real_t> x) {
|
||||||
return Evaluate(x);
|
return Evaluate(x);
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPTanhReg::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
void MLPPTanhReg::gradientDescent(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
cost_prev = Cost(y_hat, outputSet);
|
cost_prev = Cost(y_hat, outputSet);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputSet);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputSet);
|
||||||
|
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputSet), alg.hadamard_product(error, avn.tanh(z, 1)))));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputSet), alg.hadamard_product(error, avn.tanh(z, 1)))));
|
||||||
weights = regularization.regWeights(weights, lambda, alpha, reg);
|
weights = regularization.regWeights(weights, lambda, alpha, reg);
|
||||||
@ -64,10 +64,10 @@ void MLPPTanhReg::gradientDescent(double learning_rate, int max_epoch, bool UI)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPTanhReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
void MLPPTanhReg::SGD(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
@ -76,10 +76,10 @@ void MLPPTanhReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
std::uniform_int_distribution<int> distribution(0, int(n - 1));
|
||||||
int outputIndex = distribution(generator);
|
int outputIndex = distribution(generator);
|
||||||
|
|
||||||
double y_hat = Evaluate(inputSet[outputIndex]);
|
real_t y_hat = Evaluate(inputSet[outputIndex]);
|
||||||
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
cost_prev = Cost({ y_hat }, { outputSet[outputIndex] });
|
||||||
|
|
||||||
double error = y_hat - outputSet[outputIndex];
|
real_t error = y_hat - outputSet[outputIndex];
|
||||||
|
|
||||||
// Weight Updation
|
// Weight Updation
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate * error * (1 - y_hat * y_hat), inputSet[outputIndex]));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate * error * (1 - y_hat * y_hat), inputSet[outputIndex]));
|
||||||
@ -103,11 +103,11 @@ void MLPPTanhReg::SGD(double learning_rate, int max_epoch, bool UI) {
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPTanhReg::MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
void MLPPTanhReg::MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI) {
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
@ -116,11 +116,11 @@ void MLPPTanhReg::MBGD(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<double> y_hat = Evaluate(inputMiniBatches[i]);
|
std::vector<real_t> y_hat = Evaluate(inputMiniBatches[i]);
|
||||||
std::vector<double> z = propagate(inputMiniBatches[i]);
|
std::vector<real_t> z = propagate(inputMiniBatches[i]);
|
||||||
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
cost_prev = Cost(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
std::vector<double> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
std::vector<real_t> error = alg.subtraction(y_hat, outputMiniBatches[i]);
|
||||||
|
|
||||||
// Calculating the weight gradients
|
// Calculating the weight gradients
|
||||||
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), alg.hadamard_product(error, avn.tanh(z, 1)))));
|
weights = alg.subtraction(weights, alg.scalarMultiply(learning_rate / n, alg.mat_vec_mult(alg.transpose(inputMiniBatches[i]), alg.hadamard_product(error, avn.tanh(z, 1)))));
|
||||||
@ -146,7 +146,7 @@ void MLPPTanhReg::MBGD(double learning_rate, int max_epoch, int mini_batch_size,
|
|||||||
forwardPass();
|
forwardPass();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPTanhReg::score() {
|
real_t MLPPTanhReg::score() {
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
return util.performance(y_hat, outputSet);
|
return util.performance(y_hat, outputSet);
|
||||||
}
|
}
|
||||||
@ -156,30 +156,30 @@ void MLPPTanhReg::save(std::string fileName) {
|
|||||||
util.saveParameters(fileName, weights, bias);
|
util.saveParameters(fileName, weights, bias);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPTanhReg::Cost(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPTanhReg::Cost(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
return cost.MSE(y_hat, y) + regularization.regTerm(weights, lambda, alpha, reg);
|
return cost.MSE(y_hat, y) + regularization.regTerm(weights, lambda, alpha, reg);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPTanhReg::Evaluate(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPTanhReg::Evaluate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
return avn.tanh(alg.scalarAdd(bias, alg.mat_vec_mult(X, weights)));
|
return avn.tanh(alg.scalarAdd(bias, alg.mat_vec_mult(X, weights)));
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPTanhReg::propagate(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPTanhReg::propagate(std::vector<std::vector<real_t>> X) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.scalarAdd(bias, alg.mat_vec_mult(X, weights));
|
return alg.scalarAdd(bias, alg.mat_vec_mult(X, weights));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPTanhReg::Evaluate(std::vector<double> x) {
|
real_t MLPPTanhReg::Evaluate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
return avn.tanh(alg.dot(weights, x) + bias);
|
return avn.tanh(alg.dot(weights, x) + bias);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPTanhReg::propagate(std::vector<double> x) {
|
real_t MLPPTanhReg::propagate(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.dot(weights, x) + bias;
|
return alg.dot(weights, x) + bias;
|
||||||
}
|
}
|
||||||
|
@ -8,6 +8,8 @@
|
|||||||
// Created by Marc Melikyan on 10/2/20.
|
// Created by Marc Melikyan on 10/2/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
@ -15,41 +17,41 @@
|
|||||||
|
|
||||||
class MLPPTanhReg {
|
class MLPPTanhReg {
|
||||||
public:
|
public:
|
||||||
MLPPTanhReg(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, std::string reg = "None", double lambda = 0.5, double alpha = 0.5);
|
MLPPTanhReg(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
||||||
std::vector<double> modelSetTest(std::vector<std::vector<double>> X);
|
std::vector<real_t> modelSetTest(std::vector<std::vector<real_t>> X);
|
||||||
double modelTest(std::vector<double> x);
|
real_t modelTest(std::vector<real_t> x);
|
||||||
void gradientDescent(double learning_rate, int max_epoch, bool UI = 1);
|
void gradientDescent(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void SGD(double learning_rate, int max_epoch, bool UI = 1);
|
void SGD(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
void MBGD(double learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
void MBGD(real_t learning_rate, int max_epoch, int mini_batch_size, bool UI = 1);
|
||||||
double score();
|
real_t score();
|
||||||
void save(std::string fileName);
|
void save(std::string fileName);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
double Cost(std::vector<double> y_hat, std::vector<double> y);
|
real_t Cost(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
|
|
||||||
std::vector<double> Evaluate(std::vector<std::vector<double>> X);
|
std::vector<real_t> Evaluate(std::vector<std::vector<real_t>> X);
|
||||||
std::vector<double> propagate(std::vector<std::vector<double>> X);
|
std::vector<real_t> propagate(std::vector<std::vector<real_t>> X);
|
||||||
double Evaluate(std::vector<double> x);
|
real_t Evaluate(std::vector<real_t> x);
|
||||||
double propagate(std::vector<double> x);
|
real_t propagate(std::vector<real_t> x);
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet;
|
std::vector<std::vector<real_t>> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
std::vector<double> z;
|
std::vector<real_t> z;
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
std::vector<double> weights;
|
std::vector<real_t> weights;
|
||||||
double bias;
|
real_t bias;
|
||||||
|
|
||||||
int n;
|
int n;
|
||||||
int k;
|
int k;
|
||||||
|
|
||||||
// UI Portion
|
// UI Portion
|
||||||
void UI(int epoch, double cost_prev);
|
void UI(int epoch, real_t cost_prev);
|
||||||
|
|
||||||
// Regularization Params
|
// Regularization Params
|
||||||
std::string reg;
|
std::string reg;
|
||||||
double lambda;
|
real_t lambda;
|
||||||
double alpha; /* This is the controlling param for Elastic Net*/
|
real_t alpha; /* This is the controlling param for Elastic Net*/
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -14,11 +14,11 @@
|
|||||||
|
|
||||||
// DCT ii.
|
// DCT ii.
|
||||||
// https://www.mathworks.com/help/images/discrete-cosine-transform.html
|
// https://www.mathworks.com/help/images/discrete-cosine-transform.html
|
||||||
std::vector<std::vector<double>> MLPPTransforms::discreteCosineTransform(std::vector<std::vector<double>> A) {
|
std::vector<std::vector<real_t>> MLPPTransforms::discreteCosineTransform(std::vector<std::vector<real_t>> A) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
A = alg.scalarAdd(-128, A); // Center around 0.
|
A = alg.scalarAdd(-128, A); // Center around 0.
|
||||||
|
|
||||||
std::vector<std::vector<double>> B;
|
std::vector<std::vector<real_t>> B;
|
||||||
B.resize(A.size());
|
B.resize(A.size());
|
||||||
for (int i = 0; i < B.size(); i++) {
|
for (int i = 0; i < B.size(); i++) {
|
||||||
B[i].resize(A[i].size());
|
B[i].resize(A[i].size());
|
||||||
@ -28,18 +28,18 @@ std::vector<std::vector<double>> MLPPTransforms::discreteCosineTransform(std::ve
|
|||||||
|
|
||||||
for (int i = 0; i < B.size(); i++) {
|
for (int i = 0; i < B.size(); i++) {
|
||||||
for (int j = 0; j < B[i].size(); j++) {
|
for (int j = 0; j < B[i].size(); j++) {
|
||||||
double sum = 0;
|
real_t sum = 0;
|
||||||
double alphaI;
|
real_t alphaI;
|
||||||
if (i == 0) {
|
if (i == 0) {
|
||||||
alphaI = 1 / std::sqrt(M);
|
alphaI = 1 / std::sqrt(M);
|
||||||
} else {
|
} else {
|
||||||
alphaI = std::sqrt(double(2) / double(M));
|
alphaI = std::sqrt(real_t(2) / real_t(M));
|
||||||
}
|
}
|
||||||
double alphaJ;
|
real_t alphaJ;
|
||||||
if (j == 0) {
|
if (j == 0) {
|
||||||
alphaJ = 1 / std::sqrt(M);
|
alphaJ = 1 / std::sqrt(M);
|
||||||
} else {
|
} else {
|
||||||
alphaJ = std::sqrt(double(2) / double(M));
|
alphaJ = std::sqrt(real_t(2) / real_t(M));
|
||||||
}
|
}
|
||||||
|
|
||||||
for (int k = 0; k < B.size(); k++) {
|
for (int k = 0; k < B.size(); k++) {
|
||||||
|
@ -7,13 +7,15 @@
|
|||||||
//
|
//
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
|
|
||||||
class MLPPTransforms {
|
class MLPPTransforms {
|
||||||
public:
|
public:
|
||||||
std::vector<std::vector<double>> discreteCosineTransform(std::vector<std::vector<double>> A);
|
std::vector<std::vector<real_t>> discreteCosineTransform(std::vector<std::vector<real_t>> A);
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -16,19 +16,19 @@
|
|||||||
// ŷ = b0 + b1x1
|
// ŷ = b0 + b1x1
|
||||||
|
|
||||||
|
|
||||||
MLPPUniLinReg::MLPPUniLinReg(std::vector<double> x, std::vector<double> y) :
|
MLPPUniLinReg::MLPPUniLinReg(std::vector<real_t> x, std::vector<real_t> y) :
|
||||||
inputSet(x), outputSet(y) {
|
inputSet(x), outputSet(y) {
|
||||||
MLPPStat estimator;
|
MLPPStat estimator;
|
||||||
b1 = estimator.b1Estimation(inputSet, outputSet);
|
b1 = estimator.b1Estimation(inputSet, outputSet);
|
||||||
b0 = estimator.b0Estimation(inputSet, outputSet);
|
b0 = estimator.b0Estimation(inputSet, outputSet);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPUniLinReg::modelSetTest(std::vector<double> x) {
|
std::vector<real_t> MLPPUniLinReg::modelSetTest(std::vector<real_t> x) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return alg.scalarAdd(b0, alg.scalarMultiply(b1, x));
|
return alg.scalarAdd(b0, alg.scalarMultiply(b1, x));
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPUniLinReg::modelTest(double input) {
|
real_t MLPPUniLinReg::modelTest(real_t input) {
|
||||||
return b0 + b1 * input;
|
return b0 + b1 * input;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -8,21 +8,23 @@
|
|||||||
// Created by Marc Melikyan on 9/29/20.
|
// Created by Marc Melikyan on 9/29/20.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
|
|
||||||
class MLPPUniLinReg {
|
class MLPPUniLinReg {
|
||||||
public:
|
public:
|
||||||
MLPPUniLinReg(std::vector<double> x, std::vector<double> y);
|
MLPPUniLinReg(std::vector<real_t> x, std::vector<real_t> y);
|
||||||
std::vector<double> modelSetTest(std::vector<double> x);
|
std::vector<real_t> modelSetTest(std::vector<real_t> x);
|
||||||
double modelTest(double x);
|
real_t modelTest(real_t x);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
std::vector<double> inputSet;
|
std::vector<real_t> inputSet;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
|
|
||||||
double b0;
|
real_t b0;
|
||||||
double b1;
|
real_t b1;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
||||||
|
@ -12,81 +12,81 @@
|
|||||||
|
|
||||||
|
|
||||||
|
|
||||||
std::vector<double> MLPPUtilities::weightInitialization(int n, std::string type) {
|
std::vector<real_t> MLPPUtilities::weightInitialization(int n, std::string type) {
|
||||||
std::random_device rd;
|
std::random_device rd;
|
||||||
std::default_random_engine generator(rd());
|
std::default_random_engine generator(rd());
|
||||||
|
|
||||||
std::vector<double> weights;
|
std::vector<real_t> weights;
|
||||||
for (int i = 0; i < n; i++) {
|
for (int i = 0; i < n; i++) {
|
||||||
if (type == "XavierNormal") {
|
if (type == "XavierNormal") {
|
||||||
std::normal_distribution<double> distribution(0, sqrt(2 / (n + 1)));
|
std::normal_distribution<real_t> distribution(0, sqrt(2 / (n + 1)));
|
||||||
weights.push_back(distribution(generator));
|
weights.push_back(distribution(generator));
|
||||||
} else if (type == "XavierUniform") {
|
} else if (type == "XavierUniform") {
|
||||||
std::uniform_real_distribution<double> distribution(-sqrt(6 / (n + 1)), sqrt(6 / (n + 1)));
|
std::uniform_real_distribution<real_t> distribution(-sqrt(6 / (n + 1)), sqrt(6 / (n + 1)));
|
||||||
weights.push_back(distribution(generator));
|
weights.push_back(distribution(generator));
|
||||||
} else if (type == "HeNormal") {
|
} else if (type == "HeNormal") {
|
||||||
std::normal_distribution<double> distribution(0, sqrt(2 / n));
|
std::normal_distribution<real_t> distribution(0, sqrt(2 / n));
|
||||||
weights.push_back(distribution(generator));
|
weights.push_back(distribution(generator));
|
||||||
} else if (type == "HeUniform") {
|
} else if (type == "HeUniform") {
|
||||||
std::uniform_real_distribution<double> distribution(-sqrt(6 / n), sqrt(6 / n));
|
std::uniform_real_distribution<real_t> distribution(-sqrt(6 / n), sqrt(6 / n));
|
||||||
weights.push_back(distribution(generator));
|
weights.push_back(distribution(generator));
|
||||||
} else if (type == "LeCunNormal") {
|
} else if (type == "LeCunNormal") {
|
||||||
std::normal_distribution<double> distribution(0, sqrt(1 / n));
|
std::normal_distribution<real_t> distribution(0, sqrt(1 / n));
|
||||||
weights.push_back(distribution(generator));
|
weights.push_back(distribution(generator));
|
||||||
} else if (type == "LeCunUniform") {
|
} else if (type == "LeCunUniform") {
|
||||||
std::uniform_real_distribution<double> distribution(-sqrt(3 / n), sqrt(3 / n));
|
std::uniform_real_distribution<real_t> distribution(-sqrt(3 / n), sqrt(3 / n));
|
||||||
weights.push_back(distribution(generator));
|
weights.push_back(distribution(generator));
|
||||||
} else if (type == "Uniform") {
|
} else if (type == "Uniform") {
|
||||||
std::uniform_real_distribution<double> distribution(-1 / sqrt(n), 1 / sqrt(n));
|
std::uniform_real_distribution<real_t> distribution(-1 / sqrt(n), 1 / sqrt(n));
|
||||||
weights.push_back(distribution(generator));
|
weights.push_back(distribution(generator));
|
||||||
} else {
|
} else {
|
||||||
std::uniform_real_distribution<double> distribution(0, 1);
|
std::uniform_real_distribution<real_t> distribution(0, 1);
|
||||||
weights.push_back(distribution(generator));
|
weights.push_back(distribution(generator));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return weights;
|
return weights;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPUtilities::biasInitialization() {
|
real_t MLPPUtilities::biasInitialization() {
|
||||||
std::random_device rd;
|
std::random_device rd;
|
||||||
std::default_random_engine generator(rd());
|
std::default_random_engine generator(rd());
|
||||||
std::uniform_real_distribution<double> distribution(0, 1);
|
std::uniform_real_distribution<real_t> distribution(0, 1);
|
||||||
|
|
||||||
return distribution(generator);
|
return distribution(generator);
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPUtilities::weightInitialization(int n, int m, std::string type) {
|
std::vector<std::vector<real_t>> MLPPUtilities::weightInitialization(int n, int m, std::string type) {
|
||||||
std::random_device rd;
|
std::random_device rd;
|
||||||
std::default_random_engine generator(rd());
|
std::default_random_engine generator(rd());
|
||||||
|
|
||||||
std::vector<std::vector<double>> weights;
|
std::vector<std::vector<real_t>> weights;
|
||||||
weights.resize(n);
|
weights.resize(n);
|
||||||
|
|
||||||
for (int i = 0; i < n; i++) {
|
for (int i = 0; i < n; i++) {
|
||||||
for (int j = 0; j < m; j++) {
|
for (int j = 0; j < m; j++) {
|
||||||
if (type == "XavierNormal") {
|
if (type == "XavierNormal") {
|
||||||
std::normal_distribution<double> distribution(0, sqrt(2 / (n + m)));
|
std::normal_distribution<real_t> distribution(0, sqrt(2 / (n + m)));
|
||||||
weights[i].push_back(distribution(generator));
|
weights[i].push_back(distribution(generator));
|
||||||
} else if (type == "XavierUniform") {
|
} else if (type == "XavierUniform") {
|
||||||
std::uniform_real_distribution<double> distribution(-sqrt(6 / (n + m)), sqrt(6 / (n + m)));
|
std::uniform_real_distribution<real_t> distribution(-sqrt(6 / (n + m)), sqrt(6 / (n + m)));
|
||||||
weights[i].push_back(distribution(generator));
|
weights[i].push_back(distribution(generator));
|
||||||
} else if (type == "HeNormal") {
|
} else if (type == "HeNormal") {
|
||||||
std::normal_distribution<double> distribution(0, sqrt(2 / n));
|
std::normal_distribution<real_t> distribution(0, sqrt(2 / n));
|
||||||
weights[i].push_back(distribution(generator));
|
weights[i].push_back(distribution(generator));
|
||||||
} else if (type == "HeUniform") {
|
} else if (type == "HeUniform") {
|
||||||
std::uniform_real_distribution<double> distribution(-sqrt(6 / n), sqrt(6 / n));
|
std::uniform_real_distribution<real_t> distribution(-sqrt(6 / n), sqrt(6 / n));
|
||||||
weights[i].push_back(distribution(generator));
|
weights[i].push_back(distribution(generator));
|
||||||
} else if (type == "LeCunNormal") {
|
} else if (type == "LeCunNormal") {
|
||||||
std::normal_distribution<double> distribution(0, sqrt(1 / n));
|
std::normal_distribution<real_t> distribution(0, sqrt(1 / n));
|
||||||
weights[i].push_back(distribution(generator));
|
weights[i].push_back(distribution(generator));
|
||||||
} else if (type == "LeCunUniform") {
|
} else if (type == "LeCunUniform") {
|
||||||
std::uniform_real_distribution<double> distribution(-sqrt(3 / n), sqrt(3 / n));
|
std::uniform_real_distribution<real_t> distribution(-sqrt(3 / n), sqrt(3 / n));
|
||||||
weights[i].push_back(distribution(generator));
|
weights[i].push_back(distribution(generator));
|
||||||
} else if (type == "Uniform") {
|
} else if (type == "Uniform") {
|
||||||
std::uniform_real_distribution<double> distribution(-1 / sqrt(n), 1 / sqrt(n));
|
std::uniform_real_distribution<real_t> distribution(-1 / sqrt(n), 1 / sqrt(n));
|
||||||
weights[i].push_back(distribution(generator));
|
weights[i].push_back(distribution(generator));
|
||||||
} else {
|
} else {
|
||||||
std::uniform_real_distribution<double> distribution(0, 1);
|
std::uniform_real_distribution<real_t> distribution(0, 1);
|
||||||
weights[i].push_back(distribution(generator));
|
weights[i].push_back(distribution(generator));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -94,11 +94,11 @@ std::vector<std::vector<double>> MLPPUtilities::weightInitialization(int n, int
|
|||||||
return weights;
|
return weights;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPUtilities::biasInitialization(int n) {
|
std::vector<real_t> MLPPUtilities::biasInitialization(int n) {
|
||||||
std::vector<double> bias;
|
std::vector<real_t> bias;
|
||||||
std::random_device rd;
|
std::random_device rd;
|
||||||
std::default_random_engine generator(rd());
|
std::default_random_engine generator(rd());
|
||||||
std::uniform_real_distribution<double> distribution(0, 1);
|
std::uniform_real_distribution<real_t> distribution(0, 1);
|
||||||
|
|
||||||
for (int i = 0; i < n; i++) {
|
for (int i = 0; i < n; i++) {
|
||||||
bias.push_back(distribution(generator));
|
bias.push_back(distribution(generator));
|
||||||
@ -106,8 +106,8 @@ std::vector<double> MLPPUtilities::biasInitialization(int n) {
|
|||||||
return bias;
|
return bias;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPUtilities::performance(std::vector<double> y_hat, std::vector<double> outputSet) {
|
real_t MLPPUtilities::performance(std::vector<real_t> y_hat, std::vector<real_t> outputSet) {
|
||||||
double correct = 0;
|
real_t correct = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
if (std::round(y_hat[i]) == outputSet[i]) {
|
if (std::round(y_hat[i]) == outputSet[i]) {
|
||||||
correct++;
|
correct++;
|
||||||
@ -116,8 +116,8 @@ double MLPPUtilities::performance(std::vector<double> y_hat, std::vector<double>
|
|||||||
return correct / y_hat.size();
|
return correct / y_hat.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPUtilities::performance(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y) {
|
real_t MLPPUtilities::performance(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y) {
|
||||||
double correct = 0;
|
real_t correct = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
int sub_correct = 0;
|
int sub_correct = 0;
|
||||||
for (int j = 0; j < y_hat[i].size(); j++) {
|
for (int j = 0; j < y_hat[i].size(); j++) {
|
||||||
@ -132,7 +132,7 @@ double MLPPUtilities::performance(std::vector<std::vector<double>> y_hat, std::v
|
|||||||
return correct / y_hat.size();
|
return correct / y_hat.size();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPUtilities::saveParameters(std::string fileName, std::vector<double> weights, double bias, bool app, int layer) {
|
void MLPPUtilities::saveParameters(std::string fileName, std::vector<real_t> weights, real_t bias, bool app, int layer) {
|
||||||
std::string layer_info = "";
|
std::string layer_info = "";
|
||||||
std::ofstream saveFile;
|
std::ofstream saveFile;
|
||||||
|
|
||||||
@ -160,7 +160,7 @@ void MLPPUtilities::saveParameters(std::string fileName, std::vector<double> wei
|
|||||||
saveFile.close();
|
saveFile.close();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPUtilities::saveParameters(std::string fileName, std::vector<double> weights, std::vector<double> initial, double bias, bool app, int layer) {
|
void MLPPUtilities::saveParameters(std::string fileName, std::vector<real_t> weights, std::vector<real_t> initial, real_t bias, bool app, int layer) {
|
||||||
std::string layer_info = "";
|
std::string layer_info = "";
|
||||||
std::ofstream saveFile;
|
std::ofstream saveFile;
|
||||||
|
|
||||||
@ -194,7 +194,7 @@ void MLPPUtilities::saveParameters(std::string fileName, std::vector<double> wei
|
|||||||
saveFile.close();
|
saveFile.close();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPUtilities::saveParameters(std::string fileName, std::vector<std::vector<double>> weights, std::vector<double> bias, bool app, int layer) {
|
void MLPPUtilities::saveParameters(std::string fileName, std::vector<std::vector<real_t>> weights, std::vector<real_t> bias, bool app, int layer) {
|
||||||
std::string layer_info = "";
|
std::string layer_info = "";
|
||||||
std::ofstream saveFile;
|
std::ofstream saveFile;
|
||||||
|
|
||||||
@ -226,7 +226,7 @@ void MLPPUtilities::saveParameters(std::string fileName, std::vector<std::vector
|
|||||||
saveFile.close();
|
saveFile.close();
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPUtilities::UI(std::vector<double> weights, double bias) {
|
void MLPPUtilities::UI(std::vector<real_t> weights, real_t bias) {
|
||||||
std::cout << "Values of the weight(s):" << std::endl;
|
std::cout << "Values of the weight(s):" << std::endl;
|
||||||
for (int i = 0; i < weights.size(); i++) {
|
for (int i = 0; i < weights.size(); i++) {
|
||||||
std::cout << weights[i] << std::endl;
|
std::cout << weights[i] << std::endl;
|
||||||
@ -235,7 +235,7 @@ void MLPPUtilities::UI(std::vector<double> weights, double bias) {
|
|||||||
std::cout << bias << std::endl;
|
std::cout << bias << std::endl;
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPUtilities::UI(std::vector<std::vector<double>> weights, std::vector<double> bias) {
|
void MLPPUtilities::UI(std::vector<std::vector<real_t>> weights, std::vector<real_t> bias) {
|
||||||
std::cout << "Values of the weight(s):" << std::endl;
|
std::cout << "Values of the weight(s):" << std::endl;
|
||||||
for (int i = 0; i < weights.size(); i++) {
|
for (int i = 0; i < weights.size(); i++) {
|
||||||
for (int j = 0; j < weights[i].size(); j++) {
|
for (int j = 0; j < weights[i].size(); j++) {
|
||||||
@ -248,7 +248,7 @@ void MLPPUtilities::UI(std::vector<std::vector<double>> weights, std::vector<dou
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPUtilities::UI(std::vector<double> weights, std::vector<double> initial, double bias) {
|
void MLPPUtilities::UI(std::vector<real_t> weights, std::vector<real_t> initial, real_t bias) {
|
||||||
std::cout << "Values of the weight(s):" << std::endl;
|
std::cout << "Values of the weight(s):" << std::endl;
|
||||||
for (int i = 0; i < weights.size(); i++) {
|
for (int i = 0; i < weights.size(); i++) {
|
||||||
std::cout << weights[i] << std::endl;
|
std::cout << weights[i] << std::endl;
|
||||||
@ -261,7 +261,7 @@ void MLPPUtilities::UI(std::vector<double> weights, std::vector<double> initial,
|
|||||||
std::cout << bias << std::endl;
|
std::cout << bias << std::endl;
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPUtilities::CostInfo(int epoch, double cost_prev, double Cost) {
|
void MLPPUtilities::CostInfo(int epoch, real_t cost_prev, real_t Cost) {
|
||||||
std::cout << "-----------------------------------" << std::endl;
|
std::cout << "-----------------------------------" << std::endl;
|
||||||
std::cout << "This is epoch: " << epoch << std::endl;
|
std::cout << "This is epoch: " << epoch << std::endl;
|
||||||
std::cout << "The cost function has been minimized by " << cost_prev - Cost << std::endl;
|
std::cout << "The cost function has been minimized by " << cost_prev - Cost << std::endl;
|
||||||
@ -269,21 +269,21 @@ void MLPPUtilities::CostInfo(int epoch, double cost_prev, double Cost) {
|
|||||||
std::cout << Cost << std::endl;
|
std::cout << Cost << std::endl;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> MLPPUtilities::createMiniBatches(std::vector<std::vector<double>> inputSet, int n_mini_batch) {
|
std::vector<std::vector<std::vector<real_t>>> MLPPUtilities::createMiniBatches(std::vector<std::vector<real_t>> inputSet, int n_mini_batch) {
|
||||||
int n = inputSet.size();
|
int n = inputSet.size();
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> inputMiniBatches;
|
std::vector<std::vector<std::vector<real_t>>> inputMiniBatches;
|
||||||
|
|
||||||
// Creating the mini-batches
|
// Creating the mini-batches
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<std::vector<double>> currentInputSet;
|
std::vector<std::vector<real_t>> currentInputSet;
|
||||||
for (int j = 0; j < n / n_mini_batch; j++) {
|
for (int j = 0; j < n / n_mini_batch; j++) {
|
||||||
currentInputSet.push_back(inputSet[n / n_mini_batch * i + j]);
|
currentInputSet.push_back(inputSet[n / n_mini_batch * i + j]);
|
||||||
}
|
}
|
||||||
inputMiniBatches.push_back(currentInputSet);
|
inputMiniBatches.push_back(currentInputSet);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (double(n) / double(n_mini_batch) - int(n / n_mini_batch) != 0) {
|
if (real_t(n) / real_t(n_mini_batch) - int(n / n_mini_batch) != 0) {
|
||||||
for (int i = 0; i < n - n / n_mini_batch * n_mini_batch; i++) {
|
for (int i = 0; i < n - n / n_mini_batch * n_mini_batch; i++) {
|
||||||
inputMiniBatches[n_mini_batch - 1].push_back(inputSet[n / n_mini_batch * n_mini_batch + i]);
|
inputMiniBatches[n_mini_batch - 1].push_back(inputSet[n / n_mini_batch * n_mini_batch + i]);
|
||||||
}
|
}
|
||||||
@ -291,15 +291,15 @@ std::vector<std::vector<std::vector<double>>> MLPPUtilities::createMiniBatches(s
|
|||||||
return inputMiniBatches;
|
return inputMiniBatches;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<std::vector<double>>> MLPPUtilities::createMiniBatches(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, int n_mini_batch) {
|
std::tuple<std::vector<std::vector<std::vector<real_t>>>, std::vector<std::vector<real_t>>> MLPPUtilities::createMiniBatches(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, int n_mini_batch) {
|
||||||
int n = inputSet.size();
|
int n = inputSet.size();
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> inputMiniBatches;
|
std::vector<std::vector<std::vector<real_t>>> inputMiniBatches;
|
||||||
std::vector<std::vector<double>> outputMiniBatches;
|
std::vector<std::vector<real_t>> outputMiniBatches;
|
||||||
|
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<std::vector<double>> currentInputSet;
|
std::vector<std::vector<real_t>> currentInputSet;
|
||||||
std::vector<double> currentOutputSet;
|
std::vector<real_t> currentOutputSet;
|
||||||
for (int j = 0; j < n / n_mini_batch; j++) {
|
for (int j = 0; j < n / n_mini_batch; j++) {
|
||||||
currentInputSet.push_back(inputSet[n / n_mini_batch * i + j]);
|
currentInputSet.push_back(inputSet[n / n_mini_batch * i + j]);
|
||||||
currentOutputSet.push_back(outputSet[n / n_mini_batch * i + j]);
|
currentOutputSet.push_back(outputSet[n / n_mini_batch * i + j]);
|
||||||
@ -308,7 +308,7 @@ std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<std::vecto
|
|||||||
outputMiniBatches.push_back(currentOutputSet);
|
outputMiniBatches.push_back(currentOutputSet);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (double(n) / double(n_mini_batch) - int(n / n_mini_batch) != 0) {
|
if (real_t(n) / real_t(n_mini_batch) - int(n / n_mini_batch) != 0) {
|
||||||
for (int i = 0; i < n - n / n_mini_batch * n_mini_batch; i++) {
|
for (int i = 0; i < n - n / n_mini_batch * n_mini_batch; i++) {
|
||||||
inputMiniBatches[n_mini_batch - 1].push_back(inputSet[n / n_mini_batch * n_mini_batch + i]);
|
inputMiniBatches[n_mini_batch - 1].push_back(inputSet[n / n_mini_batch * n_mini_batch + i]);
|
||||||
outputMiniBatches[n_mini_batch - 1].push_back(outputSet[n / n_mini_batch * n_mini_batch + i]);
|
outputMiniBatches[n_mini_batch - 1].push_back(outputSet[n / n_mini_batch * n_mini_batch + i]);
|
||||||
@ -317,15 +317,15 @@ std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<std::vecto
|
|||||||
return { inputMiniBatches, outputMiniBatches };
|
return { inputMiniBatches, outputMiniBatches };
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<std::vector<std::vector<double>>>> MLPPUtilities::createMiniBatches(std::vector<std::vector<double>> inputSet, std::vector<std::vector<double>> outputSet, int n_mini_batch) {
|
std::tuple<std::vector<std::vector<std::vector<real_t>>>, std::vector<std::vector<std::vector<real_t>>>> MLPPUtilities::createMiniBatches(std::vector<std::vector<real_t>> inputSet, std::vector<std::vector<real_t>> outputSet, int n_mini_batch) {
|
||||||
int n = inputSet.size();
|
int n = inputSet.size();
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> inputMiniBatches;
|
std::vector<std::vector<std::vector<real_t>>> inputMiniBatches;
|
||||||
std::vector<std::vector<std::vector<double>>> outputMiniBatches;
|
std::vector<std::vector<std::vector<real_t>>> outputMiniBatches;
|
||||||
|
|
||||||
for (int i = 0; i < n_mini_batch; i++) {
|
for (int i = 0; i < n_mini_batch; i++) {
|
||||||
std::vector<std::vector<double>> currentInputSet;
|
std::vector<std::vector<real_t>> currentInputSet;
|
||||||
std::vector<std::vector<double>> currentOutputSet;
|
std::vector<std::vector<real_t>> currentOutputSet;
|
||||||
for (int j = 0; j < n / n_mini_batch; j++) {
|
for (int j = 0; j < n / n_mini_batch; j++) {
|
||||||
currentInputSet.push_back(inputSet[n / n_mini_batch * i + j]);
|
currentInputSet.push_back(inputSet[n / n_mini_batch * i + j]);
|
||||||
currentOutputSet.push_back(outputSet[n / n_mini_batch * i + j]);
|
currentOutputSet.push_back(outputSet[n / n_mini_batch * i + j]);
|
||||||
@ -334,7 +334,7 @@ std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<std::vecto
|
|||||||
outputMiniBatches.push_back(currentOutputSet);
|
outputMiniBatches.push_back(currentOutputSet);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (double(n) / double(n_mini_batch) - int(n / n_mini_batch) != 0) {
|
if (real_t(n) / real_t(n_mini_batch) - int(n / n_mini_batch) != 0) {
|
||||||
for (int i = 0; i < n - n / n_mini_batch * n_mini_batch; i++) {
|
for (int i = 0; i < n - n / n_mini_batch * n_mini_batch; i++) {
|
||||||
inputMiniBatches[n_mini_batch - 1].push_back(inputSet[n / n_mini_batch * n_mini_batch + i]);
|
inputMiniBatches[n_mini_batch - 1].push_back(inputSet[n / n_mini_batch * n_mini_batch + i]);
|
||||||
outputMiniBatches[n_mini_batch - 1].push_back(outputSet[n / n_mini_batch * n_mini_batch + i]);
|
outputMiniBatches[n_mini_batch - 1].push_back(outputSet[n / n_mini_batch * n_mini_batch + i]);
|
||||||
@ -343,8 +343,8 @@ std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<std::vecto
|
|||||||
return { inputMiniBatches, outputMiniBatches };
|
return { inputMiniBatches, outputMiniBatches };
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<double, double, double, double> MLPPUtilities::TF_PN(std::vector<double> y_hat, std::vector<double> y) {
|
std::tuple<real_t, real_t, real_t, real_t> MLPPUtilities::TF_PN(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
double TP, FP, TN, FN = 0;
|
real_t TP, FP, TN, FN = 0;
|
||||||
for (int i = 0; i < y_hat.size(); i++) {
|
for (int i = 0; i < y_hat.size(); i++) {
|
||||||
if (y_hat[i] == y[i]) {
|
if (y_hat[i] == y[i]) {
|
||||||
if (y_hat[i] == 1) {
|
if (y_hat[i] == 1) {
|
||||||
@ -363,20 +363,20 @@ std::tuple<double, double, double, double> MLPPUtilities::TF_PN(std::vector<doub
|
|||||||
return { TP, FP, TN, FN };
|
return { TP, FP, TN, FN };
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPUtilities::recall(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPUtilities::recall(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
auto [TP, FP, TN, FN] = TF_PN(y_hat, y);
|
auto [TP, FP, TN, FN] = TF_PN(y_hat, y);
|
||||||
return TP / (TP + FN);
|
return TP / (TP + FN);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPUtilities::precision(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPUtilities::precision(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
auto [TP, FP, TN, FN] = TF_PN(y_hat, y);
|
auto [TP, FP, TN, FN] = TF_PN(y_hat, y);
|
||||||
return TP / (TP + FP);
|
return TP / (TP + FP);
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPUtilities::accuracy(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPUtilities::accuracy(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
auto [TP, FP, TN, FN] = TF_PN(y_hat, y);
|
auto [TP, FP, TN, FN] = TF_PN(y_hat, y);
|
||||||
return (TP + TN) / (TP + FP + FN + TN);
|
return (TP + TN) / (TP + FP + FN + TN);
|
||||||
}
|
}
|
||||||
double MLPPUtilities::f1_score(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPUtilities::f1_score(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
return 2 * precision(y_hat, y) * recall(y_hat, y) / (precision(y_hat, y) + recall(y_hat, y));
|
return 2 * precision(y_hat, y) * recall(y_hat, y) / (precision(y_hat, y) + recall(y_hat, y));
|
||||||
}
|
}
|
||||||
|
@ -8,6 +8,9 @@
|
|||||||
// Created by Marc Melikyan on 1/16/21.
|
// Created by Marc Melikyan on 1/16/21.
|
||||||
//
|
//
|
||||||
|
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
#include <tuple>
|
#include <tuple>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
@ -16,37 +19,37 @@
|
|||||||
class MLPPUtilities {
|
class MLPPUtilities {
|
||||||
public:
|
public:
|
||||||
// Weight Init
|
// Weight Init
|
||||||
static std::vector<double> weightInitialization(int n, std::string type = "Default");
|
static std::vector<real_t> weightInitialization(int n, std::string type = "Default");
|
||||||
static double biasInitialization();
|
static real_t biasInitialization();
|
||||||
|
|
||||||
static std::vector<std::vector<double>> weightInitialization(int n, int m, std::string type = "Default");
|
static std::vector<std::vector<real_t>> weightInitialization(int n, int m, std::string type = "Default");
|
||||||
static std::vector<double> biasInitialization(int n);
|
static std::vector<real_t> biasInitialization(int n);
|
||||||
|
|
||||||
// Cost/Performance related Functions
|
// Cost/Performance related Functions
|
||||||
double performance(std::vector<double> y_hat, std::vector<double> y);
|
real_t performance(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
double performance(std::vector<std::vector<double>> y_hat, std::vector<std::vector<double>> y);
|
real_t performance(std::vector<std::vector<real_t>> y_hat, std::vector<std::vector<real_t>> y);
|
||||||
|
|
||||||
// Parameter Saving Functions
|
// Parameter Saving Functions
|
||||||
void saveParameters(std::string fileName, std::vector<double> weights, double bias, bool app = 0, int layer = -1);
|
void saveParameters(std::string fileName, std::vector<real_t> weights, real_t bias, bool app = 0, int layer = -1);
|
||||||
void saveParameters(std::string fileName, std::vector<double> weights, std::vector<double> initial, double bias, bool app = 0, int layer = -1);
|
void saveParameters(std::string fileName, std::vector<real_t> weights, std::vector<real_t> initial, real_t bias, bool app = 0, int layer = -1);
|
||||||
void saveParameters(std::string fileName, std::vector<std::vector<double>> weights, std::vector<double> bias, bool app = 0, int layer = -1);
|
void saveParameters(std::string fileName, std::vector<std::vector<real_t>> weights, std::vector<real_t> bias, bool app = 0, int layer = -1);
|
||||||
|
|
||||||
// Gradient Descent related
|
// Gradient Descent related
|
||||||
static void UI(std::vector<double> weights, double bias);
|
static void UI(std::vector<real_t> weights, real_t bias);
|
||||||
static void UI(std::vector<double> weights, std::vector<double> initial, double bias);
|
static void UI(std::vector<real_t> weights, std::vector<real_t> initial, real_t bias);
|
||||||
static void UI(std::vector<std::vector<double>>, std::vector<double> bias);
|
static void UI(std::vector<std::vector<real_t>>, std::vector<real_t> bias);
|
||||||
static void CostInfo(int epoch, double cost_prev, double Cost);
|
static void CostInfo(int epoch, real_t cost_prev, real_t Cost);
|
||||||
|
|
||||||
static std::vector<std::vector<std::vector<double>>> createMiniBatches(std::vector<std::vector<double>> inputSet, int n_mini_batch);
|
static std::vector<std::vector<std::vector<real_t>>> createMiniBatches(std::vector<std::vector<real_t>> inputSet, int n_mini_batch);
|
||||||
static std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<std::vector<double>>> createMiniBatches(std::vector<std::vector<double>> inputSet, std::vector<double> outputSet, int n_mini_batch);
|
static std::tuple<std::vector<std::vector<std::vector<real_t>>>, std::vector<std::vector<real_t>>> createMiniBatches(std::vector<std::vector<real_t>> inputSet, std::vector<real_t> outputSet, int n_mini_batch);
|
||||||
static std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<std::vector<std::vector<double>>>> createMiniBatches(std::vector<std::vector<double>> inputSet, std::vector<std::vector<double>> outputSet, int n_mini_batch);
|
static std::tuple<std::vector<std::vector<std::vector<real_t>>>, std::vector<std::vector<std::vector<real_t>>>> createMiniBatches(std::vector<std::vector<real_t>> inputSet, std::vector<std::vector<real_t>> outputSet, int n_mini_batch);
|
||||||
|
|
||||||
// F1 score, Precision/Recall, TP, FP, TN, FN, etc.
|
// F1 score, Precision/Recall, TP, FP, TN, FN, etc.
|
||||||
std::tuple<double, double, double, double> TF_PN(std::vector<double> y_hat, std::vector<double> y); //TF_PN = "True", "False", "Positive", "Negative"
|
std::tuple<real_t, real_t, real_t, real_t> TF_PN(std::vector<real_t> y_hat, std::vector<real_t> y); //TF_PN = "True", "False", "Positive", "Negative"
|
||||||
double recall(std::vector<double> y_hat, std::vector<double> y);
|
real_t recall(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
double precision(std::vector<double> y_hat, std::vector<double> y);
|
real_t precision(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
double accuracy(std::vector<double> y_hat, std::vector<double> y);
|
real_t accuracy(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
double f1_score(std::vector<double> y_hat, std::vector<double> y);
|
real_t f1_score(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
};
|
};
|
||||||
|
@ -15,7 +15,7 @@
|
|||||||
#include <iostream>
|
#include <iostream>
|
||||||
|
|
||||||
|
|
||||||
MLPPWGAN::MLPPWGAN(double k, std::vector<std::vector<double>> outputSet) :
|
MLPPWGAN::MLPPWGAN(real_t k, std::vector<std::vector<real_t>> outputSet) :
|
||||||
outputSet(outputSet), n(outputSet.size()), k(k) {
|
outputSet(outputSet), n(outputSet.size()), k(k) {
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -23,15 +23,15 @@ MLPPWGAN::~MLPPWGAN() {
|
|||||||
delete outputLayer;
|
delete outputLayer;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPWGAN::generateExample(int n) {
|
std::vector<std::vector<real_t>> MLPPWGAN::generateExample(int n) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
return modelSetTestGenerator(alg.gaussianNoise(n, k));
|
return modelSetTestGenerator(alg.gaussianNoise(n, k));
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPWGAN::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
void MLPPWGAN::gradientDescent(real_t learning_rate, int max_epoch, bool UI) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
double cost_prev = 0;
|
real_t cost_prev = 0;
|
||||||
int epoch = 1;
|
int epoch = 1;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
|
|
||||||
@ -40,11 +40,11 @@ void MLPPWGAN::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
while (true) {
|
while (true) {
|
||||||
cost_prev = Cost(y_hat, alg.onevec(n));
|
cost_prev = Cost(y_hat, alg.onevec(n));
|
||||||
|
|
||||||
std::vector<std::vector<double>> generatorInputSet;
|
std::vector<std::vector<real_t>> generatorInputSet;
|
||||||
std::vector<std::vector<double>> discriminatorInputSet;
|
std::vector<std::vector<real_t>> discriminatorInputSet;
|
||||||
|
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
std::vector<double> outputSet;
|
std::vector<real_t> outputSet;
|
||||||
|
|
||||||
// Training of the discriminator.
|
// Training of the discriminator.
|
||||||
for (int i = 0; i < CRITIC_INTERATIONS; i++) {
|
for (int i = 0; i < CRITIC_INTERATIONS; i++) {
|
||||||
@ -54,7 +54,7 @@ void MLPPWGAN::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
|
|
||||||
y_hat = modelSetTestDiscriminator(discriminatorInputSet);
|
y_hat = modelSetTestDiscriminator(discriminatorInputSet);
|
||||||
outputSet = alg.scalarMultiply(-1, alg.onevec(n)); // WGAN changes y_i = 1 and y_i = 0 to y_i = 1 and y_i = -1
|
outputSet = alg.scalarMultiply(-1, alg.onevec(n)); // WGAN changes y_i = 1 and y_i = 0 to y_i = 1 and y_i = -1
|
||||||
std::vector<double> outputSetReal = alg.onevec(n);
|
std::vector<real_t> outputSetReal = alg.onevec(n);
|
||||||
outputSet.insert(outputSet.end(), outputSetReal.begin(), outputSetReal.end()); // Fake + real output scores.
|
outputSet.insert(outputSet.end(), outputSetReal.begin(), outputSetReal.end()); // Fake + real output scores.
|
||||||
|
|
||||||
auto [cumulativeDiscriminatorHiddenLayerWGrad, outputDiscriminatorWGrad] = computeDiscriminatorGradients(y_hat, outputSet);
|
auto [cumulativeDiscriminatorHiddenLayerWGrad, outputDiscriminatorWGrad] = computeDiscriminatorGradients(y_hat, outputSet);
|
||||||
@ -69,7 +69,7 @@ void MLPPWGAN::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
y_hat = modelSetTestDiscriminator(discriminatorInputSet);
|
y_hat = modelSetTestDiscriminator(discriminatorInputSet);
|
||||||
outputSet = alg.onevec(n);
|
outputSet = alg.onevec(n);
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> cumulativeGeneratorHiddenLayerWGrad = computeGeneratorGradients(y_hat, outputSet);
|
std::vector<std::vector<std::vector<real_t>>> cumulativeGeneratorHiddenLayerWGrad = computeGeneratorGradients(y_hat, outputSet);
|
||||||
cumulativeGeneratorHiddenLayerWGrad = alg.scalarMultiply(learning_rate / n, cumulativeGeneratorHiddenLayerWGrad);
|
cumulativeGeneratorHiddenLayerWGrad = alg.scalarMultiply(learning_rate / n, cumulativeGeneratorHiddenLayerWGrad);
|
||||||
updateGeneratorParameters(cumulativeGeneratorHiddenLayerWGrad, learning_rate);
|
updateGeneratorParameters(cumulativeGeneratorHiddenLayerWGrad, learning_rate);
|
||||||
|
|
||||||
@ -85,7 +85,7 @@ void MLPPWGAN::gradientDescent(double learning_rate, int max_epoch, bool UI) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPWGAN::score() {
|
real_t MLPPWGAN::score() {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPUtilities util;
|
MLPPUtilities util;
|
||||||
forwardPass();
|
forwardPass();
|
||||||
@ -105,7 +105,7 @@ void MLPPWGAN::save(std::string fileName) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPWGAN::addLayer(int n_hidden, std::string activation, std::string weightInit, std::string reg, double lambda, double alpha) {
|
void MLPPWGAN::addLayer(int n_hidden, std::string activation, std::string weightInit, std::string reg, real_t lambda, real_t alpha) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (network.empty()) {
|
if (network.empty()) {
|
||||||
network.push_back(MLPPHiddenLayer(n_hidden, activation, alg.gaussianNoise(n, k), weightInit, reg, lambda, alpha));
|
network.push_back(MLPPHiddenLayer(n_hidden, activation, alg.gaussianNoise(n, k), weightInit, reg, lambda, alpha));
|
||||||
@ -116,7 +116,7 @@ void MLPPWGAN::addLayer(int n_hidden, std::string activation, std::string weight
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPWGAN::addOutputLayer(std::string weightInit, std::string reg, double lambda, double alpha) {
|
void MLPPWGAN::addOutputLayer(std::string weightInit, std::string reg, real_t lambda, real_t alpha) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
outputLayer = new MLPPOutputLayer(network[network.size() - 1].n_hidden, "Linear", "WassersteinLoss", network[network.size() - 1].a, weightInit, "WeightClipping", -0.01, 0.01);
|
outputLayer = new MLPPOutputLayer(network[network.size() - 1].n_hidden, "Linear", "WassersteinLoss", network[network.size() - 1].a, weightInit, "WeightClipping", -0.01, 0.01);
|
||||||
@ -125,7 +125,7 @@ void MLPPWGAN::addOutputLayer(std::string weightInit, std::string reg, double la
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<double>> MLPPWGAN::modelSetTestGenerator(std::vector<std::vector<double>> X) {
|
std::vector<std::vector<real_t>> MLPPWGAN::modelSetTestGenerator(std::vector<std::vector<real_t>> X) {
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
network[0].input = X;
|
network[0].input = X;
|
||||||
network[0].forwardPass();
|
network[0].forwardPass();
|
||||||
@ -138,7 +138,7 @@ std::vector<std::vector<double>> MLPPWGAN::modelSetTestGenerator(std::vector<std
|
|||||||
return network[network.size() / 2].a;
|
return network[network.size() / 2].a;
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<double> MLPPWGAN::modelSetTestDiscriminator(std::vector<std::vector<double>> X) {
|
std::vector<real_t> MLPPWGAN::modelSetTestDiscriminator(std::vector<std::vector<real_t>> X) {
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
for (int i = network.size() / 2 + 1; i < network.size(); i++) {
|
for (int i = network.size() / 2 + 1; i < network.size(); i++) {
|
||||||
if (i == network.size() / 2 + 1) {
|
if (i == network.size() / 2 + 1) {
|
||||||
@ -154,10 +154,10 @@ std::vector<double> MLPPWGAN::modelSetTestDiscriminator(std::vector<std::vector<
|
|||||||
return outputLayer->a;
|
return outputLayer->a;
|
||||||
}
|
}
|
||||||
|
|
||||||
double MLPPWGAN::Cost(std::vector<double> y_hat, std::vector<double> y) {
|
real_t MLPPWGAN::Cost(std::vector<real_t> y_hat, std::vector<real_t> y) {
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
double totalRegTerm = 0;
|
real_t totalRegTerm = 0;
|
||||||
|
|
||||||
auto cost_function = outputLayer->cost_map[outputLayer->cost];
|
auto cost_function = outputLayer->cost_map[outputLayer->cost];
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
@ -186,7 +186,7 @@ void MLPPWGAN::forwardPass() {
|
|||||||
y_hat = outputLayer->a;
|
y_hat = outputLayer->a;
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPWGAN::updateDiscriminatorParameters(std::vector<std::vector<std::vector<double>>> hiddenLayerUpdations, std::vector<double> outputLayerUpdation, double learning_rate) {
|
void MLPPWGAN::updateDiscriminatorParameters(std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations, std::vector<real_t> outputLayerUpdation, real_t learning_rate) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
outputLayer->weights = alg.subtraction(outputLayer->weights, outputLayerUpdation);
|
outputLayer->weights = alg.subtraction(outputLayer->weights, outputLayerUpdation);
|
||||||
@ -203,7 +203,7 @@ void MLPPWGAN::updateDiscriminatorParameters(std::vector<std::vector<std::vector
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPWGAN::updateGeneratorParameters(std::vector<std::vector<std::vector<double>>> hiddenLayerUpdations, double learning_rate) {
|
void MLPPWGAN::updateGeneratorParameters(std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations, real_t learning_rate) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
@ -216,25 +216,25 @@ void MLPPWGAN::updateGeneratorParameters(std::vector<std::vector<std::vector<dou
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<double>> MLPPWGAN::computeDiscriminatorGradients(std::vector<double> y_hat, std::vector<double> outputSet) {
|
std::tuple<std::vector<std::vector<std::vector<real_t>>>, std::vector<real_t>> MLPPWGAN::computeDiscriminatorGradients(std::vector<real_t> y_hat, std::vector<real_t> outputSet) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> cumulativeHiddenLayerWGrad; // Tensor containing ALL hidden grads.
|
std::vector<std::vector<std::vector<real_t>>> cumulativeHiddenLayerWGrad; // Tensor containing ALL hidden grads.
|
||||||
|
|
||||||
auto costDeriv = outputLayer->costDeriv_map[outputLayer->cost];
|
auto costDeriv = outputLayer->costDeriv_map[outputLayer->cost];
|
||||||
auto outputAvn = outputLayer->activation_map[outputLayer->activation];
|
auto outputAvn = outputLayer->activation_map[outputLayer->activation];
|
||||||
outputLayer->delta = alg.hadamard_product((cost.*costDeriv)(y_hat, outputSet), (avn.*outputAvn)(outputLayer->z, 1));
|
outputLayer->delta = alg.hadamard_product((cost.*costDeriv)(y_hat, outputSet), (avn.*outputAvn)(outputLayer->z, 1));
|
||||||
std::vector<double> outputWGrad = alg.mat_vec_mult(alg.transpose(outputLayer->input), outputLayer->delta);
|
std::vector<real_t> outputWGrad = alg.mat_vec_mult(alg.transpose(outputLayer->input), outputLayer->delta);
|
||||||
outputWGrad = alg.addition(outputWGrad, regularization.regDerivTerm(outputLayer->weights, outputLayer->lambda, outputLayer->alpha, outputLayer->reg));
|
outputWGrad = alg.addition(outputWGrad, regularization.regDerivTerm(outputLayer->weights, outputLayer->lambda, outputLayer->alpha, outputLayer->reg));
|
||||||
|
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
auto hiddenLayerAvn = network[network.size() - 1].activation_map[network[network.size() - 1].activation];
|
auto hiddenLayerAvn = network[network.size() - 1].activation_map[network[network.size() - 1].activation];
|
||||||
|
|
||||||
network[network.size() - 1].delta = alg.hadamard_product(alg.outerProduct(outputLayer->delta, outputLayer->weights), (avn.*hiddenLayerAvn)(network[network.size() - 1].z, 1));
|
network[network.size() - 1].delta = alg.hadamard_product(alg.outerProduct(outputLayer->delta, outputLayer->weights), (avn.*hiddenLayerAvn)(network[network.size() - 1].z, 1));
|
||||||
std::vector<std::vector<double>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[network.size() - 1].input), network[network.size() - 1].delta);
|
std::vector<std::vector<real_t>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[network.size() - 1].input), network[network.size() - 1].delta);
|
||||||
|
|
||||||
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[network.size() - 1].weights, network[network.size() - 1].lambda, network[network.size() - 1].alpha, network[network.size() - 1].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[network.size() - 1].weights, network[network.size() - 1].lambda, network[network.size() - 1].alpha, network[network.size() - 1].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
||||||
|
|
||||||
@ -244,7 +244,7 @@ std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<double>> M
|
|||||||
for (int i = network.size() - 2; i > network.size() / 2; i--) {
|
for (int i = network.size() - 2; i > network.size() / 2; i--) {
|
||||||
auto hiddenLayerAvn = network[i].activation_map[network[i].activation];
|
auto hiddenLayerAvn = network[i].activation_map[network[i].activation];
|
||||||
network[i].delta = alg.hadamard_product(alg.matmult(network[i + 1].delta, alg.transpose(network[i + 1].weights)), (avn.*hiddenLayerAvn)(network[i].z, 1));
|
network[i].delta = alg.hadamard_product(alg.matmult(network[i + 1].delta, alg.transpose(network[i + 1].weights)), (avn.*hiddenLayerAvn)(network[i].z, 1));
|
||||||
std::vector<std::vector<double>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[i].input), network[i].delta);
|
std::vector<std::vector<real_t>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[i].input), network[i].delta);
|
||||||
|
|
||||||
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[i].weights, network[i].lambda, network[i].alpha, network[i].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[i].weights, network[i].lambda, network[i].alpha, network[i].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
||||||
}
|
}
|
||||||
@ -252,36 +252,36 @@ std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<double>> M
|
|||||||
return { cumulativeHiddenLayerWGrad, outputWGrad };
|
return { cumulativeHiddenLayerWGrad, outputWGrad };
|
||||||
}
|
}
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> MLPPWGAN::computeGeneratorGradients(std::vector<double> y_hat, std::vector<double> outputSet) {
|
std::vector<std::vector<std::vector<real_t>>> MLPPWGAN::computeGeneratorGradients(std::vector<real_t> y_hat, std::vector<real_t> outputSet) {
|
||||||
class MLPPCost cost;
|
class MLPPCost cost;
|
||||||
MLPPActivation avn;
|
MLPPActivation avn;
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
MLPPReg regularization;
|
MLPPReg regularization;
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> cumulativeHiddenLayerWGrad; // Tensor containing ALL hidden grads.
|
std::vector<std::vector<std::vector<real_t>>> cumulativeHiddenLayerWGrad; // Tensor containing ALL hidden grads.
|
||||||
|
|
||||||
auto costDeriv = outputLayer->costDeriv_map[outputLayer->cost];
|
auto costDeriv = outputLayer->costDeriv_map[outputLayer->cost];
|
||||||
auto outputAvn = outputLayer->activation_map[outputLayer->activation];
|
auto outputAvn = outputLayer->activation_map[outputLayer->activation];
|
||||||
outputLayer->delta = alg.hadamard_product((cost.*costDeriv)(y_hat, outputSet), (avn.*outputAvn)(outputLayer->z, 1));
|
outputLayer->delta = alg.hadamard_product((cost.*costDeriv)(y_hat, outputSet), (avn.*outputAvn)(outputLayer->z, 1));
|
||||||
std::vector<double> outputWGrad = alg.mat_vec_mult(alg.transpose(outputLayer->input), outputLayer->delta);
|
std::vector<real_t> outputWGrad = alg.mat_vec_mult(alg.transpose(outputLayer->input), outputLayer->delta);
|
||||||
outputWGrad = alg.addition(outputWGrad, regularization.regDerivTerm(outputLayer->weights, outputLayer->lambda, outputLayer->alpha, outputLayer->reg));
|
outputWGrad = alg.addition(outputWGrad, regularization.regDerivTerm(outputLayer->weights, outputLayer->lambda, outputLayer->alpha, outputLayer->reg));
|
||||||
if (!network.empty()) {
|
if (!network.empty()) {
|
||||||
auto hiddenLayerAvn = network[network.size() - 1].activation_map[network[network.size() - 1].activation];
|
auto hiddenLayerAvn = network[network.size() - 1].activation_map[network[network.size() - 1].activation];
|
||||||
network[network.size() - 1].delta = alg.hadamard_product(alg.outerProduct(outputLayer->delta, outputLayer->weights), (avn.*hiddenLayerAvn)(network[network.size() - 1].z, 1));
|
network[network.size() - 1].delta = alg.hadamard_product(alg.outerProduct(outputLayer->delta, outputLayer->weights), (avn.*hiddenLayerAvn)(network[network.size() - 1].z, 1));
|
||||||
std::vector<std::vector<double>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[network.size() - 1].input), network[network.size() - 1].delta);
|
std::vector<std::vector<real_t>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[network.size() - 1].input), network[network.size() - 1].delta);
|
||||||
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[network.size() - 1].weights, network[network.size() - 1].lambda, network[network.size() - 1].alpha, network[network.size() - 1].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[network.size() - 1].weights, network[network.size() - 1].lambda, network[network.size() - 1].alpha, network[network.size() - 1].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
||||||
|
|
||||||
for (int i = network.size() - 2; i >= 0; i--) {
|
for (int i = network.size() - 2; i >= 0; i--) {
|
||||||
auto hiddenLayerAvn = network[i].activation_map[network[i].activation];
|
auto hiddenLayerAvn = network[i].activation_map[network[i].activation];
|
||||||
network[i].delta = alg.hadamard_product(alg.matmult(network[i + 1].delta, alg.transpose(network[i + 1].weights)), (avn.*hiddenLayerAvn)(network[i].z, 1));
|
network[i].delta = alg.hadamard_product(alg.matmult(network[i + 1].delta, alg.transpose(network[i + 1].weights)), (avn.*hiddenLayerAvn)(network[i].z, 1));
|
||||||
std::vector<std::vector<double>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[i].input), network[i].delta);
|
std::vector<std::vector<real_t>> hiddenLayerWGrad = alg.matmult(alg.transpose(network[i].input), network[i].delta);
|
||||||
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[i].weights, network[i].lambda, network[i].alpha, network[i].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
cumulativeHiddenLayerWGrad.push_back(alg.addition(hiddenLayerWGrad, regularization.regDerivTerm(network[i].weights, network[i].lambda, network[i].alpha, network[i].reg))); // Adding to our cumulative hidden layer grads. Maintain reg terms as well.
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return cumulativeHiddenLayerWGrad;
|
return cumulativeHiddenLayerWGrad;
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPWGAN::UI(int epoch, double cost_prev, std::vector<double> y_hat, std::vector<double> outputSet) {
|
void MLPPWGAN::UI(int epoch, real_t cost_prev, std::vector<real_t> y_hat, std::vector<real_t> outputSet) {
|
||||||
MLPPUtilities::CostInfo(epoch, cost_prev, Cost(y_hat, outputSet));
|
MLPPUtilities::CostInfo(epoch, cost_prev, Cost(y_hat, outputSet));
|
||||||
std::cout << "Layer " << network.size() + 1 << ": " << std::endl;
|
std::cout << "Layer " << network.size() + 1 << ": " << std::endl;
|
||||||
MLPPUtilities::UI(outputLayer->weights, outputLayer->bias);
|
MLPPUtilities::UI(outputLayer->weights, outputLayer->bias);
|
||||||
|
@ -19,32 +19,32 @@
|
|||||||
|
|
||||||
class MLPPWGAN {
|
class MLPPWGAN {
|
||||||
public:
|
public:
|
||||||
MLPPWGAN(double k, std::vector<std::vector<double>> outputSet);
|
MLPPWGAN(real_t k, std::vector<std::vector<real_t>> outputSet);
|
||||||
~MLPPWGAN();
|
~MLPPWGAN();
|
||||||
std::vector<std::vector<double>> generateExample(int n);
|
std::vector<std::vector<real_t>> generateExample(int n);
|
||||||
void gradientDescent(double learning_rate, int max_epoch, bool UI = 1);
|
void gradientDescent(real_t learning_rate, int max_epoch, bool UI = 1);
|
||||||
double score();
|
real_t score();
|
||||||
void save(std::string fileName);
|
void save(std::string fileName);
|
||||||
|
|
||||||
void addLayer(int n_hidden, std::string activation, std::string weightInit = "Default", std::string reg = "None", double lambda = 0.5, double alpha = 0.5);
|
void addLayer(int n_hidden, std::string activation, std::string weightInit = "Default", std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
||||||
void addOutputLayer(std::string weightInit = "Default", std::string reg = "None", double lambda = 0.5, double alpha = 0.5);
|
void addOutputLayer(std::string weightInit = "Default", std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
||||||
|
|
||||||
private:
|
private:
|
||||||
std::vector<std::vector<double>> modelSetTestGenerator(std::vector<std::vector<double>> X); // Evaluator for the generator of the WGAN.
|
std::vector<std::vector<real_t>> modelSetTestGenerator(std::vector<std::vector<real_t>> X); // Evaluator for the generator of the WGAN.
|
||||||
std::vector<double> modelSetTestDiscriminator(std::vector<std::vector<double>> X); // Evaluator for the discriminator of the WGAN.
|
std::vector<real_t> modelSetTestDiscriminator(std::vector<std::vector<real_t>> X); // Evaluator for the discriminator of the WGAN.
|
||||||
|
|
||||||
double Cost(std::vector<double> y_hat, std::vector<double> y);
|
real_t Cost(std::vector<real_t> y_hat, std::vector<real_t> y);
|
||||||
|
|
||||||
void forwardPass();
|
void forwardPass();
|
||||||
void updateDiscriminatorParameters(std::vector<std::vector<std::vector<double>>> hiddenLayerUpdations, std::vector<double> outputLayerUpdation, double learning_rate);
|
void updateDiscriminatorParameters(std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations, std::vector<real_t> outputLayerUpdation, real_t learning_rate);
|
||||||
void updateGeneratorParameters(std::vector<std::vector<std::vector<double>>> hiddenLayerUpdations, double learning_rate);
|
void updateGeneratorParameters(std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations, real_t learning_rate);
|
||||||
std::tuple<std::vector<std::vector<std::vector<double>>>, std::vector<double>> computeDiscriminatorGradients(std::vector<double> y_hat, std::vector<double> outputSet);
|
std::tuple<std::vector<std::vector<std::vector<real_t>>>, std::vector<real_t>> computeDiscriminatorGradients(std::vector<real_t> y_hat, std::vector<real_t> outputSet);
|
||||||
std::vector<std::vector<std::vector<double>>> computeGeneratorGradients(std::vector<double> y_hat, std::vector<double> outputSet);
|
std::vector<std::vector<std::vector<real_t>>> computeGeneratorGradients(std::vector<real_t> y_hat, std::vector<real_t> outputSet);
|
||||||
|
|
||||||
void UI(int epoch, double cost_prev, std::vector<double> y_hat, std::vector<double> outputSet);
|
void UI(int epoch, real_t cost_prev, std::vector<real_t> y_hat, std::vector<real_t> outputSet);
|
||||||
|
|
||||||
std::vector<std::vector<double>> outputSet;
|
std::vector<std::vector<real_t>> outputSet;
|
||||||
std::vector<double> y_hat;
|
std::vector<real_t> y_hat;
|
||||||
|
|
||||||
std::vector<MLPPHiddenLayer> network;
|
std::vector<MLPPHiddenLayer> network;
|
||||||
MLPPOutputLayer *outputLayer;
|
MLPPOutputLayer *outputLayer;
|
||||||
|
@ -45,11 +45,11 @@
|
|||||||
#include "../mlpp/uni_lin_reg/uni_lin_reg.h"
|
#include "../mlpp/uni_lin_reg/uni_lin_reg.h"
|
||||||
#include "../mlpp/wgan/wgan.h"
|
#include "../mlpp/wgan/wgan.h"
|
||||||
|
|
||||||
Vector<double> dstd_vec_to_vec(const std::vector<double> &in) {
|
Vector<real_t> dstd_vec_to_vec(const std::vector<real_t> &in) {
|
||||||
Vector<double> r;
|
Vector<real_t> r;
|
||||||
|
|
||||||
r.resize(static_cast<int>(in.size()));
|
r.resize(static_cast<int>(in.size()));
|
||||||
double *darr = r.ptrw();
|
real_t *darr = r.ptrw();
|
||||||
|
|
||||||
for (uint32_t i = 0; i < in.size(); ++i) {
|
for (uint32_t i = 0; i < in.size(); ++i) {
|
||||||
darr[i] = in[i];
|
darr[i] = in[i];
|
||||||
@ -58,8 +58,8 @@ Vector<double> dstd_vec_to_vec(const std::vector<double> &in) {
|
|||||||
return r;
|
return r;
|
||||||
}
|
}
|
||||||
|
|
||||||
Vector<Vector<double>> dstd_mat_to_mat(const std::vector<std::vector<double>> &in) {
|
Vector<Vector<real_t>> dstd_mat_to_mat(const std::vector<std::vector<real_t>> &in) {
|
||||||
Vector<Vector<double>> r;
|
Vector<Vector<real_t>> r;
|
||||||
|
|
||||||
for (uint32_t i = 0; i < in.size(); ++i) {
|
for (uint32_t i = 0; i < in.size(); ++i) {
|
||||||
r.push_back(dstd_vec_to_vec(in[i]));
|
r.push_back(dstd_vec_to_vec(in[i]));
|
||||||
@ -75,9 +75,9 @@ void MLPPTests::test_statistics() {
|
|||||||
MLPPConvolutions conv;
|
MLPPConvolutions conv;
|
||||||
|
|
||||||
// STATISTICS
|
// STATISTICS
|
||||||
std::vector<double> x = { 1, 2, 3, 4, 5, 6, 7, 8, 9, 10 };
|
std::vector<real_t> x = { 1, 2, 3, 4, 5, 6, 7, 8, 9, 10 };
|
||||||
std::vector<double> y = { 10, 9, 8, 7, 6, 5, 4, 3, 2, 1 };
|
std::vector<real_t> y = { 10, 9, 8, 7, 6, 5, 4, 3, 2, 1 };
|
||||||
std::vector<double> w = { 0.1, 0.1, 0.1, 0.1, 0.1, 0.1, 0.1, 0.1, 0.1, 0.1 };
|
std::vector<real_t> w = { 0.1, 0.1, 0.1, 0.1, 0.1, 0.1, 0.1, 0.1, 0.1, 0.1 };
|
||||||
|
|
||||||
is_approx_equalsd(stat.mean(x), 5.5, "Arithmetic Mean");
|
is_approx_equalsd(stat.mean(x), 5.5, "Arithmetic Mean");
|
||||||
is_approx_equalsd(stat.mean(x), 5.5, "Median");
|
is_approx_equalsd(stat.mean(x), 5.5, "Median");
|
||||||
@ -117,19 +117,19 @@ void MLPPTests::test_statistics() {
|
|||||||
void MLPPTests::test_linear_algebra() {
|
void MLPPTests::test_linear_algebra() {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
std::vector<std::vector<double>> square = { { 1, 1 }, { -1, 1 }, { 1, -1 }, { -1, -1 } };
|
std::vector<std::vector<real_t>> square = { { 1, 1 }, { -1, 1 }, { 1, -1 }, { -1, -1 } };
|
||||||
std::vector<std::vector<double>> square_rot_res = { { 1.41421, 1.11022e-16 }, { -1.11022e-16, 1.41421 }, { 1.11022e-16, -1.41421 }, { -1.41421, -1.11022e-16 } };
|
std::vector<std::vector<real_t>> square_rot_res = { { 1.41421, 1.11022e-16 }, { -1.11022e-16, 1.41421 }, { 1.11022e-16, -1.41421 }, { -1.41421, -1.11022e-16 } };
|
||||||
|
|
||||||
is_approx_equals_dmat(dstd_mat_to_mat(alg.rotate(square, M_PI / 4)), dstd_mat_to_mat(square_rot_res), "alg.rotate(square, M_PI / 4)");
|
is_approx_equals_dmat(dstd_mat_to_mat(alg.rotate(square, M_PI / 4)), dstd_mat_to_mat(square_rot_res), "alg.rotate(square, M_PI / 4)");
|
||||||
|
|
||||||
std::vector<std::vector<double>> A = {
|
std::vector<std::vector<real_t>> A = {
|
||||||
{ 1, 2, 3, 4, 5, 6, 7, 8, 9, 10 },
|
{ 1, 2, 3, 4, 5, 6, 7, 8, 9, 10 },
|
||||||
{ 1, 2, 3, 4, 5, 6, 7, 8, 9, 10 },
|
{ 1, 2, 3, 4, 5, 6, 7, 8, 9, 10 },
|
||||||
};
|
};
|
||||||
std::vector<double> a = { 4, 3, 1, 3 };
|
std::vector<real_t> a = { 4, 3, 1, 3 };
|
||||||
std::vector<double> b = { 3, 5, 6, 1 };
|
std::vector<real_t> b = { 3, 5, 6, 1 };
|
||||||
|
|
||||||
std::vector<std::vector<double>> mmtr_res = {
|
std::vector<std::vector<real_t>> mmtr_res = {
|
||||||
{ 2, 4, 6, 8, 10, 12, 14, 16, 18, 20 },
|
{ 2, 4, 6, 8, 10, 12, 14, 16, 18, 20 },
|
||||||
{ 4, 8, 12, 16, 20, 24, 28, 32, 36, 40 },
|
{ 4, 8, 12, 16, 20, 24, 28, 32, 36, 40 },
|
||||||
{ 6, 12, 18, 24, 30, 36, 42, 48, 54, 60 },
|
{ 6, 12, 18, 24, 30, 36, 42, 48, 54, 60 },
|
||||||
@ -146,14 +146,14 @@ void MLPPTests::test_linear_algebra() {
|
|||||||
|
|
||||||
is_approx_equalsd(alg.dot(a, b), 36, "alg.dot(a, b)");
|
is_approx_equalsd(alg.dot(a, b), 36, "alg.dot(a, b)");
|
||||||
|
|
||||||
std::vector<std::vector<double>> had_prod_res = {
|
std::vector<std::vector<real_t>> had_prod_res = {
|
||||||
{ 1, 4, 9, 16, 25, 36, 49, 64, 81, 100 },
|
{ 1, 4, 9, 16, 25, 36, 49, 64, 81, 100 },
|
||||||
{ 1, 4, 9, 16, 25, 36, 49, 64, 81, 100 }
|
{ 1, 4, 9, 16, 25, 36, 49, 64, 81, 100 }
|
||||||
};
|
};
|
||||||
|
|
||||||
is_approx_equals_dmat(dstd_mat_to_mat(alg.hadamard_product(A, A)), dstd_mat_to_mat(had_prod_res), "alg.hadamard_product(A, A)");
|
is_approx_equals_dmat(dstd_mat_to_mat(alg.hadamard_product(A, A)), dstd_mat_to_mat(had_prod_res), "alg.hadamard_product(A, A)");
|
||||||
|
|
||||||
std::vector<std::vector<double>> id_10_res = {
|
std::vector<std::vector<real_t>> id_10_res = {
|
||||||
{ 1, 0, 0, 0, 0, 0, 0, 0, 0, 0 },
|
{ 1, 0, 0, 0, 0, 0, 0, 0, 0, 0 },
|
||||||
{ 0, 1, 0, 0, 0, 0, 0, 0, 0, 0 },
|
{ 0, 1, 0, 0, 0, 0, 0, 0, 0, 0 },
|
||||||
{ 0, 0, 1, 0, 0, 0, 0, 0, 0, 0 },
|
{ 0, 0, 1, 0, 0, 0, 0, 0, 0, 0 },
|
||||||
@ -177,7 +177,7 @@ void MLPPTests::test_univariate_linear_regression() {
|
|||||||
|
|
||||||
MLPPUniLinReg model(ds->input, ds->output);
|
MLPPUniLinReg model(ds->input, ds->output);
|
||||||
|
|
||||||
std::vector<double> slr_res = {
|
std::vector<real_t> slr_res = {
|
||||||
24.1095, 28.4829, 29.8082, 26.0974, 27.2902, 61.0851, 30.4709, 25.0372, 25.5673, 35.9046,
|
24.1095, 28.4829, 29.8082, 26.0974, 27.2902, 61.0851, 30.4709, 25.0372, 25.5673, 35.9046,
|
||||||
54.4587, 18.8083, 23.4468, 18.5432, 19.2059, 21.1938, 23.0492, 18.8083, 25.4348, 35.9046,
|
54.4587, 18.8083, 23.4468, 18.5432, 19.2059, 21.1938, 23.0492, 18.8083, 25.4348, 35.9046,
|
||||||
37.76, 40.278, 63.8683, 68.5068, 40.4106, 46.772, 32.0612, 23.3143, 44.784, 44.519,
|
37.76, 40.278, 63.8683, 68.5068, 40.4106, 46.772, 32.0612, 23.3143, 44.784, 44.519,
|
||||||
@ -254,8 +254,8 @@ void MLPPTests::test_multivariate_linear_regression_score_sgd_adam(bool ui) {
|
|||||||
|
|
||||||
const int TRIAL_NUM = 1000;
|
const int TRIAL_NUM = 1000;
|
||||||
|
|
||||||
double scoreSGD = 0;
|
real_t scoreSGD = 0;
|
||||||
double scoreADAM = 0;
|
real_t scoreADAM = 0;
|
||||||
for (int i = 0; i < TRIAL_NUM; i++) {
|
for (int i = 0; i < TRIAL_NUM; i++) {
|
||||||
MLPPLinReg modelf(alg.transpose(ds->input), ds->output);
|
MLPPLinReg modelf(alg.transpose(ds->input), ds->output);
|
||||||
modelf.MBGD(0.001, 5, 1, ui);
|
modelf.MBGD(0.001, 5, 1, ui);
|
||||||
@ -327,8 +327,8 @@ void MLPPTests::test_c_log_log_regression(bool ui) {
|
|||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
// CLOGLOG REGRESSION
|
// CLOGLOG REGRESSION
|
||||||
std::vector<std::vector<double>> inputSet = { { 1, 2, 3, 4, 5, 6, 7, 8 }, { 0, 0, 0, 0, 1, 1, 1, 1 } };
|
std::vector<std::vector<real_t>> inputSet = { { 1, 2, 3, 4, 5, 6, 7, 8 }, { 0, 0, 0, 0, 1, 1, 1, 1 } };
|
||||||
std::vector<double> outputSet = { 0, 0, 0, 0, 1, 1, 1, 1 };
|
std::vector<real_t> outputSet = { 0, 0, 0, 0, 1, 1, 1, 1 };
|
||||||
MLPPCLogLogReg model(alg.transpose(inputSet), outputSet);
|
MLPPCLogLogReg model(alg.transpose(inputSet), outputSet);
|
||||||
model.SGD(0.1, 10000, ui);
|
model.SGD(0.1, 10000, ui);
|
||||||
alg.printVector(model.modelSetTest(alg.transpose(inputSet)));
|
alg.printVector(model.modelSetTest(alg.transpose(inputSet)));
|
||||||
@ -338,8 +338,8 @@ void MLPPTests::test_exp_reg_regression(bool ui) {
|
|||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
// EXPREG REGRESSION
|
// EXPREG REGRESSION
|
||||||
std::vector<std::vector<double>> inputSet = { { 0, 1, 2, 3, 4 } };
|
std::vector<std::vector<real_t>> inputSet = { { 0, 1, 2, 3, 4 } };
|
||||||
std::vector<double> outputSet = { 1, 2, 4, 8, 16 };
|
std::vector<real_t> outputSet = { 1, 2, 4, 8, 16 };
|
||||||
MLPPExpReg model(alg.transpose(inputSet), outputSet);
|
MLPPExpReg model(alg.transpose(inputSet), outputSet);
|
||||||
model.SGD(0.001, 10000, ui);
|
model.SGD(0.001, 10000, ui);
|
||||||
alg.printVector(model.modelSetTest(alg.transpose(inputSet)));
|
alg.printVector(model.modelSetTest(alg.transpose(inputSet)));
|
||||||
@ -349,8 +349,8 @@ void MLPPTests::test_tanh_regression(bool ui) {
|
|||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
// TANH REGRESSION
|
// TANH REGRESSION
|
||||||
std::vector<std::vector<double>> inputSet = { { 4, 3, 0, -3, -4 }, { 0, 0, 0, 1, 1 } };
|
std::vector<std::vector<real_t>> inputSet = { { 4, 3, 0, -3, -4 }, { 0, 0, 0, 1, 1 } };
|
||||||
std::vector<double> outputSet = { 1, 1, 0, -1, -1 };
|
std::vector<real_t> outputSet = { 1, 1, 0, -1, -1 };
|
||||||
MLPPTanhReg model(alg.transpose(inputSet), outputSet);
|
MLPPTanhReg model(alg.transpose(inputSet), outputSet);
|
||||||
model.SGD(0.1, 10000, ui);
|
model.SGD(0.1, 10000, ui);
|
||||||
alg.printVector(model.modelSetTest(alg.transpose(inputSet)));
|
alg.printVector(model.modelSetTest(alg.transpose(inputSet)));
|
||||||
@ -387,9 +387,9 @@ void MLPPTests::test_mlp(bool ui) {
|
|||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
// MLP
|
// MLP
|
||||||
std::vector<std::vector<double>> inputSet = { { 0, 0, 1, 1 }, { 0, 1, 0, 1 } };
|
std::vector<std::vector<real_t>> inputSet = { { 0, 0, 1, 1 }, { 0, 1, 0, 1 } };
|
||||||
inputSet = alg.transpose(inputSet);
|
inputSet = alg.transpose(inputSet);
|
||||||
std::vector<double> outputSet = { 0, 1, 1, 0 };
|
std::vector<real_t> outputSet = { 0, 1, 1, 0 };
|
||||||
|
|
||||||
MLPPMLP model(inputSet, outputSet, 2);
|
MLPPMLP model(inputSet, outputSet, 2);
|
||||||
model.gradientDescent(0.1, 10000, ui);
|
model.gradientDescent(0.1, 10000, ui);
|
||||||
@ -412,7 +412,7 @@ void MLPPTests::test_autoencoder(bool ui) {
|
|||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
// AUTOENCODER
|
// AUTOENCODER
|
||||||
std::vector<std::vector<double>> inputSet = { { 1, 2, 3, 4, 5, 6, 7, 8, 9, 10 }, { 3, 5, 9, 12, 15, 18, 21, 24, 27, 30 } };
|
std::vector<std::vector<real_t>> inputSet = { { 1, 2, 3, 4, 5, 6, 7, 8, 9, 10 }, { 3, 5, 9, 12, 15, 18, 21, 24, 27, 30 } };
|
||||||
MLPPAutoEncoder model(alg.transpose(inputSet), 5);
|
MLPPAutoEncoder model(alg.transpose(inputSet), 5);
|
||||||
model.SGD(0.001, 300000, ui);
|
model.SGD(0.001, 300000, ui);
|
||||||
alg.printMatrix(model.modelSetTest(alg.transpose(inputSet)));
|
alg.printMatrix(model.modelSetTest(alg.transpose(inputSet)));
|
||||||
@ -425,8 +425,8 @@ void MLPPTests::test_dynamically_sized_ann(bool ui) {
|
|||||||
// Possible Weight Init Methods: Default, Uniform, HeNormal, HeUniform, XavierNormal, XavierUniform
|
// Possible Weight Init Methods: Default, Uniform, HeNormal, HeUniform, XavierNormal, XavierUniform
|
||||||
// Possible Activations: Linear, Sigmoid, Swish, Softplus, Softsign, CLogLog, Ar{Sinh, Cosh, Tanh, Csch, Sech, Coth}, GaussianCDF, GELU, UnitStep
|
// Possible Activations: Linear, Sigmoid, Swish, Softplus, Softsign, CLogLog, Ar{Sinh, Cosh, Tanh, Csch, Sech, Coth}, GaussianCDF, GELU, UnitStep
|
||||||
// Possible Loss Functions: MSE, RMSE, MBE, LogLoss, CrossEntropy, HingeLoss
|
// Possible Loss Functions: MSE, RMSE, MBE, LogLoss, CrossEntropy, HingeLoss
|
||||||
std::vector<std::vector<double>> inputSet = { { 0, 0, 1, 1 }, { 0, 1, 0, 1 } };
|
std::vector<std::vector<real_t>> inputSet = { { 0, 0, 1, 1 }, { 0, 1, 0, 1 } };
|
||||||
std::vector<double> outputSet = { 0, 1, 1, 0 };
|
std::vector<real_t> outputSet = { 0, 1, 1, 0 };
|
||||||
MLPPANN ann(alg.transpose(inputSet), outputSet);
|
MLPPANN ann(alg.transpose(inputSet), outputSet);
|
||||||
ann.addLayer(2, "Cosh");
|
ann.addLayer(2, "Cosh");
|
||||||
ann.addOutputLayer("Sigmoid", "LogLoss");
|
ann.addOutputLayer("Sigmoid", "LogLoss");
|
||||||
@ -448,7 +448,7 @@ void MLPPTests::test_wgan(bool ui) {
|
|||||||
//MLPPData data;
|
//MLPPData data;
|
||||||
//MLPPConvolutions conv;
|
//MLPPConvolutions conv;
|
||||||
|
|
||||||
std::vector<std::vector<double>> outputSet = {
|
std::vector<std::vector<real_t>> outputSet = {
|
||||||
{ 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20 },
|
{ 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20 },
|
||||||
{ 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30, 32, 34, 36, 38, 40 }
|
{ 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30, 32, 34, 36, 38, 40 }
|
||||||
};
|
};
|
||||||
@ -465,8 +465,8 @@ void MLPPTests::test_wgan(bool ui) {
|
|||||||
void MLPPTests::test_ann(bool ui) {
|
void MLPPTests::test_ann(bool ui) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet = { { 0, 0 }, { 0, 1 }, { 1, 0 }, { 1, 1 } }; // XOR
|
std::vector<std::vector<real_t>> inputSet = { { 0, 0 }, { 0, 1 }, { 1, 0 }, { 1, 1 } }; // XOR
|
||||||
std::vector<double> outputSet = { 0, 1, 1, 0 };
|
std::vector<real_t> outputSet = { 0, 1, 1, 0 };
|
||||||
|
|
||||||
MLPPANN ann(inputSet, outputSet);
|
MLPPANN ann(inputSet, outputSet);
|
||||||
ann.addLayer(5, "Sigmoid");
|
ann.addLayer(5, "Sigmoid");
|
||||||
@ -474,7 +474,7 @@ void MLPPTests::test_ann(bool ui) {
|
|||||||
ann.addOutputLayer("Sigmoid", "LogLoss");
|
ann.addOutputLayer("Sigmoid", "LogLoss");
|
||||||
ann.gradientDescent(1, 20000, ui);
|
ann.gradientDescent(1, 20000, ui);
|
||||||
|
|
||||||
std::vector<double> predictions = ann.modelSetTest(inputSet);
|
std::vector<real_t> predictions = ann.modelSetTest(inputSet);
|
||||||
alg.printVector(predictions); // Testing out the model's preds for train set.
|
alg.printVector(predictions); // Testing out the model's preds for train set.
|
||||||
std::cout << "ACCURACY: " << 100 * ann.score() << "%" << std::endl; // Accuracy.
|
std::cout << "ACCURACY: " << 100 * ann.score() << "%" << std::endl; // Accuracy.
|
||||||
}
|
}
|
||||||
@ -483,8 +483,8 @@ void MLPPTests::test_dynamically_sized_mann(bool ui) {
|
|||||||
MLPPData data;
|
MLPPData data;
|
||||||
|
|
||||||
// DYNAMICALLY SIZED MANN (Multidimensional Output ANN)
|
// DYNAMICALLY SIZED MANN (Multidimensional Output ANN)
|
||||||
std::vector<std::vector<double>> inputSet = { { 1, 2, 3 }, { 2, 4, 6 }, { 3, 6, 9 }, { 4, 8, 12 } };
|
std::vector<std::vector<real_t>> inputSet = { { 1, 2, 3 }, { 2, 4, 6 }, { 3, 6, 9 }, { 4, 8, 12 } };
|
||||||
std::vector<std::vector<double>> outputSet = { { 1, 5 }, { 2, 10 }, { 3, 15 }, { 4, 20 } };
|
std::vector<std::vector<real_t>> outputSet = { { 1, 5 }, { 2, 10 }, { 3, 15 }, { 4, 20 } };
|
||||||
|
|
||||||
MLPPMANN mann(inputSet, outputSet);
|
MLPPMANN mann(inputSet, outputSet);
|
||||||
mann.addOutputLayer("Linear", "MSE");
|
mann.addOutputLayer("Linear", "MSE");
|
||||||
@ -497,8 +497,8 @@ void MLPPTests::test_train_test_split_mann(bool ui) {
|
|||||||
MLPPData data;
|
MLPPData data;
|
||||||
|
|
||||||
// TRAIN TEST SPLIT CHECK
|
// TRAIN TEST SPLIT CHECK
|
||||||
std::vector<std::vector<double>> inputSet1 = { { 1, 2, 3, 4, 5, 6, 7, 8, 9, 10 }, { 3, 5, 9, 12, 15, 18, 21, 24, 27, 30 } };
|
std::vector<std::vector<real_t>> inputSet1 = { { 1, 2, 3, 4, 5, 6, 7, 8, 9, 10 }, { 3, 5, 9, 12, 15, 18, 21, 24, 27, 30 } };
|
||||||
std::vector<std::vector<double>> outputSet1 = { { 2, 4, 6, 8, 10, 12, 14, 16, 18, 20 } };
|
std::vector<std::vector<real_t>> outputSet1 = { { 2, 4, 6, 8, 10, 12, 14, 16, 18, 20 } };
|
||||||
|
|
||||||
Ref<MLPPDataComplex> d;
|
Ref<MLPPDataComplex> d;
|
||||||
d.instance();
|
d.instance();
|
||||||
@ -525,8 +525,8 @@ void MLPPTests::test_naive_bayes() {
|
|||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
// NAIVE BAYES
|
// NAIVE BAYES
|
||||||
std::vector<std::vector<double>> inputSet = { { 1, 1, 1, 1, 1 }, { 0, 0, 1, 1, 1 }, { 0, 0, 1, 0, 1 } };
|
std::vector<std::vector<real_t>> inputSet = { { 1, 1, 1, 1, 1 }, { 0, 0, 1, 1, 1 }, { 0, 0, 1, 0, 1 } };
|
||||||
std::vector<double> outputSet = { 0, 1, 0, 1, 1 };
|
std::vector<real_t> outputSet = { 0, 1, 0, 1, 1 };
|
||||||
|
|
||||||
MLPPMultinomialNB MNB(alg.transpose(inputSet), outputSet, 2);
|
MLPPMultinomialNB MNB(alg.transpose(inputSet), outputSet, 2);
|
||||||
alg.printVector(MNB.modelSetTest(alg.transpose(inputSet)));
|
alg.printVector(MNB.modelSetTest(alg.transpose(inputSet)));
|
||||||
@ -541,7 +541,7 @@ void MLPPTests::test_k_means(bool ui) {
|
|||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
// KMeans
|
// KMeans
|
||||||
std::vector<std::vector<double>> inputSet = { { 32, 0, 7 }, { 2, 28, 17 }, { 0, 9, 23 } };
|
std::vector<std::vector<real_t>> inputSet = { { 32, 0, 7 }, { 2, 28, 17 }, { 0, 9, 23 } };
|
||||||
MLPPKMeans kmeans(inputSet, 3, "KMeans++");
|
MLPPKMeans kmeans(inputSet, 3, "KMeans++");
|
||||||
kmeans.train(3, ui);
|
kmeans.train(3, ui);
|
||||||
std::cout << std::endl;
|
std::cout << std::endl;
|
||||||
@ -553,8 +553,8 @@ void MLPPTests::test_knn(bool ui) {
|
|||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
// kNN
|
// kNN
|
||||||
std::vector<std::vector<double>> inputSet = { { 1, 2, 3, 4, 5, 6, 7, 8 }, { 0, 0, 0, 0, 1, 1, 1, 1 } };
|
std::vector<std::vector<real_t>> inputSet = { { 1, 2, 3, 4, 5, 6, 7, 8 }, { 0, 0, 0, 0, 1, 1, 1, 1 } };
|
||||||
std::vector<double> outputSet = { 0, 0, 0, 0, 1, 1, 1, 1 };
|
std::vector<real_t> outputSet = { 0, 0, 0, 0, 1, 1, 1, 1 };
|
||||||
MLPPKNN knn(alg.transpose(inputSet), outputSet, 8);
|
MLPPKNN knn(alg.transpose(inputSet), outputSet, 8);
|
||||||
alg.printVector(knn.modelSetTest(alg.transpose(inputSet)));
|
alg.printVector(knn.modelSetTest(alg.transpose(inputSet)));
|
||||||
std::cout << "ACCURACY: " << 100 * knn.score() << "%" << std::endl;
|
std::cout << "ACCURACY: " << 100 * knn.score() << "%" << std::endl;
|
||||||
@ -566,18 +566,18 @@ void MLPPTests::test_convolution_tensors_etc() {
|
|||||||
MLPPConvolutions conv;
|
MLPPConvolutions conv;
|
||||||
|
|
||||||
// CONVOLUTION, POOLING, ETC..
|
// CONVOLUTION, POOLING, ETC..
|
||||||
std::vector<std::vector<double>> input = {
|
std::vector<std::vector<real_t>> input = {
|
||||||
{ 1 },
|
{ 1 },
|
||||||
};
|
};
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> tensorSet;
|
std::vector<std::vector<std::vector<real_t>>> tensorSet;
|
||||||
tensorSet.push_back(input);
|
tensorSet.push_back(input);
|
||||||
tensorSet.push_back(input);
|
tensorSet.push_back(input);
|
||||||
tensorSet.push_back(input);
|
tensorSet.push_back(input);
|
||||||
|
|
||||||
alg.printTensor(data.rgb2xyz(tensorSet));
|
alg.printTensor(data.rgb2xyz(tensorSet));
|
||||||
|
|
||||||
std::vector<std::vector<double>> input2 = {
|
std::vector<std::vector<real_t>> input2 = {
|
||||||
{ 62, 55, 55, 54, 49, 48, 47, 55 },
|
{ 62, 55, 55, 54, 49, 48, 47, 55 },
|
||||||
{ 62, 57, 54, 52, 48, 47, 48, 53 },
|
{ 62, 57, 54, 52, 48, 47, 48, 53 },
|
||||||
{ 61, 60, 52, 49, 48, 47, 49, 54 },
|
{ 61, 60, 52, 49, 48, 47, 49, 54 },
|
||||||
@ -595,19 +595,19 @@ void MLPPTests::test_convolution_tensors_etc() {
|
|||||||
alg.printMatrix(conv.convolve(input2, conv.getPrewittVertical(), 1)); // Can use padding
|
alg.printMatrix(conv.convolve(input2, conv.getPrewittVertical(), 1)); // Can use padding
|
||||||
alg.printMatrix(conv.pool(input2, 4, 4, "Max")); // Can use Max, Min, or Average pooling.
|
alg.printMatrix(conv.pool(input2, 4, 4, "Max")); // Can use Max, Min, or Average pooling.
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> tensorSet2;
|
std::vector<std::vector<std::vector<real_t>>> tensorSet2;
|
||||||
tensorSet2.push_back(input2);
|
tensorSet2.push_back(input2);
|
||||||
tensorSet2.push_back(input2);
|
tensorSet2.push_back(input2);
|
||||||
alg.printVector(conv.globalPool(tensorSet2, "Average")); // Can use Max, Min, or Average global pooling.
|
alg.printVector(conv.globalPool(tensorSet2, "Average")); // Can use Max, Min, or Average global pooling.
|
||||||
|
|
||||||
std::vector<std::vector<double>> laplacian = { { 1, 1, 1 }, { 1, -4, 1 }, { 1, 1, 1 } };
|
std::vector<std::vector<real_t>> laplacian = { { 1, 1, 1 }, { 1, -4, 1 }, { 1, 1, 1 } };
|
||||||
alg.printMatrix(conv.convolve(conv.gaussianFilter2D(5, 1), laplacian, 1));
|
alg.printMatrix(conv.convolve(conv.gaussianFilter2D(5, 1), laplacian, 1));
|
||||||
}
|
}
|
||||||
void MLPPTests::test_pca_svd_eigenvalues_eigenvectors(bool ui) {
|
void MLPPTests::test_pca_svd_eigenvalues_eigenvectors(bool ui) {
|
||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
// PCA, SVD, eigenvalues & eigenvectors
|
// PCA, SVD, eigenvalues & eigenvectors
|
||||||
std::vector<std::vector<double>> inputSet = { { 1, 1 }, { 1, 1 } };
|
std::vector<std::vector<real_t>> inputSet = { { 1, 1 }, { 1, 1 } };
|
||||||
|
|
||||||
MLPPLinAlg::EigenResult eigen = alg.eigen(inputSet);
|
MLPPLinAlg::EigenResult eigen = alg.eigen(inputSet);
|
||||||
|
|
||||||
@ -675,7 +675,7 @@ void MLPPTests::test_nlp_and_data(bool ui) {
|
|||||||
//alg.printMatrix(data.BOW(textArchive, "Default"));
|
//alg.printMatrix(data.BOW(textArchive, "Default"));
|
||||||
std::cout << std::endl;
|
std::cout << std::endl;
|
||||||
|
|
||||||
std::vector<std::vector<double>> inputSet = { { 1, 2 }, { 2, 3 }, { 3, 4 }, { 4, 5 }, { 5, 6 } };
|
std::vector<std::vector<real_t>> inputSet = { { 1, 2 }, { 2, 3 }, { 3, 4 }, { 4, 5 }, { 5, 6 } };
|
||||||
std::cout << "Feature Scaling Example:" << std::endl;
|
std::cout << "Feature Scaling Example:" << std::endl;
|
||||||
alg.printMatrix(data.featureScaling(inputSet));
|
alg.printMatrix(data.featureScaling(inputSet));
|
||||||
std::cout << std::endl;
|
std::cout << std::endl;
|
||||||
@ -692,7 +692,8 @@ void MLPPTests::test_outlier_finder(bool ui) {
|
|||||||
MLPPLinAlg alg;
|
MLPPLinAlg alg;
|
||||||
|
|
||||||
// Outlier Finder
|
// Outlier Finder
|
||||||
std::vector<double> inputSet = { 1, 2, 3, 4, 5, 6, 7, 8, 9, 23554332523523 };
|
//std::vector<real_t> inputSet = { 1, 2, 3, 4, 5, 6, 7, 8, 9, 23554332523523 };
|
||||||
|
std::vector<real_t> inputSet = { 1, 2, 3, 4, 5, 6, 7, 8, 9, 23554332 };
|
||||||
MLPPOutlierFinder outlierFinder(2); // Any datapoint outside of 2 stds from the mean is marked as an outlier.
|
MLPPOutlierFinder outlierFinder(2); // Any datapoint outside of 2 stds from the mean is marked as an outlier.
|
||||||
alg.printVector(outlierFinder.modelTest(inputSet));
|
alg.printVector(outlierFinder.modelTest(inputSet));
|
||||||
}
|
}
|
||||||
@ -702,15 +703,15 @@ void MLPPTests::test_new_math_functions() {
|
|||||||
MLPPData data;
|
MLPPData data;
|
||||||
|
|
||||||
// Testing new Functions
|
// Testing new Functions
|
||||||
double z_s = 0.001;
|
real_t z_s = 0.001;
|
||||||
std::cout << avn.logit(z_s) << std::endl;
|
std::cout << avn.logit(z_s) << std::endl;
|
||||||
std::cout << avn.logit(z_s, 1) << std::endl;
|
std::cout << avn.logit(z_s, 1) << std::endl;
|
||||||
|
|
||||||
std::vector<double> z_v = { 0.001 };
|
std::vector<real_t> z_v = { 0.001 };
|
||||||
alg.printVector(avn.logit(z_v));
|
alg.printVector(avn.logit(z_v));
|
||||||
alg.printVector(avn.logit(z_v, 1));
|
alg.printVector(avn.logit(z_v, 1));
|
||||||
|
|
||||||
std::vector<std::vector<double>> Z_m = { { 0.001 } };
|
std::vector<std::vector<real_t>> Z_m = { { 0.001 } };
|
||||||
alg.printMatrix(avn.logit(Z_m));
|
alg.printMatrix(avn.logit(Z_m));
|
||||||
alg.printMatrix(avn.logit(Z_m, 1));
|
alg.printMatrix(avn.logit(Z_m, 1));
|
||||||
|
|
||||||
@ -721,18 +722,18 @@ void MLPPTests::test_new_math_functions() {
|
|||||||
alg.printMatrix(alg.matrixPower({ { 5, 5 }, { 5, 5 } }, 2));
|
alg.printMatrix(alg.matrixPower({ { 5, 5 }, { 5, 5 } }, 2));
|
||||||
alg.printVector(alg.solve({ { 1, 1 }, { 1.5, 4.0 } }, { 2200, 5050 }));
|
alg.printVector(alg.solve({ { 1, 1 }, { 1.5, 4.0 } }, { 2200, 5050 }));
|
||||||
|
|
||||||
std::vector<std::vector<double>> matrixOfCubes = { { 1, 2, 64, 27 } };
|
std::vector<std::vector<real_t>> matrixOfCubes = { { 1, 2, 64, 27 } };
|
||||||
std::vector<double> vectorOfCubes = { 1, 2, 64, 27 };
|
std::vector<real_t> vectorOfCubes = { 1, 2, 64, 27 };
|
||||||
alg.printMatrix(alg.cbrt(matrixOfCubes));
|
alg.printMatrix(alg.cbrt(matrixOfCubes));
|
||||||
alg.printVector(alg.cbrt(vectorOfCubes));
|
alg.printVector(alg.cbrt(vectorOfCubes));
|
||||||
std::cout << alg.max({ { 1, 2, 3, 4, 5 }, { 6, 5, 3, 4, 1 }, { 9, 9, 9, 9, 9 } }) << std::endl;
|
std::cout << alg.max({ { 1, 2, 3, 4, 5 }, { 6, 5, 3, 4, 1 }, { 9, 9, 9, 9, 9 } }) << std::endl;
|
||||||
std::cout << alg.min({ { 1, 2, 3, 4, 5 }, { 6, 5, 3, 4, 1 }, { 9, 9, 9, 9, 9 } }) << std::endl;
|
std::cout << alg.min({ { 1, 2, 3, 4, 5 }, { 6, 5, 3, 4, 1 }, { 9, 9, 9, 9, 9 } }) << std::endl;
|
||||||
|
|
||||||
//std::vector<double> chicken;
|
//std::vector<real_t> chicken;
|
||||||
//data.getImage("../../Data/apple.jpeg", chicken);
|
//data.getImage("../../Data/apple.jpeg", chicken);
|
||||||
//alg.printVector(chicken);
|
//alg.printVector(chicken);
|
||||||
|
|
||||||
std::vector<std::vector<double>> P = { { 12, -51, 4 }, { 6, 167, -68 }, { -4, 24, -41 } };
|
std::vector<std::vector<real_t>> P = { { 12, -51, 4 }, { 6, 167, -68 }, { -4, 24, -41 } };
|
||||||
alg.printMatrix(P);
|
alg.printMatrix(P);
|
||||||
|
|
||||||
alg.printMatrix(alg.gramSchmidtProcess(P));
|
alg.printMatrix(alg.gramSchmidtProcess(P));
|
||||||
@ -752,7 +753,7 @@ void MLPPTests::test_positive_definiteness_checker() {
|
|||||||
//MLPPConvolutions conv;
|
//MLPPConvolutions conv;
|
||||||
|
|
||||||
// Checking positive-definiteness checker. For Cholesky Decomp.
|
// Checking positive-definiteness checker. For Cholesky Decomp.
|
||||||
std::vector<std::vector<double>> A = {
|
std::vector<std::vector<real_t>> A = {
|
||||||
{ 1, -1, -1, -1 },
|
{ 1, -1, -1, -1 },
|
||||||
{ -1, 2, 2, 2 },
|
{ -1, 2, 2, 2 },
|
||||||
{ -1, 2, 3, 1 },
|
{ -1, 2, 3, 1 },
|
||||||
@ -765,19 +766,19 @@ void MLPPTests::test_positive_definiteness_checker() {
|
|||||||
alg.printMatrix(chres.Lt);
|
alg.printMatrix(chres.Lt);
|
||||||
}
|
}
|
||||||
|
|
||||||
// double f(double x){
|
// real_t f(real_t x){
|
||||||
// return x*x*x + 2*x - 2;
|
// return x*x*x + 2*x - 2;
|
||||||
// }
|
// }
|
||||||
|
|
||||||
double f(double x) {
|
real_t f(real_t x) {
|
||||||
return sin(x);
|
return sin(x);
|
||||||
}
|
}
|
||||||
|
|
||||||
double f_prime(double x) {
|
real_t f_prime(real_t x) {
|
||||||
return 2 * x;
|
return 2 * x;
|
||||||
}
|
}
|
||||||
|
|
||||||
double f_prime_2var(std::vector<double> x) {
|
real_t f_prime_2var(std::vector<real_t> x) {
|
||||||
return 2 * x[0] + x[1];
|
return 2 * x[0] + x[1];
|
||||||
}
|
}
|
||||||
/*
|
/*
|
||||||
@ -787,7 +788,7 @@ double f_prime_2var(std::vector<double> x) {
|
|||||||
y''(2) = 12
|
y''(2) = 12
|
||||||
*/
|
*/
|
||||||
|
|
||||||
// double f_mv(std::vector<double> x){
|
// real_t f_mv(std::vector<real_t> x){
|
||||||
// return x[0] * x[0] + x[0] * x[1] * x[1] + x[1] + 5;
|
// return x[0] * x[0] + x[0] * x[1] * x[1] + x[1] + 5;
|
||||||
// }
|
// }
|
||||||
|
|
||||||
@ -798,7 +799,7 @@ double f_prime_2var(std::vector<double> x) {
|
|||||||
∂^2f/∂x∂y = 2
|
∂^2f/∂x∂y = 2
|
||||||
*/
|
*/
|
||||||
|
|
||||||
double f_mv(std::vector<double> x) {
|
real_t f_mv(std::vector<real_t> x) {
|
||||||
return x[0] * x[0] * x[0] + x[0] + x[1] * x[1] * x[1] * x[0] + x[2] * x[2] * x[1];
|
return x[0] * x[0] * x[0] + x[0] + x[1] * x[1] * x[1] * x[0] + x[2] * x[2] * x[1];
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -865,7 +866,7 @@ void MLPPTests::test_numerical_analysis() {
|
|||||||
|
|
||||||
std::cout << numAn.laplacian(f_mv, { 1, 1, 1 }) << std::endl;
|
std::cout << numAn.laplacian(f_mv, { 1, 1, 1 }) << std::endl;
|
||||||
|
|
||||||
std::vector<std::vector<std::vector<double>>> tensor;
|
std::vector<std::vector<std::vector<real_t>>> tensor;
|
||||||
tensor.push_back({ { 1, 2 }, { 1, 2 }, { 1, 2 } });
|
tensor.push_back({ { 1, 2 }, { 1, 2 }, { 1, 2 } });
|
||||||
tensor.push_back({ { 1, 2 }, { 1, 2 }, { 1, 2 } });
|
tensor.push_back({ { 1, 2 }, { 1, 2 }, { 1, 2 } });
|
||||||
|
|
||||||
@ -877,7 +878,7 @@ void MLPPTests::test_numerical_analysis() {
|
|||||||
std::cout << numAn.eulerianMethod(f_prime, { 1, 1 }, 1.5, 0.000001) << std::endl;
|
std::cout << numAn.eulerianMethod(f_prime, { 1, 1 }, 1.5, 0.000001) << std::endl;
|
||||||
std::cout << numAn.eulerianMethod(f_prime_2var, { 2, 3 }, 2.5, 0.00000001) << std::endl;
|
std::cout << numAn.eulerianMethod(f_prime_2var, { 2, 3 }, 2.5, 0.00000001) << std::endl;
|
||||||
|
|
||||||
std::vector<std::vector<double>> A = {
|
std::vector<std::vector<real_t>> A = {
|
||||||
{ 1, 0, 0, 0 },
|
{ 1, 0, 0, 0 },
|
||||||
{ 0, 0, 0, 0 },
|
{ 0, 0, 0, 0 },
|
||||||
{ 0, 0, 0, 0 },
|
{ 0, 0, 0, 0 },
|
||||||
@ -898,8 +899,8 @@ void MLPPTests::test_numerical_analysis() {
|
|||||||
std::cout << std::endl;
|
std::cout << std::endl;
|
||||||
} // Harris detector works. Life is good!
|
} // Harris detector works. Life is good!
|
||||||
|
|
||||||
std::vector<double> a = { 3, 4, 4 };
|
std::vector<real_t> a = { 3, 4, 4 };
|
||||||
std::vector<double> b = { 4, 4, 4 };
|
std::vector<real_t> b = { 4, 4, 4 };
|
||||||
alg.printVector(alg.cross(a, b));
|
alg.printVector(alg.cross(a, b));
|
||||||
}
|
}
|
||||||
void MLPPTests::test_support_vector_classification_kernel(bool ui) {
|
void MLPPTests::test_support_vector_classification_kernel(bool ui) {
|
||||||
@ -913,16 +914,16 @@ void MLPPTests::test_support_vector_classification_kernel(bool ui) {
|
|||||||
kernelSVM.gradientDescent(0.0001, 20, ui);
|
kernelSVM.gradientDescent(0.0001, 20, ui);
|
||||||
std::cout << "SCORE: " << kernelSVM.score() << std::endl;
|
std::cout << "SCORE: " << kernelSVM.score() << std::endl;
|
||||||
|
|
||||||
std::vector<std::vector<double>> linearlyIndependentMat = {
|
std::vector<std::vector<real_t>> linearlyIndependentMat = {
|
||||||
{ 1, 2, 3, 4 },
|
{ 1, 2, 3, 4 },
|
||||||
{ 234538495, 4444, 6111, 55 }
|
{ 2345384, 4444, 6111, 55 }
|
||||||
};
|
};
|
||||||
|
|
||||||
std::cout << "True of false: linearly independent?: " << std::boolalpha << alg.linearIndependenceChecker(linearlyIndependentMat) << std::endl;
|
std::cout << "True of false: linearly independent?: " << std::boolalpha << alg.linearIndependenceChecker(linearlyIndependentMat) << std::endl;
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPTests::test_mlpp_vector() {
|
void MLPPTests::test_mlpp_vector() {
|
||||||
std::vector<double> a = { 4, 3, 1, 3 };
|
std::vector<real_t> a = { 4, 3, 1, 3 };
|
||||||
|
|
||||||
Ref<MLPPVector> rv;
|
Ref<MLPPVector> rv;
|
||||||
rv.instance();
|
rv.instance();
|
||||||
@ -939,7 +940,7 @@ void MLPPTests::test_mlpp_vector() {
|
|||||||
is_approx_equals_vec(rv, rv2, "re-set_from_std_vectors test.");
|
is_approx_equals_vec(rv, rv2, "re-set_from_std_vectors test.");
|
||||||
}
|
}
|
||||||
void MLPPTests::test_mlpp_matrix() {
|
void MLPPTests::test_mlpp_matrix() {
|
||||||
std::vector<std::vector<double>> A = {
|
std::vector<std::vector<real_t>> A = {
|
||||||
{ 1, 0, 0, 0 },
|
{ 1, 0, 0, 0 },
|
||||||
{ 0, 1, 0, 0 },
|
{ 0, 1, 0, 0 },
|
||||||
{ 0, 0, 1, 0 },
|
{ 0, 0, 1, 0 },
|
||||||
@ -961,13 +962,13 @@ void MLPPTests::test_mlpp_matrix() {
|
|||||||
is_approx_equals_mat(rmat, rmat2, "re-set_from_std_vectors test.");
|
is_approx_equals_mat(rmat, rmat2, "re-set_from_std_vectors test.");
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPTests::is_approx_equalsd(double a, double b, const String &str) {
|
void MLPPTests::is_approx_equalsd(real_t a, real_t b, const String &str) {
|
||||||
if (!Math::is_equal_approx(a, b)) {
|
if (!Math::is_equal_approx(a, b)) {
|
||||||
ERR_PRINT("TEST FAILED: " + str + " Got: " + String::num(a) + " Should be: " + String::num(b));
|
ERR_PRINT("TEST FAILED: " + str + " Got: " + String::num(a) + " Should be: " + String::num(b));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPTests::is_approx_equals_dvec(const Vector<double> &a, const Vector<double> &b, const String &str) {
|
void MLPPTests::is_approx_equals_dvec(const Vector<real_t> &a, const Vector<real_t> &b, const String &str) {
|
||||||
if (a.size() != b.size()) {
|
if (a.size() != b.size()) {
|
||||||
goto IAEDVEC_FAILED;
|
goto IAEDVEC_FAILED;
|
||||||
}
|
}
|
||||||
@ -1003,7 +1004,7 @@ IAEDVEC_FAILED:
|
|||||||
ERR_PRINT(fail_str);
|
ERR_PRINT(fail_str);
|
||||||
}
|
}
|
||||||
|
|
||||||
String vmat_to_str(const Vector<Vector<double>> &a) {
|
String vmat_to_str(const Vector<Vector<real_t>> &a) {
|
||||||
String str;
|
String str;
|
||||||
|
|
||||||
str += "[ \n";
|
str += "[ \n";
|
||||||
@ -1011,7 +1012,7 @@ String vmat_to_str(const Vector<Vector<double>> &a) {
|
|||||||
for (int i = 0; i < a.size(); ++i) {
|
for (int i = 0; i < a.size(); ++i) {
|
||||||
str += " [ ";
|
str += " [ ";
|
||||||
|
|
||||||
const Vector<double> &aa = a[i];
|
const Vector<real_t> &aa = a[i];
|
||||||
|
|
||||||
for (int j = 0; j < aa.size(); ++j) {
|
for (int j = 0; j < aa.size(); ++j) {
|
||||||
str += String::num(aa[j]);
|
str += String::num(aa[j]);
|
||||||
@ -1026,14 +1027,14 @@ String vmat_to_str(const Vector<Vector<double>> &a) {
|
|||||||
return str;
|
return str;
|
||||||
}
|
}
|
||||||
|
|
||||||
void MLPPTests::is_approx_equals_dmat(const Vector<Vector<double>> &a, const Vector<Vector<double>> &b, const String &str) {
|
void MLPPTests::is_approx_equals_dmat(const Vector<Vector<real_t>> &a, const Vector<Vector<real_t>> &b, const String &str) {
|
||||||
if (a.size() != b.size()) {
|
if (a.size() != b.size()) {
|
||||||
goto IAEDMAT_FAILED;
|
goto IAEDMAT_FAILED;
|
||||||
}
|
}
|
||||||
|
|
||||||
for (int i = 0; i < a.size(); ++i) {
|
for (int i = 0; i < a.size(); ++i) {
|
||||||
const Vector<double> &aa = a[i];
|
const Vector<real_t> &aa = a[i];
|
||||||
const Vector<double> &bb = b[i];
|
const Vector<real_t> &bb = b[i];
|
||||||
|
|
||||||
if (aa.size() != bb.size()) {
|
if (aa.size() != bb.size()) {
|
||||||
goto IAEDMAT_FAILED;
|
goto IAEDMAT_FAILED;
|
||||||
@ -1066,8 +1067,8 @@ void MLPPTests::is_approx_equals_mat(Ref<MLPPMatrix> a, Ref<MLPPMatrix> b, const
|
|||||||
|
|
||||||
int ds = a->data_size();
|
int ds = a->data_size();
|
||||||
|
|
||||||
const double *aa = a->ptr();
|
const real_t *aa = a->ptr();
|
||||||
const double *bb = b->ptr();
|
const real_t *bb = b->ptr();
|
||||||
|
|
||||||
if (a->size() != b->size()) {
|
if (a->size() != b->size()) {
|
||||||
goto IAEMAT_FAILED;
|
goto IAEMAT_FAILED;
|
||||||
|
@ -4,6 +4,8 @@
|
|||||||
// TODO port this class to use the test module once it's working
|
// TODO port this class to use the test module once it's working
|
||||||
// Also don't forget to remove it's bindings
|
// Also don't forget to remove it's bindings
|
||||||
|
|
||||||
|
#include "core/math/math_defs.h"
|
||||||
|
|
||||||
#include "core/containers/vector.h"
|
#include "core/containers/vector.h"
|
||||||
|
|
||||||
#include "core/object/reference.h"
|
#include "core/object/reference.h"
|
||||||
@ -64,9 +66,9 @@ public:
|
|||||||
void test_mlpp_vector();
|
void test_mlpp_vector();
|
||||||
void test_mlpp_matrix();
|
void test_mlpp_matrix();
|
||||||
|
|
||||||
void is_approx_equalsd(double a, double b, const String &str);
|
void is_approx_equalsd(real_t a, real_t b, const String &str);
|
||||||
void is_approx_equals_dvec(const Vector<double> &a, const Vector<double> &b, const String &str);
|
void is_approx_equals_dvec(const Vector<real_t> &a, const Vector<real_t> &b, const String &str);
|
||||||
void is_approx_equals_dmat(const Vector<Vector<double>> &a, const Vector<Vector<double>> &b, const String &str);
|
void is_approx_equals_dmat(const Vector<Vector<real_t>> &a, const Vector<Vector<real_t>> &b, const String &str);
|
||||||
|
|
||||||
void is_approx_equals_mat(Ref<MLPPMatrix> a, Ref<MLPPMatrix> b, const String &str);
|
void is_approx_equals_mat(Ref<MLPPMatrix> a, Ref<MLPPMatrix> b, const String &str);
|
||||||
void is_approx_equals_vec(Ref<MLPPVector> a, Ref<MLPPVector> b, const String &str);
|
void is_approx_equals_vec(Ref<MLPPVector> a, Ref<MLPPVector> b, const String &str);
|
||||||
|
Loading…
Reference in New Issue
Block a user