pmlpp/mlpp/output_layer/output_layer.h

57 lines
1.6 KiB
C++

//
// OutputLayer.hpp
//
// Created by Marc Melikyan on 11/4/20.
//
#ifndef OutputLayer_hpp
#define OutputLayer_hpp
#include "Activation/Activation.hpp"
#include "Cost/Cost.hpp"
#include <vector>
#include <map>
#include <string>
namespace MLPP {
class OutputLayer{
public:
OutputLayer(int n_hidden, std::string activation, std::string cost, std::vector<std::vector<double>> input, std::string weightInit, std::string reg, double lambda, double alpha);
int n_hidden;
std::string activation;
std::string cost;
std::vector<std::vector<double>> input;
std::vector<double> weights;
double bias;
std::vector<double> z;
std::vector<double> a;
std::map<std::string, std::vector<double> (Activation::*)(std::vector<double>, bool)> activation_map;
std::map<std::string, double (Activation::*)(double, bool)> activationTest_map;
std::map<std::string, double (Cost::*)(std::vector<double>, std::vector<double>)> cost_map;
std::map<std::string, std::vector<double> (Cost::*)(std::vector<double>, std::vector<double>)> costDeriv_map;
double z_test;
double a_test;
std::vector<double> delta;
// Regularization Params
std::string reg;
double lambda; /* Regularization Parameter */
double alpha; /* This is the controlling param for Elastic Net*/
std::string weightInit;
void forwardPass();
void Test(std::vector<double> x);
};
}
#endif /* OutputLayer_hpp */