#ifndef MLPP_MANN_H #define MLPP_MANN_H // // MANN.hpp // // Created by Marc Melikyan on 11/4/20. // #include "core/math/math_defs.h" #include "core/object/reference.h" #include "../regularization/reg.h" #include "../lin_alg/mlpp_matrix.h" #include "../lin_alg/mlpp_vector.h" #include "../hidden_layer/hidden_layer.h" #include "../multi_output_layer/multi_output_layer.h" class MLPPMANN : public Reference { GDCLASS(MLPPMANN, Reference); public: /* Ref get_input_set(); void set_input_set(const Ref &val); Ref get_output_set(); void set_output_set(const Ref &val); */ Ref model_set_test(const Ref &X); Ref model_test(const Ref &x); void gradient_descent(real_t learning_rate, int max_epoch, bool ui = false); real_t score(); void save(const String &file_name); void add_layer(int n_hidden, MLPPActivation::ActivationFunction activation, MLPPUtilities::WeightDistributionType weight_init = MLPPUtilities::WEIGHT_DISTRIBUTION_TYPE_DEFAULT, MLPPReg::RegularizationType reg = MLPPReg::REGULARIZATION_TYPE_NONE, real_t lambda = 0.5, real_t alpha = 0.5); void add_output_layer(MLPPActivation::ActivationFunction activation, MLPPCost::CostTypes loss, MLPPUtilities::WeightDistributionType weight_init = MLPPUtilities::WEIGHT_DISTRIBUTION_TYPE_DEFAULT, MLPPReg::RegularizationType reg = MLPPReg::REGULARIZATION_TYPE_NONE, real_t lambda = 0.5, real_t alpha = 0.5); bool is_initialized(); void initialize(); MLPPMANN(const Ref &p_input_set, const Ref &p_output_set); MLPPMANN(); ~MLPPMANN(); private: real_t cost(const Ref &y_hat, const Ref &y); void forward_pass(); static void _bind_methods(); Ref _input_set; Ref _output_set; Ref _y_hat; Vector> _network; Ref _output_layer; int _n; int _k; int _n_output; bool _initialized; }; #endif /* MANN_hpp */