2023-01-24 18:57:18 +01:00
|
|
|
|
|
|
|
#ifndef MLPP_WGAN_H
|
|
|
|
#define MLPP_WGAN_H
|
|
|
|
|
2023-01-23 21:13:26 +01:00
|
|
|
//
|
|
|
|
// WGAN.hpp
|
|
|
|
//
|
|
|
|
// Created by Marc Melikyan on 11/4/20.
|
|
|
|
//
|
|
|
|
|
2023-02-05 18:46:12 +01:00
|
|
|
#include "core/containers/vector.h"
|
|
|
|
#include "core/math/math_defs.h"
|
|
|
|
#include "core/string/ustring.h"
|
|
|
|
|
|
|
|
#include "core/object/reference.h"
|
|
|
|
|
|
|
|
#include "../lin_alg/mlpp_matrix.h"
|
|
|
|
#include "../lin_alg/mlpp_vector.h"
|
|
|
|
|
2023-01-24 18:12:23 +01:00
|
|
|
#include "../hidden_layer/hidden_layer.h"
|
|
|
|
#include "../output_layer/output_layer.h"
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-02-06 02:36:22 +01:00
|
|
|
#include "../activation/activation.h"
|
|
|
|
#include "../cost/cost.h"
|
|
|
|
#include "../regularization/reg.h"
|
|
|
|
#include "../utilities/utilities.h"
|
|
|
|
|
2023-01-23 21:13:26 +01:00
|
|
|
#include <string>
|
2023-01-24 19:00:54 +01:00
|
|
|
#include <tuple>
|
|
|
|
#include <vector>
|
|
|
|
|
2023-02-05 18:46:12 +01:00
|
|
|
class MLPPWGAN : public Reference {
|
|
|
|
GDCLASS(MLPPWGAN, Reference);
|
|
|
|
|
2023-02-05 18:16:34 +01:00
|
|
|
public:
|
2023-02-06 12:20:52 +01:00
|
|
|
Ref<MLPPMatrix> get_output_set();
|
|
|
|
void set_output_set(const Ref<MLPPMatrix> &val);
|
|
|
|
|
|
|
|
int get_k() const;
|
|
|
|
void set_k(const int val);
|
|
|
|
|
2023-02-06 02:36:22 +01:00
|
|
|
Ref<MLPPMatrix> generate_example(int n);
|
|
|
|
void gradient_descent(real_t learning_rate, int max_epoch, bool ui = false);
|
2023-02-05 18:16:34 +01:00
|
|
|
real_t score();
|
2023-02-06 02:36:22 +01:00
|
|
|
void save(const String &file_name);
|
2023-02-05 18:16:34 +01:00
|
|
|
|
2023-02-06 02:36:22 +01:00
|
|
|
void add_layer(int n_hidden, MLPPActivation::ActivationFunction activation, MLPPUtilities::WeightDistributionType weight_init = MLPPUtilities::WEIGHT_DISTRIBUTION_TYPE_DEFAULT, MLPPReg::RegularizationType reg = MLPPReg::REGULARIZATION_TYPE_NONE, real_t lambda = 0.5, real_t alpha = 0.5);
|
|
|
|
void add_output_layer(MLPPUtilities::WeightDistributionType weight_init = MLPPUtilities::WEIGHT_DISTRIBUTION_TYPE_DEFAULT, MLPPReg::RegularizationType reg = MLPPReg::REGULARIZATION_TYPE_NONE, real_t lambda = 0.5, real_t alpha = 0.5);
|
2023-02-05 18:16:34 +01:00
|
|
|
|
2023-02-06 02:36:22 +01:00
|
|
|
MLPPWGAN(real_t k, const Ref<MLPPMatrix> &output_set);
|
2023-02-05 18:46:12 +01:00
|
|
|
|
|
|
|
MLPPWGAN();
|
|
|
|
~MLPPWGAN();
|
|
|
|
|
|
|
|
protected:
|
2023-02-06 02:36:22 +01:00
|
|
|
Ref<MLPPMatrix> model_set_test_generator(const Ref<MLPPMatrix> &X); // Evaluator for the generator of the WGAN.
|
|
|
|
Ref<MLPPVector> model_set_test_discriminator(const Ref<MLPPMatrix> &X); // Evaluator for the discriminator of the WGAN.
|
2023-02-05 18:46:12 +01:00
|
|
|
|
2023-02-06 02:36:22 +01:00
|
|
|
real_t cost(const Ref<MLPPVector> &y_hat, const Ref<MLPPVector> &y);
|
2023-02-05 18:16:34 +01:00
|
|
|
|
2023-02-05 18:46:12 +01:00
|
|
|
void forward_pass();
|
2023-02-06 02:36:22 +01:00
|
|
|
void update_discriminator_parameters(Vector<Ref<MLPPMatrix>> hidden_layer_updations, const Ref<MLPPVector> &output_layer_updation, real_t learning_rate);
|
|
|
|
void update_generator_parameters(Vector<Ref<MLPPMatrix>> hidden_layer_updations, real_t learning_rate);
|
|
|
|
|
|
|
|
struct DiscriminatorGradientResult {
|
|
|
|
Vector<Ref<MLPPMatrix>> cumulative_hidden_layer_w_grad; // Tensor containing ALL hidden grads.
|
|
|
|
Ref<MLPPVector> output_w_grad;
|
|
|
|
};
|
|
|
|
|
|
|
|
DiscriminatorGradientResult compute_discriminator_gradients(const Ref<MLPPVector> &y_hat, const Ref<MLPPVector> &output_set);
|
|
|
|
Vector<Ref<MLPPMatrix>> compute_generator_gradients(const Ref<MLPPVector> &y_hat, const Ref<MLPPVector> &output_set);
|
2023-02-05 18:16:34 +01:00
|
|
|
|
2023-02-06 02:36:22 +01:00
|
|
|
void handle_ui(int epoch, real_t cost_prev, const Ref<MLPPVector> &y_hat, const Ref<MLPPVector> &output_set);
|
2023-02-05 18:16:34 +01:00
|
|
|
|
2023-02-05 18:46:12 +01:00
|
|
|
static void _bind_methods();
|
2023-02-05 18:16:34 +01:00
|
|
|
|
2023-02-06 02:36:22 +01:00
|
|
|
Ref<MLPPMatrix> output_set;
|
|
|
|
Ref<MLPPVector> y_hat;
|
2023-02-05 18:16:34 +01:00
|
|
|
|
2023-02-06 02:36:22 +01:00
|
|
|
Vector<Ref<MLPPHiddenLayer>> network;
|
|
|
|
Ref<MLPPOutputLayer> output_layer;
|
2023-02-05 18:16:34 +01:00
|
|
|
|
|
|
|
int n;
|
|
|
|
int k;
|
|
|
|
};
|
|
|
|
|
2023-02-05 17:05:46 +01:00
|
|
|
class MLPPWGANOld {
|
2023-01-24 19:00:54 +01:00
|
|
|
public:
|
2023-02-05 17:05:46 +01:00
|
|
|
MLPPWGANOld(real_t k, std::vector<std::vector<real_t>> outputSet);
|
|
|
|
~MLPPWGANOld();
|
2023-01-27 13:01:16 +01:00
|
|
|
std::vector<std::vector<real_t>> generateExample(int n);
|
2023-02-04 13:59:26 +01:00
|
|
|
void gradientDescent(real_t learning_rate, int max_epoch, bool UI = false);
|
2023-01-27 13:01:16 +01:00
|
|
|
real_t score();
|
2023-01-24 19:00:54 +01:00
|
|
|
void save(std::string fileName);
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-27 13:01:16 +01:00
|
|
|
void addLayer(int n_hidden, std::string activation, std::string weightInit = "Default", std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
|
|
|
void addOutputLayer(std::string weightInit = "Default", std::string reg = "None", real_t lambda = 0.5, real_t alpha = 0.5);
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-24 19:00:54 +01:00
|
|
|
private:
|
2023-01-27 13:01:16 +01:00
|
|
|
std::vector<std::vector<real_t>> modelSetTestGenerator(std::vector<std::vector<real_t>> X); // Evaluator for the generator of the WGAN.
|
|
|
|
std::vector<real_t> modelSetTestDiscriminator(std::vector<std::vector<real_t>> X); // Evaluator for the discriminator of the WGAN.
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-27 13:01:16 +01:00
|
|
|
real_t Cost(std::vector<real_t> y_hat, std::vector<real_t> y);
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-24 19:00:54 +01:00
|
|
|
void forwardPass();
|
2023-01-27 13:01:16 +01:00
|
|
|
void updateDiscriminatorParameters(std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations, std::vector<real_t> outputLayerUpdation, real_t learning_rate);
|
|
|
|
void updateGeneratorParameters(std::vector<std::vector<std::vector<real_t>>> hiddenLayerUpdations, real_t learning_rate);
|
|
|
|
std::tuple<std::vector<std::vector<std::vector<real_t>>>, std::vector<real_t>> computeDiscriminatorGradients(std::vector<real_t> y_hat, std::vector<real_t> outputSet);
|
|
|
|
std::vector<std::vector<std::vector<real_t>>> computeGeneratorGradients(std::vector<real_t> y_hat, std::vector<real_t> outputSet);
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-27 13:01:16 +01:00
|
|
|
void UI(int epoch, real_t cost_prev, std::vector<real_t> y_hat, std::vector<real_t> outputSet);
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-27 13:01:16 +01:00
|
|
|
std::vector<std::vector<real_t>> outputSet;
|
|
|
|
std::vector<real_t> y_hat;
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-30 16:56:16 +01:00
|
|
|
std::vector<MLPPOldHiddenLayer> network;
|
|
|
|
MLPPOldOutputLayer *outputLayer;
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-24 19:00:54 +01:00
|
|
|
int n;
|
|
|
|
int k;
|
|
|
|
};
|
2023-01-24 19:20:18 +01:00
|
|
|
|
2023-01-23 21:13:26 +01:00
|
|
|
#endif /* WGAN_hpp */
|