pmlpp/gan/gan.h

119 lines
5.0 KiB
C
Raw Normal View History

#ifndef MLPP_GAN_H
#define MLPP_GAN_H
2023-01-24 18:57:18 +01:00
2023-12-30 00:41:59 +01:00
/*************************************************************************/
/* gan.h */
/*************************************************************************/
/* This file is part of: */
/* PMLPP Machine Learning Library */
/* https://github.com/Relintai/pmlpp */
/*************************************************************************/
2023-12-30 00:43:39 +01:00
/* Copyright (c) 2023-present Péter Magyar. */
2023-12-30 00:41:59 +01:00
/* Copyright (c) 2022-2023 Marc Melikyan */
/* */
/* Permission is hereby granted, free of charge, to any person obtaining */
/* a copy of this software and associated documentation files (the */
/* "Software"), to deal in the Software without restriction, including */
/* without limitation the rights to use, copy, modify, merge, publish, */
/* distribute, sublicense, and/or sell copies of the Software, and to */
/* permit persons to whom the Software is furnished to do so, subject to */
/* the following conditions: */
/* */
/* The above copyright notice and this permission notice shall be */
/* included in all copies or substantial portions of the Software. */
/* */
/* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, */
/* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF */
/* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.*/
/* IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY */
/* CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, */
/* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE */
/* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */
/*************************************************************************/
2024-01-25 13:42:45 +01:00
#ifdef USING_SFW
#include "sfw.h"
#else
2023-01-27 13:01:16 +01:00
#include "core/math/math_defs.h"
2023-02-12 16:28:00 +01:00
#include "core/object/reference.h"
2024-01-25 13:42:45 +01:00
#endif
2023-02-12 16:28:00 +01:00
2023-01-24 18:12:23 +01:00
#include "../hidden_layer/hidden_layer.h"
#include "../output_layer/output_layer.h"
2023-04-30 17:39:00 +02:00
#include "../lin_alg/mlpp_tensor3.h"
2023-02-16 17:32:35 +01:00
#include "../activation/activation.h"
#include "../utilities/utilities.h"
2023-01-24 19:00:54 +01:00
2023-02-12 16:28:00 +01:00
class MLPPGAN : public Reference {
GDCLASS(MLPPGAN, Reference);
2023-01-24 19:00:54 +01:00
public:
2023-02-12 10:05:17 +01:00
/*
Ref<MLPPMatrix> get_input_set();
void set_input_set(const Ref<MLPPMatrix> &val);
Ref<MLPPVector> get_output_set();
void set_output_set(const Ref<MLPPVector> &val);
int get_k();
void set_k(const int val);
*/
2023-02-16 17:32:35 +01:00
Ref<MLPPMatrix> generate_example(int n);
2023-02-12 10:05:17 +01:00
void gradient_descent(real_t learning_rate, int max_epoch, bool ui = false);
2023-01-27 13:01:16 +01:00
real_t score();
2023-02-16 17:32:35 +01:00
void save(const String &file_name);
2023-02-12 10:05:17 +01:00
2023-02-16 17:32:35 +01:00
void add_layer(int n_hidden, MLPPActivation::ActivationFunction activation, MLPPUtilities::WeightDistributionType weight_init = MLPPUtilities::WEIGHT_DISTRIBUTION_TYPE_DEFAULT, MLPPReg::RegularizationType reg = MLPPReg::REGULARIZATION_TYPE_NONE, real_t lambda = 0.5, real_t alpha = 0.5);
void add_output_layer(MLPPUtilities::WeightDistributionType weight_init = MLPPUtilities::WEIGHT_DISTRIBUTION_TYPE_DEFAULT, MLPPReg::RegularizationType reg = MLPPReg::REGULARIZATION_TYPE_NONE, real_t lambda = 0.5, real_t alpha = 0.5);
2023-02-12 10:05:17 +01:00
2023-02-16 17:32:35 +01:00
MLPPGAN(real_t k, const Ref<MLPPMatrix> &output_set);
2023-02-12 10:05:17 +01:00
MLPPGAN();
~MLPPGAN();
protected:
2023-02-16 17:32:35 +01:00
Ref<MLPPMatrix> model_set_test_generator(const Ref<MLPPMatrix> &X); // Evaluator for the generator of the gan.
Ref<MLPPVector> model_set_test_discriminator(const Ref<MLPPMatrix> &X); // Evaluator for the discriminator of the gan.
2023-02-12 10:05:17 +01:00
2023-02-16 17:32:35 +01:00
real_t cost(const Ref<MLPPVector> &y_hat, const Ref<MLPPVector> &y);
2023-02-12 10:05:17 +01:00
void forward_pass();
2023-04-30 17:39:00 +02:00
void update_discriminator_parameters(const Ref<MLPPTensor3> &hidden_layer_updations, const Ref<MLPPVector> &output_layer_updation, real_t learning_rate);
void update_generator_parameters(const Ref<MLPPTensor3> &hidden_layer_updations, real_t learning_rate);
2023-02-16 17:32:35 +01:00
struct ComputeDiscriminatorGradientsResult {
2023-04-30 17:39:00 +02:00
Ref<MLPPTensor3> cumulative_hidden_layer_w_grad; // Tensor containing ALL hidden grads.
2023-02-16 17:32:35 +01:00
Ref<MLPPVector> output_w_grad;
2023-04-30 17:39:00 +02:00
ComputeDiscriminatorGradientsResult() {
cumulative_hidden_layer_w_grad.instance();
output_w_grad.instance();
}
2023-02-16 17:32:35 +01:00
};
2023-02-16 17:32:35 +01:00
ComputeDiscriminatorGradientsResult compute_discriminator_gradients(const Ref<MLPPVector> &y_hat, const Ref<MLPPVector> &output_set);
2023-04-30 17:39:00 +02:00
Ref<MLPPTensor3> compute_generator_gradients(const Ref<MLPPVector> &y_hat, const Ref<MLPPVector> &output_set);
2023-02-16 17:32:35 +01:00
void print_ui(int epoch, real_t cost_prev, const Ref<MLPPVector> &y_hat, const Ref<MLPPVector> &output_set);
2023-02-12 10:05:17 +01:00
static void _bind_methods();
2023-02-16 17:32:35 +01:00
Ref<MLPPMatrix> _output_set;
Ref<MLPPVector> _y_hat;
2023-02-16 17:32:35 +01:00
Vector<Ref<MLPPHiddenLayer>> _network;
Ref<MLPPOutputLayer> _output_layer;
2023-02-12 10:05:17 +01:00
int _n;
int _k;
2023-01-24 19:00:54 +01:00
};
2023-01-24 19:20:18 +01:00
#endif /* GAN_hpp */