2023-01-23 21:13:26 +01:00
|
|
|
//
|
|
|
|
// Reg.cpp
|
|
|
|
//
|
|
|
|
// Created by Marc Melikyan on 1/16/21.
|
|
|
|
//
|
|
|
|
|
2023-01-24 18:12:23 +01:00
|
|
|
#include "reg.h"
|
|
|
|
#include "../activation/activation.h"
|
2023-01-24 19:00:54 +01:00
|
|
|
#include "../lin_alg/lin_alg.h"
|
|
|
|
#include <iostream>
|
|
|
|
#include <random>
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-24 19:20:18 +01:00
|
|
|
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-27 13:01:16 +01:00
|
|
|
real_t MLPPReg::regTerm(std::vector<real_t> weights, real_t lambda, real_t alpha, std::string reg) {
|
2023-01-24 19:00:54 +01:00
|
|
|
if (reg == "Ridge") {
|
2023-01-27 13:01:16 +01:00
|
|
|
real_t reg = 0;
|
2023-01-24 19:00:54 +01:00
|
|
|
for (int i = 0; i < weights.size(); i++) {
|
|
|
|
reg += weights[i] * weights[i];
|
|
|
|
}
|
|
|
|
return reg * lambda / 2;
|
|
|
|
} else if (reg == "Lasso") {
|
2023-01-27 13:01:16 +01:00
|
|
|
real_t reg = 0;
|
2023-01-24 19:00:54 +01:00
|
|
|
for (int i = 0; i < weights.size(); i++) {
|
|
|
|
reg += abs(weights[i]);
|
|
|
|
}
|
|
|
|
return reg * lambda;
|
|
|
|
} else if (reg == "ElasticNet") {
|
2023-01-27 13:01:16 +01:00
|
|
|
real_t reg = 0;
|
2023-01-24 19:00:54 +01:00
|
|
|
for (int i = 0; i < weights.size(); i++) {
|
|
|
|
reg += alpha * abs(weights[i]); // Lasso Reg
|
|
|
|
reg += ((1 - alpha) / 2) * weights[i] * weights[i]; // Ridge Reg
|
|
|
|
}
|
|
|
|
return reg * lambda;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-27 13:01:16 +01:00
|
|
|
real_t MLPPReg::regTerm(std::vector<std::vector<real_t>> weights, real_t lambda, real_t alpha, std::string reg) {
|
2023-01-24 19:00:54 +01:00
|
|
|
if (reg == "Ridge") {
|
2023-01-27 13:01:16 +01:00
|
|
|
real_t reg = 0;
|
2023-01-24 19:00:54 +01:00
|
|
|
for (int i = 0; i < weights.size(); i++) {
|
|
|
|
for (int j = 0; j < weights[i].size(); j++) {
|
|
|
|
reg += weights[i][j] * weights[i][j];
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return reg * lambda / 2;
|
|
|
|
} else if (reg == "Lasso") {
|
2023-01-27 13:01:16 +01:00
|
|
|
real_t reg = 0;
|
2023-01-24 19:00:54 +01:00
|
|
|
for (int i = 0; i < weights.size(); i++) {
|
|
|
|
for (int j = 0; j < weights[i].size(); j++) {
|
|
|
|
reg += abs(weights[i][j]);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return reg * lambda;
|
|
|
|
} else if (reg == "ElasticNet") {
|
2023-01-27 13:01:16 +01:00
|
|
|
real_t reg = 0;
|
2023-01-24 19:00:54 +01:00
|
|
|
for (int i = 0; i < weights.size(); i++) {
|
|
|
|
for (int j = 0; j < weights[i].size(); j++) {
|
|
|
|
reg += alpha * abs(weights[i][j]); // Lasso Reg
|
|
|
|
reg += ((1 - alpha) / 2) * weights[i][j] * weights[i][j]; // Ridge Reg
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return reg * lambda;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-27 13:01:16 +01:00
|
|
|
std::vector<real_t> MLPPReg::regWeights(std::vector<real_t> weights, real_t lambda, real_t alpha, std::string reg) {
|
2023-01-25 00:29:02 +01:00
|
|
|
MLPPLinAlg alg;
|
2023-01-24 19:00:54 +01:00
|
|
|
if (reg == "WeightClipping") {
|
|
|
|
return regDerivTerm(weights, lambda, alpha, reg);
|
|
|
|
}
|
|
|
|
return alg.subtraction(weights, regDerivTerm(weights, lambda, alpha, reg));
|
|
|
|
// for(int i = 0; i < weights.size(); i++){
|
|
|
|
// weights[i] -= regDerivTerm(weights, lambda, alpha, reg, i);
|
|
|
|
// }
|
|
|
|
// return weights;
|
|
|
|
}
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-27 13:01:16 +01:00
|
|
|
std::vector<std::vector<real_t>> MLPPReg::regWeights(std::vector<std::vector<real_t>> weights, real_t lambda, real_t alpha, std::string reg) {
|
2023-01-25 00:29:02 +01:00
|
|
|
MLPPLinAlg alg;
|
2023-01-24 19:00:54 +01:00
|
|
|
if (reg == "WeightClipping") {
|
|
|
|
return regDerivTerm(weights, lambda, alpha, reg);
|
|
|
|
}
|
|
|
|
return alg.subtraction(weights, regDerivTerm(weights, lambda, alpha, reg));
|
|
|
|
// for(int i = 0; i < weights.size(); i++){
|
|
|
|
// for(int j = 0; j < weights[i].size(); j++){
|
|
|
|
// weights[i][j] -= regDerivTerm(weights, lambda, alpha, reg, i, j);
|
|
|
|
// }
|
|
|
|
// }
|
|
|
|
// return weights;
|
|
|
|
}
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-27 13:01:16 +01:00
|
|
|
std::vector<real_t> MLPPReg::regDerivTerm(std::vector<real_t> weights, real_t lambda, real_t alpha, std::string reg) {
|
|
|
|
std::vector<real_t> regDeriv;
|
2023-01-24 19:00:54 +01:00
|
|
|
regDeriv.resize(weights.size());
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-24 19:00:54 +01:00
|
|
|
for (int i = 0; i < regDeriv.size(); i++) {
|
|
|
|
regDeriv[i] = regDerivTerm(weights, lambda, alpha, reg, i);
|
|
|
|
}
|
|
|
|
return regDeriv;
|
|
|
|
}
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-27 13:01:16 +01:00
|
|
|
std::vector<std::vector<real_t>> MLPPReg::regDerivTerm(std::vector<std::vector<real_t>> weights, real_t lambda, real_t alpha, std::string reg) {
|
|
|
|
std::vector<std::vector<real_t>> regDeriv;
|
2023-01-24 19:00:54 +01:00
|
|
|
regDeriv.resize(weights.size());
|
|
|
|
for (int i = 0; i < regDeriv.size(); i++) {
|
|
|
|
regDeriv[i].resize(weights[0].size());
|
|
|
|
}
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-24 19:00:54 +01:00
|
|
|
for (int i = 0; i < regDeriv.size(); i++) {
|
|
|
|
for (int j = 0; j < regDeriv[i].size(); j++) {
|
|
|
|
regDeriv[i][j] = regDerivTerm(weights, lambda, alpha, reg, i, j);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return regDeriv;
|
|
|
|
}
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-27 13:01:16 +01:00
|
|
|
real_t MLPPReg::regDerivTerm(std::vector<real_t> weights, real_t lambda, real_t alpha, std::string reg, int j) {
|
2023-01-24 19:23:30 +01:00
|
|
|
MLPPActivation act;
|
2023-01-24 19:00:54 +01:00
|
|
|
if (reg == "Ridge") {
|
|
|
|
return lambda * weights[j];
|
|
|
|
} else if (reg == "Lasso") {
|
|
|
|
return lambda * act.sign(weights[j]);
|
|
|
|
} else if (reg == "ElasticNet") {
|
|
|
|
return alpha * lambda * act.sign(weights[j]) + (1 - alpha) * lambda * weights[j];
|
|
|
|
} else if (reg == "WeightClipping") { // Preparation for Wasserstein GANs.
|
|
|
|
// We assume lambda is the lower clipping threshold, while alpha is the higher clipping threshold.
|
|
|
|
// alpha > lambda.
|
|
|
|
if (weights[j] > alpha) {
|
|
|
|
return alpha;
|
|
|
|
} else if (weights[j] < lambda) {
|
|
|
|
return lambda;
|
|
|
|
} else {
|
|
|
|
return weights[j];
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
2023-01-23 21:13:26 +01:00
|
|
|
|
2023-01-27 13:01:16 +01:00
|
|
|
real_t MLPPReg::regDerivTerm(std::vector<std::vector<real_t>> weights, real_t lambda, real_t alpha, std::string reg, int i, int j) {
|
2023-01-24 19:23:30 +01:00
|
|
|
MLPPActivation act;
|
2023-01-24 19:00:54 +01:00
|
|
|
if (reg == "Ridge") {
|
|
|
|
return lambda * weights[i][j];
|
|
|
|
} else if (reg == "Lasso") {
|
|
|
|
return lambda * act.sign(weights[i][j]);
|
|
|
|
} else if (reg == "ElasticNet") {
|
|
|
|
return alpha * lambda * act.sign(weights[i][j]) + (1 - alpha) * lambda * weights[i][j];
|
|
|
|
} else if (reg == "WeightClipping") { // Preparation for Wasserstein GANs.
|
|
|
|
// We assume lambda is the lower clipping threshold, while alpha is the higher clipping threshold.
|
|
|
|
// alpha > lambda.
|
|
|
|
if (weights[i][j] > alpha) {
|
|
|
|
return alpha;
|
|
|
|
} else if (weights[i][j] < lambda) {
|
|
|
|
return lambda;
|
|
|
|
} else {
|
|
|
|
return weights[i][j];
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
return 0;
|
|
|
|
}
|
2023-01-23 21:13:26 +01:00
|
|
|
}
|
2023-01-24 19:20:18 +01:00
|
|
|
|