Skip to content
Snippets Groups Projects
Commit 649836a8 authored by Cédric Traizet's avatar Cédric Traizet
Browse files

m_Dimension is now an attribute of ML models, it is now possible to change the dim of a pca model

parent 717f6f93
No related branches found
No related tags found
No related merge requests found
...@@ -105,6 +105,7 @@ private: ...@@ -105,6 +105,7 @@ private:
MeasurementType meanMeasurementVector; MeasurementType meanMeasurementVector;
MeasurementType stddevMeasurementVector; MeasurementType stddevMeasurementVector;
if (HasValue("io.stats") && IsParameterEnabled("io.stats")) if (HasValue("io.stats") && IsParameterEnabled("io.stats"))
{ {
StatisticsReader::Pointer statisticsReader = StatisticsReader::New(); StatisticsReader::Pointer statisticsReader = StatisticsReader::New();
......
...@@ -107,6 +107,10 @@ class CbDimensionalityReductionVector : public Application ...@@ -107,6 +107,10 @@ class CbDimensionalityReductionVector : public Application
"If not given, the input vector data file is updated."); "If not given, the input vector data file is updated.");
MandatoryOff("out"); MandatoryOff("out");
AddParameter(ParameterType_Int, "pcadim", "Principal component"); //
SetParameterDescription("pcadim","This optional parameter can be set to reduce the number of eignevectors used in the PCA model file."); //
MandatoryOff("pcadim");
// Doc example parameter settings // Doc example parameter settings
SetDocExampleParameterValue("in", "vectorData.shp"); SetDocExampleParameterValue("in", "vectorData.shp");
SetDocExampleParameterValue("instat", "meanVar.xml"); SetDocExampleParameterValue("instat", "meanVar.xml");
...@@ -211,13 +215,21 @@ class CbDimensionalityReductionVector : public Application ...@@ -211,13 +215,21 @@ class CbDimensionalityReductionVector : public Application
/** Read the model */ /** Read the model */
std::cout << "create the fact ?" << std::endl;
m_Model = DimensionalityReductionModelFactoryType::CreateDimensionalityReductionModel(GetParameterString("model"), m_Model = DimensionalityReductionModelFactoryType::CreateDimensionalityReductionModel(GetParameterString("model"),
DimensionalityReductionModelFactoryType::ReadMode); DimensionalityReductionModelFactoryType::ReadMode);
if (m_Model.IsNull()) if (m_Model.IsNull())
{ {
otbAppLogFATAL(<< "Error when loading model " << GetParameterString("model") << " : unsupported model type"); otbAppLogFATAL(<< "Error when loading model " << GetParameterString("model") << " : unsupported model type");
} }
if (HasValue("pcadim") && IsParameterEnabled("pcadim"))
{
int dimension = GetParameterInt("pcadim");
m_Model->SetDimension(dimension );
std::cout << "yo" << std::endl;
}
m_Model->Load(GetParameterString("model")); m_Model->Load(GetParameterString("model"));
otbAppLogINFO("Model loaded"); otbAppLogINFO("Model loaded");
......
...@@ -4,6 +4,8 @@ ...@@ -4,6 +4,8 @@
#include "otbMachineLearningModelTraits.h" #include "otbMachineLearningModelTraits.h"
#include "otbMachineLearningModel.h" #include "otbMachineLearningModel.h"
#include <shark/Algorithms/StoppingCriteria/AbstractStoppingCriterion.h>
namespace otb namespace otb
{ {
template <class TInputValue, class AutoencoderType> template <class TInputValue, class AutoencoderType>
...@@ -35,12 +37,15 @@ public: ...@@ -35,12 +37,15 @@ public:
itkNewMacro(Self); itkNewMacro(Self);
itkTypeMacro(AutoencoderModel, DimensionalityReductionModel); itkTypeMacro(AutoencoderModel, DimensionalityReductionModel);
unsigned int GetDimension() {return m_NumberOfHiddenNeurons[m_net.size()-1];}; // Override the Dimensionality Reduction model method, it is used in the dimensionality reduction filter to set the output image size //unsigned int GetDimension() {return m_NumberOfHiddenNeurons[m_net.size()-1];}; // Override the Dimensionality Reduction model method, it is used in the dimensionality reduction filter to set the output image size
itkGetMacro(NumberOfHiddenNeurons,itk::Array<unsigned int>); itkGetMacro(NumberOfHiddenNeurons,itk::Array<unsigned int>);
itkSetMacro(NumberOfHiddenNeurons,itk::Array<unsigned int>); itkSetMacro(NumberOfHiddenNeurons,itk::Array<unsigned int>);
itkGetMacro(NumberOfIterations,unsigned int); itkGetMacro(NumberOfIterations,unsigned int);
itkSetMacro(NumberOfIterations,unsigned int); itkSetMacro(NumberOfIterations,unsigned int);
itkGetMacro(Epsilon,double);
itkSetMacro(Epsilon,double);
itkGetMacro(Regularization,itk::Array<double>); itkGetMacro(Regularization,itk::Array<double>);
itkSetMacro(Regularization,itk::Array<double>); itkSetMacro(Regularization,itk::Array<double>);
...@@ -61,8 +66,12 @@ public: ...@@ -61,8 +66,12 @@ public:
void Load(const std::string & filename, const std::string & name="") ITK_OVERRIDE; void Load(const std::string & filename, const std::string & name="") ITK_OVERRIDE;
void Train() ITK_OVERRIDE; void Train() ITK_OVERRIDE;
void TrainOneLayer(unsigned int,double, double, shark::Data<shark::RealVector> &);
void TrainOneSparseLayer(unsigned int,double, double,double, shark::Data<shark::RealVector> &); template <class T>
void TrainOneLayer(shark::AbstractStoppingCriterion<T> & criterion, unsigned int,double, double, shark::Data<shark::RealVector> &);
template <class T>
void TrainOneSparseLayer(shark::AbstractStoppingCriterion<T> & criterion, unsigned int,double, double,double, shark::Data<shark::RealVector> &);
protected: protected:
AutoencoderModel(); AutoencoderModel();
...@@ -80,7 +89,8 @@ private: ...@@ -80,7 +89,8 @@ private:
itk::Array<unsigned int> m_NumberOfHiddenNeurons; itk::Array<unsigned int> m_NumberOfHiddenNeurons;
/** Training parameters */ /** Training parameters */
unsigned int m_NumberOfIterations; unsigned int m_NumberOfIterations; // stop the training after a fixed number of iterations
double m_Epsilon; // Stops the training when the training error seems to converge
itk::Array<double> m_Regularization; // L2 Regularization parameter itk::Array<double> m_Regularization; // L2 Regularization parameter
itk::Array<double> m_Noise; // probability for an input to be set to 0 (denosing autoencoder) itk::Array<double> m_Noise; // probability for an input to be set to 0 (denosing autoencoder)
itk::Array<double> m_Rho; // Sparsity parameter itk::Array<double> m_Rho; // Sparsity parameter
......
...@@ -16,6 +16,8 @@ ...@@ -16,6 +16,8 @@
#include <shark/Models/ImpulseNoiseModel.h> //noise source to corrupt the inputs #include <shark/Models/ImpulseNoiseModel.h> //noise source to corrupt the inputs
#include <shark/Models/ConcatenatedModel.h>//to concatenate the noise with the model #include <shark/Models/ConcatenatedModel.h>//to concatenate the noise with the model
#include <shark/Algorithms/StoppingCriteria/MaxIterations.h> //A simple stopping criterion that stops after a fixed number of iterations
#include <shark/Algorithms/StoppingCriteria/TrainingProgress.h> //Stops when the algorithm seems to converge, Tracks the progress of the training error over a period of time
namespace otb namespace otb
{ {
...@@ -35,24 +37,52 @@ template <class TInputValue, class AutoencoderType> ...@@ -35,24 +37,52 @@ template <class TInputValue, class AutoencoderType>
void AutoencoderModel<TInputValue,AutoencoderType>::Train() void AutoencoderModel<TInputValue,AutoencoderType>::Train()
{ {
std::vector<shark::RealVector> features; std::vector<shark::RealVector> features;
std::cout << "converting the input ListSample to Shark vector" << std::endl;
Shark::ListSampleToSharkVector(this->GetInputListSample(), features); Shark::ListSampleToSharkVector(this->GetInputListSample(), features);
std::cout << "creating the data vector" << std::endl;
shark::Data<shark::RealVector> inputSamples = shark::createDataFromRange( features ); shark::Data<shark::RealVector> inputSamples = shark::createDataFromRange( features );
for (unsigned int i = 0 ; i < m_NumberOfHiddenNeurons.Size(); ++i)
{ if (m_Epsilon > 0){
if (m_Noise[i] != 0) // Shark doesn't allow to train a layer using a sparsity term AND a noisy input. (shark::SparseAutoencoderError takes an autoen shark::TrainingProgress<> criterion(5,m_Epsilon);
for (unsigned int i = 0 ; i < m_NumberOfHiddenNeurons.Size(); ++i)
{ {
TrainOneLayer( m_NumberOfHiddenNeurons[i],m_Noise[i],m_Regularization[i], inputSamples); if (m_Noise[i] != 0) // Shark doesn't allow to train a layer using a sparsity term AND a noisy input. (shark::SparseAutoencoderError takes an autoen
{
TrainOneLayer(criterion, m_NumberOfHiddenNeurons[i],m_Noise[i],m_Regularization[i], inputSamples);
}
else
{
TrainOneSparseLayer( criterion,m_NumberOfHiddenNeurons[i],m_Rho[i],m_Beta[i],m_Regularization[i], inputSamples);
}
criterion.reset();
} }
else
}
else {
shark::MaxIterations<> criterion(m_NumberOfIterations);
for (unsigned int i = 0 ; i < m_NumberOfHiddenNeurons.Size(); ++i)
{ {
TrainOneSparseLayer( m_NumberOfHiddenNeurons[i],m_Rho[i],m_Beta[i],m_Regularization[i], inputSamples); if (m_Noise[i] != 0) // Shark doesn't allow to train a layer using a sparsity term AND a noisy input. (shark::SparseAutoencoderError takes an autoen
{
TrainOneLayer(criterion, m_NumberOfHiddenNeurons[i],m_Noise[i],m_Regularization[i], inputSamples);
}
else
{
TrainOneSparseLayer(criterion, m_NumberOfHiddenNeurons[i],m_Rho[i],m_Beta[i],m_Regularization[i], inputSamples);
}
criterion.reset();
} }
} }
} }
template <class TInputValue, class AutoencoderType> template <class TInputValue, class AutoencoderType>
void AutoencoderModel<TInputValue,AutoencoderType>::TrainOneLayer(unsigned int nbneuron,double noise_strength,double regularization, shark::Data<shark::RealVector> &samples) template <class T>
void AutoencoderModel<TInputValue,AutoencoderType>::TrainOneLayer(shark::AbstractStoppingCriterion<T> & criterion, unsigned int nbneuron,double noise_strength,double regularization, shark::Data<shark::RealVector> &samples)
{ {
AutoencoderType net; AutoencoderType net;
...@@ -64,19 +94,22 @@ void AutoencoderModel<TInputValue,AutoencoderType>::TrainOneLayer(unsigned int n ...@@ -64,19 +94,22 @@ void AutoencoderModel<TInputValue,AutoencoderType>::TrainOneLayer(unsigned int n
shark::LabeledData<shark::RealVector,shark::RealVector> trainSet(samples,samples);//labels identical to inputs shark::LabeledData<shark::RealVector,shark::RealVector> trainSet(samples,samples);//labels identical to inputs
shark::SquaredLoss<shark::RealVector> loss; shark::SquaredLoss<shark::RealVector> loss;
shark::ErrorFunction error(trainSet, &model, &loss); shark::ErrorFunction error(trainSet, &model, &loss);
//shark::SparseAutoencoderError error(trainSet,&model, &loss, m_Rho, m_Beta);
//shark::SparseAutoencoderError error(trainSet,&net, &loss, 0.1, 0.1);
shark::TwoNormRegularizer regularizer(error.numberOfVariables()); shark::TwoNormRegularizer regularizer(error.numberOfVariables());
error.setRegularizer(regularization,&regularizer); error.setRegularizer(regularization,&regularizer);
shark::IRpropPlusFull optimizer; shark::IRpropPlusFull optimizer;
error.init(); error.init();
optimizer.init(error); optimizer.init(error);
std::cout<<"Optimizing model: "+net.name()<<std::endl;
for(std::size_t i = 0; i != m_NumberOfIterations; ++i){ std::cout<<"error before training : " << optimizer.solution().value<<std::endl;
unsigned int i=0;
do{
i++;
optimizer.step(error); optimizer.step(error);
std::cout<<i<<" "<<optimizer.solution().value<<std::endl; } while( !criterion.stop( optimizer.solution() ) );
} std::cout<<"error after " << i << "iterations : " << optimizer.solution().value<<std::endl;
net.setParameterVector(optimizer.solution().point); net.setParameterVector(optimizer.solution().point);
m_net.push_back(net); m_net.push_back(net);
samples = net.encode(samples); samples = net.encode(samples);
...@@ -84,7 +117,8 @@ void AutoencoderModel<TInputValue,AutoencoderType>::TrainOneLayer(unsigned int n ...@@ -84,7 +117,8 @@ void AutoencoderModel<TInputValue,AutoencoderType>::TrainOneLayer(unsigned int n
template <class TInputValue, class AutoencoderType> template <class TInputValue, class AutoencoderType>
void AutoencoderModel<TInputValue,AutoencoderType>::TrainOneSparseLayer(unsigned int nbneuron,double rho,double beta, double regularization, shark::Data<shark::RealVector> &samples) template <class T>
void AutoencoderModel<TInputValue,AutoencoderType>::TrainOneSparseLayer(shark::AbstractStoppingCriterion<T> & criterion, unsigned int nbneuron,double rho,double beta, double regularization, shark::Data<shark::RealVector> &samples)
{ {
AutoencoderType net; AutoencoderType net;
...@@ -102,11 +136,14 @@ void AutoencoderModel<TInputValue,AutoencoderType>::TrainOneSparseLayer(unsigned ...@@ -102,11 +136,14 @@ void AutoencoderModel<TInputValue,AutoencoderType>::TrainOneSparseLayer(unsigned
shark::IRpropPlusFull optimizer; shark::IRpropPlusFull optimizer;
error.init(); error.init();
optimizer.init(error); optimizer.init(error);
std::cout<<"Optimizing model: "+net.name()<<std::endl; std::cout<<"error before training : " << optimizer.solution().value<<std::endl;
for(std::size_t i = 0; i != m_NumberOfIterations; ++i){ unsigned int i=0;
do{
i++;
optimizer.step(error); optimizer.step(error);
std::cout<<i<<" "<<optimizer.solution().value<<std::endl; } while( !criterion.stop( optimizer.solution() ) );
} std::cout<<"error after " << i << "iterations : " << optimizer.solution().value<<std::endl;
net.setParameterVector(optimizer.solution().point); net.setParameterVector(optimizer.solution().point);
m_net.push_back(net); m_net.push_back(net);
samples = net.encode(samples); samples = net.encode(samples);
...@@ -165,7 +202,7 @@ void AutoencoderModel<TInputValue,AutoencoderType>::Load(const std::string & fil ...@@ -165,7 +202,7 @@ void AutoencoderModel<TInputValue,AutoencoderType>::Load(const std::string & fil
for (int i=0; i<m_net.size(); i++){ for (int i=0; i<m_net.size(); i++){
m_NumberOfHiddenNeurons[i] = m_net[i].numberOfHiddenNeurons(); m_NumberOfHiddenNeurons[i] = m_net[i].numberOfHiddenNeurons();
} }
this->m_Dimension = m_NumberOfHiddenNeurons[m_net.size()-1];
} }
......
...@@ -35,11 +35,11 @@ namespace otb ...@@ -35,11 +35,11 @@ namespace otb
{ {
template <class TInputValue, class TTargetValue> template <class TInputValue, class TTargetValue>
using AutoencoderModelFactory = AutoencoderModelFactoryBase<TInputValue, TTargetValue, shark::Autoencoder<shark::TanhNeuron, shark::LinearNeuron>> ; using AutoencoderModelFactory = AutoencoderModelFactoryBase<TInputValue, TTargetValue, shark::Autoencoder<shark::LogisticNeuron, shark::LogisticNeuron>> ;
template <class TInputValue, class TTargetValue> template <class TInputValue, class TTargetValue>
using TiedAutoencoderModelFactory = AutoencoderModelFactoryBase<TInputValue, TTargetValue, shark::TiedAutoencoder< shark::TanhNeuron, shark::LinearNeuron>> ; using TiedAutoencoderModelFactory = AutoencoderModelFactoryBase<TInputValue, TTargetValue, shark::TiedAutoencoder< shark::LogisticNeuron, shark::LogisticNeuron>> ;
template <class TInputValue, class TTargetValue> template <class TInputValue, class TTargetValue>
......
...@@ -36,10 +36,10 @@ public: ...@@ -36,10 +36,10 @@ public:
itkNewMacro(Self); itkNewMacro(Self);
itkTypeMacro(PCAModel, DimensionalityReductionModel); itkTypeMacro(PCAModel, DimensionalityReductionModel);
/*
unsigned int GetDimension() {return m_Dimension;}; unsigned int GetDimension() {return m_Dimension;};
itkSetMacro(Dimension,unsigned int); itkSetMacro(Dimension,unsigned int);
*/
itkSetMacro(Do_resize_flag,bool); itkSetMacro(Do_resize_flag,bool);
bool CanReadFile(const std::string & filename); bool CanReadFile(const std::string & filename);
...@@ -64,7 +64,7 @@ private: ...@@ -64,7 +64,7 @@ private:
shark::LinearModel<> m_encoder; shark::LinearModel<> m_encoder;
shark::LinearModel<> m_decoder; shark::LinearModel<> m_decoder;
shark::PCA m_pca; shark::PCA m_pca;
unsigned int m_Dimension; //unsigned int m_Dimension;
bool m_Do_resize_flag; bool m_Do_resize_flag;
}; };
} // end namespace otb } // end namespace otb
......
...@@ -20,6 +20,7 @@ template <class TInputValue> ...@@ -20,6 +20,7 @@ template <class TInputValue>
PCAModel<TInputValue>::PCAModel() PCAModel<TInputValue>::PCAModel()
{ {
this->m_IsDoPredictBatchMultiThreaded = true; this->m_IsDoPredictBatchMultiThreaded = true;
this->m_Dimension = 0;
} }
...@@ -39,9 +40,9 @@ void PCAModel<TInputValue>::Train() ...@@ -39,9 +40,9 @@ void PCAModel<TInputValue>::Train()
shark::Data<shark::RealVector> inputSamples = shark::createDataFromRange( features ); shark::Data<shark::RealVector> inputSamples = shark::createDataFromRange( features );
m_pca.setData(inputSamples); m_pca.setData(inputSamples);
m_pca.encoder(m_encoder, m_Dimension); m_pca.encoder(m_encoder, this->m_Dimension);
std::cout << m_encoder.matrix() << std::endl; std::cout << m_encoder.matrix() << std::endl;
m_pca.decoder(m_decoder, m_Dimension); m_pca.decoder(m_decoder, this->m_Dimension);
} }
...@@ -94,11 +95,19 @@ void PCAModel<TInputValue>::Load(const std::string & filename, const std::string ...@@ -94,11 +95,19 @@ void PCAModel<TInputValue>::Load(const std::string & filename, const std::string
boost::archive::polymorphic_text_iarchive ia(ifs); boost::archive::polymorphic_text_iarchive ia(ifs);
m_encoder.read(ia); m_encoder.read(ia);
ifs.close(); ifs.close();
m_Dimension = m_encoder.outputSize(); if (this->m_Dimension ==0)
{
this->m_Dimension = m_encoder.outputSize();
}
else
{
std::cout << "yo" << std::endl;
}
auto eigenvectors = m_encoder.matrix(); auto eigenvectors = m_encoder.matrix();
eigenvectors.resize(2,m_encoder.inputSize()); eigenvectors.resize(this->m_Dimension,m_encoder.inputSize());
m_encoder.setStructure(eigenvectors, m_encoder.offset() ); m_encoder.setStructure(eigenvectors, m_encoder.offset() );
std::cout << m_encoder.matrix() << std::endl; std::cout << m_encoder.matrix() << "end" << std::endl;
//this->m_Size = m_NumberOfHiddenNeurons; //this->m_Size = m_NumberOfHiddenNeurons;
} }
...@@ -120,9 +129,9 @@ PCAModel<TInputValue>::DoPredict(const InputSampleType & value, ConfidenceValueT ...@@ -120,9 +129,9 @@ PCAModel<TInputValue>::DoPredict(const InputSampleType & value, ConfidenceValueT
data = m_encoder(data); data = m_encoder(data);
TargetSampleType target; TargetSampleType target;
target.SetSize(m_Dimension); target.SetSize(this->m_Dimension);
for(unsigned int a = 0; a < m_Dimension; ++a){ for(unsigned int a = 0; a < this->m_Dimension; ++a){
target[a]=data.element(0)[a]; target[a]=data.element(0)[a];
} }
return target; return target;
...@@ -140,10 +149,10 @@ void PCAModel<TInputValue> ...@@ -140,10 +149,10 @@ void PCAModel<TInputValue>
TargetSampleType target; TargetSampleType target;
data = m_encoder(data); data = m_encoder(data);
unsigned int id = startIndex; unsigned int id = startIndex;
target.SetSize(m_Dimension); target.SetSize(this->m_Dimension);
for(const auto& p : data.elements()){ for(const auto& p : data.elements()){
for(unsigned int a = 0; a < m_Dimension; ++a){ for(unsigned int a = 0; a < this->m_Dimension; ++a){
target[a]=p[a]; target[a]=p[a];
//target[a]=1; //target[a]=1;
......
...@@ -85,7 +85,7 @@ public: ...@@ -85,7 +85,7 @@ public:
void Train() ITK_OVERRIDE; void Train() ITK_OVERRIDE;
//void Dimensionality_reduction() {}; // Dimensionality reduction is done by DoPredict //void Dimensionality_reduction() {}; // Dimensionality reduction is done by DoPredict
unsigned int GetDimension() { return MapType::ImageDimension;}; //unsigned int GetDimension() { return MapType::ImageDimension;};
protected: protected:
SOMModel(); SOMModel();
~SOMModel() ITK_OVERRIDE; ~SOMModel() ITK_OVERRIDE;
......
...@@ -22,6 +22,7 @@ namespace otb ...@@ -22,6 +22,7 @@ namespace otb
template <class TInputValue, unsigned int MapDimension> template <class TInputValue, unsigned int MapDimension>
SOMModel<TInputValue, MapDimension>::SOMModel() SOMModel<TInputValue, MapDimension>::SOMModel()
{ {
//m_Dimension = typename MapType::ImageDimension;
} }
...@@ -171,6 +172,7 @@ void SOMModel<TInputValue, MapDimension>::Load(const std::string & filename, con ...@@ -171,6 +172,7 @@ void SOMModel<TInputValue, MapDimension>::Load(const std::string & filename, con
} }
ifs.close(); ifs.close();
this->m_Dimension = MapType::ImageDimension;
} }
......
...@@ -101,10 +101,10 @@ public: ...@@ -101,10 +101,10 @@ public:
#ifdef OTB_USE_SHARK #ifdef OTB_USE_SHARK
typedef shark::Autoencoder< shark::TanhNeuron, shark::LinearNeuron> AutoencoderType; typedef shark::Autoencoder< shark::LogisticNeuron, shark::LogisticNeuron> AutoencoderType;
typedef otb::AutoencoderModel<InputValueType, AutoencoderType> AutoencoderModelType; typedef otb::AutoencoderModel<InputValueType, AutoencoderType> AutoencoderModelType;
typedef shark::TiedAutoencoder< shark::TanhNeuron, shark::LinearNeuron> TiedAutoencoderType; typedef shark::TiedAutoencoder< shark::LogisticNeuron, shark::LogisticNeuron> TiedAutoencoderType;
typedef otb::AutoencoderModel<InputValueType, TiedAutoencoderType> TiedAutoencoderModelType; typedef otb::AutoencoderModel<InputValueType, TiedAutoencoderType> TiedAutoencoderModelType;
typedef otb::PCAModel<InputValueType> PCAModelType; typedef otb::PCAModel<InputValueType> PCAModelType;
......
...@@ -44,6 +44,13 @@ cbLearningApplicationBaseDR<TInputValue,TOutputValue> ...@@ -44,6 +44,13 @@ cbLearningApplicationBaseDR<TInputValue,TOutputValue>
"model.autoencoder.nbiter", "model.autoencoder.nbiter",
"The maximum number of iterations used during training."); "The maximum number of iterations used during training.");
AddParameter(ParameterType_Float, "model.autoencoder.epsilon",
" ");
SetParameterFloat("model.autoencoder.epsilon",0, false);
SetParameterDescription(
"model.autoencoder.epsilon",
" ");
//Number Of Hidden Neurons //Number Of Hidden Neurons
AddParameter(ParameterType_StringList , "model.autoencoder.nbneuron", "Size"); AddParameter(ParameterType_StringList , "model.autoencoder.nbneuron", "Size");
...@@ -134,6 +141,7 @@ void cbLearningApplicationBaseDR<TInputValue,TOutputValue> ...@@ -134,6 +141,7 @@ void cbLearningApplicationBaseDR<TInputValue,TOutputValue>
std::cout << nb_neuron << std::endl; std::cout << nb_neuron << std::endl;
dimredTrainer->SetNumberOfHiddenNeurons(nb_neuron); dimredTrainer->SetNumberOfHiddenNeurons(nb_neuron);
dimredTrainer->SetNumberOfIterations(GetParameterInt("model.autoencoder.nbiter")); dimredTrainer->SetNumberOfIterations(GetParameterInt("model.autoencoder.nbiter"));
dimredTrainer->SetEpsilon(GetParameterFloat("model.autoencoder.epsilon"));
dimredTrainer->SetRegularization(regularization); dimredTrainer->SetRegularization(regularization);
dimredTrainer->SetNoise(noise); dimredTrainer->SetNoise(noise);
dimredTrainer->SetRho(rho); dimredTrainer->SetRho(rho);
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment