mirror of
https://github.com/davisking/dlib.git
synced 2024-11-01 10:14:53 +08:00
149 lines
6.7 KiB
C++
149 lines
6.7 KiB
C++
// The contents of this file are in the public domain. See LICENSE_FOR_EXAMPLE_PROGRAMS.txt
|
|
/*
|
|
|
|
This is an example that shows how you can perform model selection with the
|
|
dlib C++ Library.
|
|
|
|
It will create a simple dataset and show you how to use cross validation and
|
|
global optimization to determine good parameters for the purpose of training
|
|
an svm to classify the data.
|
|
|
|
The data used in this example will be 2 dimensional data and will come from a
|
|
distribution where points with a distance less than 10 from the origin are
|
|
labeled +1 and all other points are labeled as -1.
|
|
|
|
|
|
As an side, you should probably read the svm_ex.cpp and matrix_ex.cpp example
|
|
programs before you read this one.
|
|
*/
|
|
|
|
|
|
#include <iostream>
|
|
#include <dlib/svm.h>
|
|
#include <dlib/global_optimization.h>
|
|
|
|
using namespace std;
|
|
using namespace dlib;
|
|
|
|
|
|
int main() try
|
|
{
|
|
// The svm functions use column vectors to contain a lot of the data on which they
|
|
// operate. So the first thing we do here is declare a convenient typedef.
|
|
|
|
// This typedef declares a matrix with 2 rows and 1 column. It will be the
|
|
// object that contains each of our 2 dimensional samples.
|
|
typedef matrix<double, 2, 1> sample_type;
|
|
|
|
|
|
|
|
// Now we make objects to contain our samples and their respective labels.
|
|
std::vector<sample_type> samples;
|
|
std::vector<double> labels;
|
|
|
|
// Now let's put some data into our samples and labels objects. We do this
|
|
// by looping over a bunch of points and labeling them according to their
|
|
// distance from the origin.
|
|
for (double r = -20; r <= 20; r += 0.8)
|
|
{
|
|
for (double c = -20; c <= 20; c += 0.8)
|
|
{
|
|
sample_type samp;
|
|
samp(0) = r;
|
|
samp(1) = c;
|
|
samples.push_back(samp);
|
|
|
|
// if this point is less than 10 from the origin
|
|
if (sqrt(r*r + c*c) <= 10)
|
|
labels.push_back(+1);
|
|
else
|
|
labels.push_back(-1);
|
|
}
|
|
}
|
|
|
|
cout << "Generated " << samples.size() << " points" << endl;
|
|
|
|
|
|
// Here we normalize all the samples by subtracting their mean and dividing by their
|
|
// standard deviation. This is generally a good idea since it often heads off
|
|
// numerical stability problems and also prevents one large feature from smothering
|
|
// others. Doing this doesn't matter much in this example so I'm just doing this here
|
|
// so you can see an easy way to accomplish this with the library.
|
|
vector_normalizer<sample_type> normalizer;
|
|
// let the normalizer learn the mean and standard deviation of the samples
|
|
normalizer.train(samples);
|
|
// now normalize each sample
|
|
for (unsigned long i = 0; i < samples.size(); ++i)
|
|
samples[i] = normalizer(samples[i]);
|
|
|
|
|
|
// Now that we have some data we want to train on it. We are going to train a
|
|
// binary SVM with the RBF kernel to classify the data. However, there are
|
|
// three parameters to the training. These are the SVM C parameters for each
|
|
// class and the RBF kernel's gamma parameter. Our choice for these
|
|
// parameters will influence how good the resulting decision function is. To
|
|
// test how good a particular choice of these parameters is we can use the
|
|
// cross_validate_trainer() function to perform n-fold cross validation on our
|
|
// training data. However, there is a problem with the way we have sampled
|
|
// our distribution above. The problem is that there is a definite ordering
|
|
// to the samples. That is, the first half of the samples look like they are
|
|
// from a different distribution than the second half. This would screw up
|
|
// the cross validation process, but we can fix it by randomizing the order of
|
|
// the samples with the following function call.
|
|
randomize_samples(samples, labels);
|
|
|
|
|
|
// And now we get to the important bit. Here we define a function,
|
|
// cross_validation_score(), that will do the cross-validation we
|
|
// mentioned and return a number indicating how good a particular setting
|
|
// of gamma, c1, and c2 is.
|
|
auto cross_validation_score = [&](const double gamma, const double c1, const double c2)
|
|
{
|
|
// Make a RBF SVM trainer and tell it what the parameters are supposed to be.
|
|
typedef radial_basis_kernel<sample_type> kernel_type;
|
|
svm_c_trainer<kernel_type> trainer;
|
|
trainer.set_kernel(kernel_type(gamma));
|
|
trainer.set_c_class1(c1);
|
|
trainer.set_c_class2(c2);
|
|
|
|
// Finally, perform 10-fold cross validation and then print and return the results.
|
|
matrix<double> result = cross_validate_trainer(trainer, samples, labels, 10);
|
|
cout << "gamma: " << setw(11) << gamma << " c1: " << setw(11) << c1 << " c2: " << setw(11) << c2 << " cross validation accuracy: " << result;
|
|
|
|
// Now return a number indicating how good the parameters are. Bigger is
|
|
// better in this example. Here I'm returning the harmonic mean between the
|
|
// accuracies of each class. However, you could do something else. For
|
|
// example, you might care a lot more about correctly predicting the +1 class,
|
|
// so you could penalize results that didn't obtain a high accuracy on that
|
|
// class. You might do this by using something like a weighted version of the
|
|
// F1-score (see http://en.wikipedia.org/wiki/F1_score).
|
|
return 2*prod(result)/sum(result);
|
|
};
|
|
|
|
|
|
// And finally, we call this global optimizer that will search for the best parameters.
|
|
// It will call cross_validation_score() 50 times with different settings and return
|
|
// the best parameter setting it finds. find_max_global() uses a global optimization
|
|
// method based on a combination of non-parametric global function modeling and
|
|
// quadratic trust region modeling to efficiently find a global maximizer. It usually
|
|
// does a good job with a relatively small number of calls to cross_validation_score().
|
|
// In this example, you should observe that it finds settings that give perfect binary
|
|
// classification of the data.
|
|
auto result = find_max_global(cross_validation_score,
|
|
{1e-5, 1e-5, 1e-5}, // lower bound constraints on gamma, c1, and c2, respectively
|
|
{100, 1e6, 1e6}, // upper bound constraints on gamma, c1, and c2, respectively
|
|
max_function_calls(50));
|
|
|
|
double best_gamma = result.x(0);
|
|
double best_c1 = result.x(1);
|
|
double best_c2 = result.x(2);
|
|
|
|
cout << " best cross-validation score: " << result.y << endl;
|
|
cout << " best gamma: " << best_gamma << " best c1: " << best_c1 << " best c2: "<< best_c2 << endl;
|
|
}
|
|
catch (exception& e)
|
|
{
|
|
cout << e.what() << endl;
|
|
}
|
|
|