Commit 82a3e625 authored by Davis King's avatar Davis King

Changed the interface to the ridge regression trainer objects so that

they report the entire set of LOO prediction values rather than a
summary statistic like mean squared error.
parent 9b863dc9
...@@ -189,7 +189,8 @@ namespace dlib ...@@ -189,7 +189,8 @@ namespace dlib
const in_scalar_vector_type& y const in_scalar_vector_type& y
) const ) const
{ {
scalar_type temp, temp2; std::vector<scalar_type> temp;
scalar_type temp2;
return do_train(vector_to_matrix(x), vector_to_matrix(y), false, temp, temp2); return do_train(vector_to_matrix(x), vector_to_matrix(y), false, temp, temp2);
} }
...@@ -200,11 +201,11 @@ namespace dlib ...@@ -200,11 +201,11 @@ namespace dlib
const decision_function<kernel_type> train ( const decision_function<kernel_type> train (
const in_sample_vector_type& x, const in_sample_vector_type& x,
const in_scalar_vector_type& y, const in_scalar_vector_type& y,
scalar_type& looe std::vector<scalar_type>& loo_values
) const ) const
{ {
scalar_type temp; scalar_type temp;
return do_train(vector_to_matrix(x), vector_to_matrix(y), true, looe, temp); return do_train(vector_to_matrix(x), vector_to_matrix(y), true, loo_values, temp);
} }
template < template <
...@@ -214,11 +215,11 @@ namespace dlib ...@@ -214,11 +215,11 @@ namespace dlib
const decision_function<kernel_type> train ( const decision_function<kernel_type> train (
const in_sample_vector_type& x, const in_sample_vector_type& x,
const in_scalar_vector_type& y, const in_scalar_vector_type& y,
scalar_type& looe, std::vector<scalar_type>& loo_values,
scalar_type& lambda_used scalar_type& lambda_used
) const ) const
{ {
return do_train(vector_to_matrix(x), vector_to_matrix(y), true, looe, lambda_used); return do_train(vector_to_matrix(x), vector_to_matrix(y), true, loo_values, lambda_used);
} }
...@@ -231,8 +232,8 @@ namespace dlib ...@@ -231,8 +232,8 @@ namespace dlib
const decision_function<kernel_type> do_train ( const decision_function<kernel_type> do_train (
const in_sample_vector_type& x, const in_sample_vector_type& x,
const in_scalar_vector_type& y, const in_scalar_vector_type& y,
bool output_looe, const bool output_loo_values,
scalar_type& best_looe, std::vector<scalar_type>& loo_values,
scalar_type& the_lambda scalar_type& the_lambda
) const ) const
{ {
...@@ -311,8 +312,8 @@ namespace dlib ...@@ -311,8 +312,8 @@ namespace dlib
decision_function<linear_kernel<matrix<scalar_type,0,0,mem_manager_type> > > lin_df; decision_function<linear_kernel<matrix<scalar_type,0,0,mem_manager_type> > > lin_df;
if (output_looe) if (output_loo_values)
lin_df = trainer.train(proj_x,y, best_looe, the_lambda); lin_df = trainer.train(proj_x,y, loo_values, the_lambda);
else else
lin_df = trainer.train(proj_x,y); lin_df = trainer.train(proj_x,y);
......
...@@ -275,7 +275,7 @@ namespace dlib ...@@ -275,7 +275,7 @@ namespace dlib
const decision_function<kernel_type> train ( const decision_function<kernel_type> train (
const in_sample_vector_type& x, const in_sample_vector_type& x,
const in_scalar_vector_type& y, const in_scalar_vector_type& y,
scalar_type& looe std::vector<scalar_type>& loo_values
) const; ) const;
/*! /*!
requires requires
...@@ -283,12 +283,10 @@ namespace dlib ...@@ -283,12 +283,10 @@ namespace dlib
ensures ensures
- returns train(x,y) - returns train(x,y)
(i.e. executes train(x,y) and returns its result) (i.e. executes train(x,y) and returns its result)
- if (will_use_regression_loss_for_loo_cv()) - #loo_values.size() == y.size()
- #looe == the mean squared error as determined by leave-one-out - for all valid i:
cross-validation. - #loo_values[i] == leave-one-out prediction for the value of y(i) based
- else on all the training samples other than (x(i),y(i)).
- #looe == the fraction of samples misclassified as determined by
leave-one-out cross-validation.
!*/ !*/
template < template <
...@@ -298,7 +296,7 @@ namespace dlib ...@@ -298,7 +296,7 @@ namespace dlib
const decision_function<kernel_type> train ( const decision_function<kernel_type> train (
const in_sample_vector_type& x, const in_sample_vector_type& x,
const in_scalar_vector_type& y, const in_scalar_vector_type& y,
scalar_type& looe, std::vector<scalar_type>& loo_values,
scalar_type& lambda_used scalar_type& lambda_used
) const; ) const;
/*! /*!
...@@ -307,12 +305,10 @@ namespace dlib ...@@ -307,12 +305,10 @@ namespace dlib
ensures ensures
- returns train(x,y) - returns train(x,y)
(i.e. executes train(x,y) and returns its result) (i.e. executes train(x,y) and returns its result)
- if (will_use_regression_loss_for_loo_cv()) - #loo_values.size() == y.size()
- #looe == the mean squared error as determined by leave-one-out - for all valid i:
cross-validation. - #loo_values[i] == leave-one-out prediction for the value of y(i) based
- else on all the training samples other than (x(i),y(i)).
- #looe == the fraction of samples misclassified as determined by
leave-one-out cross-validation.
- #lambda_used == the value of lambda used to generate the - #lambda_used == the value of lambda used to generate the
decision_function. Note that this lambda value is always decision_function. Note that this lambda value is always
equal to get_lambda() if get_lambda() isn't 0. equal to get_lambda() if get_lambda() isn't 0.
......
...@@ -133,7 +133,8 @@ namespace dlib ...@@ -133,7 +133,8 @@ namespace dlib
const in_scalar_vector_type& y const in_scalar_vector_type& y
) const ) const
{ {
scalar_type temp, temp2; std::vector<scalar_type> temp;
scalar_type temp2;
return do_train(vector_to_matrix(x), vector_to_matrix(y), false, temp, temp2); return do_train(vector_to_matrix(x), vector_to_matrix(y), false, temp, temp2);
} }
...@@ -144,11 +145,11 @@ namespace dlib ...@@ -144,11 +145,11 @@ namespace dlib
const decision_function<kernel_type> train ( const decision_function<kernel_type> train (
const in_sample_vector_type& x, const in_sample_vector_type& x,
const in_scalar_vector_type& y, const in_scalar_vector_type& y,
scalar_type& looe std::vector<scalar_type>& loo_values
) const ) const
{ {
scalar_type temp; scalar_type temp;
return do_train(vector_to_matrix(x), vector_to_matrix(y), true, looe, temp); return do_train(vector_to_matrix(x), vector_to_matrix(y), true, loo_values, temp);
} }
template < template <
...@@ -158,11 +159,11 @@ namespace dlib ...@@ -158,11 +159,11 @@ namespace dlib
const decision_function<kernel_type> train ( const decision_function<kernel_type> train (
const in_sample_vector_type& x, const in_sample_vector_type& x,
const in_scalar_vector_type& y, const in_scalar_vector_type& y,
scalar_type& looe, std::vector<scalar_type>& loo_values,
scalar_type& lambda_used scalar_type& lambda_used
) const ) const
{ {
return do_train(vector_to_matrix(x), vector_to_matrix(y), true, looe, lambda_used); return do_train(vector_to_matrix(x), vector_to_matrix(y), true, loo_values, lambda_used);
} }
...@@ -175,8 +176,8 @@ namespace dlib ...@@ -175,8 +176,8 @@ namespace dlib
const decision_function<kernel_type> do_train ( const decision_function<kernel_type> do_train (
const in_sample_vector_type& x, const in_sample_vector_type& x,
const in_scalar_vector_type& y, const in_scalar_vector_type& y,
bool output_looe, const bool output_loo_values,
scalar_type& best_looe, std::vector<scalar_type>& loo_values,
scalar_type& the_lambda scalar_type& the_lambda
) const ) const
{ {
...@@ -279,7 +280,7 @@ namespace dlib ...@@ -279,7 +280,7 @@ namespace dlib
// We can save some work by pre-multiplying the x vectors by trans(V) // We can save some work by pre-multiplying the x vectors by trans(V)
// and saving the result so we don't have to recompute it over and over later. // and saving the result so we don't have to recompute it over and over later.
matrix<column_matrix_type,0,1,mem_manager_type > Vx; matrix<column_matrix_type,0,1,mem_manager_type > Vx;
if (lambda == 0 || output_looe) if (lambda == 0 || output_loo_values)
{ {
// Save the transpose of V into a temporary because the subsequent matrix // Save the transpose of V into a temporary because the subsequent matrix
// vector multiplies will be faster (because of better cache locality). // vector multiplies will be faster (because of better cache locality).
...@@ -299,9 +300,11 @@ namespace dlib ...@@ -299,9 +300,11 @@ namespace dlib
// If we need to automatically select a lambda then do so using the LOOE trick described // If we need to automatically select a lambda then do so using the LOOE trick described
// above. // above.
bool did_loov = false;
scalar_type best_looe = std::numeric_limits<scalar_type>::max();
if (lambda == 0) if (lambda == 0)
{ {
best_looe = std::numeric_limits<scalar_type>::max(); did_loov = true;
// Compute leave one out errors for a bunch of different lambdas and pick the best one. // Compute leave one out errors for a bunch of different lambdas and pick the best one.
for (long idx = 0; idx < lams.size(); ++idx) for (long idx = 0; idx < lams.size(); ++idx)
...@@ -342,16 +345,7 @@ namespace dlib ...@@ -342,16 +345,7 @@ namespace dlib
} }
} }
// mark that we saved the looe to best_looe already
output_looe = false;
best_looe /= x.size(); best_looe /= x.size();
if (verbose)
{
using namespace std;
cout << "Using lambda: " << the_lambda << endl;
cout << "LOO Error: " << best_looe << endl;
}
} }
...@@ -370,8 +364,10 @@ namespace dlib ...@@ -370,8 +364,10 @@ namespace dlib
// If we haven't done this already and we are supposed to then compute the LOO error rate for // If we haven't done this already and we are supposed to then compute the LOO error rate for
// the current lambda and store the result in best_looe. // the current lambda and store the result in best_looe.
if (output_looe) if (output_loo_values)
{ {
loo_values.resize(x.size());
did_loov = true;
best_looe = 0; best_looe = 0;
for (long i = 0; i < x.size(); ++i) for (long i = 0; i < x.size(); ++i)
{ {
...@@ -385,18 +381,26 @@ namespace dlib ...@@ -385,18 +381,26 @@ namespace dlib
loov = 0; loov = 0;
best_looe += loss(loov, y(i)); best_looe += loss(loov, y(i));
loo_values[i] = loov;
} }
best_looe /= x.size(); best_looe /= x.size();
if (verbose) }
{ else
using namespace std; {
cout << "Using lambda: " << the_lambda << endl; loo_values.clear();
cout << "LOO Error: " << best_looe << endl;
}
} }
if (verbose && did_loov)
{
using namespace std;
cout << "Using lambda: " << the_lambda << endl;
if (use_regression_loss)
cout << "LOO Mean Squared Error: " << best_looe << endl;
else
cout << "LOO Classification Accuracy: " << best_looe << endl;
}
// convert w into a proper decision function // convert w into a proper decision function
decision_function<kernel_type> df; decision_function<kernel_type> df;
......
...@@ -208,7 +208,7 @@ namespace dlib ...@@ -208,7 +208,7 @@ namespace dlib
const decision_function<kernel_type> train ( const decision_function<kernel_type> train (
const in_sample_vector_type& x, const in_sample_vector_type& x,
const in_scalar_vector_type& y, const in_scalar_vector_type& y,
scalar_type& looe std::vector<scalar_type>& loo_values
) const; ) const;
/*! /*!
requires requires
...@@ -216,12 +216,10 @@ namespace dlib ...@@ -216,12 +216,10 @@ namespace dlib
ensures ensures
- returns train(x,y) - returns train(x,y)
(i.e. executes train(x,y) and returns its result) (i.e. executes train(x,y) and returns its result)
- if (will_use_regression_loss_for_loo_cv()) - #loo_values.size() == y.size()
- #looe == the mean squared error as determined by leave-one-out - for all valid i:
cross-validation. - #loo_values[i] == leave-one-out prediction for the value of y(i) based
- else on all the training samples other than (x(i),y(i)).
- #looe == the fraction of samples misclassified as determined by
leave-one-out cross-validation.
!*/ !*/
template < template <
...@@ -231,7 +229,7 @@ namespace dlib ...@@ -231,7 +229,7 @@ namespace dlib
const decision_function<kernel_type> train ( const decision_function<kernel_type> train (
const in_sample_vector_type& x, const in_sample_vector_type& x,
const in_scalar_vector_type& y, const in_scalar_vector_type& y,
scalar_type& looe, std::vector<scalar_type>& loo_values,
scalar_type& lambda_used scalar_type& lambda_used
) const; ) const;
/*! /*!
...@@ -240,12 +238,10 @@ namespace dlib ...@@ -240,12 +238,10 @@ namespace dlib
ensures ensures
- returns train(x,y) - returns train(x,y)
(i.e. executes train(x,y) and returns its result) (i.e. executes train(x,y) and returns its result)
- if (will_use_regression_loss_for_loo_cv()) - #loo_values.size() == y.size()
- #looe == the mean squared error as determined by leave-one-out - for all valid i:
cross-validation. - #loo_values[i] == leave-one-out prediction for the value of y(i) based
- else on all the training samples other than (x(i),y(i)).
- #looe == the fraction of samples misclassified as determined by
leave-one-out cross-validation.
- #lambda_used == the value of lambda used to generate the - #lambda_used == the value of lambda used to generate the
decision_function. Note that this lambda value is always decision_function. Note that this lambda value is always
equal to get_lambda() if get_lambda() isn't 0. equal to get_lambda() if get_lambda() isn't 0.
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment