summaryrefslogtreecommitdiffstats
path: root/ml/dlib/dlib/svm/svm_rank_trainer_abstract.h
diff options
context:
space:
mode:
Diffstat (limited to 'ml/dlib/dlib/svm/svm_rank_trainer_abstract.h')
-rw-r--r--ml/dlib/dlib/svm/svm_rank_trainer_abstract.h298
1 files changed, 298 insertions, 0 deletions
diff --git a/ml/dlib/dlib/svm/svm_rank_trainer_abstract.h b/ml/dlib/dlib/svm/svm_rank_trainer_abstract.h
new file mode 100644
index 000000000..4658d950f
--- /dev/null
+++ b/ml/dlib/dlib/svm/svm_rank_trainer_abstract.h
@@ -0,0 +1,298 @@
+// Copyright (C) 2012 Davis E. King (davis@dlib.net)
+// License: Boost Software License See LICENSE.txt for the full license.
+#undef DLIB_SVM_RANK_TrAINER_ABSTRACT_Hh_
+#ifdef DLIB_SVM_RANK_TrAINER_ABSTRACT_Hh_
+
+#include "ranking_tools_abstract.h"
+#include "sparse_vector_abstract.h"
+#include "function_abstract.h"
+#include "kernel_abstract.h"
+#include "../algs.h"
+
+namespace dlib
+{
+
+// ----------------------------------------------------------------------------------------
+
+ template <
+ typename K
+ >
+ class svm_rank_trainer
+ {
+ /*!
+ REQUIREMENTS ON K
+ Is either linear_kernel or sparse_linear_kernel.
+
+ WHAT THIS OBJECT REPRESENTS
+ This object represents a tool for training a ranking support vector machine
+ using linear kernels. In particular, this object is a tool for training
+ the Ranking SVM described in the paper:
+ Optimizing Search Engines using Clickthrough Data by Thorsten Joachims
+
+ Note that we normalize the C parameter by multiplying it by 1/(number of ranking pairs).
+ Therefore, to make an exact comparison between this object and Equation 12
+ in the paper you must multiply C by the appropriate normalizing quantity.
+
+ Finally, note that the implementation of this object is done using the oca
+ optimizer and count_ranking_inversions() method. This means that it runs
+ in O(n*log(n)) time, making it suitable for use with large datasets.
+ !*/
+
+ public:
+ typedef K kernel_type;
+ typedef typename kernel_type::scalar_type scalar_type;
+ typedef typename kernel_type::sample_type sample_type;
+ typedef typename kernel_type::mem_manager_type mem_manager_type;
+ typedef decision_function<kernel_type> trained_function_type;
+
+ svm_rank_trainer (
+ );
+ /*!
+ ensures
+ - This object is properly initialized and ready to be used to train a
+ ranking support vector machine.
+ - #get_oca() == oca() (i.e. an instance of oca with default parameters)
+ - #get_c() == 1
+ - #get_epsilon() == 0.001
+ - this object will not be verbose unless be_verbose() is called
+ - #get_max_iterations() == 10000
+ - #learns_nonnegative_weights() == false
+ - #forces_last_weight_to_1() == false
+ - #has_prior() == false
+ !*/
+
+ explicit svm_rank_trainer (
+ const scalar_type& C
+ );
+ /*!
+ requires
+ - C > 0
+ ensures
+ - This object is properly initialized and ready to be used to train a
+ ranking support vector machine.
+ - #get_oca() == oca() (i.e. an instance of oca with default parameters)
+ - #get_c() == C
+ - #get_epsilon() == 0.001
+ - this object will not be verbose unless be_verbose() is called
+ - #get_max_iterations() == 10000
+ - #learns_nonnegative_weights() == false
+ - #forces_last_weight_to_1() == false
+ - #has_prior() == false
+ !*/
+
+ void set_epsilon (
+ scalar_type eps
+ );
+ /*!
+ requires
+ - eps > 0
+ ensures
+ - #get_epsilon() == eps
+ !*/
+
+ const scalar_type get_epsilon (
+ );
+ /*!
+ ensures
+ - returns the error epsilon that determines when training should stop.
+ Smaller values may result in a more accurate solution but take longer to
+ train. You can think of this epsilon value as saying "solve the
+ optimization problem until the average ranking accuracy is within epsilon
+ of its optimal value". Here we mean "ranking accuracy" in the same sense
+ used by test_ranking_function() and cross_validate_ranking_trainer().
+ !*/
+
+ unsigned long get_max_iterations (
+ ) const;
+ /*!
+ ensures
+ - returns the maximum number of iterations the SVM optimizer is allowed to
+ run before it is required to stop and return a result.
+ !*/
+
+ void set_max_iterations (
+ unsigned long max_iter
+ );
+ /*!
+ ensures
+ - #get_max_iterations() == max_iter
+ !*/
+
+ void be_verbose (
+ );
+ /*!
+ ensures
+ - This object will print status messages to standard out so that a user can
+ observe the progress of the algorithm.
+ !*/
+
+ void be_quiet (
+ );
+ /*!
+ ensures
+ - this object will not print anything to standard out
+ !*/
+
+ bool forces_last_weight_to_1 (
+ ) const;
+ /*!
+ ensures
+ - returns true if this trainer has the constraint that the last weight in
+ the learned parameter vector must be 1. This is the weight corresponding
+ to the feature in the training vectors with the highest dimension.
+ !*/
+
+ void force_last_weight_to_1 (
+ bool should_last_weight_be_1
+ );
+ /*!
+ ensures
+ - #forces_last_weight_to_1() == should_last_weight_be_1
+ - if (should_last_weight_be_1 == true) then
+ - #has_prior() == false
+ !*/
+
+ void set_oca (
+ const oca& item
+ );
+ /*!
+ ensures
+ - #get_oca() == item
+ !*/
+
+ const oca get_oca (
+ ) const;
+ /*!
+ ensures
+ - returns a copy of the optimizer used to solve the SVM problem.
+ !*/
+
+ const kernel_type get_kernel (
+ ) const;
+ /*!
+ ensures
+ - returns a copy of the kernel function in use by this object. Since the
+ linear kernels don't have any parameters this function just returns
+ kernel_type()
+ !*/
+
+ bool learns_nonnegative_weights (
+ ) const;
+ /*!
+ ensures
+ - The output of training is a weight vector that defines the behavior of
+ the resulting decision function. That is, the decision function simply
+ takes the dot product between the learned weight vector and a test sample
+ and returns the result. Therefore, if learns_nonnegative_weights() == true
+ then the resulting learned weight vector will always have non-negative
+ entries.
+ !*/
+
+ void set_learns_nonnegative_weights (
+ bool value
+ );
+ /*!
+ ensures
+ - #learns_nonnegative_weights() == value
+ - if (value == true) then
+ - #has_prior() == false
+ !*/
+
+ void set_prior (
+ const trained_function_type& prior
+ );
+ /*!
+ requires
+ - prior == a function produced by a call to this class's train() function.
+ Therefore, it must be the case that:
+ - prior.basis_vectors.size() == 1
+ - prior.alpha(0) == 1
+ ensures
+ - Subsequent calls to train() will try to learn a function similar to the
+ given prior.
+ - #has_prior() == true
+ - #learns_nonnegative_weights() == false
+ - #forces_last_weight_to_1() == false
+ !*/
+
+ bool has_prior (
+ ) const
+ /*!
+ ensures
+ - returns true if a prior has been set and false otherwise. Having a prior
+ set means that you have called set_prior() and supplied a previously
+ trained function as a reference. In this case, any call to train() will
+ try to learn a function that matches the behavior of the prior as close
+ as possible but also fits the supplied training data. In more technical
+ detail, having a prior means we replace the ||w||^2 regularizer with one
+ of the form ||w-prior||^2 where w is the set of parameters for a learned
+ function.
+ !*/
+
+ void set_c (
+ scalar_type C
+ );
+ /*!
+ requires
+ - C > 0
+ ensures
+ - #get_c() == C
+ !*/
+
+ const scalar_type get_c (
+ ) const;
+ /*!
+ ensures
+ - returns the SVM regularization parameter. It is the parameter that
+ determines the trade off between trying to fit the training data exactly
+ or allowing more errors but hopefully improving the generalization of the
+ resulting classifier. Larger values encourage exact fitting while
+ smaller values of C may encourage better generalization.
+ !*/
+
+ const decision_function<kernel_type> train (
+ const std::vector<ranking_pair<sample_type> >& samples
+ ) const;
+ /*!
+ requires
+ - is_ranking_problem(samples) == true
+ - if (has_prior()) then
+ - The vectors in samples must have the same dimensionality as the
+ vectors used to train the prior given to set_prior().
+ ensures
+ - trains a ranking support vector classifier given the training samples.
+ - returns a decision function F with the following properties:
+ - F.alpha.size() == 1
+ - F.basis_vectors.size() == 1
+ - F.alpha(0) == 1
+ - Given two vectors, A and B, then A is predicted to come before B
+ in the learned ranking if and only if F(A) > F(B).
+ - Based on the contents of samples, F will attempt to give relevant
+ vectors higher scores than non-relevant vectors.
+ !*/
+
+ const decision_function<kernel_type> train (
+ const ranking_pair<sample_type>& sample
+ ) const;
+ /*!
+ requires
+ - is_ranking_problem(std::vector<ranking_pair<sample_type> >(1, sample)) == true
+ - if (has_prior()) then
+ - The vectors in samples must have the same dimensionality as the
+ vectors used to train the prior given to set_prior().
+ ensures
+ - This is just a convenience routine for calling the above train()
+ function. That is, it just copies sample into a std::vector object and
+ invokes the above train() method. This means that calling this function
+ is equivalent to invoking:
+ return train(std::vector<ranking_pair<sample_type> >(1, sample));
+ !*/
+
+ };
+
+// ----------------------------------------------------------------------------------------
+
+}
+
+#endif // DLIB_SVM_RANK_TrAINER_ABSTRACT_Hh_
+