Commit abd3a9e2 authored by Davis King's avatar Davis King

Made spec more clear

--HG--
extra : convert_revision : svn%3Afdd8eb12-d10e-0410-9acb-85c331704f74/trunk%403633
parent 0c2cb034
...@@ -38,7 +38,7 @@ namespace dlib ...@@ -38,7 +38,7 @@ namespace dlib
In the above setting, all the training data consists of labeled samples. In the above setting, all the training data consists of labeled samples.
However, it would be nice to be able to benefit from unlabeled data. However, it would be nice to be able to benefit from unlabeled data.
The idea of manifold regularization is to extract useful information from The idea of manifold regularization is to extract useful information from
unlabeled data by defining which data samples are "close" to each other unlabeled data by first defining which data samples are "close" to each other
(perhaps by using their 3 nearest neighbors) and then adding a term to (perhaps by using their 3 nearest neighbors) and then adding a term to
the loss function that penalizes any decision rule which produces the loss function that penalizes any decision rule which produces
different outputs on data samples which we have designated as being close. different outputs on data samples which we have designated as being close.
...@@ -91,8 +91,8 @@ namespace dlib ...@@ -91,8 +91,8 @@ namespace dlib
ensures ensures
- #dimensionality() == samples[0].size() - #dimensionality() == samples[0].size()
- This function sets up the transformation matrix describe above. The manifold - This function sets up the transformation matrix describe above. The manifold
regularization is done assuming that the following samples are meant to regularization is done assuming that the samples are meant to be "close"
be "close" according to the graph defined by the given edges. I.e: according to the graph defined by the given edges. I.e:
- for all valid i: samples[edges[i].index1()] is close to samples[edges[i].index2()]. - for all valid i: samples[edges[i].index1()] is close to samples[edges[i].index2()].
How much we care about these two samples having similar outputs according How much we care about these two samples having similar outputs according
to the learned rule is given by weight_funct(edges[i]). Bigger weights mean to the learned rule is given by weight_funct(edges[i]). Bigger weights mean
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment