ABSTRACT
It is common in classification methods to first place data in a vector space and then learn decision boundaries. We propose reversing that process: for fixed decision boundaries, we "learn" the location of the data. This way we (i) do not need a metric (or even stronger structure) - pairwise dissimilarities suffice; and additionally (ii) produce low-dimensional embeddings that can be analyzed visually. We achieve this by combining an entropy-based embedding method with an entropy-based version of semi-supervised logistic regression. We present results for clustering and semi-supervised classification.
- Belkin, M., & Niyogi, P. (2004). Semi-supervised learning on Riemannian manifolds. Machine Learning, 56, 209--239. Google ScholarDigital Library
- Bishop, C. M. (1995). Neural networks for pattern recognition. Oxford: Clarendon Press. Google ScholarDigital Library
- Chapelle, O., & Zien, A. (2005). Semi-supervised classification by low density separation. Tenth International Workshop on Artificial Intelligence and Statistics.Google Scholar
- Grandvalet, Y., & Bengio, Y. (2005). Semi-supervised learning by entropy minimization. In L. K. Saul, Y. Weiss and L. Bottou (Eds.), Advances in neural information processing systems 17, vol. 17. Cambridge, MA: MIT Press.Google Scholar
- Hinton, G., & Roweis, S. (2003). Stochastic neighbor embedding. Advances in Neural Information Processing Systems 15. Cambridge, MA: MIT Press.Google Scholar
- Iwata, T., Saito, K., Ueda, N., Stromsten, S., Griffiths, T. L., & Tenenbaum, J. B. (2005). Parametric embedding for class visualization. In L. K. Saul. Y. Weiss and L. Bottou (Eds.), Advances in neural information processing systems 17. Cambridge, MA: MIT Press.Google Scholar
- Joachims, T. (1999). Transductive inference for text classification using support vector machines. International Conference on Machine Learning. Google ScholarDigital Library
- Kohonen, T. (1988). Self-organization and associative memory. Berlin, Germany: Springer. Google ScholarDigital Library
- Lin, J. (1991). Divergence measures based on the Shannon entropy. IEEE Transactions on Information Theory, 37, 145--151.Google ScholarDigital Library
- Nam, K., Je, H., & Choi, S. (2004). Fast stochastic neighbor embedding: A trust-region algorithm. Proc. Int'l Joint Conf. Neural Networks (IJCNN) (pp. 123--128). Budapest, Hungary.Google Scholar
- Roweis, S., & Saul, L. (2000). Nonlinear dimensionality reduction by locally linear embedding. Science, 290, 2323 2326.Google ScholarCross Ref
- Tenenbaum, J. B., de Silva, V., & Langford, J. C. (2000). A global geometric framework for nonlinear dimensionality reduction. Science, 290, 2319--2323.Google ScholarCross Ref
Recommendations
Max-margin embedding for multi-label learning
Multi-label learning refers to methods for learning a classification function that predicts a set of relevant labels for an instance. Label embedding seeks a transformation which maps labels into a latent space where regression is performed to predict a ...
Local margin based semi-supervised discriminant embedding for visual recognition
Most manifold learning algorithms adopt the k nearest neighbors function to construct the adjacency graph. However, severe bias may be introduced in this case if the samples are not uniformly distributed in the ambient space. In this paper a semi-...
Comments