Access the full text.
Sign up today, get DeepDyve free for 14 days.
Jonathan Baxter (1998)
Learning to Learn
S. Roweis, L. Saul (2000)
Nonlinear dimensionality reduction by locally linear embedding.Science, 290 5500
Haotong Zhang, A. Berg, M. Maire, Jitendra Malik (2006)
SVM-KNN: Discriminative Nearest Neighbor Classification for Visual Category Recognition2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06), 2
S. Lazebnik, C. Schmid, J. Ponce (2006)
Beyond Bags of Features: Spatial Pyramid Matching for Recognizing Natural Scene Categories2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06), 2
K. Tsuda, Taishin Kin, K. Asai (2002)
Marginalized kernels for biological sequencesBioinformatics, 18 Suppl 1
B. Olshausen, D. Field (1996)
Emergence of simple-cell receptive field properties by learning a sparse code for natural imagesNature, 381
R. Ando, Tong Zhang (2005)
A Framework for Learning Predictive Structures from Multiple Tasks and Unlabeled DataJ. Mach. Learn. Res., 6
Honglak Lee, Alexis Battle, Rajat Raina, A. Ng (2006)
Efficient sparse coding algorithms
Kamal Nigam, Tom Mitchell (1998)
Text Classi(cid:12)cation from Labeled and Unlabeled Documents using EM
Y. Shirai (1984)
Model-Based VisionJournal of the Robotics Society of Japan, 2
T. Jaakkola, D. Haussler (1998)
Exploiting Generative Models in Discriminative Classifiers
D. Blei, A. Ng, Michael Jordan (2009)
Latent Dirichlet Allocation
R. Tibshirani (1996)
Regression Shrinkage and Selection via the LassoJournal of the royal statistical society series b-methodological, 58
K. Nigam, A. McCallum, S. Thrun, Tom Mitchell (2000)
Text Classification from Labeled and Unlabeled Documents using EMMachine Learning, 39
R. Caruana (1997)
Multitask LearningMachine Learning, 28
Alex Holub, M. Welling, P. Perona (2005)
Combining generative models and Fisher kernels for object recognitionTenth IEEE International Conference on Computer Vision (ICCV'05) Volume 1, 1
A. Ng (2004)
Feature selection, L1 vs. L2 regularization, and rotational invarianceProceedings of the twenty-first international conference on Machine learning
S. Thrun (1995)
Is Learning The n-th Thing Any Easier Than Learning The First?
J. Tenenbaum, V. Silva, J. Langford (2000)
A global geometric framework for nonlinear dimensionality reduction.Science, 290 5500
Geoffrey Hinton, R. Salakhutdinov (2006)
Supporting Online Material for Reducing the Dimensionality of Data with Neural Networks
T. Landauer, S. Dumais (2008)
Latent semantic analysisScholarpedia, 3
Thomas Serre, Lior Wolf, T. Poggio (2005)
Object recognition with features inspired by visual cortex2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05), 2
P. Hoyer (2004)
Non-negative Matrix Factorization with Sparseness ConstraintsArXiv, cs.LG/0408058
Li Fei-Fei, R. Fergus, P. Perona (2004)
Learning Generative Visual Models from Few Training Examples: An Incremental Bayesian Approach Tested on 101 Object Categories2004 Conference on Computer Vision and Pattern Recognition Workshop
© Institute of Mathematical Statistics, 2004 LEAST ANGLE REGRESSION
Self-taught Learning: Transfer Learning from Unlabeled Data Rajat Raina Alexis Battle Honglak Lee Benjamin Packer Andrew Y. Ng Computer Science Department, Stanford University, CA 94305 USA [email protected] [email protected] [email protected] [email protected] [email protected] Abstract We present a new machine learning framework called self-taught learning for using unlabeled data in supervised classi cation tasks. We do not assume that the unlabeled data follows the same class labels or generative distribution as the labeled data. Thus, we would like to use a large number of unlabeled images (or audio samples, or text documents) randomly downloaded from the Internet to improve performance on a given image (or audio, or text) classi cation task. Such unlabeled data is signi cantly easier to obtain than in typical semi-supervised or transfer learning settings, making selftaught learning widely applicable to many practical learning problems. We describe an approach to self-taught learning that uses sparse coding to construct higher-level features using the unlabeled data. These features form a succinct input representation and signi cantly improve classi cation performance. When using an SVM for classi cation, we further show how a Fisher kernel can be learned for this representation. 1. Introduction Labeled data for machine learning is often
Read and print from thousands of top scholarly journals.
Already have an account? Log in
Bookmark this article. You can see your Bookmarks on your DeepDyve Library.
To save an article, log in first, or sign up for a DeepDyve account if you don’t already have one.
Copy and paste the desired citation format or use the link below to download a file formatted for EndNote
Access the full text.
Sign up today, get DeepDyve free for 14 days.
All DeepDyve websites use cookies to improve your online experience. They were placed on your computer when you launched this website. You can change your cookie settings through your browser.