TY - JOUR AU - Weston, Jason AB - Curriculum Learning Yoshua Bengio1 Yoshua.Bengio@umontreal.ca J´rˆme Louradour1,2 e o jeromelouradour@gmail.com Ronan Collobert3 ronan@collobert.com Jason Weston3 jasonw@nec-labs.com (1) U. Montreal, P.O. Box 6128, Montreal, Canada (2) A2iA SA, 40bis Fabert, Paris, France (3) NEC Laboratories America, 4 Independence Way, Princeton, NJ, USA Abstract Humans and animals learn much better when the examples are not randomly presented but organized in a meaningful order which illustrates gradually more concepts, and gradually more complex ones. Here, we formalize such training strategies in the context of machine learning, and call them œcurriculum learning . In the context of recent research studying the di ƒculty of training in the presence of non-convex training criteria (for deep deterministic and stochastic neural networks), we explore curriculum learning in various set-ups. The experiments show that signi cant improvements in generalization can be achieved. We hypothesize that curriculum learning has both an e €ect on the speed of convergence of the training process to a minimum and, in the case of non-convex criteria, on the quality of the local minima obtained: curriculum learning can be seen as a particular form of continuation method (a general strategy for global optimization of non-convex functions). training and remarkably increase the speed TI - Curriculum learning DA - 2009-06-14 UR - https://www.deepdyve.com/lp/association-for-computing-machinery/curriculum-learning-1Bi5YD2QKX DP - DeepDyve ER -