I'm seeking papers on "effects of training examples' order to the classifiers' performance and accuracy". Do you know any literature on this? Jeffrey Elman has a paper related to Neural Networks(in origins of language paper),-actually this isn't a ML paper- but I couldn't find any other useful research on this.

Edit: The algorithm doesn't matter much for me. But I'm curious about how the performance and the accuracy is effected.

asked Dec 06 '10 at 07:01

cglr's gravatar image

cglr
1954811

edited Dec 06 '10 at 07:34


2 Answers:

The first paper (and in my opinion a quite interesting one) that comes to mind is the one by Bengio et al. (2009) on curriculum learning. The paper by LeCun et al. 1998 explains why for stochastic gradient descend it is good to make sure that the datapoints are presented in a random order.

It all depends on the kind of training algorithm you use but I assume you are thinking about stochastic gradient descend because of the reference to Elman's paper.

answered Dec 06 '10 at 07:53

Philemon%20Brakel's gravatar image

Philemon Brakel
2445103560

Actually I don't have any presupposition about the algorithm but while googling, it was surprising to me not to see any papers on this topic. Because data is nearly everything in this field.

(Dec 06 '10 at 08:05) cglr

I might be wrong but I have the impression that this is an issue that did not get the attention it deserves yet from the machine learning community.

(Dec 06 '10 at 08:09) Philemon Brakel

By the Elman paper I'm guessing you mean Learning and development in neural networks: the importance of starting small, and that you're interested in ideas like show the simplest examples first, then more complicated, and so on? There's a 2006 paper by Cornuejols, MACHINE LEARNING: The Necessity of Order (is order in order?) which is an overview of issues in this area, but not really a thorough survey.

It's an important notion, which went out of fashion as machine learning became more and more identified with batch mode optimization. Now that online learning, lifetime learning, multi-task learning, streaming algorithms, and the like are on the ascendance I suspect we'll see more attention to this.

As Philemon said, there's the more mundane issue of the importance of randomizing training data order in stochastic gradient descent. In fact I just posted a question to MA about that.

answered Dec 06 '10 at 12:32

Dave%20Lewis's gravatar image

Dave Lewis
890202846

edited Dec 06 '10 at 12:46

Thanks for the paper Dave :) I agree with you especially for the online algorithms, data order is a crucial.

(Dec 06 '10 at 12:42) cglr
Your answer
toggle preview

powered by OSQA

User submitted content is under Creative Commons: Attribution - Share Alike; Other things copyright (C) 2010, MetaOptimize LLC.