![]() | This article is rated Start-class on Wikipedia's
content assessment scale. It is of interest to the following WikiProjects: | ||||||||||||||||||||||||||||||
|
![]() | Text and/or other creative content from Training_set was copied or moved into Test_set with this edit. The former page's history now serves to provide attribution for that content in the latter page, and it must not be deleted as long as the latter page exists. |
There is absolutely no value added of having two articles Training set and Test set separately when neither can be discussed alone. The concept is Training and test sets with references to information science, statistics, data mining, biostatistics, etc. Currently the two articles are near duplicates (or could be based on the available information. Can we imagine some information for either which is not relevant for the other? Sda030 ( talk) 22:53, 27 February 2014 (UTC)
Totally agree with the suggestion - training set, testing set and validation set are all parts of one whole and should be presented in one topic. (MM-Professor of QM & MIS, WWU-USA)
A training set is also called a discovery set, right? (See for example <DOI: 10.1056/NEJMoa1406498>.) Perhaps a link should be created so that looking up "discovery set" redirects to here. Now, "discovery set" just gets a bunch of mostly-irrelevant search results. 73.53.61.168 ( talk) 11:17, 13 December 2015 (UTC)
I have seen the term "gold standard" been used at a few places in connection with articles about machine learning. On the page Gold standard (disambiguation), it says that in statistics and machine learning, gold standard is "a manually annotated training set or test set". What does it mean that the test set is manually annotated? And is "gold standard" a term that is important enough to be mentioned in this article perhaps? — Kri ( talk) 16:00, 19 January 2016 (UTC)
Lots of mentions in ML literature. Wqwt ( talk) 20:52, 22 March 2018 (UTC)
It's not clear in _whose_ practice this terms are flipped. In lots of posts by recognized practitioners (e.g. [1]) they're not flipped. — Preceding unsigned comment added by FabianMontescu ( talk • contribs) 19:35, 21 September 2018 (UTC)
Should this page make some reference to the way in which data is sampled - split into training/validation/test sets? This article is written in the context of Machine Learning, and often when training/validation/test sets are sampled from the main data source they are done so either randomly or in a stratified way. I think that this is worthy of mention in this article, even if not in detail.
aricooperdavis ( talk) 14:18, 13 November 2018 (UTC)
I can't seem to find here or in other places the earliest source for this method. it seems the holdout method was separately proposed by Highleyman in 1962, and cross validation was separately proposed by Stone in 1974, but the mixture of those two method resulting the train/validation/test is yet to be credited to one person. is this the truth? earliest source here is the Bishop book in 1995, but I don't think he is the one responsible for proposing this in literature
... no explanation is given, let alone in bold. Can someone please rectify? Thanks. 92.27.180.78 ( talk) 20:39, 3 May 2020 (UTC)