Changes

Jump to: navigation, search

Supervised Learning

806 bytes added, 12:49, 16 November 2019
no edit summary
'''[[Main Page|Home]] * [[Learning]] * Supervised Learning'''
 
[[FILE:Supervised machine learning in a nutshell.svg|480px|border|right|thumb| Supervised Learning <ref>A data flow diagram shows the machine learning process in summary, by [https://en.wikipedia.org/wiki/User:EpochFail EpochFail], November 15, 2015, [https://en.wikipedia.org/wiki/Wikimedia_Commons Wikimedia Commons]</ref> ]]
'''Supervised Learning''',<br/>
is learning from examples provided by a knowledgable external [https://en.wikipedia.org/wiki/Supervisor supervisor].
In machine learning, supervised learning is a technique for deducing a function from [https://en.wikipedia.org/wiki/Training,_validation,_and_test_sets training data]. The training data consist of pairs of input objects and desired outputs . After parameter adjustment and learning, the performance of the resulting function should be measured on a test set that is separate from the training set <ref>[https://en.wikipedia.org/wiki/Supervised_learning Supervised learning from Wikipedia]</ref>. In computer games and chess, supervised learning techniques were used in [[Automated Tuning|automated tuning]] or to train [[Neural Networks|neural network]] game and chess programs. Input objects are [[Chess Position|chess positions]]. The desired output is either the supervisor's move choice in that position ([[Automated Tuning#MoveAdaption|move adaption]]), or a [[Score|score]] provided by an [[Oracle|oracle]] ([[Automated Tuning#ValueAdaption|value adaption]]).
=Move Adaption=
* [[Michael Buro]] ('''2002'''). ''Improving Mini-max Search by Supervised Learning.'' [https://en.wikipedia.org/wiki/Artificial_Intelligence_(journal) Artificial Intelligence], Vol. 134, No. 1, [http://www.cs.ualberta.ca/%7Emburo/ps/logaij.pdf pdf]
* [[Dave Gomboc]], [[Michael Buro]], [[Tony Marsland]] ('''2005'''). ''Tuning Evaluation Functions by Maximizing Concordance''. [https://en.wikipedia.org/wiki/Theoretical_Computer_Science_%28journal%29 Theoretical Computer Science], Vol. 349, No. 2, [http://www.cs.ualberta.ca/%7Emburo/ps/tcs-learn.pdf pdf]
* [[Amos Storkey]], [https://www.k.u-tokyo.ac.jp/pros-e/person/masashi_sugiyama/masashi_sugiyama.htm Masashi Sugiyama] ('''2006'''). ''[http://papers.neurips.cc/paper/3019-mixture-regression-for-covariate-shift Mixture Regression for Covariate Shift]''. [https://dblp.uni-trier.de/db/conf/nips/nips2006.html NIPS 2006]
==2010 ...==
* [[Tor Lattimore]], [[Marcus Hutter]] ('''2011'''). ''No Free Lunch versus Occam's Razor in Supervised Learning''. [https://en.wikipedia.org/wiki/Ray_Solomonoff Solomonoff] Memorial, [https://en.wikipedia.org/wiki/Lecture_Notes_in_Computer_Science Lecture Notes in Computer Science], [https://en.wikipedia.org/wiki/Springer-Verlag Springer], [https://arxiv.org/abs/1111.3846 arXiv:1111.3846] <ref>[https://en.wikipedia.org/wiki/No_free_lunch_in_search_and_optimization No free lunch in search and optimization - Wikipedia]</ref> <ref>[https://en.wikipedia.org/wiki/Occam%27s_razor Occam's razor from Wikipedia]</ref>

Navigation menu