Difference between revisions of "Michael L. Littman"

From Chessprogramming wiki
Jump to: navigation, search
Line 58: Line 58:
 
==2010 ...==  
 
==2010 ...==  
 
* [[Thomas J. Walsh]], [[Sergiu Goschin]], [[Michael L. Littman]] ('''2010'''). ''Integrating sample-based planning and model-based reinforcement learning.'' [[AAAI]], [https://pdfs.semanticscholar.org/bdc9/bfb6ecc6fb5afb684df03d7220c46ebdbf4e.pdf pdf] » [[Monte-Carlo Tree Search|MCTS]], [[UCT]], [[Reinforcement Learning]]
 
* [[Thomas J. Walsh]], [[Sergiu Goschin]], [[Michael L. Littman]] ('''2010'''). ''Integrating sample-based planning and model-based reinforcement learning.'' [[AAAI]], [https://pdfs.semanticscholar.org/bdc9/bfb6ecc6fb5afb684df03d7220c46ebdbf4e.pdf pdf] » [[Monte-Carlo Tree Search|MCTS]], [[UCT]], [[Reinforcement Learning]]
 +
* [[Thomas J. Walsh]], [[István Szita]], [[Carlos Diuk]], [[Michael L. Littman]] ('''2012'''). ''Exploring compact reinforcement-learning representations with linear regression''. [https://arxiv.org/abs/1205.2606 arXiv:1205.2606]
 
* [[Michael L. Littman]] ('''2012'''). ''Technical Perspective: A New Way to Search Game Trees''. [[ACM#Communications|Communications of the ACM]], Vol. 55, No. 3
 
* [[Michael L. Littman]] ('''2012'''). ''Technical Perspective: A New Way to Search Game Trees''. [[ACM#Communications|Communications of the ACM]], Vol. 55, No. 3
 
* [[Ari Weinstein]], [[Michael L. Littman]] ('''2012'''). ''[https://www.aaai.org/ocs/index.php/ICAPS/ICAPS12/paper/view/4697 Bandit-Based Planning and Learning in Continuous-Action Markov Decision Processes]''. [http://dblp.uni-trier.de/db/conf/aips/icaps2012.html ICAPS 2012]
 
* [[Ari Weinstein]], [[Michael L. Littman]] ('''2012'''). ''[https://www.aaai.org/ocs/index.php/ICAPS/ICAPS12/paper/view/4697 Bandit-Based Planning and Learning in Continuous-Action Markov Decision Processes]''. [http://dblp.uni-trier.de/db/conf/aips/icaps2012.html ICAPS 2012]
* [[Mathematician#MKearns|Michael Kearns]], [[Michael L. Littman]], [[Mathematician#SSingh|Satinder Singh]] ('''2013'''). ''Graphical Models for Game Theory''. [http://dblp.uni-trier.de/db/journals/corr/corr1301.html#abs-1301-2281 CoRR, January 2013], [https://www.cis.upenn.edu/~mkearns/papers/graphgames.pdf pdf]
+
* [[Mathematician#MKearns|Michael Kearns]], [[Michael L. Littman]], [[Mathematician#SSingh|Satinder Singh]] ('''2013'''). ''Graphical Models for Game Theory''. [https://arxiv.org/abs/1301.2281 arXiv:1301.2281]
 
* [[Anthony R. Cassandra]], [[Michael L. Littman]], [[Nevin L. Zhang]] ('''2013'''). ''Incremental Pruning: A Simple, Fast, Exact Method for Partially Observable Markov Decision Processes''. [http://dblp.uni-trier.de/db/journals/corr/corr1302.html#abs-1302-1525 CoRR, February 2013], [http://arxiv.org/ftp/arxiv/papers/1302/1302.1525.pdf pdf]
 
* [[Anthony R. Cassandra]], [[Michael L. Littman]], [[Nevin L. Zhang]] ('''2013'''). ''Incremental Pruning: A Simple, Fast, Exact Method for Partially Observable Markov Decision Processes''. [http://dblp.uni-trier.de/db/journals/corr/corr1302.html#abs-1302-1525 CoRR, February 2013], [http://arxiv.org/ftp/arxiv/papers/1302/1302.1525.pdf pdf]
 
* [[Ari Weinstein]], [[Michael L. Littman]], [[Sergiu Goschin]] ('''2013'''). ''[http://proceedings.mlr.press/v24/weinstein12a.html Rollout-based Game-tree Search Outprunes Traditional Alpha-beta]''. [http://proceedings.mlr.press/ PMLR], Vol. 24 » [[Monte-Carlo Tree Search|MCTS]], [[UCT]]
 
* [[Ari Weinstein]], [[Michael L. Littman]], [[Sergiu Goschin]] ('''2013'''). ''[http://proceedings.mlr.press/v24/weinstein12a.html Rollout-based Game-tree Search Outprunes Traditional Alpha-beta]''. [http://proceedings.mlr.press/ PMLR], Vol. 24 » [[Monte-Carlo Tree Search|MCTS]], [[UCT]]

Revision as of 14:05, 23 October 2018

Home * People * Michael L. Littman

Michael L. Littman [1]

Michael Lederman Littman,
an American mathematician, computer scientist and professor of CS at Brown University, and before at Rutgers University and Duke University. He received his Ph.D. in 1996 at Brown University on Algorithms for Sequential Decision Making under advisor Leslie P. Kaelbling [2] . His research interests focus on stochastic games and reinforcement learning along with the related Markov Models for decision making, such as Markov Chain, Hidden Markov Model (HMM), Markov Decision Process (MDP), and Partially Observable Markov Decision Process (POMDP).

Crossword Solver

During his time at Duke University in the late 90s, Michael L. Littman worked on an automated crossword solver Proverb [3] [4], which won an Outstanding Paper Award in 1999 from AAAI [5] [6] and competed in the American Crossword Puzzle Tournament [7] [8].

See also

Markov Models

Michael Littman's explanatory grid on Markov Models [9] :

Markov
Models
Do we have control
over the state transitons?
NO YES
Are the states
completely observable?
YES Markov Chain MDP
NO HMM POMDP

Selected Publications

[10]

1990 ...

2000 ...

2010 ...

External Links

References

Up one Level