Difference between revisions of "Timothy Lillicrap"

From Chessprogramming wiki
Jump to: navigation, search
Line 18: Line 18:
 
* [[Shixiang Gu]], [[Timothy Lillicrap]], [[Ilya Sutskever]], [[Sergey Levine]] ('''2016'''). ''Continuous Deep Q-Learning with Model-based Acceleration''. [https://arxiv.org/abs/1603.00748 arXiv:1603.00748] <ref>[https://en.wikipedia.org/wiki/Q-learning Q-learning from Wikipedia]</ref>
 
* [[Shixiang Gu]], [[Timothy Lillicrap]], [[Ilya Sutskever]], [[Sergey Levine]] ('''2016'''). ''Continuous Deep Q-Learning with Model-based Acceleration''. [https://arxiv.org/abs/1603.00748 arXiv:1603.00748] <ref>[https://en.wikipedia.org/wiki/Q-learning Q-learning from Wikipedia]</ref>
 
* [[Shixiang Gu]], [[Ethan Holly]], [[Timothy Lillicrap]], [[Sergey Levine]] ('''2016'''). ''Deep Reinforcement Learning for Robotic Manipulation with Asynchronous Off-Policy Updates''. [https://arxiv.org/abs/1610.00633 arXiv:1610.00633]
 
* [[Shixiang Gu]], [[Ethan Holly]], [[Timothy Lillicrap]], [[Sergey Levine]] ('''2016'''). ''Deep Reinforcement Learning for Robotic Manipulation with Asynchronous Off-Policy Updates''. [https://arxiv.org/abs/1610.00633 arXiv:1610.00633]
* [[Shixiang Gu]], [[Timothy Lillicrap]], [[Zoubin Ghahramani]], [[Richard E. Turner]], [[Sergey Levine]] ('''2016'''). ''Q-Prop: Sample-Efficient Policy Gradient with An Off-Policy Critic''. [https://arxiv.org/abs/1611.02247 arXiv:1611.02247]
+
* [[Shixiang Gu]], [[Timothy Lillicrap]], [[Mathematician#ZGhahramani|Zoubin Ghahramani]], [[Richard E. Turner]], [[Sergey Levine]] ('''2016'''). ''Q-Prop: Sample-Efficient Policy Gradient with An Off-Policy Critic''. [https://arxiv.org/abs/1611.02247 arXiv:1611.02247]
 
'''2017'''
 
'''2017'''
 
* [[Yutian Chen]], [[Matthew W. Hoffman]], [[Sergio Gomez Colmenarejo]], [[Misha Denil]], [[Timothy Lillicrap]], [[Matthew Botvinick]], [[Nando de Freitas]] ('''2017'''). ''Learning to Learn without Gradient Descent by Gradient Descent''. [https://arxiv.org/abs/1611.03824v6 arXiv:1611.03824v6], [http://dblp.uni-trier.de/db/conf/icml/icml2017.html ICML 2017]
 
* [[Yutian Chen]], [[Matthew W. Hoffman]], [[Sergio Gomez Colmenarejo]], [[Misha Denil]], [[Timothy Lillicrap]], [[Matthew Botvinick]], [[Nando de Freitas]] ('''2017'''). ''Learning to Learn without Gradient Descent by Gradient Descent''. [https://arxiv.org/abs/1611.03824v6 arXiv:1611.03824v6], [http://dblp.uni-trier.de/db/conf/icml/icml2017.html ICML 2017]

Revision as of 20:58, 14 November 2019

Home * People * Timothy Lillicrap

Timothy Lillicrap [1]

Timothy P. (Tim) Lillicrap,
a Canadian neuroscientist an AI researcher, adjunct professor at University College London, and staff research scientist at Google, DeepMind, where he is involved in the AlphaGo and AlphaZero projects mastering the games of Go, chess and Shogi. He holds a B.Sc. in cognitive science and artificial intelligence from University of Toronto in 2005, and a Ph.D. in systems neuroscience from Queen's University in 2014 under Stephen H. Scott [2] [3]. His research focuses on machine learning and statistics for optimal control and decision making, as well as using these mathematical frameworks to understand how the brain learns. He has developed algorithms and approaches for exploiting deep neural networks in the context of reinforcement learning, and new recurrent memory architectures for one-shot learning [4].

Selected Publications

[5]

2014

2015 ...

2016

2017

2018

External Links

References

Up one level