Difference between revisions of "Timothy Lillicrap"

From Chessprogramming wiki
Jump to: navigation, search
 
Line 26: Line 26:
 
* [https://dblp.org/pers/hd/r/Rae:Jack_W= Jack W. Rae], [https://scholar.google.com/citations?user=W2DsnAkAAAAJ&hl=en Chris Dyer], [[Peter Dayan]], [[Timothy Lillicrap]] ('''2018'''). ''Fast Parametric Learning with Activation Memorization''. [https://arxiv.org/abs/1803.10049 arXiv:1803.10049]
 
* [https://dblp.org/pers/hd/r/Rae:Jack_W= Jack W. Rae], [https://scholar.google.com/citations?user=W2DsnAkAAAAJ&hl=en Chris Dyer], [[Peter Dayan]], [[Timothy Lillicrap]] ('''2018'''). ''Fast Parametric Learning with Activation Memorization''. [https://arxiv.org/abs/1803.10049 arXiv:1803.10049]
 
* [[David Silver]], [[Thomas Hubert]], [[Julian Schrittwieser]], [[Ioannis Antonoglou]], [[Matthew Lai]], [[Arthur Guez]], [[Marc Lanctot]], [[Laurent Sifre]], [[Dharshan Kumaran]], [[Thore Graepel]], [[Timothy Lillicrap]], [[Karen Simonyan]], [[Demis Hassabis]] ('''2018'''). ''[http://science.sciencemag.org/content/362/6419/1140 A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play]''. [https://en.wikipedia.org/wiki/Science_(journal) Science], Vol. 362, No. 6419 <ref>[https://deepmind.com/blog/alphazero-shedding-new-light-grand-games-chess-shogi-and-go/ AlphaZero: Shedding new light on the grand games of chess, shogi and Go] by [[David Silver]], [[Thomas Hubert]], [[Julian Schrittwieser]] and [[Demis Hassabis]], [[DeepMind]], December 03, 2018</ref>
 
* [[David Silver]], [[Thomas Hubert]], [[Julian Schrittwieser]], [[Ioannis Antonoglou]], [[Matthew Lai]], [[Arthur Guez]], [[Marc Lanctot]], [[Laurent Sifre]], [[Dharshan Kumaran]], [[Thore Graepel]], [[Timothy Lillicrap]], [[Karen Simonyan]], [[Demis Hassabis]] ('''2018'''). ''[http://science.sciencemag.org/content/362/6419/1140 A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play]''. [https://en.wikipedia.org/wiki/Science_(journal) Science], Vol. 362, No. 6419 <ref>[https://deepmind.com/blog/alphazero-shedding-new-light-grand-games-chess-shogi-and-go/ AlphaZero: Shedding new light on the grand games of chess, shogi and Go] by [[David Silver]], [[Thomas Hubert]], [[Julian Schrittwieser]] and [[Demis Hassabis]], [[DeepMind]], December 03, 2018</ref>
 +
* [[Vinícius Flores Zambaldi]], [[David Raposo]], [[Adam Santoro]], [[Victor Bapst]], [[Yujia Li]], [[Igor Babuschkin]], [[Karl Tuyls]], [[David P. Reichert]], [[Timothy Lillicrap]], [[Edward Lockhart]], [[Murray Shanahan]], [[Victoria Langston]], [[Razvan Pascanu]], [[Matthew Botvinick]], [[Oriol Vinyals]], [[Peter W. Battaglia]] ('''2018'''). ''Relational Deep Reinforcement Learning''. [https://arxiv.org/abs/1806.01830 arXiv:1806.01830]
 
'''2019'''
 
'''2019'''
 
* [[Julian Schrittwieser]], [[Ioannis Antonoglou]], [[Thomas Hubert]], [[Karen Simonyan]], [[Laurent Sifre]], [[Simon Schmitt]], [[Arthur Guez]], [[Edward Lockhart]], [[Demis Hassabis]], [[Thore Graepel]], [[Timothy Lillicrap]], [[David Silver]] ('''2019'''). ''Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model''. [https://arxiv.org/abs/1911.08265 arXiv:1911.08265]
 
* [[Julian Schrittwieser]], [[Ioannis Antonoglou]], [[Thomas Hubert]], [[Karen Simonyan]], [[Laurent Sifre]], [[Simon Schmitt]], [[Arthur Guez]], [[Edward Lockhart]], [[Demis Hassabis]], [[Thore Graepel]], [[Timothy Lillicrap]], [[David Silver]] ('''2019'''). ''Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model''. [https://arxiv.org/abs/1911.08265 arXiv:1911.08265]

Latest revision as of 08:53, 17 April 2021

Home * People * Timothy Lillicrap

Timothy Lillicrap [1]

Timothy P. (Tim) Lillicrap,
a Canadian neuroscientist an AI researcher, adjunct professor at University College London, and staff research scientist at Google, DeepMind, where he is involved in the AlphaGo and AlphaZero projects mastering the games of Go, chess and Shogi. He holds a B.Sc. in cognitive science and artificial intelligence from University of Toronto in 2005, and a Ph.D. in systems neuroscience from Queen's University in 2014 under Stephen H. Scott [2] [3]. His research focuses on machine learning and statistics for optimal control and decision making, as well as using these mathematical frameworks to understand how the brain learns. He has developed algorithms and approaches for exploiting deep neural networks in the context of reinforcement learning, and new recurrent memory architectures for one-shot learning [4].

Selected Publications

[5]

2014

2015 ...

2016

2017

2018

2019

2020 ...

External Links

References

Up one level