Difference between revisions of "Edward Lockhart"

From Chessprogramming wiki
Jump to: navigation, search
Line 12: Line 12:
 
==2018 ...==
 
==2018 ...==
 
* [[Vinícius Flores Zambaldi]], [[David Raposo]], [[Adam Santoro]], [[Victor Bapst]], [[Yujia Li]], [[Igor Babuschkin]], [[Karl Tuyls]], [[David P. Reichert]], [[Timothy Lillicrap]], [[Edward Lockhart]], [[Murray Shanahan]], [[Victoria Langston]], [[Razvan Pascanu]], [[Matthew Botvinick]], [[Oriol Vinyals]], [[Peter W. Battaglia]] ('''2018'''). ''Relational Deep Reinforcement Learning''. [https://arxiv.org/abs/1806.01830 arXiv:1806.01830]
 
* [[Vinícius Flores Zambaldi]], [[David Raposo]], [[Adam Santoro]], [[Victor Bapst]], [[Yujia Li]], [[Igor Babuschkin]], [[Karl Tuyls]], [[David P. Reichert]], [[Timothy Lillicrap]], [[Edward Lockhart]], [[Murray Shanahan]], [[Victoria Langston]], [[Razvan Pascanu]], [[Matthew Botvinick]], [[Oriol Vinyals]], [[Peter W. Battaglia]] ('''2018'''). ''Relational Deep Reinforcement Learning''. [https://arxiv.org/abs/1806.01830 arXiv:1806.01830]
* [[Marc Lanctot]], [[Edward Lockhart]], [[Jean-Baptiste Lespiau]], [[Vinicius Zambaldi]], [[Satyaki Upadhyay]], [[Julien Pérolat]], [[Sriram Srinivasan]], [[Finbarr Timbers]], [[Karl Tuyls]], [[Shayegan Omidshafiei]], [[Daniel Hennes]], [[Dustin Morrill]], [[Paul Muller]], [[Timo Ewalds]], [[Ryan Faulkner]], [[János Kramár]], [[Bart De Vylder]], [[Brennan Saeta]], [[James Bradbury]], [[David Ding]], [[Sebastian Borgeaud]], [[Matthew Lai]], [[Julian Schrittwieser]], [[Thomas Anthony]], [[Edward Hughes]], [[Ivo Danihelka]], [[Jonah Ryan-Davis]] ('''2019'''). ''OpenSpiel: A Framework for Reinforcement Learning in Games''. [https://arxiv.org/abs/1908.09453 arXiv:1908.09453] <ref>[https://github.com/deepmind/open_spiel/blob/master/docs/contributing.md open_spiel/contributing.md at master · deepmind/open_spiel · GitHub]</ref>
+
* [[Marc Lanctot]], [[Edward Lockhart]], [[Jean-Baptiste Lespiau]], [[Vinícius Flores Zambaldi]], [[Satyaki Upadhyay]], [[Julien Pérolat]], [[Sriram Srinivasan]], [[Finbarr Timbers]], [[Karl Tuyls]], [[Shayegan Omidshafiei]], [[Daniel Hennes]], [[Dustin Morrill]], [[Paul Muller]], [[Timo Ewalds]], [[Ryan Faulkner]], [[János Kramár]], [[Bart De Vylder]], [[Brennan Saeta]], [[James Bradbury]], [[David Ding]], [[Sebastian Borgeaud]], [[Matthew Lai]], [[Julian Schrittwieser]], [[Thomas Anthony]], [[Edward Hughes]], [[Ivo Danihelka]], [[Jonah Ryan-Davis]] ('''2019'''). ''OpenSpiel: A Framework for Reinforcement Learning in Games''. [https://arxiv.org/abs/1908.09453 arXiv:1908.09453] <ref>[https://github.com/deepmind/open_spiel/blob/master/docs/contributing.md open_spiel/contributing.md at master · deepmind/open_spiel · GitHub]</ref>
 
* [[Julian Schrittwieser]], [[Ioannis Antonoglou]], [[Thomas Hubert]], [[Karen Simonyan]], [[Laurent Sifre]], [[Simon Schmitt]], [[Arthur Guez]], [[Edward Lockhart]], [[Demis Hassabis]], [[Thore Graepel]], [[Timothy Lillicrap]], [[David Silver]] ('''2019'''). ''Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model''. [https://arxiv.org/abs/1911.08265 arXiv:1911.08265]
 
* [[Julian Schrittwieser]], [[Ioannis Antonoglou]], [[Thomas Hubert]], [[Karen Simonyan]], [[Laurent Sifre]], [[Simon Schmitt]], [[Arthur Guez]], [[Edward Lockhart]], [[Demis Hassabis]], [[Thore Graepel]], [[Timothy Lillicrap]], [[David Silver]] ('''2019'''). ''Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model''. [https://arxiv.org/abs/1911.08265 arXiv:1911.08265]
 
* [[Edward Lockhart]], [[Marc Lanctot]], [[Julien Pérolat]], [[Jean-Baptiste Lespiau]], [[Dustin Morrill]], [[Finbarr Timbers]], [[Karl Tuyls]] ('''2019'''). ''Computing Approximate Equilibria in Sequential Adversarial Games by Exploitability Descent''. [https://arxiv.org/abs/1903.05614 arXiv:1903.05614]
 
* [[Edward Lockhart]], [[Marc Lanctot]], [[Julien Pérolat]], [[Jean-Baptiste Lespiau]], [[Dustin Morrill]], [[Finbarr Timbers]], [[Karl Tuyls]] ('''2019'''). ''Computing Approximate Equilibria in Sequential Adversarial Games by Exploitability Descent''. [https://arxiv.org/abs/1903.05614 arXiv:1903.05614]

Revision as of 09:27, 17 April 2021

Home * People * Edward Lockhart

Edward Lockhart [1]

Edward Lockhart,
a British computer scientist and reaearch engineer at DeepMind and head of its AI components. He holds a MA in mathematics from University of Cambridge in 1996 [2]. His current research focus is on sampling algorithms for equilibrium computation and decision-making. Edward Lockhart contributed to various reinforcement learning projects, such as OpenSpiel and MuZero [3].

Selected Publications

[4]

2018 ...

2020 ...

External Links

References

Up one level