Amos James Storkey
Born (1971-02-14) 14 February 1971 (age 52)
NationalityBritish
Alma materTrinity College, Cambridge
Known forStorkey Learning Rule
First Convolutional Network for Learning Go
Parent(s)Alan Storkey, Elaine Storkey
Scientific career
FieldsMachine learning, artificial intelligence, computer science
InstitutionsUniversity of Edinburgh

Amos James Storkey (born 1971) is Professor of Machine Learning and Artificial Intelligence at the School of Informatics, University of Edinburgh.

Storkey studied mathematics at Trinity College, Cambridge and obtained his doctorate from Imperial College, London. In 1997 during his PhD, he worked on the Hopfield Network a form of recurrent artificial neural network popularized by John Hopfield in 1982. Hopfield nets serve as content-addressable ("associative") memory systems with binary threshold nodes and Storkey developed what became known as the "Storkey Learning Rule".[1][2][3][4]

Subsequently, he has worked on approximate Bayesian methods, machine learning in astronomy,[5] graphical models, inference and sampling, and neural networks. Storkey joined the School of Informatics at the University of Edinburgh in 1999, was Microsoft Research Fellow from 2003 to 2004, appointed as reader in 2012, and to a personal chair in 2018. He is currently a Member of Institute for Adaptive and Neural Computation, Director of CDT in Data Science [2014-22] leading the Bayesian and Neural Systems Group.[6] In December 2014, Clark and Storkey together published an innovative paper "Teaching Deep Convolutional Neural Networks to Play Go". Convolutional neural network (CNN, or ConvNet) is a class of deep neural networks, most commonly applied to analyzing visual imagery. Their paper showed that a Convolutional Neural Network trained by supervised learning from a database of human professional games could outperform GNU Go and win some games against Monte Carlo tree search Fuego 1.1 in a fraction of the time it took Fuego to play.[7][8][9][10][circular reference]

Most cited work

References

  1. ^ Aggarwal, Charu C. "Neural Networks and Deep Learning" p240
  2. ^ Leveraging Different Learning Rules in Hopfield Nets for Multiclass Classification saiconference.com
  3. ^ Storkey, Amos. "Increasing the capacity of a Hopfield network without sacrificing functionality." Artificial Neural Networks – ICANN'97 (1997): 451-456.
  4. ^ Storkey, Amos. "Efficient Covariance Matrix Methods for Bayesian Gaussian Processes and Hopfield Neural Networks". PhD Thesis. University of London. (1999)
  5. ^ "One giant scrapheap for mankind". BBC News. 15 April 2004.
  6. ^ "Home". bayeswatch.com.
  7. ^ arXiv, Emerging Technology from the. "Why Neural Networks Look Set to Thrash the Best Human Go Players for the First Time". MIT Technology Review.
  8. ^ Chris J Maddison, 'Move Evaluation in Go' Madhttp://www0.cs.ucl.ac.uk/staff/d.silver/web/Applications_files/deepgo.pdf
  9. ^ Clark, Christopher; Storkey, Amos (2014). "Teaching Deep Convolutional Neural Networks to Play Go". arXiv:1412.3409 [cs.AI].
  10. ^ Convolutional neural network
  11. ^ a b c d e https://scholar.google.com/scholar?hl=en&as_sdt=0%2C33&q=Amos+storkey&btnG= Google Scholar Author page, Accessed June 14, 2021
This article needs additional or more specific categories. Please help out by adding categories to it so that it can be listed with similar articles. (July 2021)