Andrew G. Barto (born c. 1948) is an American computer scientist, currently Professor Emeritus of computer science at University of Massachusetts Amherst. Barto is best known for his foundational contributions to the field of modern computational reinforcement learning.[1]

Early life and education

Barto received his B.S. with distinction in mathematics from the University of Michigan in 1970, after having initially majored in naval architecture and engineering. After reading work by Michael Arbib and McCulloch and Pitts he became interested in using computers and mathematics to model the brain, and five years later was awarded a Ph.D. in computer science for a thesis on cellular automata.[2]

Career

In 1977, Barto joined the College of Information and Computer Sciences at the University of Massachusetts Amherst as a postdoctoral research associate, was promoted to associate professor in 1982, and full professor in 1991. He was department chair from 2007 to 2011 and a core faculty member of the Neuroscience and Behavior program.[3]

During this time at UMass, Barto co-directed the Autonomous Learning Laboratory (initially the Adaptive Network Laboratory), which generated several key ideas in reinforcement learning. Richard Sutton, with whom he co-authored the influential book Reinforcement Learning: An Introduction (MIT Press 1998; 2nd edition 2018), was his first PhD student. Barto graduated 27 PhD students, thirteen of which went on to become professors.[3]

Reinforcement learning

When Barto started at UMass, he joined a group of researchers trying to explore the behavior of neurons in the human brain as the basis for human intelligence, a concept that had been advanced by computer scientist A. Harry Klopf. Barto was joined by his doctoral student, Sutton in using mathematics toward furthering the concept and using it as the basis for artificial intelligence. This concept became known as reinforcement learning and went on to becoming a key part of artificial intelligence techniques.[4]

Barto and Sutton used Markov decision processes (MDP) as the mathematical foundation to explain how agents (algorithmic entities) made decisions when in a stochastic or random environment, receiving rewards at the end of every action. Traditional MDP theory assumed the agents knew all information about the MDPs in their attempt toward maximizing their cumulative rewards. Barto and Sutton's reinforcement learning techniques allowed for both the environment and the rewards to be unknown, and thus allowed for these category of algorithms to be applied to a wide array of problems.[5]

Barto built a lab in UMass Amherst toward developing the ideas on reinforcement learning while Sutton returned to Canada. Reinforcement learning as a topic continued to develop in academic circles until one of its first major real world applications saw Google's AlphaGo program built on this concept defeating the then prevailing human champion.[4] Barto and Sutton have widely been credited and accepted as pioneers of modern reinforcement learning, with the technique itself being foundational to the modern AI boom.[6]

Barto published over one hundred papers or chapters in journals, books, and conference and workshop proceedings. He is co-author with Richard Sutton of the book Reinforcement Learning: An Introduction, MIT Press 1998 (2nd edition 2018), and co-editor with Jennie Si, Warren Powell, and Don Wunch II of the Handbook of Learning and Approximate Dynamic Programming, Wiley-IEEE Press, 2004.[7]

Awards and honors

Barto is a Fellow of the American Association for the Advancement of Science, a Fellow and Senior Member of the IEEE,[8] and a member of the American Association for Artificial Intelligence and the Society for Neuroscience.[9]

Barto was awarded the UMass Neurosciences Lifetime Achievement Award, 2019, the IEEE Neural Network Society Pioneer Award in 2004,[10] and the IJCAI Award for Research Excellence, 2017. His citation for the latter read: "Professor Barto is recognized for his groundbreaking and impactful research in both the theory and application of reinforcement learning."[1]

In 2025, he received the Turing Award from the Association for Computing Machinery together with his former doctoral student Richard S. Sutton for their work on reinforcement learning; the citation of the award read: "For developing the conceptual and algorithmic foundations of reinforcement learning."[11][12][13]

References

  1. ^ a b "IJCAI 2017 Awards". 19 August 2017. Retrieved September 6, 2022.
  2. ^ "Virtual History Interview". International Neural Network Society. 7 January 2022. Retrieved September 6, 2022.
  3. ^ a b "Andrew G. Barto". University of Massachusetts Amherst. 17 February 2008. Retrieved October 18, 2020.
  4. ^ a b Metz, Cade (2025-03-05). "Turing Award Goes to 2 Pioneers of Artificial Intelligence". The New York Times. ISSN 0362-4331. Retrieved 2025-03-08.
  5. ^ "A.M. Turing Award". amturing.acm.org. Retrieved 2025-03-08.
  6. ^ "AI pioneers Andrew Barto and Richard Sutton win 2025 Turing Award for groundbreaking contributions to reinforcement learning | NSF - National Science Foundation". www.nsf.gov. 2025-03-05. Retrieved 2025-03-08.
  7. ^ UMass Amherst: Department of Computer Science Archived September 2, 2006, at the Wayback Machine
  8. ^ "Barto elected IEEE fellow". University of Massachusetts Amherst. November 22, 2005. Archived from the original on December 3, 2019. Retrieved December 3, 2019.
  9. ^ "CMU CS - AI Seminar". www.cs.cmu.edu. Retrieved 2025-03-07.
  10. ^ ""IEEE Computational Intelligence Society Past Recipients"". 6 September 2022. Retrieved September 6, 2022.
  11. ^ Metz, Cade (2025-03-05). "Turing Award Goes to 2 Pioneers of Artificial Intelligence". The New York Times. ISSN 0362-4331. Retrieved 2025-03-08.
  12. ^ "Turing Awardees - Directorate for Computer and Information Science and Engineering (CISE) | NSF - National Science Foundation". www.nsf.gov. 2025-03-05. Retrieved 2025-03-08.
  13. ^ Metz, Cade (2025-03-05). "Turing Award Goes to 2 Pioneers of Artificial Intelligence". The New York Times. ISSN 0362-4331. Retrieved 2025-03-08.
No tags for this post.