News
Now that the 2020 Tea Time Talks are on Youtube, you can always have time for tea with Amii and the RLAI Lab! Hosted by Amii’s Chief Scientific Advisory Dr. Richard S. Sutton, these 20-minute talks on technical topics are delivered by students, faculty and guests. The talks are a relaxed and informal way of hearing leaders in AI discuss future lines of research they may explore, with topics ranging from ideas starting to take root to fully-finished projects.
Week seven of the Tea Time Talks features:
Predictions, specifically those of general value functions (GVFs), have led to many lines of research and thought at the RLAI lab. While there have been many new algorithms for learning GVFs in recent years, there are still many questions around their use. In this talk, Matthew introduces the core concepts of hierarchical predictive coding (Rao, 1999), a scheme that uses predictions to inhibit feed-forward signals through corrective feedback. He also discusses an instantiation of the hierarchical predictive coding model using techniques from deep learning.
Environments come preconfigured with hyper-parameters, such as discretization rates and frame-skips, that determine an agent's window of temporal abstraction. In turn, this temporal window influences the magnitude of the action gap and greatly impacts learning. Alex discusses ongoing work that uses a recurrent neural network to flexibly learn action sequences within a temporal window.
In this talk, Shibhansh explores the interplay of generate-and-test and gradient-descent techniques for solving supervised learning problems. He starts by introducing a novel idealized setting in which the target function is stationary but much more complex than the learner, and in which the distribution of input is slowly varying. Then, he shows that if the target function is more complex than the approximator, tracking is better than any fixed set of weights. Finally, he explains that conventional backpropagation performs poorly in this setting, but its performance can be improved if we use random-search to replace low utility features.
In his talk, Dhawal explores the possibility of using adaptive stepsize techniques from the deep learning community for the use of temporal difference (TD) learning. Do the adaptive step size methods offer respite in TD learning divergence issues, mainly because of behavioural and target policy mismatch? Is this even something which merits looking into, or should completely separate stepsize techniques for TD learning be developed?
The Tea Time Talks have now concluded for the year, but stay tuned as we will be uploading the remaining talks in the weeks ahead. In the meantime, you can rewatch or catch up on previous talks on our Youtube playlist.
Nov 7th 2024
News
Amii partners with pipikwan pêhtâkwan and its startup company wâsikan kisewâtisiwin, to harness AI in efforts to challenge misinformation about Indigenous People and include Indigenous People in the development of AI. The project is supported by the PrairiesCan commitment to accelerate AI adoption among SMEs in the Prairie region.
Nov 7th 2024
News
Amii Fellow and Canada CIFAR AI Chair Russ Greiner and University of Alberta researcher and collaborator David Wishart were awarded the Brockhouse Canada Prize for Interdisciplinary Research in Science and Engineering from the National Sciences and Engineering Research Council of Canada (NSERC).
Nov 6th 2024
News
Amii founding member Jonathan Schaeffer has spent 40 years making huge impacts in game theory and AI. Now he’s retiring from academia and sharing some of the insights he’s gained over his impressive career.
Looking to build AI capacity? Need a speaker at your event?