Talk 2: Recurrent neural nets and differentiable memory mechanism
Títol de la revista
ISSN de la revista
Títol del volum
Autors
Col·laborador
Editor
Tribunal avaluador
Realitzat a/amb
Càtedra / Departament / Institut
Tipus de document
Data publicació
Editor
Part de
Condicions d'accés
item.page.rightslicense
Datasets relacionats
Projecte CCD
Abstract
This past year, RNNs have seen a lot of attention as powerful models that are able to decode sequences from signals. The key component of such methods are the use of a recurrent neural network architecture that is trained end-to-end to optimize the probability of the output sequence given those signals. In this talk, I’ll define the architecture and review some recent successes in my group on machine translation, image understanding, and beyond. On the second part of the talk, I will introduce a new paradigm — differentiable memory — that has enabled learning programs (e.g., planar Traveling Salesman Problem) using training instances via a powerful extension of RNNs with memory. This effectively turns a machine learning model into a “differentiable computer”. I will conclude the talk giving a few examples (e.g., AlphaGo) on how these recent Machine Learning advances have been the main catalyst in Artificial Intelligence in the past years.


