Known for his work on Neural Networks and Natural Language Processing. One of his most notable contributions is the development of the attention mechanism in neural networks, along with his colleagues Kyunghyun Cho and Yoshua Bengio.

The attention mechanism, introduced in their 2014 paper “Neural Machine Translation by Jointly Learning to Align and Translate,” has been a groundbreaking development in the field of deep learning. It significantly improved the performance of neural machine translation systems by allowing the model to focus on different parts of the input sequence when predicting each word of the output sequence, akin to how human attention works when we focus on specific aspects of a given task.