Learning Long-Time Dependencies with RNNs

EPISODE 484
LISTEN
Banner Image: Thorben Konstantin Rusch - Podcast Interview

Join our list for notifications and early access to events

About this Episode

Today we conclude our 2021 ICLR coverage joined by Konstantin Rusch, a PhD Student at ETH Zurich. In our conversation with Konstantin, we explore his recent papers, titled coRNN and uniCORNN respectively, which focus on a novel architecture of recurrent neural networks for learning long-time dependencies. We explore the inspiration he drew from neuroscience when tackling this problem, how the performance results compared to networks like LSTMs and others that have been proven to work on this problem, and Konstantin's future research goals.
Connect with Thorben Konstantin
Read More
Connect with Konstantin

Thanks to our sponsor Qualcomm AI Research

Qualcomm AI Research is dedicated to advancing AI to make its core capabilities — perception, reasoning, and action — ubiquitous across devices. Their work makes it possible for billions of users around the world to have AI-enhanced experiences on devices powered by Qualcomm Technologies. To learn more about what Qualcomm Technologies is up to on the research front, visit twimlai.com/qualcomm.

Qualcomm Technologies Logo

More from TWIML

Leave a Reply

Your email address will not be published. Required fields are marked *