InfoQ Homepage Presentations Creating Robust Interpretable NLP Systems with Attention
Creating Robust Interpretable NLP Systems with Attention
Summary
Alexander Wolf introduces Attention, an interpretable type of neural network layer that is loosely based on attention in human, explaining why and how it has been utilized to revolutionize NLP.
Bio
Alexander Wolf is a Data Scientist at Dataiku, working with clients around the world to organize their data infrastructures and deploy data-driven products into production. Prior to that, he worked on software and business development in the tech industry. He's passionate about the latest developments in Deep Learning/Tech and works at enriching Dataiku's NLP features.
About the conference
PAPIs is the conference made by and for ML practitioners. Anyone can submit talk proposals (blind-reviewed by our committee) or ask questions to take discussions further. Breaks between sessions are perfect to get to know fellow attendees, speakers, and top ML companies.