#3: Bandits and Simulators for Recommenders with Olivier Jeunen
In episode three I am joined by Olivier Jeunen, who is a postdoctoral scientist at Amazon. Olivier obtained his PhD from University of Antwerp with his work "Offline Approaches to Recommendation with Online Success". His work concentrates on Bandits, Reinforcement Learning and Causal Inference for Recommender Systems.
In episode three I am joined by Olivier Jeunen, who is a postdoctoral scientist at Amazon. Olivier obtained his PhD from University of Antwerp with his work "Offline Approaches to Recommendation with Online Success". His work concentrates on Bandits, Reinforcement Learning and Causal Inference for Recommender Systems.
We talk about methods for evaluating online performance of recommender systems in an offline fashion and based on rich logging data. These methods stem from fields like bandit theory and reinforcement learning. They heavily rely on simulators whose benefits, requirements and limitations we discuss in greater detail. We further discuss the differences between organic and bandit feedback as well as what sets recommenders apart from advertising. We also talk about the right target for optimization and receive some advice to continue livelong learning as a researcher, be it in academia or industry.
Olivier has published multiple papers at RecSys, NeurIPS, WSDM, UMAP, and WWW. He also won the RecoGym challenge with his team from University of Antwerp. With research internships at Criteo, Facebook and Spotify Research he brings significant experience to the table.
Enjoy this enriching episode of RECSPERTS - Recommender Systems Experts.
Links from this Episode:
We talk about methods for evaluating online performance of recommender systems in an offline fashion and based on rich logging data. These methods stem from fields like bandit theory and reinforcement learning. They heavily rely on simulators whose benefits, requirements and limitations we discuss in greater detail. We further discuss the differences between organic and bandit feedback as well as what sets recommenders apart from advertising. We also talk about the right target for optimization and receive some advice to continue livelong learning as a researcher, be it in academia or industry.
Olivier has published multiple papers at RecSys, NeurIPS, WSDM, UMAP, and WWW. He also won the RecoGym challenge with his team from University of Antwerp. With research internships at Criteo, Facebook and Spotify Research he brings significant experience to the table.
Enjoy this enriching episode of RECSPERTS - Recommender Systems Experts.
Links from this Episode:
- Olivier's Website
- Olivier Jeunen on LinkedIn and Twitter
- Simulators:
- Blogpost: Lessons Learned from Winning the RecoGym Challenge
- RecSys 2020 REVEAL Workshop on Bandit and Reinforcement Learning from User Interactions
- RecSys 2021 Tutorial on Counterfactual Learning and Evaluation for Recommender Systems
- NeurIPS 2021 Workshop on Causal Inference and Machine Learning
Thesis and Papers:
- Dissertation: Offline Approaches to Recommendation with Online Success
- Chen et al. (2018): Top-K Off-Policy Correction for a REINFORCE Recommender System
- Jeunen et al. (2021): Disentangling Causal Effects from Sets of Interventions in the Presence of Unobserved Confounders
- Jeunen et al. (2021): Top-𝐾 Contextual Bandits with Equity of Exposure
General Links:
- Follow me on Twitter: https://twitter.com/LivesInAnalogia
- Send me your comments, questions and suggestions to marcel@recsperts.com
- Podcast Website: https://www.recsperts.com/