From A to B cover image

Are Multi-Armed Bandits the Next Thing to "Kill CRO" ?

From A to B

00:00

How to Motivate Yourself for a Race

The next competitor, stepping up on the stage to try and kill experimentation is multi arm bandits. This is something that I've been seeing a lot of on LinkedIn. It's effectively using machine learning to auto allocate traffic based on whatever variation you have. That's winning the most. If it's 50 50, like I don't think I've ever actually seen a multi embedded that runs at 50 50.

Transcript
Play full episode

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app