Sequential Decision Making: Contextual Bandits

A talk by James Sorrell
Senior Data Scientist, WooliesX

Register to watch this content

By submitting your email you agree to the Terms of Service and Privacy Statement
Watch this content now

Stages covered by this talk

About this talk

In this presentation, James walkthroughs how Contextual Bandits and the Sequential Decision Making framework are helping him better understand his customers as a Senior Data Scientist at WooliesX.

Sequential Decision Making is an intuitive framework that we can use to map many problems that we face into an agent that acts and an environment which it interacts with. Learning to efficiently navigate these problems presents exciting opportunities in an active and exciting area of research: Decision Science.

The origins of these frameworks date back to temporal-difference learning papers Minky (1954) who may have been the first to release that these principles could be important for artificial learning systems. Exciting projects such as AlphaGo, AlphaStar, OpenAI Five and tremendous advancements in Atari Benchmarks through algorithmic innovation have brought tremendous activity to this area of research. Contextual Bandits are a particularly useful version of the Multi-Arm Bandit problem which aims to relate contextual information to reward distributions for various actions.

Have you got yours yet?

Our All-Access Passes are a must if you want to get the most out of this event.

Check them out

Learn from amazing companies like these

Woolworths Group

Proudly supported by

Want to sponsor this event? contact us.