Threshold learning for optimal decision making

Research output: Chapter in Book/Report/Conference proceedingChapter in a book

78 Downloads (Pure)

Abstract

Decision making under uncertainty is commonly modelled as a process of competitive stochastic evidence accumulation to threshold (the drift-diffusion model). However, it is unknown how animals learn these decision thresholds. We examine threshold learning by constructing a reward function that averages over many trials to Wald's cost function that defines decision optimality. These rewards are highly stochastic and hence challenging to optimize, which we address in two ways: first, a simple two-factor reward-modulated learning rule derived from Williams' REINFORCE method for neural networks; and second, Bayesian optimization of the reward function with a Gaussian process. Bayesian optimization converges in fewer trials than REINFORCE but is slower computationally with greater variance. The REINFORCE method is also a better model of acquisition behaviour in animals and a similar learning rule has been proposed for modelling basal ganglia function.
Original languageEnglish
Title of host publicationAdvances in Neural Information Processing Systems 29 (NIPS 2016)
PublisherNeural Information Processing Systems (NIPS)
Publication statusPublished - Dec 2016

Fingerprint Dive into the research topics of 'Threshold learning for optimal decision making'. Together they form a unique fingerprint.

Cite this