Asset Details
MbrlCatalogueTitleDetail
Do you wish to reserve the book?
Contextual Bandits in a Survey Experiment on Charitable Giving: Within-Experiment Outcomes versus Policy Learning
by
Athey, Susan
, Leung, Weiwen
, Williams, Joseph Jay
, Krishnamurthy, Sanath Kumar
, Byambadalai, Undral
, Hadad, Vitor
in
Algorithms
/ Data collection
/ Decay rate
/ Design optimization
/ Experimentation
/ Experiments
/ Heuristic
/ Lower bounds
/ Machine learning
/ Microbalances
/ Randomization
2022
Hey, we have placed the reservation for you!
By the way, why not check out events that you can attend while you pick your title.
You are currently in the queue to collect this book. You will be notified once it is your turn to collect the book.
Oops! Something went wrong.
Looks like we were not able to place the reservation. Kindly try again later.
Are you sure you want to remove the book from the shelf?
Contextual Bandits in a Survey Experiment on Charitable Giving: Within-Experiment Outcomes versus Policy Learning
by
Athey, Susan
, Leung, Weiwen
, Williams, Joseph Jay
, Krishnamurthy, Sanath Kumar
, Byambadalai, Undral
, Hadad, Vitor
in
Algorithms
/ Data collection
/ Decay rate
/ Design optimization
/ Experimentation
/ Experiments
/ Heuristic
/ Lower bounds
/ Machine learning
/ Microbalances
/ Randomization
2022
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
Do you wish to request the book?
Contextual Bandits in a Survey Experiment on Charitable Giving: Within-Experiment Outcomes versus Policy Learning
by
Athey, Susan
, Leung, Weiwen
, Williams, Joseph Jay
, Krishnamurthy, Sanath Kumar
, Byambadalai, Undral
, Hadad, Vitor
in
Algorithms
/ Data collection
/ Decay rate
/ Design optimization
/ Experimentation
/ Experiments
/ Heuristic
/ Lower bounds
/ Machine learning
/ Microbalances
/ Randomization
2022
Please be aware that the book you have requested cannot be checked out. If you would like to checkout this book, you can reserve another copy
We have requested the book for you!
Your request is successful and it will be processed during the Library working hours. Please check the status of your request in My Requests.
Oops! Something went wrong.
Looks like we were not able to place your request. Kindly try again later.
Contextual Bandits in a Survey Experiment on Charitable Giving: Within-Experiment Outcomes versus Policy Learning
Paper
Contextual Bandits in a Survey Experiment on Charitable Giving: Within-Experiment Outcomes versus Policy Learning
2022
Request Book From Autostore
and Choose the Collection Method
Overview
We design and implement an adaptive experiment (a ``contextual bandit'') to learn a targeted treatment assignment policy, where the goal is to use a participant's survey responses to determine which charity to expose them to in a donation solicitation. The design balances two competing objectives: optimizing the outcomes for the subjects in the experiment (``cumulative regret minimization'') and gathering data that will be most useful for policy learning, that is, for learning an assignment rule that will maximize welfare if used after the experiment (``simple regret minimization''). We evaluate alternative experimental designs by collecting pilot data and then conducting a simulation study. Next, we implement our selected algorithm. Finally, we perform a second simulation study anchored to the collected data that evaluates the benefits of the algorithm we chose. Our first result is that the value of a learned policy in this setting is higher when data is collected via a uniform randomization rather than collected adaptively using standard cumulative regret minimization or policy learning algorithms. We propose a simple heuristic for adaptive experimentation that improves upon uniform randomization from the perspective of policy learning at the expense of increasing cumulative regret relative to alternative bandit algorithms. The heuristic modifies an existing contextual bandit algorithm by (i) imposing a lower bound on assignment probabilities that decay slowly so that no arm is discarded too quickly, and (ii) after adaptively collecting data, restricting policy learning to select from arms where sufficient data has been gathered.
Publisher
Cornell University Library, arXiv.org
Subject
This website uses cookies to ensure you get the best experience on our website.