File Download

There are no files associated with this item.

  • Find it @ UNIST can give you direct access to the published full text of this article. (UNISTARs only)

Views & Downloads

Detailed Information

Cited time in webofscience Cited time in scopus
Metadata Downloads

Full metadata record

DC Field Value Language
dc.contributor.advisor Kim, Gi-Soo -
dc.contributor.author Kwon, Youngin -
dc.date.accessioned 2025-04-04T13:48:01Z -
dc.date.available 2025-04-04T13:48:01Z -
dc.date.issued 2025-02 -
dc.description.abstract Interests have grown in discovering the causality in marketing and medical domains. For instance, online retail companies aim to expose the ad that has greatest effectiveness of exposure, while physicians practice precision medicine by applying a medical treatment only to patients who have genetic sequence that have the largest improvement in health. Both effectiveness and improvement imply causality, which can be measured by comparing outcomes under intervention versus baseline. In this work, we propose a bandit algorithm which achieves the aforementioned goals in an online manner from repeated sequences of choosing an arm (ad or genetic sequence) and observing outcomes under intervention or baseline. The reward, i.e., the causal effect, is not fully observed, which presents a new challenge compared to conventional bandit settings. We aim to do best arm identification and regret minimization simultaneously. We consider two types of regrets and propose two different algorithms, each utilizing a distinct intervention/baseline allocation policy to minimize one of the two regrets.We show that with high probability, each algorithm identifies the best arm after a number of iterations that closely matches known lower bound. We also show that the high-probability upper bound of the regret closely matches known lower bound. -
dc.description.degree Master -
dc.description Graduate School of Artificial Intelligence -
dc.identifier.uri https://scholarworks.unist.ac.kr/handle/201301/86359 -
dc.identifier.uri http://unist.dcollection.net/common/orgView/200000848271 -
dc.language ENG -
dc.publisher Ulsan National Institute of Science and Technology -
dc.subject Bandits -
dc.subject Causal Inference -
dc.subject Best Arm Identification -
dc.title Bandit Algorithm for Optimizing Causal Interventions -
dc.type Thesis -

qrcode

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.