Skip to yearly menu bar Skip to main content


Poster

Contextual Decision-Making with Knapsacks beyond the Worst Case

Rui Ai · Zhaohua Chen · Xiaotie Deng · Yuqi Pan · Chang Wang · Mingwei Yang

[ ]
Thu 12 Dec 4:30 p.m. PST — 7:30 p.m. PST

Abstract: We study the framework of a dynamic decision-making scenario with resource constraints.In this framework, an agent, whose target is to maximize the total reward under the initial inventory, selects an action in each round upon observing a random request, leading to a reward and resource consumptions that are further associated with an unknown random external factor.While previous research has already established an $\widetilde{O}(\sqrt{T})$ worst-case regret for this problem, this work offers two results that go beyond the worst-case perspective: one for worst-case locations and another for logarithmic regret rates.We first show that an $\Omega(\sqrt{T})$ worst-case regret is unavoidable when the fluid LP has a degenerate optimal solution.On the algorithmic side, we merge the re-solving heuristic with distribution estimation skills and propose an algorithm that achieves an $\widetilde{O}(1)$ regret as long as the fluid LP has a unique and non-degenerate solution.Furthermore, we prove that our algorithm maintains a near-optimal $\widetilde{O}(\sqrt{T})$ regret even in the worst cases and extend these results to the setting where the request and external factor are continuous.Regarding information, our regret results are obtained under two feedback models, respectively, where the algorithm accesses the external factor at the end of each round and at the end of a round only when a non-null action is executed.

Live content is unavailable. Log in and register to view live content