Cited By
View all- Singh NSingh S(2024)Reinforcement Learning in Bug TriagingAdvancing Software Engineering Through AI, Federated Learning, and Large Language Models10.4018/979-8-3693-3502-4.ch011(162-182)Online publication date: 21-Jun-2024
We introduce ballooning multi-armed bandits (BL-MAB), a novel extension to the classical stochastic MAB model. In the BL-MAB model, the set of available arms grows (or balloons) over time. The regret in a BL-MAB setting is computed with respect to the ...
A multi-armed bandit episode consists of n trials, each allowing selection of one of K arms, resulting in payoff from a distribution over [0,1] associated with that arm. We assume contextual side information is available at the start of the episode. ...
We consider a budgeted combinatorial multi-armed bandit setting where, in every round, the algorithm selects a super-arm consisting of one or more arms. The goal is to minimize the total expected regret after all rounds within a limited budget. Existing ...
Association for Computing Machinery
New York, NY, United States
Check if you have access through your login credentials or your institution to get full access on this article.
Sign inView or Download as a PDF file.
PDFView online with eReader.
eReaderView this article in HTML Format.
HTML Format