A k-Armed Bandit Maximization Task is an online rewards-maximization task where the decision-making agent must make a finite sequence of choices against [math]k[/math] independent systems such that rewards are maximized.

## References

### 2015

• (Wikipedia, 2015) ⇒ http://en.wikipedia.org/wiki/multi-armed_bandit Retrieved:2015-6-20.
• In probability theory, the 'multi-armed bandit problem (sometimes called the K-[1] or N-armed bandit problem ) is a problem in which a gambler at a row of slot machines (sometimes known as "one-armed bandits") has to decide which machines to play, how many times to play each machine and in which order to play them.[2] When played, each machine provides a random reward from a distribution specific to that machine. The objective of the gambler is to maximize the sum of rewards earned through a sequence of lever pulls.[3][4] Robbins in 1952, realizing the importance of the problem, constructed convergent population selection strategies in "some aspects of the sequential design of experiments". A theorem, the Gittins index published first by John C. Gittins gives an optimal policy in the Markov setting for maximizing the expected discounted reward.

In practice, multi-armed bandits have been used to model the problem of managing research projects in a large organization, like a science foundation or a pharmaceutical company. Given a fixed budget, the problem is to allocate resources among the competing projects, whose properties are only partially known at the time of allocation, but which may become better understood as time passes.[3][4]

In early versions of the multi-armed bandit problem, the gambler has no initial knowledge about the machines. The crucial tradeoff the gambler faces at each trial is between "exploitation" of the machine that has the highest expected payoff and "exploration" to get more information about the expected payoffs of the other machines. The trade-off between exploration and exploitation is also faced in reinforcement learning.

1. Cite error: Invalid `<ref>` tag; no text was provided for refs named `doi10.1023/A:1013689704352`
2. Cite error: Invalid `<ref>` tag; no text was provided for refs named `weber`
3. Cite error: Invalid `<ref>` tag; no text was provided for refs named `Gittins89`
4. Cite error: Invalid `<ref>` tag; no text was provided for refs named `BF`

### 1989

• (Gittins, 1989) ⇒ J. C. Gittins. (1989). “Multi-Armed Bandit Allocation Indices." John Wiley & Sons, Ltd., ISBN 0-471-92059-2.

### 1985

• (Berry & Fristedt) ⇒ Donald A. Berry, and Bert Fristedt. (1985). “Bandit Problems: Sequential allocation of experiments." Chapman & Hall, ISBN 0-412-24810-7.