1.Introduction
In recent years, contextual multi-armed bandits (CMBA) have been widely used in complex decision-making processes, where agents make a sequence of decisions from a set of arms whose reward depends on contextual information. The time in the decision-making process goes like 1, 2, ..., T, and the agent needs to make a decision to pull one arm among a set of K arms at each time T. Each arm has an unknown reward distribution that depends on the context, and the purpose of the agent is to learn which arm to select in order to maximize cumulative rewards over time. The reward of the arm being pulled will be received and that of the other arms remains unknown. In terms of the distribution of the reward, it is sampled from an unknown distribution in a stochastic setting while chosen by an adversary in an adversarial setting. It is worth noting that in various contexts, the arm getting the highest reward may also be different.
Applications of CMAB span various domains, from online advertising to healthcare treatment optimization [1]. By adapting strategies based on observed contexts and rewards, CMAB algorithms offer an intelligent approach to maximize cumulative rewards over time. As an overview, Table 1 summarizes all the algorithms discussed in this paper. In Table 1, the number of distinct contexts is represented by C, the number of policies is represented by N, the number of arms is represented by K, and the dimension of contexts is represented by d.
Through a review, this paper aims to inspire researchers' interest in Contextual Multi-Armed Bandits, fostering further exploration and innovation to address the evolving decision challenges in the real world.
Table 1. A comparison between all the algorithms of contextual bandits discussed in this paper.
Algorithm |
Regret |
LinUCB |
O(d\( \sqrt[]{Tln((1+T)/δ)} \)) |
LinTS |
\( O({d^{2}}{δ^{-1}}\sqrt[]{{T^{1+δ}}}) \) |
D-LinUCB |
O((1/\( {∈^{2}} \)+1/∆)log(T)) |
D-RandLinUCB |
\( O({d^{\frac{2}{3}}}\bar{B_{T}^{\frac{1}{3}}}{T^{\frac{2}{3}}}) \) |
D-LinTS |
\( O\left( \frac{d^{\frac{2}{3}} \left( \log K \right)^{\frac{1}{3}} B_T^{\frac{1}{3}}}{T^{\frac{2}{3}}} \right) \) |
TS for High-Dimensionl |
\( \begin{array}{c} O({s^{*}}\sqrt[]{Tlogd}) \\ O({s^{*2}}[logd+logT]logT) \end{array} \) |
2.Contextual bandits feature representation
In the context of CMAB problems, an essential component is the process of feature extraction. CMAB problems involve an intelligent agent making decisions within various contextual scenarios to optimize cumulative rewards. The feature extraction step plays a pivotal role in mapping contextual information to appropriate action choices. In this regard, several established methods for feature extraction in CMAB scenarios are worth noting.
Firstly, the use of one-hot encoding is a prevalent technique [2]. It involves converting discrete contextual features, such as user demographics (e.g., gender or age groups), into binary vectors. Each unique feature value corresponds to a distinct dimension within the vector space. This method is particularly effective for scenarios with a finite set of discrete features.
Another approach is the application of embeddings [2]. Embeddings are advantageous when dealing with contextual features characterized by a multitude of potential values. They facilitate the transformation of such features into lower-dimensional continuous vector spaces, allowing for the capture of intricate relationships among features.
In addition, the normalization and standardization of continuous features are customary preprocessing steps [2]. These steps ensure that continuous features maintain consistent scales and ranges, which can be crucial for effective learning in CMAB settings.
Considering temporal dynamics, incorporating time-related features, such as timestamps or time intervals, can be essential. These features enable the model to account for temporal dependencies, which can be particularly relevant in scenarios where time plays a significant role.
Deep learning models, including Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs), offer a sophisticated approach to feature extraction. These models are adept at automatically deriving complex feature representations from contextual information.
Lastly, the combination of contextual features can yield more intricate representations. Techniques such as cross-feature interactions and polynomial feature engineering can be applied to enrich the feature set.
The choice of a specific feature extraction method within the CMAB framework depends on the problem's characteristics and the nature of the data at hand. It is imperative to experiment and fine-tune these methods to ascertain the most effective approach for maximizing rewards in CMAB scenarios.
3.Contextual bandits algorithms
3.1.Stochastic contextual bandits
Stochastic contextual bandit algorithms typically operate under the assumption that the reward associated with each arm conforms to an undisclosed probability distribution. Certain algorithms even extend this assumption to posit that the distribution adheres to a sub-Gaussian distribution with unspecified parameters. Under the linear realizability assumption, this section discusses stochastic contextual bandit algorithms.
3.1.1.LinUCB. In the LinUCB algorithm [3], we assume that on each arm, the feature vector\( {x_{t,a}}∈{R^{d}} \)is linear with the expected reward:
\( E[{r_{t,a}}∣{x_{t,a}}]=x_{t,a}^{⊤}{θ^{*}} \)(1)
Where\( {θ^{*}} \)is the true coefficient vector. Assume that at time t, the best arm is\( a_{t}^{*}=arg{max_{a}}x_{t,a}^{⊤}{θ^{*}} \), then the regret of LinUCB after t round is defined as
\( \begin{matrix} & {R_{T}}=E[\sum _{t=1}^{T}{r_{t,a_{t}^{*}}}-\sum _{t=1}^{T}{r_{t,{a_{t}}}}] \\ & =\sum _{t=1}^{T}x_{t,a_{t}^{*}}^{⊤}{θ^{*}}-\sum _{t=1}^{T}x_{t,{a_{t}}}^{⊤}{θ^{*}} \\ \end{matrix} \)(2)
Let\( {D_{t}} \)represent the feature vector of the arm that is pulled at each time. Let\( {c_{t}} \)represent the corresponding reward. If the sample\( ({x_{t,a}},{r_{t,{a_{t}}}}) \)is independent, then a closed-form estimator of\( {θ^{*}} \)is obtained through ridge regression
\( {\hat{θ}_{t}}=(D_{t}^{⊤}{D_{t}}+λ{I_{d}}{)^{-1}}D_{t}^{⊤}{c_{t}} \)(3)
For a prediction\( x_{t,a}^{⊤}{\hat{θ}_{t}} \), the upper confidence bound should be:
Assume that the rewards\( {r_{t,a}} \)are independent random variables with means\( E[{r_{t,a}}]=x_{t,a}^{⊤}{θ^{*}} \), let\( ϵ=\sqrt[]{\frac{1}{2}ln\frac{2TK}{δ}} \), and\( {A_{t}}=D_{t}^{⊤}{D_{t}}+{I_{d}} \), then with the probability\( 1-δ/T \), we got
\( ∣x_{t,a}^{⊤}{\hat{θ}_{t}}-x_{t,a}^{⊤}{θ^{*}}∣≤(ϵ+1)\sqrt[]{x_{t,a}^{⊤}A_{t}^{-1}{x_{t,a}}} \)(4)
However, in the LinUCB algorithm, samples from the previous round are used to estimate\( {θ^{*}} \)and a sample for the present round is chosen, thus the samples are not independent.
3.1.2.LinTS. As an effective approach for balancing exploration and exploitation, Thompson sampling produces good empirical results with respect to display ads and news recommendations [4]. Multi-armed bandit problems can be solved using Thompson sampling as well. For contextual bandits, Agrawal et al. and Li et al. provide a Thompson sampling algorithm with linear payoffs [5,6]. Assume that in the bandit problem, there are K arms, and at time t, each arm "a" is linked to a d-dimensional feature vector\( {x_{t,a}} \). The context selection is not assumed to follow a specific distribution and can be determined by an adversary. A d-dimensional parameter\( μ∈{R^{d}} \)is used to define a linear predictor and predict the mean reward of arm a by\( μ\cdot {x_{t,a}} \). We assume an unknown underlying parameter\( {μ^{*}}∈{R^{d}} \), therefore, at time t, the expected reward for the arm a is given by\( {\bar{r}_{t,a}}={μ^{*}}\cdot {x_{t,a}} \). The actual reward\( {r_{t,a}} \)is obtained from choosing the arm a at time t with an unknown distribution with mean\( {\bar{r}_{t,a}} \). The Thompson sampling algorithm selects an arm\( {a_{t}} \)at each time t\( ∈ \){1,...,T} and receives a reward\( {r_{t}} \). Let\( {a^{*}} \)be the optimal arm at time t:
\( a_{t}^{*}=argmax{\bar{r}_{t,a}} \)(5)
Let\( {Δ_{t,a}} \)represent the difference of the expected reward between the optimal arm and arm a:
\( {Δ_{t,a}}={\bar{r}_{t,{a^{*}}}}-{\bar{r}_{t,a}} \)(6)
Mathematically, the regret can be expressed as:
\( {R_{T}}=\sum _{t=1}^{T} {Δ_{t,{a_{t}}}} \)(7)
In the mentioned paper, the assumption
\( {δ_{t,a}}={r_{t,a}}-{\bar{r}_{t,a}} \)is conditionally R-sub-Gaussian, which implies that for a constant R≥0,\( {r_{t,a}}∈[{\bar{r}_{t,a}}-R,{\bar{r}_{r,t}}+R] \).
There are various likelihood distributions that satisfy this condition, but for the sake of simplicity, the paper assumes a Gaussian likelihood and Gaussian prior.
Thus, the likelihood of reward\( {\bar{r}_{t,a}} \), given the context\( {x_{t,a}} \), is modeled using the probability density function\( N(x_{t,a}^{⊤}{μ^{*}},{v^{2}}) \)of a Gaussian distribution. Mathematically,\( v \)is defined as\( v=R\sqrt[]{\frac{24}{ϵ}dln(\frac{t}{δ})} \), where the algorithm parameter is\( ϵ∈(0,1) \), and δ is a parameter that controls the high probability regret bound.
Similar to the closed-form of linear regression, we define
\( {B_{t}}={I_{d}}+\sum _{τ=1}^{t-1} {x_{τ,a}}x_{τ,a}^{⊤} \)(8)
\( {\hat{μ}_{t}}=B_{t}^{-1}(\sum _{τ=1}^{t-1} {x_{τ,a}}{r_{τ,a}}) \)(9)
With probability 1−δ, the regret is bounded by:
\( {R_{T}}=O(\frac{{d^{2}}}{ϵ}\sqrt[]{{T^{1+ϵ}}}(ln(Td)ln\frac{1}{δ})) \)(10)
3.1.3.D-LinUCB. LinUCB is an optimistic algorithm for non-stationary environments in a stochastic linear bandit model [7]. To handle non-stationarity and forget past observations smoothly, it utilizes discounted linear regression and exponential weights\( {w_{t}}={γ^{-t}} \), where 0< γ<1 is the discount factor. The algorithm incorporates regularization and computes the upper confidence bound(UCB) for each action based on estimated regression parameters and uncertainty. The action with the highest UCB is selected to play, and the algorithm updates the parameter estimation and uncertainty based on observed rewards and chosen actions.
At each step, the algorithm receives a set of available actions\( {A_{t}} \)and computes an upper confidence bound(UCB)\( UCB(a)={a^{⊤}}\hat{θ}+{β_{t-1}}\sqrt[]{{a^{⊤}}{V^{-1}}\widetilde{V}{V^{-1}}a} \)for each action based on the current estimates of the unknown regression parameter\( {\hat{θ}_{t}}=arg{min_{θ∈{R^{d}}}}(\sum _{s=1}^{t}{w_{s}}({X_{s}}-⟨{A_{s}},θ⟩{)^{2}}+{λ_{t}}∥θ∥_{2}^{2}) \)and the uncertainty in the estimates. The algorithm selects the action with the highest UCB and plays it, receiving a reward\( {X_{t}}=⟨{A_{t}},θ_{t}^{⋆}⟩+{η_{t}}, \)The algorithm then updates its estimates of\( θ \)and the uncertainty in the estimates based on the played action and reward.
In the D-LinUCB algorithm, we introduce a confidence ellipsoid\( {C_{t}} \)which is defined as\( \lbrace θ:{∥θ-{\hat{θ}_{t-1}}∥_{{V_{t-1}}\widetilde{V}_{t-1}^{-1}{V_{t-1}}}}≤{β_{t-1}}\rbrace \), and let
\( {β_{t}}=\sqrt[]{λ}S+σ\sqrt[]{2log(1/δ)+dlog(1+\frac{{L^{2}}(1-{γ^{2t}})}{λd(1-{γ^{2}})})} \)(11)
Based on the remark above regarding to scale-invariance, we can easily conclude that at time t, our D-LinUCB algorithm chooses the action\( {A_{t}} \)that maximizes\( ⟨a,θ⟩ \)for\( a∈{A_{t}} \)and\( θ∈{C_{t}} \).
Assuming that\( \sum _{s=1}^{T-1}∥θ_{s}^{⋆}-θ_{s+1}^{⋆}{∥_{2}}≤{B_{T}} \), the regret of the D-LinUCB algorithm is bounded for all\( γ∈(0,1) \)and D\( ≥1 \)then with the probability\( 1-δ/T \), we got
\( {R_{T}}≤2LD{B_{T}}+\frac{4{L^{3}}S}{λ}\frac{{γ^{D}}}{1-γ}T+2\sqrt[]{2}{β_{T}}\sqrt[]{dT}\sqrt[]{Tlog(1/γ)+log(1+\frac{{L^{2}}}{dλ(1-γ)})} \)(12)
3.1.4.D-RandLinUCB. D-LinUCB algorithm follows the optimism in the face of the uncertainty principle and picks actions through maximizing the UCB bound of expected reward based on confidence level a and\( \hat{θ}_{t}^{wls} \)[8]. However, in a non-stationary linear bandit environment, the D-RandLinUCB algorithm replaces the confidence level a with a random variable\( {Z_{t}}∼D \).
D-LinUCB:\( {X_{t}}=arg\underset{x∈{X_{t}}}{max}⟨x,\hat{θ}_{t}^{wls}⟩+a∥x{∥_{V_{t}^{-1}}} \)
D-RandLinUCB:\( {X_{t}}=arg\underset{x∈{X_{t}}}{max}⟨x,\hat{θ}_{t}^{wls}⟩+{Z_{t}}∥x{∥_{V_{t}^{-1}}}. \)
In the non-stationary linear bandit environment, dynamic regret is introduced to quantify the cumulative regret incurred by an algorithm over a sequence of time steps, which allow us to evaluate the performance of various algorithm in non-stationary bandit settings.
Suppose that at time t, algorithm A chooses arm\( {X_{t}}=arg{max_{{X_{t}}}}{\widetilde{f}_{t}}(x) \). The corresponding expected dynamic regret is bounded for integer D>0,
\( \begin{matrix}E[R(T)] & ≤({c_{1}}+{c_{2}})(1+\frac{2}{{p_{3}}-{p_{2}}})\sqrt[]{{c_{3}}T} \\ & +T({p_{1}}+{p_{2}})+d+2D{B_{T}}+\frac{4}{λ}\frac{{γ^{D}}}{1-γ}T. \\ \end{matrix} \)(13)
\( {c_{1}}=\sqrt[]{2logT+dlog(1+\frac{1-{γ^{2(T-1)}}}{λd(1-{γ^{2}})})}+{λ^{1/2}}, \)(14)
\( {c_{2}}=a\sqrt[]{2log{(\frac{T}{2})}},and {a^{2}}=14c_{1}^{2} \)(15)
If we choose\( D=\frac{log{T}}{1-γ},γ=1-({B_{T}}/(dT){)^{2/3}} \), the expected dynamic regret is asymptotically upper bounded by\( O({d^{2/3}}B_{T}^{1/3}{T^{2/3}}) \)as\( T→∞. \)
In the D-RandLinUCB algorithm which is designed to overcome conservatism issues faced by optimism-based algorithms in practice, we use the weighted method with exponentially discounting factor to adjust the non-stationary linear bandit environment. Since the action set\( {X_{t}} \)changes from time t and has infinite arms, and the true parameter\( θ_{t}^{⋆} \)varies within total variation\( {B_{T}} \).
D-RandLinUCB achieves statistical optimality in terms of dynamic regret, but it has a trade-off with computational efficiency compared to another algorithm called Discounted Linear Thompson Sampling (D-LinTS).
3.1.5.D-LinTS. In the previous D-LinUCB algorithm [7], the random perturbations were injected by replacing optimism with simple randomization when deciding the confidence level. However, the D-LinTS algorithm chooses to perturb estimates before the expected rewards are maximized [8]. We use a weighted least-squares estimator\( \hat{θ}_{t}^{wls} \)along with the corresponding matrix\( {V_{t}}={W_{t,λ}}\widetilde{W}_{t,λ}^{-1}{W_{t,λ}} \)instead of\( \hat{θ}_{t}^{ls} \)and the Gram matrix\( {V_{t,λ}} \). Since the random perturbations are not shared on each arm, the D-LinTS algorithm has more variation and corresponding larger regret bounds than the previous D-RandLinUCB algorithm.
In the D-LinTS algorithm [8], we choose\( D=logT/(1-γ) \)and\( γ=1-({B_{T}}/(dT\sqrt[]{logK}){)^{2/3}} \), the expected dynamic regret is asymptotically upper bounded by\( O({d^{2/3}}(logK{)^{1/3}}B_{T}^{1/3}{T^{2/3}}) \)as\( T→∞. \)
To ensure that the randomly chosen confidence bound of D-RandLinUCB belongs to that of D-LinUCB with high probability, D-RandLinUCB uses a truncated normal distribution with zero mean and standard deviation 2/5 over [0,∞) as D. On the other hand, when implementing both LinTS and D-LinTS, D-LinTS uses a non-inflated version by setting a=1. According to Li et al. [3], in all scenarios, both randomized algorithms outperform the non-randomized D-LinUCB. However, D-LinTS not only outperforms D-RandLinUCB in all scenarios but also works as well as Oracle Restart LinTS considering the high dimension and big action space.
3.2.High-dimensionl contextual bandits
In the domain of high-dimensional contextual bandits, we move beyond the traditional constraints of fixed probability distributions for arm rewards. Here, the arms are not bound by predefined distributions and can be strategically selected by an adversary, challenging the decision-making process. To navigate this complex landscape, advanced techniques tailored to high-dimensional contexts are employed. These methods empower agents to leverage the rich contextual information, often encoded in high-dimensional feature spaces, to make informed arm selections. The agent's strategy adapts dynamically based on observed rewards, allowing for intricate adjustments in the high-dimensional space to optimize decision-making in this adversarial environment.
The Thompson sampling algorithm is a popular Bayesian algorithm for solving the contextual bandit problem. In the high-dimensional and sparse contextual bandit problem, the algorithm uses special classes of sparsity-inducing priors [9], such as spike-and-slab priors, to model the unknown parameter. The algorithm works by first initializing the prior distribution over the unknown parameter, and then at each round, it samples a parameter from the posterior distribution, which is updated based on the observed rewards and contexts. The algorithm then selects the action that maximizes the expected reward under the sampled parameter. By using sparsity-inducing priors, the algorithm can effectively handle high-dimensional and sparse contexts, and by using Bayesian inference, it can provide a probabilistic estimate of the unknown parameter.
The algorithm used the sparsity-inducing prior proposed in the research of Russac et al. [10] for posterior sampling and established posterior contraction results for non-i.i.d. observations coming from a bandit environment and for a wide class of noise distributions. Using the posterior contraction result, an almost dimension-free regret bound is established for the proposed TS algorithm under different arm-separation regimes parameterized by\( ω \). The algorithm enjoys minimax optimal performance for\( ω∈[0,1) \). In addition, the prior allows us to design a computationally efficient TS algorithm based on Variational Bayes.
First, a dimension s is selected from a prior\( {π_{d}} \)on the set [d]; next, a random subset S⊂[d] of size |S|=s, and finally, given S, a set of nonzero values\( {β_{S}}:=\lbrace {β_{i}}:i∈S\rbrace \)from a prior density\( {g_{S}} \)for\( {R^{S}} \). In the tth round of the algorithm, a specific prior Π is set on β, and it is updated sequentially based on the observed rewards and contexts. In particular, it chooses the prior described in
\( (S,β)↦{π_{d}}(|S|)\frac{1}{(\frac{d}{|S|})}{g_{S}}({β_{S}}){δ_{0}}({β_{{S^{c}}}}) \)(16)
with an appropriate choice of round-specific prior scaling\( {λ_{t}} \)and updates the posterior using the observed rewards and contexts until (t−1)th round. Then a sample is generated from the posterior and an arm\( {a_{t}} \)is chosen greedily based on the generated sample.
Since\( C=Θ({ϕ_{u}}{ϑ^{2}}ξKlogK) \), and\( K≥2,d≥T \). Define the quantity
\( κ(ξ,ϑ,K):=min\lbrace (4{c_{3}}Kξ{ϑ^{2}}{)^{-1}},1/2\rbrace \)(17)
Rewards where c3 is a universal positive constant. Also, set the prior scaling\( {λ_{t}} \)as follows:
\( (5/3){\overset{¯}{λ}_{t}}≤{λ_{t}}≤2{\overset{¯}{λ}_{t}},{\overset{¯}{λ}_{t}}={x_{тах}}\sqrt[]{2t(logd+logt)} \)(18)
Then there exists a universal constant C0>0 such that we have the following regret bound for the algorithm:
\( E\lbrace R(T)\rbrace ≲{I_{b}}+{I_{ω}} \)(19)
where,
\( {I_{b}}=\lbrace \frac{{b_{max}}{x_{max}}{ϕ_{u}}{ϑ^{2}}ξ(KlogK)}{min\lbrace {κ^{2}}(ξ,ϑ,K),logK\rbrace }\rbrace {s^{*}}log(Kd) \)(20)
\( {I_{ω}}=\begin{cases} \begin{array}{c} {Φ^{1+ω}}(\frac{{s^{*1+ω}}{(log{d})^{\frac{1+ω}{2}}}{T^{\frac{1-ω}{2}}}}{Δ_{*}^{ω}}), forω∈[0,1), \\ {Φ^{2}}(\frac{{s^{*2}}[logd+logT]logT}{{Δ_{*}}}), forω=1, \\ \frac{{Φ^{2}}}{(ω-1)}(\frac{{s^{*2}}[logd+logT]}{{Δ_{*}}}), forω∈(1,∞) \\ {Φ^{2}}(\frac{{s^{*2}}[logd+logT]}{{Δ_{*}}}), forω=∞, \end{array} \end{cases} \)(21)
and
\( Φ=σx_{max}^{2}ξK(2+40A_{4}^{-1}+{C_{0}}Kξx_{max}^{2}A_{4}^{-1}) \)(22)
4.Practical applications
The nature of contextual bandit problems makes them suitable for various real-life situation and application (see Table 2). In particular, they can be beneficial when collecting data for assessing treatment effectiveness on animal models throughout different disease stages. Traditionally, conducting such assessments using conventional random treatment allocation procedures can be challenging. This is because administering poor treatments can lead to a deterioration of the subject's health, making data collection difficult. To address this issue, Durand et al. [11] intend to develop an adaptive allocation strategy that allocates more samples, so as to enhance the data-collection efficiency and explore promising treatments. In their work, the authors approach this application as a contextual bandit problem and introduce a practical algorithm for exploration and exploitation within this framework.
Table 2. Bandit for Real Life Application [12].
CMAB |
Non-stationary CMAB |
|
Healthcare |
√ |
|
Recommendation system |
√ |
√ |
Dialogue system |
√ |
In addition to real life applications like clinical trials, contextual bandits could also be used to improve various machine learning algorithms (see Table 3).
Bouneffouf et al. [13] explore this idea by introducing a novel active learning strategy that models the active learning problem as a contextual bandit problem. Their proposed method, called Active Thompson Sampling (ATS), adopts a sequential algorithmic approach. In each round of the algorithm, ATS assigns a sampling distribution on a pool of available unlabeled data points. From this distribution, it samples one point and queries the oracle for the corresponding label of the sampled point. This active learning strategy effectively balances exploration and exploitation by making informed decisions on which data points to query for label information.
Noothigattu et al. [14] conducted a study that focuses on a scenario where an agent can observe the behavioral traces of individuals within a society but lacks access to explicit constraints governing the observed behaviors. To address this challenge, the authors employ inverse reinforcement learning to learn these potential constraints. Once the constraints are learned, they are combined with a potentially unrelated value function using a contextual bandit-based orchestrator. This orchestrator plays a critical role in selecting a contextually-appropriate choice between two policies: the constraint-based policy and the environment reward-based policy. When making decisions, the agent can now mix policies in new approaches, selecting the best actions from either a reward-maximizing policy or a constrained policy.
Table 3. Bandit in Machine Learning [12].
CMAB |
Non-stationary CMAB |
|
Active Learning |
√ |
|
Reinforcement learning |
√ |
5.Conclusion
This paper reviews some of the most notable algorithms of contextual multi-armed bandits and summarizes it, in an organized way (Table 1). The LinUCB algorithm is a linear bandit algorithm that balances exploration and exploitation by modeling the uncertainty of each arm's reward using a linear regression approach [3]. The LinTS algorithm is a linear bandit algorithm that uses Bayesian methods to estimate the uncertainty of each arm, achieving a balance between exploration and exploitation [5]. The D-LinUCB algorithm is an enhanced linear upper confidence bound algorithm that improves the accuracy of uncertainty estimation for each arm by introducing covariance information, achieving a better balance between exploration and exploitation [6]. The D-RandLinUCB algorithm combines randomization and LinUCB approach to achieve improved exploration-exploitation trade-off by effectively estimating the arm rewards and uncertainties [7]. Similarly, the D-LinTS algorithm enhances the LinTS approach by incorporating randomization, leading to more effective uncertainty modeling and a better balance between exploration and exploitation in linear bandit problems [7]. In a high-dimensional environment, the TS algorithm [8] usually employs Bayesian methods to estimate uncertainty for multi-dimensional arms, achieving a more precise balance between exploration and exploitation.
In conclusion, this review has highlighted key insights and trends in contextual multi-armed bandits. It is evident that the traditional UCB algorithm and TS algorithm are no longer sufficient to address the decision-making challenges posed by complex environments in multi-armed bandit problems and that this field has seen significant developments in recent years. While we have discussed important contributions, it's essential to acknowledge the existing uncertainties and the need for further research. Looking ahead, future research in this area could explore extension in multimodal environments or integration of deep learning and reinforcement learning, and the findings presented here are expected to have a lasting impact on many Recommendation Systems like medical decision-making and automated trading.
References
[1]. Bietti, A., Agarwal, A. and Langford, J. (2021). A contextual bandit bake-off, arXiv.org. Available at: https://arxiv.org/abs/1802.04064 (Accessed: 08 September 2023).
[2]. Lin, B. et al. (2020). Contextual bandit with adaptive feature extraction, arXiv.org. Available at: https://arxiv.org/abs/1802.00981 (Accessed: 08 September 2023).
[3]. Li, L. et al. (2010). A contextual-bandit approach to personalized news article recommendation. Proceedings of the 19th international conference on World wide web [Preprint]. doi:10.1145/1772690.1772758.
[4]. Chapelle, O. and Li, L. (2011). An empirical evaluation of thompson sampling. Advances in neural information processing systems, 24.
[5]. Agrawal, S. and Goyal, N. (2014). Thompson sampling for contextual bandits with linear payoffs, arXiv.org. Available at: https://arxiv.org/abs/1209.3352 (Accessed: 08 September 2023).
[6]. Li, L., Chu, W., Langford, J. and Schapire, R. E. (2010). A Contextual-Bandit Approach to Personalized News Article Recommendation. arXiv.org. https://doi.org/10.1145/1772690.1772758.
[7]. Russac, Y., Vernade, C. and Cappé, O. (2020). Weighted Linear Bandits for non-stationary environments, arXiv.org. Available at: https://arxiv.org/abs/1909.09146 (Accessed: 08 September 2023).
[8]. Kim, B. and Tewari, A. (2021). Randomized exploration for non-stationary stochastic linear bandits, arXiv.org. Available at: https://arxiv.org/abs/1912.05695 (Accessed: 08 September 2023).
[9]. Chakraborty, S., Roy, S. and Tewari, A. (2023). Thompson sampling for high-dimensional sparse linear contextual bandits, PMLR. Available at: https://proceedings.mlr.press/v202/chakraborty23b.html (Accessed: 08 September 2023).
[10]. Castillo, I., Schmidt-Hieber, J., and Van der Vaart, A. (2015). Bayesian linear regression with sparse priors. The Annals of Statistics, 43(5), 1986–2018.
[11]. Durand, A., Achilleos, C., Lacovides, D., Strati, K., Mitsis, G. D. and Pineau, J. (2018). Contextual bandits for adapting treatment in a mouse model of de novo carcinogenesis. In Machine Learning for Healthcare Conference, 67–82.
[12]. Bouneffouf, D. and Rish, I. (2019). A survey on practical applications of multi-armed and Contextual Bandits, arXiv.org. Available at: https://arxiv.org/abs/1904.10040 (Accessed: 08 September 2023).
[13]. Bouneffouf, D., Laroche, R., Urvoy, T., Feraud, R. and Allesiardo, R. (2014). Contextual bandit for active learning: Active thompson sampling. In International Conference on Neural Information Processing, 405–412. Springer.
[14]. Noothigattu, R., Bouneffouf, D., Mattei, N., Chandra, R., Madan, P., Varshney, K., Campbell, M., Singh, M. and Rossi, F. (2018). Interpretable multi-objective reinforcement learning through policy orchestration. arXiv preprint arXiv:1809.08343.
Cite this article
Chen,Q. (2024). A survey on contextual multi-armed bandits. Applied and Computational Engineering,53,287-295.
Data availability
The datasets used and/or analyzed during the current study will be available from the authors upon reasonable request.
Disclaimer/Publisher's Note
The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of EWA Publishing and/or the editor(s). EWA Publishing and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
About volume
Volume title: Proceedings of the 4th International Conference on Signal Processing and Machine Learning
© 2024 by the author(s). Licensee EWA Publishing, Oxford, UK. This article is an open access article distributed under the terms and
conditions of the Creative Commons Attribution (CC BY) license. Authors who
publish this series agree to the following terms:
1. Authors retain copyright and grant the series right of first publication with the work simultaneously licensed under a Creative Commons
Attribution License that allows others to share the work with an acknowledgment of the work's authorship and initial publication in this
series.
2. Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the series's published
version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgment of its initial
publication in this series.
3. Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and
during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See
Open access policy for details).
References
[1]. Bietti, A., Agarwal, A. and Langford, J. (2021). A contextual bandit bake-off, arXiv.org. Available at: https://arxiv.org/abs/1802.04064 (Accessed: 08 September 2023).
[2]. Lin, B. et al. (2020). Contextual bandit with adaptive feature extraction, arXiv.org. Available at: https://arxiv.org/abs/1802.00981 (Accessed: 08 September 2023).
[3]. Li, L. et al. (2010). A contextual-bandit approach to personalized news article recommendation. Proceedings of the 19th international conference on World wide web [Preprint]. doi:10.1145/1772690.1772758.
[4]. Chapelle, O. and Li, L. (2011). An empirical evaluation of thompson sampling. Advances in neural information processing systems, 24.
[5]. Agrawal, S. and Goyal, N. (2014). Thompson sampling for contextual bandits with linear payoffs, arXiv.org. Available at: https://arxiv.org/abs/1209.3352 (Accessed: 08 September 2023).
[6]. Li, L., Chu, W., Langford, J. and Schapire, R. E. (2010). A Contextual-Bandit Approach to Personalized News Article Recommendation. arXiv.org. https://doi.org/10.1145/1772690.1772758.
[7]. Russac, Y., Vernade, C. and Cappé, O. (2020). Weighted Linear Bandits for non-stationary environments, arXiv.org. Available at: https://arxiv.org/abs/1909.09146 (Accessed: 08 September 2023).
[8]. Kim, B. and Tewari, A. (2021). Randomized exploration for non-stationary stochastic linear bandits, arXiv.org. Available at: https://arxiv.org/abs/1912.05695 (Accessed: 08 September 2023).
[9]. Chakraborty, S., Roy, S. and Tewari, A. (2023). Thompson sampling for high-dimensional sparse linear contextual bandits, PMLR. Available at: https://proceedings.mlr.press/v202/chakraborty23b.html (Accessed: 08 September 2023).
[10]. Castillo, I., Schmidt-Hieber, J., and Van der Vaart, A. (2015). Bayesian linear regression with sparse priors. The Annals of Statistics, 43(5), 1986–2018.
[11]. Durand, A., Achilleos, C., Lacovides, D., Strati, K., Mitsis, G. D. and Pineau, J. (2018). Contextual bandits for adapting treatment in a mouse model of de novo carcinogenesis. In Machine Learning for Healthcare Conference, 67–82.
[12]. Bouneffouf, D. and Rish, I. (2019). A survey on practical applications of multi-armed and Contextual Bandits, arXiv.org. Available at: https://arxiv.org/abs/1904.10040 (Accessed: 08 September 2023).
[13]. Bouneffouf, D., Laroche, R., Urvoy, T., Feraud, R. and Allesiardo, R. (2014). Contextual bandit for active learning: Active thompson sampling. In International Conference on Neural Information Processing, 405–412. Springer.
[14]. Noothigattu, R., Bouneffouf, D., Mattei, N., Chandra, R., Madan, P., Varshney, K., Campbell, M., Singh, M. and Rossi, F. (2018). Interpretable multi-objective reinforcement learning through policy orchestration. arXiv preprint arXiv:1809.08343.