Hostname: page-component-78c5997874-j824f Total loading time: 0 Render date: 2024-11-10T17:11:27.378Z Has data issue: false hasContentIssue false

A regret lower bound for assortment optimization under the capacitated MNL model with arbitrary revenue parameters

Published online by Cambridge University Press:  01 September 2021

Yannik Peeters
Affiliation:
Amsterdam Business School, University of Amsterdam, Plantage Muidergracht 12, 1018 TV Amsterdam, The Netherlands. E-mail: y.a.peeters@uva.nl
Arnoud V. den Boer
Affiliation:
Amsterdam Business School, University of Amsterdam, Plantage Muidergracht 12, 1018 TV Amsterdam, The Netherlands. E-mail: y.a.peeters@uva.nl Korteweg-de Vries Institute for Mathematics, University of Amsterdam, Science Park 105-107, 1098 XG Amsterdam, The Netherlands
Rights & Permissions [Opens in a new window]

Abstract

In this note, we consider dynamic assortment optimization with incomplete information under the capacitated multinomial logit choice model. Recently, it has been shown that the regret (the cumulative expected revenue loss caused by offering suboptimal assortments) that any decision policy endures is bounded from below by a constant times $\sqrt {NT}$, where $N$ denotes the number of products and $T$ denotes the time horizon. This result is shown under the assumption that the product revenues are constant, and thus leaves the question open whether a lower regret rate can be achieved for nonconstant revenue parameters. In this note, we show that this is not the case: we show that, for any vector of product revenues there is a positive constant such that the regret of any policy is bounded from below by this constant times $\sqrt {N T}$. Our result implies that policies that achieve ${{\mathcal {O}}}(\sqrt {NT})$ regret are asymptotically optimal for all product revenue parameters.

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
Copyright © The Author(s), 2021. Published by Cambridge University Press

1. Introduction

In this note, we consider the problem of assortment optimization under the multinomial logit (MNL) choice model with a capacity constraint on the size of the assortment and incomplete information about the model parameters. This problem has recently received considerable attention in the literature (see, e.g., [Reference Agrawal, Avadhanula, Goyal and Zeevi1Reference Kallus and Udell7,Reference Rusmevichientong, Shen and Shmoys9,Reference Sauré and Zeevi10]).

Two notable recent contributions are from Agrawal et al. [Reference Agrawal, Avadhanula, Goyal and Zeevi1,Reference Agrawal, Avadhanula, Goyal and Zeevi2], who construct decision policies based on Thompson Sampling and Upper Confidence Bounds, respectively, and show that the regret of these policies—the cumulative expected revenue loss compared with the benchmark of always offering an optimal assortment—is, up to logarithmic terms, bounded by a constant times $\sqrt {N T}$, where $N$ denotes the number of products and $T\geqslant N$ denotes the length of the time horizon. These upper bounds are complemented by the recent work from Chen and Wang [Reference Chen and Wang3], who show that the regret of any policy is bounded from below by a positive constant times $\sqrt {N T}$, implying that the policies by Agrawal et al. [Reference Agrawal, Avadhanula, Goyal and Zeevi1,Reference Agrawal, Avadhanula, Goyal and Zeevi2] are (up to logarithmic terms) asymptotically optimal.

The lower bound by Chen and Wang [Reference Chen and Wang3] is proven under the assumption that the product revenues are constant—that is, each product generates the same amount of revenue when sold. In practice, it often happens that different products have different marginal revenues, and it is a priori not completely clear whether the policies by Agrawal et al. [Reference Agrawal, Avadhanula, Goyal and Zeevi1,Reference Agrawal, Avadhanula, Goyal and Zeevi2] are still asymptotically optimal or that a lower regret can be achieved. In addition, Chen and Wang [Reference Chen and Wang3] assume that $K$, the maximum number of products allowed in an assortment, is bounded by $\tfrac {1}{4} \cdot N$, but point out that this constant $\tfrac {1}{4}$ can probably be increased.

In this note, we settle this open question by proving a $\sqrt {N T}$ regret lower bound for any given vector of product revenues. This implies that policies with ${{\mathcal {O}}}(\sqrt {N T})$ regret are asymptotically optimal regardless of the product revenue parameters. Furthermore, our result is valid for all $K < \tfrac {1}{2} N$, thereby confirming the intuition of Chen and Wang [Reference Chen and Wang3] that the constraint $K \leqslant \tfrac {1}{4} N$ is not tight.

2. Model and main result

We consider the problem of dynamic assortment optimization under the MNL choice model. In this model, the number of products is $N \in {{\mathbb {N}}}$. Henceforth, we abbreviate the set of products $\{1\ldots ,N\}$ as $[N]$. Each product $i\in [N]$ yields a known marginal revenue to the seller of $w_i> 0$. Without loss of generality due to scaling, we can assume that $w_i\leqslant 1$ for all $i\in [N]$. Each product $i\in [N]$ is associated with a preference parameter $v_i\geqslant 0$, unknown to the seller. Each offered assortment $S\subseteq [N]$ must satisfy a capacity constraint, that is, $|S|\leqslant K$ for capacity constraint $K \in {{\mathbb {N}}}$, $K\leqslant N$. For notational convenience, we write

$$\mathcal{A}_K := \{S\subseteq[N]:|S|\leqslant K\}$$

for the collection of all assortments of size at most $K$, and

$$\mathcal{S}_K := \{S\subseteq[N]:|S|= K\}$$

for the collection of all assortments of exact size $K$. Let $T \in {{\mathbb {N}}}$ denote a finite time horizon. Then, at each time $t\in [T]$, the seller selects an assortment $S_t\in \mathcal {A}_K$ based on the purchase information available up to and including time $t-1$. Thereafter, the seller observes a single purchase $Y_t\in S_t\cup \{0\}$, where product $0$ indicates a no-purchase. The purchase probabilities within the MNL model are given by

$${{\mathbb{P}}}(Y_t = i\,|\,S_t=S) = \frac{v_i}{1+\sum_{j\in S} v_j},$$

for all $t \in [T]$ and $i \in S \cup \{0\}$, where we write $v_0 := 1$. The assortment decisions of the seller are described by his/her policy: a collection of probability distributions $\pi = (\pi (\,\cdot \,|\,h) : h \in H)$ on $\mathcal {A}_K$, where

$$H:= \bigcup_{t \in [T]} \{ (S, Y) : Y \in S \cup \{0\}, S \in \mathcal{A}_K\}^{t-1}$$

is the set of possible histories. Then, conditionally on $h=(S_1, Y_1, \ldots , S_{t-1}, Y_{t-1})$, assortment $S_t$ has distribution $\pi (\,\cdot \,|\,h)$, for all $h \in H$ and all $t \in [T]$. Let $\mathbb {E}^{\pi }_v$ be the expectation operator under policy $\pi$ and preference vector $v\in \mathcal {V} := [0, \infty )^{N}$. The objective for the seller is to find a policy $\pi$ that maximizes the total accumulated revenue or, equivalently, minimizes the accumulated regret:

$$\Delta_{\pi}(T, v) := T \cdot \max_{S \in \mathcal{A}_K} r(S, v) - \sum_{t=1}^{T} \mathbb{E}^{\pi}_v [ r(S_t, v)],$$

where $r(S,v)$ is the expected revenue of assortment $S\subseteq [N]$ under preference vector $v \in \mathcal {V}$:

$$r(S,v) := \frac{\sum_{i\in S} v_i w_i}{1+\sum_{i\in S} v_i}.$$

In addition, we define the worst-case regret:

$$\Delta_{\pi}(T) := \sup_{v \in \mathcal{V}} \Delta_{\pi}(T, v).$$

The main result, presented below, states that the regret of any policy can uniformly be bounded from below by a constant times $\sqrt {NT}$.

Theorem 1 Suppose that $K < N / 2$. Then, there exists a constant $c_1 > 0$ such that, for all $T \geqslant N$ and for all policies $\pi$,

$$\Delta_\pi(T) \geqslant c_1\,\sqrt{NT}.$$

3. Proof of Theorem 1

3.1. Proof outline

The proof of Theorem 1 can be broken up into four steps. First, we define a baseline preference vector $v^{0}\in \mathcal {V}$ and we show that under $v^{0}$ any assortment $S\in \mathcal {S}_K$ is optimal. Second, for each $S\in \mathcal {A}_K$, we define a preference vector $v^{S}\in \mathcal {V}$ by

(3.1)\begin{equation} v_i^{S} :=\left\{\begin{array}{ll} v^{0}_i(1+\epsilon), & \text{if }i\in S, \\ v^{0}_i, & \text{otherwise}, \end{array}\right. \end{equation}

for some $\epsilon \in (0,1]$. For each such $v^{S}$, we show that the instantaneous regret from offering a suboptimal assortment $S_t$ is bounded from below by a constant times the number of products $|S \setminus S_t|$ not in $S$; cf. Lemma 1 below. This lower bound takes into account how much the assortments $S_1,\ldots ,S_T$ overlap with $S$ when the preference vector is $v^{S}$. Third, let $N_i$ denote the number of times product $i\in [N]$ is contained $S_1,\ldots ,S_T$, that is,

$$N_i:= \sum_{t=1}^{T} \textbf{1}\{i\in S_t\}.$$

Then, we use the Kullback–Leibler (KL) divergence and Pinsker's inequality to upper bound the difference between the expected value of $N_i$ under $v^{S}$ and $v^{S\backslash \{i\}}$, see Lemma 2. Fourth, we apply a randomization argument over $\{v^{S}:S\in \mathcal {S}_K\}$, we combine the previous steps, and we set $\epsilon$ accordingly to conclude the proof.

The novelty of this work is concentrated in the first two steps. The third and fourth step closely follow the work of Chen and Wang [Reference Chen and Wang3]. These last steps are included (1) because of slight deviations in our setup, (2) for the sake of completeness, and (3) since the proof techniques are extended to the case where $K/N<1/2$. In the work of Chen and Wang [Reference Chen and Wang3], the lower bound is shown for $K/N\leqslant 1/4$, but the authors already mention that this constraint can probably be relaxed. Our proof confirms that this is indeed the case.

3.2. Step 1: Construction of the baseline preference vector

Let $\underline {w} := \min _{i\in [N]} w_i>0$ and define the constant

$$s:= \frac{\underline{w}^{2}}{3+2\underline{w}}.$$

Note that $s<\underline {w}$. The baseline preference vector is formally defined as

$$v^{0}_i := \frac{s}{K(w_i-s)},\quad \text{for all } i\in[N].$$

Now, the expected revenue for any $S\in \mathcal {A}_K$ under $v^{0}$ can be rewritten as

$$r(S,v^{0}) = \frac{\sum_{i\in S}v^{0}_i w_i}{1 + \sum_{i\in S}v^{0}_i} = \frac{ s\sum_{i\in S} \frac{w_i}{w_i-s}}{ K + \sum_{i\in S} \frac{s}{w_i-s}} = \frac{ s\sum_{i\in S} \frac{w_i}{w_i-s}}{ K - |S| + \sum_{i\in S} \frac{w_i}{w_i-s}}.$$

The expression on the right-hand side is only maximized by assortments $S$ with maximal size $|S| = K$, in which case

$$r(S,v^{0}) = \max_{S'\in\mathcal{A}_K} r(S',v^{0}) = s.$$

It follows that all assortments $S$ with size $K$ are optimal.

3.3. Step 2: Lower bound on the instantaneous regret of $v^{S}$

For the second step, we bound the instantaneous regret under $v^{S}$.

Lemma 1 Let $S\in \mathcal {S}_K$. Then, there exists a constant $c_2>0$, only depending on $\underline {w}$ and $s$, such that, for all $t\in [T]$ and $S_t\in \mathcal {A}_K$,

$$\max_{S'\in \mathcal{A}_K} r(S', v^{S}) - r(S_t,v^{S}) \geqslant c_2\frac{\epsilon\,|S\backslash S_t|}{K}.$$

As a consequence,

(3.2)\begin{equation} T \cdot \max_{S'\in \mathcal{A}_K} r(S', v^{S}) - \sum_{t=1}^{T} r(S_t,v^{S}) \geqslant c_2\,\epsilon\left(T - \frac{1}{K}\sum_{i\in S}N_i\right). \end{equation}

Proof. Fix $S\in \mathcal {S}_K$. First, note that since $\epsilon \leqslant 1$, for any $S'\in \mathcal {A}_K$, it holds that

(3.3)\begin{equation} \sum_{i\in S'}v^{S}_i \leqslant \frac{2s}{\underline{w}-s}. \end{equation}

Second, let $S^{*}\in \textrm {arg max}_{S'\in \mathcal {A}_K}\,r(S',v^{S})$ and $\varrho ^{*} = r(S^{*},v^{S})$. By rewriting the inequality $\varrho ^{*}\geqslant r(S',v^{S})$ for all $S'\in \mathcal {A}_K$, we find that for all $S'\in \mathcal {A}_K$

(3.4)\begin{equation} \varrho^{*} \geqslant \sum_{i\in S'} v^{S}_i(w_i-\varrho^{*}). \end{equation}

Let $t\in [T]$ and $S_t\in \mathcal {A}_K$. Then, it holds that

\begin{align*} r(S^{*},v^{S}) - r(S_t,v^{S}) & = \varrho^{*} - \frac{\sum_{i\in S_t} v^{S}_i w_i}{1 + \sum_{i\in S_t} v^{S}_i} \\ & = \frac{1}{1 + \sum_{i\in S_t} v^{S}_i}\left(\varrho^{*} + \sum_{i\in S_t} v^{S}_i\varrho^{*} - \sum_{i\in S_t} v^{S}_i w_i\right)\\ & \geqslant \frac{\underline{w}-s}{\underline{w}+s}\left(\varrho^{*} - \sum_{i\in S_t} v^{S}_i(w_i-\varrho^{*})\right)\\ & \geqslant \frac{\underline{w}-s}{\underline{w}+s}\left(\sum_{i\in S} v^{S}_i(w_i-\varrho^{*}) - \sum_{i\in S_t} v^{S}_i(w_i-\varrho^{*})\right) \\ & = \frac{\underline{w}-s}{\underline{w}+s} \left(\vphantom{\sum_{i\in S}}\right. \underbrace{\sum_{i\in S} v^{S}_i(w_i-s) - \sum_{i\in S_t} v^{S}_i(w_i-s)}_{(a)} \\ & \quad- \underbrace{(\varrho^{*}-s)\left( \sum_{i\in S} v^{S}_i - \sum_{i\in S_t} v^{S}_i\right)}_{(b)} \left.\vphantom{\sum_{i\in S}}\right). \end{align*}

Here, the first inequality is due to (3.3) and the second inequality follows from (3.4) with $S'=S$. Next, note that since $|S_t|\leqslant K$ and $|S|=K$, we find that

(3.5)\begin{equation} |S_t\backslash S|\leqslant |S\backslash S_t|. \end{equation}

Now, term $(a)$ can be bounded from below as

(3.6)\begin{align} (a) & = \sum_{i\in S\backslash S_t} v^{S}_i(w_i-s) - \sum_{i\in S_t\backslash S} v^{S}_i(w_i-s) \nonumber\\ & = \frac{s}{K}((1+\epsilon)|S\backslash S_t| - |S_t\backslash S|) \nonumber\\ & \geqslant s \,\frac{\epsilon\,|S\backslash S_t|}{K}. \end{align}

Here, at the final inequality, we used (3.5). Next, term $(b)$ can be bounded from above as

$$(b) \leqslant \underbrace{|\varrho^{*}-s|}_{(c)}\,\underbrace{\left| \sum_{i\in S} v^{S}_i - \sum_{i\in S_t} v^{S}_i\right|}_{(d)}.$$

Now, for term $(c)$, we note that $v^{S}_i\geqslant v^{0}_i$ for all $i\in [N]$. In addition, since $r(S^{*},v^{0})\leqslant s$,

\begin{align*} \varrho^{*} - s & \leqslant \frac{\sum_{i\in S^{*}}v^{S}_i w_i}{1 + \sum_{i\in S^{*}}v^{S}_i} - \frac{\sum_{i\in S^{*}}v^{0}_i w_i}{1 + \sum_{i\in S^{*}}v^{0}_i}\\ & \leqslant \frac{1}{1 + \sum_{i\in S^{*}}v^{0}_i}\sum_{i\in S^{*}}(v^{S}_i-v^{0}_i)w_i\\ & \leqslant \sum_{i=1}^{N}(v^{S}_i-v^{0}_i) = \epsilon\sum_{i\in S}v^{0}_i \leqslant \frac{s}{\underline{w} - s} \epsilon. \end{align*}

This entails an upper bound for $(c)$. Term $(d)$ is bounded from above as

\begin{align*} (d)& \leqslant \sum_{i\in S\backslash S_t} v^{S}_i + \sum_{i\in S_t\backslash S} v^{S}_i \\ & \leqslant (1+\epsilon)\sum_{i\in S\backslash S_t} v^{0}_i + \sum_{i\in S_t\backslash S} v^{0}_i \\ & \leqslant (1+\epsilon)\frac{s}{K(\underline{w} - s)}|S\backslash S_t| + \frac{s}{K(\underline{w} - s)}|S_t\backslash S|\\ & \leqslant \frac{3s}{\underline{w} - s} \frac{|S\backslash S_t|}{K}. \end{align*}

Here, at the final inequality, we used (3.5) and the fact that $\epsilon \leqslant 1$. Now, we combine the upper bounds of $(c)$ and $(d)$ to find that

(3.7)\begin{equation} (b) \leqslant \frac{3s^{2}}{(\underline{w} -s)^{2}}\cdot \frac{\epsilon\,|S\backslash S_t|}{K}. \end{equation}

It follows from (3.6) and (3.7) that

\begin{align*} r(S^{*},v^{S}) - r(S_t,v^{S}) & \geqslant \frac{\underline{w}-s}{\underline{w}+s}\left(s -\frac{3s^{2}}{(\underline{w} -s)^{2}}\right)\frac{\epsilon |S\backslash S_t|}{K}\\ & \geqslant c_2 \frac{\epsilon|S\backslash S_t|}{K}, \end{align*}

where

$$c_2 := \frac{\underline{w}-s}{\underline{w}+s}\left(s -\frac{3s^{2}}{(\underline{w} -s)^{2}}\right).$$

Note that the constant $c_2$ is positive if $(\underline {w}-s)^{2}>3s$. This follows from $s=\underline {w}^{2}/(3+2\underline {w})$ since

$$(\underline{w}-s)^{2} - 3s > \underline{w}^{2} - s(3+2\underline{w}).$$

Statement (3.2) follows from the additional observation

$$\sum_{t=1}^{T} |S\backslash S_t| = TK - \sum_{t=1}^{T} |S\cap S_t|= TK - \sum_{i\in S}N_i.$$

3.4. Step 3: KL divergence and Pinsker's inequality

We denote the dependence of the expected value and the probability on the preference vector $v^{S}$ as $\mathbb {E}_S[\,\cdot \,]$ and ${{\mathbb {P}}}_S(\cdot )$ for $S\in \mathcal {A}_K$. In addition, we write $S \backslash i$ instead of $S \backslash \{i\}$. The lemma below states an upper bound on the KL divergence of ${{\mathbb {P}}}_S$ and ${{\mathbb {P}}}_{S\backslash i}$ and uses Pinsker's inequality to derive an upper bound on the absolute difference between the expected value of $N_i$ under $v^{S}$ and $v^{S\backslash i}$.

Lemma 2 Let $S\in \mathcal {S}_K$, $S'\in \mathcal {A}_K$, and $i\in S$. Then, there exists a constant $c_3$, only depending on $\underline {w}$ and $s$, such that

$${\textrm{KL}}({{\mathbb{P}}}_S(\,\cdot\,|S') \,|\!|\, {{\mathbb{P}}}_{S\backslash i}(\,\cdot\,|S')) \leqslant c_3\frac{\epsilon^{2}}{K}.$$

As a consequence,

(3.8)\begin{equation} |\mathbb{E}_S [N_i] - \mathbb{E}_{S\backslash i} [N_i]| \leqslant \sqrt{2c_3}\frac{\epsilon T^{3/2}}{\sqrt{K}}. \end{equation}

Proof. Let ${{\mathbb {P}}}$ and ${{\mathbb {Q}}}$ be arbitrary probability measures on $S'\cup \{0\}$. It can be shown, see, for example, Lemma 3 from Chen and Wang [Reference Chen and Wang3], that

$${\textrm{KL}}({{\mathbb{P}}}\,|\!|\,{{\mathbb{Q}}}) \leqslant \sum_{j\in S'\cup\{0\}} \frac{(p_j-q_j)^{2}}{q_j},$$

where $p_j$ and $q_j$ are the probabilities of outcome $j$ under ${{\mathbb {P}}}$ and ${{\mathbb {Q}}}$, respectively. We apply this result for $p_j$ and $q_j$ defined as

$$p_j := \frac{v^{S}_j}{1+\sum_{\ell\in S'}v^{S}_\ell} \quad\text{and}\quad q_j := \frac{v^{S\backslash i}_j}{1+\sum_{\ell\in S'}v^{S\backslash i}_\ell},$$

for $j\in S'\cup \{0\}$. First, note that by (3.3), for all $j\in S'\cup \{0\}$,

$$q_j\geqslant \frac{v^{0}_j}{1 + 2\frac{s}{\underline{w} - s}} = \frac{\underline{w} - s}{\underline{w}+s}v^{0}_j.$$

Now, we bound $|p_j-q_j|$ from above for $j\in S'\cup \{0\}$. Note that for $j=0$ it holds that

\begin{align*} |p_0-q_0| & = \frac{|\sum_{\ell\in S'}v^{S}_\ell - \sum_{\ell\in S'}v^{S\backslash i}_\ell|}{(1 + \sum_{\ell\in S'}v^{S}_\ell)(1 + \sum_{\ell\in S'}v^{S\backslash i}_\ell)}\\ & \leqslant |(1+\epsilon)v^{0}_i - v^{0}_i| = v^{0}_i \epsilon. \end{align*}

For $j\neq i$, since $\epsilon \leqslant 1$, we find that

$$|p_j-q_j| = v_j^{S}|p_0-q_0| \leqslant 2 v^{0}_j v^{0}_i \epsilon.$$

For $j=i$, we find that

\begin{align*} |p_i-q_i| & = v^{0}_i|p_0 - q_0 + \epsilon p_0|\\ & \leqslant v^{0}_i(|p_0 - q_0| + \epsilon p_0) \\ & \leqslant v^{0}_i(v^{0}_i + 1)\epsilon. \end{align*}

Therefore, we conclude that

\begin{align*} {\textrm{KL}}({{\mathbb{P}}}_S(\,\cdot\,|S')\,|\!|\,{{\mathbb{P}}}_{S\backslash i}(\,\cdot\,|S')) & \leqslant \sum_{j\in S'\cup\{0\}} \frac{(p_j-q_j)^{2}}{q_j}\\ & \leqslant \frac{(p_0-q_0)^{2}}{q_0} + \sum_{j\in S':j\neq i} \frac{(p_j-q_j)^{2}}{q_j} + \frac{(p_i-q_i)^{2}}{q_i}\\ & \leqslant \frac{\underline{w}+s}{\underline{w}-s} \left((v_i^{0}\epsilon)^{2} + 4(v^{0}_i\epsilon)^{2}\sum_{j\in S':j\neq i}v_j^{0} + v_i^{0}(v_i^{0}+1)^{2} \epsilon^{2} \right)\\ & \leqslant \frac{s(\underline{w}+s)}{(\underline{w}-s)^{2}}\left(\frac{s}{\underline{w} -s } + \frac{4s^{2}}{(\underline{w}-s)^{2}} + \left(\frac{\underline{w}}{\underline{w}-s}\right)^{2}\right)\frac{\epsilon^{2}}{K}\\ & = c_3\frac{\epsilon^{2}}{K}, \end{align*}

where

$$c_3 := \frac{s(\underline{w}+s)}{(\underline{w}-s)^{2}}\left(\frac{s}{\underline{w} -s } + \frac{4s^{2} + \underline{w}^{2}}{(\underline{w}-s)^{2}}\right).$$

Next, note that the entire probability measures ${{\mathbb {P}}}_S$ and ${{\mathbb {P}}}_{S\backslash i}$ depend on $T$. Then, as a consequence of the chain rule of the KL divergence, we find that

$${\textrm{KL}}({{\mathbb{P}}}_S \,|\!|\,{{\mathbb{P}}}_{S\backslash i})\leqslant c_3\,\frac{\epsilon^{2}\,T}{K}.$$

Now, statement (3.8) follows from

(3.9)\begin{align} |\mathbb{E}_S [N_i] - \mathbb{E}_{S\backslash i} [N_i]|& \leqslant \sum_{n=0}^{T} n |{{\mathbb{P}}}_S(N_i=n) - {{\mathbb{P}}}_{S\backslash i} (N_i=n)| \nonumber\\ & \leqslant T\sum_{n=0}^{T} |{{\mathbb{P}}}_S(N_i=n) - {{\mathbb{P}}}_{S\backslash i} (N_i=n)| \nonumber\\ & = 2T\max_{n=0,\ldots,T}|{{\mathbb{P}}}_S(N_i=n) - {{\mathbb{P}}}_{S\backslash i} (N_i=n)| \nonumber\\ & \leqslant 2T\sup_{A} |{{\mathbb{P}}}_S(A) - {{\mathbb{P}}}_{S\backslash i} (A)| \nonumber\\ & \leqslant T\sqrt{2\,{\textrm{KL}}({{\mathbb{P}}}_S \,|\!|\,{{\mathbb{P}}}_{S\backslash i})}, \end{align}

where the step in (3.9) follows from, for example, Proposition 4.2 from Levin et al. [Reference Levin, Peres and Wilmer8] and we used Pinsker's inequality at the final inequality.

3.5. Step 4: Proving the main result

With all the established ingredients, we can finalize the proof of the lower bound on the regret.

Proof of Theorem 1. Since $v^{S}\in \mathcal {V}$ for all $S\in \mathcal {S}_K$ and by Lemma 1, we know that

(3.10)\begin{align} \Delta_{\pi}(T) & \geqslant \frac{1}{|\mathcal{S}_K|} \sum_{S \in \mathcal{S}_K} \Delta_{\pi}(T,v^{S}) \nonumber\\ & \geqslant c_2\,\epsilon \left(\vphantom{\sum_{S \in \mathcal{S}_K}}\right. T - \underbrace{\frac{1}{|\mathcal{S}_K|} \sum_{S \in \mathcal{S}_K} \frac{1}{K} \sum_{i \in S} \mathbb{E}_S [N_i]}_{(a)}\left.\vphantom{\sum_{S \in \mathcal{S}_K}}\right). \end{align}

We decompose $(a)$ into two terms:

$$(a) = \underbrace{\frac{1}{|\mathcal{S}_K|} \sum_{S \in \mathcal{S}_K} \frac{1}{K} \sum_{i \in S} \mathbb{E}_{S \backslash i} [N_i] }_{(b)} + \underbrace{\frac{1}{|\mathcal{S}_K|} \sum_{S \in \mathcal{S}_K} \frac{1}{K} \sum_{i \in S} ( \mathbb{E}_{S} [N_i] - \mathbb{E}_{S \backslash i} [N_i] ) }_{(c)}.$$

Recall that $c = K/N\in (0,1/2)$. By summing over $S' = S \backslash i$ instead of over $S$, we bound $(b)$ from above by

$$(b) = \frac{1}{|\mathcal{S}_K|} \sum_{S' \in \mathcal{S}_{K-1}} \frac{1}{K} \sum_{i \notin S'} \mathbb{E}_{S'} [N_i] \leqslant \frac{|\mathcal{S}_{K-1}|}{|\mathcal{S}_K|} T \leqslant \frac{c}{1-c} T,$$

where the first inequality follows from $\sum _{i \in [N]} \mathbb {E}_{S' }[N_i] \leqslant T K$, and the second inequality from

$$\frac{ |\mathcal{S}_{K-1}| }{ |\mathcal{S}_K|} = \frac{\binom{N}{K-1} }{ \binom{N}{K} } = \frac{K}{ N-K+1} \leqslant \frac{ K/N}{1 - K/N}.$$

Now, $(c)$ can be bounded by applying Lemma 2:

$$(c)\leqslant \sqrt{2c_3}\frac{\epsilon T^{3/2}}{\sqrt{K}} = \frac{\sqrt{2c_3}}{\sqrt{c}} \frac{\epsilon T^{3/2}}{\sqrt{N}}.$$

By plugging the upper bounds on $(b)$ and $(c)$ in (3.10), we obtain

\begin{align*} \Delta_{\pi}(T) & \geqslant c_2\epsilon \left( T - \frac{c}{1 - c} T - \frac{\sqrt{2c_3}}{\sqrt{c}} \frac{\epsilon T^{3/2}}{\sqrt{N}} \right)\\ & = c_2 \epsilon \left( \frac{1-2c}{1 - c} T - \frac{\sqrt{2c_3}}{\sqrt{c}} \frac{\epsilon T^{3/2}}{\sqrt{N}} \right). \end{align*}

Now, we set $\epsilon$ as

$$\epsilon = \min\left\{ 1, \frac{(1-2c)\sqrt{c}}{2(1 - c)\sqrt{2c_3}}\sqrt{N/ T} \right\}.$$

This yields, for all $T \geqslant N$,

$$\Delta_{\pi}(T) \geqslant \min\left\{\frac{c_2\sqrt{2c_3}}{\sqrt{c}} \,T, \frac{c_2(1 - 2c)^{2}c}{8(1-c)\sqrt{2c_3}} \sqrt{N T} \right\}.$$

Finally, note that for $T\geqslant N$ it follows that $T\geqslant \sqrt {NT}$ and therefore

$$\Delta_{\pi}(T) \geqslant c_1 \sqrt{N T},$$

where

$$c_1:=\min\left\{\frac{c_2\sqrt{2c_3}}{\sqrt{c}}, \frac{c_2(1 - 2c)^{2}c}{8(1-c)\sqrt{2c_3}}\right\}>0.$$

Acknowledgments

The authors thank the Editor in Chief and an anonymous reviewer for their positive remarks and useful comments.

Competing interest

The authors declare no conflict of interest.

References

Agrawal, S., Avadhanula, V., Goyal, V., & Zeevi, A. (2017). Thompson sampling for the MNL-bandit. In Conference on Learning Theory (COLT), pp. 76–78.Google Scholar
Agrawal, S., Avadhanula, V., Goyal, V., & Zeevi, A. (2019). MNL-bandit: a dynamic learning approach to assortment selection. Operations Research 67(5): 14531485.CrossRefGoogle Scholar
Chen, X. & Wang, Y. (2018). A note on a tight lower bound for MNL-bandit assortment selection models. Operations Research Letters 46(5): 534537.CrossRefGoogle Scholar
Chen, X., Wang, Y., & Zhou, Y. (2018). An optimal policy for dynamic assortment planning under uncapacitated multinomial logit models. ArXiv e-print. https://arxiv.org/abs/1805.04785.Google Scholar
Cheung, W.C. & Simchi-Levi, D. (2017). Assortment optimization under unknown multinomial logit choice models. ArXiv e-print. https://arxiv.org/abs/1704.00108.Google Scholar
Farias, V.F., Jagabathula, S., & Shah, D. (2013). A nonparametric approach to modeling choice with limited data. Management Science 59(2): 305322.CrossRefGoogle Scholar
Kallus, N. & Udell, M. (2016). Dynamic assortment personalization in high dimensions. ArXiv e-print. https://arxiv.org/abs/1610.05604.Google Scholar
Levin, D.A., Peres, Y., & Wilmer, E.L. (2017). Markov chains and mixing times, 2nd ed. Providence, RI: American Mathematical Society.CrossRefGoogle Scholar
Rusmevichientong, P., Shen, Z.-J.M., & Shmoys, D.B. (2010). Dynamic assortment optimization with a multinomial logit choice model and capacity constraint. Operations Research 58(6): 16661680.CrossRefGoogle Scholar
Sauré, D. & Zeevi, A. (2013). Optimal dynamic assortment planning with demand learning. Manufacturing & Service Operations Management 15(3): 387404.CrossRefGoogle Scholar