Abstract
This paper proposes a new family of algorithms for the online optimisation of composite objectives. The algorithms can be interpreted as the combination of the exponentiated gradient and p-norm algorithm. Combined with algorithmic ideas of adaptivity and optimism, the proposed algorithms achieve a sequence-dependent regret upper bound, matching the best-known bounds for sparse target decision variables. Furthermore, the algorithms have efficient implementations for popular composite objectives and constraints and can be converted to stochastic optimisation algorithms with the optimal accelerated rate for smooth objectives.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
Many machine learning problems involve minimising high dimensional composite objectives (Dhurandhar et al., 2018; Lu et al., 2014; Ribeiro et al., 2016; Xie et al., 2018). For example, in the task of explaining predictions of an image classifier (Dhurandhar et al., 2018; Ribeiro et al., 2016), we need to find a sufficiently small set of features explaining the prediction by solving the following constrained optimisation problem
where l is a function relating to the classifier, \(\lambda _{1}\) controls the sparsity of the feature set, \(\lambda _{2}\) controls the complexity of the feature set, and \(c_{1},\ldots ,c_{d}\) are the ranges of the features. For l with a complicated structure and large d, it is practical to solve the problem by optimising the first-order approximation of the objective function (Lan, 2020). However, the first-order methods can not attain optimal performance due to the non-smooth component \(\lambda _{1}\Vert \cdot \Vert _{1}\). Furthermore, the purpose of introducing the \(\ell _{1}\) regularisation is to ensure the sparsity of the decision variable. Applying the first-order algorithms directly on the subgradient of \(\lambda _{1}\Vert \cdot \Vert _{1}\) does not lead to sparse updates (Duchi et al., 2010). We refer to the objective function consisting of a loss with a complicated structure and a simple (possibly non-smooth) convex regularisation term as a composite objective.
This paper focuses on the more general online convex optimisation (OCO), which can be considered as an iterative game between a player and an adversary. In each round t of the game, the player makes a decision \(x_{t}\in {\mathcal{K}}\). Next, the adversary selects and reveals a convex loss \(l_{t}\) to the player, who then suffers the composite loss \(f_{t}(x)=l_{t}(x)+r_{t}(x)\), where \(l_{t}:{\mathcal{K}}\rightarrow {\mathbb{R}}\) is a convex function revealed at each iteration and \(r_{t}:{\mathbb{X}}\rightarrow {\mathbb{R}}_{\ge 0}\) is a known closed convex function. The target is to develop algorithms minimising the regret of not choosing the best decision \(x\in {\mathcal{K}}\)
An online optimisation algorithm can be converted into a stochastic optimisation algorithm using the online-to-batch conversion technique (Cesa-Bianchi et al., 2004), which is our primary motivation. In addition to that, online optimisation also has many direct applications, such as recommender systems (Song et al., 2014) and time series prediction (Anava et al., 2013).
Given a sequence of subgradients \(\{g_{t}\}\) of \(\{l_{t}\}\), we are interested in the so-called adaptive algorithms ensuring regret bounds of the form \({\mathcal{O}}(\sqrt{\sum _{t=1}^{T}\Vert g_{t}\Vert _{*} ^{2}})\). The adaptive algorithms are worst-case optimal in the online setting (McMahan & Streeter, 2010) and can be converted into stochastic optimisation algorithms with optimal convergence rates (Cutkosky, 2019; Joulani et al., 2020; Kavis et al., 2019; Levy et al., 2018). The adaptive subgradient methods (AdaGrad) (Duchi et al., 2011) and their variants (Alacaoglu et al., 2020; Duchi et al., 2011; Orabona & Pál, 2018; Orabona et al., 2015) have become the most popular adaptive algorithms in recent years. They are often applied to estimating deep learning models and outperform standard optimisation algorithms when the gradient vectors are sparse. However, such property can not be expected in every problem. If the decision variables are in an \(\ell _{1}\) ball and gradient vectors are dense, the Adagrad-style algorithms do not have an optimal theoretical guarantee due to the sub-linear regret dependence on the dimensionality.
The exponentiated gradient (EG) methods (Arora et al., 2012; Kivinen & Warmuth, 1997), which are designed for estimating weights in the positive orthant, enjoy the regret bound growing logarithmically with the dimensionality. The \({{\textit{EG}}^{\pm }}\) algorithm generalises this idea to negative weights (Kivinen & Warmuth, 1997; Warmuth, 2007). Given d dimensional problems with the maximum norm of the gradient bounded by G, the regret of \({{\textit{EG}}^{\pm }}\) is upper bounded by \({\mathcal{O}}(G\sqrt{T\ln d})\). As the performance of the \({{\textit{EG}}^{\pm }}\) algorithm depends strongly on the choice of hyperparameters, the p-norm algorithm (Gentile, 2003), which is less sensitive to the tuning of hyperparameters, is introduced to approach the logarithmic behaviour of \({{\textit{EG}}^{\pm }}\). Kakade et al. (2012) further extends the p-norm algorithm to learning with matrices. An adaptive version of the p-norm algorithm is analysed in Orabona et al. (2015), which has a regret upper bound proportional to \(\Vert x\Vert _{p,*}^{2}\sqrt{\sum _{t=1}^{T}\Vert g_{t}\Vert _{p}^{2}}\) for a given sequence of gradients \(\{g_{t}\}\). By choosing \(p=2\ln d\), a regret upper bound \({\mathcal{O}}(\Vert x\Vert _{1}^{2}\sqrt{\ln d \sum _{t=1}^{T}\Vert g_{t}\Vert _{\infty }^{2}})\) can be achieved. However, tuning hyperparameters is still required to attain the optimal regret \({\mathcal{O}}(\Vert x\Vert _{1}\sqrt{\ln d \sum _{t=1}^{T}\Vert g_{t}\Vert _{\infty }^{2}})\).
Recently, Ghai et al. (2020) has introduced a hyperbolic regulariser for online mirror descent update (HU), which can be viewed as an interpolation between gradient descent and EG. It has a logarithmic behaviour as in EG and a stepsize that can be flexibly scheduled as gradient descent. However, many optimisation problems with sparse targets have an \(\ell _{1}\) or nuclear regulariser in the objective function. Otherwise, the optimisation algorithm has to pick a decision variable from a compact decision set. Due to the hyperbolic regulariser, it is difficult to derive a closed-form solution for either case. Ghai et al. (2020) has proposed a workaround by tuning a temperature-like hyperparameter to normalise the decision variable at each iteration, which is equivalent to the \({{\textit{EG}}^{\pm }}\) algorithm and leads to a performance dependence on the tuning.
This paper proposes a family of algorithms for the online optimisation of composite objectives. The algorithms employ an entropy-like regulariser combined with algorithmic ideas of adaptivity and optimism. Equipped with the regulariser, the online mirror descent (OMD) and the follow-the-regulariser-leader (FTRL) algorithms update the absolute value of the scalar components of the decision variable in the same way as EG in the positive orthant. The directions of the decision variables are set in the same way as the p-norm algorithm. To derive the regret upper bound, we first show that the regulariser is strongly convex with respect to the \(\ell _{1}\)-norm over the \(\ell _{1}\) ball. Then we analyse the algorithms in the comprehensive framework for optimistic algorithms with adaptive regularisers (Joulani et al., 2017). Given the radius of decision set D, sequences of gradients \(\{g_{t}\}\) and hints \(\{h_{t}\}\), the proposed algorithms achieve a regret upper bound in the form of \({\mathcal{O}}(D\sqrt{\ln d\sum _{t=1}^{T}\Vert g_{t}-h_{t}\Vert ^{2}_{\infty }})\). With the techniques introduced in Ghai et al. (2020), a spectral analogue of the entropy-like regulariser can be found and proved to be strongly convex with respect to the nuclear norm over the nuclear ball, from which the best-known regret upper bound depending on \(\sqrt{\ln (\min \{m,n\})}\) for problems in \({\mathbb{R}}^{m,n}\) follows.
Furthermore, the algorithms have closed-form solutions for the \(\ell _{1}\) and nuclear regularised objective functions. For the \(\ell _{2}\) and Frobenius regularised objectives, the update rules involve values of the principal branch of the Lambert function, which can be well approximated. We propose a sorting based procedure projecting the solution to the decision set for the \(\ell _{1}\) or nuclear ball constrained problems. Finally, the proposed online algorithms can be converted into algorithms for stochastic optimisation with the technique introduced in Joulani et al. (2020). We show that the converted algorithms guarantee an optimal accelerated convergence rate for smooth objective functions. The convergence rate depends logarithmically on the dimensionality of the problem, which suggests its advantage compared to the accelerated AdaGrad-Style algorithms (Cutkosky, 2019; Joulani et al., 2020; Levy et al., 2018).
The rest of the paper is organised as follows. Section 2 reviews the existing work. Section 3 introduces the notation and preliminary concepts. Next, we present and analyse our algorithms in Sect. 4. In Sect. 5, we derive efficient implementations for some popular choices of composite objectives, constraints and stochastic optimisation. Section 6 demonstrates the empirical evaluations using both synthetic and real-world data. Finally, we conclude our work in Sect. 7.
2 Related work
Our primary motivation is to solve the optimisation problems with an elastic net regulariser in their objective function, which are highly involved in attacking (Cancela et al., 2021; Carlini & Wagner, 2017; Chen et al., 2018) and explaining (Dhurandhar et al., 2018; Ribeiro et al., 2016) deep neural networks. The proximal gradient method (PGD) (Nesterov, 2003) and its accelerated variants (Beck & Teboulle, 2009) are usually applied to solving the problem. However, these algorithms are not practical since they require prior knowledge about the smoothness of the objective function to ensure their convergence.
The AdaGrad-style algorithms (Alacaoglu et al., 2020; Duchi et al., 2011; Orabona & Pál, 2018; Orabona et al., 2015) have become popular in the machine learning community in recent years. Given the gradient vectors \(g_{1},\ldots , g_{t}\) received at iteration t, the core idea of these algorithms is to set the stepsizes proportional to \(\frac{1}{\sqrt{\sum _{s=1}^{t-1}\Vert g_{s}\Vert _{*} ^{2}}}\) to ensure a regret upper bounded by \({\mathcal{O}}(\sqrt{\sum _{t=1}^{T}\Vert g_{t}\Vert _{*} ^{2}})\) after T iterations. Online learning algorithms with this adaptive regret can be directly applied to the stochastic optimisation problems (Alacaoglu et al., 2020; Li & Orabona, 2019) or can be converted into a stochastic algorithm (Cesa-Bianchi & Gentile, 2008) with a convergence rate \({\mathcal{O}}(\frac{1}{\sqrt{T}})\). This rate can be further improved to \({\mathcal{O}}(\frac{1}{T^{2}})\) for unconstrained problems with smooth loss functions by applying the acceleration techniques (Cutkosky, 2019; Kavis et al., 2019; Levy et al., 2018). These acceleration techniques do not require prior knowledge about the smoothness of the loss function and a guarantee convergence rate of \({\mathcal{O}}(\frac{1}{\sqrt{T}})\) for non-smooth functions. Joulani et al. (2020) has proposed a simple approach to accelerate optimistic online optimisation algorithms with adaptive regret bound.
Given a d-dimensional problem, the algorithms mentioned above have a regret upper bound depending (sub-) linearly on d. We are interested in a logarithmic regret dependence on the dimensionality, which can be attained by the \({\textit{EG}}\) family algorithms (Arora et al., 2012; Kivinen & Warmuth, 1997; Warmuth, 2007) and their adaptive optimistic extension (Steinhardt & Liang, 2014). However, these algorithms work only for decision sets in the form of cross-polytopes and require prior knowledge about the radius of the decision set for general convex optimisation problems. The p-norm algorithm (Gentile, 2003; Kakade et al., 2012) does not have the limitation mentioned above; however, it still requires prior knowledge about the problem to attain optimal performance (Orabona et al., 2015). The HU algorithm (Ghai et al., 2020), which interpolates gradient descent and EG, can theoretically be applied to loss functions with elastic net regularisers and decision sets other than cross-polytopes. However, it is not practical due to the complex projection step.
Following the idea of HU, we propose more practical algorithms interpolating EG and the p-norm algorithm. The core of our algorithm is a symmetric logarithmic function. Orabona (2013) first introduced the idea of composing the single-dimensional symmetric logarithmic function and a norm to generalise EG to the infinite-dimensional space. It has become popular for parameter-free optimisation (Cutkosky & Boahen, 2016, 2017a, b; Kempka et al., 2019) since one can easily construct an adaptive regulariser with this composition (Cutkosky & Boahen, 2017a). In this paper, instead of using the composition, we apply the symmetric logarithmic function directly to each entry of a vector to construct a symmetric entropy-like function that is strongly convex with respect to the \(\ell _{1}\) norm. We analyse MD and FTRL with the entropy-like function in the framework developed in Joulani et al. (2017). The analysis of the spectral analogue of the entropy-like function follows the idea proposed in Ghai et al. (2020).
3 Preliminary
The focus of this paper is OCO with the decision variable taken from a compact convex subset \({\mathcal{K}}\subseteq {\mathbb{X}}\) of finite dimensional vector space equipped with a norm \(\Vert \cdot \Vert\). Given a sequence of vectors \(\{v_{t}\}\), we use the compressed-sum notation \(v_{1:t}= \sum _{s=1}^{t}v_{s}\) for simplicity. We denote by \({\mathbb{X}}_{*}\) the dual space with the dual norm \(\Vert \cdot \Vert _{*}\). The bi-linear map combining vectors in \({\mathbb{X}}_{*}\) and \({\mathbb{X}}\) is denoted by
For \({\mathbb{X}}={\mathbb{R}}^{d}\), we denote by \(\Vert \cdot \Vert _{1}\) the \(\ell _{1}\) norm, the dual norm of which is the maximum norm denoted by \(\Vert \cdot \Vert _{\infty }\). It is well known that the \(\ell _{2}\) norm denoted by \(\Vert \cdot \Vert _{2}\) is self-dual. In case \({\mathbb{X}}\) is the space of the matrices, for simplicity, we also use \(\Vert \cdot \Vert _{1}\), \(\Vert \cdot \Vert _{2}\) and \(\Vert \cdot \Vert _{\infty }\) for the nuclear, Frobenius and spectral norm, respectively.
Let \(\sigma :{\mathbb{R}}^{m,n}\rightarrow {\mathbb{R}}^{\min \{m,n\}}\) be the function mapping a matrix to its singular values. Define
with
Clearly, the singular value decomposition (SVD) of a matrix X can be expressed as
Similarly, we write the eigendecomposition of a symmetric matrix X as
where we denote by \(\lambda :{\mathbb{S}}^{d}\mapsto {\mathbb{R}}^{d}\) the function mapping a symmetric matrix to its spectrum.
Given a convex set \({\mathcal{K}}\subseteq {\mathbb{X}}\) and a convex function \(f:{\mathcal{K}}\rightarrow {\mathbb{R}}\) defined on \({\mathcal{K}}\), we denote by \(\partial f(y)=\{g\in {\mathbb{X}}_{*}|\forall y\in {\mathcal{K}}.f(x)-f(y)\ge \langle g,x-y\rangle \}\) the subgradient of f at y. We refer to \(\triangledown f(y)\) any element in \(\partial f(y)\). A function is \(\eta\)-strongly convex with respect to \(\Vert \cdot \Vert\) over \({\mathcal{K}}\) if
holds for all \(x,y\in {\mathcal{K}}\) and \(\triangledown f(y)\in \partial f(y)\).
4 Algorithms and analysis
In this section, we present and analyse our algorithms, which begins with a short review on EG and the p-norm algorithm for the case \(f_{t}= l_{t}\). The EG algorithm can be considered as an instance of OMD, the update rules of which is given by
where \(g_{t}\in \partial f_{t}(x_{t})\) is the subgradient, and \(\eta > 0\) is the stepsize. Although the algorithm has the expected logarithmic dependence on the dimensionality, its update rule is applicable only to the decision variables on the standard simplex. For the problem with decision variables taken from an \(\ell _{1}\) ball \(\{x|\Vert x\Vert _{1}\le D\}\), one can apply the \({\textit{EG}}^{\pm }\) trick, i.e. use the vector \([\frac{D}{2}g_{t}^{\top },-\frac{D}{2}g_{t}^{\top }]^{\top }\) to update \([x_{t+1,+}^{\top },x_{t+1,-}^{\top }]^{\top }\) at iteration t and choose the decision variable \(x_{t+1,+}-x_{t+1,-}\). However, if the decision set is implicitly given by a regularisation term, the parameter D has to be tuned. Since applying an overestimated D increases regret, while using an underestimated D decreases the freedom of the model, the algorithm is sensitive to tuning. For composite objectives, EG is not practical due to its update rule.
Compared to EG, the p-norm algorithm, the update rule of which is given by
is better applicable for unknown D. To combine the ideas of EG and the p-norm algorithm, we consider the following generalised entropy function
In the next lemma, we show the twice differentiability and strict convexity of \(\phi\), based on which a strongly convex potential function for OMD in a compact decision set can be constructed.
Lemma 1
\(\phi\) is twice continuous differentiable and strictly convex with
-
1.
\(\phi '(x)=\alpha \ln \left( \frac{\vert x |}{\beta }+1\right) {\text{sgn}}(x)\)
-
2.
\(\phi ''(x)=\frac{\alpha }{\vert x |+\beta }\).
Furthermore, the convex conjugate given by \(\phi ^{*}:{\mathbb{R}}\rightarrow {\mathbb{R}}, \theta \mapsto \alpha \beta \exp \frac{\vert \theta |}{\alpha }-\beta \vert \theta |-\alpha \beta\) is also twice continuous differentiable with
-
1.
\(\phi ^{*\prime }(\theta )=\left( \beta \exp \frac{\vert \theta |}{\alpha }-\beta \right) {\text{sgn}}(\theta )\)
-
2.
\(\phi ^{*\prime \prime }(\theta )=\frac{\beta }{\alpha } \exp \frac{\vert \theta |}{\alpha }.\)
Since we can expand the natural logarithm as \(\ln (\frac{\vert x |}{\beta }+1)=\frac{\vert x |}{\beta }-\frac{\vert x |^{2}}{2\beta ^{2}}+\frac{\vert x |^{3}}{3\beta ^{3}}-\cdots\), \(\phi (x)\) can be intuitively considered as an interpolation between the absolute value and square. As observed in Fig. 1a, it is closer to the absolute value compared to the hyperbolic entropy introduced in Ghai et al. (2020). Moreover, running OMD with regulariser \(x\mapsto \sum _{i=1}^{d}\phi (x_{i})\) yields an update rule
which sets the signs of coordinates like the p-norm algorithm and updates the scale similarly to EG. As illustrated in Fig. 1b, the mirror map \(\triangledown \phi ^{*}\) is close to the mirror map of EG, while the behavior of HU is more similar to the gradient descent update.
4.1 Algorithms in the Euclidean space
To obtain an adaptive and optimistic algorithm, we define the following time varying function
and apply it to the adaptive optimistic OMD (AO-OMD) given by
for the sequence of subgradients \(\{g_{t}\}\) and hints \(\{h_{t}\}\). In a bounded domain, \(\phi _{t}\) is strongly convex with respect to \(\Vert \cdot \Vert _{1}\), which is shown in the next lemma.
Lemma 2
Let \({\mathcal{K}}\subseteq {\mathbb{R}}^{d}\) be convex and bounded such that \(\Vert x\Vert _{1}\le D\) for all \(x\in {\mathcal{K}}\). Then we have for all \(x,y\in {\mathcal{K}}\)
With the property of the strong convexity, the regret of AO-OMD with regulariser (2) can be analysed in the framework of optimistic algorithm (Joulani et al., 2017) and is upper bounded by the following theorem.
Theorem 1
Let \({\mathcal{K}}\subseteq {\mathbb{R}}^{d}\) be a compact convex set. Assume that there is some \(D>0\) such that \(\Vert x\Vert _{1}\le D\) holds for all \(x\in {\mathcal{K}}\). Let \(\{x_{t}\}\) be the sequence generated by update rule (3) with regulariser (2). Setting \(\beta =\frac{1}{d}\), \(\eta =\sqrt{\frac{1}{\ln (D+1)+\ln d}}\), and \(\alpha _{t}=\eta \sqrt{\sum _{s=1}^{t-1}\Vert g_{s}-h_{s}\Vert ^{2}_{\infty }}\), we obtain
for some \(c(d,D)\in {\mathcal{O}}(D\sqrt{\ln (D+1)+\ln d})\).
EG can also be considered as an instance of FTRL with a constant stepsize. The update rule of the adaptive optimistic FTRL (AO-FTRL) is given by
The regret of AO-FTRL is upper bounded by the following theorem.
Theorem 2
Let \({\mathcal{K}}\subseteq {\mathbb{R}}^{d}\) be a compact convex set with \(d>e\). Assume that there is some \(D\ge 1\) such that \(\Vert x\Vert _{1}\le D\) holds for all \(x\in {\mathcal{K}}\subseteq {\mathbb{R}}^{d}\). Let \(\{x_{t}\}\) be the sequence generated by updating rule (4) with regulariser (2) at iteration t. Setting \(\beta =\frac{1}{d}\), \(\eta =\sqrt{\frac{1}{\ln (D+1)+\ln d}}\) and \(\alpha _{t}=\eta \sqrt{\sum _{s=1}^{t-1}\Vert g_{s}-h_{s}\Vert _{\infty }^{2}}\), we obtain
for some \(c(d,D)\in {\mathcal{O}}(D\sqrt{\ln (D+1)+\ln d})\).
4.2 Spectral algorithms
We now consider the setting in which the decision variables are matrices taken from a compact convex set \({\mathcal{K}}\subseteq {\mathbb{R}}^{m,n}\). A direct attempt to solve this problem is to apply the updating rule (3) or (4) to the vectorised matrices. A regret bound of \({\mathcal{O}}(D\sqrt{T\ln (mn)})\) can be guaranteed if the \(\ell _{1}\) norm of the vectorised matrices from \({\mathcal{K}}\) are bounded by D, which is not optimal. In many applications, elements in \({\mathcal{K}}\) are assumed to have bounded nuclear norm, for which the regulariser
can be applied. The next theorem gives the strong convexity of \(\Phi _{t}\) with respect to \(\Vert \cdot \Vert _{1}\) over \({\mathcal{K}}\), which allows us to use \(\{\Phi _{t}\}\) as the potential functions in OMD and FTRL.
Theorem 3
Let \(\sigma :{\mathbb{R}}^{m,n}\rightarrow {\mathbb{R}}^{d}\) be the function mapping a matrix to its singular values. Then the function \(\Phi _{t}=\phi _{t}\circ \sigma\) is \(\frac{\alpha _{t}}{2(D+\min \{m,n\}\beta )}\)-strongly convex with respect to the nuclear norm over the nuclear ball with radius D.
The proof of Theorem 3 follows the idea introduced in Ghai et al. (2020). Define the operator
The set \({\mathcal{X}}=\{S(X)|\in {\mathbb{R}}^{m,n}\}\) is a finite dimensional linear subspace of the space of symmetric matrices \({\mathbb{S}}^{m+n}\). Its dual space \({\mathcal{X}}_{*}\) determined by the Frobenius inner product can be represented by \({\mathcal{X}}\) itself. For any \(S(X)\in {\mathcal{X}}\), the set of eigenvalues of S(X) consists of the singular values and the negative singular values of X. Since \(\phi\) is even, we have \(\sum _{i=1}^{d}\phi (\sigma _{i}(X))=\sum _{i=1}^{d}\phi (\lambda _{i}(X))\) for symmetric X. The next lemma shows that both \(\Phi _{t}|_{\mathcal{X}}\) and \(\Phi ^{*}_{t}|_{\mathcal{X}}\) are twice differentiable.
Lemma 3
Let \(f:{\mathbb{R}}\rightarrow {\mathbb{R}}\) be twice continuously differentiable. Then the function given by
is twice differentiable. Furthermore, let \(X\in {\mathbb{S}}^{d}\) be a symmetric matrix with eigenvalue decomposition
Define the matrix of the divided difference \(\Gamma (f,X)=[\gamma (f,X)_{ij}]\) with
Then for any \(G,H\in {\mathbb{S}}^{d}\), we have
where \(\tilde{g}_{ij}\) and \(\tilde{h}_{ij}\) are the elements of the i-th row and j-th column of the matrix \(U^{\top } G U\) and \(U^{\top } H U\), respectively.
Lemma 3 implies the unsurprising positive semidefiniteness of \(D^{2}F(X)\) for convex f. Furthermore, the exact expression of the second differential allows us to show the local smoothness of \(\Phi ^{*}_{t}\) using the local smoothness of \(\phi ^{*}\). Together with Lemma 4, the locally strong convexity of \(\Phi _{t}|_{\mathcal{X}}\) can be proved.
Lemma 4
Let \(\Phi :{\mathbb{X}}\rightarrow {\mathbb{R}}\) be a closed convex function such that \(\Phi ^{*}\) is twice differentiable at some \(\theta \in {\mathbb{X}}_{*}\) with positive definite \(D^{2}\Phi ^{*}(\theta )\in {\mathcal{L}}({\mathbb{X}}_{*},{\mathcal{L}}({\mathbb{X}}_{*},{\mathbb{R}}))\). Suppose that \(D^{2}\Phi ^{*}(\theta )(v,v)\le \Vert v\Vert _{*} ^{2}\) holds for all \(v\in {\mathbb{X}}_{*}\). Then we have \(D^{2}\Phi (D\Phi ^{*}(\theta ))(x,x)\ge \Vert x\Vert ^{2}\) for all \(x\in {\mathbb{X}}\).
Lemma 4 can be considered as a generalised version of the local duality of smoothness and convexity proved in Ghai et al. (2020). The required positive definiteness of \(D^{2}\Phi ^{*}_{t}(\theta )\) is guaranteed by the exact expression of the second differential described in Lemma 3 and the fact \(\phi ^{*\prime \prime }(\theta )> 0\) for all \(\theta \in {\mathbb{R}}\). Finally, using the construction of \({\mathcal{X}}\), the locally strong convexity of \(\Phi _{t}|_{\mathcal{X}}\) can be extended to \(\Phi _{t}\). The complete proofs of Theorem 3 and the technical lemmata can be found in “Appendix 2.1”.
With the property of the strong convexity, the regret of applying (5) to AO-OMD and AO-FTRL can be upper bounded by the following theorems.
Theorem 4
Let \({\mathcal{K}}\subseteq {\mathbb{R}}^{m,n}\) be a compact convex set. Assume that there is some \(D>0\) such that \(\Vert x\Vert _{1}\le D\) holds for all \(x\in {\mathcal{K}}\). Let \(\{x_{t}\}\) be the sequence generated by update rule (3) with regulariser (5) at iteration t. Setting \(\beta =\frac{1}{\min \{m,n\}}\), \(\eta =\sqrt{\frac{1}{\ln (D+1)+\ln \min \{m,n\}}}\), and \(\alpha _{t}=\eta \sqrt{\sum _{s=1}^{t-1}\Vert g_{s}-h_{s}\Vert ^{2}_{\infty }}\), we obtain
with \(c(m,n,D)\in {\mathcal{O}}(D\sqrt{\ln (D+1)+\ln \min \{m,n\}})\).
Theorem 5
Let \({\mathcal{K}}\subseteq {\mathbb{R}}^{\min \{m,n\}}\) be a compact convex set with \(\min \{m,n\}>e\). Assume that there is some \(D\ge 1\) such that \(\Vert x\Vert _{1}\le D\) holds for all \(x\in {\mathcal{K}}\). Let \(\{x_{t}\}\) be the sequence generated by updating rule (4) with time varying regulariser (5). Setting \(\beta =\frac{1}{\min \{m,n\}}\), \(\eta =\sqrt{\frac{1}{\ln (D+1)+\ln \min \{m,n\}}}\) and \(\alpha _{t}=\eta \sqrt{\sum _{s=1}^{t-1}\Vert g_{s}-h_{s}\Vert _{\infty }^{2}}\), we obtain
with \(c(m,n,D)\in {\mathcal{O}}(D\sqrt{\ln (D+1)+\ln \min \{m,n\}})\).
With regulariser (5), both AO-OMD and AO-FTRL guarantee a regret upper bound proportional to \(\sqrt{\ln \min \{m,n\}}\), which is the best known dependence on the size of the matrices.
5 Derived algorithms
Given \(z_{t+1}\in {\mathbb{X}}_{*}\) and a time varying closed convex function \(R_{t+1}:{\mathcal{K}}\rightarrow {\mathbb{R}}\), we consider the following updating rule
It is easy to verify that (6) is equivalent to
Setting \(z_{t+1}=\triangledown \phi _{t+1}(x_{t})-g_{t}+h_{t}-h_{t+1}\) and \(R_{t+1}=r_{t+1}\), we obtain the AO-OMD update
Setting \(z_{t+1}=-\triangledown \phi _{t+1}(x_{1})+g_{1:t}+h_{t+1}\) and \(R_{t+1}=r_{1:t+1}\), we obtain the AO-FTRL update
The rest of this section focuses on solving the second line of (6) for some popular choices of \(r\) and \({\mathcal{K}}\).
5.1 Elastic net regularisation
We first consider the setting of \({\mathcal{K}}={\mathbb{R}}^{d}\) and \(R_{t+1}(x)=\gamma _{1} \Vert x\Vert _{1}+\frac{\gamma _{2}}{2}\Vert x\Vert ^{2}_{2}\), which has countless applications in machine learning. It is easy to verify that the Bregman divergence associated with \(\psi _{t+1}\) is given by
The minimiser of
in \({\mathbb{R}}^{d}\) can be simply obtained by setting the subgradient to 0. For \(\ln (\frac{\vert y_{i,t+1} |}{\beta }+1)\le \frac{\gamma _{1}}{\alpha _{t+1}}\), we set \(x_{i,t+1}=0\). Otherwise, the 0 subgradient implies \({\text{sgn}}(x_{i,t+1})={\text{sgn}}(y_{i,t+1})\) and \(\vert x_{i,t+1} |\) given by the root of
for \(i=1,\ldots , d\). For simplicity, we set \(a=\beta\), \(b=\frac{\gamma _{2}}{\alpha _{t+1}}\) and \(c=\frac{\gamma _{1}}{\alpha _{t+1}}-\ln (\frac{\vert y_{i,t+1} |}{\beta }+1)\). It can be verified that \(\vert x_{i,t+1} |\) is given by
where \(W_{0}\) is the principal branch of the Lambert function and can be well approximated. For \(\gamma _{2}=0\), i.e. the \(\ell _{1}\) regularised problem, \(\vert x_{i,t+1} |\) has the closed form solution
The implementation is described in Algorithm 1.
5.2 Nuclear and Frobenius regularisation
Similarly, we consider \({\mathcal{K}}={\mathbb{R}}^{m,n}\) with a regulariser \(R_{t+1}(x)=\gamma _{1} \Vert x\Vert _{1}+\frac{\gamma _{2}}{2}\Vert x\Vert ^{2}_{2}\) mixed with the nuclear and Frobenius norm. The second line of update rule (6) can be implemented as follows
Let \(y_{t+1}\) and \(\tilde{y}_{t+1}\) be as defined in (9). It is easy to verify
From the characterisation of subgradient, it follows
and
where \(x=U{\text{diag}}(\sigma (x))V^{\top }\) is SVD of x. Similar to the case in \({\mathbb{R}}^{d}\), \(\tilde{x}_{t+1}\) is the root of
The subgradient of the objective (10) at \(x_{t+1}=U_{t+1}{\text{diag}}(\tilde{x}_{t+1})V_{t+1}^{\top }\) is clearly 0.
5.3 Projection onto the cross-polytope
Next, we consider the setting where \(r_{t}\) is the zero function and \({\mathcal{K}}\) is the \(\ell _{1}\) ball with radius D. Clearly, we simply set \(x_{t+1}=y_{t+1}\) for \(\Vert y_{t+1}\Vert _{1}\le D\). Otherwise, Algorithm 2 describes a sorting based procedure projecting \(y_{t+1}\) onto the \(\ell _{1}\) ball with time complexity \({\mathcal{O}}(d\log d)\). The correctness of the algorithm is shown in the next lemma.
Lemma 5
Let \(y\in {\mathbb{R}}^{d}\) with \(\Vert y\Vert _{1}>D\) and \(x^{*}\) as returned by Algorithm 2, then we have
For the case that \({\mathcal{K}}\subseteq {\mathbb{R}}^{m,n}\) is the nuclear ball with radius D and \(\Vert y_{t+1}\Vert _{1}> D\), we need to solve the problem
where the constant part of the Bregman divergence is removed. From the von Neumann’s trace inequality, the Frobenius inner product is upper bounded by
The equality holds when x and \(U_{t+1}\triangledown \phi _{t+1}(\tilde{y}_{t+1})V_{t+1}^{\top }\) share a simultaneous SVD, i.e. the minimiser has an SVD of the form
Thus the problem is reduced to
which can be solved by Algorithm 2. Thus, the projection of update rule (6) can be implemented as follows
5.4 Stochastic acceleration
Finally, we consider the stochastic optimisation problem of the form
where \(l:{\mathbb{X}}\rightarrow {\mathbb{R}}\) and \(r:{\mathcal{K}}\rightarrow {\mathbb{R}}_{\ge 0}\) are closed convex functions. In the stochastic setting, instead of having a direct access to \(\triangledown l\), we query a stochastic gradient \(g_{t}\) of l at \(z_{t}\) in each iteration t with \({\mathbb{E}}[g_{t}|z_{t}]\in \partial l(z_{t})\). Algorithms with a regret bound of the form \({\mathcal{O}}(\sqrt{\sum _{t=1}^{T}\Vert g_{t}-h_{t}\Vert _{*} ^{2}})\) can be easily converted into a stochastic optimisation algorithm by applying the update rule to the scaled stochastic gradient \(a_{t}g_{t}\) and hint \(a_{t+1}g_{t}\), which is described in Algorithm 3. Joulani et al. (2020) has shown the convergence of accelerating Adagrad for the problem in \({\mathbb{R}}^{d}\). We extend the result to any finite dimensional normed vector space in the following corollary.
Corollary 1
Let \(({\mathbb{X}},\Vert \cdot \Vert )\) be a finite dimensional normed vector space and \({\mathcal{K}}\subseteq {\mathbb{X}}\) a compact convex set. Denote by \(\mathcal{A}\) be some optimistic algorithm generating \(x_{t}\in {\mathcal{K}}\) at iteration t. Denote by
the variance. If \(\mathcal{A}\) has a regret upper bound in the form of
then there is some \(L>0\) such that the error incurred by Algorithm 3 is upper bounded by
Furthermore, if l is M-smooth, then we have
Setting \(\alpha _{t}=t\), we obtain a convergence of \({\mathcal{O}}(\frac{c_{2}}{\sqrt{T}})\) in general case, and \({\mathcal{O}}(\frac{c_{2}}{T^{2}}+\frac{c_{2}\max _{t}\nu _{t}}{\sqrt{T}})\) for smooth loss function. Applying update rule (3) or (4) with regulariser (2) or (5) to Algorithm 3, the constant \(c_{2}\) is proportional to \(\sqrt{\ln d}\) and \(\sqrt{\ln (\min \{m,n\})}\) for \({\mathbb{X}}={\mathbb{R}}^{d}\) and \({\mathbb{X}}={\mathbb{R}}^{m,n}\) respectively, while the accelerated AdaGrad has a linear dependence on the dimensionality (Joulani et al., 2020).
6 Experiments
This section shows the empirical evaluation of the developed algorithms. We carry out experiments on both synthetic and real-world data and demonstrate the performances of the OMD (Exp-MD) and FTRL (Exp-FTRL) based on the exponentiated update.
6.1 Online logistic regression
For a sanity check, we simulate an d-dimensional online logistic regression problem, in which the model parameter \(w^{*}\) has a \(99\%\) sparsity and the non-zero values are randomly drawn from the uniform distribution over \([-1,1]\). At each iteration t, we sample a random feature vector \(x_{t}\) from a uniform distribution over \([-1,1]^{d}\) and generate a label \(y_{t}\in \{-1,1\}\) using a logit model, i.e. \({\text{Pr}}[y_{t}=1]=(1+\exp (-w^{\top } x_{t}))^{-1}\). The goal is to minimise the cumulative regret
with \(l_{t}(w)=\ln (1+\exp (-y_{t}w^{\top } x_{t}))\). We choose \(d=10\),000 and compare our algorithms with AdaGrad, AdaFTRL (Duchi et al., 2011) and HU (Ghai et al., 2020). For both AdaGrad and AdaFTRL, we set the i-th entry of the proximal matrix \(H_{t}\) to \(h_{ii}=10^{-6}+\sum _{s=1}^{t-1}g_{s,i}^{2}\) as their theory suggested (Duchi et al., 2011). The stepsize of HU is set to \(\sqrt{\frac{1}{\sum _{s=1}^{t-1}\Vert g_{s}\Vert _{\infty }^{2}}}\) leading to an adaptive regret upper bound. All algorithms take decision variables from an \(\ell _{1}\) ball \(\{w\in {\mathbb{R}}^{d}|\Vert w\Vert _{1}\le D\}\), which is the ideal case for HU. We examine the performances of the algorithms with known, underestimated and overestimated \(\Vert w^{*}\Vert _{1}\) by setting \(D=\Vert w^{*}\Vert _{1}\), \(D=\frac{1}{2}\Vert w^{*}\Vert _{1}\) and \(D=2\Vert w^{*}\Vert _{1}\), respectively. For each choice of D, we simulate the online process of each algorithm for 10,000 iterations and repeat the experiments for 20 trials.
Figure 2 plots the curves of the average cumulative regret with the ranges of standard deviation as shaded regions. As can be observed, our algorithms have a clear and stable advantage over the AdaGrad-style algorithms and slightly outperform HU in the experiments with known \(\Vert w^{*}\Vert _{1}\). As the combination of the entropy-like regulariser and FTRL can also be used for parameter-free optimisation (Cutkosky & Boahen, 2017a), overestimating \(\Vert w^{*}\Vert _{1}\) does not have a tangible impact on the performance of Exp-FTRL, which leads to its clear advantage over the rest.
6.2 Online multitask learning
Next, we examine the performance of the developed spectral algorithms using a simulated online multi-task learning problem (Kakade et al., 2012), in which we need to solve k highly correlated d-dimensional online prediction problems simultaneously. The data are generated as follows. We first randomly draw two orthogonal matrices \(U\in {\text{GL}}(d,{\mathbb{R}})\) and \(V\in {\text{GL}}(k,{\mathbb{R}})\). Then we generate a k-dimensional vector \(\sigma\) with r non-zero values randomly drawn from a uniform distribution over [0, 10] and construct a low rank parameter matrix \(W^{*}=U{\text{diag}}(\sigma )V\). At each iteration t, k feature and label pairs \((x_{t,1},y_{t,1}),\ldots ,(x_{t,k},y_{t,k})\) are generated using k logit models with the i-th parameters taken from the i-th rows of W. The loss function is given by \(l_{t}(W)=\sum _{i=1}^{k}\ln (1+\exp (-y_{t,i}w_{i}^{\top } x_{t,i}))\). We set \(d=100\), \(k=25\) and \(r=5\), take the nuclear ball \(\{W\in {\mathbb{R}}^{d,k}|\Vert W\Vert _{1}\le D\}\) as the decision set and run the experiment as in Sect. 6.1. The average and standard deviation of the results over 20 trials are shown in Fig. 3.
Similar to the online logistic regression, our algorithms have a clear advantage over AdaGrad and AdaFTRL and slightly outperform HU in all settings. While the regret of the AdaGrad-style algorithms spread over a wider range, our algorithms yield relatively stabler results. The superiority of Exp-FTRL for the overestimated \(\Vert W^{*}\Vert _{1}\) can also be observed from Fig. 3c.
6.3 Optimisation for contrastive explanations
Generating the contrastive explanation of a machine learning model (Dhurandhar et al., 2018) is the most motivating application of this paper. Given a sample \(x_{0}\in {\mathcal{X}}\) and machine learning model \(f:{\mathcal{X}}\rightarrow {\mathbb{R}}^{K}\), the contrastive explanation consists of a set of pertinent positive (PP) features and a set of pertinent negative (PN) features, which can be found by solving the following optimisation problem (Dhurandhar et al., 2018)
Let \(\kappa \ge 0\) be a constant and define \(k_{0}=\arg \max _{i}f(x_{0})_{i}\). The loss function for finding PP is given by
which imposes a penalty on the features that do not justify the prediction. PN is the set of features altering the final classification and is modelled by the following loss function
In the experiment, we first train a ResNet20 model (He et al., 2016) on the CIFAR-10 dataset (Krizhevsky, 2009), which attains a test accuracy of \(91.49\%\). For each class of the images, we randomly pick 100 correctly classified images from the test dataset and generate PP and PN for them. For PP, we take the set of all feasible images as the decision set, while for PN, we take the set of tensors x, such that \(x_{0}+x\) is a feasible image.
We first consider the white-box setting, in which we have the access to \(\triangledown l_{x_{0}}\). Our goal is to demonstrate the performance of the accelerated AO-OMD and AO-FTRL based on the exponentiated update (AccAOExpMD and AccAOExpFTRL). In Dhurandhar et al. (2018), the fast iterative shrinkage-thresholding algorithm (FISTA) (Beck & Teboulle, 2009) is applied to finding the PP and PN. Therefore, we take FISTA as our baseline. In addition, our algorithms are also compared with the accelerated AO-OMD and AO-FTRL with AdaGrad-style stepsizes (AccAOMD and AccAOFTRL) (Joulani et al., 2020).
We pick \(\lambda _{1}=\lambda _{2}=\frac{1}{2}\), which is the largest value from the set \(\{2^{-i}|i\in {\mathbb{N}}\}\) allowing FISTA to attain a negative loss \(l_{x_{0}}\) for 10 randomly selected images. All algorithms start from \(x_{1}=0\). Figure 4 plots the convergence behaviour of the five algorithms, averaged over the 1000 images. In the experiment for PP, our algorithms are obviously better than the AdaGrad-style algorithms. Although FISTA converges faster at the first 100 iterations, it does not make further progress afterwards due to the tiny stepsize found by the backtracking rule. In the experiment for PN, all algorithms behave similarly. It is worth pointing out that the backtracking rule of FISTA requires multiple function evaluations, which are expensive for explaining deep neural networks.
Next, we consider the black-box setting, in which the gradient is estimated through the two-points estimation
where \(\delta\), \(\mu\) are constants and \(v_{i}\) is a random vector. Following Chen et al. (2019), we set \(\delta =d\) and sample \(v_{i}\) independently from the uniform distribution over the unit sphere for AdaGrad-style algorithms. Since the convergence of our algorithms depends on the variance of the gradient estimation in \(({\mathbb{R}}^{d}, \Vert \cdot \Vert _{\infty })\), we set \(\delta =1\) and sample \(\nu _{i,1},\ldots ,\nu _{i,d}\) independently from Rademacher distribution according to Corollary 3 in Duchi et al. (2015). To ensure a small bias of the gradient estimation, we set \(\mu =\frac{1}{\sqrt{dT}}\), which is the recommended value for non-convex and constrained optimisation in Chen et al. (2019). The performances of the algorithms are examined in the high and low variance settings with \(b=1\) and \(b=\sqrt{T}\), respectively. Since the problem is stochastic, FISTA, which searches for the stepsize at each iteration, is not practical. Thus, we remove it from the comparison.
Figure 5 plots the convergence behaviour of the algorithms in the high variance setting. Our algorithms outperform the AdaGrad-style algorithms for generating both PP and PN. Furthermore, the FTRL based algorithms have higher convergence rates than the MD based ones at the first few iterations, leading to overall better performance. The experimental results of the low variance setting are plotted in Fig. 6. Though AccAOExpFTRL yields the smallest objective value at the beginning of the experiments, it gets stuck in the local minimum around 0 and is outperformed by AccAOExpMD and AccAOFTRL at the later iterations. Overall, the algorithms based on the exponentiated update have an advantage over the AdaGrad-style algorithms for both high and low variance settings.
7 Conclusion
This paper proposes and analyses a family of online optimisation algorithms based on an entropy-like regulariser combined with the ideas of optimism and adaptivity. The proposed algorithms have adaptive regret bounds depending logarithmically on the dimensionality of the problem, can handle popular composite objectives and can be easily converted into stochastic optimisation algorithms with optimal accelerated convergence rates for smooth function. As a future research direction, we plan to analyse the convergence of the proposed algorithms together with variance reduction techniques for non-convex stochastic optimisation and analyse their empirical performance for training deep neural networks.
Availability of data and materials
The source code generating synthetic data, creating neural networks and model training are available on GitHub https://github.com/mrdexteritas/exp_grad. The CIFAR-10 data are collected from https://www.cs.toronto.edu/~kriz/cifar.html.
Code availability
The implementation of the experiments and all algorithms involved in the experiments are available on GitHub https://github.com/mrdexteritas/exp_grad.
Change history
19 November 2022
Missing Open Access funding information has been added in the Funding Note.
References
Alacaoglu, A., Malitsky, Y., Mertikopoulos, P., & Cevher, V. (2020). A new regret analysis for Adam-type algorithms. In International conference on machine learning (pp. 202–210).
Allen-Zhu, Z., & Orecchia, L. (2017). Linear coupling: An ultimate unification of gradient and mirror descent. In 8th Innovations in theoretical computer science conference (ITCS 2017).
Anava, O., Hazan, E., Mannor, S., & Shamir, O. (2013). Online learning for time series prediction. In Conference on learning theory (pp. 172–184).
Arora, S., Hazan, E., & Kale, S. (2012). The multiplicative weights update method: A meta-algorithm and applications. Theory of Computing, 8(1), 121–164.
Barbu, V., & Precupanu, T. (2012). Convexity and optimization in banach spaces. Berlin: Springer.
Beck, A., & Teboulle, M. (2009). A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM Journal on Imaging Sciences, 2(1), 183–202.
Bhatia, R. (2013). Matrix analysis (Vol. 169). Berlin: Springer.
Cancela, B., Bolón-Canedo, V., & Alonso-Betanzos, A. (2021). A delayed elastic-net approach for performing adversarial attacks. In 2020 25th International conference on pattern recognition (ICPR) (pp. 378–384). https://doi.org/10.1109/ICPR48806.2021.9413170.
Carlini, N., & Wagner, D. (2017). Towards evaluating the robustness of neural networks. In 2017 IEEE symposium on security and privacy (SP) (pp. 39–57).
Cesa-Bianchi, N., Conconi, A., & Gentile, C. (2004). On the generalization ability of on-line learning algorithms. IEEE Transactions on Information Theory, 50(9), 2050–2057.
Cesa-Bianchi, N., & Gentile, C. (2008). Improved risk tail bounds for on-line algorithms. IEEE Transactions on Information Theory, 54(1), 386–390.
Chen, P.-Y., Sharma, Y., Zhang, H., Yi, J., & Hsieh, C.-J. (2018). Ead: Elastic-net attacks to deep neural networks via adversarial examples. In Thirty-second AAAI conference on artificial intelligence.
Chen, X., Liu, S., Xu, K., Li, X., Lin, X., Hong, M., & Cox, D. (2019). Zo-adamm: Zeroth-order adaptive momentum method for black-box optimization. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d’Alché-Buc, E. Fox, & R. Garnett (Eds.), Advances in neural information processing systems (Vol. 32). Berlin: Curran Associates, Inc.
Cutkosky, A. (2019). Anytime online-to-batch, optimism and acceleration. In International conference on machine learning (pp. 1446–1454).
Cutkosky, A., & Boahen, K. (2017a). Online learning without prior information. In Conference on learning theory (pp. 643–677).
Cutkosky, A., & Boahen, K. A. (2016). Online convex optimization with unconstrained domains and losses. In D. Lee, M. Sugiyama, U. Luxburg, I. Guyon, & R. Garnett (Eds.), Advances in neural information processing systems (Vol. 29). Berlin: Curran Associates, Inc.
Cutkosky, A., & Boahen, K. A. (2017b). Stochastic and adversarial online learning without hyperparameters. In I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, & R. Garnett (Eds.), Advances in neural information processing systems (Vol. 30). Berlin: Curran Associates, Inc.
Dhurandhar, A., Chen, P.-Y., Luss, R., Tu, C.-C., Ting, P., Shanmugam, K., & Das, P. (2018). Explanations based on the missing: Towards contrastive explanations with pertinent negatives. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, & R. Garnett (Eds.), Advances in neural information processing systems. (Vol. 31). Berlin: Curran Associates Inc.
Duchi, J., Hazan, E., & Singer, Y. (2011). Adaptive subgradient methods for online learning and stochastic optimization. Journal of Machine Learning Research, 12, 2121–2159.
Duchi, J. C., Jordan, M. I., Wainwright, M. J., & Wibisono, A. (2015). Optimal rates for zero-order convex optimization: The power of two function evaluations. IEEE Transactions on Information Theory, 61(5), 2788–2806.
Duchi, J. C., Shalev-Shwartz, S., Singer, Y., & Tewari, A. (2010). Composite objective mirror descent. In A. T. Kalai & M. Mohri (Eds.), COLT 2010—The 23rd conference on learning theory, Haifa, Israel, June 27–29, 2010 (pp. 14–26). Omnipress.
Gentile, C. (2003). The robustness of the p-norm algorithms. Machine Learning, 53(3), 265–299.
Ghai, U., Hazan, E., & Singer, Y. (2020). Exponentiated gradient meets gradient descent. In Algorithmic learning theory (pp. 386–407).
He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In 2016 IEEE conference on computer vision and pattern recognition (CVPR) (pp. 770-778). https://doi.org/10.1109/CVPR.2016.90.
Joulani, P., György, A., & Szepesvári, C. (2017). A modular analysis of adaptive (non-) convex optimization: Optimism, composite objectives, and variational bounds. Journal of Machine Learning Research, 1, 40.
Joulani, P., Raj, A., Gyorgy, A., & Szepesvári, C. (2020). A simpler approach to accelerated optimization: Iterative averaging meets optimism. In International conference on machine learning (pp. 4984–4993).
Kakade, S. M., Shalev-Shwartz, S., & Tewari, A. (2012). Regularization techniques for learning with matrices. The Journal of Machine Learning Research, 13(1), 1865–1890.
Kavis, A., Levy, K. Y., Bach, F., & Cevher, V. (2019). Unixgrad: A universal, adaptive algorithm with optimal guarantees for constrained optimization. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d’Alché-Buc, E. Fox, & R. Garnett (Eds.), Advances in neural information processing systems (pp. 6260–6269). Berlin: Curran Associates Inc.
Kempka, M., Kotlowski, W., & Warmuth, M. K. (2019). Adaptive scale-invariant online algorithms for learning linear models. In International conference on machine learning (pp. 3321–3330).
Kivinen, J., & Warmuth, M. K. (1997). Exponentiated gradient versus gradient descent for linear predictors. Information and Computation, 132(1), 1–63.
Krizhevsky, A. (2009). Learning multiple layers of features from tiny images. Master’s thesis, University of Tront.
Lan, G. (2020). First-order and stochastic optimization methods for machine learning. Berlin: Springer.
Levy, Y. K., Yurtsever, A., & Cevher, V. (2018). Online adaptive methods, universality and acceleration. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, & R. Garnett (Eds.), Advances in neural information processing systems (pp. 6500–6509). Berlin: Curran Associates Inc.
Lewis, A. S. (1995). The convex analysis of unitarily invariant matrix functions. Journal of Convex Analysis, 2(1), 173–183.
Li, X., & Orabona, F. (2019). On the convergence of stochastic gradient descent with adaptive stepsizes. In The 22nd international conference on artificial intelligence and statistics (pp. 983–992).
Lu, C., Lin, Z., & Yan, S. (2014). Smoothed low rank and sparse matrix recovery by iteratively reweighted least squares minimization. IEEE Transactions on Image Processing, 24(2), 646–654.
McMahan, H. B., & Streeter, M. J. (2010). Adaptive bound optimization for online convex optimization. In A. T. Kalai & M. Mohri (Eds.), COLT 2010—The 23rd conference on learning theory, Haifa, Israel, June 27–29, 2010 (pp. 244–256). Omnipress.
Nesterov, Y. (2003). Introductory lectures on convex optimization: A basic course (Vol. 87). Berlin: Springer.
Orabona, F. (2013). Dimension-free exponentiated gradient. In NIPS (pp. 1806–1814).
Orabona, F., Crammer, K., & Cesa-Bianchi, N. (2015). A generalized online mirror descent with applications to classification and regression. Machine Learning, 99(3), 411–435.
Orabona, F., & Pál, D. (2018). Scale-free online learning. Theoretical Computer Science, 716, 50–69.
Ribeiro, M. T., Singh, S., Guestrin, C. (2016). “Why should I trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining (pp. 1135–1144).
Song, L., Tekin, C., & Van Der Schaar, M. (2014). Online learning in large-scale contextual recommender systems. IEEE Transactions on Services Computing, 9(3), 433–445.
Steinhardt, J., & Liang, P. (2014). Adaptivity and optimism: An improved exponentiated gradient algorithm. In International conference on machine learning (pp. 1593–1601).
Warmuth, M. K. (2007). Winnowing subspaces. In Proceedings of the 24th international conference on machine learning (pp. 999–1006).
Xie, C., Bijral, A., & Ferres, J. L. (2018). Nonstop: A nonstationary online prediction method for time series. IEEE Signal Processing Letters, 25(10), 1545–1549.
Funding
Open Access funding enabled and organized by Projekt DEAL. The research leading to these results received funding from the German Federal Ministry for Economic Affairs and Climate Action under Grant Agreement No. 01MK20002C.
Author information
Authors and Affiliations
Contributions
Conceptualization: WS; Methodology: WS; Formal analysis and investigation: WS; Software: WS; Validation: WS, FS; Visualization: WS; Writing - original draft preparation: WS; Writing - review and editing: WS, FS; Funding acquisition: SA; Resources: SA; Supervision: FS, SA.
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest or competing interests.
Ethics approval
Not applicable.
Consent to participate
Not applicable.
Consent for publication
Not applicable.
Additional information
Editors: Krzysztof Dembczynski and Emilie Devijver.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix 1: Missing proofs of section 3.1
1.1 Appendix 1.1: Proof of Lemma 1
Proof
It is straightforward that \(\phi\) is differentiable at \(x\ne 0\) with
For any \(h\in {\mathbb{R}}\), we have
where the first inequality uses the fact \(\ln x\le x-1\). Further more, we have
where the first inequality uses the fact \(\ln x\ge 1-\frac{1}{x}\). Thus, we have
for \(h>0\) and
for \(h<0\), from which it follows \(\lim _{h\rightarrow 0} \frac{\phi (0+h)-\phi (0)}{h}=0\). Similarly, we have for \(x\ne 0\)
Let \(h\ne 0\). Then we have
From the inequalities of the logarithm, it follows
Thus, we obtain \(\phi ''(0)=\frac{\alpha }{\beta }\). By the definition of the convex conjugate we have
which is differentiable. The maximiser y satisfies
Since \(\ln \left( \frac{\vert y |}{\beta }+1\right) \ge 0\) holds, we have \({\text{sgn}}(y)={\text{sgn}}(\theta )\) and
Thus, we obtain the maximiser \(y=\phi ^{*\prime }(\theta )\) by setting
Combining (12) and (13), we obtain
To prove that \(\phi ^{*}\) is twice differentiable, it suffices to show that \(\phi ^{*\prime }\) is differentiable at 0. For any \(h\ne 0\), we have
Applying the inequalities of the logarithm, we obtain
from which it follows \(\phi ^{*}\) is twice differentiable at 0 and
\(\square\)
1.2 Appendix 1.2: Proof of Lemma 2
Proof
Let \(x\in {\mathcal{K}}\) be arbitrary. We have
for all \(v\in {\mathbb{R}}^{d}\), where the first inequality follows from the Cauchy-Schwarz inequality. This leads clearly to the strong convexity for a twice differentiable function. \(\square\)
1.3 Appendix 1.3: Proof of Theorem 1
Proposition 1
Let \({\mathcal{K}}\subseteq {\mathbb{X}}\) be a convex set. Assume that \(r_{t}:{\mathcal{K}}\rightarrow {\mathbb{R}}_{\ge 0}\) is closed convex function defined on \({\mathcal{K}}\) and \(\psi _{t}:{\mathcal{K}}\mapsto {\mathbb{R}}\) is \(\eta _{t}\)-strongly convex w.r.t. \(\Vert \cdot \Vert\) over \({\mathcal{K}}\). Then the sequence \(\{x_{t}\}\) generated by (3) with regulariser \(\{\psi _{t}\}\) guarantees
Proof
From the optimality condition, it follows that for all \(x\in {\mathcal{K}}\)
Then, we have
Adding up from 1 to T , we obtain
\(h_{1}\), \(h_{T+1}\) and \(x_{T+1}\), which are artifacts of the analysis, can be set to 0. Then, we simply obtain
Since \(r_{T+1}\) is not involved in the regret, we assume without loss of generality \(r_{1}=r_{T+1}\). From the \(\eta _{t}\)-strong convexity of \(\phi _{t}\) we have
where the second inequality uses the definition of dual norm, the third inequality follows from the fact \(ab\le \frac{a^{2}}{2}+\frac{b^{2}}{2}\). The claimed the result follows. \(\square\)
Proof of Theorem 1
Proposition 1 can be directly applied, and we obtain
Using Lemma 8, we bound the first term of (14)
Using Lemma 6, the second term of (14) can be bounded as
The third term of (14) is simply 0 since we set \(\alpha _{1}=0\). Setting \(\eta =\sqrt{\frac{1}{\ln (D+1)+\ln d}}\) and combining the inequalities above, we obtain the claimed result. \(\square\)
1.4 Appendix 1.4: Proof of Theorem 2
Proposition 2
Let \({\mathcal{K}}\subseteq {\mathbb{X}}\) be a compact convex set such that \(\Vert x\Vert \le D\) holds for all \(x\in {\mathcal{K}}\), \(r_{t}:{\mathcal{K}}\rightarrow {\mathbb{R}}_{\ge 0}\) and \(\phi _{t}:{\mathcal{K}}\mapsto {\mathbb{R}}\) closed convex function defined on \({\mathcal{K}}\). Assume \(\phi _{t}\) is \(\eta _{t}\)-strongly convex w.r.t. \(\Vert \cdot \Vert\) over \({\mathcal{K}}\) and \(\phi _{t}\le \phi _{t+1}\) for all \(t=1,\ldots ,T\). Then the sequence \(\{x_{t}\}\) generated by (4) with guarantees
Proof of Proposition 2
First, define \(\psi _{t}=r_{1:t}+\phi _{t}\). Then, we have
Setting the artifacts \(h_{T+1}\) to 0, rearranging and adding \(\sum _{t=1}^{T}\langle g_{t},w_{t}\rangle\) to both sides, we obtain
From the definition of \(\psi _{t}\), it follows
where we assumed \(r_{T+1}\equiv 0\), since it is not involved in the regret. Furthermore, we have for \(t\ge 1\)
where the first inequality uses the definition of convex conjugate and the second inequality follows from the fact \(\phi _{t+1}\le \phi _{t}\). Adding up from 1 to T, we obtain
where we use \(r_{T+1}\equiv 0\) and \(h_{T+1}=0\). Combining the inequality above and rearranging, we have
Next, by the definition of the Bregman divergence, we have
Since \(\phi _{t}\) is \(\eta _{t}\) strongly convex, we have
We also have
Putting (17) and (18) together, we have
Combining the inequalities above, we obtain
\(\square\)
Proof of Theorem 2
We take the Bregman divergence \({\mathcal{B}}_{\phi _{t}}(x,x_{1})\) as the regulariser at iteration t. Since \({\mathcal{B}}_{\phi _{t}}(x,x_{1})\) is non-negative, increasing with t and \(\frac{2\alpha _{t}}{D+\beta d}\) strongly-convex w.r.t. \(\Vert \cdot \Vert _{1}\), Proposition 2 can be directly applied, and we get
Setting \(\beta =\frac{1}{d}\) and \(\eta =\frac{1}{\sqrt{\ln (D+1)+\ln d}}\), we have
where the inequality uses the assumptions \(D\ge 1\) and \(d>e\). Adding up from 1 to T, we obtain
The first term can be bounded by Lemma 8
Combining the inequality above, we obtain
with \(c(D,d)\in {\mathcal{O}}(D\sqrt{\ln (D+1)+\ln d})\), which is the claimed result. \(\square\)
Appendix 2: Missing Proofs of section 3.2
1.1 Appendix 2.1: Proof of Theorem 3
The Proof of Theorem 3 is based on the idea of Ghai et al. (2020). We first revise some technical lemmata.
Proof of Lemma 3
Define \(\tilde{F}:{\mathbb{S}}^{d}\rightarrow {\mathbb{S}}^{d}, X\mapsto U{\text{diag}}(f(\lambda _{1}(X)),\ldots ,f(\lambda _{d}(X)))U^{\top }\). Apparently, we have \(F(X)={\text{Tr}}\tilde{F}(X)\). From the Theorem V.3.3 in Bhatia (2013), it follows that \(\tilde{F}\) is differentiable and
Using the linearity of the trace and the chain rule, F is differentiable and the directional derivative at X in H is given by
where \(\tilde{h}_{ii}\) is the i-th element in the diagonal of the matrix \(U^{\top } HU\). Next, define
And we have
Applying Theorem V.3.3 in Bhatia (2013) again, we obtain the differentiability of \(\bar{F}\) and
Note that \(X\mapsto {\text{Tr}}(X(\cdot ))\) is a linear map between finite dimensional spaces. Thus F is twice differentiable. From the linearity of the trace operator and matrix multiplication, it follows that \(D_{H}F(X)\) is differentiable. Applying the chain rule, we obtain
which is the claimed result. \(\square\)
Proof of Lemma 4
Since \(D^{2}\Phi ^{*}(\theta )\in {\mathcal{L}}({\mathbb{X}}_{*},{\mathcal{L}}({\mathbb{X}}_{*},{\mathbb{R}}))\) is positive definite and \({\mathbb{X}}\) is finite dimensional, the map
is invertible. Furthermore, defining \(\psi _{\theta }:{\mathbb{X}}_{*}\rightarrow {\mathbb{R}}, v\mapsto \frac{1}{2} D^{2}\Phi ^{*}(\theta )(v,v)\), we have
Thus, we obtain the convex conjugate \(\psi _{\theta }^{*}\)
by setting \(x=D\psi _{\theta }(v)\). Denote by \(I:{\mathbb{X}}\rightarrow {\mathbb{X}}, x\mapsto x\) the identity function. From \(D\Phi ^{*}=D\Phi ^{-1}\), it follows
for \(\theta =D\Phi (v)\) and all \(x\in {\mathbb{X}}\). Thus, we have \(f_{\theta }^{-1}= D^{2}\Phi (D\Phi ^{*}(\theta ))\) and
Finally, since \(\psi _{\theta }(v)\le \frac{1}{2}\Vert v\Vert _{*} ^{2}\) holds for all \(v\in {\mathbb{X}}_{*}\), we can reverse the order by applying Proposition 2.19 in Barbu and Precupanu (2012) and obtain for all \(x\in {\mathbb{X}}\)
which is the claimed result. \(\square\)
Finally, we can prove Theorem 3.
Proof of Theorem 3
We start the proof by introducing the required definitions. Define the operator
The set \({\mathcal{X}}=\{S(X)|X\in {\mathbb{R}}^{m,n}\}\) is a finite dimensional linear subspace of the space of symmetric matrices \({\mathbb{S}}^{m+n}\), and thus \(({\mathcal{X}},\Vert \cdot \Vert _{1})\) is a finite dimensional Banach space. Its dual space \({\mathcal{X}}_{*}\) determined by the Frobenius inner product can be represented by \({\mathcal{X}}\) itself. Denote by \({\mathbb{B}}(D)=\{X\in {\mathbb{R}}^{m,n}|\Vert X\Vert _{1}\le D\}\) the nuclear ball with radius D. Then the set \({\mathcal{K}}=\{S(X)|X\in {\mathbb{B}}(D)\}\) is a nuclear ball in \({\mathcal{X}}\) with radius 2D, since \(\Vert S(X)\Vert _{1}=2\Vert X\Vert _{1}\) for all \(X\in {\mathbb{R}}^{m,n}\).
Let \(S(X)\in {\mathcal{K}}\) be arbitrary. Denote by \(F_{t}=\Phi _{t}\vert _{{\mathcal{X}}}\) the restriction of \(\Phi _{t}\) to \({\mathcal{X}}\). Next, we show the strong convexity of \(F_{t}\) over \({\mathcal{K}}\). From the conjugacy formula of Theorem 2.4 in Lewis (1995) and Lemma 1, it follows
where the second equality follows from the fact that \(\Phi ^{*}_{t}\) is absolutely symmetric. By Lemmas 1 and 3, \(F_{t}^{*}\) is twice differentiable. Let \(X\in {\mathcal{K}}\) be arbitrary and \(\Theta =DF_{t}(X)\in {\mathcal{X}}_{*}\). For simplicity, we define
Then, for all \(H\in {\mathcal{X}}\),
where \(\Gamma (f_{t}',\Theta )=[\gamma (f_{t}',\Theta )_{ij}]\) is the matrix of the second divided difference with
\(D^{2}F_{t}^{*}(\Theta )\) is clearly positive definite over \({\mathbb{S}}^{m+n}\), since \(\gamma (f_{t}',\Theta )_{ij}>0\) for all i and j. Furthermore, from the mean value theorem and the convexity of \(f_{t}^{\prime \prime }\), there is a \(c_{ij}\in (0,1)\) such that
holds for all \(\lambda _{i}(\Theta )\ne \lambda _{j}(\Theta )\). Thus, we obtain
where the last line uses von Neumann’s trace inequality and the fact that the rank of \(H\in {\mathcal{X}}\) and \(\Theta\) is at most \(2\min \{m,n\}\). Since \(H^{2}\) is positive semi-definite, \(\sigma _{i}(H^{2})=\sigma _{i}(H)^{2}\) holds for all i. Furthermore, \(f_{t}''(x)\ge 0\) holds for all \(x\in {\mathbb{R}}\). Thus, the last line of (19) can be rewritten into
Recall \(\Theta =DF_{t}(S(X))\) for \(S(X)\in {\mathcal{K}}\). Together with Lemma 1, we obtain
By the construction of \({\mathcal{K}}\), it is clear that \(\sum _{i=1}^{2\min \{m,n\}}\vert \lambda _{i}(S(X)) |\le 2D\). Thus, (20) can be simply further upper bounded by
Finally, applying Lemma 4, we obtain
which implies the \(\frac{\alpha _{t}}{4(D+\min \{m,n\}\beta )}\)-strong convexity of \(F_{t}\) over \({\mathcal{K}}\).
Finally, we prove the strongly convexity of \(\Phi _{t}\) over \(B(D)\in {\mathbb{R}}^{m+n}\). Let \(X,Y\in B(D)\) be arbitrary matrices in the nuclear ball. The following inequality can be obtained
which implies the \(\frac{\alpha _{t}}{2(D+\min \{m,n\}\beta )}\)-strong convexity of \(\Phi _{t}\) as desired. \(\square\)
1.2 Appendix 2.2: Proof of Theorem 4
Proof
The proof is almost identical to the proof of Theorem 1. From the strong convexity of \(\Phi _{t}\) shown in Theorem 3 and the general upper bound in Proposition 1, we obtain
Using Lemma 8, we have
Furthermore, from Lemma 6, it follows
The claimed result is obtained by combining the inequalities above. \(\square\)
1.3 Appendix 2.3: Proof of Theorem 5
Proof
Since \({\mathcal{B}}_{\Phi _{t}}(x,x_{1})\) is non-negative, increasing and \(\frac{2\alpha _{t}}{D+\beta d}\) strongly-convex w.r.t. \(\Vert \cdot \Vert _{1}\), Proposition 2 can be directly applied, and we get
Setting \(\beta =\frac{1}{\min \{m,n\}}\) and \(\eta =\frac{1}{\sqrt{\ln (D+1)+\ln \min \{m,n\}}}\), we have
where the inequality uses the assumptions \(D\ge 1\) and \(\min \{m,n\}>e\). Adding up from 1 to T, we obtain
The first term can be bounded by Lemma 8
Combining the inequalities above, we obtain
with \(c(D,m,n)\in {\mathcal{O}}(D\sqrt{\ln (D+1)+\ln \min \{m,n\}})\), which is the claimed result. \(\square\)
Appendix 3: Missing Proofs of section 3.4
1.1 Appendix 3.1: Proof of Lemma 5
Proof of Lemma 5
Let \(x^{*}\) be the minimiser of \({\mathcal{B}}_{\psi _{t+1}}(x,y_{t+1})\) in \({\mathcal{K}}\). Using the the fact \(\ln a\ge 1-\frac{1}{a}\), we obtain
and
Thus, \(y_{i}=0\) implies \(x^{*}_{i}=0\). Furthermore \({\text{sgn}}(x^{*}_{i})={\text{sgn}}(y_{i})\) must hold for all i with \(y_{i}\ne 0\), since otherwise we can always flip the sign of \(x^{*}_{i}\) to obtain smaller objective value. So we assume without loss of generality that \(y_{i}\ge 0\). We claim that \(\sum _{i=1}^{d} x^{*}_{i}=D\) holds for the minimiser \(x^{*}\). If it is not the case, there must be some i with \(x^{*}_{i}< y_{i}\), and increasing \(x^{*}_{i}\) by a small enough amount can decrease the objective function. Thus minimising the Bregman divergence can be rewritten into
Using Lagrange multipliers for \(x\in {\mathbb{R}}^{d}\), \(\lambda \in {\mathbb{R}}\) and \(\nu \in {\mathbb{R}}^{d}_+\)
Setting \(\frac{\partial {\mathcal{L}}}{\partial x_{i}}=0\), we obtain
From the complementary slackness, we have \(\nu _{i}=0\) for \(x_{i}\ne 0\), which implies
where \(z=\exp (\lambda )\). Let \(x^{*}\) be the minimiser and \({\mathcal{I}}=\{i:x^{*}_{i}>0\}\) the support of \(x^{*}\). Then we have
Let p be a permutation of \(\{1,\ldots ,d\}\) such that \(y_{p(i)}\le y_{p(i+1)}\). Define
It follows from
that \(\theta (j)\) is increasing in j. Let \(\rho =\min \{i|\theta (i) >0\}\). For all \(j<\rho\), p(j) is not in the support \({\mathcal{I}}\), since otherwise it would imply \(x^{*}_{p(j)}\le 0\). Thus the minimisation problem (22) is equivalent to
Define function \(R:{\mathbb{R}}_{>0}\rightarrow {\mathbb{R}}, x\mapsto x\ln x\). It can be verified that R is convex. The objective function in (23) can be further rewritten into
where the inequality follows from the Jensen’s inequality. The minimum is attained if and only if \(\frac{x_{p(i)}+\beta }{y_{p(i)}+\beta }\) are equal for all i. This is only possible when \(\sigma (i)\) is in the support \({\mathcal{I}}\) for all \(i\ge \rho\). Thus we can set \(z=\frac{\sum _{i=\rho }^{d}(\vert y_{p(i)} |+\beta )}{D+(d-\rho +1)\beta }\) and obtain \(x^{*}_{i}=\max \{\frac{(\vert y_{i} |+\beta )-\beta }{z},0\}{\text{sgn}}(y_{i})\) for \(i=1\ldots d\), which is the claimed result. \(\square\)
1.2 Appendix 3.2: Proof of Corollary 1
Proposition 3
Let \(\{x_{t}\}\) be any sequences and \(\{y_{t}\}\) be the sequence produced by \(y_{t+1}=\frac{a_{t}}{a_{1:t}}x_{t}+(1-\frac{a_{t}}{a_{1:t}})y_{t}\). Choosing \(a_{t}>0\) , we have, for all \(x\in {\mathcal{W}}\)
with \({\mathcal{R}}_{1:T}=\sum _{t=1}^{T}a_{t}(\langle g_{t},x_{t}-x\rangle +r(x_{t})-r(x))\).
Proof
It is interesting to see that the average scheme can be considered as an instance of the linear coupling introduced in Allen-Zhu and Orecchia (2017). For any sequence \(\{x_{t}\}\), \(\{y_{t}\}\) and \(z_{t}=\frac{a_{t}}{a_{1:t}}x_{t}+(1-\frac{a_{t}}{a_{1:t}})y_{t}\), we start the proof by bounding \(a_{t}(f(y_{t+1})-f(x))\) as follows
Denote by \(\tau _{t}=\frac{a_{t}}{a_{1:t}}\) the weight. The first term of the the inequality above can be further bounded by
Next, we have
Combining (24), (25) and (26), we have
Simply setting \(y_{t+1}:=z_{t}\) makes the first term above 0 and implies \(z_{t}=\frac{\sum _{s=1}^{t}a_{s}x_{s}}{a_{1:t}}\). Furthermore it follows from the convexity of \(r\)
Combining the inequalities above and rearranging, we obtain
Furthermore, we have
Finally, we we obtain
which is the claimed result. \(\square\)
Proof of Corollary 1
First of all, we have
For all t, we have
Since \(z_{t-1}\) is fixed when \(z_{t}\) is given, the first term above can be bounded by
Since \({\mathcal{K}}\) is compact, there is some \(L>0\) such that \(\Vert \triangledown l(z)\Vert _{*} \le L\) for all \(z\in {\mathbb{X}}\). Thus the second term of (28) can be bounded by
Combining (27), (28) and (29), we have
and combining with Proposition 3, we obtain
If l is M-smooth, then for \(t\ge 2\), we have
Using fact \(2ab-a^{2}\le b^{2}\), we have
Combining (27), (28) and (31), we have
which implies
\(\square\)
Appendix 4: Technical lemmata
Lemma 6
For positive values \(a_{1},\ldots ,a_{n}\) the following holds:
-
1.
$$\begin{aligned} \sum _{i=1}^{n}\frac{a_{i}}{\sum _{k=1}^{i}a_{k}+1}\le \log \left( \sum _{i=1}^{n}a_{i}+1\right) \end{aligned}$$
-
2.
$$\begin{aligned} \sqrt{\sum _{i=1}^{n}a_{i}}\le \sum _{i=1}^{n}\frac{a_{i}}{\sqrt{\sum _{j=1}^ia_{j}^{2}}}\le 2\sqrt{\sum _{i=1}^{n}a_{i}}. \end{aligned}$$
Proof
The proof of (1) can be found in Lemma A.2 in Levy et al. (2018) For (2), we define \(A_{0}=1\) and \(A_{i}=\sum _{k=1}^{i}a_{i}+1\) for \(i>0\). Then we have
where the inequality follows from the concavity of \(\log\). \(\square\)
Lemma 7
Let l be convex and M-smooth over \({\mathbb{X}}\), i.e.
Then
holds for all \(x,y\in {\mathbb{X}}\).
Proof
Let \(x,y\in {\mathbb{X}}\) be arbitrary. Define \(h:{\mathbb{X}}\rightarrow {\mathbb{R}}, z\mapsto l(z)-\langle \triangledown l(y),z\rangle\). Clearly, h is M-smooth and minimised at y. Thus we have
where the first inequality uses the M-smoothness of h, and the second uses \(\langle \triangledown h(x),z-x\rangle \ge -\Vert \triangledown h(x)\Vert _{*} \Vert z-x\Vert\), for which we choose z such that the equality holds. This implies
and the desired result follows. \(\square\)
Lemma 8
Define \(\psi :{\mathbb{R}}^{d}\rightarrow {\mathbb{R}}, x\mapsto \sum _{i=1}^{d}\phi (x_{i})\) for \(\phi\) be as defined in (1). Assume \(\Vert x\Vert _{1}\le D\) for all \(x\in {\mathcal{K}}\subseteq {\mathbb{R}}^{d}\). Setting \(\beta =\frac{1}{d}\), we obtain for all \(x,y\in {\mathcal{K}}\)
Similarly, we define \(\Psi :{\mathbb{R}}^{m,n}\rightarrow {\mathbb{R}}, x\mapsto \psi \circ \sigma (x)\). Assume \(\Vert x\Vert _{1}\le D\) for all \(x\in {\mathcal{K}}\subseteq {\mathbb{R}}^{m,n}\). Setting \(\beta =\frac{1}{\min \{m,n\}}\), we obtain for all \(x,y\in {\mathcal{K}}\)
Proof
From the definition of the Bregman divergence it follows for all \(x,y\in {\mathcal{K}}\)
Using the closed form of \(\Vert \triangledown \psi (x)\Vert _{\infty }\), we have for \(x\in {\mathcal{K}}\)
Combining the inequalities above and choosing \(\beta = \frac{1}{d}\), we obtain
Using the same argument, we have for all \(x,y\in {\mathcal{K}}\subseteq {\mathbb{R}}^{m,n}\)
From the characterisation of subgradient, it follows for \(x\in {\mathcal{K}}\)
Combine the inequalities above and choose \(\beta = \frac{1}{\min \{m,n\}}\), we obtain
\(\square\)
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Shao, W., Sivrikaya, F. & Albayrak, S. Optimistic optimisation of composite objective with exponentiated update. Mach Learn 111, 4719–4764 (2022). https://doi.org/10.1007/s10994-022-06229-1
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10994-022-06229-1