1 Introduction and terminology

Let X be a real linear space. A set \(D\subset X\) is said to be convex if

$$\begin{aligned} tx+(1-t)y\in D \quad \text {for every}\ x, y\in D\ \text {and}\ t\in [0,1]. \end{aligned}$$

A point \(x_0\in D\) is said to be algebraically internal for a set \(D\subset X\), and we will write \(x_0\in algint(D)\), if for every \(x\in X\) there exists a number \(\delta >0\) such that

$$\begin{aligned} x_0+t x\in D\quad \text {for}\ t \in (-\delta ,\delta ). \end{aligned}$$

A set D is algebraically open if \(algint(D)=D\).

Let us recall that a function \(f:D\rightarrow \mathbb R\) is said to be convex if

$$\begin{aligned} f(tx+(1-t)y)\le t f(x)+(1-t)f(y),\quad x, y\in D,\ t\in [0,1]. \end{aligned}$$

If the above inequality holds for all \(x, y\in D\) with \(t=\frac{1}{2}\) then f is said to be Jensen-convex.

A relation of majorization was introduced by Schur [20] in 1923 (see also [5]) in the following manner: for \(x, y\in \mathbb R^n\)

$$\begin{aligned} x\prec y \quad \Leftrightarrow \quad \sum _{i=1}^{k}x_{[i]}\le \sum _{i=1}^{k}y_{[i]}\ \ \text {for}\ k=1,\ldots ,n-1 \quad \text {and}\quad \sum _{i=1}^{n}x_{[i]}= \sum _{i=1}^{n}y_{[i]}, \end{aligned}$$

where, for any \(x=(x_{1},\ldots ,x_{n})\in {\mathbb {R}}^n,\) \((x_{[1]},\ldots ,x_{[n]})\) denotes the components of x in decreasing order: \(x_{[1]}\ge \cdots \ge x_{[n]}\). When \(x\prec y\), x is said to be majorized by y. The relation of majorization defined above turns out to be a preordering relation i.e. it is reflexive and transitive. The fact \(x\prec y\) is equivalent (see [2, 5]) to the existence of a doubly stochastic matrix (i.e. a square matrix containing nonnegative elements with all rows and columns summing up to 1) \(T\in \mathbb R^{n}_{n}\) such that

$$\begin{aligned} x=Ty. \end{aligned}$$

Particularly interesting examples of doubly stochastic matrices are provided by the permutation matrices. Recall that, a given matrix is said to be a permutation matrix if each row and column has a single unit entry, and all other entries are zero.

The functions that preserve the order of majorization (in Schur’s honor who first considered them) are said to be convex in the sense of Schur. Thus we say that a function \(f:W\rightarrow \mathbb R\), where \(W\subset \mathbb R^n\) is Schur-convex, if for all \(x, y\in W\) the implication

$$\begin{aligned} x\prec y \Rightarrow f(x)\le f(y) \end{aligned}$$

holds. In the case, where \(W=I^n\) with some interval \(I\subseteq \mathbb R\) the above condition is equivalent to the following one

$$\begin{aligned} f(Tx)\le f(x) \quad \text {for}\ x\in I^n \end{aligned}$$

and for all doubly stochastic matrices \(T\in \mathbb R^{n}_{n}\). A survey of results concerning majorization and Schur convex-functions may be found in an extensive monograph by Arnold et al. [2].

It is known from the classical papers of Schur [20], Hardy–Littlewood–Pólya [5] and Karamata [7] that if a function \(f:I\rightarrow \mathbb R\) is convex then it generates so-called Schur-convex sums, that is the function \(F:I^n\rightarrow \mathbb R\) defined by

$$\begin{aligned} F(x_1,\ldots ,x_n)=f(x_1)+\cdots +f(x_n) \end{aligned}$$

is Schur-convex. It is also known that the convexity of f is a sufficient but not necessary condition under which F is Schur-convex.

In 1954 E. M. Wright [21] introduced a new convexity property. A function \(f:D \rightarrow \mathbb R\) is called Wright-convex if

$$\begin{aligned} f(tx+(1-t)y)+f((1-t)x+ty)\le f(x)+f(y)\quad \text {for} \ x,y\in D,\ t\in [0,1]. \end{aligned}$$

Clearly, each convex and additive function is Wright-convex, and each Wright-convex functions is convex in the sense of Jensen.

In 1987 C. T. Ng [12] gave a full characterization of functions generating Schur-convex sums. The following theorem of Ng shows also the connection between the functions generated by Schur-convex sums and Wright-convex functions.

Theorem 1

[12] Let \(D\subset \mathbb R^m\) be a nonempty open and convex set, \(f:D\rightarrow \mathbb R\) and \(F(x_1,\ldots ,x_n)=\sum _{j=1}^{n}f(x_j)\). The following conditions are equivalent to each other:

  1. (a)

    F is Schur-convex for some \(n\ge 2\),

  2. (b)

    F is Schur-convex for every \(n\ge 2\),

  3. (c)

    f is convex in the sense of Wright,

  4. (d)

    f admits the representation

$$\begin{aligned} f(x)=w(x)+a(x),\ \ x\in D, \end{aligned}$$

where \(w:D\rightarrow \mathbb R\) is a convex function, and \(a:\mathbb R^m\rightarrow \mathbb R\) is an additive function.

Ng proved his decomposition theorem using de Bruijn’s result [3] related to functions having continuous differences. Several other proofs of this theorem are known in the literature, c.f. Kominek [9] used a support type result based on Rode’s theorem [19], whereas Nikodem in [13] applied a local stability theorem to the Jensen equation. All the above mentioned proofs of Ng’s decomposition theorem are based on results that depend on the axiom of choice, i.e. they use transfinite induction. In the paper [18] Páles obtained a new and elementary proof without using transfinite induction.

In this paper we use the version of Ng’s theorem proved by Nikodem and Páles. In [14], among other results, they gave a decomposition type theorem for approximately Wright-convex functions (see [14, Theorem 2, p. 145]). Applying this result for \(\varepsilon =0\) we obtain Ng’s characterization of Wright-convex functions (as a sum of convex and additive functions) defined on a convex subset of a real linear space with a nonempty algebraic interior.

For an overview about the other properties and generalizations of Wright-convexity we refer to the papers [1, 4, 8,9,10,11,12,13,14, 16, 18, 21].

2 A generalized version of Ng’s theorem

During this section D denotes a convex subset of a given real linear space X. Now, we extend Ng’s decomposition theorem for a function \(F:D^n\rightarrow \mathbb R\) of the form:

$$\begin{aligned} F(x_1,\ldots ,x_n)= \sum _{j=1}^{n}f_j(x_j), \end{aligned}$$

where \(f_1,\ldots ,f_n:D\rightarrow \mathbb R\) and \(n\in {\mathbb {N}},\ n\ge 2\). In the proof of a generalized version of Ng’s theorem we use the following result from [15].

Lemma 1

[15, Lemma 1] Let a map \(H:D^n\rightarrow Y\) be of the form

$$\begin{aligned} H(x_1,\ldots ,x_n)=\sum _{j=1}^{n}F_j(x_j), \end{aligned}$$

where \(F_j:D\rightarrow Y\), for \(j=1,\ldots ,n\) and Y stands for an abelian group. Then H is symmetric if and only if there exist a map \(F:D\rightarrow Y\) and constants \(C_1,\ldots ,C_n\in Y\) such that

$$\begin{aligned} F_j(x_j)=F(x_j)+C_j,\quad x_j\in D,\ j=1,\ldots ,n. \end{aligned}$$

The following theorem generalizes Ng’s decomposition theorem:

Theorem 2

Let D be a convex subset of a real linear space X such that \(algint(D)\ne \emptyset \) and let \(f_1,\ldots ,f_n:D\rightarrow \mathbb R\) be given functions. Then the following conditions are pairwise equivalent:

  1. (i)

    a function \(F:D^n\rightarrow \mathbb R\) given by the formula

    $$\begin{aligned} F(x_1,\ldots ,x_n)=\sum _{j=1}^{n}f_j(x_j) \end{aligned}$$
    (1)

    is Schur-convex;

  2. (ii)

    each function \(f_j\) is Wright-convex i.e.

    $$\begin{aligned} f_j(tx+(1-t)y)+f_j((1-t)x+ty)\le f_j(x)+f_j(y),\quad x, y\in D,\ t\in [0,1] \end{aligned}$$

    for \(j=1,\ldots ,n\), moreover, \(f_i-f_j=const\);

  3. (iii)

    there exist a convex function \(w:D\rightarrow \mathbb R\), an additive map \(a:X\rightarrow \mathbb R\) and constants \(c_1,\ldots ,c_n \in \mathbb R\) such that

    $$\begin{aligned} f_j(x)=w(x)+a(x)+c_j,\quad x\in D,\ j=1,\ldots ,n. \end{aligned}$$

Proof

\((i)\Rightarrow (ii)\) Assume that the function F given by the formula (1) is Schur-convex. In particular it is also symmetric. Indeed, because for any permutation matrix P its inverse \(P^{-1}\) is also a permutation matrix, by the Schur-convexity of F we obtain

$$\begin{aligned} F(x)=F(P^{-1}(Px))\le F(Px)\le F(x), \end{aligned}$$

and consequently \(F(Px)=F(x)\). Applying Lemma 1 we infer that

$$\begin{aligned} f_j(x)=f(x)+c_j,\quad j=1,\ldots ,n \end{aligned}$$

for some function \(f:D\rightarrow \mathbb R\) and constants \(c_1,\ldots ,c_n\in \mathbb R\). Clearly,

$$\begin{aligned} f_j(x)-f_i(x)=c_j-c_i=const,\quad x\in D. \end{aligned}$$

Put \(c:=c_1+\cdots +c_n\) and fix \(j\in \{1,\ldots ,n\},\ x, y, z\in D,\ t\in [0,1]\) arbitrarily. Then on account of the Schur-convexity of F and the fact that

$$\begin{aligned} (tx+(1-t)y, (1-t)x+ty,z,\ldots ,z)\prec (x,y,z,\ldots ,z) \end{aligned}$$

we obtain

$$\begin{aligned}{} & {} f_j(tx+(1-t)y)+f_j((1-t)x+ty)=f(tx+(1-t)y)+f((1-t)x+ty)+2c_j \\{} & {} \quad =F(tx+(1-t)y,(1-t)x+ty,z,\ldots ,z)-(n-2)f(z)+2c_j-c \\{} & {} \quad \le F(x,y,z,\ldots ,z)-(n-2)f(z)+2c_j-c \\{} & {} \quad =f(x)+f(y)+2c_j=f_j(x)+f_j(y). \end{aligned}$$

\((ii)\Rightarrow (iii)\) By putting

$$\begin{aligned} \overline{c}_{ij}:=f_i(x)-f_j(x) \end{aligned}$$

and \(f=f_1,\ c_j:=\overline{c}_{1,j},\ j=1,\ldots ,n\) we obtain the representation

$$\begin{aligned} f_j(x)=f(x)+c_j,\quad j=1,\ldots ,n. \end{aligned}$$

Obviously, since \(f=f_j-c_j\), f is convex in the sense of Wright and by Theorem 2 from [14] it has the form

$$\begin{aligned} f(x)=w(x)+a(x),\quad x\in D \end{aligned}$$
(2)

where \(w:D\rightarrow \mathbb R\) is a convex function and \(a:X\rightarrow \mathbb R\) is an additive map.

\((iii)\Rightarrow (i)\) Fix \(x, y\in D^n,\ x\prec y\). There exists a doubly stochastic matrix T such that \(x=Ty\). Since

$$\begin{aligned} x_j=\sum _{i=1}^{n}y_i t_{i j}, \end{aligned}$$

then using the representation (2) we obtain

$$\begin{aligned} F(x)=F(Ty)= & {} F\Big (\sum _{j=1}^{n}t_{1 j}y_j,\ldots ,\sum _{j=1}^{n}t_{n j}y_j\Big ) \\= & {} \sum _{i=1}^{n}w\Big (\sum _{j=1}^{n}t_{i j}y_j\Big )+\sum _{i=1}^{n}a\Big (\sum _{j=1}^{n}t_{i j}y_j\Big )+\sum _{i=1}^{n}c_i \\\le & {} \sum _{i=1}^{n}\sum _{j=1}^{n}t_{i j}w(y_j)+\sum _{i=1}^{n}a\Big (\sum _{j=1}^{n}t_{i j}y_j\Big )+\sum _{i=1}^{n}c_i \\= & {} \sum _{j=1}^{n}w(y_j)+\sum _{j=1}^{n}a(y_j)+\sum _{i=1}^{n}c_i =F(y) \end{aligned}$$

\(\square \)

As an immediate consequence of the implication \((i)\Rightarrow (ii)\) from the above theorem we obtain.

Corollary 1

Under the assumptions of the previous theorem if a function \(F:D^n\rightarrow \mathbb R\) of the form

$$\begin{aligned} F(x_1,\ldots ,x_n)=\sum _{j=1}^{n}f_j(x_j) \end{aligned}$$

is Schur-convex then its diagonalization i.e. a function \(f:D\rightarrow \mathbb R\) of the form

$$\begin{aligned} f(x):=\sum _{j=1}^{n}f_j(x) \end{aligned}$$

is Wright-convex.

3 The problem of Zsolt Páles

In this section we will focus on some problems posed by Zsolt Páles concerning the Wright-convexity. During the whole section \(I\subset \mathbb R\) stands for an interval with nonempty interior and for a number \(h>0\) let

$$\begin{aligned} I_h:=I\cap (I-h). \end{aligned}$$

Clearly, \(I_h\) is an interval (possibly an empty set), moreover, for \(0<h_1<h_2\) we have

$$\begin{aligned} I_{h_2}\subset I_{h_1}. \end{aligned}$$

In the sequel the symbol LBV(I) will stand for the class of all functions \(f:I\rightarrow \mathbb R\) with locally bounded variation on I i.e.

$$\begin{aligned} LBV(I):=\{f:I\rightarrow \mathbb R\ |\ \ V_a^b(f)<\infty \ \ \text {for all}\ a, b\in I,\ a<b\}, \end{aligned}$$

where the total variation \(V_{a}^{b}(f)\) of a function \(f:[a,b]\rightarrow \mathbb R\) over [ab] is defined by the formula

$$\begin{aligned} V_{a}^{b}(f):=\sup \Big \{\sum _{j=1}^{n}|f(x_j)-f(x_{j-1})|:\ (x_0,\dots ,x_n)\in {\mathcal {P}}_{[a,b]}\Big \}, \end{aligned}$$

and \({\mathcal {P}}_{[a,b]}\) stands for the family of all partitions of the interval [ab]. A function f is of bounded total variation on [ab] if \(V_{a}^{b}(f)<\infty .\)

The concept of bounded variation functions was introduced in 1881 by Camille Jordan [6] for real functions defined on a closed interval \([a,b]\subset \mathbb R\) in his study of Fourier series. It is well-known that any function \(f\in LBV(I)\) can be decomposed as the difference \(f=f_1-f_2\), where \(f_1, f_2:I\rightarrow \mathbb R\) are increasing functions. The functions \(f_1, f_2:I\rightarrow \mathbb R\) can be given for example by the formulas:

$$\begin{aligned} f_1(x)=\left\{ \begin{array}{rcl} V^{x}_{a}(f),&{} x>a\\ 0,&{}x=a\\ V_{x}^{a}(f),&{} x<a \end{array} \right. \quad \text {and}\quad f_2(x)=f_1(x)-f(x),\quad x\in I, \end{aligned}$$

where the point \(a\in I\) is fixed. This representation is known as the Jordan decomposition.

It is well-known (see for instant [4]) that the Wright-convexity of a function \(f:I\rightarrow \mathbb R\) defined on a nonempty interval \(I\subset \mathbb R\) is equivalent to the condition, that for an arbitrary \(h>0\) a function

$$\begin{aligned} I_h \ni x \longrightarrow f(x+h)-f(x) \end{aligned}$$

is increasing.

In 2004 during the 42nd International Symposium on Functional Equations Zs. Páles posed the following problem connected with the above property:

(Zs. Páles, [17], Problem 22) Let \(I\subset \mathbb R\) be an open interval. Given functions \(f, g:I\rightarrow \mathbb R\) characterize the situation (in terms of the properties, or decompositions of f and g) when

$$\begin{aligned} I_h\ni x\longrightarrow f(x+h)-g(x) \end{aligned}$$
(3)

is increasing for all \(h>0\).

The following proposition gives a characterization of a pair of functions (fg) satisfying the condition (3) as a solution of some conditional functional inequality.

Proposition 1

Let \(I\subset \mathbb R\) be a nonempty interval and let \(f, g: I\rightarrow \mathbb R\). Then a function

$$\begin{aligned} I_h\ni x\longrightarrow f(x+h)-g(x) \end{aligned}$$

is increasing for all \(h>0\) if and only if the inequality

$$\begin{aligned} f(tx+(1-t)y)+g((1-t)x+ty)\le f(y)+g(x), \end{aligned}$$
(4)

holds for all \(x, y\in I, x<y\).

Proof

First, assume that (3) holds. Pick \(x, y\in I, x<y,\ t\in (0,1)\) arbitrarily. Put \(h:=(1-t)(y-x)>0\). Obviously, \(x<y-h\) and by (3) we get

$$\begin{aligned}{} & {} f(x+(1-t)(y-x))-g(x)\le f(y-(1-t)(y-x)+(1-t)(y-x))\\{} & {} \quad -g(y-(1-t)(y-x)), \end{aligned}$$

or equivalently,

$$\begin{aligned} f(tx+(1-t)y)+g((1-t)x+ty)\le f(y)+g(x). \end{aligned}$$

Conversely, suppose that the pair of functions (fg) satisfies the inequality (4). Fix an \(h>0\) and \(x, y\in I_h,\ x<y\) arbitrarily. Since \(x+h, y \in (x,y+h)\) and \((x+h)+y=x+(y+h)\) there exists a number \(t\in (0,1)\) such that

$$\begin{aligned} x+h=tx+(1-t)(y+h)\quad \text {and}\quad y=(1-t)x+t(y+h). \end{aligned}$$

By (4) we get

$$\begin{aligned} f(x+h)+g(y)\le f(y+h)+g(x), \end{aligned}$$

or equivalently,

$$\begin{aligned} f(x+h)-g(x)\le f(y+h)-g(y). \end{aligned}$$

\(\square \)

The idea of the proof of our main result is inspired by the method used in [10]. The below presented Definition 1 and Lemma 2 are a slight modification of the Definition and Lemma 4 from [10].

Definition 1

Let \(f:I\rightarrow \mathbb R\) be fixed and let \(I_0\subset I\) be a subinterval of positive length. We say that f is decomposable on \(I_0\) if there exist a locally bounded variation function \(f_0:I_0\rightarrow \mathbb R\) and an additive function \(a_0:\mathbb R\rightarrow \mathbb R\) such that \(a_0(\mathbb Q)=\{0\}\) and \(f=f_0+a_0\) on \(I_0\).

The proof of the following, technical lemma is similar to the proof of Lemma 2 from [10], so it will be omitted.

Lemma 2

Let \(f:I\rightarrow \mathbb R\) be a given function. Then

  1. (i)

    If f is decomposable on the subintervals \(I_1, I_2\) of I and the interval \(I_1\cap I_2\) has positive length then f is decomposable on \(I_1\cup I_2\).

  2. (ii)

    If \(\{I_n\}_{n\in \mathbb N}\) is an increasing sequence of subintervals of positive length of I on which f is decomposable then f is also decomposable on \(\bigcup _{k=1}^{\infty }I_k\).

  3. (iii)

    If, for all \(a, b\in I,\ a<b\), f is decomposable on \(\left[ a,\frac{a+b}{2}\right] \) then f is also decomposable on I.

In the proof of our main result we also use the following theorem proved by de Bruijn in [3] (Theorem 6.1, p.213).

Theorem 3

[3] Let f(x) have the period 1, and assume that, for any value of h, the difference \(\Delta _hf(x)\) has bounded variation over \(0\le x\le 1\). Then f(x) can be written in the form \(g(x)+H(x)\), where g(x) has bounded variation, and H(x) is additive.

The main result of the paper reads as follows.

Theorem 4

Let \(f, g:I\rightarrow \mathbb R\). If the mapping

$$\begin{aligned} I_h\ni x \longrightarrow f(x+h)-g(x) \end{aligned}$$

is increasing for all \(h>0\) then

$$\begin{aligned} f(x)=a(x)+f_0(x),\quad g(x)=a(x)+g_0(x),\quad x\in I, \end{aligned}$$
(5)

where \(f_0, g_0:I\rightarrow \mathbb R\) are functions of locally bounded variation, and \(a:\mathbb R\rightarrow \mathbb R\) is an additive function such that \(a(\mathbb Q)=\{0\}\). Moreover, under the assumption \(a(\mathbb Q)=\{0\}\) the decomposition (5) is unique.

Proof

By the assumption for arbitrary \(h>0\) the functions

$$\begin{aligned} I_h\ni x \longrightarrow f(x+h)-g(x) \quad \text {and}\quad I_{2h}\ni x \longrightarrow f(x+2h)-g(x), \end{aligned}$$

as well as

$$\begin{aligned} I_{2h}\ni x \longrightarrow f(x+2h)-g(x+h), \end{aligned}$$

are increasing which implies that the functions

$$\begin{aligned} I_{2h}\ni x\longrightarrow f(x+2h)-f(x+h),\ \text {and}\ I_{2h}\ni x\longrightarrow g(x+h)-g(x) \end{aligned}$$
(6)

are the differences of two increasing functions on the interval \(I_{2h}\) and consequently belong to the family \(LBV(I_{2h})\). We show that \(f=a_1+f_1\) where \(f_1\in LBV(I_0)\), for some interval of positive length \(I_0\subset I\) and \(a_1:\mathbb R\rightarrow \mathbb R\) is an additive function. The corresponding proof for the function g runs in a similar way. Changing x to \(x-h\) in the formula (6) for f we get that

$$\begin{aligned} \Delta _h f\in LBV(I_{-h}\cap I_h). \end{aligned}$$

Now, choose a number \(\delta >0\) in such a way that the interval \(I_0:=I_{-\delta }\cap I_\delta \) has a non-empty interior. We will show that f is decomposable on \(I_0\). On account of Lemma 2 it is enough to prove that for all \(a, b\in I_0,\ a<b\) f is decomposable on \(\left[ a,\frac{a+b}{2}\right] \). To see it fix arbitrary \(a, b\in I_0,\ a<b\) and consider a map \(\phi :\mathbb R\rightarrow \mathbb R\) given by the formula

$$\begin{aligned} \phi (x)=a+\frac{1}{2}(b-a)x,\quad x\in \mathbb R. \end{aligned}$$

Put \(J:=\phi ^{-1}(I_0)\). Clearly, \([0,2]\subset J\). Let us define a function \(f_1: J\rightarrow \mathbb R\) via the formula

$$\begin{aligned} f_1(x):=f(\phi (x))-x\left[ f\left( \frac{a+b}{2}\right) -f(a)\right] ,\quad x\in J. \end{aligned}$$

Obviously, \(f_1(0)=f_1(1)\). For any \(h>0\) and \(x\in J_h:=J\cap (J-h)\) we obtain

$$\begin{aligned} \Delta _hf_1(x)= & {} f_1(x+h)-f_1(x)=f(\phi (x+h))-f(\phi (x))-h\left[ f\left( \frac{a+b}{2}\right) -f(a)\right] \\= & {} f(\phi (x)+\phi (h)-a)-f(\phi (x))-h\left[ f\left( \frac{a+b}{2}\right) -f(a)\right] \\= & {} \Delta _{\phi (h)-a}f(\phi (x))-h\left[ f\left( \frac{a+b}{2}\right) -f(a)\right] . \end{aligned}$$

We have shown that \(\Delta _hf_1\in LBV(J_h)\). Let a function \(p:\mathbb R\rightarrow \mathbb R\) be the 1-periodic extension of the restriction \(f{_{1}}{_{|[0,1]}}\) i.e.

$$\begin{aligned} p(x)= & {} f(x),\quad x\in [0,1], \\ p(x+1)= & {} p(x),\quad x\in \mathbb R, \end{aligned}$$

and let us define a function \(f_2:J\rightarrow \mathbb R\) via the formula

$$\begin{aligned} f_2(x):=f_1(x)-p(x). \end{aligned}$$

Observe that for all \(x\in J\)

$$\begin{aligned} f_2(x)= & {} f_1(x)-p(x)=f_1([x]+\{x\})-p([x]+\{x\}) \\= & {} f_1(\{x\}+[x])-p(\{x\})=\Delta _{[x]}f_1(\{x\}), \end{aligned}$$

where [x] and \(\{x\}\) stand for an integer and the fractional part of the number x, respectively. Since for all \(x, h\in [0,1]\) we have

$$\begin{aligned} \Delta _hp(x)=\Delta _hf_1(x)-\Delta _hf_2(x), \end{aligned}$$

therefore, \(\Delta _h p\in LBV([0,1])\) for all \(h\in [0,1]\).

Now, we prove that \(\Delta _h p\in LBV(\mathbb R)\) for all \(h\in [0,1]\). To show it fix arbitrary \(c, d\in \mathbb R,\ c<d\) and \(h\in [0,1]\). Let us consider three cases:

  1. 1.

    If \([c]+1<[d]\) then

    $$\begin{aligned} V_c^d(\Delta _h p)= & {} V_c^{[c]+1}(\Delta _h p)+\sum _{j=[c]+1}^{[d]-1}V_j^{j+1}(\Delta _h p)+V_{[d]}^d(\Delta _h p) \\= & {} V_{\{c\}}^1(\Delta _h p)+\sum _{j=[c]+1}^{[d]-1} V_0^1(\Delta _h p)+V_0^{\{d\}}(\Delta _h p)<\infty . \end{aligned}$$
  2. 2.

    If \([c]+1>[d]\) then \([c]=[d]\) and consequently, \(c, d\in ([c],[c]+1)\). Therefore,

    $$\begin{aligned} V_c^d(\Delta _h p)=V_{\{c\}}^{\{d\}}(\Delta _h p)<\infty . \end{aligned}$$
  3. 3.

    In the case where \([c]+1=[d]\) we obtain

    $$\begin{aligned} V_c^d(\Delta _h p)= & {} V_c^{[c]+1}(\Delta _h p)+V_{[d]}^d(\Delta _h p) =V_{\{c\}}^1(\Delta _h p)+V_0^{\{d\}}(\Delta _h p)<\infty . \end{aligned}$$

Now, because for any \(h>0\) and natural number n such that \(n>h\) the formula

$$\begin{aligned} \Delta _h p(x)=\sum _{j=1}^{n-1}\Delta _{\frac{h}{n}} p\left( x+\frac{j}{n} h\right) , \end{aligned}$$

holds, \(\Delta _h p\in LBV(\mathbb R),\ h>0\). Finally, for \(h<0\) it is enough to apply the identity

$$\begin{aligned} \Delta _{-h}p(x)=-\Delta _h p(x-h) \end{aligned}$$

to obtain that \(\Delta _h p\in LBV(\mathbb R),\ h\in \mathbb R\). Applying de Bruijn’s theorem we infer that p is a sum of an additive function and a function of locally bounded variation which, due to the definition of p implies that f also has this property on the interval \(\left[ a,\frac{a+b}{2}\right] \), i.e.

$$\begin{aligned} f(x)={\bar{a}}(x)+{\bar{f}}(x), \quad x\in \left[ a,\frac{a+b}{2}\right] , \end{aligned}$$

where \({\bar{f}}:\left[ a,\frac{a+b}{2}\right] \rightarrow \mathbb R\) is a function of bounded variation and \({\bar{a}}:\mathbb R\rightarrow \mathbb R\) is an additive function. Now, put

$$\begin{aligned} f_0(x)={\bar{f}}(x)+{\bar{a}}(1)x,\quad x\in \left[ a,\frac{a+b}{2}\right] \end{aligned}$$

and

$$\begin{aligned} a_1(x):={\bar{a}}(x)-{\bar{a}}(1) x,\qquad x\in \mathbb R. \end{aligned}$$

We obtain immediately from the definition that \(a_1(\mathbb Q)=\{0\}\) and \(f=f_0+a_0\) on \(\left[ a,\frac{a+b}{2}\right] \), so f is decomposable on \(\left[ a,\frac{a+b}{2}\right] \) and by Lemma 2f is decomposable on I.

Analogously, we can prove that g is decomposable on I i.e. there exist a function \(g_0\) with locally bounded variation and an additive function \(a_2:\mathbb R\rightarrow \mathbb R\) such that \(a_2(\mathbb Q)=\{0\}\) and \(g=g_0+a_2\) on I.

Having in mind that for all \(h>0\) the function

$$\begin{aligned} I_h\ni x\longrightarrow f(x+h)-g(x) \end{aligned}$$

is increasing we infer that the additive function

$$\begin{aligned} \mathbb R\ni x\longrightarrow a_1(x)-a_2(x) \end{aligned}$$

is continuous, therefore it has the form

$$\begin{aligned} a_1(x)-a_2(x)=c x,\quad x\in \mathbb R, \end{aligned}$$

for some constant \(c\in \mathbb R\). Since for all \(\alpha \in \mathbb Q\) we have

$$\begin{aligned} 0=a_1(\alpha )-a_2(\alpha )=c\alpha , \end{aligned}$$

\(c=0\) and consequently \(a_1=a_2\).

For the proof of uniqueness of the representation (5) let us assume that

$$\begin{aligned} f(x)=f_1(x)+a_1(x)=f_2(x)+a_2(x),\quad x\in I, \end{aligned}$$

where \(f_1, f_2\in LBV(I)\) and \(a_1, a_2:\mathbb R\rightarrow \mathbb R\) are additive functions satisfying the property \(a_1(\mathbb Q)=a_2(\mathbb Q)=\{0\}\). Then

$$\begin{aligned} f_1(x)-f_2(x)=a_2(x)-a_1(x),\quad x\in I. \end{aligned}$$

Therefore, the additive function \(a_2-a_1\) has to be continuous which implies that \(a_1=a_2\) and consequently \(f_1=f_2\). The proof for the function g runs in a similar way. \(\square \)

The following example below shows that the functions f and g in the statement of the Theorem 4 need not be convex.

Example 1

Let \(f, g:\mathbb R\rightarrow \mathbb R\) be given by the formulas

$$\begin{aligned} f(x):=\sin x +3x,\quad g(x):=\cos x,\ x\in \mathbb R. \end{aligned}$$

An easy calculation shows that for arbitrary \(h>0\) the function

$$\begin{aligned} \mathbb R\ni x\longrightarrow \sin (x+h)+3(x+h)-\cos x \end{aligned}$$

is increasing but neither f nor g is convex.

Let us observe that if additionally we assume in Theorem 4 that for all \(h>0\) also the function

$$\begin{aligned} I_h\ni x \longrightarrow g(x+h)-f(x) \end{aligned}$$

is increasing then we obtain the expected representation. Namely, the following theorem holds true.

Theorem 5

Let \(I\subset \mathbb R\) be an open interval and let \(f,g:I\rightarrow \mathbb R\). Then the following conditions are pairwise equivalent:

  1. (i)

    the mappings

    $$\begin{aligned} I_h\ni x \longrightarrow f(x+h)-g(x) \ \ \text {and}\ \ I_h\ni x \longrightarrow g(x+h)-f(x) \end{aligned}$$

    are increasing for all \(h>0\);

  2. (ii)

    the inequality

    $$\begin{aligned} g(tx+(1-t)y)+f((1-t)x+ty)\le g(x)+f(y) \end{aligned}$$

    holds for all \(x, y\in I\) and all \(t\in [0,1]\);

  3. (iii)

    a function \(F:\mathbb R^2\rightarrow \mathbb R\) given by the formula

    $$\begin{aligned} F(x,y)=f(x)+g(y) \end{aligned}$$

    is Schur-convex;

  4. (iv)

    both functions f and g are Wright-convex and

    $$\begin{aligned} f(x)-g(x)=c,\quad x\in I, \end{aligned}$$

    for some constant \(c\in \mathbb R\);

  5. (v)

    there exist: a convex function \(w:I\rightarrow \mathbb R\), an additive function \(a:\mathbb R\rightarrow \mathbb R\) and a constant \(c\in \mathbb R\) such that

    $$\begin{aligned} f(x)=a(x)+w(x)\quad \text {and}\quad g(x)=a(x)+w(x)+c,\ x\in I. \end{aligned}$$

Proof

\((i)\Rightarrow (ii)\) Fix \(x, y\in I\) and \(t\in (0,1)\), arbitrarily. Assume that \(x<y\) and put \(h:=(1-t)(y-x)\). From the assumption (i)

$$\begin{aligned}{} & {} f(x+(1-t)(y-x))-g(x)\le f(y-(1-t)(y-x)+(1-t)(y-x))\\{} & {} \quad -g(y-(1-t)(y-x)), \end{aligned}$$

or equivalently,

$$\begin{aligned} f(tx+(1-t)y)+g((1-t)x+ty)\le f(y)+g(x). \end{aligned}$$

Analogously,

$$\begin{aligned}{} & {} g(x+(1-t)(y-x))-f(x)\le g(y-(1-t)(y-x)+(1-t)(y-x))\\{} & {} \quad -f(y-(1-t)(y-x)), \end{aligned}$$

therefore,

$$\begin{aligned} g(tx+(1-t)y)+f((1-t)x+ty)\le g(y)+f(x). \end{aligned}$$

We have shown that for all \(x, y\in I\) and \(t\in [0,1]\) we have

$$\begin{aligned} g(tx+(1-t)y)+f((1-t)x+ty)\le g(y)+f(x), \end{aligned}$$

which implies (ii).

\((ii)\Rightarrow (iii)\) Assume that (ii) is fulfilled. We can rewrite the inequality in condition (ii) as

$$\begin{aligned} f(Tx)\le F(x),\quad x\in I\times I, \end{aligned}$$

where

$$\begin{aligned} T=\left[ \begin{array}{cc} t&{}1-t\\ 1-t&{}t \end{array}\right] \end{aligned}$$

and \(t\in [0,1]\). Since any \(2\times 2\) double stochastic matrix is of the above form, (iii) holds.

The proofs of the implications: \((iii)\Rightarrow (iv)\) and \((iv)\Rightarrow (v)\) are direct consequences of Theorem 2.

\((v)\Rightarrow (i)\) Assume that

$$\begin{aligned} f(x)=a(x)+w(x),\quad \text {and}\quad g(x)=a(x)+w(x)+c, \quad x\in I \end{aligned}$$

where \(a:\mathbb R\rightarrow \mathbb R\) is an additive function, \(w:I\rightarrow \mathbb R\) is a convex function and \(c\in \mathbb R\) is a constant. Because every convex function \(w:I\rightarrow \mathbb R\) has the property that for any \(h>0\) the function

$$\begin{aligned} I_h\ni x \longrightarrow w(x+h)-w(x) \end{aligned}$$

is increasing and for any additive function \(a:\mathbb R\rightarrow \mathbb R\) we have

$$\begin{aligned} \mathbb R\ni x\longrightarrow a(x+h)-a(x)=a(h) \end{aligned}$$

then we get condition (i). The proof of the theorem is finished. \(\square \)