1 Introduction

One of the main open problems in sub-Riemannian geometry is the regularity of length-minimizing curves. Minimizers may be abnormal extremals, in the language of Geometric Control Theory, and R. Montgomery showed in [5] an example of abnormal extremal that is length minimizing. Liu and Sussmann proved later that the existence of length-minimizing abnormal extremals is typical of rank 2 distributions [3].

All known examples of length-minimizing curves are smooth. On the other hand, there is no regularity theory of a general character for sub-Riemannian “geodesics”, apart from the partial results of [2] and [6].

During the meeting Geometric control and sub-Riemannian geometry held in Cortona in May 2012, A. Agrachev and J. P. Gauthier suggested the following situation, in order to find a nonsmooth length-minimizing curve.

Consider in \({\mathbb{R }}^4\) the vector fields

$$\begin{aligned} X_1 = \frac{\partial }{\partial x_1} + 2 x_2 \frac{\partial }{\partial x_3} + x_3^2 \frac{\partial }{\partial x_4},\quad X_2 = \frac{\partial }{\partial x_2} - 2 x_1 \frac{\partial }{\partial x_3}, \end{aligned}$$
(1.1)

and denote by \(\mathcal{D }\) the distribution of \(2\)-planes in \({\mathbb{R }}^4\) spanned pointwise by \(X_1\) and \(X_2\):

$$\begin{aligned} \mathcal{D }(x) = \mathrm{span} \{ X_1(x), X_2(x)\},\quad x\in {\mathbb{R }}^4. \end{aligned}$$
(1.2)

A Lipschitz curve \(\gamma :[0,1]\rightarrow {\mathbb{R }}^4\) is horizontal if \({\dot{\gamma }}(t) \in \mathcal{D }(\gamma (t))\) for a.e. \(t\in [0,1]\). Namely, \(\gamma \) is horizontal if there exist two functions \(h_1,h_1\in L^{\infty }(0,1)\) such that

$$\begin{aligned} {\dot{\gamma }} = h_1 X_1(\gamma ) + h_2 X_2(\gamma )\quad \text{ a.e. } \text{ on }\, [0,1]. \end{aligned}$$
(1.3)

The length of \(\gamma \) is then defined as

$$\begin{aligned} L(\gamma ) = \int \limits _0^1 g_\gamma ({\dot{\gamma }},{\dot{\gamma }})^{1/2} {\text{ d }}t, \end{aligned}$$
(1.4)

where \(g_x\) is a metric on \({\mathcal{D }}(x), x\in {\mathbb{R }}^4\).

Fix a parameter \(\alpha >0\) and consider the initial and final points \(L = (-1,\alpha ,0,0)\in {\mathbb{R }}^4\) and \(R = (1,\alpha ,0,0)\in {\mathbb{R }}^4\). Let \({\bar{\gamma }}:[-1,1]\rightarrow {\mathbb{R }}^4\) be the curve

$$\begin{aligned} {\bar{\gamma }}_1(t) = t, \quad {\bar{\gamma }}_2(t) = \alpha |t|,\quad {\bar{\gamma }}_3(t) = 0,\quad {\bar{\gamma }}_4(t)=0, \quad t\in [-1,1]. \end{aligned}$$
(1.5)

The curve \({\bar{\gamma }}\) is horizontal and joins \(L\) to \(R\). Moreover, it can be easily checked that \({\bar{\gamma }}\) is an abnormal extremal in the sense of Geometric Control Theory. The question is whether the curve \({\bar{{\gamma }}}\) is length minimizing or not, especially for small \(\alpha >0\).

The interest of this question arises from the following consideration. Let \(M\) be an \(n\)-dimensional smooth manifold with \(n\ge 3\), and let \({\mathcal{D }}\) be a completely nonintegrable (i.e., bracket generating) distribution on \(M\). Let \({\mathcal{D }}_1={\mathcal{D }}\) and \({\mathcal{D }}_i = [{\mathcal{D }}_1,{\mathcal{D }}_{i-1}]\), i.e., \({\mathcal{D }}_i\) is the linear span of all commutators \([{\mathcal{D }}_1,{\mathcal{D }}_{i-1}]\). We let \(\mathcal{L }_0=\{0\}\) and \(\mathcal{L }_i = {\mathcal{D }}_1+\cdots +{\mathcal{D }}_i, i\ge 1\). By the nonintegrability condition, for any \(p\in M\), there exists \(r\in {\mathbb{N }}\) such that \(\mathcal{L }_r (p) = T_p M\), the tangent space of \(M\) at \(p\). Assume that \({\mathcal{D }}\) is equiregular, i.e., assume that for each \(i=1,\ldots ,r\)

$$\begin{aligned} \mathrm{dim}\, \big (\mathcal{L }_i(p) / \mathcal{L }_{i-1}(p)\big ) \,\text{ is } \text{ constant } \text{ for }\, p\in M. \end{aligned}$$
(1.6)

In [2], Leonardi and the author proved the following theorem.

Theorem 1.1

Let \((M,{\mathcal{D }},g)\) be a sub-Riemannian manifold, where \(g\) is a metric on \({\mathcal{D }}\), that satisfies (1.6) and

$$\begin{aligned}{}[\mathcal{L }_i,\mathcal{L }_j]\subset \mathcal{L }_{i+j-1}, \quad i, j\ge 2,\, i+j>4. \end{aligned}$$
(1.7)

Then, any curve in \(M\) with a corner is not length minimizing in \((M,{\mathcal{D }},g)\).

A “curve with a corner” is a \(\mathcal{D }\)-horizontal curve \(\gamma :[0,1]\rightarrow M\) such that at some point \(t\in (0,1)\) the left and right derivatives \({\dot{\gamma }}_L(t)\ne {\dot{\gamma }}_{R}(t)\) exist and are different and nonzero.

In view of Theorem 1.1, it is natural to look for a nonsmooth length-minimizing curve in a sub-Riemannian manifold violating (1.7). The step \(r\) associated with \((M, \mathcal{D })\) has to be at least \(5\), because condition (1.7) is automatically satisfied if \(r\le 4\). A first attempt could be to consider the free Carnot group of step \(5\) and rank \(2\). This structure is diffeomorphic to \({\mathbb{R }}^8\) and, by the results in [1], all abnormal extremals are computable.

If we drop the equiregularity condition (1.6), however, the search for structures violating (1.7) is easier. The manifold \(M= {\mathbb{R }}^4\) with the distribution \(\mathcal{D }\) spanned by the vector fields (1.1) is one such example.

For \(n\in {\mathbb{N }}\), we define the set of multi-indexes \(\mathcal{I }_n = \{1,2\}^n = \big \{(\beta _1,\ldots ,\beta _n)\in {\mathbb{N }}^n : \beta _1,\ldots ,\beta _n=1,2 \big \}\) and, for any \(\beta \in \mathcal{I }_n\), we let

$$\begin{aligned} X_\beta = [ X_{\beta _1},[\ldots ,[X_{\beta _{n-1}},X_{\beta _n}]\ldots ]]. \end{aligned}$$
(1.8)

We define the length of the commutator \(X_\beta \) as \(\mathrm{len}(X_\beta )= n\) if and only if \(\beta \in \mathcal{I }_n\).

For any \(\beta \in \mathcal{I }_3\cup \mathcal{I }_4\), we have \(X_\beta (0) = 0\). On the other hand, when \(\beta \in \mathcal{I }_5\), we have the following situation:

$$\begin{aligned} \frac{1}{16} X_\beta&= \frac{\partial }{\partial x_4},\quad \mathrm{when} \,\beta = (1,1,2,2,1)\, \text{ or }\, \beta =(1,2,1,2,1) \nonumber \\ \frac{1}{32} X_\beta&= - \frac{\partial }{\partial x_4},\quad \mathrm{when} \,\beta =(2,1,1,2,1) \nonumber \\ X_\beta&= 0 \quad \mathrm{otherwise}. \end{aligned}$$
(1.9)

In particular, there holds

$$\begin{aligned} \frac{1}{48} [[X_2,X_1],[[X_2,X_1],X_1]] = \frac{\partial }{\partial x_4}, \end{aligned}$$

and thus, (1.7) is violated with \(i=2\) and \(j=3\).

We show that the curve \({\bar{{\gamma }}}\) in (1.5) is not length minimizing for any small \(\alpha \), thus answering in the negative to the question raised by Agrachev and Gauthier. Namely, we prove the following.

Theorem 1.2

For any \(\alpha >0\) with \(\alpha \ne 1\), the curve \({\bar{\gamma }}\) in (1.5) is not length minimizing in \(({\mathbb{R }}^4,\mathcal{D }, g)\), for any choice of metric \(g\) on \({\mathcal{D }}\).

The proof of the theorem is constructive. For \(\alpha \ne 1\), we construct a horizontal curve joining the left and right end-points \(L\) and \(R\) that is strictly shorter than \({\bar{{\gamma }}}\).

When \(\alpha =1\), the construction does not work, and our shorter curves for \(\alpha \ne 1\) converge to the curve \({\bar{\gamma }}\) as \(\alpha \rightarrow 1\) (see Remark 3.1). The problem of establishing whether \({\bar{{\gamma }}}\) for \(\alpha =1\) is length minimizing or not is open.

In the second part of the paper, we extend Theorem 1.2 to \(4\)-dimensional sub-Riemannian manifolds having at the corner point of the involved curve the same infinitesimal structure as \(({\mathbb{R }}^4,\mathcal{D })\).

Let \(M\) be a \(4\)-dimensional smooth manifold, and let \(\mathcal{D }\subset TM\) be a distribution of \(2\)-planes on \(M\). Locally, we have \(\mathcal{D } = \mathrm{span} \{Y_1,Y_2\}\), where \(Y_1\) and \(Y_2\) are linearly independent smooth vector fields on \(M\). We denote by \(\mathcal{L }_2 =\mathrm{span} \{ Y_i, [Y_j,Y_k] : i,j,k=1,2\}\) the pointwise linear span of \(\mathcal{D }\) and of the brackets of \(\mathcal{D }\).

Fix a point \(p\in M\). We make the following three assumptions:

  1. (H1)

    The vector fields

    $$\begin{aligned} Y_1,\,Y_2,\, Y_3 = \frac{1}{4}[Y_2,Y_1],\, \text{ and }\, Y_{4} =\frac{1}{16} [Y_1,[Y_1,[Y_2,[Y_2,Y_1]]]] \end{aligned}$$
    (1.10)

    are linearly independent at the point \(p\).

  2. (H2)

    For all \(\beta \in \mathcal{I }_3\cup \mathcal{I }_4\), we have

    $$\begin{aligned} Y_{\beta } (p) = 0 \,\mathrm{mod} \, \mathcal{L }_{2}(p), \end{aligned}$$
    (1.11)

    where \( \mathcal{L }_{2}(p)= \{ Y(p) : Y\in \mathcal{L }_2 \}\).

  3. (H3)

    For all \(\beta \in \mathcal{I }_{5}\) with \(\beta =(*,*,*,2,1)\), we have

    $$\begin{aligned} Y_{\beta } (p)&= 16 Y_{4}(p)\, \mathrm{mod} \, \mathcal{L }_{2}(p),\, \text{ for } \,\beta = (1,2,1,2,1)\, \text{ or }\, \beta =(1,1,2,2,1), \nonumber \\ Y_{\beta } (p)&= -32 Y_4 (p) \, \mathrm{mod} \, \mathcal{L }_{2}(p),\, \text{ for }\, \beta = (2,1,1,2,1),\, \text{ and } \nonumber \\ Y_{\beta } (p)&= 0 \,\mathrm{mod}\, \mathcal{L }_{2}(p) \, \text{ otherwise }. \end{aligned}$$
    (1.12)

Finally, let \(g\) be a metric on \(\mathcal{D }\) making the vector fields \(Y_1\) and \(Y_2\) orthogonal at \(p\),

$$\begin{aligned} g_p(Y_1,Y_2) = 0. \end{aligned}$$
(1.13)

In Sect. 5, we prove the following result. Below, we denote by \({\dot{\gamma }}_L(t)\) and \({\dot{\gamma }}_R(t)\) the left and right derivative of a curve \(\gamma :[0,1]\rightarrow M\) at the point \(t\in (0,1)\). The curve \(\gamma \) has a corner at the point \(p=\gamma (t)\), if the left and right derivatives at \(t\) exist, do not vanish, and \({\dot{\gamma }}_L(t)\ne {\dot{\gamma }}_R(t)\).

Theorem 1.3

Let \((M,\mathcal{D }, g)\) be \(4\)-dimensional sub-Riemannian manifold satisfying (H1)–(H3) and (1.13) at the point \(p\in M\). Let \(\gamma :[0,1]\rightarrow M\) be a horizontal curve in \((M,\mathcal{D })\) having a corner at the point \(p=\gamma (t), t\in (0,1)\), such that \(g_p({\dot{\gamma }}_L(t),{\dot{\gamma }}_R(t))\ne 0\). Then, \(\gamma \) is not length minimizing.

The proof of Theorem 1.3 relies upon a blow-up argument reducing the situation to the one of the Theorem 1.2. Theorem 1.3 is proved in Sect. 5, and the preliminary nilpotent approximation is explained in Sect. 4. In Sect. 2, we set up the construction of the shorter curve used to prove Theorem 1.2. In Sect. 3, we solve a system of end-point equations, concluding the proof of Theorem 1.2.

We thank D. Vittone for his careful reading of an early version of this paper. We also thank the referee for suggesting to us possible extensions of Theorem 1.2.

2 Construction of the competing curve

In this section, we construct the competing curve used to prove Theorem 1.2. In a first step, we cut the corner of \({\bar{{\gamma }}}\) at \(t=0\) in the \(x_1x_2\)-plane. Lifting the new curve in \({\mathbb{R }}^2\) to a horizontal curve in \({\mathbb{R }}^4\) produces an error on the third and fourth coordinates of the final point. To restore the final point, we use two devices. We describe the intuition behind the construction. The vector fields \(X_1\) and \(X_2\) in (1.1) have a Heisenberg group structure in the \(x_1x_2x_3\) space. Thus, to restore the third coordinate, we have to add an “area” in the \(x_1x_2\)-plane equaling the “area” cut at the corner. We do this by means of a “rectangle” having basis of fixed length and variable height \({\varepsilon }_1\), the first parameter.

Restoring the fourth coordinate is more delicate. This is due to the fact that the coefficient of \(\partial /\partial x_4\) of \(X_1\) in (1.1), the power \(x_3^2\), is always nonnegative. This means that the cut produces a positive error, see (2.7). To correct it, the perturbation of \({\bar{{\gamma }}}\) must contain arcs where the first coordinate is decreasing. To do this, we add a square with side length \({\varepsilon }_2\), the second parameter, at the final point \(R\) in the \(x_1x_2\) plane.

Our goal is to prove that the final end-point can be adjusted via this construction. Technically, this means that we have to find solutions \({\varepsilon }_1,{\varepsilon }_2\) to a system of two equations depending on the cut parameter. Moreover, we need to check that the cost of length of “rectangle” and square does not exceed the gain of length of the cut. This will be done in Sect. 3.

We begin the construction of the shorter curve.

2.1 Cut

Fix a cut parameter \(0<\eta <1/4\), and let

$$\begin{aligned} T_\eta = \big \{ (x_1,x_2)\in {\mathbb{R }}^ 2 :\alpha |x_1|<x_2<\alpha \eta \big \}. \end{aligned}$$

The curve \({\bar{\gamma }}\) is deviated along the cut, i.e., along the side of the triangle \(T_\eta \) parallel to the \(x_1\)-axis. In other words, we cut the corner. The cut produces an error on the final point, namely on the third and fourth coordinates. To correct these errors, we use two devices, a “rectangle” and a square.

2.2 Rectangle

We deviate \({\bar{\gamma }}\) along a “rectangle” put along \({\bar{\gamma }}\). The length of the basis is fixed. The height is a variable parameter. Namely, for \({\varepsilon }_1>0\), we let

$$\begin{aligned} R_{{\varepsilon }_1} = \Big \{(x_1,x_2) \in {\mathbb{R }}^2 : \frac{1}{4} <x_1<\frac{1}{2},\, \alpha x_1 <x_2<\alpha x_1 +{\varepsilon }_1\Big \}. \end{aligned}$$

The curve \({\bar{{\gamma }}}\) is deviated following clockwise three sides of the rectangle. When \({\varepsilon }_1<0\), the construction is analogous, but the rectangle is below the curve, and we follow its boundary counterclockwise.

2.3 Square

Next, we use a square with bottom-left vertex at \(R\), the final point. Namely, for any \({\varepsilon }_2\in {\mathbb{R }}\), we let:

$$\begin{aligned} Q_{{\varepsilon }_2} = \Big \{(x_1,x_2) \in {\mathbb{R }}^2 : 1<x_1<1+|{\varepsilon }_2|, \alpha < x_2<\alpha +|{\varepsilon }_2| \Big \}. \end{aligned}$$

When \({\varepsilon }_2>0\), we follow the boundary of the square clockwise. When \({\varepsilon }_2<0\), we follow the boundary counterclockwise.

The devices \(R_{{\varepsilon }_1}\) and \(Q_{{\varepsilon }_2}\) produce effects in the coordinates 3 and 4 of \({\bar{\gamma }}\) after the cut. We call \(\gamma :[-1,1]\rightarrow {\mathbb{R }}^4\) the curve obtained after using \(T_\eta , R_{{\varepsilon }_1}\), and \(Q_{{\varepsilon }_2}\). To construct the coordinate \(\gamma _3\), we use the formula

$$\begin{aligned} {\gamma }_3(t) = 2 \int \limits _{-1}^t ({\gamma }_2\dot{\gamma }_1- {\gamma }_1\dot{\gamma }_2){\text{ d }}s , \quad t\in [0,1]. \end{aligned}$$
(2.1)

To construct \(\gamma _4\), we use the formula

$$\begin{aligned} \gamma _4(t) = \int \limits _{-1}^t {\dot{\gamma }}_1 \gamma _3 ^2\, {\text{ d }}s,\quad t\in [0,1]. \end{aligned}$$
(2.2)

On suitable subintervals, we shall use different parameterizations for \(\gamma \).

We construct step by step the horizontal lift of \(\gamma \).

2.4 Effect of \(T_\eta \)

We parameterize on the interval \([-\eta ,\eta ]\) the piece of \(\gamma \) that is on the horizontal part of \(\partial T_\eta \). We have, by (2.1),

$$\begin{aligned} {\gamma }_3(t) = \int \limits _{-\eta }^t 2\alpha \eta {\text{ d }}s = 2\alpha \eta (t+\eta ),\quad t\in [-\eta ,\eta ]. \end{aligned}$$

The final value of \(\gamma _3\) is \(\gamma _3(\eta ) = 4\alpha \eta ^2\). The final value of \(\gamma _4\) is, by (2.2),

$$\begin{aligned} \gamma _4(\eta ) = \int \limits _{-\eta }^\eta 4\alpha ^2 \eta ^2 (t+\eta )^2 {\text{ d }}t = \frac{32}{3}\alpha ^2 \eta ^5. \end{aligned}$$

2.5 First segment

Now, we have to follow for a time \(t\in [\eta ,1/4]\) the piece of the original curve \({\bar{{\gamma }}}\) connecting the cut to the rectangle \(R_{{\varepsilon }_1}\). Here, \({\gamma }_3\) is constant: \(\gamma _3 \equiv 4\alpha \eta ^2\). The final value of \(\gamma _4\) is

$$\begin{aligned} \gamma _4(1/4) = \frac{32}{3}\alpha ^2 \eta ^5+ \int \limits _\eta ^{1/4} 16\alpha ^2 \eta ^4 \, {\text{ d }}t = 16\alpha ^2 \eta ^4 \Big (\frac{1}{4} -\frac{\eta }{3}\Big ). \end{aligned}$$

2.6 Rectangle

1 \(^{\mathbf{st}}\) side. We compute \({\gamma }_3\) along the first vertical side \(\{(1/4,\alpha /4+t)\in {\mathbb{R }}^2: t\in [0,{\varepsilon }_1]\big \}\) of \(R_{{\varepsilon }_1}\):

$$\begin{aligned} \gamma _3(t) = 4\alpha \eta ^2-\frac{t}{2},\quad t\in [0,{\varepsilon }_1]. \end{aligned}$$

The final value is \(\gamma _3({\varepsilon }_1) = 4\alpha \eta ^2-\frac{{\varepsilon }_1}{ 2}\). This value is correct also when \({\varepsilon }_1<0\). Along this piece of curve, \(\gamma _4\) stays constant because \(\dot{\gamma }_1=0\).

2 \(^{\mathbf{nd}}\) side. Along the second side of the rectangle, we have

$$\begin{aligned} {\gamma }_1(t) =\frac{1}{4} +t,\quad {\gamma }_2(t) = \frac{{\alpha }}{4} +{\varepsilon }_1 +\alpha t,\quad t\in [0,1/4]. \end{aligned}$$

Then, the third coordinate is

$$\begin{aligned} {\gamma }_3(t)&= 4\alpha \eta ^2 -\frac{{\varepsilon }_1}{2} +2{\varepsilon }_1 t,\quad t\in [0,1/4], \\ {\gamma }_3(1/4)&= 4\alpha \eta ^2. \end{aligned}$$

The final value of \(\gamma _4\) at the end of this side is

$$\begin{aligned} \gamma _4(1/4)&= 16\alpha ^2 \eta ^4 \Big (\frac{1}{4} -\frac{\eta }{3}\Big ) + \int \limits _0^{1/4} \big ( 4\alpha \eta ^2 -\frac{{\varepsilon }_1}{2} + 2 {\varepsilon }_1 t\big )^2 {\text{ d }}t \nonumber \\&= 16\alpha ^2 \eta ^4 \Big (\frac{1}{4} -\frac{\eta }{3}\Big ) +4\alpha ^2 \eta ^4 -\frac{1}{2} \alpha \eta ^2 {\varepsilon }_1 +\frac{1}{48}{\varepsilon }_1^2 \nonumber \\&= \varrho . \end{aligned}$$
(2.3)

This is the final value of \(\gamma _4\) after the rectangle \(R_{{\varepsilon }_1}\). The number \(\varrho \) is defined via the last identity in (2.3).

3 \(^{\mathbf{rd}}\) side. We compute \({\gamma }_3\) along the second vertical side \(\{(1/2,\alpha /2+t)\in {\mathbb{R }}^2: t\in [0,{\varepsilon }_1]\big \}\) of the rectangle \(R_{{\varepsilon }_1}\). We have to follow \(-X_2\), and we get

$$\begin{aligned} \gamma _3(t)&= 4\alpha \eta ^2 +t,\quad t\in [0,{\varepsilon }_1],\\ \gamma _3({\varepsilon }_1)&= 4\alpha \eta ^2 +{\varepsilon }_{1}. \end{aligned}$$

Here, \(\gamma _4\) is constant because \(\dot{\gamma }_1=0\) on this side of the rectangle.

2.7 Second diagonal segment

Now, we have to follow for a time \(t\in [1/2,1]\) the piece of the original curve \({\bar{{\gamma }}}\) connecting \(R_{{\varepsilon }_1}\) to the final point \(R\). Here, \({\gamma }_3\) is constant: \(\gamma _3 \equiv 4\alpha \eta ^2+{\varepsilon }_1\). We call \(\Delta _0({\varepsilon }_1;\eta )\) the final value of \(\gamma _4\) before the square \(Q_{{\varepsilon }_2}\), and namely [recall the definition of \(\varrho \) in (2.3)],

$$\begin{aligned} \Delta _0({\varepsilon }_1;\eta )&= \varrho + \frac{1}{2} (4\alpha \eta ^2+{\varepsilon }_1)^2 \nonumber \\&= 16 \alpha ^2 \eta ^4\Big (1 -\frac{\eta }{3}\Big ) +\frac{7}{2} \alpha \eta ^2 {\varepsilon }_1 +\frac{25}{48} {\varepsilon }_1^2. \end{aligned}$$
(2.4)

2.8 Square

Case \({\varepsilon }_2>0\). We compute the effect on the third and fourth coordinates of the square \(Q_{{\varepsilon }_2}\). Here, we do the computations for the case \({\varepsilon }_2>0\). In this case, we follow the boundary \(\partial Q_{{\varepsilon }_2}\) clockwise.

1 \(^{\mathbf{st}}\) vertical side. We follow the segment \(\big \{(1,{\alpha }+t)\in {\mathbb{R }}^2: t\in [0,{\varepsilon }_2]\big \}\). The coordinate \(\gamma _3\) is

$$\begin{aligned} {\gamma }_3(t)&= 4\alpha \eta ^2+{\varepsilon }_1-2t,\quad t\in [0,{\varepsilon }_2],\\ {\gamma }_3({\varepsilon }_2)&= 4\alpha \eta ^2+{\varepsilon }_1-2{\varepsilon }_2. \end{aligned}$$

The coordinate \({\gamma }_4\) stays constant along this side, \(\gamma _4\equiv \Delta _0({\varepsilon }_1;\eta )\).

2 \(^{\mathbf{nd}}\) (horizontal) side. We follow the line segment \(\big \{(1+t,\alpha +{\varepsilon }_2)\in {\mathbb{R }}^2 : t\in [0,{\varepsilon }_2]\big \}\). The coordinate \({\gamma }_3\) is

$$\begin{aligned} {\gamma }_3(t)&= 4\alpha \eta ^2+{\varepsilon }_1-2{\varepsilon }_2+2 (\alpha +{\varepsilon }_2) t,\quad t\in [0,{\varepsilon }_2]\\ {\gamma }_3({\varepsilon }_2)&= 4\alpha \eta ^2+{\varepsilon }_1-2{\varepsilon }_2+2 (\alpha +{\varepsilon }_2) {\varepsilon }_{2}. \end{aligned}$$

Here, there is an important contribution to the fourth coordinate:

$$\begin{aligned} {\gamma }_4({\varepsilon }_2)&= \Delta _0({\varepsilon }_1;\eta ) + \int \limits _0^{{\varepsilon }_2} \big (4\alpha \eta ^2+{\varepsilon }_1-2{\varepsilon }_2+2 (\alpha +{\varepsilon }_2) t \big )^2 {\text{ d }}t \\&= \Delta _0({\varepsilon }_1;\eta ) +(4\alpha \eta ^2 + {\varepsilon }_1 -2{\varepsilon }_2)^2{\varepsilon }_2\\&\quad + 2(4\alpha \eta ^2 +{\varepsilon }_1-2{\varepsilon }_2)({\alpha }+{\varepsilon }_2){\varepsilon }_2^2 + \frac{4}{3} ({\alpha }+{\varepsilon }_2)^2 {\varepsilon }_{2}^3\\&= \Delta _1({\varepsilon }_2,{\varepsilon }_2;\eta ). \end{aligned}$$

Above, \(\Delta _1({\varepsilon }_2,{\varepsilon }_2;\eta )\) is defined via the last identity.

3 \(^{\mathbf{rd}}\) (vertical) side. We have to follow the vertical side \(\big \{ (1+{\varepsilon }_2, {\alpha }+t)\in {\mathbb{R }}^2: t\in [0,{\varepsilon }_2]\big \}\) along the vector field \(-X_2\). The final coordinate \({\gamma }_3\) is:

$$\begin{aligned} {\gamma }_3({\varepsilon }_2) = 4 \alpha \eta ^2 +{\varepsilon }_1 +2{\alpha }{\varepsilon }_2 +4{\varepsilon }_2^2. \end{aligned}$$

The coordinate \({\gamma }_4\) does not change.

4 \(^{\mathbf{th}}\) (horizontal) side. We have to follow the horizontal side \(\big \{ (1+t, \alpha )\in {\mathbb{R }}^2: t\in [0,{\varepsilon }_2]\big \}\) along the vector field \(-X_1\). The coordinate \({\gamma }_3\) is as follows:

$$\begin{aligned} {\gamma }_3(t)&= 4 \alpha \eta ^2 +{\varepsilon }_1 +2{\alpha }{\varepsilon }_2 +4{\varepsilon }_2^2-2\alpha t,\quad t\in [0,{\varepsilon }_2] \\ {\gamma }_3({\varepsilon }_2)&= 4 \alpha \eta ^2 +{\varepsilon }_1 +4{\varepsilon }_2^2. \end{aligned}$$

This is the final value of \({\gamma }_3\) after the entire construction, when \({\varepsilon }_2>0\). When \({\varepsilon }_2<0\), we have to replace \({\varepsilon }_2^2\) above with \(\mathrm{sgn}({\varepsilon }_2) |{\varepsilon }_2|^2\). There is a change of sign. We call \(\Gamma _3({\varepsilon }_1,{\varepsilon }_2;\eta )\) the final value of \(\gamma _3\) after the entire construction, and namely,

$$\begin{aligned} \Gamma _3({\varepsilon }_1,{\varepsilon }_2;\eta ) = 4 \alpha \eta ^2 +{\varepsilon }_1 +4\mathrm{sgn}({\varepsilon }_2)|{\varepsilon }_2|^2. \end{aligned}$$
(2.5)

The final value of \(\gamma _4\) is as follows:

$$\begin{aligned} {\gamma }_4({\varepsilon }_2)&= \Delta _1\big ({\varepsilon }_2,{\varepsilon }_2;\eta \big ) -\int \limits _0^{{\varepsilon }_2} \Big (4 \alpha \eta ^2 +{\varepsilon }_1 +2\alpha {\varepsilon }_2 +4{\varepsilon }_2^2-2\alpha t \Big )^2 {\text{ d }}t \\&= \Delta _1({\varepsilon }_2,{\varepsilon }_2;\eta ) -\Big [ \big ( 4\alpha \eta ^2+{\varepsilon }_1+4{\varepsilon }_2^2\big )^2{\varepsilon }_2 +2\big (4\alpha \eta ^2 +{\varepsilon }_1 +4{\varepsilon }_2^2\big ){\alpha }{\varepsilon }_2^2+\frac{4}{3} {\alpha }^2 {\varepsilon }_2^3\Big ] \\&= \Delta _0\big ({\varepsilon }_1;\eta \big )+ 4 {\varepsilon }_2^2 \big (1+2{\varepsilon }_2\big ) \Big [ {\varepsilon }_2\big (1-2{\varepsilon }_2\big )-4{\alpha }\eta ^2-{\varepsilon }_1\Big ] \\&+2{\varepsilon }_2^3 \Big [4{\alpha }\eta ^2 +{\varepsilon }_1- 2 \big ({\alpha }+{\varepsilon }_2\big )- 4{\alpha }{\varepsilon }_2 \Big ] + \frac{4}{3} {\varepsilon }_2^4 \big (2{\alpha }+{\varepsilon }_2\big ). \end{aligned}$$

Recalling (2.4), we let for \({\varepsilon }_2>0\)

$$\begin{aligned} \Gamma _4({\varepsilon }_1,{\varepsilon }_2;\eta )&= 16 \alpha ^2 \eta ^4\Big (1 -\frac{\eta }{3}\Big ) +\frac{7}{2} \alpha \eta ^2 {\varepsilon }_1 +\frac{25}{48} {\varepsilon }_1^2 \nonumber \\&+ 4 {\varepsilon }_2^2 (1+2{\varepsilon }_2) \big [ {\varepsilon }_2(1-2{\varepsilon }_2)-4{\alpha }\eta ^2-{\varepsilon }_1\big ] \nonumber \\&+2{\varepsilon }_2^3 \big [4{\alpha }\eta ^2 +{\varepsilon }_1- 2 ({\alpha }+{\varepsilon }_2)- 4{\alpha }{\varepsilon }_2 \big ] + \frac{4}{3} {\varepsilon }_2^4 (2{\alpha }+{\varepsilon }_2). \end{aligned}$$
(2.6)

2.9 Square

Case \({\varepsilon }_2<0\). We compute the effect of the square \(Q_{{\varepsilon }_2}\) on the fourth coordinate when \({\varepsilon }_2<0\). In this case, we follow \(\partial Q_{{\varepsilon }_2}\) counterclockwise.

1 \(^{\mathbf{st}}\) horizontal side. We follow the segment \(\big \{(1+t,{\alpha })\in {\mathbb{R }}^2: t\in [0,|{\varepsilon }_2|]\big \}\). The coordinate \(\gamma _3\) is

$$\begin{aligned} {\gamma }_3(t)&= 4\alpha \eta ^2+{\varepsilon }_1+2\alpha t,\quad t\in [0,|{\varepsilon }_2|],\\ {\gamma }_3(|{\varepsilon }_2|)&= 4\alpha \eta ^2+{\varepsilon }_1+2 \alpha |{\varepsilon }_2|. \end{aligned}$$

The final value of the coordinate \(\gamma _4\) is

$$\begin{aligned} {\gamma }_4(|{\varepsilon }_2|)&= \Delta _0({\varepsilon }_1;\eta )+ \int \limits _0^{|{\varepsilon }_2|} \big ( 4\alpha \eta ^2+{\varepsilon }_1+2\alpha t \big ) ^2{\text{ d }}t \\&= \Delta _0({\varepsilon }_1;\eta ) + |{\varepsilon }_2| (4\alpha \eta ^2 +{\varepsilon }_1)^2 +2\alpha |{\varepsilon }_2|^2 (4\alpha \eta ^2+{\varepsilon }_1)+\frac{4}{3} \alpha ^2 |{\varepsilon }_2|^3 \\&= \Delta _1({\varepsilon }_1,{\varepsilon }_2;\eta ). \end{aligned}$$

Here, \(\Delta _1({\varepsilon }_1,{\varepsilon }_2)\) for \({\varepsilon }_2<0\) is defined via the last identity.

2 \(^{\mathbf{nd}}\) (vertical) side. We follow the line segment \(\big \{(1+|{\varepsilon }_2|,\alpha +t)\in {\mathbb{R }}^2 : t\in [0,|{\varepsilon }_2|]\big \}\). The coordinate \({\gamma }_3\) is

$$\begin{aligned} {\gamma }_3(t)&= 4\alpha \eta ^2+{\varepsilon }_1+2\alpha |{\varepsilon }_2| - 2 (1+|{\varepsilon }_2|) t,\quad t\in [0,|{\varepsilon }_2|] \\ {\gamma }_3(|{\varepsilon }_2|)&= 4\alpha \eta ^2+{\varepsilon }_1+2\alpha |{\varepsilon }_2| - 2 (1+|{\varepsilon }_2|)|{\varepsilon }_2|. \end{aligned}$$

The fourth coordinate does not change.

3 \(^{\mathbf{rd}}\) (horizontal) side. We have to follow the horizontal side \(\big \{ (1+t, {\alpha }+|{\varepsilon }_2|)\in {\mathbb{R }}^2: t\in [0,|{\varepsilon }_2|]\big \}\) along the vector field \(-X_1\). The coordinate \({\gamma }_3\) is:

$$\begin{aligned} {\gamma }_3(t)&= 4\alpha \eta ^2+{\varepsilon }_1+2\alpha |{\varepsilon }_2| - 2 (1+|{\varepsilon }_2|)|{\varepsilon }_2|-2(\alpha +|{\varepsilon }_2|)t,\\ {\gamma }_3(|{\varepsilon }_2|)&= 4\alpha \eta ^2+{\varepsilon }_1+2\alpha |{\varepsilon }_2| - 2 (1+|{\varepsilon }_2|)|{\varepsilon }_2|-2(\alpha +|{\varepsilon }_2|)|{\varepsilon }_2|, \end{aligned}$$

The final value of the coordinate \(\gamma _4\) is

$$\begin{aligned} \gamma _4(|{\varepsilon }_2|)&= \Delta _1 ({\varepsilon }_1,{\varepsilon }_2;\eta )-\int \limits _0^{|{\varepsilon }_2|} \big ( 4\alpha \eta ^2+{\varepsilon }_1+2\alpha |{\varepsilon }_2| - 2 (1+|{\varepsilon }_2|)|{\varepsilon }_2|-2(\alpha +|{\varepsilon }_2|)t \big ) ^2 {\text{ d }}t \\&= \Delta _1({\varepsilon }_1,{\varepsilon }_2;\eta ) -|{\varepsilon }_2| \Big ( 4\alpha \eta ^2 +{\varepsilon }_1+ 2\alpha |{\varepsilon }_2|-2(1+|{\varepsilon }_2|)|{\varepsilon }_2|\Big )^2 \\&\quad + 2 (\alpha +|{\varepsilon }_2|) |{\varepsilon }_2|^2 \Big ( 4\alpha \eta ^2 +{\varepsilon }_1+ 2\alpha |{\varepsilon }_2|-2(1+|{\varepsilon }_2|)|{\varepsilon }_2|\Big ) -\frac{4}{3} (\alpha +|{\varepsilon }_2|)^2 |{\varepsilon }_2|^3. \end{aligned}$$

This is the final value of \(\gamma _4\) after the entire construction when \({\varepsilon }_2<0\). Then, for \({\varepsilon }_2<0\) we let

$$\begin{aligned} \Gamma _4({\varepsilon }_1,{\varepsilon }_2;\eta )&= 16 \alpha ^2 \eta ^4\Big (1 -\frac{\eta }{3}\Big ) +\frac{7}{2} \alpha \eta ^2 {\varepsilon }_1 +\frac{25}{48} {\varepsilon }_1^{2} \nonumber \\&\quad + |{\varepsilon }_2| ( 4\alpha \eta ^2 +{\varepsilon }_1)^2 +2\alpha |{\varepsilon }_2|^2 (4\alpha \eta ^2+{\varepsilon }_1)+\frac{4}{3} \alpha ^2 |{\varepsilon }_2|^3 \nonumber \\&\quad -|{\varepsilon }_2| \Big ( 4\alpha \eta ^2 +{\varepsilon }_1+ 2\alpha |{\varepsilon }_2|-2(1+|{\varepsilon }_2|)|{\varepsilon }_2|\Big )^2 \nonumber \\&\quad + 2 (\alpha +|{\varepsilon }_2|) |{\varepsilon }_2|^2 \Big ( 4\alpha \eta ^2 +{\varepsilon }_1+ 2\alpha |{\varepsilon }_2|-2(1+|{\varepsilon }_2|)|{\varepsilon }_2|\Big ) \nonumber \\&\quad -\frac{4}{3} (\alpha +|{\varepsilon }_2|)^2 |{\varepsilon }_2|^3. \end{aligned}$$
(2.7)

3 Proof of Theorem 1.2

In this section, we complete the proof of Theorem 1.2. We show that the system of end-point equations has a solution, and we prove that, for a small \(\eta >0\), this solution provides a curve that is shorter than \({\bar{\gamma }}\).

To correct the coordinates \(\gamma _3\) and \(\gamma _4\), we have to find solutions \({\varepsilon }_1\) and \({\varepsilon }_2\), depending on the (small) parameter \(\eta >0\), to the following system of equations

$$\begin{aligned} \left\{ \begin{array}{l} \Gamma _3({\varepsilon }_1,{\varepsilon }_2;\eta )=0 \\ \Gamma _4({\varepsilon }_1,{\varepsilon }_2;\eta )=0. \end{array} \right. \end{aligned}$$
(3.1)

Here, \(\Gamma _{4}({\varepsilon }_1,{\varepsilon }_2;\eta )\) is defined in (2.6) and (2.7), for \({\varepsilon }_2>0\) and \({\varepsilon }_2<0\), respectively.

By (2.5), the first equation \(\Gamma _3({\varepsilon }_1,{\varepsilon }_2;\eta )=0\) is

$$\begin{aligned} 4 \alpha \eta ^2 +{\varepsilon }_1 +4\mathrm{sgn}({\varepsilon }_2) |{\varepsilon }_2|^2=0. \end{aligned}$$
(3.2)

We compute \({\varepsilon }_1\) using Eq. (3.2) and replace this value into the second equation \(\Gamma _4({\varepsilon }_1,{\varepsilon }_2;\eta )=0\).

3.1 Case \(\alpha >1\)

In this case, we look for a solution \({\varepsilon }_2>0\), and we use formula (2.6) for \(\Gamma _4({\varepsilon }_1,{\varepsilon }_2;\eta )\). Letting \({\varepsilon }_2 =\sigma > 0\), we obtain the following equation in \(\sigma \)

$$\begin{aligned}&\frac{1}{3} \Big ( 31 \alpha ^2\eta ^4 -16\alpha ^2\eta ^5+8\alpha \eta ^2\sigma ^2 +25 \sigma ^4 \Big ) \nonumber \\&\quad + 4 \sigma ^3 \big (1-\alpha +\sigma (3-2\alpha )+2\sigma ^2\big ) + \frac{4}{3} \sigma ^4 (2{\alpha }+\sigma ) =0. \end{aligned}$$
(3.3)

When \(\alpha >1\) Eq. (3.3) reads

$$\begin{aligned} \phi (\sigma ) = \alpha ^2\eta ^4 (31-16\eta ) +8\alpha \eta ^2\sigma ^2+12(1-\alpha )\sigma ^3+o(\sigma ^3)=0, \end{aligned}$$
(3.4)

where \(o(\sigma ^3)\) contains terms with \(\sigma ^4\) and \(\sigma ^5\). Notice that the coefficient of \(\sigma ^3\) is negative, because \(\alpha >1\). The function \(\phi \) introduced in (3.4) is continuous, and moreover,

$$\begin{aligned} \phi (0) = \alpha ^2 \eta ^4 (31-16\eta )>0, \end{aligned}$$

as soon as \(\eta <31/16\). In fact, we have \(\eta <1/4\). On the other hand, we have

$$\begin{aligned} \phi \Bigg ( C\root 3 \of {\frac{\alpha ^2}{\alpha -1}}\eta ^{4/3}\Bigg ) = \alpha ^2 \eta ^4 \big (31-12 C^3\big )+o\big (\eta ^4\big )<0, \end{aligned}$$

provided that we choose a constant \(C\) such that \(12 C^3>31\) and \(\eta >0\) is small enough.

By the theorem of zeros for continuous functions, for any \(\eta >0\) that is small enough, Eq. (3.4) has a solution \({\varepsilon }_2=\sigma >0\) such that

$$\begin{aligned} {\varepsilon }_2 = \sigma \le C \root 3 \of { \frac{ \alpha ^2}{ \alpha -1 } } \eta ^{4/3}. \end{aligned}$$
(3.5)

Moreover, by (3.2),

$$\begin{aligned} {\varepsilon }_1 =- 4 \alpha \eta ^2 - 4\mathrm{sgn}({\varepsilon }_2) |{\varepsilon }_2|^2 =- 4 \alpha \eta ^2 - 4{\varepsilon }_2^2. \end{aligned}$$

In particular, \({\varepsilon }_1\) is asymptotic to \(-4\alpha \eta ^2\), for small \(\eta >0\), and thus, we have

$$\begin{aligned} |{\varepsilon }_1|\le 5\alpha \eta ^2, \end{aligned}$$
(3.6)

for all \(\eta >0\) that are sufficiently small.

3.2 Case \(0<\alpha <1\)

In this case, we look for solutions \({\varepsilon }_1,{\varepsilon }_2\) to the system (3.1) such that \({\varepsilon }_2<0\). The unknown \({\varepsilon }_1\) is determined by (3.2). We use formula (2.7) for \(\Gamma _4({\varepsilon }_1,{\varepsilon }_2;\eta )\). Letting \(\sigma =|{\varepsilon }_2|> 0\), equation \(\Gamma _4({\varepsilon }_1,{\varepsilon }_2;\eta )=0\) reads

$$\begin{aligned}&16 \alpha ^2 \eta ^4\Big (1 -\frac{\eta }{3}\Big ) +14 \alpha \eta ^2\big ( \sigma ^2- \alpha \eta ^2 \big ) +\frac{25}{3} \big ( \sigma ^2- \alpha \eta ^2 \big )^2 + 16\sigma ^5 +8\alpha \sigma ^4\nonumber \\&+\frac{4}{3} \alpha ^2 \sigma ^3 -4\sigma ^3 \big ( \sigma + \alpha - 1\big )^2 + 4 (\alpha +\sigma ) \sigma ^3 \big ( \sigma + \alpha -1 \big ) -\frac{4}{3} (\alpha +\sigma )^2 \sigma ^3=0.\nonumber \\ \end{aligned}$$
(3.7)

We may shorten Eq. (3.7) in the following way:

$$\begin{aligned} \alpha ^2 \eta ^4 (31-16\eta ) - 8 \alpha \eta ^2 \sigma ^2 + 12\big (\alpha -1\big ) \sigma ^3 +o\big (\sigma ^3\big ) =0. \end{aligned}$$
(3.8)

Compare Eqs. (3.8) and (3.4). Notice that the coefficient of \(\sigma ^3\) in (3.8) is negative. The same argument used in the case \(\alpha >1\) proves that there exists a solution \(|{\varepsilon }_2| = \sigma \) to Eq. (3.8) such that

$$\begin{aligned} 0< |{\varepsilon }_2| = \sigma \le C \root 3 \of { \frac{ \alpha ^2}{1- \alpha } } \eta ^{4/3}, \end{aligned}$$
(3.9)

where \(C\) is the same constant as in the case \(\alpha >1\). As above, we deduce that

$$\begin{aligned} |{\varepsilon }_1|\le 5\alpha \eta ^{2}. \end{aligned}$$
(3.10)

3.3 Gain of length

We compute the difference between the length of the original curve \({\bar{{\gamma }}}\) and the curve \(\gamma \) obtained after applying the cut \(T_\eta \), the “rectangle” \(R_{{\varepsilon }_1}\), and the square \(Q_{{\varepsilon }_2}\)

Without loss of generality, we can assume that the metric \(g\) on \({\mathcal{D }}\) is the one making \(X_1\) and \(X_1\) orthonormal. In this case, the length \(L(\gamma )\) of a horizontal curve \(\gamma \) as in (1.4) is

$$\begin{aligned} L(\gamma ) = \int \limits _0^1 |h(t)| {\text{ d }}t, \quad h=(h_1,h_2). \end{aligned}$$
(3.11)

By (3.11), the gain of length \(G(\eta ) \) obtained through the cut \(T_\eta \) is

$$\begin{aligned} G(\eta ) = 2\eta \big ( \sqrt{1+\alpha ^2}-1\big ). \end{aligned}$$
(3.12)

By (3.6) and (3.9), the cost of length \(C_R(\varepsilon _1)\) of the correction made through the “rectangle” \(R_{{\varepsilon }_1} \) is

$$\begin{aligned} C_R({\varepsilon }_1) = 2 |{\varepsilon }_1 |\le 10\alpha \eta ^2. \end{aligned}$$

By (3.5) and (3.9), the cost of length \(C_Q(\varepsilon _2)\) of the correction made through the square \(Q_{{\varepsilon }_2} \) is

$$\begin{aligned} C_Q({\varepsilon }_2) = 4 |{\varepsilon }_2 |\le 4 C \root 3 \of { \frac{ \alpha ^2}{ \alpha -1 } } \eta ^{4/3}. \end{aligned}$$

We conclude that

$$\begin{aligned} G(\eta ) - C_R({\varepsilon }_1)-C_Q({\varepsilon }_2) \ge 2\eta \big ( \sqrt{1+\alpha ^2}-1\big )-10\alpha \eta ^2 - 4 C \root 3 \of { \frac{ \alpha ^2}{ \alpha -1 } } \eta ^{4/3} > 0, \end{aligned}$$

for all \(\eta >0\) that are sufficiently small.

Thus, the construction provides a horizontal curve \(\gamma \) shorter than \({\bar{\gamma }}\) and joining the same end-points \( L\) and \(R\). This ends the proof of Theorem 1.2.

Remark 3.1

When \(\alpha \rightarrow 1\), our solutions \({\varepsilon }_1,{\varepsilon }_2\) to the system (3.1), along with the parameter \(\eta \), converge to \(0\) and the shorter curve \(\gamma \) converges to \({\bar{\gamma }}\).

We may look for solutions to (3.1) in the case \(\alpha =1\). However, when \(\alpha =1\), Eq. (3.3) for the case \(\sigma ={\varepsilon }_2>0\) reads

$$\begin{aligned} \eta ^4 (31-16\eta ) +8\eta ^2\sigma ^2+45 \sigma ^4 +o\big (\sigma ^4\big )=0. \end{aligned}$$

This equation has no small solution \(\sigma >0\) for small \(\eta \).

On the other hand, when \(\alpha =1\), Eq. (3.7) for the case \(\sigma = |{\varepsilon }_2|\) with \({\varepsilon }_2<0\) reads

$$\begin{aligned} \eta ^4(31-16\eta ) -8\eta ^2\sigma ^2 +53\sigma ^4 + o\big (\sigma ^4\big )=0. \end{aligned}$$

And also this equation has for small \(\eta >0\) no small solution \(\sigma >0\).

Remark 3.2

The shorter curve \(\gamma \) constructed above has a “curl” at the end-point \(R\). When \(\alpha >1\), this curl is oriented clockwise (i.e., \({\varepsilon }_2>0\)). When \(0<\alpha <1\), it is oriented counterclockwise (i.e., \({\varepsilon }_2<0\)). This suggests that the geodesic joining \(L\) to \(R\) displays a similar behavior. One may wonder what is the behavior at the point \(R\) of the length-minimizing curve joining \(L\) to \(R\) when \(\alpha =1\).

Remark 3.3

The proof of Theorem 1.2 shows why the problem of extending Theorem 1.1 to situations where (1.6) fails is technically complicated. The presence of coefficients depending on nonhorizontal coordinates [as the coefficient \(x_3^2\) in the vector field \(X_1\) in (1.1)] makes the construction of horizontal competitors complicated.

4 Nilpotent approximation

Let \((M,\mathcal{D })\) be a \(4\)-dimensional sub-Riemannian manifold satisfying (H1)-(H3) at the point \(p\in M\), where \(\mathcal{D } = \mathrm{span}\{Y_1,Y_2\}\), and let \(Y_1,\ldots ,Y_4\) be linearly independent vector fields at the point \(p\), as in (1.10). In this section, we construct the homogeneous nilpotent approximation of \((M,\mathcal{D })\) at \(p\).

In a neighborhood of \(p\in M\), we fix the exponential coordinates of the first type induced by the frame \(Y_1,\ldots ,Y_4\) starting from \(p\). Then, we can identify \(M\) with \({\mathbb{R }}^4, Y_1,\ldots ,Y_4\) with vector fields on \({\mathbb{R }}^4\), and \(p\) with \(0\in {\mathbb{R }}^4\). We have

$$\begin{aligned} x = \exp \Bigg (\sum _{i=1}^4 x_i Y_i\Bigg )(0),\quad x = (x_1,\ldots ,x_4)\in {\mathbb{R }}^4. \end{aligned}$$
(4.1)

Here, the exponential mapping is defined by \(\exp (Y)(0) = \gamma (1)\) where \(\gamma \) is the solution of \({\dot{\gamma }} = Y(\gamma )\) and \(\gamma (0)=0\).

We assign to the coordinate \(x_1\) the weight \(w_1=1\), to \(x_2\) the weight \(w_2=1\), to \(x_3\) the weight \(w_3=2\), and to \(x_4\) the weight \(w_4=5\). In fact, the length of \(Y_3\) is \(\mathrm{len}(Y_3)=2\), and the length of \(Y_4\) is \(\mathrm{len}(Y_4)=5\). The natural dilations adapted to the frame \(Y_1,\ldots ,Y_4\) are

$$\begin{aligned} \delta _\lambda (x) = \big (\lambda x_1, \lambda x_2, \lambda ^2 x_3, \lambda ^5 x_4\big ),\quad x\in {\mathbb{R }}^4,\, \lambda >0. \end{aligned}$$

Let \(Y=Y_\beta \) be a commutator of \(Y_1,\ldots ,Y_4\), with the notation (1.8) for iterated commutators. Then, we have

$$\begin{aligned} Y = \sum _{i=1}^4 a_i (x) \frac{\partial }{\partial x_i}, \end{aligned}$$
(4.2)

where \(a_i\in C^{\infty }({\mathbb{R }}^4), i=1,\ldots ,4\), are smooth functions that have the structure described in the following proposition.

Proposition 4.1

Assume that \((H1)\) and \((H2)\) hold. There are polynomials \(p_i:{\mathbb{R }}^4\rightarrow {\mathbb{R }}\) and functions \(r_i :{\mathbb{R }}^4\rightarrow {\mathbb{R }}, i=1,\ldots ,4\), such that:

  1. i)

    \(a_i(x) = p_i(x) + r_i(x), x\in {\mathbb{R }}^4\);

  2. ii)

    \(p_i(\delta _\lambda (x)) = \lambda ^{w_i-\mathrm{len}(Y)} p_i(x)\), where \(\mathrm{len}(Y)\) is the length of \(Y\);

  3. iii)

    \(\displaystyle \lim _{\lambda \rightarrow \infty } \lambda ^{w_i -\mathrm{len}(Y)} r_i(\delta _{1/\lambda } (x))=0, x\in {\mathbb{R }}^4\).

Proposition 4.1 can be proved as in [4, on page 306]. We omit the details, here.

Let \(Y= Y_\beta \) be a vector field as in (4.2). For \(\lambda >0\), we let

$$\begin{aligned} Y^\lambda (x) = \sum _{i=1}^4 \lambda ^{w_i -\mathrm{len}(Y) } a_i (\delta _{1/\lambda } (x))\frac{\partial }{\partial x_i},\quad x\in {\mathbb{R }}^4. \end{aligned}$$
(4.3)

The mapping \(Y\mapsto Y^\lambda \) is bracket-preserving. Namely, for any multi-index \(\beta \), and for \(i=1,2\), we have

$$\begin{aligned}{}[Y_i,Y_\beta ]^\lambda = [Y_i^\lambda , Y_\beta ^\lambda ],\quad \lambda >0. \end{aligned}$$
(4.4)

The vector fields \(Y_1^\lambda ,\ldots , Y_4^\lambda \) induce on \({\mathbb{R }}^4\) exponential coordinates of the first type starting from \(0\):

$$\begin{aligned} x = \exp \Bigg (\sum _{i=1}^4 x_i Y_i^\lambda \Bigg )(0),\quad x \in {\mathbb{R }}^4. \end{aligned}$$
(4.5)

The proof of (4.5) relies upon the fact that if a curve \(\gamma \) solves \({\dot{\gamma }} =\sum _{i=1}^4 x_i Y_i(\gamma )\), then the curve \(\gamma _ \lambda = \delta _\lambda (\gamma )\) solves \({\dot{\gamma }}_\lambda =\sum _{i=1}^4 \lambda ^{ w_i}x_i Y_i^\lambda (\gamma _\lambda )\).

By Proposition 4.1, for any \(Y=Y_\beta \) as in (4.2), we can define the vector field \(Y^\infty \) in \({\mathbb{R }}^4\)

$$\begin{aligned} Y^\infty (x) = \lim _{\lambda \rightarrow \infty } Y^\lambda (x) = \sum _{i=1}^4 p_i(x) \frac{\partial }{\partial x_i},\quad x\in {\mathbb{R }}^4, \end{aligned}$$

where \(p_i, i=1,\ldots ,4\), are polynomials such that \(p_i\circ \delta _\lambda = \lambda ^{w_i-\mathrm{len}(Y)}p_i\). In particular, if \(w_i<\mathrm{len} (Y)\), then \(p_i = 0\). Passing to the limit as \(\lambda \rightarrow \infty \) in (4.4), we see that also the mapping \(Y\mapsto Y^\infty \) is bracket-preserving

$$\begin{aligned}{}[Y_i,Y_\beta ]^\infty = [Y_i^\infty , Y_\beta ^\infty ],\quad \lambda >0. \end{aligned}$$
(4.6)

Moreover, passing to the limit as \(\lambda \rightarrow \infty \) in (4.5), we see that \(Y_1^\infty ,\ldots ,Y_4^\infty \) induce exponential coordinates of the first type:

$$\begin{aligned} x = \exp \Bigg (\sum _{i=1}^4 x_i Y_i^\infty \Bigg )(0),\quad x \in {\mathbb{R }}^4. \end{aligned}$$
(4.7)

Then, the vector fields \(Y_1^\infty ,\ldots ,Y_4^\infty \) are

$$\begin{aligned} Y^\infty _1&= \frac{\partial }{\partial x_1 } + q_1 \frac{\partial }{\partial x_3 }+ p_1 \frac{\partial }{\partial x_4 }, \nonumber \\ Y^\infty _2&= \frac{\partial }{\partial x_2 } + q_2 \frac{\partial }{\partial x_3 }+ p_2 \frac{\partial }{\partial x_4 }, \nonumber \\ Y^\infty _3&= \frac{1}{4} [Y_2^\infty , Y_1^\infty ]= \frac{\partial }{\partial x_3 } + p_3 \frac{\partial }{\partial x_4 }, \nonumber \\ Y^\infty _4&= \frac{1}{16} [Y_1^\infty ,[Y_1^\infty ,[Y_2^\infty ,[Y_2^\infty ,Y_1^\infty ]]]]= \frac{\partial }{\partial x_4 }. \end{aligned}$$
(4.8)

Above, \(q_i\) and \(p_i\) are polynomials such that \(q_i\circ \delta _\lambda =\lambda q_i\) and \(p_i\circ \delta _\lambda = \lambda ^4 p_i\) for \(i=1,2\) and \(\lambda >0\). In particular, for \(i=1,2\), we have

$$\begin{aligned} q_i = a_{i1} x_1 + a_{i2} x_2\quad \mathrm{and} \quad p_i = p_{i1} + x_3 p_{i2} + d_i x_3^2, \end{aligned}$$

where \(a_{ij},d_i\in {\mathbb{R }}\) are constants, \(p_{i1}\) are homogeneous polynomials of degree \(4\) in the variables \(x_1,x_2\), and \(p_{i2}\) are homogeneous polynomials of degree \(2\) in the variables \(x_1,x_2\). By the relation \(Y^\infty _3 = \frac{1}{4} [Y^\infty _2,Y_2^\infty ]\), we deduce that the polynomial \(p_3\) is

$$\begin{aligned} p_3 = \frac{1}{4} \Big \{ \Big ( \frac{\partial }{\partial x_2 } + q_2 \frac{\partial }{\partial x_3 } \Big ) p_1 -\Big ( \frac{\partial }{\partial x_1 } + q_1 \frac{\partial }{\partial x_3 } \Big ) p_2 \Big \}. \end{aligned}$$
(4.9)

If \(Y_1,\ldots ,Y_4\) satisfy (1.12), then the vector fields \(Y_1^\infty ,\ldots ,Y_4^\infty \) satisfy, for all \(\beta \in \mathcal{I }_5\) with \(\beta =(*,*,*,2,1)\),

$$\begin{aligned} Y_\beta ^{\infty }(0)&= 16 Y_4^\infty (0)\, \mathrm{for} \,\beta = (1,2,1,2,1) \,\text{ or }\, \beta =(1,1,2,2,1), \nonumber \\ Y_\beta ^\infty (0)&= -32 Y_4 ^{\infty }(0) \,\mathrm{for}\, \beta = (2,1,1,2,1),\, \text{ and } \nonumber \\ Y_{\beta }^{\infty } (0)&= 0 \,\mathrm{otherwise}. \end{aligned}$$
(4.10)

We claim that

$$\begin{aligned} Y_1^\infty = X_1\quad \mathrm{and}\quad Y_2^\infty = X_2, \end{aligned}$$
(4.11)

where \(X_1\) and \(X_2\) are the vector fields (1.1).

For any \(x\in {\mathbb{R }}^4\), let \(\gamma :[0,1]\rightarrow {\mathbb{R }}^4\) be the curve such that \({\dot{\gamma }} = \mathrm{exp}\big (\sum _{i=1}^4 x_ i Y^\infty _i(\gamma )\big ) (0)\) and \(\gamma (0)=0\). Condition (4.7) implies that \(\gamma (t) = tx\) for all \(t\in [0,1]\). Differentiating this identity at \(t=1\), we obtain

$$\begin{aligned} \sum _{i=1}^ 4 x_i \frac{\partial }{\partial x_i} = \sum _{i=1}^4 x_i Y^\infty _i (x), \quad x\in {\mathbb{R }}^4. \end{aligned}$$
(4.12)

The equation for the coordinate \(i=3\) in (4.12) is \(x_1 q_1+ x_2 q_2=0\), i.e.,

$$\begin{aligned} \sum _{i,j=1}^2 a_{ij} x_i x_j = 0. \end{aligned}$$

This equation along with the relation \( \frac{1}{4} [Y_2^\infty , Y_1^\infty ](0) = {\partial }/ {\partial x_3}\) gives

$$\begin{aligned} q_1 =2 x_1 \quad \mathrm{and}\quad q_2 =-2x_1. \end{aligned}$$
(4.13)

The equation for the coordinate \(i=4\) in (4.12) is \(x_1p_1 + x_2 p_2+ x_3 p_3=0\). By (4.13) and (4.9), this equation reads

$$\begin{aligned} x_1 p_1 + x_2 p_2 +\frac{x_3}{4} \Big ( \partial _2 p_1 - 2 x_1 \partial _3 p_1 - \partial _1 p_2 - 2 x_2 \partial _3 p_2\Big ) = 0, \end{aligned}$$

or equivalently,

$$\begin{aligned} x_1 p_{11} \!+\! x_2 p_{21} \!+\!\frac{x_3}{4} \Big ( 2 x_1 p_{12} \!+\! 2 x_2 p_{22} \!+\!\partial _2 p_{11} \!-\!\partial _1 p_{21}\Big ) +\frac{x_3^2}{4}\Big ( \partial _2 p_{12} -\partial _1 p_{22}\Big )=0.\nonumber \\ \end{aligned}$$
(4.14)

Setting to \(0\) the coefficients of the powers of \(x_3\) in (4.14), we obtain the following equations:

$$\begin{aligned}&x_1 p_{11} + x_2 p_{21} = 0, \end{aligned}$$
(4.15)
$$\begin{aligned}&2 x_1 p_{12} + 2 x_2 p_{22} + \partial _2 p_{11} - \partial _1 p_{21} = 0, \end{aligned}$$
(4.16)
$$\begin{aligned}&\partial _2 p_{12} -\partial _1 p_{22} = 0. \end{aligned}$$
(4.17)

Equation (4.15) implies that

$$\begin{aligned} p_{11} = x_2 p \quad \mathrm{and} \quad p_{21} = - x_1 p, \end{aligned}$$
(4.18)

where

$$\begin{aligned} p\, {\text{ is } \text{ a } \text{ homogeneous } \text{ polynomial } \text{ of } \text{ degree } \text{3 } \text{ in } \text{ the } \text{ variables }}\, x_1,x_2. \end{aligned}$$
(4.19)

Now, Eq. (4.16) reads

$$\begin{aligned} 5 p+ 2 x_1 p_{12} + 2x_2 p_{22}=0. \end{aligned}$$
(4.20)

The vector fields \(Y_1^\infty \) and \(Y_2^\infty \) are

$$\begin{aligned} Y_1^\infty = X_1 + w_1 \frac{\partial }{\partial x_4} \quad \mathrm{and} \quad Y_2^\infty = X_2 + w_2 \frac{\partial }{\partial x_4}, \end{aligned}$$

where \(X_1,X_2\) are the vector fields (1.1), and

$$\begin{aligned} w_1&= x_2 p + x_3 p_{12} + (d_1-1) x_3^2, \nonumber \\ w_2&= - x_1 p + x_3 p_{22} + d_2 x_3^2. \end{aligned}$$
(4.21)

We claim that \(w_1=w_2=0\). Once this claim is proved, the main claim (4.11) will follow.

We compute commutators of length \(5\) of \(Y_1^\infty \) and \(Y_2^\infty \), as functions of \(X_1,X_1,w_1\), and \(w_2\). First of all, we have

$$\begin{aligned}{}[Y_2^\infty , Y_1^\infty ] = [X_2,X_1] +( X_2 w_1- X_1 w_2) \frac{\partial }{\partial x_4}. \end{aligned}$$

Then, for any \(i,j,k=1,2\), we have

$$\begin{aligned}{}[Y_i^\infty ,[Y_j^\infty ,[Y_k^\infty ,[Y_2^\infty , Y_1^\infty ]]]]&= [X_i,[X_j,[X_k,[X_2,X_1]]]] \nonumber \\&\quad +\big ( X_i X_j X_k X_2 w_1 - X_iX_j X_k X_1 w_2 - 4 X_i X_j \partial _3 w_k \big ) \frac{\partial }{\partial x_4}.\nonumber \\ \end{aligned}$$
(4.22)

By (1.9) and (4.10), commutators (of length \(5\)) of \(Y_1^\infty \) and \(Y_2^\infty \), and commutators of \(X_1\) and \(X_2\) satisfy the same relations at the point \(x=0\). From (4.22), we deduce that we have for all \(i,j,k=1,2\)

$$\begin{aligned} X_i X_j X_k X_2 w_1 - X_iX_j X_k X_1 w_2 - 4 X_i X_j \partial _3 w_k=0, \quad \mathrm{at}\,x=0. \end{aligned}$$
(4.23)

Notice that \( X_i X_j X_k X_2 w_1 - X_iX_j X_k X_1 w_2 - 4 X_i X_j \partial _3 w_k\) is a homogeneous polynomial of degree \(0\) for the dilations \((x_1,x_2,x_3)\mapsto (\lambda x_1, \lambda x_2, \lambda ^2 x_3), \lambda >0\). Thus, Eq. (4.23) holds for all \(x\in {\mathbb{R }}^4\). By integration, we deduce that \( X_j X_k X_2 w_1 - X_j X_k X_1 w_2 - 4 X_j \partial _3 w_k\) is constant in \({\mathbb{R }}^3\). For \( X_j X_k X_2 w_1 - X_j X_k X_1 w_2 - 4 X_j \partial _3 w_k\) is a homogeneous polynomial of degree \(1\), we deduce that it is identically zero. Repeating the same argument, we conclude that for \(k=1,2\)

$$\begin{aligned} X_k X_2 w_1 - X_k X_1 w_2 - 4 \partial _3 w_k =0 \quad \mathrm{in}\, {\mathbb{R }}^{3}. \end{aligned}$$
(4.24)

In order to analyze Eq. (4.24), we preliminarily compute

$$\begin{aligned} X_2 w_1&= p + x_ 2 \partial _2 p +x _3 \partial _2 p_{12} - 2 x_1 p_{12} - 4 (d_1-1) x_1 x_3,\\ X_1 w_2&= - p - x_1 \partial _1 p + x_3 \partial _1 p_{22} + 2 x_ 2 p_{22} + 4 d_ 2 x_{2} x_{3}. \end{aligned}$$

When \(k=1\), Eq. (4.24) reads

$$\begin{aligned}&3 \partial _1 p + x_1 \partial _1\partial _1 p + x_2 \partial _1 \partial _2 p - 6p_{12} - 2x_1 \partial _1 p_{12}-2 x_2 \partial _1 p_{22}+ 2 x_2 (\partial _2 p_{12}-\partial _1 p_{22})+ \\&\quad - 8 ( d_1-1) x_1 x_2 - 8 d_2 x_2^2 + x_3\left( \partial _1\partial _2 p_{12} -\partial _1\partial _1 p_{22} - 12 (d_1-1)\right) = 0. \end{aligned}$$

As \(\partial _{1} p\) is a homogeneous polynomial of degree \(2\), we have \(x_1 \partial _1\partial _1 p + x_2 \partial _2 \partial _1 p = 2\partial _1 p\). Also using the fact that \(p_{12}\) is a homogeneous polynomial of degree 2, and using identity (4.17), the previous equation reads

$$\begin{aligned} 5 \partial _1 p - 10p_{12} - 8 ( d_1-1) x_1 x_2 - 8 d_2 x_2^2 - 12 (d_1-1)x_3 = 0. \end{aligned}$$

This implies

$$\begin{aligned} d_1-1 =0 \quad \mathrm{and} \quad \partial _1 p - 2p_{12} - \frac{8}{5} d_2 x_2^2 =0. \end{aligned}$$
(4.25)

When \(k=2\), Eq. (4.24) reads

$$\begin{aligned}&3\partial _2 p + x_1 \partial _2\partial _1 p + x_2 \partial _2\partial _2 p-6p_{22} -2x_1 \partial _2 p_{12} - 2 x_2 \partial _2 p_{22} + 2 x_1( \partial _1 p_{22} -\partial _2 p_{12})\\&\quad +8(d_1-1) x_1^2 + 8 d_2 x_1 x_2 +x_3(\partial _2 \partial _2 p_{12} -\partial _2 \partial _1 p_{22} - 12 d_2) =0. \end{aligned}$$

Using the fact that \(\partial _2 p\) is homogeneous of degree \(2\), identity (4.17), the fact that \(p_{22}\) is homogeneous of degree \(2\), and \(d_1=1\), the previous equation reads

$$\begin{aligned} 5 \partial _2 p -10 p_{22} + 8 d_2 x_1 x_2 -12 d_2 x_3=0. \end{aligned}$$

This implies \(d_2 =0\) and \( \partial _2 p -2 p_{22} =0\). We multiply the latter equation by \(x_2\), we multiply by \(x_1\) the second equation in (4.25) (with \(d_2=0\)), we sum the two resulting equations, and we obtain \(x_1\partial _1 p + x_2\partial _2 p -2x_1 p_{12}-2x_2 p_{22}=0\). Using the fact that \(p\) is homogeneous of degree \(3\), we finally obtain

$$\begin{aligned} 3 p - 2x_1 p_{12}-2x_2 p_{22}=0. \end{aligned}$$
(4.26)

From (4.26) and (4.20), we deduce that \(p=0\), and thus, \(p_{12}= p_{22}=0\). We conclude that \(w_1=w_2=0\), and this proves the main claim (4.11).

5 Proof of Theorem 1.3

In this section, we prove Theorem 1.3. As in Sect.  4, we assume that \(M= {\mathbb{R }}^4, p=0, \mathcal{D } =\mathrm{span}\{ Y_1, Y_2\}\) and that the vector fields \(Y_1,\ldots ,Y_4\) satisfy (H1)-(H3) and induce exponential coordinates of the first type in \({\mathbb{R }}^4\), as in (4.1). On \(\mathcal{D }\), we fix a metric \(g \) satisfying (1.13).

For \(\lambda >0\), let \(Y_1^\lambda \) and \(Y_2^\lambda \) be the vector fields in (4.3), and let \(\mathcal{D }^\lambda = \mathrm{span}\{Y_1^\lambda , Y_2^\lambda \}\). Let \(g^\lambda \) be the metric on \(\mathcal{D }^\lambda \) defined by

$$\begin{aligned} g^\lambda _x(Y_i^\lambda ,Y_j^\lambda ) = g_{\delta _{1/\lambda } (x)}(Y_i,Y_j),\quad x\in {\mathbb{R }}^4,\quad i,j=1,2. \end{aligned}$$

When \(\lambda =\infty \), we let \(\mathcal{D }^\infty =\mathrm{span}\{Y_1^\infty , Y_2^\infty \}\), where, by the discussion of Sect. 4, we have \(Y_1^\infty = X_1\) and \(Y_2^\infty = X_2\). On \(\mathcal{D }^\infty \), we define the metric \( g^\infty \)

$$\begin{aligned} g^\infty _x(Y_i^\infty ,Y_j^\infty ) = \lim _{\lambda \rightarrow \infty } g^\lambda _x(Y_i^\lambda , Y_j^\lambda ) = \lim _{\lambda \rightarrow \infty } g_{\delta _{1/\lambda }(x)}(Y_i, Y_j) = g_0(Y_i,Y_j),\quad x\in {\mathbb{R }}^4. \end{aligned}$$

Remark 5.1

From (1.13), it follows that

$$\begin{aligned} g^\infty _x (Y_1^\infty ,Y_2^\infty ) = g_0(Y_1,Y_2) = 0 ,\quad x\in {\mathbb{R }}^4. \end{aligned}$$

Without loss of generality, we also assume that \(g^\infty _x(Y_i^\infty ,Y_i^\infty )=1\) for all \(x\in {\mathbb{R }}^4\) and \(i=1,2\). In other words, the metric \(g^\infty \) makes the vector fields \(Y_1\) and \(Y_2\) orthonormal. It follows that linear mappings \(T:{\mathbb{R }}^4\rightarrow {\mathbb{R }}^4\) of the form

$$\begin{aligned} T(x) = \big ( U(x_1,x_2), \det (U) x_3, x_4\big ),\quad x=(x_1,\ldots ,x_4)\in {\mathbb{R }}^4, \end{aligned}$$

where \(U\in O(2)\) is an orthogonal mapping in \({\mathbb{R }}^2\), are isometries of \(({\mathbb{R }}^4,\mathcal{D }^\infty ,g^\infty )\).

Let \(\gamma :[-1,1]\rightarrow M={\mathbb{R }}^4\) be a \(\mathcal{D }\)-horizontal curve such that \(\gamma (0)=0\) and

$$\begin{aligned} {\dot{\gamma }}_L(0) \ne \dot{\gamma }_R(0), \end{aligned}$$
(5.1)

where \({\dot{\gamma }}_L(0)\ne 0\) and \(\dot{\gamma }_R(0)\ne 0\) are the left and right derivatives of \(\gamma \) at \(t=0\).

For \(\lambda >0\), the curves \(\gamma ^\lambda (t) = \delta _\lambda \gamma (t/\lambda ), t\in [-\lambda ,\lambda ]\), are \(\mathcal{D }^\lambda \)-horizontal. Moreover, the limit curve \(\gamma ^\infty :(-\infty ,\infty ) \rightarrow {\mathbb{R }}^4\)

$$\begin{aligned} \gamma ^\infty (t) = \left\{ \begin{array}{ll} \displaystyle \lim _{\lambda \rightarrow \infty } \gamma ^\lambda (t) = {\dot{\gamma }}_R(0) t &{} t\ge 0, \\ \displaystyle \lim _{\lambda \rightarrow \infty } \gamma ^\lambda (t) = {\dot{\gamma }}_L(0) t &{} t< 0, \end{array} \right. \end{aligned}$$
(5.2)

is \(\mathcal{D }^\infty \)-horizontal. Here, the vectors \({\dot{\gamma }}_L(0)\) and \({\dot{\gamma }}_R(0)\) are identified with vectors of \({\mathbb{R }}^4\) and have the form \((*,*,0,0)\). The curve \(\gamma ^\infty \) lies, therefore, in the \(x_1x_2\)-plane.

Proposition 5.1

Assume that the curve \(\gamma \) is length minimizing in \(({\mathbb{R }}^4,\mathcal{D }, g)\). Then, the curve \(\gamma ^\infty \) is length minimizing in \(({\mathbb{R }}^4,\mathcal{D }^\infty ,g^\infty )\).

Proposition 5.1 is proved in [2], Proposition 2.4.

We conclude the proof of Theorem 1.3. Assume by contradiction that there exists a curve \(\gamma \) as in Theorem 1.3 that is length minimizing in \(({\mathbb{R }}^4,\mathcal{D }, g)\). By Proposition 5.1, the curve \(\gamma ^\infty \) in (5.2) is length minimizing in \(({\mathbb{R }}^4, \mathcal{D }^\infty ,g^\infty )\). By assumption, we have \(g_0({\dot{\gamma }}_L(0),{\dot{\gamma }}_R(0))\ne 0\). It follows that

$$\begin{aligned} g^\infty _0({\dot{\gamma }}_L^\infty (0),{\dot{\gamma }}_R^\infty (0)) = \lim _{\lambda \rightarrow \infty } g_0^\lambda ({\dot{\gamma }}^\lambda _L(0),{\dot{\gamma }}^\lambda _R(0)) = g_0({\dot{\gamma }}_L(0),{\dot{\gamma }}_R(0))\ne 0. \end{aligned}$$
(5.3)

By Remark 5.1 and (5.1), we can assume that \(\gamma ^\infty \) has the form (1.5) with \(\alpha >0\) (up to re-parameterization), and by (5.3), we have \(\alpha \ne 1\). By Theorem 1.2, the curve \(\gamma ^\infty \) is not length minimizing. This is a contradiction.