1 Introduction

A Diophantine n-tuple is a set \( \left\{ a_1,\ldots ,a_n \right\} \) consisting of n positive integers with the property that \( a_i a_j + 1 \) is a perfect square for all \( 1 \le i < j \le n \). More generally, for a given integer m, a D(m) -n-tuple is a set \( \left\{ a_1,\ldots ,a_n \right\} \) consisting of n positive integers with the property that \( a_i a_j + m \) is a perfect square for all \( 1 \le i<j \le n \). For \( m = 1 \) we get the above defined Diophantine n-tuples. The study of Diophantine tuples started more than two thousand years ago and is well known in the meantime. Early contributions are due to Diophantus of Alexandria, Fermat and Euler. Later, many others contributed to this beautiful subject. The original result of Diophantus is that \( \left\{ 1,33,68,105 \right\} \) is a D(256) -quadruple. The first D(1) -quadruple, the set \( \left\{ 1,3,8,120 \right\} \), is due to Fermat. The question usually asked is, for which n do there exist any Diophantine n-tuples, and if one exists, whether there are finitely or infinitely many. It was proved recently (cf. [19]) that \( n \le 4 \) for a D(1) -n-tuple and this bound is sharp. By Euler, there are even infinitely many Diophantine quadruples. For arbitrary m the situation is much less clear. Dujella proved in [5, 6] an upper bound for n depending (logarithmically) on m. More precise results are only available for special values of m, e.g. for D(4) -tuples (cf. [1]) or \( D(-1) \)-tuples (cf. [2]).

One can consider many other variants of Diophantine tuples, e.g. by considering algebraic integers or polynomials instead of integers, higher or perfect powers instead of squares, etc.. For a summary about Diophantine tuples and its variants we refer to [4].

In what follows we shall look at complex polynomials. Let \( p \in \mathbb {C}[X] \) be given. A D(p) -n-tuple in \( \mathbb {C}[X] \) is defined as a set \( \left\{ a_1,\ldots ,a_n \right\} \) consisting of n non-zero complex polynomials such that \( a_i a_j + p \) is a perfect square for all \( 1 \le i<j \le n \) and such that there is no polynomial \( q \in \mathbb {C}[X] \) with \( p/q^2\), \(a_1/q\), ..., \(a_n/q \) are constant polynomials. Again, one is interested in the same questions as above. In [7] it is proven that \( n \le 10 \) for a D(1) -n-tuple in \( \mathbb {C}[X] \). This bound is reduced to 7 in [10]. It is not clear what the expected true upper bound is. For arbitrary p there is no upper bound known so far. This changes if one looks e.g. at linear polynomials p, see [11], or at integer polynomials, cf. for example [5, 8, 9, 20].

Since the sequence of squares can be written as a linear recurrence sequence, one can ask the questions about existence of Diophantine tuples and finiteness of their number not only in the case of squares but also if we restrict \( a_i a_j + 1 \) to take values in an arbitrary fixed linear recurrence sequence. This situation was considered in [18] for the first time. For instance, in [15], which is based on [14] and [16], the authors consider sets \( \left\{ a,b,c \right\} \) of positive integers satisfying \( 1< a< b < c \) such that \( ab+1, ac+1, bc+1 \) are values in a linear recurrence sequence of Pisot type with Binet formula \( G_n=f_1\alpha _1^n + \cdots + f_k\alpha _k^n \). They prove three independent conditions under which there are only finitely many such sets. One of these conditions allows neither \( f_1 \) nor \( f_1 \alpha _1 \) to be a perfect square. This condition will be used in our statement, too. To give an idea about the origin of the conditions appearing in the statement (and in particular the condition that we also use), we very roughly sketch the argument. It starts by assuming on the contrary that there are infinitely many sets \( \left\{ a,b,c \right\} \) such that \( ab+1 = G_x, ac+1 = G_y, bc+1 = G_z \). It follows that \( \min \left( x,y,z \right) \) goes to infinity while xyz are all of the same size. On the other hand we have infinitely many squares in the multi-recurrence \( ((G_y-1)(G_z-1)/(G_x-1)) \). Using the Subspace theorem-approach, this is only possible if a functional identity in terms of xyz holds (compare e.g. with Theorem 2 in [3]). Comparing “leading terms” gives a contradiction unless one of the assumptions is satisfied. If neither \( f_1 \) nor \( f_1 \alpha _1 \), which are the possible leadings coefficients in the above multi-recurrence, are perfect squares, we get a contradiction at once (see again Theorem 2 in [3]). For large k one has much freedom in the functional identity coming from the irreducibility of the characteristic polynomial so that other arguments work. This last argument cannot be transferred to our situation as we will see below.

In the present paper we consider a function field variant of Diophantine tuples taking values in recurrences. Namely, we study sets \( \left\{ a,b,c \right\} \) consisting of three non-zero complex polynomials with the property that \( ab+1, ac+1, bc+1 \) are elements of a given linear recurrence sequence of polynomials. It will be proven that under some conditions on the recurrence sequence there are only finitely many such sets. In fact we are going to prove even more: the same result still holds if we add an arbitrary fixed polynomial p instead of 1. We mention that the same holds true if we replace \( \mathbb {C}\) by an arbitrary algebraically closed field of characteristic 0.

2 Results

For basic concepts and notions related to linear recurrences we refer to [24]. We call \( (G_n)_{n=0}^{\infty } \) a polynomial power sum if it is a simple linear recurrence sequence of complex polynomials with power sum representation

$$\begin{aligned} G_n = f_1\alpha _1^n + \cdots + f_k\alpha _k^n \end{aligned}$$
(2.1)

such that \( \alpha _1, \ldots , \alpha _k \in \mathbb {C}[X] \) are polynomials and \( f_1,\ldots ,f_k \in \mathbb {C}(X) \). Our main theorem that we are going to prove in this paper is the following statement:

Theorem 2.1

Let \( (G_n)_{n=0}^{\infty } \) be a polynomial power sum given as in (2.1). Assume either that the order of this sequence is \( k \ge 3 \) and the dominant root condition \( \deg \alpha _1> \deg \alpha _2 > \deg \alpha _3 \ge \deg \alpha _4 \ge \cdots \ge \deg \alpha _k \) is fulfilled, or that the order is \( k = 2 \) and we have \( \deg \alpha _1> \deg \alpha _2 > 0 \). Moreover, let \( p \in \mathbb {C}[X] \) be a given polynomial. If neither \( f_1 \) nor \( f_1 \alpha _1 \) is a square in \( \mathbb {C}(X) \), then there are only finitely many sets \( \left\{ a,b,c \right\} \) consisting of three non-zero polynomials such that \( ab+p\), \(ac+p\), \(bc+p \) are all elements of \( (G_n)_{n=0}^{\infty } \).

Note that we do not require that the linear recurrence sequence is non-degenerate, i.e. there may be two indices i and j such that \( \alpha _i \) and \( \alpha _j \) differ only be a constant factor. The only non-degeneracy properties we need are given by the dominant root condition.

Moreover, the dominant root condition is really necessary in this situation. If we omit the dominant root condition, then the statement of the theorem does not hold any more as the following example illustrates: Let \( A_n\), \( B_n\), \( C_n \) be three linear recurrence sequences of polynomials. Then the products \( A_nB_n\), \(A_nC_n\), \(B_nC_n \) are also linear recurrence sequences of polynomials. Let \( A_nB_n = s_1 \sigma _1^n + \cdots + s_k \sigma _k^n \) be the Binet representation. Consider now a new linear recurrence sequence \( D_n \) generated from \( A_nB_n \) by the following procedure: Replace each summand \( s_i \sigma _i^n \) with the term

$$\begin{aligned} \frac{1}{3} s_i \sigma _i^n + \frac{1}{3} s_i (\zeta _3 \sigma _i)^n + \frac{1}{3} s_i (\zeta _3^2 \sigma _i)^n \end{aligned}$$

where \( \zeta _3 \) is a primitive third root of unity. This new sequence \( D_n \) has for indices of the shape 3u the same values as \( A_nB_n \) and is zero otherwise. In other words \( D_{3u} = A_{3u}B_{3u} \) and \( D_{3u+1} = 0 = D_{3u+2} \). In the same manner we construct linear recurrence sequences \( E_n \) and \( F_n \) such that \( E_{3u+1} = A_{3u}C_{3u} \) and \( E_{3u} = 0 = E_{3u+2} \) as well as \( F_{3u+2} = B_{3u}C_{3u} \) and \( F_{3u} = 0 = F_{3u+1} \). Last but not least we define the sequence \( G_n := D_n + E_n + F_n + 1 \). Thus we have

$$\begin{aligned} G_n = {\left\{ \begin{array}{ll} A_{3u}B_{3u} + 1 &{}\text { if } n=3u \\ A_{3u}C_{3u} + 1 &{}\text { if } n=3u+1 \\ B_{3u}C_{3u} + 1 &{}\text { if } n=3u+2 \end{array}\right. }. \end{aligned}$$

First note that \( G_n \) has no dominant root since by the construction above for each characteristic root (possibly except 1) there are two other ones which differ only by the factor \( \zeta _3 \) and \( \zeta _3^2 \), respectively, and therefore have the same degree. If we choose the simple sequences \( A_n,B_n,C_n \) in such a way that all characteristic roots are squares in \( \mathbb {C}[X] \), then all characteristic roots of \( G_n \) are squares in \( \mathbb {C}[X] \) as well. Thus, writing \( G_n \) in the form (2.1), \( f_1 \) is a square in \( \mathbb {C}(X) \) if and only if \( f_1 \alpha _1 \) is a square in \( \mathbb {C}(X) \). Moreover, choose \( A_n,B_n,C_n \) such that all occurring characteristic roots are pairwise distinct, non-constant and have pairwise no common root. Furthermore, let all coefficients in the Binet-formulas of \( A_n,B_n,C_n \) be non-constant polynomials without multiple roots. Assume that these coefficients are pairwise distinct and have pairwise no common root. Additionally, no complex number \( z \in \mathbb {C}\) should be a root of both, an arbitrary root and an arbitrary coefficient. According to this construction no coefficient of a non-constant characteristic root of \( G_n \) is a square in \( \mathbb {C}(X) \). Nevertheless, there are obviously infinitely many sets \( \left\{ a,b,c \right\} \) such that \( ab+1, ac+1, bc+1 \) are all elements of \( (G_n)_{n=0}^{\infty } \). For instance, one can choose

$$\begin{aligned} A_n&= (x-1) \left( (x-2)^4 \right) ^n + (x-3) \left( (x-4)^2 \right) ^n, \\ B_n&= (x-5) \left( (x-6)^4 \right) ^n + (x-7) \left( (x-8)^2 \right) ^n, \\ C_n&= (x-9) \left( (x-10)^4 \right) ^n + (x-11) \left( (x-12)^2 \right) ^n \end{aligned}$$

to get an explicit example of the above described generic construction.

We remark that our preliminary assumptions are somewhat the opposite of those in [15] since there the characteristic polynomial is irreducible whereas in our case the characteristic polynomial splits in linear factors over the ground field. Concerning the conclusion, Theorem 2.1 can be seen as a function field analogue of Theorem 2 in [15]. We are unable to prove the result under the condition of Theorem 3 in [15], i.e. for any large k.

Our result is ineffective in the sense that our method of proof does neither produce an upper bound for the number of solutions nor gives a method to actually locate them.

Finally, we mention that we follow very closely the method of proof and the arguments used in [15]. The application of the Subspace theorem is replaced by an application of Proposition 3.2 below. A difference is that we work directly with the identity \( (G_x-1)(G_y-1)(G_z-1) = (abc)^2 \), similarly to [18], which better resembles the dominant root situation in our case. Moreover, another simplification appears since we do not have to parametrize xyz with one parameter to identify “leading terms”. Finally, the contradiction follows at once since the coefficients in our functional identity are (complex) constants.

3 Preliminaries

We start with a few quick remarks on the height in function fields in one variable over \( \mathbb {C}\), i.e. finite extensions of the rational function field \( \mathbb {C}(X) \). As such function fields correspond uniquely to regular projective irreducible algebraic curves, it is no surprise that the genus will play a role in the proposition below.

For the convenience of the reader we give a short wrap-up of the notion of valuations and of the height that can e.g. also be found in [12] and [13]. As references for the theory of heights and for other number theoretic aspects in function fields we refer to [21, 22] or [23].

For \( c \in \mathbb {C}\) and \( f(X) \in \mathbb {C}(X) \) denote by \( \nu _c(f) \) the unique integer such that \( f(X) = (X-c)^{\nu _c(f)} p(X) / q(X) \) with \( p(X),q(X) \in \mathbb {C}[X] \) such that \( p(c)q(c) \ne 0 \). Further denote by \( \nu _{\infty }(f) = \deg q - \deg p \) if \( f(X) = p(X) / q(X) \). These functions \( \nu \) are up to equivalence all valuations in \( \mathbb {C}(X) \). If \( \nu _c(f) > 0 \), then c is called a zero of f, and if \( \nu _c(f) < 0 \), then c is called a pole of f. For a finite extension F of \( \mathbb {C}(X) \) each valuation in \( \mathbb {C}(X) \) can be extended to no more (up to equivalence) than \( [F : \mathbb {C}(X)] \) valuations in F. This again gives up to equivalence all valuations in F. Both, in \( \mathbb {C}(X) \) as well as in F the sum-formula

$$\begin{aligned} \sum _{\nu } \nu (f) = 0 \end{aligned}$$

holds, where \( \sum _{\nu } \) means that the sum is taken over all valuations up to equivalence (i.e. over the above defined functions) in the considered function field. Each valuation (again up to equivalence) in a function field corresponds to a place and vice versa. The places can be thought as the equivalence classes of valuations. Moreover, we write \( \deg f = -\nu _{\infty }(f) \) for all \( f(X) \in \mathbb {C}(X) \).

The proof in the next section will make use of height functions in function fields. Let us therefore define the height of an element \( f \in F^* \) by

$$\begin{aligned} \mathcal {H}(f) := - \sum _{\nu } \min \left( 0, \nu (f) \right) = \sum _{\nu } \max \left( 0, \nu (f) \right) \end{aligned}$$

where the sum is taken over all valuations (up to equivalence) on the function field \( F/\mathbb {C}\). Additionally we define \( \mathcal {H}(0) = \infty \). This height function satisfies some basic properties that are listed in the lemma below, which is proven in [17]:

Lemma 3.1

Denote as above by \( \mathcal {H}\) the height on \( F/\mathbb {C}\). Then for \( f,g \in F^* \) the following properties hold:

  1. (a)

    \( \mathcal {H}(f) \ge 0 \) and \( \mathcal {H}(f) = \mathcal {H}(1/f) \),

  2. (b)

    \( \mathcal {H}(f) - \mathcal {H}(g) \le \mathcal {H}(f+g) \le \mathcal {H}(f) + \mathcal {H}(g) \),

  3. (c)

    \( \mathcal {H}(f) - \mathcal {H}(g) \le \mathcal {H}(fg) \le \mathcal {H}(f) + \mathcal {H}(g) \),

  4. (d)

    \( \mathcal {H}(f^n) = \left| n \right| \cdot \mathcal {H}(f) \),

  5. (e)

    \( \mathcal {H}(f) = 0 \iff f \in \mathbb {C}^* \),

  6. (f)

    \( \mathcal {H}(A(f)) = \deg A \cdot \mathcal {H}(f) \) for any \( A \in \mathbb {C}[T] {\setminus } \left\{ 0 \right\} \).

When proving our theorem, we will use the following function field analogue of the Schmidt subspace theorem. A proof for this proposition can be found in [25]:

Proposition 3.2

(Zannier) Let \( F/\mathbb {C}\) be a function field in one variable, of genus \( \mathfrak {g}\), let \( \varphi _1, \ldots , \varphi _n \in F \) be linearly independent over \( \mathbb {C}\) and let \( r \in \left\{ 0,1, \ldots , n \right\} \). Let S be a finite set of places of F containing all the poles of \( \varphi _1, \ldots , \varphi _n \) and all the zeros of \( \varphi _1, \ldots , \varphi _r \). Put \( \sigma = \sum _{i=1}^{n} \varphi _i \). Then

$$\begin{aligned} \sum _{\nu \in S} \left( \nu (\sigma ) - \min _{i=1, \ldots , n} \nu (\varphi _i) \right) \le \begin{pmatrix}{n}\\ {2}\end{pmatrix} (\left| S \right| + 2\mathfrak {g}- 2) + \sum _{i=r+1}^{n} \mathcal {H}(\varphi _i). \end{aligned}$$

4 Proof

In this section we will prove Theorem 2.1. Before we begin the proof, let us remark that we will use some ideas of [15] that are quite useful also in our situation. We note that, since some arguments are easier in the function field setting we are in and the proof thus is not too long, it is feasible to write down the proof without interruption.

Proof of Theorem 2.1

We are going to prove the statement indirectly and assume therefore that there are infinitely many sets with the required properties. First note that without loss of generality we can assume that for a still infinite set of sets the inequality \( \deg a \le \deg b \le \deg c \) holds. For any such triple exist non-negative integers xyz such that

$$\begin{aligned} ab+p = G_x, \qquad ac+p = G_y, \qquad bc+p = G_z. \end{aligned}$$
(4.1)

If all three parameters xyz were bounded by a constant, then there can be only finitely many sets \( \left\{ a,b,c \right\} \). Hence it must hold that \( \max \left\{ x,y,z \right\} \rightarrow \infty \).

Since \( \alpha _1 \) is the dominant root, for large enough n, i.e. for \( n \ge n_0 \), the degree satisfies the (in)equality

$$\begin{aligned} \deg G_n = \deg f_1 + n \deg \alpha _1 > \deg p. \end{aligned}$$
(4.2)

Now, as \( \max \left\{ x,y,z \right\} \rightarrow \infty \), we get \( \max \left\{ \deg G_x, \deg G_y, \deg G_z \right\} \rightarrow \infty \). Thus \( \deg c \rightarrow \infty \), and consequently we have \( z \rightarrow \infty \) as well as \( y \rightarrow \infty \). Therefore for a still infinite subset of sets we can assume that both, z and y, are not smaller than \( n_0 \), which implies that Eq. (4.2) is applicable.

Recalling the ordering \( \deg a \le \deg b \le \deg c \), we get \( \deg G_x \le \deg G_y \le \deg G_z \). Now again using (4.2) and taking into account that abc are pairwise distinct yields \( x< y < z \).

Let us for the moment assume that x is bounded by a constant. So \( G_x \) is the same fixed value for infinitely many sets. This implies that a and b are fixed for infinitely many sets. Therefore

$$\begin{aligned} bG_y - aG_z = abc + bp - abc - ap = (b-a)p \end{aligned}$$

is constant for infinitely many sets. Dividing by \( f_1 \) and using the power sum representation of the recurrence sequence, we can rewrite this equation in the form

$$\begin{aligned} (b - a\alpha _1^{z-y}) \alpha _1^y = - b\frac{f_2}{f_1} \alpha _2^y - \cdots - b\frac{f_k}{f_1} \alpha _k^y + a\frac{f_2}{f_1} \alpha _2^z + \cdots + a\frac{f_k}{f_1} \alpha _k^z + \frac{(b-a)p}{f_1}. \end{aligned}$$
(4.3)

If the left hand side of Eq. (4.3) is non-zero, then it has degree at least \( y \deg \alpha _1 \). However, the degree of the right hand side is at most \( C_0 + z \deg \alpha _2 \), where \( C_0 \) is a constant. Since \( \deg G_y = \deg a + \deg c \) and \( \deg G_z = \deg b + \deg c \) the equation

$$\begin{aligned} (z-y) \deg \alpha _1 = \deg G_z - \deg G_y = \deg b - \deg a \end{aligned}$$

is satisfied which implies that \( \rho := z-y \) is constant. Consequently, the degree of the right hand side of Eq. (4.3) is at most \( C_1 + y \deg \alpha _2 \) for a new constant \( C_1 \). The only way this can work is that both sides of Eq. (4.3) are zero. Therefore, \( b = a \alpha _1^{\rho } \). Considering the equation

$$\begin{aligned} (b-a)p = bG_y - aG_z = b \left( f_1\alpha _1^y + \cdots + f_k\alpha _k^y \right) - a \left( f_1\alpha _1^{y+\rho } + \cdots + f_k\alpha _k^{y+\rho } \right) , \end{aligned}$$

dividing this by a and replacing b by \( a \alpha _1^{\rho } \) yields

$$\begin{aligned} (\alpha _1^{\rho } - 1) p = \alpha _1^{\rho } \left( f_2\alpha _2^y + \cdots + f_k\alpha _k^y \right) - \left( f_2\alpha _2^{y+\rho } + \cdots + f_k\alpha _k^{y+\rho } \right) . \end{aligned}$$
(4.4)

The left hand side of Eq. (4.4) has constant degree whereas the degree of the right hand side is \( \rho \deg \alpha _1 + \deg f_2 + y \deg \alpha _2 \). This is a contradiction for large y, implying that x cannot be bounded by a constant.

Overall, there is a still infinite set of sets such that all three indices xyz are always greater than an arbitrary fixed constant. In particular, we can assume that no index is smaller than \( n_0 \).

We have already mentioned above, that \( x< y < z \). In the next step it will be proven that z cannot grow much faster than x. For doing so let

$$\begin{aligned} g := \gcd \left( G_y-p, G_z-p \right) \end{aligned}$$

be the greatest common divisor of these two polynomials. Now we distinguish between two cases. Firstly, assume \( y \le \kappa z \), where \( \kappa \) is a rational number in the interval (0, 1) which will be determined later. It holds that

$$\begin{aligned} \deg g&\le \deg (G_y-p) = \deg f_1 + y \deg \alpha _1 \\&\le \deg f_1 + \kappa z \deg \alpha _1 \le C_2 + \kappa z \deg \alpha _1. \end{aligned}$$

Secondly, we assume in the other case that \( y > \kappa z \). Thus we have \( z-y < z - \kappa z = (1-\kappa ) z \). By the definition of g as greatest common divisor, it immediately follows that g is also a divisor of \( (G_z-p) - \alpha _1^{z-y} (G_y-p) \) and therefore

$$\begin{aligned} \deg g&\le \deg \left( (G_z-p) - \alpha _1^{z-y} (G_y-p) \right) \\&= \deg \left( (f_2 \alpha _2^z + \cdots + f_k \alpha _k^z - p) - (f_2 \alpha _1^{z-y} \alpha _2^y + \cdots + f_k \alpha _1^{z-y} \alpha _k^y - \alpha _1^{z-y} p) \right) \\&= \deg f_2 + (z-y) \deg \alpha _1 + y \deg \alpha _2 \\&\le \deg f_2 + (z-y) \deg \alpha _1 + y (\deg \alpha _1 - 1) \\&= C_3 + z \deg \alpha _1 - y < C_3 + z \deg \alpha _1 - \kappa z = C_3 + z (\deg \alpha _1 - \kappa ). \end{aligned}$$

We want to choose \( \kappa \) in such a way that \( \kappa \deg \alpha _1 = \deg \alpha _1 - \kappa \). Hence we set

$$\begin{aligned} \kappa = \frac{\deg \alpha _1}{1 + \deg \alpha _1} \end{aligned}$$

which yields in both of our cases

$$\begin{aligned} \deg g \le C_4 + z \kappa \deg \alpha _1. \end{aligned}$$
(4.5)

If we denote \( \widetilde{g} = \gcd \left( G_x-p, G_z-p \right) \), then \( G_z-p \) is a divisor of \( g \widetilde{g} \) since \( c \mid g \), \( b \mid \widetilde{g} \) and \( G_z-p = bc \). This gives us

$$\begin{aligned} \deg f_1 + x \deg \alpha _1&= \deg (f_1 \alpha _1^x) = \deg G_x = \deg (G_x - p) \ge \deg \widetilde{g} \\&\ge \deg (G_z-p) - \deg g = \deg f_1 + z \deg \alpha _1 - \deg g \end{aligned}$$

as well as by using inequality (4.5) that

$$\begin{aligned} x \deg \alpha _1 \ge z \deg \alpha _1 - \deg g \ge z \deg \alpha _1 - C_4 - z \kappa \deg \alpha _1 \end{aligned}$$

and

$$\begin{aligned} x \ge z - z \kappa - \frac{C_4}{\deg \alpha _1} = (1-\kappa ) z - C_5 > C_6 z. \end{aligned}$$
(4.6)

Thus z is bounded above by \( x/C_6 \). This fact means that the three indices grow with a similar rate.

Combining the three equations in (4.1), we can express the polynomials abc by elements of the linear recurrence sequence \( (G_n)_{n=0}^{\infty } \) in the following way:

$$\begin{aligned} a&= \frac{\sqrt{G_x-p} \cdot \sqrt{G_y-p}}{\sqrt{G_z-p}}, \\ b&= \frac{\sqrt{G_x-p} \cdot \sqrt{G_z-p}}{\sqrt{G_y-p}}, \\ c&= \frac{\sqrt{G_y-p} \cdot \sqrt{G_z-p}}{\sqrt{G_x-p}}. \end{aligned}$$

By using the square root symbol in an equation we mean that the equation holds for a suitable choose of one of the two possible square roots, which differ only by the factor \( -1 \). This choose can vary from one equation to another.

For this reason, we aim for rewriting the expression \( \sqrt{G_n-p} \) in a more suitable manner. This will be done by applying the (formal) multinomial series expansion to the power sum representation of our recurrence sequence:

$$\begin{aligned} \sqrt{G_n-p}&= \sqrt{f_1\alpha _1^n + \cdots + f_k\alpha _k^n - p} \\&= \sqrt{f_1} \alpha _1^{n/2} \sqrt{1 + \frac{f_2}{f_1} \left( \frac{\alpha _2}{\alpha _1} \right) ^n + \cdots + \frac{f_k}{f_1} \left( \frac{\alpha _k}{\alpha _1} \right) ^n - \frac{p}{f_1} \left( \frac{1}{\alpha _1} \right) ^n} \\&= \sqrt{f_1} \alpha _1^{n/2} \sum _{h_1,\ldots ,h_k=0}^{\infty } \gamma _{h_1,\ldots ,h_k} \left( \frac{-p}{f_1} \right) ^{h_1} \left( \frac{1}{\alpha _1} \right) ^{nh_1} \left( \prod _{i=2}^{k} \left( \frac{f_i}{f_1} \right) ^{h_i} \left( \frac{\alpha _i}{\alpha _1} \right) ^{nh_i} \right) \\&= \sqrt{f_1} \alpha _1^{n/2} \sum _{h_1,\ldots ,h_k=0}^{\infty } t_{h_1,\ldots ,h_k}^{(n)} = \sqrt{f_1} \alpha _1^{n/2} \sum _{\underline{h}=0}^{\infty } t_{\underline{h}}^{(n)} \end{aligned}$$

where we use the notation \( \underline{h} = (h_1,\ldots ,h_k) \) and

$$\begin{aligned} t_{h_1,\ldots ,h_k}^{(n)} = \gamma _{h_1,\ldots ,h_k} \left( \frac{-p}{f_1} \right) ^{h_1} \left( \frac{1}{\alpha _1} \right) ^{nh_1} \left( \prod _{i=2}^{k} \left( \frac{f_i}{f_1} \right) ^{h_i} \left( \frac{\alpha _i}{\alpha _1} \right) ^{nh_i} \right) . \end{aligned}$$

The next step is now to calculate a lower bound for the valuation \( \nu _{\infty } \) of the quantity above, which we will need later on. Since \( \gamma _{h_1,\ldots ,h_k} \in \mathbb {C}\), we get

$$\begin{aligned} \nu _{\infty } \left( t_{\underline{h}}^{(n)} \right)&= h_1 \left( n \deg \alpha _1 + \deg \frac{f_1}{p} \right) + \sum _{i=2}^{k} h_i \left( n (\deg \alpha _1 - \deg \alpha _i) + \deg \frac{f_1}{f_i} \right) \\&\ge h_1 (n + C_7) + \sum _{i=2}^{k} h_i (n + C_7) = \left( \sum _{i=1}^{k} h_i \right) (n + C_7) \end{aligned}$$

where \( C_7 = \min \left\{ \deg f_1 - \deg f_2, \ldots , \deg f_1 - \deg f_k, \deg f_1 - \deg p \right\} \). Note that for our purpose we can assume \( n + C_7 > 0 \).

Combining the representations of abc and \( \sqrt{G_n-p} \) we have so far, yields the following representation of the product abc of the elements in a triple:

$$\begin{aligned} abc&= \sqrt{G_x-p} \sqrt{G_y-p} \sqrt{G_z-p} \\&= f_1^{3/2} \alpha _1^{(x+y+z)/2} \sum _{\underline{h}^{(x)}=0}^{\infty } t_{\underline{h}^{(x)}}^{(x)} \sum _{\underline{h}^{(y)}=0}^{\infty } t_{\underline{h}^{(y)}}^{(y)} \sum _{\underline{h}^{(z)}=0}^{\infty } t_{\underline{h}^{(z)}}^{(z)} \\&= f_1^{3/2} \alpha _1^{(x+y+z)/2} \sum _{\underline{h}^{(x)}, \underline{h}^{(y)}, \underline{h}^{(z)}=0}^{\infty } t_{\underline{h}^{(x)}}^{(x)} t_{\underline{h}^{(y)}}^{(y)} t_{\underline{h}^{(z)}}^{(z)}. \end{aligned}$$

For the valuation we get the lower bound

$$\begin{aligned} \nu _{\infty } \left( t_{\underline{h}^{(x)}}^{(x)} t_{\underline{h}^{(y)}}^{(y)} t_{\underline{h}^{(z)}}^{(z)} \right)&= \nu _{\infty } \left( t_{\underline{h}^{(x)}}^{(x)} \right) + \nu _{\infty } \left( t_{\underline{h}^{(y)}}^{(y)} \right) + \nu _{\infty } \left( t_{\underline{h}^{(z)}}^{(z)} \right) \\&\ge \left( \sum _{i=1}^{k} h_i^{(x)} + \sum _{i=1}^{k} h_i^{(y)} + \sum _{i=1}^{k} h_i^{(z)} \right) (x + C_7). \end{aligned}$$

Let \( J > 0 \) be a number to be fixed later. Then there exists a natural number L, depending on J, such that

$$\begin{aligned} abc= & {} f_1^{3/2} \alpha _1^{(x+y+z)/2} t_1^{(x,y,z)} + \cdots + f_1^{3/2} \alpha _1^{(x+y+z)/2} t_L^{(x,y,z)} \\&+ \sum _{\nu _{\infty } \left( t_{\underline{h}^{(x)}}^{(x)} t_{\underline{h}^{(y)}}^{(y)} t_{\underline{h}^{(z)}}^{(z)} \right) \ge J(x + C_7)} f_1^{3/2} \alpha _1^{(x+y+z)/2} t_{\underline{h}^{(x)}}^{(x)} t_{\underline{h}^{(y)}}^{(y)} t_{\underline{h}^{(z)}}^{(z)}. \end{aligned}$$

Define \( \varphi _0 = abc \) as well as

$$\begin{aligned} \varphi _j = - f_1^{3/2} \alpha _1^{(x+y+z)/2} t_j^{(x,y,z)} \end{aligned}$$

for \( j=1,\ldots ,L \). Furthermore, put

$$\begin{aligned} \sigma = \sum _{j=0}^{L} \varphi _j. \end{aligned}$$

Let S be a finite set of places of \( F = \mathbb {C}(X, \sqrt{\alpha _1}, \sqrt{f_1}) \) containing all places lying over the zeros of \( \alpha _1,\ldots ,\alpha _k \), the zeros and poles of \( f_1,\ldots ,f_k \), the zeros of p, and all infinite places. Note that F contains always both possible values of the square roots if we require that one of them is contained since they differ only by the factor \( -1 \). By applying Proposition 3.2, if \( \varphi _0, \ldots , \varphi _L \) are linearly independent over \( \mathbb {C}\), we get the inequality

$$\begin{aligned} \nu _{\infty } (\sigma ) - \min _{i=0, \ldots , L} \nu _{\infty } (\varphi _i)&\le C_8 + \mathcal {H}(abc) \\&\le C_8 + C_9 \deg (abc) \\&= C_8 + \frac{C_9}{2} \deg (a^2b^2c^2) \\&= C_8 + \frac{C_9}{2} \deg ((G_x-p)(G_y-p)(G_z-p)) \\&= C_8 + \frac{C_9}{2} (3 \deg f_1 + (x+y+z) \deg \alpha _1) \\&\le C_{10} + C_{11} z \deg \alpha _1. \end{aligned}$$

In order to get also a lower bound for this expression we look at

$$\begin{aligned} \min _{i=0, \ldots , L} \nu _{\infty } (\varphi _i) \le \nu _{\infty } (abc) = - \deg (abc) = - \frac{3}{2} \deg f_1 - \frac{x+y+z}{2} \deg \alpha _1, \end{aligned}$$

which gives us

$$\begin{aligned} \nu _{\infty } (\sigma ) - \min _{i=0, \ldots , L} \nu _{\infty } (\varphi _i)&\ge J (x+C_7) + \nu _{\infty } \left( f_1^{3/2} \alpha _1^{(x+y+z)/2} \right) - \min _{i=0, \ldots , L} \nu _{\infty } (\varphi _i) \\&\ge J (x+C_7). \end{aligned}$$

Now, recalling inequality (4.6), we compare the lower with the upper bound to get

$$\begin{aligned} J (x+C_7) \le C_{10} + C_{11} z \deg \alpha _1 \le C_{10} + \frac{C_{11}}{C_6} x \deg \alpha _1. \end{aligned}$$

Therefore we set \( J := 1 + \frac{C_{11}}{C_6} \deg \alpha _1 \) and note that the right hand side does not depend on J. Plugging this definition into the last inequality yields

$$\begin{aligned} x \le C_{10} - JC_7 \end{aligned}$$

which is a contradiction since we have already proven that x cannot be bounded by a constant and that therefore we may assume that all three indices are greater than any fixed constant.

Thus \( \varphi _0, \ldots , \varphi _L \) must be linearly dependent over \( \mathbb {C}\). Without loss of generality we may assume that \( \varphi _1, \ldots , \varphi _L \) are linearly independent since otherwise we could group them together before doing the previous step. Hence, in a relation of linear dependence, there must be a non-zero coefficient in front of abc. So we can write

$$\begin{aligned} abc = \sum _{j=1}^{L} \lambda _j f_1^{3/2} \alpha _1^{(x+y+z)/2} t_j^{(x,y,z)} = f_1^{3/2} \alpha _1^{(x+y+z)/2} \sum _{j=1}^{L} \lambda _j t_j^{(x,y,z)} \end{aligned}$$

for \( \lambda _j \in \mathbb {C}\).

We distinguish between two cases considering the parity of \( x+y+z \). If \( x+y+z \) is even, then we have

$$\begin{aligned} f_1^{1/2} = \frac{abc}{f_1 \alpha _1^{(x+y+z)/2} \sum _{j=1}^{L} \lambda _j t_j^{(x,y,z)}} \in \mathbb {C}(X) \end{aligned}$$

which contradicts the assumption in our theorem that \( f_1 \) is no square in \( \mathbb {C}(X) \). When \( x+y+z \) is odd, we get

$$\begin{aligned} (f_1 \alpha _1)^{1/2} = \frac{abc}{f_1 \alpha _1^{(x+y+z-1)/2} \sum _{j=1}^{L} \lambda _j t_j^{(x,y,z)}} \in \mathbb {C}(X) \end{aligned}$$

which contradicts the assumption in our theorem that \( f_1 \alpha _1 \) is no square in \( \mathbb {C}(X) \). All in all there can be only finitely many sets satisfying the required properties. \(\square \)