1 Introduction

Each cohomology ring of a Grassmannian or flag variety has a basis of Schubert classes indexed by the elements of the corresponding Weyl group. Classical Schubert calculus computes the cohomology rings of Grassmannians and flag varieties in terms of the Schubert classes. This paper will “do Schubert calculus” in the equivariant cohomology rings of Peterson varieties.

The Peterson variety is a subvariety of the flag variety \(G/B\) parameterized by a linear subspace \(H_{Pet} \subseteq \mathfrak {g}\) and a regular nilpotent operator \(N_0\in \mathfrak {g}\). We define the Peterson variety to be

$$\begin{aligned} Pet=\{gB\in G/B : Ad(g^{-1})N_0 \in H_{Pet}\}. \end{aligned}$$

The objects \(H_{Pet}\) and \(N_0\) are fully defined in Sect. 2.

Peterson varieties were introduced by Peterson in the 1990s. Peterson constructed the small quantum cohomology of partial flag varieties from what are now Peterson varieties. Kostant used Peterson varieties to describe the quantum cohomology of the flag manifold [10] and Rietsch gave the totally non-negative part of type \(A\) Peterson varieties [12]. Insko–Yong [9] explicitly identified the singular locus of type \(A\) Peterson varieties and intersected them with Schubert varieties. Harada–Tymoczko [7, Lemma 5.1 (3)] proved that there is a circle action \(S^1\) which preserves Peterson varieties. We study the equivariant cohomology of the Peterson variety with respect to this action.

We use GKM theory as a model for studying equivariant cohomology, but Peterson varieties are not GKM spaces under the action of \(S^1\). Nonetheless, we are able to build several structures for \(H_{S^1}^*(Pet)\) corresponding to GKM structures of \(H_T^*(G/B)\):

GKM property of \(H_T^*(G/B)\)

Corresponding property of \(H_{S^1}^*(Pet)\)

Injects into a direct sum of polynomial rings: \(H_T^*(G/B)\hookrightarrow \bigoplus \limits _{(G/B)^T} H_T^*(pt)\)

Injects into a direct sum of polynomial rings: \(H_{S^1}^*(Pet)\hookrightarrow \bigoplus \limits _{(Pet)^{S^1}} H_T^*(pt)\) see Theorem 3.1

Fixed points \((G/B)^T\) indexed by elements of the Weyl group

Fixed points \((Pet)^{S^1}\) indexed by subsets of simple roots see Proposition 2.5

Basis of Schubert classes indexed by elements of the Weyl group

Basis of Peterson Schubert classes indexed by subsets of simple roots see Theorem 3.5

Using work by Harada and Tymoczko [7] and Precup [11], we construct a basis for the \(S^1\)-equivariant cohomology of Peterson varieties in all Lie types. This construction gives a set of classes which we call Peterson Schubert classes. The name indicates that the classes are projections of Schubert classes, they do not satisfy all the classical properties of Schubert classes. From the Peterson Schubert classes, we select a collection which is not only linearly independent over \(H_{S^1}^*(pt)\), but also upper triangular when appropriately ordered. Moreover, this collection spans \(H_{S^1}^*(Pet)\), just as the full set of Peterson Schubert classes does. We choose these classes by analogy with the type \(A\) work [6]; we ask whether there is a direct geometric justification for this choice. As we will show, this particular collection has especially elegant multiplication rules.

Classical Schubert calculus asks how to multiply Schubert classes; we ask how to multiply in the basis of Peterson Schubert classes. We give a Monk’s formula for multiplying a ring generator and a module generator, and a Giambelli’s formula for expressing any Peterson Schubert class in the basis in terms of the ring generators. The type \(A\) equivariant cohomology of the Peterson variety was presented by Harada–Tymoczko who gave a basis and a Monk’s rule for the equivariant cohomology ring [6]. A type \(A\) Giambelli’s formula was given by Bayegan and Harada [1]. This paper extends those results to all Lie types.

1.1 Main results

Peterson Schubert classes are indexed by subsets of the set of simple roots. To each subset \(K\) of simple roots, we associate (1) a reduced word \(v_K\) in the Weyl group and (2) a Peterson Schubert class \(p_{v_K}\). Classes \(p_{s_i}\) corresponding to one-element sets \(K\) generate \(H_{S^1}^*(Pet)\) as a ring over \(H_{S^1}^*(pt)\) and the set \(\{p_{v_K}:K\subseteq \Delta \}\) is a module basis over \(H_{S^1}^*(pt)\).

Monk’s rule is an explicit formula for multiplying an arbitrary module generator class \(p_{v_K}\) by a ring generator class \(p_{s_i}\). For the Peterson variety, Monk’s formula gives a set of constants \(c_{i,K}^J \in H_{S^1}^*(\text {pt})\) such that

$$\begin{aligned} p_{s_i}p_{v_K}=\sum \limits _{J\subseteq \Delta } c_{i,K}^J \cdot p_{v_J}. \end{aligned}$$

We give Monk’s rule in terms of localizations of Peterson Schubert classes at fixed points \(wB\in (Pet)^{S^1}\).

Theorem

The Peterson Schubert classes satisfy

$$\begin{aligned} p_{s_i} \cdot p_{v_K} = p_{s_i}(w_K) \cdot p_{v_K} + \sum \limits _{\begin{array}{c} K\subseteq J \subseteq \Delta \\ |J|=|K|+1 \end{array}} c_{i,K}^J \cdot p_{v_J} \end{aligned}$$

where \(w_K\) is the longest word of the parabolic subgroup \(W_K\subseteq W\) and the coefficients are \(c_{i,K}^J=(p_{s_i}(w_J)-p_{s_i}(w_K)) \cdot \frac{p_{v_K}(w_J)}{p_{v_J}(w_J)}\).

Our Monk’s rule is uniform across Lie types. This formula is similar in complexity to the equivariant Monk’s rule for \(G/B\) and has positive, although occasionally non-integral, coefficients. Giambelli’s formula is much simpler for Peterson varieties than for \(G/B\).

Theorem

Giambelli’s formula for Peterson Schubert classes:

$$\begin{aligned} \begin{array}{ll} |K|! \cdot p_{v_K} = \prod \limits _{\alpha _i \in K} p_{s_i} &{} \quad \text {if}\quad K\,\text {is connected and of type} \,A_n, B_n, C_n, F_4, \text { or}\, G_2\\ \frac{|K|!}{2} \cdot p_{v_K} = \prod \limits _{\alpha _i \in K} p_{s_i} &{} \quad \text {if}\quad K\,\text { is connected and of type}\, D_n\\ \frac{|K|!}{3} \cdot p_{v_K} = \prod \limits _{\alpha _i \in K} p_{s_i} &{}\quad \text {if}\quad K\,\text { is connected and of type}\, E_n.\end{array} \end{aligned}$$

In Theorem 5.5, we state a version of Giambelli’s formula that is uniform across all Lie types.

1.2 Proof techniques

For the proof of Monk’s rule and Giambelli’s formula, we modify Billey’s formula and Ikeda–Naruse’s excited Young diagrams for the \(S^1\)-equivariant cohomology of \(Pet\). Billey’s formula evaluates an equivariant Schubert class at a fixed point [2]. The evaluation gives \(\sigma _v(w)\in H_T^*(pt)\) as a polynomial in the simple roots. The map sending each simple root to a variable \(t\) extends to a ring homomorphism from \(H_T^*(pt)\) to \(H_{S^1}^*(pt)\). We use this map to calculate \(p_v(w)\), the evaluation of the Peterson Schubert class \(p_v\) at the \(S^1\)-fixed point \(wB\in Pet\).

Ikeda–Naruse used excited Young diagrams to encode and compute Billey’s formula [8]. Their work focuses on Grassmannian permutations. We modify excited Young diagrams to work with the longest word in a Weyl group of classical type. Our proofs only require evaluating \(p_v(w)\) when \(w=w_0\) and thus these modified excited Young diagrams are sufficient for the classical Lie types.

For the proof of Giambelli’s formula in the exceptional Lie types, we give explicit computations using Sage. The complete results for types \(F_4\) and \(G_2\) are included in this work. The code for the type \(E\) computations is available at arXiv:1311.2678.

1.3 Structure of the paper

Section 2 lays out the basic definitions of Peterson varieties, defines the appropriate torus action, and discusses the relevant parts of GKM theory. In Sect. 3, we give a basis of Peterson Schubert classes for \(H_{S^1}^*(Pet)\), and in Sect. 4 we prove Monk’s rule. Giambelli’s formula for Peterson varieties is presented in Sect. 5. Section 6 describes a modification of excited Young diagrams, a major tool in the proof of Giambelli’s formula. Lastly, in Sect. 7, we prove Giambelli’s formula for each Lie type.

2 The \(S^1\) action on Peterson varieties

Fix a complex reductive linear algebraic group \(G\), a Borel subgroup \(B\), and a maximal torus \(T\subseteq B \subseteq G\). This choice gives

  • a root system \(\varPhi \)

  • positive roots \(\varPhi ^+ \subset \varPhi \)

  • simple roots \(\Delta \subset \varPhi ^+\)

  • an associated Weyl group \(W\)

  • associated Lie algebras \(\mathfrak {t}\subseteq \mathfrak {b} \subseteq \mathfrak {g}\)

  • root spaces \(\mathfrak {g_\alpha } \subset \mathfrak {g}\) for each root \(\alpha \in \varPhi \).

We also choose a basis element \(E_\alpha \in \mathfrak {g}_\alpha \) for each of the root spaces. Some of our constructions rely on a specific ordering of the roots \(\alpha _1, \alpha _2,\ldots , \alpha _{|\Delta |} \in \Delta \). This ordering can be expressed in the Dynkin diagram of \(\Delta \). Figure 1 orders the simple reflections for each Lie type.

For any Lie type, the Peterson subspace in \(\mathfrak {g}\) is the direct sum of \(\mathfrak {b}\) and the root spaces corresponding to the negative simple roots:

$$\begin{aligned} H_{Pet}= \mathfrak {b} \oplus \bigoplus \limits _{\alpha \in -\Delta } \mathfrak {g}_\alpha . \end{aligned}$$

The regular nilpotent operator \(N_0 \in \mathfrak {g}\) is

$$\begin{aligned} N_0= \sum \limits _{\alpha \in \Delta } E_\alpha . \end{aligned}$$

In type \(A\), the operator \(N_0\) is the nilpotent with one Jordan block.

Definition 2.1

The Peterson variety \(Pet\) is a subvariety of the flag variety defined by

$$\begin{aligned} Pet= \lbrace gB \in G/B : Ad(g^{-1})(N_0) \in H_{Pet} \rbrace . \end{aligned}$$

Peterson varieties are a family of regular nilpotent Hessenberg variety. They are generally irreducible and generally not smooth [9].

Fig. 1
figure 1

The Dynkin diagrams show the order on the simple reflections. The same order is imposed on the corresponding simple roots throughout this paper

2.1 GKM theory and Billey’s formula

Named for Goresky, Kottwitz, and MacPherson, GKM theory expresses the \(T\)-equivariant cohomology of certain spaces in terms of polynomials corresponding to \(T\)-fixed points [5]. The flag variety \(G/B\) with the action of a maximal torus \(T\subseteq B\subseteq G\) is such a space.

The structure of \(H_T^*(G/B)\) is encoded in the combinatorics of the Weyl group \(W\). The elements of \(W\) index the \(T\)-equivariant Schubert classes. Each class can be thought of as a tuple of polynomials, one for each \(T\)-fixed point. As the \(T\)-fixed points are also indexed by the Weyl group, any ordered pair \(v,w \in W\) determines a polynomial \(\sigma _v(w)\in \mathbb {C}[\alpha _i:\alpha _i\in \Delta ]\). This polynomial is the Schubert class \(\sigma _v\) evaluated at the fixed point \(wB\in G/B\).

Billey gave an explicit combinatorial formula for computing the polynomial \(\sigma _v(w)\) [2]. Fix a reduced word for \(w=s_{b_1}s_{b_2} \cdots s_{b_{\ell (w)}}\) and for \(j\le \ell (w)\) let \(\mathbf {r}(\mathbf {j},w)=s_{b_1}s_{b_2} \cdots s_{b_{i-1}}(\alpha _{b_i}).\) Then

$$\begin{aligned} \sigma _v(w)= \sum \limits _{\begin{array}{c} \text {reduced words} \\ v=s_{b_{j_1}}s_{b_{j_2}} \cdots s_{b_{j_{\ell (v)}}} \end{array}} \left( \prod \limits _{i=1}^{\ell (v)} \mathbf {r}(\mathbf {j_i},w) \right) . \end{aligned}$$
(1)

Proposition 2.2

(Billey [2]) Properties of the polynomial \(\sigma _v(w)\):

  1. 1.

    The polynomial \(\sigma _v(w)\) is homogeneous of degree \(\ell (v)\).

  2. 2.

    If \(v \not \le w\) then \(\sigma _v(w)=0\).

  3. 3.

    If \(v\le w\) then \(\sigma _v(w) \ne 0\).

  4. 4.

    The polynomial \(\sigma _v(w)\) has non-negative integer coefficients, i.e.,

    $$\begin{aligned} \sigma _v(w) \in \mathbb {Z}_{\ge 0} [\alpha _i:\alpha _i\in \Delta ]. \end{aligned}$$
  5. 5.

    The polynomial \(\sigma _v(w)\) does not depend on the choice of reduced word for \(w\).

Example 2.3

Let \(G/B\) have Weyl group \(W=A_2\) and let \(w=s_1s_2s_1\) and \(v=s_1\). The word \(v\) is found as a subword of \(s_1s_2s_1\) in the two places \(\mathbf {s_1} s_2s_1\) and \(s_1s_2\mathbf {s_1}\).

$$\begin{aligned} \sigma _v(w)=\mathbf {r}(\mathbf {1},s_1s_2s_1) + \mathbf {r}(\mathbf {3},s_1s_2s_1)= \alpha _1 + s_1s_2(\alpha _1)=\alpha _1+ \alpha _2. \end{aligned}$$

2.2 GKM theory and Peterson varieties

GKM theory applies to the full flag variety with the action of a maximal torus \(T\) [5]. Since the Peterson variety is a subvariety of the flag variety, it is natural to ask if it has a similar torus action. The torus \(T\) does not preserve \(Pet\) but a one-dimensional subtorus \(S^1 \subseteq T\) does. In type \(A\), this subtorus is

$$\begin{aligned} S^1= \left[ \begin{array}{c@{\quad }c@{\quad }c@{\quad }c@{\quad }c} t &{} 0 &{} \cdots &{} 0 &{}0\\ 0 &{} t^2 &{} \cdots &{} 0 &{}0\\ \vdots &{} \vdots &{} \ddots &{}\vdots &{} \vdots \\ 0&{}0&{}\cdots &{} t^{n-1} &{}0\\ 0&{}0&{}\cdots &{}0 &{}t^n\\ \end{array}\right] \subseteq \left[ \begin{array}{ccccc} t_1 &{} 0 &{} \cdots &{} 0 &{}0\\ 0 &{} t_2 &{} \cdots &{} 0 &{}0\\ \vdots &{} \vdots &{} \ddots &{}\vdots &{} \vdots \\ 0&{}0&{}\cdots &{} t_{n-1} &{}0\\ 0&{}0&{}\cdots &{}0 &{}t_n\\ \end{array}\right] =T. \end{aligned}$$

We define the one-dimensional torus \(S^1\) in general Lie type.

Definition 2.4

[7, Lemma 5.1] The characters \(\alpha _1, \ldots \alpha _n\in \mathfrak {t}^*\) are a maximal \(\mathbb {Z}\)-linearly independent set in \(\mathfrak {t}^*\). Let \(\phi :T \rightarrow (\mathbb {C}^*)^n\) be the isomorphism of linear algebraic groups \(t\mapsto (\alpha _1(t), \alpha _2(t),\ldots ,\alpha _n(t))\). Define a one-dimensional torus \(S^1\) by

$$\begin{aligned} S^1= \phi ^{-1}\left( \lbrace (c,c, \ldots , c) : c \in \mathbb {C}^* \rbrace \right) . \end{aligned}$$

Proposition 2.5

[7, Lemma 5.1] The torus \(S^1\) acts on the Peterson variety.

Any point in \(Pet\) fixed by \(T\) will also be fixed by \(S^1\). In fact these are the only points in the Peterson variety fixed by \(S^1\):

$$\begin{aligned} (Pet)^{S^1}=Pet \cap (G/B)^T. \end{aligned}$$

Harada and Tymoczko gave the \(S^1\)-fixed points of \(Pet\) explicitly. Let \(K\subseteq \Delta \) be a subset of the simple roots. Define \(W_K\subseteq W\) as the parabolic subgroup generated by \(K\) and let \(w_K\) be the longest element of \(W_K\).

Proposition 2.6

[7, Proposition 5.8] An element \(wB\in G/B\) is an \(S^1\)-fixed point of \(Pet\) if and only if \(w=w_K\) for some set \(K\subseteq \Delta \).

Notation We frequently refer to the fixed point \(w_KB\in Pet\) by the coset representative \(w_K\).

Although \(Pet\) has a torus action and torus-fixed points indexed by Weyl group elements, it is not a GKM space. For Peterson varieties, we must build the GKM-like structures we want.

3 Peterson Schubert classes as a basis of \(H_{S^1}^*(Pet)\)

Harada–Tymoczko [7, Theorem 5.4] gave a projection from \(H_T^*(G/B)\) to \(H_{S^1}^*(Pet)\) in classical Lie types. In this section, we extend their results to all Lie types. For classical Lie types, they gave the following commutative diagram which this section will extend to all Lie types.

$$\begin{aligned} \begin{array}{ccc} H_{T}^{*}(G/B) &{} \hookrightarrow &{} \bigoplus \limits _{(G/B)^T} H_T^{*}(pt) \\ \downarrow &{}&{} \downarrow \pi _1 \\ H_{S^1}^{*}(G/B)&{} \hookrightarrow &{} \bigoplus \limits _{(G/B)^{S^1}} H_{S^1}^{*}(pt)\\ \downarrow &{} &{} \downarrow \pi _2\\ H_{S^1}^{*}(Pet ) &{} \hookrightarrow &{} \bigoplus \limits _{(Pet)^{S^1}} H_{S^1}^{*}(pt)\\ \end{array} \end{aligned}$$
(2)

A priori \(H_T^*(G/B)\) is a module over \(\mathbb {C}[\alpha _i{:}\alpha _i{\in } \Delta ]\). The map \(\pi _1{:} H_T^*(pt) \rightarrow H_{S^1}^*(pt)\) is the ring homomorphism which takes simple roots \(\alpha _i \in \Delta \) to the variable \(t\). The map \(\pi _2\) forgets the \(T\)-fixed points of \(G/B\) that are not in the Peterson variety. The top two injectivities are a direct result of GKM theory [5]. We prove the bottom injectivity in Theorem 3.1.

3.1 Peterson Schubert classes

The image of a Schubert class \(\sigma _v\in \bigoplus \limits _{(G/B)^T} H_T^*(pt)\) in \(\bigoplus \limits _{(Pet)^{S^1}} H_{S^1}^*(Pet)\) is denoted \(p_v\) and called a Peterson Schubert class. The class \(p_v\) has one polynomial for each \(S^1\)-fixed point of \(Pet\) so a Peterson Schubert class can be thought of as a \(2^{|\Delta |}\)-tuple of polynomials in \(\mathbb {C}[t]\). Below is an example in type \(A_2\).

$$\begin{aligned} \begin{array}{c@{\quad }c@{\quad }c@{\quad }c@{\quad }c@{\quad }c} &{}\sigma _{s_1}&{} &{} &{} &{} {P_{s_{1}}}\\ \begin{array}{c} 1\\ s_1 \\ s_2\\ s_1s_2\\ s_2s_1\\ s_1s_2s_1\\ \end{array}&{} \left( \begin{array}{c} 0\\ \alpha _1\\ 0 \\ \alpha _1 \\ \alpha _1+\alpha _2 \\ \alpha _1+ \alpha _2\\ \end{array}\right) &{} \begin{array}{c} {\pi _{1}}\\ \longmapsto \\ \end{array}&{} \left( \begin{array}{c} 0\\ t\\ 0 \\ t \\ 2t \\ 2t\\ \end{array}\right) &{} \begin{array}{c} {\pi _{2}}\\ \longmapsto \\ \end{array}&{} \left( \begin{array}{c} 0 \\ t \\ 0 \\ \\ \\ 2t\\ \end{array}\right) \\ \end{array}. \end{aligned}$$

Theorem 3.1

The map \(H_{S^1}^*(Pet) \rightarrow \bigoplus \limits _{(Pet)^{S^1}} H_{S^1}^*(pt)\) induced by the inclusion \((Pet)^{S^1} \hookrightarrow Pet\) is an injection.

Theorem 3.1 is a generalization of Harada–Tymoczko’s Theorem 5.4 to all Lie types [7].

Proof

We start by showing that the ordinary cohomology of \(Pet\) vanishes in odd degree. Precup [11, Theorem 5.4] proved that \(Pet\) is paved by complex affines for any Lie type. Precup [11, Lemma 2.7] also showed that the compact cohomology of the Peterson variety is only supported in even dimensions. Because the Peterson variety is compact, its ordinary cohomology vanishes in odd degree.

Following Harada–Tymoczko [7, Remark 4.11], the Leray-Serre spectral sequence of the Borel-equivariant cohomology of \(Pet\) collapses and thus \(H_{S^1}^*(Pet)\) is a free \(H_{S^1}^*(pt)\)-module. Therefore, the inclusion \((Pet)^{S^1}\hookrightarrow Pet\) induces an injection

$$\begin{aligned} H_{S^1}^*(Pet)\hookrightarrow \bigoplus \limits _{(Pet)^{S^1}} H_{S^1}^*(pt). \end{aligned}$$

This concludes the proof.\(\square \)

In the terminology of Harada–Tymoczko, since \((Pet)^{S^1}= Pet \cap (G/B)^T\) the pair \((Pet, S^1)\) is GKM-compatible with \((G/B,T)\).

3.2 A basis of Peterson Schubert classes

The \(S^1\)-fixed points of \(Pet\) are indexed by subsets \(K\subseteq \Delta \) so we want to index the Peterson Schubert classes by \(K\subseteq \Delta \).

Definition 3.2

A subset of simple roots \(K\subseteq \Delta \) is called connected if the induced Dynkin diagram of \(K\) is a connected subgraph of the Dynkin diagram of \(\Delta \).

Any subset \(K\subseteq \Delta \) can be written as \(K=K_1\times \cdots \times K_m\) where each \(K_i\) is a maximally connected subset. Each connected subset corresponds to its own Lie type.

Definition 3.3

Let \(K\subseteq \Delta \) be a connected subset. We define \(v_K\in W_K\) to be

$$\begin{aligned} v_K=\prod \limits _\mathrm{{Root}_K(i)=1}^{|K|} s_i \end{aligned}$$

where \(\mathrm{{Root}}_K(i)\) is the index of the corresponding root in a root system of the same Lie type as \(K\), ordered as in Fig. 1. If \(K=K_1\times \cdots \times K_m\) and each \(K_i\) is maximally connected then \(v_K=v_{K_1}v_{K_2}\cdots v_{K_m}\).

When \(\Delta \) is not of type \(D\) or \(E\) this definition gives \(v_K=s_{a_1} s_{a_2}\cdots s_{a_m}\) where \(K=\{\alpha _{a_1},\alpha _{a_2},\ldots , \alpha _{a_m}\}\) and \(a_1<a_2<\cdots <a_m\). This is the definition given in type \(A\) by Harada–Tymoczko [6]. Example 3.4 illustrates how Definition 3.3 differs from the type \(A\) definition.

Example 3.4

Let \(\Delta =\{\alpha _1, \alpha _2,\alpha _3, \alpha _4, \alpha _5, \alpha _6\}\) be a the set of simple roots of a type \(E_6\) root system and let \(K=\Delta \setminus \{\alpha _6\}\). The subset \(K\subseteq \Delta \) represented by a marked set of vertices in the Dynkin diagram and compared to the Dynkin diagram for \(D_5\). The word \(v_K\) is \(s_1s_3s_4s_5s_2\).

figure a

Note that \(v_K\) is a Coxeter element of \(W_K\). Because of the labeling imposed on the simple roots in Fig. 1, each subset \(K\) of simple roots corresponds to exactly one word \(v_K\).

Theorem 3.5

(Basis of Peterson Schubert classes) The Peterson Schubert classes \(\lbrace p_{v_K} : K\subseteq \Delta \rbrace \) are a basis of \(H_{S^1}^*(Pet)\) as a module over the \(S^1\)-equivariant cohomology of a point, \(H_{S^1}^*(pt)\cong \mathbb {C}[t]\).

This is a version of Harada–Tymoczko’s Theorem 5.9 [7]. With Precup’s work and Lemmas 3.6 and 3.7, we extend the proof to all Peterson varieties.

Lemma 3.6

For any set of simple roots \(\Delta \) and any subsets \(J, K \subseteq \Delta \) the polynomial \(p_{v_J}(w_K)\) is zero unless \(J\subseteq K\).

Proof

Suppose \(J \not \subseteq K\) and that \(\alpha _j \in J \setminus K\). Then \(s_j \le v_J\) and \(s_j\not \le w_K\) in the Bruhat order. For \(\sigma _{v_J}(w_K)\) to be non-zero, there must be some subword of \(w_K\) that is equal to \(v_J\) and therefore \(v_J \le w_K\). But \(s_j \le v_J\) implies that \(s_j \le w_K\) which is a contradiction. Thus, \(\sigma _{v_J}(w_K)=0\) by Property 2 of Proposition 2.2 and by construction \(p_{v_J}(w_K)=0\).\(\square \)

Lemma 3.7

For any set of simple roots \(\Delta \) and any subset \(K\subseteq \Delta \)

$$\begin{aligned} p_{v_K}(w_K) \ne 0. \end{aligned}$$

Proof

Since \(v_K \in W_K\), we must have \(v_K \le w_K\). By Property 3 of Proposition 2.2, the polynomial \(\sigma _{v_K}(w_K) \in \mathbb {C}[\alpha _i:\alpha _i\in \Delta ]\) is not equal to zero. We have defined \(p_{v_K}(w_K)\) to be \(\pi _1(\sigma _{v_K}(w_K))\). Since \(\sigma _{v_K}(w_K)\) has positive integer coefficients over the simple roots by Property 4 of the same proposition, its image in \(\mathbb {C}[t]\) must also have positive integer coefficients.\(\square \)

Proof

(Basis of Peterson Schubert classes) Impose a partial order on the sets \(\lbrace K\subseteq \Delta \rbrace \) by inclusion and extend that to a total order on the sets \(K\) of simple roots. Use that total order to order the classes \(\lbrace p_{v_K}\rbrace \) and the \(S^1\)-fixed points \(w_K \in Pet\). Lemma  3.6 implies that the collection \(\lbrace p_{v_K}\rbrace \) is lower triangular and Lemma  3.7 implies that the collection has full rank. Thus, \(\lbrace p_{v_K}\rbrace \) is a linearly independent set. By the Property 1 of Billey’s formula, the polynomial degree of \(p_{v_K}\) is \(|K|\) and its cohomology degree is \(2|K|\). As there are \({n}\atopwithdelims (){|K|}\) subsets of \(\Delta \) with size \(|K|\), there are exactly \({n}\atopwithdelims (){|K|}\) Peterson Schubert varieties with cohomology degree \(2|K|\).

A paving by affines computes the Betti numbers of \(Pet\) [4, 19.1.11]. Precup’s paving by affines reveals that the dimensions of the corresponding pavings are also \(n\atopwithdelims ()|K|\) [11, Corollary 4.13]. As a linearly independent set with the right number of elements of each degree, the set \(\lbrace p_{v_K}\rbrace \) is a module basis of \(H_{S^1}^*(Pet)\) [6, Proposition A.1].\(\square \)

Example 3.8

Below is the Peterson Schubert class basis of the \(S^1\)-equivariant cohomology of \(Pet\) in Lie type \(C_3\). The classes and fixed points are indexed by the subsets \(K\subseteq \Delta \).

$$\begin{aligned} \begin{array}{ccccccccc} K &{} p_{v_\emptyset } &{} p_{v_{\{ \alpha _1 \}}} &{} p_{v_{\{ \alpha _2 \}}} &{} p_{v_{\{ \alpha _1 \}}} &{} p_{v_{\{ \alpha _1, \alpha _2 \}}} &{} p_{v_{\{ \alpha _1, \alpha _3 \}}} &{} p_{v_{\{ \alpha _2, \alpha _3 \}}} &{} p_{v_{\{ \alpha _1, \alpha _2, \alpha _3 \}}} \\ \begin{array}{c} \emptyset \\ \{ \alpha _1 \} \\ \{ \alpha _2 \} \\ \{ \alpha _3 \} \\ {\{ \alpha _1, \alpha _2 \}} \\ {\{ \alpha _1, \alpha _3 \}}\\ {\{ \alpha _2, \alpha _3 \}}\\ \{ \alpha _1, \alpha _2, \alpha _3 \}\\ \end{array}&{} \left( \begin{array}{c} 1\\ 1\\ 1\\ 1\\ 1\\ 1\\ 1\\ 1\\ \end{array}\right) &{} \left( \begin{array}{c} 0\\ t\\ 0\\ 0\\ 2t\\ t\\ 0\\ 5t\\ \end{array}\right) &{} \left( \begin{array}{c} 0\\ 0\\ t\\ 0\\ 2t\\ 0\\ 3t\\ 8t\\ \end{array}\right) &{} \left( \begin{array}{c} 0\\ 0\\ 0\\ t\\ 0\\ t\\ 4t\\ 9t\\ \end{array}\right) &{} \left( \begin{array}{c} 0\\ 0\\ 0\\ 0\\ 2t^2\\ 0\\ 0\\ 20t^2\\ \end{array}\right) &{} \left( \begin{array}{c} 0\\ 0\\ 0\\ 0\\ 0\\ t^2\\ 0\\ 45t^2\\ \end{array}\right) &{} \left( \begin{array}{c} 0\\ 0\\ 0\\ 0\\ 0\\ 0\\ 6t^2\\ 36t^2\\ \end{array}\right) &{} \left( \begin{array}{c} 0\\ 0\\ 0\\ 0\\ 0\\ 0\\ 0\\ 60t^3\\ \end{array}\right) \end{array} \end{aligned}$$

4 Monk’s formula

With the Peterson Schubert class basis for \(H_{S^1}^*(Pet)\) defined in Theorem 3.5, we can examine the structure of \(H_{S^1}^*(Pet)\) through its multiplication rules. First we determine a minimal set of Peterson Schubert classes that generate the ring \(H_{S^1}^*(Pet)\).

Lemma 4.1

The Peterson Schubert classes \(p_{s_i}\) are a minimal generating set for the ring \(H_{S^1}^*(Pet)\) over \(H_{S^1}^*(pt)\).

Proof

That the classes \(p_{s_i}\) generate \(H_{S^1}^*(Pet)\) is a consequence of Theorem 5.4. Each class \(p_{s_i}\) has polynomial degree one, so if \(p_{s_j}\) can be expressed in terms of the other degree-one Peterson Schubert classes, it is a sum

$$\begin{aligned} p_{s_j}=\sum \limits _{i\ne j} a_i \cdot p_{s_i} \end{aligned}$$

for some coefficients \(a_i\in \mathbb {C}\). But Theorem 3.5 shows that the Peterson Schubert classes are linearly independent, so the generating set \(\{p_{s_i}: \alpha _i\in \Delta \}\) is minimal.\(\square \)

Monk’s rule is an explicit formula for multiplying an arbitrary module generator class \(p_{v_K}\) by a ring generator class \(p_{s_i}\). For the Peterson variety, Monk’s formula gives a set of constants \(c_{i,K}^J \in H_{S^1}^*(\text {pt})\) such that

$$\begin{aligned} p_{s_i}p_{v_K}=\sum \limits _{J\subseteq \Delta } c_{i,K}^J \cdot p_{v_J}. \end{aligned}$$
(3)

The Peterson Schubert classes \(\lbrace p_{v_K}: K\subseteq \Delta \rbrace \) are a module basis for \(H_{S^1}^*(Pet)\) and the product of \(p_{s_i}\) and \(p_{v_K}\) is also in that module. Thus, a unique set of constants \(\lbrace c_{i,K}^J \rbrace \) solves this equation. Because \(H_{S^1}^*(pt)=\mathbb {C}[t]\) these structure constants are complex polynomials in \(t\).

Theorem 4.2

(Monk’s formula for Peterson varieties) The Peterson Schubert classes satisfy

$$\begin{aligned} p_{s_i} \cdot p_{v_K} = p_{s_i}(w_K) \cdot p_{v_K} + \sum \limits _{\begin{array}{c} J {\text { such that}}\\ K\subseteq J \subseteq \Delta \\ |J|=|K|+1 \end{array}} c_{i,K}^J \cdot p_{v_J} \end{aligned}$$

where the coefficients \(c_{i,K}^J\) are non-negative rational numbers and

$$\begin{aligned} c_{i,K}^J=(p_{s_i}(w_J)-p_{s_i}(w_K)) \cdot \frac{p_{v_K}(w_J)}{p_{v_J}(w_J)}. \end{aligned}$$

We use two lemmas to eliminate many subsets \(J\subseteq \Delta \) by showing that \(c_{i,K}^J=0\).

Lemma 4.3

If \(|J|>|K|+1\) then \(c_{i,K}^J=0\).

Proof

The polynomial degree of \(p_{v} \in \mathbb {C}[t]\) is the length of a reduced word for \(v\). Therefore, the Peterson Schubert class \(p_{v_K}\) has degree \(|K|\) and the polynomial degree of \(p_{s_i}p_{v_K}\) is \(|K|+1\). The polynomial degrees on the right- and left-hand sides of Eq. (3) must be equal. Take only the parts of each side of Eq. (3) that have degree higher than \(|K|+1\). It follows that

$$\begin{aligned} 0=\sum \limits _{\begin{array}{c} J\subseteq \Delta \\ |J|>|K|+1 \end{array}} c_{i,K}^J\cdot p_{v_J}. \end{aligned}$$

The Peterson Schubert classes \(p_{v_J}\) are linearly independent of Theorem 3.5. Therefore, whenever \(|J|>|K|+1\) the coefficient \(c_{i,K}^J=0\).\(\square \)

Lemma 4.4

The constant \(c_{i,K}^J=0\) unless \(K\subseteq J\).

Proof

Suppose that \(L\) is the smallest counterexample, i.e., \(L\subseteq \Delta \) does not contain \(K\) and for all \(H\subsetneq L\) the coefficient \(c_{i,K}^H=0\). Evaluate Monk’s formula at the \(S^1\)-fixed point \(w_L\) to get

$$\begin{aligned} p_{s_i}(w_L) \cdot p_{v_K}(w_L)=\sum \limits _{\begin{array}{c} J\subseteq \Delta \\ |J| \le |K|+1 \end{array}} c_{i,K}^J \cdot p_{v_J}(w_L). \end{aligned}$$

The word \(v_K \not \le w_L\) by hypothesis so the left-hand side is \(0\). If \(J \not \subseteq L\) then \(p_{v_J}(w_L)\) is zero and thus

$$\begin{aligned} 0=\sum \limits _{\begin{array}{c} J\subseteq L \subseteq \Delta \\ |J| \le |K|+1 \end{array}} c_{i,K}^J \cdot p_{v_J}(w_L). \end{aligned}$$

By construction, if \(J\subsetneq L\) then \(c_{i,K}^J=0\) so we are left with

$$\begin{aligned} 0= c_{i,K}^L \cdot p_{v_L}(w_L). \end{aligned}$$

By Lemma 3.7, the evaluation \(p_{v_L}(w_L) \ne 0\). Since \(H_{S^1}^*(pt)=\mathbb {C}[t]\) is an integral domain we conclude that \(c_{i,K}^L =0\).\(\square \)

Having determined which coefficients are zero, we give a third lemma addressing the non-zero coefficients.

Lemma 4.5

Consider the map \(\pi _1:H_T^*(G/B) \rightarrow H_{S^1}^*(G/B)\) from Eq. (2). Let \(v,w\) be elements of the Weyl group. The image under \(\pi _1\) of the evaluation \(\sigma _v(w)\) of a Schubert class \(\sigma _v\) at the fixed point \(w\) is a monomial \(c\cdot t^m\) where \(c\) is a non-negative integer and \(m\) is the length of \(v\).

Proof

By the Properties 1 and 4 of Billey’s formula given in Proposition 2.2, the polynomial \(\sigma _v(w)\) is homogeneous of degree \(\ell (v)\) with non-negative integer coefficients. Its image \(\pi _1(\sigma _v(w))\) is \(ct^{\ell (v)}\) where \(c\) is the sum of the integer coefficients of \(\sigma _v(w)\).

\(\square \)

We now prove Theorem 4.2.

Proof

(Monk’s formula for Peterson varieties) By Lemma 4.3, the general Monk’s formula in Eq. (3) simplifies to

$$\begin{aligned} p_{s_i} \cdot p_{v_K} = \sum \limits _{\begin{array}{c} |J|\le |K|+1 \end{array}} c_{i,K}^J \cdot p_{v_J} \end{aligned}$$

and Lemma 4.4 further refines the equation to

$$\begin{aligned} p_{s_i} \cdot p_{v_K} = c_{i,K}^K \cdot p_{v_K} + \sum \limits _{\begin{array}{c} K\subsetneq J \subseteq \Delta \\ |J|=|K|+1 \end{array}} c_{i,K}^J \cdot p_{v_J}. \end{aligned}$$
(4)

We evaluate both sides of Eq. (4) at the \(S^1\)-fixed point \(w_K\) and use the fact that \(p_{v_J}(w_K)=0\) whenever \(J\) is not a subset of \(K\) to obtain

$$\begin{aligned} p_{s_i}(w_K) \cdot p_{v_K}(w_K) = c_{i,K}^K \cdot p_{v_K}(w_K). \end{aligned}$$

The polynomial \( p_{v_K}(w_K)\) is non-zero by Lemma 3.7. Since \(\mathbb {C}[t]\) is an integral domain we may divide both sides by \( p_{v_K}(w_K)\). This leaves \( c_{i,K}^K = p_{s_i}(w_K).\) By Lemma 4.5, the polynomial \(p_{s_i}(w_K)\) is a degree-one monomial with an integer coefficient.

Next fix a subset \(L\subseteq \Delta \) such that \(K\subsetneq L \) and let \(|L|=|K|+1\). Evaluating at the \(S^1\)-fixed point \(w_L\) gives

$$\begin{aligned} p_{s_i}(w_L) \cdot p_{v_K}(w_L) = p_{s_i}(w_K) \cdot p_{v_K}(w_L) + \sum \limits _{\begin{array}{c} J \text { such that}\\ K\subsetneq J \subseteq \Delta \\ |J|=|K|+1 \end{array}} c_{i,K}^J \cdot p_{v_J}(w_L). \end{aligned}$$

But \(p_{v_J}(w_L)=0\) unless \(J\subseteq L\) by the Properties 2 and 3 of Billey’s formula so in fact

$$\begin{aligned} p_{s_i}(w_L) \cdot p_{v_K}(w_L) = p_{s_i}(w_K) \cdot p_{v_K}(w_L) + c_{i,K}^L \cdot p_{v_L}(w_L). \end{aligned}$$

Solving for \(c_{i,K}^L\) gives

$$\begin{aligned} c_{i,K}^L=(p_{s_i}(w_L)-p_{s_i}(w_K))\cdot \frac{p_{v_K}(w_L)}{p_{v_L}(w_L)}. \end{aligned}$$
(5)

If the term \((p_{s_i}(w_L)-p_{s_i}(w_K))=0\) then the constant \(c_{i,K}^L\) is non-negative and rational. Suppose that \((p_{s_i}(w_L)-p_{s_i}(w_K))\ne 0\). By Lemma 4.5, the term \((p_{s_i}(w_L)-p_{s_i}(w_K))\) has degree one. By the same lemma \(\frac{p_{v_K}(w_L)}{p_{v_L}(w_L)}\) has degree \(|K|-|L|=-1\). Thus, \(c_{i,K}^L\) is a priori a rational number. It remains to show that \(c_{i,K}^L\) is non-negative.

It suffices to show that \((p_{s_i}(w_L)-p_{s_i}(w_K))\) is non-negative because \(\frac{p_{v_K}(w_L)}{p_{v_L}(w_L)}\) will always be non-negative. The word \(w_L\) can be written as \(w_K\cdot \tilde{w}\) for some reduced word \(\tilde{w} \in W_L\) [3]. Let \(s_{b_1}s_{b_2}\cdots s_{b_{m}}\) be a reduced word for \(w_K\) and \(s_{b_{m+1}} s_{b_{m+2}} \cdots s_{b_n}\) be a reduced word for \(\tilde{w}\). The length \(\ell (s_i)=1\) for each \(i\) so Billey’s formula says

$$\begin{aligned} \begin{array}{rl} \sigma _{s_i}(w_K \cdot \tilde{w})&{}= \sum \limits _{{s_{b_j}=s_i} } \mathbf {r}(\mathbf {j},w_L) \\ &{}= \sum \limits _{\begin{array}{c} {s_{b_j}=s_i} \\ j\le m \end{array}} \mathbf {r}(\mathbf {j},w_L) + \sum \limits _{\begin{array}{c} {s_{b_j}=s_i} \\ j> m \end{array}} \mathbf {r}(\mathbf {j},w_L) \\ &{}= \sigma _{s_i}(w_K) + \sum \limits _{\begin{array}{c} {s_{b_j}=s_i} \\ j> m \end{array}} \mathbf {r}(\mathbf {j},w_L) . \end{array} \end{aligned}$$

Since \(\pi _1\) is a ring homomorphism from \(\mathbb {C}[\alpha _i:\alpha _i\in \Delta ]\) to \(\mathbb {C}[t]\), we obtain

$$\begin{aligned} p_{s_i}(w_J)-p_{s_i}(w_K)=\pi _1\left( \sigma _{s_i}(w_J)-\sigma _{s_i}(w_K)\right) =\pi _1\Big (\sum \limits _{\begin{array}{c} {s_{b_j}=s_i} \\ j> m \end{array}} \mathbf {r}(\mathbf {j},w_J)\Big ). \end{aligned}$$

By the definition of Billey’s formula, each term \( \mathbf {r}(\mathbf {j},w_L) \) is a positive root in \(\varPhi \). Therefore, its image \(\pi _1(\mathbf {r}(\mathbf {j},w_J))\) is \(c t\) for some positive integer \(c\). The \(t\) is canceled by \(\frac{p_{v_K}(w_L)}{p_{v_L}(w_L)}\) which has degree \(-1\). Thus, \((p_{s_i}(w_L)-p_{s_i}(w_K))\) is non-negative and so is the coefficient \(c_{i,K}^J\).\(\square \)

In classical Schubert calculus, the structure constants are generally non-negative integers. Frequently they are in bijection with dimensions of irreducible representations. However, structure constants for the Peterson variety are not necessarily integers. For example, in type \(D_5\), let \(K=\lbrace \alpha _1, \alpha _2,\alpha _3, \alpha _4 \rbrace \) and \(J=\Delta \). Then

$$\begin{aligned} c_{5,K}^J=\frac{5}{2}. \end{aligned}$$

Conjecture 4.6

We conjecture that in this basis, non-integral structure constants only occur in Lie types \(D\) and \(E\).

5 Giambelli’s formula

Giambelli’s formula expresses an arbitrary module-basis element in terms of the ring generators. In both ordinary and equivariant cohomology, many spaces have determinantal Giambelli’s formulae.

Giambelli’s formula for Peterson varieties, however, simplifies to a single product.

Lemma 5.1

For any Peterson Schubert class \(p_{v_K}\), there exists a constant \(C\) satisfying

$$\begin{aligned} C \cdot p_{v_K}=\prod \limits _{\alpha _i \in K} p_{s_i}. \end{aligned}$$
(6)

Proof

If \(|K|=m\) let \(K=\lbrace \alpha _{a_1}, \alpha _{a_2},\ldots ,\alpha _{a_{m}} \rbrace \). Define a sequence of nested subsets \(\emptyset = K_0 \subsetneq K_1 \subsetneq K_2 \subsetneq \cdots \subsetneq K_{m}=K\) by

$$\begin{aligned} K_i=\lbrace \alpha _{a_1}, \alpha _{a_2}, \ldots ,\alpha _{a_{i}} \rbrace . \end{aligned}$$

From Eq. (4) in the proof of Monk’s formula for Peterson varieties

$$\begin{aligned} p_{s_{a_{i+1}}}\cdot p_{v_{K_i}}= c_{a_{i+1},K_i}^{K_{i+1}}\cdot p_{v_{K_i}} + \sum \limits _{\begin{array}{c} K_i\subsetneq J \subseteq \Delta \\ |J|=|K_i|+1 \end{array}} c_{a_{i+1},K_i}^J \cdot p_{v_J}. \end{aligned}$$

Theorem 4.2 says \(c_{a_{i+1},K_i}^{K_{i+1}}= p_{s_{a_{i+1}}}(w_{K_i})\). Because \(\alpha _{a_{i+1}} \not \in K_i\) the coefficient \(p_{s_{a_{i+1}}}(w_{K_i})=0\) . If \(\alpha _{a_{i+1}} \not \in J\) the term \( p_{s_{a_{i+1}}}(w_{J})=0\). Thus, if \(J\ne K_{i+1}\), the coefficient \(c_{a_{i+1},K_i}^J =0\). Now Eq. (4) reduces to

$$\begin{aligned} p_{s_{a_{i+1}}}\cdot p_{v_{K_i}}= c_{a_{i+1},K_i}^{K_{i+1}} \cdot p_{v_{K+1}}. \end{aligned}$$

Solving for \(p_{v_{K+1}}\) gives

$$\begin{aligned} \frac{p_{s_{a_{i+1}}}\cdot p_{v_{K_i}} }{ c_{a_{i+1},K_i}^{K_{i+1}}} = p_{v_{K{i+1}}}. \end{aligned}$$

By induction on \(i\) we see

$$\begin{aligned} p_{v_K}=\frac{\prod \limits _{i=1}^{|K|} p_{s_{a_{i}}}}{\prod \limits _{i=1}^{|K|} c_{a_{i},K_{i-1}}^{K_{i}}}. \end{aligned}$$

This gives that

$$\begin{aligned} C=\prod \limits _{i=1}^{|K|} c_{a_{i+1},K_i}^{K_{i+1}}. \end{aligned}$$

\(\square \)

To find this constant \(C\) explicitly we consider the simplest non-trivial Peterson Schubert classes, those that are connected.

Definition 5.2

From Definition 3.3, a subset of simple roots \(K\subseteq \Delta \) is called connected if the induced Dynkin diagram of \(K\) is a connected subgraph of the Dynkin diagram of \(\Delta \). The class \(p_{v_K}\) is called connected whenever \(K\) is connected.

Figure 2 gives examples of sets \(K\subseteq \Delta \) which are not connected. The induced Dynkin diagrams also give the Lie type of the subsystem \(K\). Every Peterson Schubert class can be expressed in terms of connected classes.

Fig. 2
figure 2

The subsystem \(K\subseteq \Delta \) is drawn as a marked set of vertices in the Dynkin diagram. The associated Peterson Schubert class is given at right

Theorem 5.3

If \(J,K\subset \Delta \) are each connected subsets such that \(J\cup K\) is disconnected then

$$\begin{aligned} p_{v_{J \cup K}}=p_{v_J}\cdot p_{v_K}. \end{aligned}$$
(7)

Proof

We show that equality holds when Eq. (7) is evaluated at any \(S^1\)-fixed point \(w_L\). If \(L\) does not contain \(J\cup K\) we can suppose without loss of generality that \(J\not \subseteq L\). Then both \(p_{v_{J \cup K}}(w_L)\) and \(p_{v_J}(w_L)\) are zero.

Now suppose \(J \cup K \subseteq L\). Even though \(J\cup K\) is disconnected, \(L\) may be connected or disconnected. Fix a reduced word for \(w_L\)

$$\begin{aligned} \tilde{w}_L=s_{a_1}s_{a_2}\cdots s_{a_{\ell (w_L)}} \end{aligned}$$

and let \(b \prec \tilde{w}_L\) mean that \(b\) is a subword of \(\tilde{w}_L\). The indexing set of the subword \(b\) is the set \(I(b) \subseteq \lbrace 1,2, \ldots , \ell (w_L) \rbrace \) such that

$$\begin{aligned} b=s_{a_{j_1}}s_{a_{j_2}} \cdots s_{a{j_{|I(b)|}}}\quad \text {for}\quad j_1<j_2<\cdots <j_{|I(b)|} \text { with each }j_k \in I(b). \end{aligned}$$

The subwords of \(\tilde{w}_L\) are in bijection with the subsets of \(\lbrace 1,2, \ldots , \ell (w_L) \rbrace \). Given two subwords \(b_1,b_2 \prec \tilde{w}_L\) we define their union \(b_1 \cup b_2\) to be the subword

$$\begin{aligned} b_1 \cup b_2 =s_{a_{j_1}}s_{a_{j_2}} \cdots s_{a{j_{|I(b_1)\cup I(b_2)|}}} \end{aligned}$$

for \(j_1<j_2<\cdots <j_{|I(b_1)\cup I(b_2)|}\) with each \(j_k \in I(b_1)\cup I(b_2).\) Let \(b_J, b_K \prec \tilde{w}_L\) be reduced words for \(v_J\) and \(v_K\), respectively. Since \(J\) and \(K\) are disconnected \(I(b_J) \cap I(b_K) = \emptyset \) and \(v_J\) commute entirely with \(v_K\) [3]. Thus, \(b_J \cup b_K\) is a reduced word for \(v_J\cdot v_K=v_{J\cup K}\).

Conversely let \(b\prec \tilde{w}_L\) be a reduced word for \(v_{J\cup K}\). We can partition \(I(b)\) into

$$\begin{aligned} I(b)_J=\lbrace j_k\in I(b) : \alpha _{a_{j_k}}\in J \rbrace \quad \text {and}\quad I(b)_K=\lbrace j_k\in I(b) : \alpha _{a_{j_k}}\in K \rbrace . \end{aligned}$$

Since \(v_J \le v_{J\cup K}\) and \(b\) is a reduced word for \(v_{J\cup K}\), some reduced word for \(v_J\) must be a subword of \(b\). Let \(b_J \prec b\) be that subword. Since no reflections associated with \(K\) are in \(b_J\), \(I(b_J)\subseteq I(b)_J\). A parallel argument shows that there is some subword \(b_K\prec b\) equal to \(v_K\) and that \(I(b_K)\subseteq I(b)_K\).

By our previous argument \(b_J\cup b_K\) is a reduced word for \(v_J\cdot v_K =v_{J\cup K}\). So \(\ell (b_J\cup b_K)=\ell (v_{J\cup K})\) which equals \(\ell (b)\). Thus, \(I(b_J)= I(b)_J\) and \(I(b_K)= I(b)_K\) and \(b=b_J \cup b_K\).

A subword \(b\prec \tilde{w}_L\) is a reduced word for \(v_{J\cup K}\) if and only if \(b=b_J \cup b_K\) for subwords \(b_J, b_K \prec \tilde{w}_L\) reduced words for \(v_J\) and \(v_K\). Billey’s formula in Eq. (1) is a sum over such subwords. We use it to rewrite the left- and right-hand sides of Eq.  (7). The left-hand side becomes

$$\begin{aligned} p_{v_{J\cup K}}(w_L)= \sum \limits _{\begin{array}{c} b \prec \tilde{w}_L \\ b= v_{J\cup K} \\ |I(b)|=|J \cup K| \end{array}} \left( \prod \limits _{j\in I(b)} \mathbf {r}(\mathbf {j},\tilde{w}_L) \right) . \end{aligned}$$
(8)

Similarly, the right-hand side becomes

$$\begin{aligned} p_{v_{J}}(w_L)\cdot p_{v_{K}}(w_L)\!=\!\left[ \sum \limits _{\begin{array}{c} b \prec \tilde{w}_L \\ b= v_{J} \\ |I(b)|=|J | \end{array}} \left( \prod \limits _{j\in I(b)} \mathbf {r}(\mathbf {j},\tilde{w}_L) \right) \right] \cdot \left[ \sum \limits _{\begin{array}{c} b \prec \tilde{w}_L \\ b= v_{ K} \\ |I(b)|=|K| \end{array}} \left( \prod \limits _{j\in I(b)} \mathbf {r}(\mathbf {j},\tilde{w}_L) \right) \right] .\nonumber \\ \end{aligned}$$
(9)

Both Eqs. (8) and  (9) expand to the expression

$$\begin{aligned} \sum \limits _{\begin{array}{c} b_J \prec \tilde{w}_L \\ b_J= v_{J} \\ |I(b_J)|=|J | \end{array}} \sum \limits _{\begin{array}{c} b_K \prec \tilde{w}_L \\ b_K= v_{K} \\ |I(b_K)|=|K| \end{array}} \left[ \left( \prod \limits _{j\in I(b_J)} \mathbf {r}(\mathbf {j},\tilde{w}_L) \right) \cdot \left( \prod \limits _{j\in I(b_K)} \mathbf {r}(\mathbf {j},\tilde{w}_L) \right) \right] . \end{aligned}$$

\(\square \)

Any subset \(K\subset \Delta \) gives rise to a Peterson Schubert class that is the product of connected Peterson Schubert classes. Understanding the connected Peterson Schubert classes thus gives full information on all Peterson Schubert classes. The next theorem gives Giambelli’s formula explicitly for connected Peterson Schubert classes.

Theorem 5.4

If \(K\subseteq \Delta \) is a connected root subsystem of type \(A_n,B _n,C_n,F_4,\) or \(G_2\) then

$$\begin{aligned} |K|!\cdot p_{v_K} = \prod \limits _{\alpha _i \in K} p_{s_i}. \end{aligned}$$

If \(K\) is a connected root subsystem of type \(D_n\) then

$$\begin{aligned} \frac{|K|!}{2}\cdot p_{v_K} = \prod \limits _{\alpha _i \in K} p_{s_i}. \end{aligned}$$

If \(K\) is a connected root subsystem of type \(E_n\) then

$$\begin{aligned} \frac{|K|!}{3}\cdot p_{v_K} = \prod \limits _{\alpha _i \in K} p_{s_i}. \end{aligned}$$

Our proof of this theorem, given in Sect. 7, is combinatorial and treats each Lie type as its own case. Theorem 5.4 can be restated uniformly.

Theorem 5.5

If \(K\subseteq \Delta \) is a connected root subsystem of any Lie type and \(|\mathcal {R}(v_K)|\) is the number of reduced words for \(v_K\) then

$$\begin{aligned} \frac{|K|!}{|\mathcal {R}(v_K)|}\cdot p_{v_K} = \prod \limits _{\alpha _i \in K} p_{s_i}. \end{aligned}$$

The uniformity across Lie types suggests that a uniform proof exists. Such a proof might shed light on the topology of these varieties.

Proof

Given Theorem 5.4 it is sufficient to show that \(|\mathcal {R}(v_K)|=1\) if \(K\) is type \(A,B,C,F,\) or \(G\), that \(|\mathcal {R}(v_K)|=2\) for type \(D\) and that \(|\mathcal {R}(v_K)|=3\) for type \(E\). Given one reduced word any other reduced word can be obtained by a series of braid moves and commutations [3]. If \(K\) is type \(A,B,C,F,\) or \(G\) then \(s_i\) and \(s_{i+1}\) do not commute for any \(i\). Therefore, \(s_1s_2\cdots s_{n-1}s_n\) is the only reduced word for \(v_K\).

If \(K\) is of type \(D\) then \(s_i\) and \(s_{i+1}\) commute if and only if \(i=n-1\). Also \(s_{n-2}\) and \(s_n\) do not commute. The only reduced words for \(v_K\) are \(s_1s_2\cdots s_{n-2}s_{n-1}s_n\) and \(s_1s_2\cdots s_{n-2}s_{n}s_{n-1}\) so \(|\mathcal {R}(v_K)|=2\).

If \(K\) is type \(E_n\) then we start with the word \(v_K=s_1s_2s_3s_4\cdots s_n\) with the labels given as in Fig. 1. The reflection \(s_2\) commutes with \(s_1\) and \(s_3\) but not \(s_4\). The reflection \(s_3\) does not commute with \(s_1\). When \(i>2, s_i\) and \(s_{i+1}\) do not commute. Thus, \(v_K\) has exactly \(3\) reduced words: \(s_1s_2s_3s_4\cdots s_n\) and \(s_1s_3s_2s_4\cdots s_n\) and \(s_2s_1s_3s_4\cdots s_n\).\(\square \)

We can now give Giambelli’s formula explicitly for all Peterson Schubert classes.

Corollary 5.6

If \(K\subseteq \Delta \) and \(K=K_1 \times K_2 \times \cdots K_j\) where each \(K_\ell \) is maximally connected then

$$\begin{aligned} C_K \cdot p_{v_K} = \prod \limits _{i\in K} p_{s_i} \end{aligned}$$
(10)

where \(C_K=\prod \limits _{\ell =1}^j \frac{|K_\ell |!}{|\mathcal {R}(v_{K_\ell })|}\).

Proof

The corollary follows immediately from Theorems 5.3 and 5.5.\(\square \)

6 Modified excited young diagrams

To compute the constant term of Giambelli’s formula we need to evaluate the Peterson Schubert class \(p_{v}\) at \(S^1\)-fixed points of \(Pet\). Our main tool, modified excited Young diagrams, is related both to the work by Woo–Yong[13, Sect. 3] and Ikeda-Naruse [8].

We only need to evaluate at fixed points \(w_K\) where \(K\subseteq \Delta \) is a connected root system. For the remainder of this paper, \(K\) will be a connected root system identified by Lie type.

The first step is to write \(w_K\) explicitly as a skew diagram \(\lambda _{w_K}\). The \(i\)th column from the left represents the simple reflection \(s_i\). Reading left-to-right and top-to-bottom gives a reduced word for \(w_K\). Figure 3 gives several examples.

Our goal was to compute \(p_v(w_K)\). To start we use Eqs. (1) and (2) to rewrite it as

$$\begin{aligned} \begin{array}{rl} p_v(w_K)=\pi _1 (\sigma _v(w_K))&{} =\pi _1 \left( \sum \limits _{\begin{array}{c} \text {reduced words} \\ v=s_{b_{j_1}}s_{b_{j_2}} \cdots s_{b_{j_{\ell (v)}}} \end{array}} \left( \prod \limits _{i=1}^{\ell (v)} \mathbf {r}(\mathbf {j_i},w_K) \right) \right) \\ \\ &{}=\sum \limits _{\begin{array}{c} \text {reduced words} \\ v=s_{b_{j_1}}s_{b_{j_2}} \cdots s_{b_{j_{\ell (v)}}} \end{array}} \left( \prod \limits _{i=1}^{\ell (v)} \pi _1 ( \mathbf {r}(\mathbf {j_i},w_K)) \right) . \end{array} \end{aligned}$$

To help with this computation we build two labeled diagrams. The first is called \(\lambda _{w_K}\) and has boxes labeled by simple reflections. The second diagram is called \(\lambda _{p(w_K)}\) has boxes labeled by integers. We label the \(\mathbf {i}\)th box of \(\lambda _{w_K}\) with \(\frac{1}{t}\cdot \pi _1(\mathbf {r}(\mathbf {i},w_K))\). The term \(\pi _1(\mathbf {r}(\mathbf {i},w_K))\) is a degree-one monomial in \(\mathbb {C}[t]\) whose coefficient is the number of simple roots, counting multiple occurrences, that appear in the root \(\mathbf {r}(\mathbf {i},w_K)\). This number is the height in the root poset of the root \(\mathbf {r}(\mathbf {i},w_K)\). Thus, the labels are positive integers. We give an example in type \(B_3\):

figure b

A \(v\)-excitation \(\mu \) of \(\lambda _{w_K}\) is any collection of \(\ell (v)\) boxes in the labeled diagram that, when read left-to-right and top-to-bottom, gives a reduced word for \(v\). For example, if \(K\) is type \(B_3\) there are three \(s_1s_2\)-excitations of \(\lambda _{w_K}\):

figure c

If \(\mu \) is a \(v\)-excitation of \(\lambda _{w_K}\) then \(M_p(\mu )\) is the product of the entries in the boxes of \(\lambda _{p(w_K)}\) filled by \(\mu \).

figure d

For this \(s_1s_2\)-excitation \(\mu \) of \(w_{B_3}\) the coefficient is \(M_p(\mu )=(4)(5)=20. \) Now \(p_v(w_K)\) can be computed by this diagramatic construction:

$$\begin{aligned} p_v(w_K)=\sum \limits _{\begin{array}{c} \mu \text { a v-excitation}\\ \text { of } w_K \end{array}} M_p (\mu )\cdot t^{\ell (v)}. \end{aligned}$$
(11)
Fig. 3
figure 3

Skew diagrams representing the longest element in the classical Lie types. We use the reduced words for \(w_K\) given by Sage

7 Proof of the Giambelli’s formula

Theorems 5.4 and 5.5 gave two versions of the main theorem of this paper. Having used Theorem 5.4 to prove Theorem 5.5, we now prove Theorem 5.4 case by case according to Lie type. This section first gives a proof for type \(A\) which will be used for the proofs of the other classical types. Then we prove the theorem for the exceptional Lie types. This involves computer-generated proofs for the \(E\) series and explicit calculations for types \(F_4\) and \(G_2\).

Throughout this section we assume that \(K\) is a connected root system of the given Lie type.

7.1 Type A

figure e

While Giambelli’s formula for type \(A\) was fully addressed by Bayegan and Harada [1], we give a proof using modified excited Young diagrams. The diagrams \(\lambda _{w_K}\) and \(\lambda _{p(w_K)}\) are

figure f

The labels of \(\lambda _{p(w_K)}\) can be seen by considering the \(i\)th box of the \(j\)th row. By calling the word made by the top row of \(\lambda _{w_K}\) \(y_n\) and the \(j\)th row \(y_{n-j+1}\) we see that the root associated with that box is

$$\begin{aligned} y_ny_{n-1}\cdots y_{n-j+2}s_1s_2\cdots s_{i-1}(\alpha _i)&=y_ny_{n-1}\cdots y_{n-j+2}(\alpha _1+\cdots + \alpha _i)\\&=y_ny_{n-1}\cdots y_{n-j+3}(\alpha _2+\cdots + \alpha _{i+1})\\&\vdots \\&=\alpha _{j}+\cdots +\alpha _{j+i-1}. \end{aligned}$$

so the label in that box of \(\lambda _{p(w_K)}\) is \(i\). We use the diagrams to prove the type \(A\) case of Theorem 5.4.

Proposition 7.1

Theorem 5.4 holds when \(K\) is a connected type \(A\) root system.

Proof

Lemma  5.1 showed \(C\cdot p_{v_K}=\prod \limits _{1 \le i \le n} p_{s_i}\). Evaluating this equation at the fixed point \(w_K\) gives

$$\begin{aligned} C\cdot p_{v_K}(w_K)=\prod \limits _{1 \le i \le n} p_{s_i}(w_K). \end{aligned}$$
(12)

There is only one filling \(w_K\) with \(v_K=s_1s_2\cdots s_n\), specifically

figure g

Thus, \(p_{v_K}(w_K)=M(\mu )\cdot t^n=n!\cdot t^n\). We must also evaluate \(p_{s_i}(w_K)\) for each \(s_i \in K\). From \(\lambda _{p(w_K)}\) each \(s_i\)-excitation \(\mu \) of \(w_K\) has \(M(\mu )=i\). From the diagram, there are \(n-i+1\) such excitations. So

$$\begin{aligned} p_{s_i}(w_K)=\sum \limits _{\begin{array}{c} \mu \mathrm{a} s_i-\text { excitation}\\ \text { of } w_K \end{array}} M (\mu )\cdot t^{\ell (s_i)}=(n-i+1)i\cdot t. \end{aligned}$$

Solving Eq. (12) for \(C\) we obtain

$$\begin{aligned} C\cdot n! \cdot t^n=\prod \limits _{1 \le i \le n}[(n-i+1)i\cdot t]=\prod \limits _{1 \le i \le n}(n-i+1) \cdot \prod \limits _{1 \le i \le n}(i\cdot t)=(n!)^2\cdot t^n \end{aligned}$$

which gives \(C=n!=|K|!\) as desired.\(\square \)

The type \(A\) result greatly simplifies the proof for the other classical Lie types.

Lemma 7.2

Let \(K\) be a connected root system of type \(B_n,C_n,\) or \(D_n\). Then the root subsystem \(J=K\setminus \lbrace s_n \rbrace \) is type \(A_{n-1}\) and

$$\begin{aligned} c^K_{n,J} \cdot (n-1)! \cdot p_{v_K}=\prod _{\alpha _i \in K} p_{s_i} \end{aligned}$$
(13)

where \(c^K_{n,J}=p_{s_n}(w_K)\cdot \frac{p_{v_J}(w_K)}{p_{v_K}(w_K)}.\)

Proof

By the proof of Giambelli’s formula for type \(A\) and Theorem 4.2 respectively,

$$\begin{aligned} (n-1)! \cdot p_{v_J}=\prod _{\alpha _i \in J} p_{s_i} \quad \quad \text { and} \quad \quad p_{s_n}\cdot p_{v_J}=c^K_{n,J} \cdot p_{v_K}. \end{aligned}$$
(14)

Combining these gives Eq. (13). By Theorem 4.2

$$\begin{aligned} c^K_{n,J}=(p_{s_n}(w_K)-p_{s_n}(w_J))\cdot \frac{p_{v_J}(w_K)}{p_{v_K}(w_K)}. \end{aligned}$$

By construction the root \(\alpha _n\) is not in \(J\) so \(p_{s_n}(w_J)=0\) giving the desired result.\(\square \)

Now we will prove Giambelli’s formula for the other classical Lie types.

7.2 Type B

figure h

Proposition 7.3

Theorem 5.4 holds when \(K\) is a connected type \(B\) root system.

Proof

Let \(K\) be a connected type \(B_n\) root system and \(J\subset K= K\setminus \{ s_n \}\). By Lemma 7.2 showing that \(c^K_{n,J}=n\) is sufficient to prove Giambelli’s formula for type \(B\).

If \(K\) is of type \(B_n\) the diagram of the chosen reduced word for \(w_K\) is given below. Each row is labeled by the word of reflections in that row. For example, \(x_2=s_2s_3\cdots s_{n-1}s_n\).

figure i

To compute \(c_{n,J}^K\) we need to compute \(p_v(w_K)\) where \(v\) is \(s_n\), \(s_1s_2\cdots s_{n-1}\) and \(s_1s_2\cdots s_{n}\). All of the \(v\)-excitations of \(w_K\) for these words are contained in the shaded area of the \(\lambda _{w_K}\) above. So we only need the entries of \(\lambda _{p(w_K)}\) in those shaded boxes. Start with the box labeled \(s_n\) in row \(x_j\) of the diagram. The reflections that come after do not act on the root, so we look at \(x_n x_{n-1} \cdots x_j\) and calculate the root as it moves through the diagram. A bullet, \(\bullet \), marks the location of the root in the diagram at each step. The initial root is below the first diagram and we follow how it changes.

figure j

By the time the bullet gets to the position in the lower left, the root is \(\alpha _j+\cdots \alpha _n\) which is invariant under all simple reflections except \(s_j\) and \(s_{j-1}\). Neither of those reflections act on the bullet as it continues through the diagram. Thus, the label on the box \(s_n\) in row \(x_j\) of \(\lambda _{p(w_K)}\) is \(n-j+1\).

We can start the bullet in any box of the diagram. Suppose that the \(h\)th simple reflection of \(w_k\) is the \(i\)th box in row \(y_{n-1}\). Then \(\mathbf {r}(\mathbf {h},w_K)\) will be

$$\begin{aligned} x_n\cdots x_1 s_1s_2 \cdots s_{i-1}s_{i-1} (\alpha _i)= x_n\cdots x_1 \left( \sum \limits _{m=1}^i \alpha _m \right) . \end{aligned}$$

If \(i<n-1\) then applying the row \(x_1\) to \(\sum \limits _{m=1}^i \alpha _m \) gives \(\sum \limits _{m=1}^i \alpha _{m+1} \). Every subsequent \(x_j\) again increases each index by one until applying \(x_{n-i-1}\) gives \(\sum \limits _{m=n-i-1}^{n-1} \alpha _m \). If \(i=1\) we are done. Otherwise the action of \(s_n\) on this sum results in \(\left[ \sum \limits _{m=n-i-1}^{n-1} \alpha _m \right] +2\alpha _n\). Each subsequent reflection in \(x_{n-i}\) results in the double appearance of another root.

$$\begin{aligned} x_{n-i} \left( \sum \limits _{m=n-i-1}^{n-1} \alpha _m \right) = \alpha _{n-i-1}+ 2\sum \limits _{m=n-i}^{n} \alpha _m \end{aligned}$$

This root is invariant under the action of every reflection in \(x_{n-i+1}\) until the last one, \(s_{n-i+1}\) where

$$\begin{aligned} s_{n-i+1}\left( \alpha _{n-i-1}+ 2\sum \limits _{m=n-i}^{n} \alpha _m\right) =\alpha _{n-i-1}+ \alpha _{n-i}+ 2\sum \limits _{m=n-i+1}^{n} \alpha _m \end{aligned}$$

Subsequently

$$\begin{aligned} x_{n\!-\!i\!+\!1}\left( \alpha _{n-i-1}\!+\! \alpha _{n-i}\!+\! 2\sum \limits _{m=n-i+1}^{n} \alpha _m\right) \!=\!\alpha _{n-i-1}+ \alpha _{n-i}+\alpha _{n-i+1}+ 2\sum \limits _{m=n-i+2}^{n} \alpha _m \end{aligned}$$

and this process continues until the last reflection,

$$\begin{aligned} x_n\left( \alpha _{n-i-1}+ \alpha _{n-i}+\cdots +\alpha _{n-1}+ 2 \alpha _n \right) = \sum \limits _{m=1}^{i} \alpha _{n-m}. \end{aligned}$$

Thus, the entry in the corresponding box of \(\lambda _{p(w_K)}=\frac{1}{t}\pi \left( \sum \limits _{m=1}^{i} \alpha _{n-m} \right) =i\).

Let the \(h\)th simple reflection of \(w_K\) to be the \(i\)th box of row \(x_1\) for some \(i\ne n\). The root \(\mathbf {r}(\mathbf {h},w_K)\) is

$$\begin{aligned} \begin{array}{rl} x_nx_{n-1}\cdots x_{2} s_{1}s_{2} \cdots s_{i-1}(\alpha _i)&{}=x_nx_{n-1}\cdots x_{2} \left( \sum \limits _{m=1}^i \alpha _m \right) \\ &{}=\alpha _1+\alpha _2+\cdots +\alpha _{n-1}+2 \alpha _n + \sum \limits _{m=2}^{i}\alpha _{n-m+1}. \end{array} \end{aligned}$$

Thus, the entry in the corresponding box of \(\lambda _{p(w_K)}\) is \(n+i \). Now we can label the relevant boxes of \(\lambda _{p(w_K)}\):

figure k

With this labeling established, we can see that \(p_{s_n}(w_K)=\frac{n(n+1)}{2}t\). We also observe that there is only one \(v_K\)-excitation of \(w_K\) since \(v_K\) contains, in order, the reflections \(s_1\) through \(s_n\) which appear in that order in row \(x_1\) and in no other subword of \(w_K\). So

$$\begin{aligned} p_{v_K}(w_K)=(n+1)(n+2) \cdots (2n-2)(2n-1)n\cdot t^n=\frac{(2n-1)!}{(n-1)!}t^n. \end{aligned}$$

The last piece is to calculate \(p_{v_J}(w_K)\). All subwords of \(w_K\) that are reduced words for \(v_{J}\) are entirely contained within the \(x_1y_{n-1}\) subword of \(w_K\). We look at the excited Young diagrams in just those two rows of \(\lambda _{p(w_K)}\). A \(v_J\)-excitation \(\mu \) of these two rows is determined by how many boxes it takes from row \(x_1\). The excitation \(\mu \) that uses \(i\) boxes from row \(x_1\) looks like

figure l

This \(v_J\)-excitation contributes coefficient \(M(\mu )=\frac{(n+i)!}{n\cdot i!}\) to \(p_{v_J}(w_K)\). Since \(i\) ranges from 0 to \(n-1\),

$$\begin{aligned} p_{v_J}(w_K)=\sum \limits _{i=0}^{n-1} \frac{(n+i)!}{n\cdot i!} t^{n-1}. \end{aligned}$$

Putting all of the pieces together we have

$$\begin{aligned} \begin{array}{rl} c_{n,J}^{K}&{}=\frac{p_{s_n}(w_K)\cdot p_{v_{J}}(w_K)}{p_{v_K}(w_K)}\\ &{}=\frac{\frac{n(n+1)}{2}t}{\frac{(2n-1)!}{(n-1)!}t^n} \cdot \sum \limits _{i=0}^{n-1} \frac{(n+i)!}{n\cdot i!}t^{n-1}\\ &{}=\frac{(n+1)!}{2n \cdot (2n-1)!} \cdot \sum \limits _{i=0}^{n-1} \frac{(n+i)!}{i!}. \end{array} \end{aligned}$$

By combinatorial identity, the sum can be rewritten and

$$\begin{aligned} c_{n,J}^K=\frac{(n+1)!}{2n \cdot (2n-1)!} \cdot \frac{n\cdot (2n)!}{(n+1)!}=n. \end{aligned}$$

\(\square \)

7.3 Type C

figure m

Proposition 7.4

Theorem 5.4 holds when \(K\) is a connected type \(C\) root system.

Proof

This proof mirrors the proof in type \(B\). Let \(K\) be a connected type \(C_n\) root system and define \(J\subset K\) to be \(J=K\setminus \{ s_n \}\). By Lemma 7.2 showing that \(c^K_{n,J}=n\) is sufficient to prove Giambelli’s formula for type \(C\).

The longest word \(w_K\) is the same for type \(C_n\) as for type \(B_n\), the only changes from type \(B\) are the box labels of \(\lambda _{p(w_K)}\). First we find the label for the box corresponding to reflection \(s_n\) in row \(x_j\).

figure n

The root \(2\alpha _j+2\alpha _{j+1}+\cdots + 2\alpha _{n-1} +\alpha _n\) is invariant under all reflections except \(s_j\) and \(s_{j-1}\) which will not act on the root. Thus, the label in \(\lambda _{p(w_K)}\) is \(2(n-j)+1\). Adding up all the labels of these boxes gives that

$$\begin{aligned} p_{s_n}(w_K)=\sum \limits _{j=1}^n (2j-1)\cdot t =n^2\cdot t. \end{aligned}$$

To compute \(p_{v_J}(w_K)\) and \(p_{v_K}(w_K)\) we need to compute the \(\lambda _{p(w_K)}\) diagram labels of rows \(y_{n-1}\) and \(x_1\). If the \(h\)th box is box \(i\ne n\) of row \(x_1\) then

$$\begin{aligned} \mathbf {r}(\mathbf {h},w_K)=x_nx_{n-1}\cdots x_{2} s_{1}s_{2}\cdots s_{i-1}(\alpha _i)= x_nx_{n-1}\cdots x_{2} \left( \sum \limits _{m=1}^i \alpha _m\right) . \end{aligned}$$

By moving a root through the diagram, \(x_nx_{n-1}\cdots x_{2}(\alpha _1)=\alpha _1+\alpha _2+\cdots +\alpha _n\) and if \(1<i<n\) then \(x_nx_{n-1}\cdots x_{2}(\alpha _i)\) get pushed through the diagram as follows:

figure o

Since we are working in type \(C\) the reflection \(s_{n-1}\) sends \(\alpha _n\) to \(2\alpha _{n-1}+\alpha _n\) so the next row of the diagram acts like this

figure p

Row \(x_{n-i+2}\) eliminates everything except \(\alpha _{n-i+1}\) which is preserved for the rest of the diagram. So

$$\begin{aligned} \mathbf {r}(\mathbf {h},w_K)=x_nx_{n-1}\cdots x_{2} \left( \sum \limits _{m=1}^i \alpha _m\right) = \sum \limits _{m=1}^n \alpha _m +\sum \limits _{m=2}^i \alpha _{n-m+1} \end{aligned}$$

and the entry in \(\lambda _{p(w_K)}\) is \(n+i-1\). Like in type \(B\), \(x_nx_{n-1}\cdots x_2x_1(\alpha _i)= \alpha _{n-i}\) for \(i\ne n\). We can fill in the entries of rows \(x_1\) and \(y_{n-1}\) of \(\lambda _{p(w_K)}\) as follows:

figure q

A \(v_J\)-excitation of \(w_K\) is marked in light gray. Summing over the number of boxes of \(\mu \) that are in row \(x_1\) as in the previous section gives

$$\begin{aligned} p_{v_J}(w_K)=\sum \limits _{i=0}^{n-1} \frac{(n+i-1)!}{i!}t^{n-1}=\frac{(2n-1)!}{n!}t^{n-1}. \end{aligned}$$

We also see that there is only one \(v_K\)-excitation of \(w_K\) so \(p_{v_K}(w_K)=\frac{(2n-1)!}{(n-1)!}t^n\).

Putting all the pieces together we obtain

$$\begin{aligned} c_{n,J}^K=p_{s_n}(w_K)\frac{p_{v_{J}}(w_K)}{p_{v_K}(w_K)}= n^2\frac{\frac{(2n-1)!}{n!}}{\frac{(2n-1)!}{(n-1)!}}=\frac{n^2}{n}=n. \end{aligned}$$

\(\square \)

7.4 Type D

figure r

Proposition 7.5

Theorem 5.4 holds when \(K\) is a connected type \(D\) root system.

Proof

Let \(K\) be a connected type \(D_n\) root system and \(J\subset K= K\setminus \{ s_n \}\). By Lemma 7.2 it suffices to show that \(c^K_{n,J}=\frac{n}{2}\). If \(K\) is a root system of type \(D_n\) then the shape of \(\lambda _{w_K}\) depends on whether \(n\) is even or odd. Figure 4 gives the two diagrams for type \(D_n\). In each of these shapes, there is only one \(v_K\)-excitation of \(w_K\). This subword occurs in the rows \(x_1\) and \(y_{n-1}\) and looks like

figure s

The \(v_J\)-excitations of \(w_K\) are in the same two rows and there are \(n-1\) such excitations. Each excitation \(\mu \) looks like

figure t
Fig. 4
figure 4

The diagrams of \(\lambda _{w_K}\) for \(K\) a type \(D_n\) root system

We need to find the labels of these boxes in \(\lambda _{p(w_K)}\) in order to compute \(p_{v_K}(w_K)\) and \(p_{v_J}(w_K)\). Denote by \(x\) the word obtained from the first \(n-1\) rows of \(w_K\), i.e., \(x=x_{n-1}x_{n-2}\cdots x_2x_1\). We compute \(x(\alpha _i)\) for \( i < n\).

First we examine the action of \(x\) on \(\alpha _i\) for \(1<i \le n-2\). Suppose that \(i<n-2\). Then we take the action of \(x\) row by row to get to the root \(\alpha _{n-2}\). The first reflection in \(x_1\) to not preserve \(\alpha _i\) is \(s_{i+1}\) which sends it to \(\alpha _i+\alpha _{i+1}\). The next reflection, \(s_i\) then brings the root to \(\alpha _{i+1}\) which the rest of the reflections in \(x_1\) preserve. Similarly, \(x_2(\alpha _{i+1})=\alpha _{i+2}\) if \(i+1<n-2\). This pattern continues until

$$\begin{aligned} x_{n-i-2}x_{n-i-3}\cdots x_1(\alpha _i)=\alpha _{n-2} \end{aligned}$$

The action of the next three rows, \(x_{n-i-1},x_{n-i},\) and \(x_{n-i+1}\), depends on whether \(n-i\) is even or odd. If \(n-i\) is odd then the next three rows have form

figure u

The actions of these rows are

$$\begin{aligned} \begin{array}{rcl} x_{n-i-1}(\alpha _{n-2})&{}=&{}\alpha _{n-1}\\ x_{n-i}(\alpha _{n-1})&{}=&{} \alpha _{n-i}+\cdots +\alpha _{n-2}+\alpha _{n-1}\\ x_{n-i+1}(\alpha _{n-i}+\cdots +\alpha _{n-2}+\alpha _{n-1})&{}=&{}\alpha _{n-i} \end{array} \end{aligned}$$

If instead \(n-i\) is even the next three rows look like

figure v

and act by

$$\begin{aligned} \begin{array}{rcl} x_{n-i-1}(\alpha _{n-2})&{}=&{}\alpha _{n}\\ x_{n-i}(\alpha _{n-1})&{}=&{} \alpha _{n-i}+\cdots +\alpha _{n-2}+\alpha _{n}\\ x_{n-i+1}(\alpha _{n-i}+\cdots +\alpha _{n-2}+\alpha _{n})&{}=&{}\alpha _{n-i}. \end{array} \end{aligned}$$

Whether \(n-i\) is odd or even, the root \(\alpha _{n-i}\) is invariant under the action of \(s_j\) for \(j>n-i+1\) so \(x(\alpha _i)=\alpha _{n-i}\) for all \(i\) greater than \(1\) and less than \(n-2\).

If we start with the root \(\alpha _1\) then \(x_{n-3}x_{n-4}\cdots x_{1}(\alpha _1)=\alpha _{n-2}\). The rest of the computation is

$$\begin{aligned} x_{n-1}x_{n-2}(\alpha _{n-2})= {\left\{ \begin{array}{ll} s_{n-1}s_{n-2}s_n(\alpha _{n-2})=\alpha _n &{} \quad \text {if}\quad n\hbox { is odd}\\ s_{n}s_{n-2}s_{n-1}(\alpha _{n-2})=\alpha _{n-1} &{}\quad \text {if}\quad n \hbox { is even.} \end{array}\right. } \end{aligned}$$

Next we address \(x(\alpha _{n-1})\). Going row by row,

$$\begin{aligned}&x_1(\alpha _{n-1})= \alpha _1+\alpha _2+\cdots +\alpha _{n-2}+\alpha _{n-1} \text { and }\\&x_2(\alpha _1+\alpha _2+\cdots ++\alpha _{n-2}+\alpha _{n-1})=\alpha _1 \end{aligned}$$

which is invariant under \(s_j\) for \(j>2\). Thus, \(x(\alpha _{n-1})=\alpha _1\).

The result of these computations is that for \(i<n\) the word \(x\) takes each root \(\alpha _i\) to a root \(\alpha _j\) and therefore \(\pi _1(x(\alpha _i))=t\) for all \(i<n\). Since the label in the \(i\)th box of row \(y_{n-1}\) is the height of the root, \(xs_1s_2 \cdots s_{i-1}(\alpha _i)=x(\alpha _1+\cdots + \alpha _i)\) that box in \(\lambda _{p(w_K)}\) is labeled \(i\).

We also want to find the root corresponding to the \(i\)th box of row \(x_1\). A non-reduced way to write the word preceding that box in \(w_K\) is \(xs_ns_{n-2}s_{n-3}\cdots s_{i+1}s_i\) and the corresponding root is

$$\begin{aligned} \begin{array}{rl} xs_n s_{n-2} s_{n-3} \cdots s_{i+1}s_i(\alpha _i)&{}=x(-\alpha _i-\alpha _{i+1}-\cdots -\alpha _{n-2}-\alpha _n)\\ \\ &{} =-x(\alpha _i+\alpha _{i+1}+\cdots +\alpha _{n-2}) -x(\alpha _n). \end{array} \end{aligned}$$

We compute the action of \(x\) on the root \(\alpha _n\)

$$\begin{aligned}&x_1(\alpha _n)=-(\alpha _1+\alpha _2+\cdots +\alpha _{n-2}+\alpha _n)\\&s_{n-1}(-(\alpha _1+\alpha _2+\cdots +\alpha _{n-2}+\alpha _n))=-(\alpha _1+\alpha _2+\cdots +\alpha _{n-2}+\alpha _{n-1}+\alpha _n)\\&s_{n-2}(\!-\!(\alpha _1\!+\!\alpha _2\!+\!\cdots \!+\!\alpha _{n-2}\!+\!\alpha _{n-1}\!+\!\alpha _n))=\!-\!(\alpha _1\!+\!\alpha _2\!+\!\cdots \!+\!2\alpha _{n-2}+\alpha _{n-1}+\alpha _n) \end{aligned}$$

Each subsequent reflection places the coefficient \(2\) in front of another simple root until

$$\begin{aligned} x_2x_1(\alpha _n)=-(\alpha _1+2\alpha _2+\cdots +2\alpha _{n-2}+\alpha _{n-1}+\alpha _n). \end{aligned}$$

This root is invariant under the action of \(s_i\) for \(i>2\) and thus

$$\begin{aligned} x(\alpha _n)=-(\alpha _1+2\alpha _2+\cdots +2\alpha _{n-2}+\alpha _{n-1}+\alpha _n) \end{aligned}$$

Thus, the root associated with the \(i\)th box of row \(x_1\) is \( xs_n s_{n-2} s_{n-3} \cdots s_{i+1}(\alpha _i)=\)

$$\begin{aligned} -x(\alpha _i+\alpha _{i+1}+\cdots \alpha _{n-2})+ \alpha _1+2\alpha _2+\cdots +2\alpha _{n-2}+\alpha _{n-1}+\alpha _n. \end{aligned}$$

which has height in the root poset \(2n-3-(n-i-1)=n+i-2\). We can now label the rows \(x_1\) and \(y_{n-1}\) in the diagram \(\lambda _{p(w_K)}\).

figure w

This means that \(p_{v_K}(w_K)=\frac{(2n-3)!}{(n-2)!}(n-1) t^n\). Furthermore when \(\mu \) is a \(v_J\)-excitation with \(i\) boxes in the top row we see a contribution to \(p_{v_J}(w_K)\) of \(M(\mu )=(n-1)\frac{(n+i-2)!}{i!}\) . Therefore,

$$\begin{aligned} p_{v_J}(w_K)=\left[ \sum \limits _{i=0}^{n-2} (n-1)\frac{(n+i-2)!}{i!} \right] t^{n-1} = \frac{(2n-3)!}{(n-2)!}t^{n-1}. \end{aligned}$$

The constant \(c_{n,J}^K\) is

$$\begin{aligned} p_{s_n}(w_K)\frac{p_{v_J}(w_K)}{w_{v_K}(w_K)}=p_{s_n}(w_K) \frac{\frac{(2n-3)!}{(n-2)!}t^{n-1}}{\frac{(2n-3)!}{(n-2)!}(n-1) t^n}=p_{s_n}(w_K)\frac{1}{(n-1)t}. \end{aligned}$$
(15)

The polynomial \(p_{s_n}(w_K)\) is computed in the even and odd cases. In both cases, if \(m\) is less than \(n-1\) and there is a box corresponding to \(s_n\) in row \(x_m\), then

$$\begin{aligned}&x_{n-1}x_{n-2}x_{n-3}\cdots x_{m+1}s_ms_{m+1}\cdots s_{n-3}s_{n-2}(\alpha _n)\\&=x_{n-1}x_{n-2}x_{n-3}\cdots x_{m+1}(\alpha _m+\alpha _{m+1}+\cdots +\alpha _{n-3}+\alpha _{n-2} + \alpha _n)\\&=x_{n-1}x_{n-2}x_{n-3}\cdots x_{m+2}(\alpha _m+2\alpha _{m+1}+\cdots + 2\alpha _{n-3}+2\alpha _{n-2} +\alpha _{n-1}+ \alpha _n). \end{aligned}$$

The root \(\alpha _m+2\alpha _{m+1}+\cdots + 2\alpha _{n-3}+2\alpha _{n-2} +\alpha _{n-1}+ \alpha _n\) is invariant under the action of \(s_i\) for \(i>m+1\). Thus, the root corresponding to that box is

$$\begin{aligned} \alpha _m+2\alpha _{m+1}+\cdots + 2\alpha _{n-3}+2\alpha _{n-2} +\alpha _{n-1}+ \alpha _n \end{aligned}$$

and the entry in \(\lambda _{p(w_K)}\) is \(2(n-m)-1\).

If \(n\) is odd then row \(x_{n-1}\) does not contain the reflection \(s_n\) and

$$\begin{aligned} p_{s_n}(w_K)&= \left[ \sum \limits _{\begin{array}{c} 1 \le m \le n-1\\ m \text { odd} \end{array}} 2(n-m)-1 \right] \cdot t\\&= \left[ \sum \limits _{l=1}^{\frac{n}{2}-1} 2n-4l-1 \right] \cdot t\\&= \frac{n^2-n}{2} \cdot t. \end{aligned}$$

If \(n\) is even then the row \(x_{n-1}\) contains only the reflection \(s_n\) and that box corresponds to the root \(\alpha _n\). Since \(\pi (\alpha _n)=t\) we have

$$\begin{aligned} p_{s_n}(w_K)&= \left[ 1+ \sum \limits _{\begin{array}{c} 1\le m \le n-3\\ m \text { odd} \end{array}} 2(n-m)-1 \right] \cdot t\\&= \left[ 1+\sum \limits _{l=1}^{\frac{n}{2}-1} 2n-2(2l-1)-1 \right] \cdot t\\&= \left[ 1+\sum \limits _{l=1}^{\frac{n}{2}-1} 2n-4l+1 \right] \cdot t\\&= \left[ 1+ (\frac{n}{2}-1)(2n+1) - 4\frac{(\frac{n}{2}-1)(\frac{n}{2})}{2} \right] \cdot t\\&=\frac{n^2-n}{2}\cdot t \end{aligned}$$

Thus, for any \(n>3\) regardless of parity, Eq. (15) becomes

$$\begin{aligned} c_{n,J}^K=\frac{n^2-n}{2} t \cdot \frac{1}{(n-1)t}=\frac{n}{2}. \end{aligned}$$

\(\square \)

7.5 Type \(E\)

figure x

Proposition 7.6

Theorem 5.4 holds when \(K\) is a connected type \(E\) root system.

Proof

The Giambelli formula for the Peterson varieties in the exceptional types was calculated using Sage code. While the calculations for types \(F_4\) and \(G_2\) can easily be reproduced by hand, the \(E\) series computations heavily relied on computers. As such the type \(E\) computations will be given with the accompanying code to reproduce the results. Unlike for the infinite series Lie types, if \(K\) is a root system of type \(E_n\), the word \(w_K\) does not give rise to a nice diagram. We present the simple reflections and corresponding data as lists in Fig. 5.

Fig. 5
figure 5

The lists \(L_{w_K}\) and \(L_{p(w_K)}\) for \(K=E_6,E_7,\) and \(E_8\). The bold simple reflection in \(L_{w_{E_7}}\) and \(L_{w_{E_8}}\) is the last occurrence of the reflections \(s_7\) and \(s_8\), respectively

The first list \(L_{w_K}\) gives the ordered simple reflections \(s_{j_i}\) such that \(w_K=s_{j_1}s_{j_2}\cdots s_{j_\ell }\) is the reduced word for \(w_K\) given by the algebraic combinatorics platform Sage. The second list \(L_{p(w_K)}\) is created using a Sage program. For each simple reflection \(s_{j_i}\) of \(w_K\) we record \(\frac{1}{t}\cdot \pi _1(\mathbf {r}(\mathbf {i},w_K))\). The code for this program is available at arXiv:1311.2678.

The only reduced words of \(v_K\) are

$$\begin{aligned} v_K=s_1s_2s_3s_4\cdots s_n=s_2s_1s_3s_4\cdots s_n=s_1s_3s_2s_4\cdots s_n. \end{aligned}$$

Python code can find all sublists of \(L_{w_K}\) that are equal to one of the three corresponding lists. These sublists are the \(v_K\)-excitations \(\mu \) of \(w_K\). For each excitation \(\mu \), \(M(\mu )\) is the product of the entries in \(L_{p(w_K)}\) in the same positions as those in the sublist \(\mu \). We then sum over all such \(\mu \) to get

$$\begin{aligned} p_{v_K}(w_K)=\left[ \begin{array}{ll}\sum \limits _{\begin{array}{c} \mu \,\mathrm{a}\, v_{K}-\mathrm{excitation}\,\mathrm{of }\, L_{w_K} \end{array}}&M(\mu ) \end{array}\right] \cdot t^n. \end{aligned}$$

We also evaluate \(p_{s_i}(w_K)\) by summing all of the entries in \(L_{p(w_K)}\) corresponding to entries which are \(s_i\) in \(L_{w_K}\). This gives the following data for the \(E\) series:

$$\begin{aligned} \begin{array}{c@{\quad }c@{\quad }c@{\quad }c} &{} \text {Type }E_6&{} \text {Type }E_7&{} \text {Type }E_8\\ p_{v_K}(w_K) &{} 887040t^6 &{}661620960t^7 &{}11179629901440t^8\\ p_{s_1}(w_K)&{} 16t &{}34t &{}92t\\ p_{s_2}(w_K)&{}22t &{}49t &{}136t\\ p_{s_3}(w_K)&{} 30t &{}66t &{} 182t \\ p_{s_4}(w_K)&{} 42t &{} 96t &{} 270t\\ p_{s_5}(w_K)&{} 30t &{}75t &{} 220t \\ p_{s_6}(w_K)&{}16t &{} 52t &{}168t\\ p_{s_7}(w_K)&{}&{}27t&{}114t\\ p_{s_8}(w_K)&{}&{}&{}172t \end{array} \end{aligned}$$

This table along with Lemma 5.1 gives us that the \(E\) series Peterson varieties have the following Giambelli’s formula.

$$\begin{aligned} \begin{array}{cccc} E_6: &{} C \cdot p_{v_K}(w_K)=\prod \limits _{1 \le i \le 6} p_{s_i}(w_K) &{} \quad \text {where}\quad &{}C=240=\frac{6!}{3}\\ E_7: &{} C \cdot p_{v_K}(w_K)=\prod \limits _{1 \le i \le 7} p_{s_i}(w_K) &{} \quad \text {where}\quad &{}C=680=\frac{7!}{3}\\ E_8: &{} C \cdot p_{v_K}(w_K)=\prod \limits _{1 \le i \le 8} p_{s_i}(w_K) &{} \quad \text {where}\quad &{}C=13440=\frac{8!}{3}\\ \end{array} \end{aligned}$$

\(\square \)

7.6 Type \(F_4\)

figure y

Proposition 7.7

Theorem 5.4 holds when \(K\) is a connected type \(F_4\) root system.

Proof

If \(K\) is a connected root system of type \(F_4\) then \(J=K \setminus \lbrace s_4 \rbrace \) is a root subsystem of type \(B_3\) and therefore

$$\begin{aligned} c^{K}_{4,J} \cdot p_K=p_{s_4}p_{v_J}= \frac{1}{3!} \prod \limits _{i=1}^4 p_i. \end{aligned}$$

Evaluating at \(w_K=s_4s_3s_2s_3s_1s_2s_3s_4s_3s_2s_3s_1s_2s_3s_4s_3s_2s_3s_1s_2s_3s_1s_2s_1\) gives

$$\begin{aligned} \begin{array}{c@{\quad }c@{\quad }c} &{}p_K(w_K)=18480t^4=2^4\cdot 3\cdot 5 \cdot 7 \cdot 11 \cdot t^4 &{} \\ p_1(w_K)=22t &{}&{} p_2(w_K)=42t\\ p_3(w_K)= 30t&{}&{} p_4(w_K)=16t \end{array} \end{aligned}$$

Since

$$\begin{aligned} c^{K}_{4,J}\cdot p_K(w_K)=\frac{1}{3!}\prod \limits _{i=1}^4 p_i(w_K), \end{aligned}$$

we can solve for \(c^{K}_{4,J}\) to see that \(c^{K}_{4,J}=4\). Thus, \(4! \cdot p_K= (|K|)!\cdot p_K=\prod \limits _{i=1}^4 p_i\).

\(\square \)

7.7 Type \(G_2\)

figure z

Proposition 7.8

Theorem 5.4 holds when \(K\) is a connected type \(G_2\) root system.

Proof

Since in type \(G_2\) the ring \(H_{S^1}^*(Pet)\) has only four Peterson Schubert classes, we give the basis explicitly evaluated at \(e,s_1,s_2,\) and \(s_1s_2s_1s_2s_1s_2\):

$$\begin{aligned} \begin{array}{cccc} p_\emptyset &{} p_{\lbrace 1 \rbrace } &{} p_{\lbrace 2 \rbrace } &{} p_{\lbrace 1,2 \rbrace } \\ \left( \begin{array}{l} 1\\ 1\\ 1\\ 1\\ \end{array}\right) &{} \left( \begin{array}{l} 0\\ t\\ 0\\ 6t\\ \end{array}\right) &{} \left( \begin{array}{l} 0\\ 0\\ t\\ 10t\\ \end{array}\right) &{} \left( \begin{array}{l} 0\\ 0\\ 0\\ 30t^2\\ \end{array}\right) \end{array} \end{aligned}$$

From this basis we see that \(2\cdot p_K = \prod \limits _{i\in K} p_{s_i}\) and of course \(|K|!=2\).\(\square \)