1 Introduction

The main undertaking of fewnomial theory is to bound the number of connected components of the positive part of a variety defined by a system of equations solely in terms of the number of variables n and the total number of monomials N appearing in the system. Since the constitutive monograph [12] of Khovanskiĭ, fewnomial theory has often been studied alongside the theory of exponential sums. After all, the coordinatewise exponential map \(\exp :\mathbb {R}^n \rightarrow \mathbb {R}_+^n\) is a diffeomorphism, and when replacing monomials \(z^\alpha \) by exponentials \(e^{\langle w, \alpha \rangle }\) the fundamental examples (read: Descartes’ rule of signs) remain valid.

Fewnomial theory has also been studied using Gel’fand, Kapranov, and Zelevinsky’s “A-philosophy.” In this approach, one considers the family of all polynomials which can be expressed using a fixed set (the support set) A of exponent vectors \(\alpha \). This family is studied using A-discriminants and Gale duality. See, e.g., [1, 4, 9].

However, an “A-philosophy” for exponential sums has not been properly developed. We are aware only of [15], where the A-discriminant and its Horn–Kapranov uniformization were generalized to the case of exponential sums. We put ourself in this setting and consider the family

$$\begin{aligned} (\mathbb {C}^\times )^A = \left\{ \, f(w) = \sum _{\alpha \in A} c_\alpha \, e^{\langle w,\alpha \rangle } \, \bigg | \, c_\alpha \in \mathbb {C}^\times \,\right\} , \end{aligned}$$

where \(A\subset \mathbb {R}^{1+n}\) is a finite set, and \(N = \# A\). We have many names for those we love; the support set A is also known as the Bohr spectrum (or, simply, spectrum) of an exponential sum \(f \in (\mathbb {C}^\times )^A\).

The work presented in this article emerged from an innocent question of whether a theorem of Katz [11] holds also in the framework of exponential sums. The answer is affirmative, as we show in Theorem 3.5. This theorem has already seen an application in [7] to reduce a fewnomial hypersurface bound from exponential to subexponential in the dimension. Our main line of thought goes, however, in a slightly different direction.

We associate to the spectrum \(A\subset \mathbb {R}^{1+n}\) a combinatorial invariant in the form of a homogeneous polynomial \(P_A(t)\) of degree n, which we call the cuspidal form of A, see Definition 3.1. Here, t denotes the parameters of the Horn–Kapranov uniformization of the A-discriminant \(\check{X}_A\), and hence, the cuspidal form depends also on a choice of Gale dual of A. The name “cuspidal form” reflects the fact that \(P_A(t)\) describes the preimage of the cuspidal locus of \(\check{X}_A\) under the Horn–Kapranov uniformization. In particular, as one observes immediately, the configuration A is dual defective if and only if \(P_A(t)\) vanishes identically.

The core part of this work is to describe the properties of the cuspidal form \(P_A(t)\) as an invariant of the spectrum A. Our main technical results are concerned with factorizations. For example, if A is a diagonal configuration (Definition 4.3), then the cuspidal form \(P_A(t)\) factors as a product of the cuspidal forms of the diagonal configurations of A.

We spend a fair amount of energy describing linear factors of \(P_A(t)\). These are of two distinct types. The first type consists of linear factors corresponding to rows of the Gale dual. These factors correspond to points \(\alpha \in A\) such that \(A \setminus \{\alpha \}\) is dual defective. Such factors are studied in Sects. 4.1 and 5. The complementary type corresponds to discriminant varieties embedded into the cuspidal locus of \(\check{X}_A\). This generalizes results of [13], where the cuspidal form first appeared in the special case of \(n=1\) (when \(P_A(t)\) is itself a linear form).

As an application, we extend Esterov’s characterization from [5, 6] of dual-defective point configurations to the case of exponential sums: A point configuration A is dual defective if and only if it does not contain any iterated circuit (Theorem 6.4). Our proof uses in addition to the properties of the cuspidal form only the pigeonhole principle.

Finally, in Sect. 8 we study the special case \(n=2\) when \(P_A(t)\) is a quadratic form. We consider the rank and the signature of \(P_A(t)\), which are invariant under the choice of Gale dual. It turns out (Theorem 8.1) that the rank of \(P_A(t)\) is at most 3, independent of the number of variables m. The degenerate case that the rank is at most 2 occurs if and only if the spectrum A is contained in a conic section; the type of the conic section is described by the signature of \(P_A(t)\). Spectra A contained in conic sections have appeared in the study of self-dual toric varieties; see, e.g., the Togliatti surface in [3, §4.1], whose spectrum consists of six points on an ellipse.

It is possible to consider, for example, complex instead of real spectra. The A-discriminant variety, in the sense of exponential sums, remains well defined. The characterization of dual-defective point configurations as those who do not contain an iterated circuit remains valid. However, for complex spectra, the relationship to combinatorics is lost, as we cannot, for example, talk about Newton polygons. We have chosen to stay in the real world as this is the most interesting case for applications.

2 Prelude: Gale duality and the Horn–Kapranov uniformization

In this section, we introduce notation and briefly describe some standard constructions and results associated with A-discriminants and adapt them to the setting of exponential sums.

2.1 The setup

Recall that A is a finite subset of \(\mathbb {R}^{1+n}\) with N elements. By ordering the elements of A, we obtain an isomorphism \(\mathbb {C}^A \simeq \mathbb {C}^N\). With an ordering chosen, by abuse of notation, we identify A with the \((1+n)\times N\)-matrix \(A = \big (\alpha _1, \ldots , \alpha _N\big )\). Doing so, we often write \(c_k\) for the coefficient of \(e^{\langle w, \alpha _k\rangle }\) and identify an exponential sum f with its (column) coefficient vector \(c = (c_1, \ldots , c_N)^\mathrm{T}\).

The properties of the exponential sum f which we are interested in (e.g., the existence of a singular point) are invariant of linear changes of variables. That is, if we denote by \(M_{N,n}(\mathbb {R})\) the space of all such matrices A, then we are interested in the orbits under the left action of \({\text {GL}}_{1+n}(\mathbb {R})\).

If \(A \subset \mathbb {Z}^{1+n}\), then the exponential sum is said to be polynomial, and the configuration A is said to be algebraic. In this case, we can associate to f the \((1+n)\)-variate polynomial \(g(z) = \sum _{\alpha \in A} c_\alpha \, z^\alpha \). The analytic variety Z(f) is an infinite covering of the quasi-affine variety \(Z(g) \subset (\mathbb {C}^\times )^{1+n}\). We often descend to the algebraic case in examples.

Two natural assumptions are imposed. Firstly, A is assumed to be pseudo-homogeneous. That is, we assume the existence of a linear form \(\xi \in {\text {Hom}}(\mathbb {R}^{1+n}, \mathbb {R})\) such that \(\langle \xi , \alpha \rangle = 1\) for all \(\alpha \in A\). Secondly, it is assumed that \(\xi \) is unique. These two assumptions are equivalent to that the matrix A is of full rank (equal to \(1+n\)) with the all ones vector in its row span. In particular, the Newton polygon\(\mathscr {N}= {\text {Conv}}\big (\alpha _1, \ldots , \alpha _N\big )\) has dimension n. We do not assume that the columns of A are distinct.

We associate to A the map \(\exp _A\) given by

$$\begin{aligned} \exp _A:\mathbb {C}^{1+n} \rightarrow (\mathbb {C}^\times )^N, \quad w \mapsto \left( e^{\langle w, \alpha _1\rangle }, \ldots , e^{\langle w, \alpha _N\rangle }\right) ^\mathrm{T}, \end{aligned}$$

where \(\mathbb {C}^N\) should be considered as the dual space of \(\mathbb {C}^A\). In the algebraic case, the map \(\exp _A\) parametrizes a torus. By taking Zariski closure, one obtains a toric variety denoted \(X_A\). Its projectively dual \(\check{X}_A\) is known as the A-discriminantal variety, see [8]. In the case that A is algebraic and \(\check{X}_A\) is a hypersurface, its defining polynomial \(D_{A}\) is called the A-discriminant. In the algebraic case, the locus \(D_{A}(c) = 0\) is the closure of the locus of all g which has a singular point in \((\mathbb {C}^\times )^{1+n}\). In the general case, we define \(\check{X}_A \subset (\mathbb {C}^\times )^A\) by that \(f \in \check{X}_A\) if and only if the exponential sum f has a singular point in \(\mathbb {C}^{1+n}\). Notice that we do not take a Zariski closure in the general case, as \(\check{X}_A \subset (\mathbb {C}^\times )^A\) is not necessarily algebraic for real spectra.

Example 2.1

A point configuration A is said to be a pyramid if all points but one is contained in some strict affine subspace. It is straightforward to check that if A is a pyramid, then the discriminant locus \(\check{X}_A\) is empty.

Example 2.2

The codimensionm of a spectrum A is given by \(m = N - n - 1\). A point configuration of codimension 1, which is not a pyramid, is said to be a circuit. Let us consider a circuit in two dimensions. Using the \({\text {GL}}_{1+n}(\mathbb {R})\)-action, we can assume that A is of the form

$$\begin{aligned} A = \left( \begin{array}{c@{\quad }c@{\quad }c@{\quad }c} 1 &{} 1 &{} 1 &{} 1 \\ 0 &{} 1 &{} 0 &{} \alpha _{11} \\ 0 &{} 0 &{} 1 &{} \alpha _{12} \end{array}\right) . \end{aligned}$$

The formula of the discriminant \(D_{A}(c)\) in the algebraic case, from [8, p. 274], generalizes to the binomial exponential expression

$$\begin{aligned} D_{A}(c) = c_3\,\left( \frac{c_0}{\alpha _{11}+\alpha _{12} - 1}\right) ^{\alpha _{11}+\alpha _{12} - 1} - \left( \frac{-c_1}{\alpha _{11}}\right) ^{\alpha _{11}} \left( \frac{-c_2}{\alpha _{12}}\right) ^{\alpha _{12}}. \end{aligned}$$

The multivaluedness of the exponential functions requires some caution when handling this expression. For this reason, we refrain from using the A-discriminant \(D_{A}(c)\) in our analysis.

Remark 2.3

In examples, it is more convenient to consider a family of inhomogeneous n-variate exponential sums. For an exponential sum f in the variables \(z_1, \ldots , z_n\), the corresponding pseudo-homogeneous exponential sum is given by \(e^{z_0} f\). For a family of n-variate inhomogeneous polynomials with support set A, this corresponds to adjoining a top row of all ones in its matrix representation. In this case, \(\xi = (1, 0, \ldots , 0)\).

2.2 The Horn–Kapranov uniformization

Our main tool is the Horn–Kapranov uniformization of the A-discriminant hypersurface from [10]. For exponential sums, this map was deduced in [15, Thm. 1.7]. Step-by-step, the standard deduction of the Horn–Kapranov uniformization (cf. [2, Prop. 4.1]) is sufficient to cover also the case of exponential sums.

A Gale dual of A is a matrix B yielding an exact sequence of \(\mathbb {C}\)-vector spaces

figure a

Since A has full rank, the rank of B is m. Let us denote the rows of B by \(\beta _k\) for \(k= 1, \ldots , N\), and let us introduce coordinates t in \(\mathbb {C}^m\). Let \(\mathbb {C}_B^m\) denote the inverse image of \((\mathbb {C}^\times )^A\) under the map B. The exceptional locus\(\mathscr {H}_B = \mathbb {C}^m \setminus \mathbb {C}_B^m\) is the union of the linear subspaces \(Z(\beta _k)\). Assuming that A is not a pyramid (i.e., assuming that no \(\beta _k = 0\)), then \(\mathscr {H}_B\) is a central hyperplane arrangement.

Theorem 2.4

(Kapranov, Rojas–Rusak) The dual variety \(\check{X}_A\) is parametrized by the map

$$\begin{aligned} \Phi :(\mathbb {C}^\times )^{1+n}\times \mathbb {C}^m_B \rightarrow (\mathbb {C}^\times )^A, \quad \Phi (\omega ; t) = \exp _A(\omega )*(Bt), \end{aligned}$$

where \(*\) denotes componentwise multiplication.

The exponential sum \(f = \Phi (\omega ; t)\) has a singular point with coordinates \(w = -\omega \).

The map \(\Phi \) is far from injective as, for example, the parameter space and \((\mathbb {C}^\times )^A\) have the same dimensions. However, that \(\Phi \) is homogeneous in t and the existence of the linear form \(\xi \) implies that \(\Phi \) parametrizes a strict (multivalued) analytic subvariety of \((\mathbb {C}^\times )^A\).

Remark 2.5

In the algebraic case, the A-discriminantal polynomial has \(1+n\) homogeneities, arising from the matrix A. In Kapranov’s paper [10], these homogeneities were removed by composition with the map \(\exp _B:(\mathbb {C}^\times )^A \rightarrow (\mathbb {C}^\times )^m\). The composite map is, in this algebraic case, a rational function of t. To avoid real powers of linear forms, we settle for the map \(\Phi \) of Theorem 2.4.

Let us end this section with a well-known property of Gale duality. For a square matrix M, we denote its determinant by |M|.

Lemma 2.6

([14, Lem. 2.10]) Let B be any Gale dual of A, and let \(\sigma \subset [N]\) be a choice of \(1+n\) indices with sign \({\text {sgn}}(\sigma ) = \prod _{k\in \sigma }(-1)^k\). Denote by \(A_\sigma \) the maximal cofactor of A obtained by keeping the columns indexed by \(\sigma \) and let \(B_\sigma \) denote the complementary maximal cofactor of B obtained by deleting the rows indexed by \(\sigma \). Then, there is a nonzero constant C(B), independent of \(\sigma \), such that

$$\begin{aligned} |A_\sigma | = C(B){\text {sgn}}(\sigma )\, |B_\sigma |. \end{aligned}$$

Remark 2.7

In the algebraic case, it is natural to assume that \(\mathbb {Z}A = \mathbb {Z}^{1+n}\), in which case \({\bar{A}}\) can be chosen as an integer unimodular matrix. That is, also \({\bar{B}}\) is integer unimodular, and hence, B is a Gale dual in the combinatorial sense. In this case, we have that \(C(B) = \pm 1\), depending on B.

3 The cuspidal form

In this section, we define the cuspidal form \(P_A(t)\). To simplify notation, we impose the assumption that the top row of A is equal to the all ones vector, and we let \({\hat{A}}\) denote the \(n\times N\) matrix obtained from A by deleting the top row.

Let B denote a Gale dual of A. In slight deviation from the notation in Lemma 2.6, we let \(\Sigma \) denote the set of all subsets \(\sigma \subset [N]\) of cardinality n. For each \(\sigma \in \Sigma \), let \({\hat{A}}_\sigma \) denote the maximal cofactor of \({\hat{A}}\) obtained by deleting all columns corresponding to indices \(k \notin \sigma \).

Definition 3.1

We define the cuspidal form\(P_A(t)\) to be

$$\begin{aligned} P_A(t) = \sum _{\sigma \in \Sigma } \left| {\hat{A}}_\sigma \right| ^2 \, \prod _{k \in \sigma } \langle \beta _k, t\rangle . \end{aligned}$$
(1)

The vanishing locus of \(P_A(t)\) will be called the cuspidal locus in \(\mathbb {C}^m\). We say that the \(P_A(t)\) is trivial if it vanishes identically.

The cuspidal form appeared in unpublished works of Dickenstein, Feichtner, and Sturmfels (Dickenstein, personal communication, 2017) in the special case \(n=1\), it was considered in [13].

We have that \(P_A(t)\) is a homogeneous form of degree n in the coordinates t. Hence it defines, in the case that it is nontrivial, a hypersurface in \(\mathbb {P}^{m-1}\). (Except, of course, for the case \(m=1\).)

The cuspidal form \(P_A(t)\) depends on the choice of Gale dual B. As two Gale duals differ by a change of coordinates in \(\mathbb {C}^m\), we consider this dependence to be implicitly understood from that \(P_A(t)\) is written as a polynomial in the variables t. Since we can choose coordinates in \(\mathbb {C}^m\) arbitrarily, one cannot, in general, ask for a combinatorial interpretation of the coefficients of \(P_A(t)\). When there is a canonical choice of coordinates, however, there are reasonable interpretations.

Example 3.2

Let A be a pyramid. That is, there is a point \(\alpha \in A\) such that \(A\setminus \{\alpha \}\) is contained in an affine space of dimension \(n-1\). This is equivalent to that \(\beta _\alpha = 0\). Thus, if \(\alpha \in \sigma \), then \(\langle \beta _\alpha , t\rangle = 0\), and if \(\alpha \ne \sigma \) then \(|{\hat{A}}_\sigma | = 0\). Hence, each term of (1) vanishes, implying that \(P_A(t)\) vanishes identically.

Example 3.3

Let \(m = 1\), and let \(A_k\) denote the \((n+1)\times (n+1)\) submatrix of A obtained by deleting the kth entry and assume, without loss of generality, that \(\alpha _1 = (1, 0, \ldots , 0)^\mathrm{T}\). Then, each \(\sigma \in \Sigma \) containing 1 has \(|{\hat{A}}_\sigma | = 0\), and each \(\sigma \) not containing 1 has \(|{\hat{A}}_\sigma | = |A_k|\) where \(k = k(\sigma )\) is the unique index greater than 1 not contained in \(\sigma \). By Lemma 2.6, we can choose a Gale dual B according to the rule that \(\beta _k = (-1)^k |A_k|\). Hence,

$$\begin{aligned} P_A(t) = t^n \, \left( \prod _{k> 1} (-1)^k |A_k|\right) \sum _{k > 1}(-1)^k |A_k| = {\text {sgn}}(n) \, t^n\, \prod _{k = 0}^{n+2} |A_k| \end{aligned}$$

where \({\text {sgn}}(n) = (-1)^{{n \atopwithdelims ()2}}\). Notice, in particular, that \(P_A(t)\) vanishes identically if and only if \(|A_k| = 0\) for some k, which is equivalent to that A is a pyramid.

3.1 The cuspidal locus of \(\check{X}\)

Recall the assumption that the top row of A consists of the all ones vector. Let us dehomogenize \(\Phi \) by setting \(\omega _0 = 1\). That is, we consider the parametrization map \(\Phi :(\mathbb {C}^\times )^n \times \mathbb {C}^m_B \rightarrow \check{X}_A\). In standard terminology, a point \(f = \Phi (\omega ; t) \in \check{X}_A\) is a cusp of \(\check{X}_A\) if and only if the image of the pushforward

$$\begin{aligned} {\text {d}}\Phi :T_{(w; t)} \left( (\mathbb {C}^\times )^n \times \mathbb {C}^m_B\right) \rightarrow T_{f} \check{X}_A \end{aligned}$$

is an affine space of dimension at most \(N-2 = n+m-1\). That is, f is a cusp if and only if the Jacobian matrix

$$\begin{aligned} J_\Phi (\omega ; t) = \left( \begin{array}{c@{\quad }c@{\quad }c} \alpha _{11}\,\langle \beta _1, t\rangle \,e^{\langle \omega , \alpha _1\rangle }\, &{} \dots &{} \alpha _{N1}\,\langle \beta _N, t\rangle \,e^{\langle \omega , \alpha _N\rangle } \\ \vdots &{} \ddots &{} \vdots \\ \alpha _{1n}\,\langle \beta _1, t\rangle \,e^{\langle \omega , \alpha _1\rangle }\ &{} \dots &{} \alpha _{Nn}\,\langle \beta _N, t\rangle \,e^{\langle \omega , \alpha _N\rangle }\\ \beta _{11} e^{\langle \omega , \alpha _1\rangle }\ &{} \dots &{} \beta _{N1} e^{\langle \omega , \alpha _N\rangle } \\ \vdots &{} \ddots &{} \vdots \\ \beta _{1m} e^{\langle \omega , \alpha _1\rangle }\ &{} \dots &{} \beta _{Nm} e^{\langle \omega , \alpha _N\rangle } \end{array}\right) . \end{aligned}$$

has rank at most \(n+m-1\). (We drop the subindex \(\Phi \) in the notation.)

Theorem 3.4

We have that f is a cusp of \(\check{X}_A\) if and only if there exists a \((w, t) \in (\mathbb {C}^\times )^n\times \mathbb {C}_B^m\) such that \(f= \Phi (\omega ; t)\) and \(P_A(t) = 0\).

Proof

We need to determine when the maximal minors of \(J(\omega ; t)\) vanish. This does not depend on \(\omega \): the common factors \(e^{\langle \omega , \alpha _k\rangle }\) of each column of \(J(\omega ; t)\) can be factored outside of any minor. We compute the maximal minors of J(1; t).

For each submatrix \(S \subset A\) obtained by deleting any number of columns, let \({\hat{S}}\) denote the matrix S with the top row (consisting of all ones) deleted. In particular, \({\hat{\alpha }}\) denotes the column of \({\hat{A}}\) corresponding to the column \(\alpha \in A\).

Let \(J_{k}\) denote the maximal minor of J(1; t) obtained by deleting the kth column. Without loss of generality, we impose two assumptions. Firstly, as permuting the columns only alters the value of a minor by \(\pm 1\), we can assume that \(k = 1\). Secondly, we assume that \(\alpha _{11} = \cdots = \alpha _{1n} = 0\) (i.e., that \(\alpha _1 = (1, 0 \ldots , 0)^\mathrm{T}\)). Indeed, multiplying J(1; t) from the left by

$$\begin{aligned} T = \left( \begin{array}{c@{\quad }c} I_n &{} - {\hat{\alpha }}_1 t^\mathrm{T} \\ 0 &{} I_m \end{array}\right) \end{aligned}$$

corresponds to the translation \({\hat{\alpha }} \mapsto {\hat{\alpha }} - {\hat{\alpha }}_1\). As \(\left| T\right| = 1\), all maximal minors are invariant under this action.

Let \(\Sigma '\) denote the set of all subsets \(\sigma \subset [N]\setminus \{1\}\) of cardinality n, and for each \(\sigma \in \Sigma '\) set \({\bar{\sigma }} = \{1\}\cup \sigma \). By the Laplace expansion in complementary minors (of sizes \(n\times n\) and \(m\times m\))

$$\begin{aligned} J_{1} = \sum _{\sigma \in \Sigma '} {\text {sgn}}(\sigma ) \left( \prod _{k\in \sigma } \langle \beta _k, t\rangle \right) \left| {\hat{A}}_\sigma \right| \, \left| B_{{\bar{\sigma }}}\right| . \end{aligned}$$

We make three observations. Firstly, since \(\alpha _1 = (1,0, \ldots , 0)^\mathrm{T},\) we have that \(\left| A_{{\bar{\sigma }}}\right| = \left| \hat{A}_{\sigma }\right| \). Secondly, according to Lemma 2.6 (assuming that \(C(B) = 1\)), we have that \(\left| B_{\bar{\sigma }}\right| = {\text {sgn}}({\bar{\sigma }})\left| A_{{\bar{\sigma }}}\right| \). Thirdly, we have that \({\text {sgn}}(\sigma ) = {\text {sgn}}({\bar{\sigma }})\). All in all, we conclude that

$$\begin{aligned} J_{1} = \sum _{\sigma \in \Sigma '}\left| {\hat{A}}_\sigma \right| ^2 \left( \prod _{k\in \sigma } \langle \beta _k, t\rangle \right) . \end{aligned}$$

Since \({\hat{\alpha }}_1 = 0\), we have that \(|{\hat{A}}_\sigma | = 0\) for any \(\sigma \) with \(1 \in \sigma \). Hence, \(J_1 = P_A(t)\). \(\square \)

3.2 Katz’ theorem

In the algebraic case, it follows from Katz’ dimension formula [11], see also [8, §1.5], that the Jacobian \(J(\omega ; t)\) has full rank at \((\omega ; t)\) if and only if the Hessian matrix \(H_f(w)\) of \(f = \Phi (\omega ; t)\) at the point \(w = -\omega \) is singular. This statement remains true if one generalizes to exponential sums. It is possible to obtain a proof of this fact by making the appropriate modifications to the exposition in [8, §1.5]; since we have introduced the cuspidal form (1), we prefer the following more direct approach.

Theorem 3.5

Let \(f = \Phi (\omega ; t) \in \check{X}_A\). Then, the Hessian matrix \(H_f(w)\) is singular at \(w = -\omega \) if and only if \(P_A(t) = 0\). In particular, if \(\check{X}_A\) is a hypersurface and f is a smooth point on \(\check{X}_A\), then the Hessian matrix \(H_f(-\,\omega )\) is nonsingular.

Proof

We only need to prove the first part. As in the proof of Theorem 3.4, the torus action can be quoted out. Hence, it suffices to consider the case \(\omega = 0\), where the singular point is located at \(w = 0\). We have that

$$\begin{aligned} f''_{kl}(0) = \sum _{j = 1}^N \alpha _{jk}\alpha _{jl}\langle \beta _j, t\rangle , \quad 1 \le k,l \le n. \end{aligned}$$

In particular, the pth column \(H_p\) of \(H_f(1)\), for \(p = 1, \ldots , n\) is a sum of N vectors

$$\begin{aligned} H_p = \sum _{j = 1}^N\alpha _{jp}\langle \beta _j, t\rangle \, {\hat{\alpha }}_j. \end{aligned}$$

For any two such columns, the jth summands are both multiples of \({\hat{\alpha }}_j\). In particular, we can expand the determinant

$$\begin{aligned} \left| H_f(0)\right| = \sum _{\sigma \in \Sigma }\, \sum _{\pi }\, \left| \alpha _{k\pi (k)} \langle \beta _{k}, t\rangle \, \hat{\alpha }_{k}\right| _{k\in \sigma } \end{aligned}$$

where \(\pi \) runs over the set of all bijections \(\pi :\sigma \rightarrow [n]\). It follows that

$$\begin{aligned} \left| H_f(0)\right| = \sum _{\sigma \in \Sigma }\, \left( \prod _{k \in \sigma } \langle \beta _{k}, t\rangle \right) \, \sum _{\pi }\,\left| \alpha _{k\pi (k)}\, {\hat{\alpha }}_{k}\right| _{k\in \sigma } = \sum _{\sigma \in \Sigma }\, \left( \prod _{k \in \sigma } \langle \beta _{k}, t\rangle \right) \, \left| {\hat{A}}_{\sigma }\right| ^2, \end{aligned}$$

which completes the proof. \(\square \)

4 Properties of the cuspidal form

In this section, we investigate the cuspidal form \(P_A(t)\) as a function of the spectrum A.

4.1 Subconfigurations of A

We denote by \(Z(\beta )\) the hyperplane in \(\mathbb {C}^m\) with equation \(\langle \beta , t\rangle = 0\).

Theorem 4.1

Let \(\alpha \in A\) be such that \(A \setminus \{\alpha \}\) has rank \(1+n\), let \(t'\) be coordinates in \(Z(\beta )\) and \(t = (t',t_m)\) coordinates in \(\mathbb {C}^m\). Then,

$$\begin{aligned} P_{A\setminus \{\alpha \}}(t') = P_A(t)\big |_{Z(\beta )}. \end{aligned}$$

Proof

By assumption, if we write \(A = (A', \alpha )\), then \(A'\) is of rank \(1+n\) and codimension \(m-1\). We can chose a Gale dual B of A of the form

$$\begin{aligned} B = \left( \begin{array}{c@{\quad }c} B' &{} * \\ 0 &{} 1 \end{array}\right) , \end{aligned}$$

where \(B'\) is a Gale dual of \(A'\). The theorem follows from that the restriction of \(P_A(t)\) to \(Z(\beta )\) is given by \(t_m = 0\). First, for any \(\sigma \in \Sigma \) with \(N \in \sigma \), we have that the term of \(P_A(t)\) corresponding to \(\sigma \) has the monomial \(t_m\) as a factor. Second, we have that \(\langle \beta _k, t\rangle \) restricted to \(t_m = 0\) equals \(\langle \beta _k', t'\rangle \) for any \(k = 1, \ldots , N-1\). \(\square \)

Corollary 4.2

If \(A' \subset A\) is such that \(P_{A'}\) is nontrivial, then \(P_A\) is nontrivial.

4.2 Diagonal and upper triangular configurations

Definition 4.3

We say that a point configuration A is upper triangular if it can, after acting by \({\text {GL}}_{1+n}(\mathbb {R})\), be written in the form

$$\begin{aligned} A = \left( \begin{array}{c@{\quad }c@{\quad }c@{\quad }c@{\quad }c} 1 &{} 1 &{} 1 &{} \cdots &{} 1\\ 0 &{} {\tilde{A}}_1 &{} * &{} \cdots &{} *\\ 0 &{} 0 &{} {\tilde{A}}_2 &{} \cdots &{} *\\ \vdots &{} \vdots &{} \vdots &{} \ddots &{} \vdots \\ 0 &{} 0 &{} 0 &{} \cdots &{} {\tilde{A}}_m \end{array}\right) , \end{aligned}$$
(2)

where we associate to each \({\tilde{A}}_j\) the point configuration \(A_j\) defined by \({\hat{A}}_j = (0, {\tilde{A}}_j)\). The configurations \(A_j\) for \(j=1, \ldots m\) are called the diagonal configurations of A. If all elements marked by \(*\) vanish, then the point configuration is said to be diagonal.

Remark 4.4

A configuration A is an iterated circuit in the sense of Esterov [5, Def. 3.15] if and only if after applying an integer affine transformation it is an upper triangular configuration all of whose diagonal configurations are circuits.

Theorem 4.5

Let A be a diagonal point configuration with diagonal configurations \(A_1, \ldots , A_m\). Then, \(P_A(t) = P_{A_1}(t_1) \cdots P_{A_m}(t_m)\).

Proof

By induction, it suffices to consider the case \(m=2\). For a minor \(|{\hat{A}}_\sigma |\) of \({\hat{A}}\) to be nonvanishing, it must hold that the set \(\sigma \in \Sigma \) consists of \(n_j\) columns corresponding to points in \({\tilde{A}}_j\) for \(j=1,2\). Therefore, for each nonvanishing term of \(P_A(t)\) we have the factorization

$$\begin{aligned} |A_\sigma | = |A_{\sigma _1}| \, |A_{\sigma _2}| \end{aligned}$$

where \(\sigma _j \in \Sigma _j\) for \(j=1,2\) and \(\sigma = \sigma _1 \cup \sigma _2\). We can write \(A_j\) and its Gale dual in the block forms

$$\begin{aligned} A_j = \left( \begin{array}{c@{\quad }c} 1 &{} 1 \\ 0 &{} {\tilde{A}}_j \end{array}\right) \quad \text { and } \quad B_j = \left( \begin{array}{c} * \\ {\tilde{B}}_j \end{array}\right) . \end{aligned}$$

Then, a Gale dual of A can be written as

$$\begin{aligned} B = \left( \begin{array}{c@{\quad }c} * &{} * \\ {\tilde{B}}_1 &{} 0 \\ 0 &{} {\tilde{B}}_2 \end{array}\right) . \end{aligned}$$

Hence, for each \(\sigma = \sigma _1 \cup \sigma _2\) it holds that

$$\begin{aligned} \prod _{k\in \sigma } \langle \beta _k, t\rangle = \left( \prod _{k\in \sigma _1} \langle \beta _k, t_1\rangle \right) \left( \prod _{k\in \sigma _2} \langle \beta _k, t_2\rangle \right) , \end{aligned}$$

where we, by abuse of notation, interpret \(\beta _k\) as a row both of B and of \({\tilde{B}}_j\) when \(k \in \sigma _j\). \(\square \)

Remark 4.6

Let A be an iterated circuit. By Example 3.3, we have that \(P_{A_j}(t_j)\) is a nontrivial monomial in the single variable \(t_j\) of degree \(n_j\) for each \(j=1, \ldots , m\). It is straightforward to check that the coefficient of the monomial \(t_1^{n_1}t_2^{n_2}\cdots t_m^{n_m}\) is unchanged if one deletes all the elements of A marked by a star in (2). In particular, it follows from Theorem 4.5 that the monomial of \(P_A(t)\) with exponent vector \((n_1, n_2, \ldots , n_m)\) is given by \(P_{A_1}(t_1) \cdot P_{A_2}(t_2)\cdots P_{A_m}(t_m)\). In particular, \(P_A(t)\) is nontrivial in this case.

5 Linear factors

Corollary 5.1

Let \(\alpha \in A\). If the cuspidal form \(P_{A\setminus \{\alpha \}}\) vanishes identically, then \(\langle \beta _\alpha , t\rangle \) divides \(P_A(t)\).

Proof

This follows from Theorem 4.1 and Hilbert’s Nullstellensatz. \(\square \)

Example 5.2

Consider the point configuration and Gale dual

$$\begin{aligned} A = \left( \begin{array}{c@{\quad }c@{\quad }c@{\quad }c@{\quad }c} 1 &{} 1 &{} 1 &{} 1 &{} 1\\ 0 &{} 1 &{} 2 &{} 0 &{} 1\\ 0 &{} 0 &{} 0 &{} 1 &{} 2 \end{array}\right) \quad \text {and} \quad B^\mathrm{T} = \left( \begin{array}{@{\quad }r@{\quad }r@{\quad }r@{\quad }r@{\quad }r} 2 &{} -\,1 &{} 0 &{} -\,2 &{} 1 \\ 1 &{} -\,2 &{} 1 &{} 0 &{} 0 \end{array}\right) . \end{aligned}$$

Deleting any of the two last columns of A, we obtain a pyramid, which has trivial cuspidal form by Example 3.2. Notice that the corresponding rows of the Gale dual B are parallel. The cuspidal form in this case is \(P_A(t) = 4 t_1(t_2-t_1)\). In particular, Corollary 5.1 cannot be extended to a bijective correspondence between \(\alpha \in A\) with \(P_{A\setminus \{\alpha \}}\) trivial and linear factors of \(P_A(t)\).

Proposition 5.3

Let A be a point configuration such that \(P_A(t)\) is nontrivial. Assume that one (and therefore every) Gale dual B has a set of k parallel rows \(\beta _1, \ldots , \beta _k\). Let \(\beta _j = \gamma _j \beta \) for some \(\beta \ne 0\) and scalars \(\gamma _j\) for \(j= 1, \ldots , k\). Then, \(\langle \beta , t\rangle ^{k-1}\) divides \(P_A(t)\). If, in addition, \(\beta _1 + \cdots + \beta _k = 0\), then \(\langle \beta , t\rangle ^k\) divides \(P_A(t)\).

Proof

Let the columns \(\alpha _1, \ldots , \alpha _k\) of A correspond to the k rows \(\beta _1, \ldots , \beta _k\) of B. By Lemma 2.6, any maximal minor of A not containing all but at most one of the columns \(\alpha _1, \ldots , \alpha _k\) vanishes. In particular, the remaining points \({\hat{\alpha }}_{k+1}, \ldots , {\hat{\alpha }}_N\) are contained in some \(n-k+1\)-dimensional affine subspace. Since A has full rank, we can write A in the form (where we have reordered so that \(\alpha _1, \ldots , \alpha _k\) are the k last columns)

$$\begin{aligned} A = \left( \begin{array}{c@{\quad }c@{\quad }c} 1 &{} 1 &{} 1 \\ {\hat{A}}' &{} * &{} * \\ 0 &{} I_{k-1} &{} \gamma \end{array}\right) \quad \text {and} \quad B = \left( \begin{array}{c@{\quad }c} B' &{} * \\ 0 &{} - \gamma \\ 0 &{} 1\end{array}\right) . \end{aligned}$$
(3)

Here, \(A'\) is an \((n-k+2)\times (N-k)\)-matrix with Gale dual \(B'\), and \(\gamma \in \mathbb {R}^{k-1}\). Set \(|\gamma | = \gamma _1 + \cdots + \gamma _{k-1}\). We have that \(\gamma _j \ne 0\) for each j, as A is not a pyramid. Under these assumptions \(\beta = (0, \ldots , 0, 1)\), so that \(\langle \beta , t\rangle = t_m\).

To prove the first part of the proposition, we note that any maximal minor of \({\hat{A}}\) which does not contain at least \(k-1\) of the rightmost k columns vanishes. Hence, each nonvanishing term of (1) is divisible by \(t_m^{k-1}\).

Let us now compute the coefficient of \(t_m^{k-1}\). If we pick a maximal minor of \({\hat{A}}\) containing all of the k last columns, then the corresponding monomial is divisible by \(t_m^k\). Hence, we need only to consider the maximal minors of \({\hat{A}}\) containing exactly \(k-1\) of the last k columns. The determinant of a \((k-1)\times (k-1)\) submatrix of \((I_{k-1}, \gamma )\) obtained by deleting the jth column is, up to sign, equal to \(\gamma _j\) if \(j = 1, \ldots , {k-1}\), and it is equal to 1 if \(j = k\). Notice, also, that these \(k-1\) columns give a factor \(t_m^{k-1}\) of the corresponding term of \(P_A(t)\). Hence, when computing the coefficient of \(t_m^{k-1}\) in \(P_A(t)\), one should, for the remaining \(n-k+1\) columns of the maximal minor in question, replace the factor \(\langle \beta _\alpha , t\rangle \) with \(\langle \beta _\alpha ', t'\rangle \) where \(t' = (t_1, \ldots , t_{m-1})\) and \(\beta '_\alpha \) is the corresponding row of \(B'\). Let \(\Sigma '\) denote the set of all subsets of \([N-k]\) of size \(n-k+1\). All in all, we find that the coefficient of \(t_m^{k-1}\) in the cuspidal form \(P_A(t)\) is

$$\begin{aligned}&\sum _{\sigma \in \Sigma '} \, \left| {\hat{A}}'_\sigma \right| ^2\,\left( \prod _{j \in \sigma } \langle \beta '_j, t'\rangle \right) \, \left( \left( \prod _{j=1}^{k-1} (-\gamma _j) \right) + \sum _{i=1}^{k-1} \gamma _i^2 \prod _{j \ne i} (-\gamma _j) \right) \\&\qquad = (-1)^{k-1}\,\gamma _1 \cdots \gamma _{k-1} \,\left( 1 - |\gamma |\right) \, P_{{\hat{A}}}(t'). \end{aligned}$$

Thus, if \(|\gamma | = 1\), which is equivalent to the original assumption that \(\beta _1 + \cdots + \beta _k = 0\), then \(P_A(t)\) is divisible by \(t_m^k\). \(\square \)

Example 5.4

Let us stay in the situation considered in Proposition 5.3. The coefficient of \(t_m^{k-1}\) obtained at the end of that proof is

$$\begin{aligned} (-1)^{k-1}\,\gamma _1 \cdots \gamma _{k-1} \,\left( 1 - |\gamma |\right) \, P_{{\hat{A}}}(t'), \end{aligned}$$

where \(\gamma _j \ne 0\) for \(j = 1, \ldots , {k-1}\). It can happen that this coefficient vanishes even if \(|\gamma | \ne 1\), as \(P_{{\hat{A}}}(t')\) can be trivial. For example, consider the point configuration and Gale dual

$$\begin{aligned} A = \left( \begin{array}{c@{\quad }c@{\quad }c@{\quad }c@{\quad }c@{\quad }c@{\quad }c@{\quad }c@{\quad }c} 1 &{} 1 &{} 1 &{} 1 &{} 1 &{} 1 &{} 1 &{} 1 &{} 1\\ 0 &{} 1 &{} 0 &{} 0 &{} 1 &{} 1 &{} 0 &{} 0 &{} 1\\ 0 &{} 0 &{} 1 &{} 0 &{} 1 &{} 0 &{} 0 &{} 0 &{} 1\\ 0 &{} 0 &{} 0 &{} 1 &{} 0 &{} 1 &{} 0 &{} 0 &{} 1\\ 0 &{} 0 &{} 0 &{} 0 &{} 0 &{} 0 &{} 1 &{} 0 &{} 3\\ 0 &{} 0 &{} 0 &{} 0 &{} 0 &{} 0 &{} 0 &{} 1 &{} 2 \end{array}\right) \quad \text {and} \quad B = \left( \begin{array}{r@{\quad }r@{\quad }r} 1 &{} 1 &{} 7 \\ -\,1 &{} -\,1 &{} -\,1 \\ -\,1 &{} 0 &{} -\,1\\ 0 &{} -\,1 &{} -\,1\\ 1 &{} 0 &{} 0 \\ 0 &{} 1 &{} 0 \\ 0 &{} 0 &{} -\,3\\ 0 &{} 0 &{} -\,2\\ 0 &{} 0 &{} 1 \end{array}\right) . \end{aligned}$$

Here, that last three rows of the Gale dual are parallel, but they do not sum to zero. The cuspidal form corresponding to this choice of Gale dual is

$$\begin{aligned} P_A(t) = 6 \,t_3^3\, \left( 7 t_3^2 + 2t_2 t_3 - 5 t_2^2 + 2t_1t_3 + 2t_1t_2-5t_1^2\right) . \end{aligned}$$

6 Defect duals

A point configuration A and, in the algebraic case, the toric variety \(X_A\), are said to be dual defective if the dual variety \(\check{X}_A\) has codimension at least 2. The following theorem is an immediate consequence of the properties of the Jacobian matrix.

Theorem 6.1

The point configuration A is dual defective if and only if \(P_A(t)\) is trivial.

Proof

If A is dual defective, then the rank of the Jacobian matrix J(wt) is at most \(n+m-1\) for generic (wt). It follows that its maximal minors, which are polynomial in (wt), vanish for generic (wt), and hence, they vanish identically. Conversely, if \(\check{X}_A\) is a hypersurface, as the smooth locus of \(\check{X}_A\) is nonempty, we can find a point \(f = \Phi (w; t)\) for which the Jacobian matrix J(wt) has rank \(n+m\). \(\square \)

Example 6.2

Let A be a pyramid. We saw in Example 3.2 that the polynomial \(P_A(t)\) is trivial, implying (the well-known result) that \(X_A\) is dual defective.

Proposition 6.3

If A has a subset \(A'\), of full dimension, such that \(A'\) is not dual defective, then A is not dual defective.

Proof

This (well-known result) follows from Corollary 4.2. \(\square \)

Let us now recover Esterov’s results on dual-defective point configuration from [5, Lem. 3.17] and [6, Cor. 3.20] and extend it from the algebraic case to the case of exponential sums.

Theorem 6.4

A spectrum \(A \subset \mathbb {R}^{1+n}\) is dual defective if and only if A does not contain any iterated circuit.

Proof

By Theorem 6.1 and Proposition 6.3, we translate to the equivalent statement that \(P_A(t)\) is trivial if and only if A does not contain an iterated circuit. The only if-direction is proven in Remark 4.6.

Let us prove the if-direction. Assume that \(P_A(t)\) is nontrivial. We use a double induction over the codimension m and the dimension n, where the base cases \(m=1\) for arbitrary dimension are covered by Example 3.3.

If \(P_{A\setminus \{\alpha \}}(t)\) is nontrivial for some \(\alpha \in A\), then by induction on codimension, \(A \setminus \{\alpha \}\) contains an iterated circuit, implying that A contains an iterated circuit as well. Hence, it suffices to consider the case when \(P_{A\setminus \{\alpha \}}\) is trivial for all \(\alpha \in A\).

By Corollary 5.1, we have that \(\langle \beta , t \rangle \) divides \(P_A(t)\) for all rows \(\beta \) of B. However, B has N rows, while \(P_A(t)\) is a homogeneous, nontrivial polynomial of degree n. By the pigeonhole principle, the Gale dual B must have parallel rows. For a family \(\beta _1, \ldots , \beta _k\) of parallel rows of B, where \(k > 1\), let us borrow the notation from the proof of Proposition 5.3: that \(\beta _j = \gamma _j \beta \) for some scalars \(\gamma _j \ne 0\) for \(j= 1, \ldots , k\), and for some \(\beta \ne 0\). Then, again by a comparison of the number of rows of B and the degree of \(P_A(t)\), there must exist a family, \(\beta _1, \ldots , \beta _k\) of parallel rows of B, where \(k > 1\), such that \(\langle \beta , t\rangle ^k\) does not divide \(P_A(t)\). Let us fix such a family.

Let us write A and B as in (3), and let \(\gamma = (\gamma _1, \ldots , \gamma _k)^\mathrm{T}\) be as in the proof of Proposition 5.3. Since \(\langle \beta , t\rangle ^k\) does not divide \(P_A(t)\), we have that \(|\gamma _k| \ne 1\). Therefore, the \(k\times k\)-submatrix of A

$$\begin{aligned} \left( \begin{array}{c@{\quad }c} 1 &{} 1 \\ I_{k-1} &{} \gamma \end{array} \right) \end{aligned}$$

has rank k. (Recall that each \(\gamma _k \ne 0\), for otherwise A, is a pyramid.) This has two consequences. First, we have that \((0, I_{k-1}, \gamma )\) is a k-dimensional circuit. Second, by applying an integer affine transformation, we can eliminate all entries marked by \(*\) in (3). Thus, by Theorem 4.5, and with the notation of (3), it holds that \(P_{A'}(t)\) is a factor of \(P_A(t)\). In particular, \(P_{A'}(t)\) is nontrivial, and hence, \(A'\) contains an iterated circuit of dimension \(k-1\), by induction on codimension and dimension. It follows that A contains an iterated circuit. \(\square \)

7 Rationality of the cuspidal locus

Let A be algebraic and choose a Gale dual B of A. We focus in this section on the reduced A-discriminant, whose Horn–Kapranov uniformization \(\Psi _A(t)\) is given by \(t \mapsto \Phi (1; t)^B = (Bt)^B\), see [10] and Remark 2.5. We denote the reduced A-discriminant by \(\check{X}_B\).

The case \(n=1\) was studied in detail in [13], where the cuspidal form \(P_A(t)\) appeared in this special case. For \(n=1\), the cuspidal form is, for every m, a nontrivial linear form in t vanishing along some hyperplane in \(\mathbb {P}^{m-1}\). In particular, the cuspidal locus of \(\check{X}_A\) is always unirational, and if it is a subvariety of \(\check{X}_A\) of codimension one, then it is rational. Actually, a stronger statement holds: there is a point configuration E, with Gale dual F, such that the cuspidal locus of the reduced discriminant \(\check{X}_B\) is isomorphic to the reduced discriminantal variety \(\check{X}_{F}\). Let us here explain the corresponding result for general n.

Theorem 7.1

Let A be algebraic with Gale dual B. For each rational linear subspace L of \(P_A(t)\) not contained in the exceptional locus \(\mathscr {H}_B\) (see Sect. 2.2), there is an algebraic point configuration E, with Gale dual F, and a canonical morphism which embeds the reduced discriminant \(\check{X}_F\) into the cuspidal locus of \(\check{X}_{B}\).

Proof

Express L in implicit form as the image an injective linear transformation \(M:\mathbb {P}^{k-1} \rightarrow \mathbb {P}^{m-1}\), where by assumption M can be chosen as an integer matrix. Since \(ABM = 0\), we can extend A (by adding rows) to an integer matrix E yielding an exact sequence

figure b

It follows that \(\left( \Psi _A \circ M\right) ^M = \Psi _{E}\), and, hence, the morphism \(\Psi _A\circ M\circ \Psi _E^{-1}\) embeds \(\check{X}_F\) into the cuspidal locus of \(\check{X}_E\). \(\square \)

The singular locus of \(\check{X}_A\) is in general not rational. Let us give a simple example.

Example 7.2

Let \(n=3\) and \(N = 7\), so that \(P_A(t)\) is a cubic form in three variables. In particular, if \(P_A(t)\) is nonsingular, then the cuspidal locus of \(\check{X}_A\) is birational to a smooth cubic curve. Such a curve is unirational but not rational. For an explicit example when \(P_A(t)\) is nonsingular, we present the point configuration

$$\begin{aligned} A = \left( \begin{array}{c@{\quad }c@{\quad }c@{\quad }c@{\quad }c@{\quad }c@{\quad }c} 1 &{} 1 &{} 1 &{} 1 &{} 1 &{} 1 &{} 1\\ 0 &{} 1 &{} 1 &{} 2 &{} 3 &{} 3 &{} 3\\ 3 &{} 0 &{} 2 &{} 2 &{} 1 &{} 2 &{} 3\\ 0 &{} 3 &{} 2 &{} 2 &{} 2 &{} 3 &{} 0\end{array}\right) . \end{aligned}$$

8 The bivariate case: the signature of the quadratic form

In the case that \(n=2\) the polynomial \(P_A(t)\) is a quadratic form. Being defined only up to choice of coordinates, let us in this section consider the (real) invariants given by the rank and the signature of \(P_A(t)\). Recall from the literature that each real m-variate quadratic form is defined by a real symmetric \(m\times m\)-matrix Q. We write the signature \(s_A\) of \(P_A(t)\) as the triple \(s_A = (s_+, s_-; s_0)\) where \(s_\pm \) denotes the number of positive, respectively, and negative eigenvalues of Q, and \(s_0\) denotes the algebraic multiplicity of zero as an eigenvalue of Q. It holds that \(s_+ + s_- + s_0 = m\), and the rank of \(P_A(t)\) is given by \(s_+ + s_-\).

Theorem 8.1

Let \(n=2\) and let \(s_A\) denote the signature of \(P_A(t)\). In each of the below cases, it is assumed that A does not belong to any previous cases. Then,

  1. (i)

    \(s_A = (0,0; m)\) if and only if A is a pyramid.

  2. (ii)

    \(s_A = (1,0; 0)\) if and only if A is contained in a nonreal parabola.

  3. (iii)

    \(s_A = (0,1; m-1)\) if and only if A is contained in a real parabola or in two parallel lines.

  4. (iv)

    \(s_A = (1,1; m-2)\) if and only if A is contained in a hyperbola or in two nonparallel lines.

  5. (v)

    \(s_A = (0,2; m-2)\) if and only if A is contained in an ellipse.

  6. (vi)

    \(s_A = (1,2; m-3)\) in all other cases.

Remark 8.2

In the case (ii) of Theorem 8.1, it holds that \(m=1\) so that A consists of four points in the plane. As \(P_A(t)\) is nontrivial, A is a circuit. We can read of the signature of the circuit A from Example 3.3: the Newton polygon \(\mathscr {N}\) is a simplex.

More generally, as soon as A has codimension \(m=1\), then one can read of the combinatorial type of A from the cuspidal form \(P_A(t)\) using Example 3.3. Indeed, the Newton polygon \(\mathscr {N}\) is a simplex if and only if \(s_A = (1,0; 0)\), it is a quadrilateral if and only if \(s_A = (0,1; 0)\), and A is a pyramid if and only if \(s_A= (0, 0; 1)\).

To prove Theorem 8.1, we write A in the form

$$\begin{aligned} A = \left( \begin{array} {c@{\quad }c@{\quad }c@{\quad }c@{\quad }c@{\quad }c} 1 &{} 1 &{} 1 &{} 1 &{} \cdots &{} 1 \\ 0 &{} 1 &{} 0 &{} \alpha _{11} &{} \cdots &{} \alpha _{m1} \\ 0 &{} 0 &{} 1 &{} \alpha _{12} &{} \cdots &{} \alpha _{m2} \end{array}\right) . \end{aligned}$$
(4)

We write \(\alpha _k = (\alpha _{k1}, \alpha _{k2})^\mathrm{T}\), for simpler notation. We set \(|\alpha _k| = \alpha _{k1} + \alpha _{k2}\) for all k. We choose the dual matrix

$$\begin{aligned} B^\mathrm{T} = \left( \begin{array}{c@{\quad }c@{\quad }c@{\quad }c@{\quad }c@{\quad }c@{\quad }c@{\quad }c} |\alpha _1|-1 &{} -\,\alpha _{11} &{} -\,\alpha _{12} &{} 1 &{} 0 &{} \cdots &{} 0\\ |\alpha _2|-1 &{} -\,\alpha _{21} &{} -\,\alpha _{22} &{} 0 &{} 1 &{} \cdots &{} 0\\ \vdots &{} \vdots &{} \vdots &{} \vdots &{} \vdots &{} \ddots &{} \vdots \\ |\alpha _m|-1 &{} -\,\alpha _{m1} &{} -\,\alpha _{m2} &{} 0 &{} 0 &{} \cdots &{} 1 \end{array}\right) . \end{aligned}$$
(5)

Lemma 8.3

Let \(n=2\), and let A and B be as in (4) and (5). Then, the quadratic form \(P_A(t)\) is given by the matrix \(Q = \left( g(\alpha _k, \alpha _j) \right) _{k,j}\) where \(1 \le k,j \le m\) and

$$\begin{aligned} g(\alpha _k, \alpha _j)= & {} \frac{1}{2} \bigg ( \alpha _{k1} \alpha _{j2} (1-\alpha _{k1}-\alpha _{j2}) + \alpha _{k2} \alpha _{j1} (1-\alpha _{k2}-\alpha _{j1}) \\&+ \,(\alpha _{k1} \alpha _{j2} - \alpha _{k2} \alpha _{j1})^2\bigg ). \end{aligned}$$

Proof

By Theorem 4.1, it suffices to consider the case \(m=2\), which is a straightforward computation. \(\square \)

Let us introduce notation for the following \(k\times k\)-minor of the matrix Q from Lemma 8.3:

$$\begin{aligned} G_k\,\Bigg (\begin{matrix} \alpha _1, \ldots , \alpha _k \\ \delta _1, \ldots , \delta _k \end{matrix}\Bigg ) = \left| \begin{array}{ccc} g(\alpha _1, \delta _1) &{} \cdots &{} g(\alpha _1, \delta _k)\\ \vdots &{} \ddots &{} \vdots \\ g(\alpha _k, \delta _1) &{} \cdots &{} g(\alpha _k, \delta _k) \end{array}\right| . \end{aligned}$$

From this point on, proving the above statements is a matter of endurance during computations. We avoid most details in this presentation. We invite the reader to verify the following claims (preferably using a computer).

Lemma 8.4

The polynomial \(G_k\) vanishes identically if \(k \ge 4\).

Proof

The case \(k> 4\) follows from the case \(k=4\) by a Laplace expansion. The case \(k=4\) is a straightforward computation. \(\square \)

Thus, only the polynomials \(G_2\) and \(G_3\) are relevant for our investigation. The polynomial \(G_2\) has, when expanded, 96 terms. The polynomial \(G_3\) is simpler; it admits a factorization

$$\begin{aligned} G_3\,\Bigg (\begin{matrix} \alpha _1, \alpha _2, \alpha _3 \\ \delta _1, \delta _2, \delta _3 \end{matrix}\Bigg ) = \frac{1}{4} \,H(\alpha _1, \alpha _2, \alpha _3) \cdot H(\delta _1, \delta _2, \delta _3) \end{aligned}$$
(6)

where \(H(\alpha _1, \alpha _2, \alpha _3)\) is the following polynomial with (when expanded) 24 terms:

$$\begin{aligned} H(\alpha _1, \alpha _2, \alpha _3)&= \alpha _{11} \alpha _{12} \alpha _{22}\alpha _{31} (1 - \alpha _{22}) (1 - \alpha _{31})\\&\quad - \alpha _{11} \alpha _{12} \alpha _{21} \alpha _{32} (1 - \alpha _{21})(1 - \alpha _{32})\\&\quad + \alpha _{12} \alpha _{21}\alpha _{31} \alpha _{32}(1 - \alpha _{12}) (1 - \alpha _{21} )\\&\quad - \alpha _{12} \alpha _{21} \alpha _{22} \alpha _{31} (1 - \alpha _{12})(1 - \alpha _{31})\\&\quad - \alpha _{11} \alpha _{22} \alpha _{31} \alpha _{32} (1 - \alpha _{11}) (1 - \alpha _{22}) \\&\quad + \alpha _{11}\alpha _{21} \alpha _{22} \alpha _{32} (1 - \alpha _{11}) (1 - \alpha _{32}). \end{aligned}$$

Lemma 8.5

Assume that \(0, e_1, e_2, \alpha _1\), and \(\alpha _2\) are five distinct points in \(\mathbb {R}^2\). Then, the polynomial \(G_2\) vanishes for \(\delta _1 = \alpha _1\) and \(\delta _2 = \alpha _2\) if and only if there is a parabola containing the five points.

Proof

We assume that \(\alpha _{11}(1-\alpha _{11}) \ne 0\), as this is the most difficult case. The cases \(\alpha _{11}=0\) or \(\alpha _{11} = 1\) can be treated in analogous fashion. The general equation of a parabola passing through the points \(0, e_1\), and \(e_2\) is \(P(x) = 0\) where

$$\begin{aligned} P(x) = a^2 x_1^2 + 2 a b x_1 x_2 + b^2 x_2^2 - a^2 x_1 -b^2x_2. \end{aligned}$$
(7)

Requiring in addition that the parabola passes through \(\alpha _1\) gives that, in projective coordinates

$$\begin{aligned}{}[a: b ] = \big [\, \alpha _{11}\alpha _{12} \pm \sqrt{\alpha _{11}\alpha _{12}(\alpha _{11} + \alpha _{12} -1)}\, : \, \alpha _{11} ( 1- \alpha _{11}) \big ]. \end{aligned}$$
(8)

(Recall that we assume \(\alpha _{11} ( 1- \alpha _{11}) \ne 0\).) Let \(P_1(x)\) and \(P_2(x)\) denote the two parabolic equations obtained from the possible choices of signs. We leave it to the reader to verify that

$$\begin{aligned} P_1(\alpha _2)P_2(\alpha _2) = 4\, \alpha _{11}^2(1-\alpha _{11})^2\, G_2\,\Bigg (\begin{matrix} \alpha _1, \alpha _2 \\ \alpha _1, \alpha _2 \end{matrix}\Bigg ). \end{aligned}$$

\(\square \)

Remark 8.6

Assume that A, written in the form (4), is contained in a parabola C. For any indices ij,  and k, we obtain a quadratic polynomial

$$\begin{aligned} R_{ijk}(x) = G_2\Bigg (\begin{matrix} \alpha _i, \alpha _j \\ \alpha _k,\,\, x \end{matrix}\Bigg ). \end{aligned}$$

Let us assume that \(R_{121}\) is nontrivial, implying that the points \(0, e_1, e_2, \alpha _1\), and \(\alpha _2\) are in general position (in the sense that there is a unique conic passing through them). By Lemma 8.5, \(R_{121}(x)\) vanishes for \(x = \alpha _2\). It is straightforward to verify that \(R_{121}(x)\) also vanishes for \(x = 0, e_1, e_2\), and \(\alpha _1\). Hence, \(R_{121}(x)\) defines the parabola C. In particular, \(R_{121}(x)\) vanishes for all \(\alpha \in A\).

Assume now that A has at least six points, and in addition that \(R_{123}(x)\) is nontrivial. We have that \(R_{123}(x)\) vanishes at \(x = \alpha _2\) and at \(x = \alpha _3\), as it coincides (up to a constant) with \(P_{232}(\alpha _1)\), respectively, \(P_{323}(\alpha _1)\) for those values. It is straightforward to check that \(R_{123}(x)\) also vanishes for \(x = 0, e_1,\) and \(e_2\). Hence, also \(R_{123}(x)\) defines the parabola C. In particular, \(R_{123}(x)\) vanishes for all \(\alpha \in A\).

Lemma 8.7

Assume that \(0, e_1, e_2, \alpha _1\), \(\alpha _2\), and \(\alpha _3\) are six distinct points in \(\mathbb {R}^2\). Then, the polynomial \(H(\alpha _1, \alpha _2, \alpha _3)\) from (6) vanishes if and only if there is a conic containing all six points.

Proof

The general equation of a conic passing through the points \(0, e_1\), and \(e_2\) is \(P(x) = 0\) where

$$\begin{aligned} P(x) = a x_1^2 + b x_1 x_2 + c x_2^2 -a x_1 -cx_2 \end{aligned}$$
(9)

Requiring in addition that the conic passes through \(\alpha _1\) and \(\alpha _2\) gives the coefficients, up to multiplication by a constant,

$$\begin{aligned} \left\{ \begin{array}{lll} a &{} = &{}\alpha _{12} \alpha _{22} ( \alpha _{21} -\alpha _{11} - \alpha _{12} \alpha _{21} + \alpha _{11}\alpha _{22})\\ b &{} = &{}\alpha _{12} \alpha _{21}(1-\alpha _{12})(1-\alpha _{21}) - \alpha _{11} \alpha _{22}(1-\alpha _{11})(1-\alpha _{22})\\ c &{} = &{}\alpha _{11} \alpha _{21} (\alpha _{12} - \alpha _{12} \alpha _{21} - \alpha _{22} + \alpha _{11} \alpha _{22}) \end{array}\right. \end{aligned}$$

We leave it for the reader to verify that, with these coefficients, the polynomial P(x) evaluated at \(x = \alpha _{3}\) is equal to \(H(\alpha _1, \alpha _2, \alpha _3)\). \(\square \)

Proof of Theorem 8.1

It is well known that for \(n=2\) the point configuration A is dual defective if and only if A is a pyramid. This proves part (i) of the theorem.

If the codimension \(m=1\), then A consists of four points in the plane, through which two parabolas pass. It can be seen from (8) that the limiting case between real and nonreal parabolas is the case of a pyramid. Thus, this case follows from Example 3.3.

We now assume that \(m > 1\). Classical geometry says that there is a unique conic passing through five points in generic position (i.e., no four are colinear) in the plane. We allow the conic to be degenerate. Assuming that A is not a pyramid, it has a subconfiguration \(A_1\) of five points in generic position. Let \(A_1\) constitute the first five columns of (4).

Let us first prove the relaxed statement, where we only consider the rank of the cuspidal form \(P_A(t)\). It follows from Lemma 8.4 that the rank is at most three.

If the rank is at most two, then by Lemma 8.7 any choice of six points of A is contained in a conic. However, the five points of \(A_1\) determine a unique conic C. By adjoining the remaining points one by one, we conclude that A is contained in the conic C. Conversely, if A is contained in a conic, then \(G_3\) vanishes for all \(\alpha \) and \(\delta \) by (6).

If the rank is at most one, then by Lemma 8.5 any choice of five points is contained in a parabola. In particular, the five points of \(A_1\) are contained in a parabola C. Let \(\alpha \) be an additional point. Then, by Lemma 8.7, \(A_1 \cup \{\alpha \}\) is contained in a conic. But C is the unique conic containing \(A_1\), so \(\alpha \in C\). It follows that A is contained in C. Conversely, if A is contained in a parabola, then \(G_2\) vanishes for all \(\alpha \) and \(\delta \) by Lemma 8.5 and Remark 8.6.

Let us now turn to the refined statement of Theorem 8.1 regarding signatures. For each class in the above list, it suffices to consider the minimal m such that there is a point configuration in this class. Indeed, there is a “minimal” subconfiguration witnessing the class containing A, and by Theorem 4.1 we can delete the remaining points without altering the rank. In particular, parts (ii) and (iii) follow from the above discussion on codimension \(m=1\).

Any two minimal configurations of one class can be continuously deformed to each other without leaving the class in question. Therefore, it suffices to consider one representative of each class. Since a generic configuration can be continuously deformed to a configuration contained either in an ellipse or in a parabola, part (vi) follows from parts (iv) and (v) and the fact that the rank is 3 in the generic case. Hence, we finish the proof with Examples 8.8 and 8.9. \(\square \)

Example 8.8

Consider the point configuration and Gale dual

$$\begin{aligned} A = \left( \begin{array}{r@{\quad }r@{\quad }r@{\quad }r@{\quad }r} 1 &{} 1 &{} 1 &{} 1 &{} 1\\ 3 &{} -\,3 &{} 5 &{} 5 &{} -\,5\\ 0 &{} 0 &{} 4 &{} -\,4 &{} 4 \end{array}\right) \quad \text {and} \quad B^\mathrm{T} = \left( \begin{array}{r@{\quad }r@{\quad }r@{\quad }r@{\quad }r} 5 &{} -\,5 &{} -\,3 &{} 0 &{} 3 \\ -\,8 &{} 2 &{} 3 &{} 3 &{} 0 \end{array}\right) . \end{aligned}$$

We obtain the cuspidal form \(P_A(t) = -576(5t_1- 4t_2)(5t_1 +4 t_2)\), which has signature \(s_A = (1,1; 0)\).

Example 8.9

Consider the point configuration and Gale dual

$$\begin{aligned} A = \left( \begin{array}{c@{\quad }c@{\quad }c@{\quad }c@{\quad }c} 1 &{} 1 &{} 1 &{} 1 &{} 1\\ 0 &{} 1 &{} 0 &{} 1 &{} 2\\ 0 &{} 0 &{} 1 &{} 2 &{} 1 \end{array}\right) \quad \text {and} \quad B^\mathrm{T} = \left( \begin{array}{r@{\quad }r@{\quad }r@{\quad }r@{\quad }r} 2 &{} -\,2 &{} -\,1 &{} 0 &{} 1 \\ 2 &{} -\,1 &{} -\,2 &{} 1 &{} 0 \end{array}\right) . \end{aligned}$$

We obtain the cuspidal form \(P_A(t) = -4(t_1^2 + t_1t_2 + t_2^2)\), which has signature \(s_A = (0,2; 0)\).