I’ve just uploaded to the arXiv my paper “Local Bernstein theory, and lower bounds for Lebesgue constants“. This paper was initially motivated by a problem of Erdős} on Lagrange interpolation, but in the course of solving that problem, I ended up modifying some very classical arguments of Bernstein and his contemporaries (Boas, Duffin, Schaeffer, Riesz, etc.) to obtain “local” versions of these classical “Bernstein-type inequalities” that may be of independent interest.
Bernstein proved many estimates concerning the derivatives of polynomials, trigonometric polynomials, and entire functions of exponential type, but perhaps his most famous inequality in this direction is:
Lemma 1 (Bernstein’s inequality for trigonometric polynomials) Letbe a trigonometric polynomial of degree at most
, with
for all
. Then
for all
.
Similar inequalities concerning norms of derivatives of Littlewood-Paley components of functions are now ubiquitious in the modern theory of nonlinear dispersive PDE (where they are also called Bernstein estimates), but this will not be the focus of this current post.
A trigonometric polynomial of degree
is of exponential type
in the sense that
for complex
. Bernstein in fact proved a more general result:
Lemma 2 (Bernstein’s inequality for functions of exponential type) Letbe an entire function of exponential type at most
, with
for all
. Then
for all
.
There are several proofs of this lemma – see for instance this survey of Queffélec and Zarouf. In the case that is real-valued on
, there is a nice proof by Duffin and Schaeffer, which we sketch as follows. Suppose we normalize
, and adjust
by a suitable damping factor so that
actually decays slower than
as
. Then, for any
and
, one can use Rouche’s theorem to show that the function
has the same number of zeroes as
in a suitable large rectangle; but on the other hand one can use the intermediate value theorem to show that
has at least as many zeroes than
in the same rectangle. Among other things, this prevents double zeroes from occuring, which turns out to give the desired claim
after some routine calculations (in fact one obtains the stronger bound
for all real
).
The first main result of the paper is to obtain localized versions of Lemma 2 (as well as some related estimates). Roughly speaking, these estimates assert that if is holomorphic on a wide thin rectangle passing through the real axis, is bounded by
on the intersection of the real axis with this rectangle, and is “locally of exponential type” in the sense that it is bounded by
on the upper and lower edges of this rectangle (and obeys some very mild growth conditions on the remaining sides of this rectangle), then
can be bounded by
plus small errors on the real line, with some additional estimates away from the real line also available. The proof proceeds by a modification of the Duffin–Schaeffer argument, together with the two-constant theorem of Nevanlinna (and some standard estimates of harmonic measures on rectangles) to deal with the effect of the localization. (As a side note, this latter argument was provided to me by ChatGPT, as I was not previously aware of the Nevanlinna two-constant theorem.)
Once one localizes this “Bernstein theory”, it becomes suitable for the analysis of (real-rooted, monic) polynomials of a high degree
, which are not bounded globally on
(and grow polynomially rather than exponentially at infinity), but which can exhibit “local exponential type” behavior on various intervals, particularly in regions where the logarithmic potential
behaves like a smooth function (here is the empirical measure of the roots
of
). A key example is the (monic) Chebyshev polynomials
, which locally behave like sinusoids on the interval
(and are locally of exponential type above and below this interval):
This becomes relevant in the theory of Lagrange interpolation. Recall that if are real numbers and
is a polynomial of degree less than
then one has the interpolation formula
where the Lagrange basis functions are defined by the formula
In terms of the monic polynomial , we can write
The stability and convergence properties of Lagrange interpolation are closely related to the Lebesgue function
and for a given interval , the quantity
is known as the Lebesgue constant for that interval.
If one chooses the interpolation points poorly, then the Lebesgue constant can be extremely large. However, if one selects these points to be the roots of the aforementioned monic Chebyshev polynomials, then it is known that
for all fixed intervals
in
. In the case
, it was shown by Erdős} that this is the best possible value of the Lebesgue constant up to
errors for interpolation on
, thus
whenever (a more precise bound was later shown by Vertesi). Erdős and Turán then asked if the same lower bound
held for more general intervals . This is shown in our paper; a variant integral bound
is also established, answering a separate question of Erdős. These lower bounds had previously obtained up to constants by Erdős and Szabados}; the main issue was to obtain the sharp constant in the main term.
In terms of the monic polynomial , these two estimates can be written as
and
Using the intuition that should behave locally like a trigonometric polynomial, and performing some renormalizations, one can extract the following toy problem to work with first:
Problem 3 Letbe a trigonometric polynomial of degree
with
roots
in
.
It is easy to check that the lower bounds of and
are sharp by considering the case when
is a sinusoid
.
The bound (3) is immediate from Bernstein’s inequality (Lemma 1). By applying a local version of this inequality, I was able to get a weak version of the claim (1) in which was replaced with
; see this early version of the paper, which was developed through conversations with Nat Sothanaphan and Aron Bhalla. By combining this argument with ideas from the older work of Erdős}, I was able to establish (1).
The bound (2) took me longer to establish, and involved a non-trivial amount of playing around with AI tools, the story of which I would like to share here. I had discovered the toy problem (4), but initially was not able to establish this inequality; AlphaEvolve seemed to confirm it numerically (with sinusoids appearing to be the extremizer), but did not offer direct clues on how to prove this rigorously. At some point I realized that the left-hand side factorized into the expressions and
, and tried to bound these expressions separately. Perturbing around a sinusoid
, I was able to show that the
norm
was a local minimum as long as one only perturbed by lower order Fourier modes, keeping the frequency
coefficients unchanged. Guessing that this local minimum was actually a global minimum, this led me to conjecture the general lower bound
whenever was a degree
trigonometric polynomial with highest frequency components
. AlphaEvolve numerically confirmed this inequality to be likely to be true also. I still did not see how to prove this inequality, but I decided to try my luck giving it to ChatGPT Pro, which recognized it as an
approximation problem and gave me a duality-based proof (based ultimately on the Fourier expansion of the square wave). With some further discussion, I was able to adapt this proof to functions of global exponential type (replacing the Fourier manipulations with contour shifting arguments, in the spirit of the Paley-Wiener theorem), which roughly speaking gave me half of what I needed to establish (2). However, I still needed the matching lower bound
on the other factor in the toy problem. Again, AlphaEvolve could numerically confirm that this inequality was likely to be true, but now the quantity I was trying to control did not look convex or linear in , and so the previous duality method did not seem to apply. At this point I switched to pen and paper; eventually I realized that the expression almost looked like a sum of residues, and eventually after playing around with contour integrals of
using the residue theorem I was able to establish (4), and then with a bit more (human) effort I could move from the toy problem back to the original problem to obtain (2). Quite possibly AI tools would also have been able to assist with these steps, but they were not necessary here; their main value for me was in quickly confirming that the approach I had in mind was numerically plausible, and in recognizing the right technique to solve one part of the toy problem I had isolated. (I also used AI tools for several other secondary tasks, such as literature review, proofreading, and generating pictures, but these applications of AI have matured to the point where using them for this purpose is almost mundane.)