This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

On the resolution power of Fourier extensions
for oscillatory functions

Ben Adcock111Department of Mathematics, Simon Fraser University, Canada (ben_adcock@sfu.ca)  and Daan Huybrechs222Department of Computer Science, K.U.Leuven, Belgium (daan.huybrechs@cs.kuleuven.be)
Abstract

Functions that are smooth but non-periodic on a certain interval possess Fourier series that lack uniform convergence and suffer from the Gibbs phenomenon. However, they can be represented accurately by a Fourier series that is periodic on a larger interval. This is commonly called a Fourier extension. When constructed in a particular manner, Fourier extensions share many of the same features of a standard Fourier series. In particular, one can compute Fourier extensions which converge spectrally fast whenever the function is smooth, and exponentially fast if the function is analytic, much the same as the Fourier series of a smooth/analytic and periodic function.

With this in mind, the purpose of this paper is to describe, analyze and explain the observation that Fourier extensions, much like classical Fourier series, also have excellent resolution properties for representing oscillatory functions. The resolution power, or required number of degrees of freedom per wavelength, depends on a user-controlled parameter and, as we show, it varies between 22 and π\pi. The former value is optimal and is achieved by classical Fourier series for periodic functions, for example. The latter value is the resolution power of algebraic polynomial approximations. Thus, Fourier extensions with an appropriate choice of parameter are eminently suitable for problems with moderate to high degrees of oscillation.

1 Introduction

In many physical problems, one encounters the phenomenon of oscillation. When approximating the solution to such a problem with a given numerical method, this naturally leads to the question of resolution power. That is, how many degrees of freedom are required in a given scheme to resolve such oscillations? Whilst it may be impossible to answer this question in general, important heuristic information about a given approximation scheme can be gained by restricting ones interest to certain simple classes of oscillatory functions (e.g. complex exponentials for problems in bounded intervals).

Resolution power represents an a priori measure of the efficiency of a numerical scheme for a particular class of problems. Schemes with low resolution power require more degrees of freedom, and hence increased computational cost, before the onset of convergence. Conversely, schemes with high resolution power capture oscillations with fewer degrees of freedom, resulting in decreased computational expense.

Consider the case of the unit interval [1,1][-1,1] (the primary subject of this paper). Here one typically studies the question of resolution via the complex exponentials

f(x)=exp(iωπx),ω.f(x)=\exp(i\omega\pi x),\quad\omega\in\R. (1)

To this end, let ϕn(f)\phi_{n}(f), n=1,2,n=1,2,\ldots be a sequence of approximations of the function f(x)=exp(iπωx)f(x)=\exp(i\pi\omega x) which converges to ff as nn\rightarrow\infty (here nn is the number of degrees of freedom in the approximation ϕn(f)\phi_{n}(f)). For 0<ϵ<10<\epsilon<1, let n(ϵ,ω)n(\epsilon,\omega) be the minimal nn such that

fϕn(f)L[1,1]2<ϵ.\|f-\phi_{n}(f)\|_{L^{2}_{[-1,1]}}<\epsilon.

We now define the resolution constant rr of the approximation scheme {ϕn}\{\phi_{n}\} as

r=lim supϵ1limωn(ϵ,ω)ω.r=\limsup_{\epsilon\rightarrow 1^{-}}\lim_{\omega\rightarrow\infty}\frac{n(\epsilon,\omega)}{\omega}.

Note that rr need not be well defined for an arbitrary scheme {ϕn}\{\phi_{n}\} (for example, if n(ϵ,ω)n(\epsilon,\omega) were to scale superlinearly in ω\omega). However, for all schemes encountered in this paper, this will be the case.

Loosely speaking, the resolution constant rr corresponds to the required number of degrees of freedom per wavelength to capture oscillatory behaviour; a common concept in the literature on oscillatory problems [6, 20]. In particular, we say that a given scheme has high (respectively low) resolution power if it has small (large) resolution constant. It is also worth noting that, in many schemes of interest, the approximation ϕn(f)\phi_{n}(f) is based on a collocation at a particular set of nn nodes in [1,1][-1,1]. In this circumstance, the resolution constant rr is equivalent to the number of points per wavelength required to resolve an oscillatory wave (for further details, see §1.3).

With little doubt, the approximation of a smooth, periodic function via its truncated Fourier series is one of the most effective numerical methods known. Fourier series, when computed via the FFT, lead to highly efficient, stable methods for the numerical solution of a large range of problems (in particular, PDE’s with periodic boundary conditions). A simple argument leads to a resolution constant of r=2r=2 in this case (for periodic oscillations), with exponential convergence occurring once the number of Fourier coefficients exceeds 2ω2\omega.

However, the situation is altered completely once periodicity is lost. The slow pointwise convergence of the Fourier series of a nonperiodic function, as well as the presence of 𝒪(1){\mathcal{O}}(1) Gibbs oscillations near the domain boundaries, means that nonperiodic oscillations cannot be resolved by such an approximation. A standard alternative is to approximate with a sequence of orthogonal polynomials (e.g. Chebyshev polynomials). Such approximations possess exponential convergence, without periodicity, yet the resulting resolution constant increases to the value r=πr=\pi, making such an approach clearly less than ideal.

1.1 Fourier extensions

The purpose of this paper is to discuss an alternative to polynomial approximation for nonperiodic functions; the so-called Fourier extension. Our main result is to show that Fourier extensions have excellent resolution properties. In particular, there is a user-determined parameter T(1,)T\in(1,\infty) that allows for continuous variation of the resolution constant from the value 22 (in the limit T1T\rightarrow 1), the figure corresponding to Fourier series, to π\pi (TT\rightarrow\infty), the value obtained by polynomial approximations. Thus, Fourier extensions are highly suitable tools for problems with oscillations at moderate to high frequencies.

Let us now describe Fourier extensions in more detail. As discussed, Fourier series are eminently suitable for approximating smooth and periodic oscillatory functions. It follows that a potential means to recover a highly accurate approximation of a nonperiodic function f:[1,1]f:[-1,1]\rightarrow\R is to seek to extend ff to a periodic function gg defined on a larger domain [T,T][-T,T] and compute the Fourier series of gg. Unfortunately, unless ff is itself periodic, no periodic extension gg will be analytic, and hence only spectral convergence can be expected. Preferably, for analytic ff, we seek an approximation that converges exponentially fast.

To remedy this situation, rather than computing an extension gg explicitly, it was proposed in [7, 11] to directly compute a Fourier representation of ff on the domain [T,T][-T,T] via the so-called Fourier extension problem:

Problem 1.1.

Let GnG_{n} be the space of 2T2T-periodic functions of the form

gGn:g(x)=α02+k=1nαkcosπTkx+βksinπTkx.g\in G_{n}:g(x)=\frac{\alpha_{0}}{2}+\sum_{k=1}^{n}\alpha_{k}\cos\frac{\pi}{T}kx+\beta_{k}\sin\frac{\pi}{T}kx. (2)

The (continuous) Fourier extension of ff to the interval [T,T][-T,T] is the solution to the optimization problem

gn:=argmingGnfgL[1,1]2.g_{n}:=\underset{g\in G_{n}}{\operatorname{argmin}}\|f-g\|_{L^{2}_{[-1,1]}}. (3)

Note that there are infinitely many ways in which to define a smooth and periodic extension of a function f:[1,1]f:[-1,1]\rightarrow\mathbb{C}, of which (3) is but one. We say an extension gg is a Fourier extension if gGng\in G_{n} is a finite Fourier series on [T,T][-T,T]. Note, however, that even within this stipulation, there are still infinitely many ways to define gg. We refer to the extension gng_{n} defined by (3) as the continuous Fourier extension of ff (in the sense that it minimizes the continuous L2L^{2}-norm).

As numerically observed in [7, 11], the convergence of gng_{n} to ff is exponential, provided ff is analytic. When T=2T=2, this was confirmed by the analysis presented in [25]. A pivotal role is played by the map

y=cosπ2x.y=\cos\frac{\pi}{2}x. (4)

The importance of this map is that it transforms the trigonometric basis functions that span the space GnG_{n} into polynomials in yy. In this setting, Problem 1.1 reduces simply to the computation of expansions in a suitable basis of nonclassical orthogonal polynomials, since the least squares criterion corresponds exactly to an orthogonal projection in a particular weighted norm. Well-known results on orthogonal polynomials can then be used to establish convergence properties of the Fourier extension. We recall this analysis in greater detail in §2, along with its generalization to arbitrary TT.

The map (4) demonstrates the close relationship that exists between Fourier series and polynomials. Note the similarity to the classical Chebyshev map x=cosθx=\cos\theta that takes Chebyshev polynomials to trigonometric basis functions. Compared to Chebyshev expansions, however, it is important to note that the roles of polynomials and trigonometric functions are interchanged. The Chebyshev expansion of a given function ff is a polynomial approximation, which is equivalent to the Fourier series of a related function. The Fourier extension of a function on the other hand is a trigonometric approximation, which is equivalent to the polynomial expansion of a related function. In that sense, Fourier extensions and Chebyshev expansions are dual to each other.

1.2 Key results

The main result of this paper is that the resolution constant r=r(T)r=r(T) of the continuous Fourier extension satisfies

r(T)2Tsin(π2T),T(1,).r(T)\leq 2T\sin\left(\tfrac{\pi}{2T}\right),\quad T\in(1,\infty).

Accordingly, the Fourier extension gng_{n} of the function (1) will begin to converge once nn exceeds 12r(T)ω\frac{1}{2}r(T)\omega (recall that gng_{n} involves 2n+12n+1 degrees of freedom). In particular, we find that r(T)2Tr(T)\sim 2T for T1T\approx 1 and r(T)πr(T)\sim\pi as TT\rightarrow\infty. Note that the resolution of the function f(x)=exp(iωπx)f(x)=\exp(i\omega\pi x) using classical Fourier series on [T,T][-T,T] would require a minimum of 2Tω2T\omega degrees of freedom (whenever ff is periodic). Thus, Fourier extensions exhibit comparable performance for TT close to 11. However, as TT increases, ff can be resolved more efficiently via its Fourier extension than if it had been directly expanded in a Fourier series on [T,T][-T,T]. In particular, the resolution power is bounded above for all TT by π\pi, which is precisely the resolution constant for polynomial approximations.

Aside from establishing when asymptotic convergence will occur, we also show that the convergence in this regime is exponential, and that for all sufficiently large nn the rate is precisely ρ=E(T)\rho=E(T), where

E(T)=cot2(π4T),E(T)=\cot^{2}\left(\tfrac{\pi}{4T}\right),

(this result actually holds for all sufficiently analytic functions ff, not just (1)). Here, we note that E(1)=1E(1)=1, implying no exponential convergence for T=1T=1. This is of course a consequence of the Gibbs phenomenon of Fourier series on [1,1][-1,1]. Convergence is exponential for all TT greater than 11, however, and the rate increases for larger TT.

In summary, the main conclusion we draw in this paper is the following. Smaller TT yields better resolution power of the continuous Fourier extension, at a cost of slower, but still exponential, convergence. Conversely, larger TT yields faster exponential convergence, at the expense of reduced resolution power. Formally, one may also obtain a resolution constant of 22 in the limit nn\rightarrow\infty by allowing T1T\rightarrow 1 as nn\rightarrow\infty, and towards the end of the paper we shall discuss different strategies for doing this, some of which are quite effective in practice.

Unfortunately, the linear system of equations to be solved when computing (3) is severely ill-conditioned. As a result, the best attainable accuracy with a continuous Fourier extension is of order ϵ\sqrt{\epsilon}, where ϵ\epsilon is the machine precision used. To overcome this, we present a new Fourier extension (referred to as the discrete Fourier extension), based on a judicious choice of interpolation nodes, which leads to far less severe condition numbers. Numerical examples demonstrate much higher attainable accuracies with this approach, typically of order ϵ\epsilon, whilst retaining both the same rates of convergence and resolution constant.

Inherent to Fourier extensions is the concept of redundancy: namely, there are infinitely many possible extensions of a given function. As we explain, this not only leads to the ill-conditioning mentioned above, it also means that the result of a numerical computation may differ somewhat from the ‘theoretical’ (i.e. infinite precision) continuous or discrete Fourier extension. Later in the paper, we discuss this observation and its consequences in some detail.

1.3 Relation to previous work

The question of resolution power of Fourier series and Chebyshev polynomial expansions was first developed rigorously by Gottlieb and Orszag [20, p.35] (see also [33]), who introduced and popularized the concept of points per wavelength. Therein the figures of 22 and π\pi respectively were derived, the latter being generalized to arbitrary Gegenbauer polynomial expansions in [21] (in §4 we provide an alternative proof, valid for almost any orthogonal polynomial system (OPS)). Resolution was also discussed in detail in [6, chpt. 2], where, aside from Fourier and Chebyshev expansions, the extremely poor performance of finite difference schemes was noted.

One-dimensional Fourier extensions were, arguably, first studied in detail in [7, 11], where they were employed to overcome the Gibbs phenomenon in standard Fourier expansions, as well as the Runge phenomenon in equispaced polynomial interpolation (see also [9]). Application to surface parametrizations was also explored in [11].

Similar ideas (typically referred to as Fourier embeddings), were previously proposed to solve PDE’s in complex geometries. These work by embedding the domain in a larger bounding box, and computing a Fourier series approximation on this domain. See [4, 5, 8, 12, 13, 14, 15, 19, 26, 29, 34, 36]. Such methods can be shown to work well, in principle even for arbitrary regions, but they are typically prohibitively computationally expensive.

More recently, a very effective method was developed in [10, 32] to solve time-dependent PDE’s in complex geometries. This approach is based on a technique for obtaining one-dimensional Fourier extensions, known as the FE–Gram method, which is then combined with an alternating direction technique, as well as standard FFTs, to solve the PDE efficiently and accurately. This approach has also been applied to Navier–Stokes equations [3]. Interestingly, it is shown and emphasized in [32] (see also [3]) that this method leads to an absence of dispersion errors (or pollution errors) – another beneficial property for wave simulations shared with classical Fourier series, and very much related to resolution power.

Having said this, we note at this point that the FE–Gram approach is quite different to the Fourier extensions we consider in this paper. The FE–Gram approximation is not exponentially convergent, and in practice, the order of convergence is usually limited by the user to ensure stability in the overall PDE solver. However, the FE–Gram approximation is designed specifically to be computed efficiently, in 𝒪(nlogn)\mathcal{O}(n\log n) time, using only function evaluations on an equispaced grid. Although there has been some recent progress in the rapid computation of exponentially-convergent Fourier extensions of the form (2) from equispaced data [30, 31], this is not an issue we shall dwell on in this paper. Thus, the main contribution of this paper is approximation-theoretic: we show that one can construct exponentially-convergent Fourier extensions with a resolution constant arbitrarily close to optimal. In §6 we discuss ongoing and future work pertaining to these issues.

This aside, our use of Fourier extensions in this paper to resolve oscillatory functions is superficially quite similar to the Kosloff–Tal–Ezer mapping (see [27], [6, chpt 16.9] and references therein, and more recently, [24]) for improving the severe time-step restriction inherent in Chebyshev spectral methods. Such an approach also improves the very much related property of resolution power. In this technique, one replaces the standard Chebyshev interpolation nodes with a sequence of mapped points, and expands in a nonpolynomial basis defined via the particular mapping. Roughly speaking, with Fourier extensions, the situation is reversed. Rather than fixing interpolation nodes, one specifies a particular basis (i.e. the Fourier basis for a particular TT) that gives rapidly convergent expansions and good resolution properties, and chooses an appropriate means for computing the extension (for example, a particular configuration of collocation nodes) via the corresponding mapping.

Finally, we remark in passing that there are a number of commonly used alternatives to Fourier extensions. Such methods typically arise from the desire to reconstruct a function directly from its Fourier coefficients (or pointwise values on an equispaced grid), whether via re-expanding in a sequence Gegenbauer polynomials [22, 23], or by smoothing the function by implicitly matching its derivatives at the domain boundary [18]. Whilst the latter retains a resolution constant of 22, it only yields algebraic convergence of a finite order, and suffers from severe ill-conditioning. Conversely, the Gegenbauer reconstruction procedure offers exponential convergence, but with a significant deterioration in resolution power [21].

The outline of the remainder of this paper is as follows. In §2 we detail the convergence of Fourier extensions for arbitrary T>1T>1, and in §3 we address numerical computation. In §4 we consider the resolution power of polynomial expansions, and in §5 we derive the resolution constant for Fourier extensions.

2 Fourier extensions on arbitrary intervals

In this section, we recall and generalize the analysis given in [25] for the case T=2T=2. We first show a new result, namely that the continuous Fourier extension converges spectrally for all smooth functions ff and for all T>1T>1 fixed. Next, a more involved analysis in §2.2 demonstrates that the continuous Fourier extension does in fact converge exponentially whenever ff is analytic. In doing so, we repeat some of the reasoning in [25] for the clarity of presentation, as well as for establishing notation that is needed later in the paper.

Before presenting these results, a word about terminology in order to avoid confusion. Throughout this paper, we will refer to the following three types of convergence of an approximation fnf_{n} to a given function ff. We say that fnf_{n} converges algebraically fast to ff at rate kk if ffn=𝒪(nk)\|f-f_{n}\|={\mathcal{O}}(n^{-k}) as nn\rightarrow\infty. Conversely, fnf_{n} converges spectrally fast to ff if the error ffn\|f-f_{n}\| decays faster than any algebraic power of n1n^{-1}, and exponentially fast if there exists some constant ρ>1\rho>1 such that ffn=𝒪(ρn)\|f-f_{n}\|={\mathcal{O}}(\rho^{-n}) for all large nn.

2.1 Spectral convergence

Standard approximations based on orthogonal polynomials (or Fourier series in the periodic case) converge exponentially fast provided ff is analytic, and spectrally fast if ff is only smooth [16, chpts 2,5]. Whenever ff has only finite regularity, convergence is algebraic at a rate determined by the degree of smoothness: specifically, if ff is (k1)(k-1)-times continuously differentiable in [1,1][-1,1] and f(k)f^{(k)} exists almost everywhere and is square-integrable (equivalently, fHk[1,1]f\in H^{k}[-1,1] – the kkth standard Sobolev space of functions defined on [1,1][-1,1]), then fnf_{n} converges algebraically fast at rate kk . Our first result regarding Fourier extensions illustrates identical convergence in this setting:

Theorem 2.1.

Suppose that fHk[1,1]f\in H^{k}[-1,1] for some kk\in\N and that T0>1T_{0}>1. Then, for all nn\in\N and TT0T\geq T_{0},

fgnL[1,1]2ck(T0)(nπT)kfH[1,1]k,\|f-g_{n}\|_{L^{2}_{[-1,1]}}\leq c_{k}(T_{0})\left(\frac{n\pi}{T}\right)^{-k}\|f\|_{H^{k}_{[-1,1]}}, (5)

where ck(T0)>0c_{k}(T_{0})>0 is independent of nn, ff and TT, H[1,1]k\|\cdot\|_{H^{k}_{[-1,1]}} is the standard norm on Hk[1,1]H^{k}[-1,1] and gng_{n} is the continuous Fourier extension of ff on [T,T][-T,T] defined by (3).

Proof.

Recall that there exists an extension operator :Hk[1,1]Hk()\mathcal{E}:H^{k}[-1,1]\rightarrow H^{k}(\R) with f|[1,1]=f\mathcal{E}f|_{[-1,1]}=f and fHkcfH[1,1]k\|\mathcal{E}f\|_{H^{k}}\leq c\|f\|_{H^{k}_{[-1,1]}} for some positive constant cc independent of ff [1].

Let χC()\chi\in C^{\infty}(\R) be monotonically decreasing and satisfy χ(x)=0\chi(x)=0 for x>T01x>T_{0}-1 and χ(x)=1\chi(x)=1 for x<0x<0. We define the bump function C()\mathcal{B}\in C^{\infty}(\R) by

(x)={χ(x1)T0x<111x1χ(x1)1<xT00|x|>T0.\mathcal{B}(x)=\left\{\begin{array}[]{cc}\chi(-x-1)&-T_{0}\leq x<-1\\ 1&-1\leq x\leq 1\\ \chi(x-1)&1<x\leq T_{0}\\ 0&|x|>T_{0}.\end{array}\right.

Note that the function (x)f(x)Hk()\mathcal{B}(x)\mathcal{E}f(x)\in H^{k}(\R) and has support in [T0,T0][-T_{0},T_{0}]. Thus its restriction to [T,T][-T,T] is periodic for any TT0T\geq T_{0}. Since gng_{n} minimizes the L2L^{2} norm error over all functions from the set GnG_{n}, we have

fgnL[1,1]2f(f)nL[1,1]2f(f)nL[T,T]2,\|f-g_{n}\|_{L^{2}_{[-1,1]}}\leq\|f-(\mathcal{B}\mathcal{E}f)_{n}\|_{L^{2}_{[-1,1]}}\leq\|\mathcal{B}\mathcal{E}f-(\mathcal{B}\mathcal{E}f)_{n}\|_{L^{2}_{[-T,T]}},

where (f)n(\mathcal{B}\mathcal{E}f)_{n} denotes the nnth partial Fourier sum of (x)f(x)\mathcal{B}(x)\mathcal{E}f(x) on [T,T][-T,T]. Since (x)f(x)Hk[T,T]\mathcal{B}(x)\mathcal{E}f(x)\in H^{k}[-T,T] and is periodic on [T,T][-T,T], a well-known estimate [16, eqn (5.1.10)] gives

f(f)nL[T,T]2(nπT)k(f)(k)L[T,T]2.\|\mathcal{B}\mathcal{E}f-(\mathcal{B}\mathcal{E}f)_{n}\|_{L^{2}_{[-T,T]}}\leq\left(\frac{n\pi}{T}\right)^{-k}\|(\mathcal{B}\mathcal{E}f)^{(k)}\|_{L^{2}_{[-T,T]}}.

Moreover, (f)(k)L[T,T]2=(f)(k)L[T0,T0]2ck(T0)fH[1,1]k\|(\mathcal{B}\mathcal{E}f)^{(k)}\|_{L^{2}_{[-T,T]}}=\|(\mathcal{B}\mathcal{E}f)^{(k)}\|_{L^{2}_{[-T_{0},T_{0}]}}\leq c_{k}(T_{0})\|f\|_{H^{k}_{[-1,1]}} for some constant ck(T0)c_{k}(T_{0}) independent of ff. Therefore, the result follows. ∎

2.2 Exponential convergence

2.2.1 The continuous Fourier extension

As mentioned in the introduction, the continuous Fourier extension defined by (3) can be characterized in terms of certain nonclassical orthogonal polynomial expansions. This was demonstrated for the case T=2T=2 in [25], but can be shown for any T>1T>1 with relatively minor modifications.

Our approach is to construct an orthogonal basis for the space GnG_{n} of 2T2T-periodic functions. Since the least squares criterion (3) corresponds to an orthogonal projection, it then suffices to expand a given function ff in this basis in order to find its Fourier extension.

The cosine and sine functions in GnG_{n} are already mutually orthogonal and hence can be treated separately. Consider the cosines first, i.e. the set

Cn:={cosπTkx}k=0n.C_{n}:=\{\cos\frac{\pi}{T}kx\}_{k=0}^{n}.

Trigonometric functions are closely related to polynomials, through an appropriate cosine-mapping. This can be seen, for example, from the defining property of Chebyshev polynomials of the first kind TkT_{k},

coskx=Tk(cosx).\cos kx=T_{k}(\cos x).

In the same spirit, we define the map

y=cosπTx,y=\cos\frac{\pi}{T}x, (6)

and note that, since coskx\cos kx is an algebraic polynomial of degree kk in cosx\cos x, the function cosπTkx\cos\frac{\pi}{T}kx is a polynomial in yy. From this we conclude that the set of cosines CnC_{n} is a basis for the space of algebraic polynomials in yy of degree nn.

Since the functions in CnC_{n} are linearly independent, an orthogonal basis exists. Moreover, we may write the basis functions as polynomials in yy, say TkT(y)=TkT(cosπTx)T_{k}^{T}(y)=T_{k}^{T}(\cos\frac{\pi}{T}x). Orthogonality in L2[1,1]L^{2}[-1,1] implies

δkl\displaystyle\delta_{kl} =11TkT(cosπTx)TlT(cosπTx)dx\displaystyle=\int_{-1}^{1}T_{k}^{T}(\cos\frac{\pi}{T}x)T_{l}^{T}(\cos\frac{\pi}{T}x)\,{\rm d}x
=201TkT(cosπTx)TlT(cosπTx)dx\displaystyle=2\int_{0}^{1}T_{k}^{T}(\cos\frac{\pi}{T}x)T_{l}^{T}(\cos\frac{\pi}{T}x)\,{\rm d}x
=2Tπc(T)1TkT(y)TlT(y)11y2dy,\displaystyle=\frac{2T}{\pi}\int_{c(T)}^{1}T_{k}^{T}(y)T_{l}^{T}(y)\frac{1}{\sqrt{1-y^{2}}}\,{\rm d}y,

where, for ease of notation, we have defined the TT-dependent constant

c(T):=cosπT.c(T):=\cos\frac{\pi}{T}.

In the latter step above, we applied the substitution (6), which maps the interval [0,1][0,1] to [c(T),1][c(T),1] and introduces the Jacobian with the inverse square root. It follows that the TkT(y)T_{k}^{T}(y) are orthonormal polynomials on [c(T),1][c(T),1] with respect to the weight function

w1(y)=2Tπ11y2.w_{1}(y)=\frac{2T}{\pi}\frac{1}{\sqrt{1-y^{2}}}. (7)

This weight differs only by a constant factor from the typical weight function of Chebyshev polynomials of the first kind Tk(y)T_{k}(y). However, the interval of orthogonality is different from that of the Chebyshev polynomials, since [c(T),1][c(T),1] is contained within (1,1](-1,1] for T>1T>1, whereas Chebyshev polynomials are orthogonal over the whole interval [1,1][-1,1].

We now consider the set of sines in GnG_{n},

Sn:={sinπTkx}k=1n.S_{n}:=\{\sin\frac{\pi}{T}kx\}_{k=1}^{n}.

Analogously, this leads to orthogonal polynomials resembling Chebyshev polynomials of the second kind Uk(y)U_{k}(y). Indeed, from the property

sin(k+1)x=Uk(cosx)sinx\sin(k+1)x=U_{k}(\cos x)\sin x

we find that sinπTkx\sin\frac{\pi}{T}kx is also a polynomial in yy, but only up to an additional factor. This factor is

sinπTx=1y2.\sin\frac{\pi}{T}x=\sqrt{1-y^{2}}.

We therefore consider an orthogonal basis in the form

UkT(y)1y2=UkT(cosπTx)sinπTx.U_{k}^{T}(y)\sqrt{1-y^{2}}=U_{k}^{T}(\cos\frac{\pi}{T}x)\sin\frac{\pi}{T}x.

Orthogonality in L2[1,1]L^{2}[-1,1] implies

δkl\displaystyle\delta_{kl} =11UkT(cosπTx)sinπTxUlT(cosπTx)sinπTxdx\displaystyle=\int_{-1}^{1}U_{k}^{T}(\cos\frac{\pi}{T}x)\sin\frac{\pi}{T}xU_{l}^{T}(\cos\frac{\pi}{T}x)\sin\frac{\pi}{T}x\,{\rm d}x
=201UkT(cosπTx)sinπTxUlT(cosπTx)sinπTxdx\displaystyle=2\int_{0}^{1}U_{k}^{T}(\cos\frac{\pi}{T}x)\sin\frac{\pi}{T}xU_{l}^{T}(\cos\frac{\pi}{T}x)\sin\frac{\pi}{T}x\,{\rm d}x
=2Tπc(T)1UkT(y)UlT(y)1y2dy.\displaystyle=\frac{2T}{\pi}\int_{c(T)}^{1}U_{k}^{T}(y)U_{l}^{T}(y)\sqrt{1-y^{2}}\,{\rm d}y.

Thus, the UkT(y)U_{k}^{T}(y) are again orthonormal polynomials. The weight function

w2(y)=2Tπ1y2w_{2}(y)=\frac{2T}{\pi}\sqrt{1-y^{2}} (8)

corresponds to that of the Chebyshev polynomials of the second kind, but here too the interval of orthogonality [c(T),1][c(T),1] differs from the classical case in the same TT-dependent manner.

Since the functions TkT(cosπTx)T^{T}_{k}(\cos\frac{\pi}{T}x), UkT(cosπTx)sinπTxU^{T}_{k}(\cos\frac{\pi}{T}x)\sin\frac{\pi}{T}x comprise a basis for the space GnG_{n}, and since the continuous Fourier extension gng_{n} is the orthogonal projection of ff onto GnG_{n}, we now deduce the following theorem:

Theorem 2.2.

The continuous Fourier extension gng_{n} of a function ff is precisely

gn(x)=k=0nakTkT(cosπTx)+k=0n1bkUkT(cosπTx)sinπTx,g_{n}(x)=\sum_{k=0}^{n}a_{k}T_{k}^{T}\left(\cos\frac{\pi}{T}x\right)+\sum_{k=0}^{n-1}b_{k}U_{k}^{T}\left(\cos\frac{\pi}{T}x\right)\sin\frac{\pi}{T}x,

where

ak\displaystyle a_{k} =11f(x)TkT(cosπTx)dx,\displaystyle=\int_{-1}^{1}f(x)T_{k}^{T}(\cos\frac{\pi}{T}x)\,{\rm d}x, (9)
bk\displaystyle b_{k} =11f(x)UkT(cosπTx)sinπTxdx,\displaystyle=\int_{-1}^{1}f(x)U_{k}^{T}(\cos\frac{\pi}{T}x)\sin\frac{\pi}{T}x\,{\rm d}x, (10)

and the polynomials TkT(y)T_{k}^{T}(y) and UkT(y)U_{k}^{T}(y) are orthonormal on [c(T),1][c(T),1] with respect to the weights w1(y)w_{1}(y) and w2(y)w_{2}(y) given by (7) and (8) respectively.

2.2.2 Exponential convergence of the continuous Fourier extension

The functions being expanded in the theory of the continuous Fourier extension above are related to ff through the inverse of the map (6). It is necessary to distinguish between the even and odd parts of ff,

fe(x)=f(x)+f(x)2,fo(x)=f(x)f(x)2.\displaystyle f_{e}(x)=\frac{f(x)+f(-x)}{2},\quad f_{o}(x)=\frac{f(x)-f(-x)}{2}.

Since

ak\displaystyle a_{k} =11f(x)TkT(cosπTx)dx\displaystyle=\int_{-1}^{1}f(x)T_{k}^{T}(\cos\frac{\pi}{T}x)\,{\rm d}x
=201fe(x)TkT(cosπTx)dx\displaystyle=2\int_{0}^{1}f_{e}(x)T_{k}^{T}(\cos\frac{\pi}{T}x)\,{\rm d}x
=2Tπc(T)1fe(Tπcos1y)TkT(y)11y2dy,\displaystyle=\frac{2T}{\pi}\int_{c(T)}^{1}f_{e}\left(\frac{T}{\pi}\cos^{-1}y\right)T_{k}^{T}(y)\frac{1}{\sqrt{1-y^{2}}}\,{\rm d}y, (11)

we see that the even part of the Fourier extension gng_{n} is precisely the orthogonal polynomial expansion of the function

f1(y)=fe(Tπcos1y)=fe(x).f_{1}(y)=f_{e}\left(\frac{T}{\pi}\cos^{-1}y\right)=f_{e}(x).

A similar reasoning for the odd part of ff, based on the coefficients bkb_{k}, yields the second function

f2(y)=fo(Tπcos1y)1y2=fo(x)sinπTx,f_{2}(y)=\frac{f_{o}\left(\frac{T}{\pi}\cos^{-1}y\right)}{\sqrt{1-y^{2}}}=\frac{f_{o}(x)}{\sin\frac{\pi}{T}x},

with the odd part of gng_{n}, when divided by sinπTx\sin\frac{\pi}{T}x, corresponding to the expansion of f2f_{2} in the polynomials UkT(y)U^{T}_{k}(y).

Let us now recall some standard theory on orthogonal polynomial (see, for example, [6, 35, 37] for an in-depth treatment). The convergence rate of the expansion of an analytic function ff in a set of orthogonal polynomials is exponential, and for a wide class of orthogonal polynomials on the interval [1,1][-1,1] the precise rate of convergence is determined by the largest Bernstein ellipse

e(ρ):={12(ρ1eiθ+ρeiθ):θ[π,π]},ρ1,e(\rho):=\{\frac{1}{2}(\rho^{-1}e^{-\textrm{i}\theta}+\rho e^{\textrm{i}\theta}):\theta\in[-\pi,\pi]\},\quad\rho\geq 1,

within which ff is analytic. We shall use the convention ρ1\rho\geq 1 throughout. Note that the ellipse e(ρ)e(\rho) has foci ±1\pm 1, and the parameter ρ\rho coincides with the sum of the lengths of its major and minor semiaxes. Suppose now that ff is analytic within the Bernstein ellipse of radius ρ=ρmax\rho=\rho_{\max} and not analytic in any ellipse with ρ>ρmax\rho>\rho_{\max}. Then the rate convergence of the expansion of ff in a set of orthogonal polynomials on [1,1][-1,1] is precisely (ρmax)n(\rho_{\max})^{-n}.

Returning to the continuous Fourier extension, we now see that its convergence rate is determined by the nearest singularity of f1(y)f_{1}(y) or f2(y)f_{2}(y) to the interval [c(T),1][c(T),1]. Note that, even when ff is entire, singularities are introduced by the inverse cosine mapping at the points y=±1y=\pm 1. One can verify that the singularity at y=1y=1 is removable for both f1f_{1} and f2f_{2}. Thus, the nearest singularity lies at y=1y=-1.

To apply the above polynomial theory, it remains to transform the interval [c(T),1][c(T),1] to the standard interval [1,1][-1,1]. This is achieved by the affine map

m1(s)=2sc(T)1c(T)1,s[c(T),1],m^{-1}(s)=2\frac{s-c(T)}{1-c(T)}-1,\quad s\in[c(T),1], (12)

with inverse m(t)=12(1c(T))t+12(1+c(T))m(t)=\frac{1}{2}(1-c(T))t+\frac{1}{2}(1+c(T)) mapping [1,1][-1,1] to [c(T),1][c(T),1]. Note that m1m^{-1} maps the point y=1y=-1 to the point

u=3+c(T)c(T)1=12cosec2(π2T)<1,u=\frac{3+c(T)}{c(T)-1}=1-2\mathrm{cosec}^{2}\left(\tfrac{\pi}{2T}\right)<-1,

which lies on the negative real axis. Since the Bernstein ellipse e(ρ)e(\rho) crosses the negative real axis in the point 12(ρ1+ρ)-\frac{1}{2}(\rho^{-1}+\rho), equating this with uu yields the maximal value of ρ\rho. We therefore deduce the following result:

Theorem 2.3.

For all sufficiently analytic functions ff, the error in approximating ff by its continuous Fourier extension gng_{n} satisfies

|f(x)gn(x)|cfE(T)n,n,T>1,|f(x)-g_{n}(x)|\leq c_{f}E(T)^{-n},\quad\forall n\in\mathbb{N},\ T>1, (13)

uniformly for x[1,1]x\in[-1,1], where cfc_{f} is a constant depending on ff only, and

E(T)=cot2(π4T).E(T)=\cot^{2}\left(\tfrac{\pi}{4T}\right).

This theorem extends Theorem 3.14 of [25] to the case of general T>1T>1. Note that the estimate (13) holds for all nn\in\mathbb{N} and T>1T>1. For the sake of brevity, we have limited the exposition here to the case of sufficiently analytic functions ff. Yet, we make the following comments:

  • The rate of exponential convergence may be slower than E(T)E(T) when f(x)f(x) is not sufficiently analytic as a function of xx, i.e. if ff has a singularity closer than that introduced by the inverse cosine mapping.

  • The convergence may also be faster than E(T)E(T) if ff is analytic and periodic on [T,T][-T,T]. In that case, one can verify that the singularity introduced by the inverse cosine at y=1y=-1 is also removable. Thus, the convergence rate is no longer limited by the singularity introduced by the map between the xx and yy domains.

  • For the case T=2T=2 we recover the convergence rate E(2)=3+22E(2)=3+2\sqrt{2} found in [25, §3.4].

The function E(T)E(T) is depicted in Fig. 1. It is monotonically increasing on (1,)(1,\infty) as a function of TT, and behaves like 1+π(T1)1+\pi(T-1) for T1T\approx 1 and 16π2T2\frac{16}{\pi^{2}}T^{2} for T1T\gg 1. Thus, larger TT leads to more rapid exponential convergence. This is confirmed in Fig. 2, where we plot the error in approximating f(x)=exf(x)=e^{x} by fnf_{n} for various nn and TT (for this example we use high precision to avoid any numerical effects – see §3). Note the close correspondence between the observed and predicted convergence rates, as well as the significant increase in convergence rate for larger TT. Having said this, it transpires that increasing TT adversely affects the resolution power, a topic we will consider further in §5.

Refer to caption
Figure 1: Theoretical convergence rate E(T)E(T) of the continuous Fourier extension as a function of the extension parameter TT.
Refer to caption
(a) Log error
Refer to caption
(b) Convergence rate
Figure 2: The quantities en=fgnL[1,1]e_{n}=\|f-g_{n}\|_{L^{\infty}_{[-1,1]}} (solid line) and E(T)nE(T)^{-n} (dashed line) for T=2,4,6,8T=2,4,6,8. The left panel shows the values ene_{n} and E(T)nE(T)^{-n}. The right panel gives the scaled values exp(1nlogen)\exp(-\frac{1}{n}\log e_{n}) and E(T)E(T).

3 Computation of Fourier extensions

Although in §2.2.1 we introduced a new, orthogonal basis for the set GnG_{n}, this basis is never actually used in computations. Instead, we always use either complex exponentials or real sines and cosines. The reason for this is primarily simplicity (recall that the relevant orthogonal basis relates to nonstandard orthogonal polynomials, which therefore would need to be precomputed), and the fact that having a Fourier extension in the form of a standard Fourier series is certainly most convenient in practice.

With this in mind, let {ϕk}k=12n+1\{\phi_{k}\}^{2n+1}_{k=1} be an enumeration of a ‘standard’ basis for the set GnG_{n} (e.g. complex exponentials), and suppose that the continuous Fourier extension gng_{n} of a function ff has coefficients {ak}k=12n+1\{a_{k}\}^{2n+1}_{k=1} in this basis. Since gng_{n} is the solution to (3), it is straightforward to show that

Aa=B,Aa=B, (14)

where A(2n+1)×(2n+1)A\in{}^{(2n+1)\times(2n+1)} and B2n+1B\in{}^{2n+1} have entries ϕk,ϕj\langle\phi_{k},\phi_{j}\rangle and f,ϕj\langle f,\phi_{j}\rangle respectively, and aa is the vector of coefficients aka_{k}. Thus, given BB, we may compute the continuous Fourier extension gng_{n} by solving (14).

There are two issues with computing the continuous Fourier extension. First, one requires knowledge (or prior computation) of the integrals f,ϕj\langle f,\phi_{j}\rangle. Second, the condition number κ(A)E(T)2n\kappa(A)\sim E(T)^{2n} [25]. Such severe ill-conditioning typically limits the maximal achievable accuracy (by this we mean the smallest possible error fgn\|f-g_{n}\| that can be attained in finite precision for any nn) of the continuous Fourier extension to 𝒪(ϵ){\mathcal{O}}(\sqrt{\epsilon}), where ϵ\epsilon is the machine precision used [25].

Ill-conditioning of the exact extension is due to the redundancy of the set GG_{\infty}. This is explained in the next section. Fortunately, its effect can be greatly mitigated by computing a so-called discrete Fourier extension instead. This extension, which converges at precisely the same rate as the continuous Fourier extension, involves only pointwise evaluations of ff, and consequently also avoids the first issue stated above. This is discussed in §3.2.

3.1 Ill-conditioning

The reason for ill-conditioning in the exact Fourier extension is simple: although the functions exp(ikπTx)\exp(i\frac{k\pi}{T}x) form an orthogonal basis on [T,T][-T,T], they only constitute a frame when restricted to the interval [1,1][-1,1].

Lemma 3.1.

The set

Φ:={12}{12exp(ikπTx)}k\{0},\Phi:=\{\frac{1}{\sqrt{2}}\}\cup\{\frac{1}{\sqrt{2}}\exp(\textrm{i}\tfrac{k\pi}{T}x)\}_{k\in\Z\backslash\{0\}}, (15)

is a normalized tight frame for L2[1,1]L^{2}[-1,1] with frame bound TT.

Proof.

Let fL2[1,1]f\in L^{2}[-1,1] be given and define f~L2[T,T]\tilde{f}\in L^{2}[-T,T] as the extension of ff by zero to [T,T][-T,T]. Since

ak=1211f(x)exp(ikπTx)dx=12TTf~(x)exp(ikπTx)dx,a_{k}=\frac{1}{\sqrt{2}}\int^{1}_{-1}f(x)\exp(\textrm{i}\tfrac{k\pi}{T}x)\,{\rm d}x=\frac{1}{\sqrt{2}}\int^{T}_{-T}\tilde{f}(x)\exp(\textrm{i}\tfrac{k\pi}{T}x)\,{\rm d}x, (16)

we find that 1Tak\frac{1}{\sqrt{T}}a_{k} is precisely the kkth Fourier coefficient of f~\tilde{f} on [T,T][-T,T]. By Parseval’s relation

k=|ak|2=Tf~2=Tf2,\sum^{\infty}_{k=-\infty}|a_{k}|^{2}=T\|\tilde{f}\|^{2}=T\|f\|^{2},

as required. ∎

For a general introduction to the theory of frames, see [17].

Lemma 3.1 has several implications. Recall that frames are usually redundant: any given function ff typically has infinitely many representations in a particular frame. This implies that there will be approximate linear dependencies amongst the columns of the Gram matrix AA for all large nn. This makes AA ill-conditioned, and in the case of the frame (15), leads to the aforementioned exponentially large condition numbers.

Let us consider several particular representations of a smooth function ff in the frame (15). Recall that associated to any frame {fk}\{f_{k}\} is the so-called canonical dual frame {gk}\{g_{k}\} [17]. A representation of ff, the frame decomposition, is given in terms of this frame by

f=k=1f,gkfk.f=\sum^{\infty}_{k=1}\langle f,g_{k}\rangle f_{k}.

Since the frame (15) is tight, it coincides with its canonical dual up to a constant factor that is equal to the frame bound, in this case TT. Therefore its frame decomposition is precisely

k=1Takexp(ikπTx)2T,\sum^{\infty}_{k=-\infty}\frac{1}{\sqrt{T}}a_{k}\frac{\exp(\textrm{i}\tfrac{k\pi}{T}x)}{\sqrt{2T}},

where the aka_{k}’s are given by (16). However, as illustrated in the proof of the previous lemma, this is precisely the Fourier series of the discontinuous function f~\tilde{f}, and thus this infinite series converges only slowly and suffers from a Gibbs phenomenon.

On the other hand, as shown in the proof of Theorem 2.1, by extending ff more smoothly to [T,T][-T,T], one obtain representations of ff in the frame (15) that converge algebraically at arbitrarily fast rates. Thus, we conclude the following: different frame expansions of the same function may give rise to wildly different approximations.

Clearly, given its exponential rate of convergence, the continuous Fourier extension gng_{n} will not coincide with any of the aforementioned representations (recall also that gng_{n} does not typically converge outside [1,1][-1,1] [25], and therefore its limit cannot be a frame expansion in general). More importantly, however, it is not certain that the solution of the linear system (14), when computed in finite precision, will actually coincide with the continuous Fourier extension gng_{n} itself. The reason is straightforward: for large nn, AA is approximately underdetermined, and therefore there will be many approximate solutions to (14) corresponding to different frame expansions of ff. A typical linear solver will usually select an approximate solution with bounded coefficient vector aa, and there is no guarantee that this need coincide with gng_{n}.

This phenomenon has a significant potential consequence: since the numerically computed extension need not coincide with the continuous Fourier extension gng_{n}, theoretical results for the behaviour of gng_{n} are not guaranteed to be inherited by the numerical solution. Having said this, in numerical experiments, one typically witnesses exponential convergence, exactly as predicted by Theorem 2.3. However, a difference can arise when investigating other properties, such as resolution power, as we discuss and explain in §5.

The above discussion is not intended to be rigorous. A full analysis of the behaviour of ‘numerical’ Fourier extensions is outside the scope of this paper. It transpires that this can be done, and a complete analysis will appear in a future paper [2]. Instead, in the remainder of this paper, we focus mainly on approximation-theoretic properties of theoretical extensions; resolution power, in particular. Nonetheless, in §5.3 we revisit the issue of numerical extensions insomuch as it relates to resolution, and give an explanation of the phenomenon of differing resolution power mentioned above.

3.2 A discrete Fourier extension

The continuous Fourier extension is defined by a least squares criterion, and as such it suffers from the drawback of requiring computation of the integrals f,ϕk\langle f,\phi_{k}\rangle. To avoid this issue one may instead define a Fourier extension by a collocation condition. In other words, if {xi}i=12n+1\{x_{i}\}^{2n+1}_{i=1} is a set of nodes in [1,1][-1,1] we replace the matrix AA and the vector BB in (14) by A~\tilde{A} and B~\tilde{B} with entries ϕk(xj)\phi_{k}(x_{j}) and f(xj)f(x_{j}) respectively. We then define A~a=B~\tilde{A}a=\tilde{B} once more.

The key question is how to determine good nodes. Recall that gng_{n}, as defined by (3), is essentially a sum of two polynomial approximations in the variable y[c(T),1]y\in[c(T),1]. The polynomial interpolant of an analytic function in yy at Chebyshev nodes (appropriately scaled to the interval [c(T),1][c(T),1]) converges exponentially fast at the same rate ρ\rho as its orthogonal polynomial expansion. Therefore, such nodes, when mapped to the original domain via x=Tπcos1yx=\frac{T}{\pi}\cos^{-1}y, will ensure exponential convergence at rate E(T)E(T) of the resulting Fourier extension.

It is now slightly easier to redefine GnG_{n} to be the space of dimension 2n+22n+2 spanned by the functions {cosπTkx}k=0n\{\cos\frac{\pi}{T}kx\}^{n}_{k=0} and {sinπTkx}k=1n+1\{\sin\frac{\pi}{T}kx\}^{n+1}_{k=1}. The 2n+22n+2 collocation nodes in [1,1][-1,1] therefore take the form {xi}i=0n{xi}i=0n\{x_{i}\}^{n}_{i=0}\cup\{-x_{i}\}^{n}_{i=0}, where

xi=Tπcos1[12(1c(T))cos((2i+1)π2n+2)+12(1+c(T))],i=0,,n.x_{i}=\frac{T}{\pi}\cos^{-1}\left[\frac{1}{2}(1-c(T))\cos\left(\frac{(2i+1)\pi}{2n+2}\right)+\frac{1}{2}(1+c(T))\right],\quad i=0,\ldots,n. (17)

Recall that c(T)=cosπTc(T)=\cos\frac{\pi}{T}. We refer to such nodes as symmetric mapped Chebyshev nodes, and the corresponding extension gng_{n} as the discrete Fourier extension of the function ff.

Besides removing the requirement to compute the integrals f,ϕn\langle f,\phi_{n}\rangle, this choice of nodes also has the significant effect of improving conditioning, as we now explain. We first note the following:

Lemma 3.2.

Let A~\tilde{A} be the collocation matrix based on the symmetric mapped Chebyshev nodes (17), and suppose that DD is the diagonal matrix of weights ωi=πn+1\omega_{i}=\frac{\pi}{n+1}. Then the matrix AW=A~DA~A_{W}=\tilde{A}^{\top}D\tilde{A} has entries

ϕk,ϕjW=11ϕj(x)ϕk(x)W(x)dx,j,k=1,,2n+2,\langle\phi_{k},\phi_{j}\rangle_{W}=\int^{1}_{-1}\phi_{j}(x)\phi_{k}(x)W(x)\,{\rm d}x,\quad j,k=1,\ldots,2n+2,

where WW is the positive, integrable function given by

W(x)=2πTcosπ2TxcosπTxcosπT.W(x)=\frac{2\pi}{T}\frac{\cos\frac{\pi}{2T}x}{\sqrt{\cos\frac{\pi}{T}x-\cos\frac{\pi}{T}}}.
Proof.

Note that AWA_{W} is a block matrix with blocks corresponding to inner products of sines and cosines with sines and cosines. Consider the upper left block of the matrix A~DA~\tilde{A}^{\top}D\tilde{A}. In the (j,k)(j,k)th entry, using the symmetry of the collocation points, we have that

2i=0nωiϕj(xi)ϕk(xi)=2i=0nωiTj(yi)Tk(yi),2\sum^{n}_{i=0}\omega_{i}\phi_{j}(x_{i})\phi_{k}(x_{i})=2\sum^{n}_{i=0}\omega_{i}T_{j}(y_{i})T_{k}(y_{i}),

where {yi}i=0n=cos(πTxi)\{y_{i}\}^{n}_{i=0}=\cos(\frac{\pi}{T}x_{i}) are Chebyshev nodes on [c(T),1][c(T),1]. Recall that m(t)m(t) is the affine map from [1,1][-1,1] to [c(T),1][c(T),1], as defined in §2.2.2. Since the product Tj(y)Tk(y)T_{j}(y)T_{k}(y) is a polynomial in yy of degree at most 2n2n, the Gaussian quadrature rule associated with the points xix_{i} is exact and it follows that

2i=0nωiϕj(xi)ϕk(xi)\displaystyle 2\sum^{n}_{i=0}\omega_{i}\phi_{j}(x_{i})\phi_{k}(x_{i}) =211Tj(m(t))Tk(m(t))11t2dt\displaystyle=2\int^{1}_{-1}T_{j}(m(t))T_{k}(m(t))\frac{1}{\sqrt{1-t^{2}}}\,{\rm d}t
=2c(T)1Tj(y)Tk(y)w(y)dy,\displaystyle=2\int^{1}_{c(T)}T_{j}(y)T_{k}(y)w(y)\,{\rm d}y,

where w(y)w(y) is given by

w(y)=21c(T)11(2y1c(T)1c(T))2.w(y)=\frac{2}{1-c(T)}\frac{1}{\sqrt{1-\left(\frac{2y-1-c(T)}{1-c(T)}\right)^{2}}}.

The first factor is the Jacobian of the mapping to [c(T),1][c(T),1], the second factor is the scaling of the standard Chebyshev weight to that interval. The change of variables y=cosπTxy=\cos\frac{\pi}{T}x now gives

2i=0nωiϕj(xi)ϕk(xi)=2πT01ϕj(x)ϕk(x)w(cosπTx)sinπTxdx,2\sum^{n}_{i=0}\omega_{i}\phi_{j}(x_{i})\phi_{k}(x_{i})=2\frac{\pi}{T}\int^{1}_{0}\phi_{j}(x)\phi_{k}(x)w(\cos\tfrac{\pi}{T}x)\sin\tfrac{\pi}{T}x\,{\rm d}x,

which is easily found to coincide with ϕk,ϕjW\langle\phi_{k},\phi_{j}\rangle_{W}. The case corresponding to the sine functions sinπTkx\sin\frac{\pi}{T}kx is identical. ∎

This lemma demonstrates that the normal equations for the discrete Fourier extension are the equations of a continuous Fourier extension based on the weighted LW2L^{2}_{W} optimization problem. It is well-known that forming normal equations leads to worse numerical results [28, Ch.19], and exactly the same is true in this case. Indeed, since WW is an integrable weight function, we may expect that κ(AW)E(T)2n\kappa(A_{W})\approx E(T)^{2n}, i.e. exactly as in the unweighted case. Thus collocation leads to the significant reduction in condition number, with κ(A~)E(T)n\kappa(\tilde{A})\approx E(T)^{n} as opposed to E(T)2nE(T)^{2n}. As a result, one typically observes a much higher accuracy, 𝒪(ϵ){\mathcal{O}}(\epsilon) as opposed to 𝒪(ϵ){\mathcal{O}}(\sqrt{\epsilon}), with this approach.

This improvement is illustrated in Fig. 3 for the example f(x)=cos16xf(x)=\cos 16x. As shown in the left panel, the best attainable error with the continuous Fourier extension is roughly 10810^{-8}, whereas the discrete extension obtains much closer to machine epsilon. The right panel indicates the significantly milder growth in condition number.

Refer to caption
(a) Approximation error
Refer to caption
(b) Condition number
Figure 3: The approximation errors fgnL[1,1]\|f-g_{n}\|_{L^{\infty}_{[-1,1]}} (left panel) and condition numbers κ(A)\kappa(A) (right panel) for the continuous (squares) and discrete (circles) Fourier extensions with T=2T=2 and f(x)=cos16xf(x)=\cos 16x.

Such an improvement is by no means unique to this choice of nodes. A suitable alternative choice of collocation points follows from using the roots of the polynomials TkT(y)T_{k}^{T}(y). With an appropriate number of points, the matrix A~TDA~\tilde{A}^{T}D\tilde{A}, with DD containing the weights of the associated Gaussian quadrature rule on its diagonal, is precisely the matrix of the unweighted continuous Fourier extension, rather than the weighted extension that was identified in Lemma 3.2. The use of these points as Gaussian quadrature was already explored in [25]. However, these points depend on TT in a non-trivial way and, although they can easily be computed, are not available in closed form. As numerical experiments indicate that the performance of such point sets is not significantly better than the points proposed in this section, we forego a more detailed analysis.

This aside, we remark that, in some senses, the continuous and discrete Fourier extensions are analogous to the orthogonal expansion of a function in Chebyshev polynomials and its Chebyshev polynomial interpolant (sometimes known as the discrete Chebyshev expansion). Specifically, the discrete versions both arise by essentially replacing an inner product by a quadrature. However, the correspondence is not exact, since in the discrete Fourier extension the quadrature approximates the weighted inner product on LW2[1,1]L^{2}_{W}[-1,1]. As mentioned, this is done for computational convenience. Nonetheless, the approximation-theoretic properties of the discrete and continuous Fourier extensions are nearly identical, with the only key difference being that of the condition number.

We have purposefully kept this section on numerical computation of the continuous and discrete Fourier extensions short. The main conclusion is that, despite quite severe ill-conditioning (even in the discrete case we have exponentially large condition numbers), one can typically obtain very high accuracies (i.e. close to machine precision) with the discrete Fourier extension. Although this apparent contradiction can be comprehensively explained, it is beyond the scope of this paper, and will be reported in a subsequent work [2].

4 Resolution power of polynomial expansions

Having introduced and analyzed the continuous and discrete Fourier extensions, and discussed their numerical implementation, in §5 we establish their resolution power. Before doing so, since similar techniques will be used subsequently, we first derive the well-known result for orthogonal polynomial expansions.

Suppose that f(x)f(x) is defined on [1,1][-1,1] and is analytic in a neighbourhood thereof. Let fnnf_{n}\in\mathbb{P}_{n} be its nn-term expansion in orthogonal polynomials with respect to the positive and integrable weight function w(x)w(x). If Lw2[1,1]L^{2}_{w}[-1,1] is the space of weighted square integrable functions with respect to ww, then fnf_{n} is given by

fn:=argminpnfpL[1,1],w2.f_{n}:=\underset{p\in\mathbb{P}_{n}}{\operatorname{argmin}}\|f-p\|_{L^{2}_{[-1,1],w}}.

In particular, for any pn𝒫np_{n}\in\mathcal{P}_{n},

ffnL[1,1],w22fpnL[1,1],w22wL[1,1]1fpnL[1,1]2.\|f-f_{n}\|^{2}_{L^{2}_{[-1,1],w}}\leq\|f-p_{n}\|^{2}_{L^{2}_{[-1,1],w}}\leq\|w\|_{L^{1}_{[-1,1]}}\|f-p_{n}\|^{2}_{L^{\infty}_{[-1,1]}}.

Hence, we note that to study of the resolution power of any orthogonal polynomial expansion it suffices to consider only one particular example. Without loss of generality, we now focus on expansions in Chebyshev polynomials (i.e. w(x)=(1x2)12w(x)=(1-x^{2})^{-\frac{1}{2}}), in which case

pn(x)=k=0nakTk(x),f(x)=k=0akTk(x),p_{n}(x)=\sum^{n}_{k=0}\leavevmode\nobreak\ {}^{\prime}a_{k}T_{k}(x),\qquad f(x)=\sum_{k=0}^{\infty}\leavevmode\nobreak\ {}^{\prime}a_{k}T_{k}(x), (18)

where

ak=0πf(cosθ)coskθdθ,a_{k}=\int_{0}^{\pi}f(\cos\theta)\cos k\theta\,{\rm d}\theta, (19)

and indicates that the first term of the sum should be halved. Since the infinite sum (18) converges uniformly, we have the estimate

fpnL[1,1]k>n|ak|.\|f-p_{n}\|_{L^{\infty}_{[-1,1]}}\leq\sum_{k>n}|a_{k}|.

Therefore it suffices to examine the nature of the coefficients ana_{n} for the function

f(x)=exp(iπωx).f(x)=\exp(\textrm{i}\pi\omega x). (20)

We use the following standard estimate, given in [35, p.175]:

|an|2Mρρn.|a_{n}|\leq\frac{2M_{\rho}}{\rho^{n}}. (21)

Here ρ\rho corresponds to any Bernstein ellipse e(ρ)e(\rho) in which ff is analytic and MρM_{\rho} is the maximum of |f(z)||f(z)| along that ellipse.

For a given ω\omega and nn, we consider the minimum of all bounds of the form (21). Since ff is a complex exponential, ff reaches a maximum at the point on the ellipse e(ρ)e(\rho) with the smallest (negative) imaginary part. This corresponds to θ=π/2\theta=-\pi/2 and thus we have

Mρ=exp(πωρ212ρ).M_{\rho}=\exp\left(\pi\omega\frac{\rho^{2}-1}{2\rho}\right). (22)

The bound (21) now becomes

B(ω,n,ρ):=2Mρρn=2exp(πωρ212ρ)ρn.B(\omega,n,\rho):=\frac{2M_{\rho}}{\rho^{n}}=\frac{2\exp\left(\pi\omega\frac{\rho^{2}-1}{2\rho}\right)}{\rho^{n}}. (23)

For fixed ω\omega and nn, we find the minimal value of this bound by differentiating (23) with respect to ρ\rho. Equating the derivative to zero

ddρB(ω,n,ρ)=0,\frac{d}{d\rho}B(\omega,n,\rho)=0, (24)

yields two solutions

n±n2π2ω2πω.\frac{n\pm\sqrt{n^{2}-\pi^{2}\omega^{2}}}{\pi\omega}. (25)

Consider the case n<πωn<\pi\omega. Both solutions of (24) are complex-valued. Note that

B(ω,n,1)=2,B(\omega,n,1)=2,

and

ddρB(ω,n,1)=2πω2n.\frac{d}{d\rho}B(\omega,n,1)=2\pi\omega-2n.

It follows that BB is strictly increasing as a function of ρ\rho. For n<πωn<\pi\omega, the best possible bound for ana_{n} of the form (21) is 22.

Consider next the case n>πωn>\pi\omega. Since it is easily seen that the roots (25) are inverses of each other, we restrict our attention to the one greater than 11, i.e.,

ρmin:=n+n2π2ω2πω.\rho_{min}:=\frac{n+\sqrt{n^{2}-\pi^{2}\omega^{2}}}{\pi\omega}.

Since B(ω,n,ρ)B(\omega,n,\rho) initially decays at ρ=1\rho=1, ρmin\rho_{min} is the unique minimum of the bound for ρ>1\rho>1. Thus,

B(ω,n,ρmin)<B(ω,n,ρ),ρ[1,).B(\omega,n,\rho_{min})<B(\omega,n,\rho),\qquad\rho\in[1,\infty).

In particular, this implies exponential decay of the next coefficients:

|an+k|B(ω,n,ρmin)1ρmink<2ρmink.|a_{n+k}|\leq B(\omega,n,\rho_{min})\frac{1}{\rho_{min}^{k}}<\frac{2}{\rho_{min}^{k}}.

Finally, consider the value n=πωn=\pi\omega. Then ρ=1\rho=1 is a global minimum of the bound (23), since

ddρB(ω,πω,1)=d2dρ2B(ω,πω,1)=0,\frac{d}{d\rho}B(\omega,\pi\omega,1)=\frac{d^{2}}{d\rho^{2}}B(\omega,\pi\omega,1)=0,

and

d3dρ3B(ω,πω,1)=2πω>0.\frac{d^{3}}{d\rho^{3}}B(\omega,\pi\omega,1)=2\pi\omega>0.

In conclusion, we have seen that for an oscillatory function of the form (20), we need n=πωn=\pi\omega degrees of freedom before exponential decay of the coefficients ana_{n} sets in. This corresponds to π\pi degrees of freedom per wavelength – the well-known result on the resolution power of polynomials.

Note that this value was previously derived in [21] for orthogonal polynomial expansions corresponding to the Gegenbauer weight w(x)=(1x2)λ12w(x)=(1-x^{2})^{\lambda-\frac{1}{2}}, λ>12\lambda>-\frac{1}{2} (see also [20, p.35] for λ=0\lambda=0). This result was based on explicit expressions for the Gegenbauer polynomial coefficients of eixte^{\textrm{i}xt}. The previous arguments generalize this result to arbitrary weight functions w(x)w(x). In fact, we could have also used the explicit formula for the Chebyshev polynomial expansion of eixte^{\textrm{i}xt} to derive estimates similar to those given above (and possibly more accurate). However, we shall use similar techniques to those presented here in the next section, where such explicit expressions are not available.

5 Resolution power of Fourier extensions

We now consider the resolution power of the continuous and discrete Fourier extensions. As commented in §3, the continuous/discrete Fourier extension gng_{n} need not be realized in a finite precision numerical computation. Hence, we divide this section between theoretical estimates for the resolution power of the continuous/discrete extension, and its numerical realization.

5.1 Resolution power of the continuous Fourier extension

A naïve estimate for the resolution constant r(T)r(T) follows immediately from the bound (5) in Theorem 2.1. Indeed, for f(x)=exp(iωπx)f(x)=\exp(i\omega\pi x) we have

fH[1,1]k=𝒪(ωπ)k,k=1,2,,\|f\|_{H^{k}_{[-1,1]}}={\mathcal{O}}{(\omega\pi)^{k}},\quad k=1,2,\ldots,

and therefore r(T)r(T) satisfies r(T)2Tr(T)\leq 2T, with spectral convergence occurring once nn exceeds ωT\omega T.

On first viewing, this estimate seems plausible. For example, consider the special situation where the frequency of oscillation ω\omega is an integer multiple of T1T^{-1}. Then the function f(x)=exp(iπTmx)f(x)=\exp(\textrm{i}\frac{\pi}{T}mx) is precisely the mmth complex exponential in the Fourier basis on [T,T][-T,T]. Given that the continuous Fourier extension gng_{n} of ff is error minimizing amongst all functions in GnG_{n}, we can expect ff to be recovered perfectly (i.e. gnfg_{n}\equiv f) by its continuous Fourier extension whenever nm=ωTn\geq m=\omega T. Thus, for oscillations at frequencies ω=mT\omega=\frac{m}{T}, mm\in\Z, the estimate r(T)2Tr(T)\leq 2T appears correct.

However, empirical results indicate that such an estimate is only accurate for small TT: for large TT it transpires that r(T)πr(T)\sim\pi. We shall prove this result subsequently. Before doing so, however, let us note the following unexpected conclusion: when TT is large, we can actually resolve the oscillatory exponentials exp(iπTmx)\exp(\textrm{i}\frac{\pi}{T}mx) accurately on [1,1][-1,1] using Fourier extensions comprised of relatively non-oscillatory exponentials exp(iπTnx)\exp(\textrm{i}\frac{\pi}{T}nx), nmn\ll m.

To obtain an accurate estimate for r(T)r(T), we need to argue along the same lines as §2.2 and exploit the close relationship between Fourier extensions and certain orthogonal polynomial expansions. Recall that the theory of §2.2 treats even and odd cases separately. Let us first assume that ff is even, so that

f(x)=cosωπxf(x)=\cos\omega\pi x

(we consider the odd case later). Upon applying the transformation x=Tπcos1yx=\frac{T}{\pi}\cos^{-1}y we obtain

f1(y)=cosωTcos1y,y[c(T),1].f_{1}(y)=\cos\omega T\cos^{-1}y,\quad y\in[c(T),1].

The Fourier extension of f(x)f(x) is precisely the expansion of f1(y)f_{1}(y) in the orthogonal polynomials TkT(y)T^{T}_{k}(y). If we now map the domain [c(T),1][c(T),1] to [1,1][-1,1] via

u=c(T)+12yc(T)1,y=u2(1c(T))+12(1+c(T)),u=\frac{c(T)+1-2y}{c(T)-1},\quad y=\frac{u}{2}(1-c(T))+\frac{1}{2}(1+c(T)),

then this equates to an orthogonal polynomial expansion of the function

f2(u)=cos[ωTcos1(u2(1c(T))+12(1+c(T)))],u[1,1].f_{2}(u)=\cos\left[\omega T\cos^{-1}\left(\frac{u}{2}(1-c(T))+\frac{1}{2}(1+c(T))\right)\right],\quad u\in[-1,1]. (26)

Thus, as in §4, to determine the resolution power of gng_{n}, it suffices to consider the expansion of f2(u)f_{2}(u) in Chebyshev polynomials on [1,1][-1,1].

In view of the bound (21), we now seek the maximum value of f2(u)f_{2}(u) along the Bernstein ellipse e(ρ)e(\rho). We first require the following lemma:

Lemma 5.1.

For ρ<E(T)\rho<E(T) and sufficiently large ω\omega, the maximum value of |f2(u)||f_{2}(u)| on the Bernstein ellipse e(ρ)e(\rho) occurs at θ=0\theta=0.

Proof.

Let z=x+iyz=x+\textrm{i}y. Then

|cosωz|2=12(cos2ωx+cosh2ωy).|\cos\omega z|^{2}=\frac{1}{2}\left(\cos 2\omega x+\cosh 2\omega y\right).

Hence, for sufficiently large ω\omega, the maximal value of |cosωz||\cos\omega z| on some curve CC in the complex plane occurs at the point zCz\in C where |z||\Im z| is maximized.

When u=12(ρ1eiθ+ρeiθ)u=\frac{1}{2}(\rho^{-1}e^{-\textrm{i}\theta}+\rho e^{\textrm{i}\theta}), we may write f2(u)=cosωT[X(θ)+iY(θ)]f_{2}(u)=\cos\omega T[X(\theta)+\textrm{i}Y(\theta)], where X(θ)X(\theta) and Y(θ)Y(\theta) are defined by

cos[X(θ)+iY(θ)]=u2(1c(T))+12(1+c(T))=A(θ)+iB(θ),\cos\left[X(\theta)+\textrm{i}Y(\theta)\right]=\frac{u}{2}(1-c(T))+\frac{1}{2}(1+c(T))=A(\theta)+\textrm{i}B(\theta),

and, letting ρ=eη\rho=e^{\eta} for η>0\eta>0 (recall that ρ>1\rho>1),

A(θ)=12(1c(T))coshηcosθ+12(1+c(T)),B(θ)=12(1c(T))sinhηsinθ.A(\theta)=\frac{1}{2}(1-c(T))\cosh\eta\cos\theta+\frac{1}{2}(1+c(T)),\quad B(\theta)=\frac{1}{2}(1-c(T))\sinh\eta\sin\theta.

Expanding the cosine, and equating real and imaginary parts, we find that

cosX(θ)coshY(θ)=A(θ),sinX(θ)sinhY(θ)=B(θ).\cos X(\theta)\cosh Y(\theta)=A(\theta),\quad\sin X(\theta)\sinh Y(\theta)=B(\theta).

We seek to maximize |Y(θ)||Y(\theta)|. Note trivially that Y(θ)Y(\theta) cannot vanish identically for all θ\theta, and thus it suffices to consider only those θ\theta for which Y(θ)>0Y(\theta)>0. Let Z(θ)=cosh2Y(θ)Z(\theta)=\cosh^{2}Y(\theta). Then Z(θ)Z(\theta) is defined by

A2(θ)Z(θ)+B2(θ)Z(θ)1=1.\frac{A^{2}(\theta)}{Z(\theta)}+\frac{B^{2}(\theta)}{Z(\theta)-1}=1.

Rearranging,

Z2(θ)(1+A2(θ)+B2(θ))Z(θ)+A2(θ)=0.Z^{2}(\theta)-(1+A^{2}(\theta)+B^{2}(\theta))Z(\theta)+A^{2}(\theta)=0.

To complete the proof, it suffices to show that ZZ attains its maximum value at θ=0\theta=0. Suppose not. Then Z(θ0)=0Z^{\prime}(\theta_{0})=0 for some θ00\theta_{0}\neq 0 (with Z(θ0)>1Z(\theta_{0})>1), and, after differentiating the above expression, we obtain

(A(θ0)A(θ0)+B(θ0)B(θ0))Z(θ0)=A(θ0)A(θ0),\left(A^{\prime}(\theta_{0})A(\theta_{0})+B^{\prime}(\theta_{0})B(\theta_{0})\right)Z(\theta_{0})=A^{\prime}(\theta_{0})A(\theta_{0}),

i.e.

sinθ0((1c(T))\displaystyle\sin\theta_{0}\left((1-c(T))\right. cosθ0+(1+c(T))coshη)Z(θ0)\displaystyle\cos\theta_{0}\left.+(1+c(T))\cosh\eta\right)Z(\theta_{0})
=sinθ0coshη(1+c(T)(c(T)1)coshηcosθ0).\displaystyle=\sin\theta_{0}\cosh\eta\left(1+c(T)-(c(T)-1)\cosh\eta\cos\theta_{0}\right).

Let us assume first that θ00,±π\theta_{0}\neq 0,\pm\pi. Hence

((1c(T))\displaystyle\left((1-c(T))\right. cosθ0+(1+c(T))coshη)Z(θ0)\displaystyle\cos\theta_{0}\left.+(1+c(T))\cosh\eta\right)Z(\theta_{0})
=coshη(1+c(T)(c(T)1)coshηcosθ0).\displaystyle=\cosh\eta\left(1+c(T)-(c(T)-1)\cosh\eta\cos\theta_{0}\right). (27)

Suppose the left-hand side vanishes of (5.1) vanishes, i.e.

cosθ0=1+c(T)c(T)1coshη.\cos\theta_{0}=\frac{1+c(T)}{c(T)-1}\cosh\eta. (28)

Since sin(θ0)0\sin(\theta_{0})\neq 0 the right-hand side of (5.1) must also vanish, and therefore

cosθ0=1+c(T)c(T)1sechη.\cos\theta_{0}=\frac{1+c(T)}{c(T)-1}\mathrm{sech}\eta. (29)

Equations (28) and (29) cannot hold simultaneously (since η>0\eta>0), and therefore the left-hand side of (5.1) cannot vanish. We deduce that

Z(θ0)=coshη(1+c(T)(c(T)1)coshηcosθ0)(1c(T))cosθ0+(1+c(T))coshη.Z(\theta_{0})=\frac{\cosh\eta\left(1+c(T)-(c(T)-1)\cosh\eta\cos\theta_{0}\right)}{(1-c(T))\cos\theta_{0}+(1+c(T))\cosh\eta}.

We now substitute this into the quadratic equation for Z(θ)Z(\theta) to give

(cosθ0coshη)2((c(T)1)coshηcosθ0(1+c(T)))=0.\left(\cos\theta_{0}-\cosh\eta\right)^{2}\left((c(T)-1)\cosh\eta\cos\theta_{0}-(1+c(T))\right)=0.

Since η>0\eta>0, we must have that

cosθ0=1+c(T)(c(T)1)sechη.\cos\theta_{0}=\frac{1+c(T)}{(c(T)-1)}\mbox{sech}\eta.

However, substituting this into (5.1), gives Z(θ0)=0Z(\theta_{0})=0, a contradiction. Therefore θ0=0,±π\theta_{0}=0,\pm\pi. It is easy to see that θ0=±π\theta_{0}=\pm\pi leads to a smaller value of Z(θ0)Z(\theta_{0}) than θ0=0\theta_{0}=0. Hence the result follows. ∎

Using this lemma, we deduce that the nnth coefficient ana_{n} of the Chebyshev expansion of f2(u)f_{2}(u) is bounded by

B(ω,n,ρ,T)\displaystyle B(\omega,n,\rho,T) =2ρncos[ωTcos1(14(ρ+1ρ)(1c(T))+12(1+c(T)))]\displaystyle=\frac{2}{\rho^{n}}\cos\left[\omega T\cos^{-1}\left(\frac{1}{4}\left(\rho+\frac{1}{\rho}\right)(1-c(T))+\frac{1}{2}(1+c(T))\right)\right]
=2ρncos[iωTcosh1(14(ρ+1ρ)(1c(T))+12(1+c(T)))].\displaystyle=\frac{2}{\rho^{n}}\cos\left[i\omega T\cosh^{-1}\left(\frac{1}{4}\left(\rho+\frac{1}{\rho}\right)(1-c(T))+\frac{1}{2}(1+c(T))\right)\right].

We proceeded in §4 by computing the roots of the partial derivative of the bound with respect to ρ\rho. The same approach applies here, but unfortunately it does not lead to explicit expressions. However, a simple modification makes the bound more manageable. We write

cos(x)=12(eix+eix).\cos(x)=\frac{1}{2}(e^{ix}+e^{-ix}).

Since the argument of the cosine for ρ>1\rho>1 is purely imaginary, with positive imaginary part, the second exponential dominates the first. Hence, for large ω\omega, we may approximate B(w,n,ρ,T)B(w,n,\rho,T) by

B~(ω,n\displaystyle\tilde{B}(\omega,n ,ρ,T)\displaystyle,\rho,T)
:=1ρnexp[ωTcosh1(14(ρ+1ρ)(1c(T))+12(1+c(T)))].\displaystyle:=\frac{1}{\rho^{n}}\exp\left[\omega T\cosh^{-1}\left(\frac{1}{4}\left(\rho+\frac{1}{\rho}\right)\left(1-c(T)\right)+\frac{1}{2}\left(1+c(T)\right)\right)\right]. (30)

One can then explicitly find roots of the partial derivative of B~\tilde{B} with respect to ρ\rho. We have

ρ(n)\displaystyle\rho^{*}(n)
=\displaystyle= n2(c(T)+3)+ω2T2(c(T)1)±2nω2T2(c(T)21)+2n2(c(T)+1)ω2T2(c(T)1)+n2(1c(T)).\displaystyle-\frac{n^{2}(c(T)+3)+\omega^{2}T^{2}(c(T)-1)\pm 2n\sqrt{\omega^{2}T^{2}(c(T)^{2}-1)+2n^{2}(c(T)+1)}}{\omega^{2}T^{2}(c(T)-1)+n^{2}(1-c(T))}. (31)

These two roots are again each other’s inverse. One quickly verifies that the square root is real and positive if and only if

n12ωT22c(T)=12ωr(T),n\geq\frac{1}{2}\omega T\sqrt{2-2c(T)}=\frac{1}{2}\omega r(T), (32)

with

r(T):=T22c(T)=2Tsin(π2T).r(T):=T\sqrt{2-2c(T)}=2T\sin\left(\frac{\pi}{2T}\right). (33)

If the condition (32) is satisfied, selecting the ++ sign in (31) yields the root that is greater than 11.

A little care is necessary when applying the bound (30). Recall that f2(u)f_{2}(u) is only analytic in Bernstein ellipses e(ρ)e(\rho) with ρ<E(T)\rho<E(T). It may be the case that ρ(n)E(T)\rho^{*}(n)\geq E(T) and therefore we cannot use this bound directly. However, explicit computation yields the condition

E(T)ρ(n)n26+2c(T)ωT=11+cos2(π2T)ωT.E(T)\geq\rho^{*}(n)\quad\Leftrightarrow\quad n\geq\frac{2}{\sqrt{6+2c(T)}}\omega T=\frac{1}{\sqrt{1+\cos^{2}\left(\frac{\pi}{2T}\right)}}\omega T. (34)

Moreover, if r(T)r(T) is given by (33), then

12r(T)<T1+cos2(π2T),T>1,\frac{1}{2}r(T)<\frac{T}{\sqrt{1+\cos^{2}\left(\frac{\pi}{2T}\right)}},\quad\forall T>1,

With this to hand, we are now able to distinguish between the following cases:

  1. 1.

    𝐧<𝟏𝟐ω𝐫(𝐓).\mathbf{n<\frac{1}{2}\omega r(T)}. In this case, the argument of the square root in (31) is negative and hence both roots are imaginary. The function B~(ω,n,ρ,T)\tilde{B}(\omega,n,\rho,T) is either monotonically increasing or decreasing as a function of ρ\rho on [1,)[1,\infty). Reasoning as before, note that

    B~ρ|ρ=1=12ωr(T)n.\frac{\partial\tilde{B}}{\partial\rho}\big{|}_{\rho=1}=\frac{1}{2}\omega r(T)-n. (35)

    The partial derivative vanishes precisely when n=12ωr(T)n=\frac{1}{2}\omega r(T) and it is positive for smaller nn. Thus, the bound is increasing and the best possible bound we can find of the form (21) is

    |an|B~(ω,n,1,T)=2,n<12ωr(T).|a_{n}|\leq\tilde{B}(\omega,n,1,T)=2,\quad n<\frac{1}{2}\omega r(T).
  2. 2.

    𝟏𝟐ω𝐫(𝐓)<𝐧<𝐓𝟏+cos𝟐(π𝟐𝐓)ω\mathbf{\frac{1}{2}\omega r(T)<n<\frac{T}{\sqrt{1+\cos^{2}\left(\frac{\pi}{2T}\right)}}\omega}. In this case the argument of the square root in (31) is positive. Moreover, the condition n<ωTn<\omega T ensures that the overall expression for both roots ρ\rho^{*} is positive (note that the denominator switches sign at n=ωTn=\omega T). Thus, there is a minimum of B~(ω,n,ρ,T)\tilde{B}(\omega,n,\rho,T) at ρ=ρ(n)>1\rho=\rho^{*}(n)>1, and this minimum satisfies ρ(n)<E(T)\rho^{*}(n)<E(T). Therefore, we obtain the bound

    |an+k|2Mρ(n)(ρ(n))n+k1(ρ(n))kB~(ω,n,ρ(n),T)2(ρ(n))k,k.|a_{n+k}|\leq\frac{2M_{\rho^{*}(n)}}{(\rho^{*}(n))^{n+k}}\leq\frac{1}{(\rho^{*}(n))^{k}}\tilde{B}(\omega,n,\rho^{*}(n),T)\leq\frac{2}{(\rho^{*}(n))^{k}},\quad\forall k\in\N.

    Hence, we deduce exponential decay of the coefficients ana_{n} once nn exceeds n>12ωr(T)n>\frac{1}{2}\omega r(T).

  3. 3.

    𝐓𝟏+cos𝟐(π𝟐𝐓)ω<𝐧<ω𝐓\mathbf{\frac{T}{\sqrt{1+\cos^{2}\left(\frac{\pi}{2T}\right)}}\omega<n<\omega T}. The arguments of the previous case still hold, but now ρ(n)E(T)\rho^{*}(n)\geq E(T). Thus, the minimum value of B~(ω,n,ρ,T)\tilde{B}(\omega,n,\rho,T) is obtained at ρ=E(T)\rho=E(T) and therefore

    |an|2E(T)nexp[ωTcosh1(2+c(T))].|a_{n}|\leq\frac{2}{E(T)^{n}}\exp\left[\omega T\cosh^{-1}\left(2+c(T)\right)\right]. (36)

    Note that this bound is valid for all nn\in\N, not just nn in the stated range. However, it only gives an accurate portrait of the coefficient decay once nn exceeds T1+cos2(π2T)ω\frac{T}{\sqrt{1+\cos^{2}\left(\frac{\pi}{2T}\right)}}\omega.

  4. 4.

    𝐧>ω𝐓\mathbf{n>\omega T}. The denominator of (31) vanishes at n=ωTn=\omega T and switches sign for larger nn, so that ρ(n)\rho^{*}(n) becomes negative. Based on (35), we conclude that the bound is monotonically decreasing as a function of ρ\rho on [1,)[1,\infty). Once more we are limited to choosing ρE(T)\rho\leq E(T), and therefore the estimate (36) is also applicable in this case.

This derivation establishes the resolution constant r(T)=2Tsin(π2T)r(T)=2T\sin\left(\frac{\pi}{2T}\right), but only for even oscillations f(x)=cosωπxf(x)=\cos\omega\pi x. To prove the complete result, we need to obtain an equivalent statement for the odd functions

f(x)=sinωπx.f(x)=\sin\omega\pi x. (37)

We proceed along similar lines to the cosine case. First, note that the continuous Fourier extension of (37) equates to the orthogonal polynomial expansion of

f2(u)=sin[ωTcos1(u2(1c(T))+12(1+c(T))]1[u2(1c(T))+12(1+c(T))]2,u[1,1].f_{2}(u)=\frac{\sin\left[\omega T\cos^{-1}\left(\frac{u}{2}(1-c(T))+\frac{1}{2}(1+c(T)\right)\right]}{\sqrt{1-\left[\frac{u}{2}(1-c(T))+\frac{1}{2}(1+c(T))\right]^{2}}},\quad u\in[-1,1].

It is not immediately apparent that this function is analytic at u=1u=1. However, simple analysis reveals that the square-root type singularity is removable, and consequently f2(u)f_{2}(u) is analytic in any Bernstein ellipse e(ρ)e(\rho) with ρ<E(T)\rho<E(T).

Proceeding as before, we find that the nnth Chebyshev polynomial coefficient of f2(u)f_{2}(u) admits the bound |an|B(ω,n,ρ,T)|a_{n}|\leq B(\omega,n,\rho,T), ρ>1\forall\rho>1, where

B(ω,n,ρ,T)=2sinh[ωTcosh1(14(ρ+1ρ)(1c(T))+12(1+c(T)))]ρnC(ρ,T),B(\omega,n,\rho,T)=\frac{2\sinh\left[\omega T\cosh^{-1}\left(\frac{1}{4}(\rho+\frac{1}{\rho})(1-c(T))+\frac{1}{2}(1+c(T))\right)\right]}{\rho^{n}\sqrt{C(\rho,T)}},

and C(ρ,T)=[14(ρ+1ρ)(1c(T))+12(1+c(T))]21C(\rho,T)=\left[\frac{1}{4}(\rho+\frac{1}{\rho})(1-c(T))+\frac{1}{2}(1+c(T))\right]^{2}-1. Suppose first that n<12ωr(T)n<\frac{1}{2}\omega r(T). Letting ρ1+\rho\rightarrow 1^{+}, we find that

|an|limρ1+B(ω,n,ρ,T)=2ωT.|a_{n}|\leq\lim_{\rho\rightarrow 1^{+}}B(\omega,n,\rho,T)=2\omega T.

Next suppose that 12ωr(T)<n<T1+cos2(π2T)ω\frac{1}{2}\omega r(T)<n<\frac{T}{\sqrt{1+\cos^{2}\left(\frac{\pi}{2T}\right)}}\omega. Then, for sufficiently large ω\omega, we can replace B(ω,n,ρ,T)B(\omega,n,\rho,T) by

B~(ω,n,ρ,T)C(ρ,T),\frac{\tilde{B}(\omega,n,\rho,T)}{\sqrt{C(\rho,T)}},

where B~(ω,n,ρ,T)\tilde{B}(\omega,n,\rho,T) is the bound (30) of the cosine case. Hence, we deduce that

|an+k|\displaystyle|a_{n+k}| 2(ρ(n))kC(ρ(n),T)\displaystyle\leq\frac{2}{(\rho^{*}(n))^{k}\sqrt{C(\rho^{*}(n),T)}}
=2(ρ(n))kn2ω2T22(1+c(T))n2ω2T2+(c(T)21)ω4T4,k.\displaystyle=\frac{2}{(\rho^{*}(n))^{k}}\sqrt{\frac{n^{2}-\omega^{2}T^{2}}{2(1+c(T))n^{2}\omega^{2}T^{2}+(c(T)^{2}-1)\omega^{4}T^{4}}},\quad\forall k\in\N.

Note that the square root term is bounded since n>12ωr(T)n>\frac{1}{2}\omega r(T). Indeed, it attains its maximal value at n=12ωr(T)+1n=\frac{1}{2}\omega r(T)+1, and hence is bounded in both nn and ω\omega. This confirms exponential decay of the coefficients ana_{n} for n>12ωr(T)n>\frac{1}{2}\omega r(T).

In the third scenario, i.e. n>T1+cos2(π2T)ωn>\frac{T}{\sqrt{1+\cos^{2}\left(\frac{\pi}{2T}\right)}}\omega, we use the value ρ=E(T)\rho=E(T) to give

|an|2(E(T))nexp[ωTcosh1(2+c(T))](2+c(T))21.|a_{n}|\leq\frac{2}{(E(T))^{n}}\frac{\exp\left[\omega T\cosh^{-1}(2+c(T))\right]}{\sqrt{(2+c(T))^{2}-1}}.

Note the similarity of this bound with (36) for the cosine case.

To sum up, we have shown the following theorem:

Theorem 5.2.

The resolution constant r=r(T)r=r(T) of the continuous Fourier extension gng_{n} is precisely 2Tsin(π2T)2T\sin\left(\frac{\pi}{2T}\right). In particular, r(T)2r(T)\sim 2 for T1T\approx 1 and r(T)πr(T)\sim\pi for T1T\gg 1.

The resolution constant r(T)r(T) is illustrated as a function of TT in Fig. 4. In Fig. 5 we confirm this theorem with several numerical examples. As discussed in §3, the result of the numerical computation of (3) may not coincide with the exact solution gng_{n}. We shall discuss the question of resolution power of the numerical solution, as opposed to the exact solution, in detail in §5.3. For the moment, so that we can illustrate Theorem 5.2, we have removed this issue by carrying out the computations in Fig. 5 in additional precision.

Refer to caption
Figure 4: The resolution constant of Fourier extensions as a function of TT.
Refer to caption
(a) T=2T=\sqrt{2}
Refer to caption
(b) T=5T=5
Figure 5: The error fgnL[1,1]\|f-g_{n}\|_{L^{\infty}_{[-1,1]}}, where f(x)=exp(iπωx)f(x)=\exp(\textrm{i}\pi\omega x) and ω=10,20,40\omega=10,20,40 (squares, circles and crosses respectively).

As shown in the right panel of Fig. 5, the resolution constant r(T)πr(T)\sim\pi for T1T\gg 1. Incidentally, the fact that r(T)r(T) is independent of TT for large TT can be seen by the studying the quantity f2(u)f_{2}(u). For fixed ω\omega, we find that

f2(u)=cos(πw21u)+𝒪(T2),T.f_{2}(u)=\cos\left(\frac{\pi w}{\sqrt{2}}\sqrt{1-u}\right)+{\mathcal{O}}{(T^{-2})},\quad T\rightarrow\infty.

The leading term is independent of TT, thus we expect r(T)r(T) to approach a constant value as TT\rightarrow\infty. It is also natural to expect that this constant is the same as the resolution constant of polynomial approximations. Indeed, for large TT the functions cosπTkx\cos\frac{\pi}{T}kx and sinπTkx\sin\frac{\pi}{T}kx are not at all oscillatory on [1,1][-1,1]. In particular, they are well approximated on [1,1][-1,1] by their Taylor series. As such, the span of the first nn such functions is very close to the span of their Taylor series, which is exactly the space of polynomials of degree nn. Thus, for fixed nn and sufficiently large TT, the Fourier extension gng_{n} of an arbitrary function ff closely resembles the best polynomial approximation to ff in the L2L^{2} sense; in other words, the Legendre polynomial expansion. The value of π\pi for the resolution constant arises from the discussion in §4.

As commented at the start of this section, oscillations at ‘periodic’ frequencies ω=mT\omega=\frac{m}{T}, mm\in\Z, are informative in that they illustrate when the Fourier extension behaves roughly like a classical Fourier series on [T,T][-T,T], that is, when T1T\approx 1, and when it does not, i.e. T1T\gg 1. The decay of the coefficients ana_{n} for these oscillations is also quite special. Since f(x)=exp(imTπx)f(x)=\exp(\textrm{i}\frac{m}{T}\pi x) is entire and periodic on [T,T][-T,T], the corresponding function f2(u)f_{2}(u) given by (26) is also entire in the variable uu. This means that f2(u)f_{2}(u) is analytic in any Bernstein ellipse e(r)e(r), and hence the corresponding Chebyshev coefficients ana_{n} (respectively, the errors fgn\|f-g_{n}\|) decay superexponentially fast, as opposed to merely exponentially fast at rate E(T)E(T). Comparison of the left and right panels of Fig. 5 illustrates this difference.

We now summarize this observation, along with the other convergence characteristics derived above, in the following theorem:

Theorem 5.3.

The error fgn\|f-g_{n}\| in approximating f(x)=exp(iπωx)f(x)=\exp(\textrm{i}\pi\omega x) by its continuous Fourier extension gng_{n} is 𝒪(1){\mathcal{O}}(1) for n<12ωr(T)n<\frac{1}{2}\omega r(T). Once nn exceeds 12ωr(T)\frac{1}{2}\omega r(T) the error begins to decay exponentially, and when n>T1+cos2(π2T)ωn>\frac{T}{\sqrt{1+\cos^{2}\left(\frac{\pi}{2T}\right)}}\omega, the rate of exponential convergence is precisely E(T)E(T), unless ω=mT\omega=\frac{m}{T} for some mm\in\Z, in which case it is superexponential.

5.2 Points per wavelength and resolution power of the discrete Fourier extension

The discrete Fourier extension possesses exactly the same resolution power as its continuous counterpart. This is a consequence of the fact that the error committed by the polynomial interpolant of a function at Chebyshev nodes can be bounded by a logarithmically growing factor (the Lebesgue constant) multiplied by the error of its expansion in Chebyshev polynomials [35].

However, the discrete Fourier extension does conveniently connect the concept of resolution power with the notion of points per wavelength (see §1). As shown by Theorem 5.2, Fourier extensions require r(T)=2Tsin(π2T)r(T)=2T\sin\left(\frac{\pi}{2T}\right) points per wavelength to resolve oscillatory behaviour, provided the corresponding nodes are of the form (17).

As it transpires, the exact value for the resolution constant can also be heuristically obtained by looking at the maximal spacing between the mapped symmetric Chebyshev nodes xix_{i} (as defined in (17)). Intuitively speaking, if the maximal node spacing for a set of nn nodes is hh, then one can only expect to be able to resolve oscillations of frequency ω<1h\omega<\frac{1}{h}. For the mapped symmetric Chebyshev nodes, one can show that the maximal spacing

h=xnxn112nr(T),n.h=x_{n}-x_{n-1}\sim\frac{1}{2n}r(T),\quad n\rightarrow\infty.

Given that there are a total of 2n+22n+2 nodes, one also obtains the stipulated value for the resolution constant in this manner. Note that identical arguments based on either equispaced or Chebyshev nodes in [1,1][-1,1] give maximal grid spacings h=2nh=\frac{2}{n} and h=πnh=\frac{\pi}{n} respectively. These correspond to the resolution constants of Fourier series and polynomial approximations, in agreement with previous discussions.

5.3 Resolution power of numerical approximations

As mentioned, it is not necessarily the case that the numerical solution of (14) (or its discrete counterpart) will coincide with the continuous (discrete) Fourier extension. Therefore, the predictions of Theorem 5.2 may not be witnessed in computations. In Fig. 6 we give results for the numerical computation of the continuous/discrete Fourier extensions. For small TT there is little difference with the left panel of Fig. 5 (except, as mentioned in §5.3, the computation of the continuous extension only attains around eight digits of accuracy). However, when T1T\gg 1 the numerically computed extension appears to possess a much larger resolution constant than the value r(T)πr(T)\sim\pi predicted by Theorem 5.2. In fact, as demonstrated in Fig. 7, the numerical Fourier extension appears to have a resolution constant of 2T2T for all TT, much like the naïve estimate given at the start of §5.1. This observation is further corroborated in Fig. 8, where a linear dependence of the numerical resolution constant with TT is witnessed.

Refer to caption
(a) continuous, T=2T=\sqrt{2}
Refer to caption
(b) continuous, T=5T=5
Refer to caption
(c) discrete, T=2T=\sqrt{2}
Refer to caption
(d) discrete, T=5T=5
Figure 6: The error fgnL[1,1]\|f-g_{n}\|_{L^{\infty}_{[-1,1]}}, where f(x)=exp(iπωx)f(x)=\exp(\textrm{i}\pi\omega x) and ω=10,20,40\omega=10,20,40.
Refer to caption
(a) continuous
Refer to caption
(b) discrete
Figure 7: The error fgnL[1,1]\|f-g_{n}\|_{L^{\infty}_{[-1,1]}}, where f(x)=exp(50iπx)f(x)=\exp(50\textrm{i}\pi x) and T=2,3,4,5T=2,3,4,5 (squares, circles, crosses and diamonds respectively).
Refer to caption
Figure 8: The error fg160L[1,1]\|f-g_{160}\|_{L^{\infty}_{[-1,1]}} against ω=1,2,,100\omega=1,2,\ldots,100, where gng_{n} is the discrete Fourier extension of f(x)=exp(iωπx)f(x)=\exp(\textrm{i}\omega\pi x) and T=2,4,8,16T=2,4,8,16 (squares, circles, crosses and diamonds respectively).

This difference can be explained intuitively. The improvement of Theorem 5.2 over the naïve estimate given at the beginning of §5.1 revolves around the observation that slowly oscillatory functions in the frame can be recombined to approximate functions with higher frequencies with exponential accuracy. This is due to the redundancy of the frame. However, such combinations necessarily yield large coefficients: the orthogonal polynomials that give the exact solution grow rapidly outside [1,1][-1,1] (see [25]) and, hence their coefficients when represented in the frame consisting of exponentials that are bounded on [T,T][-T,T] must be large. In fact, one has the following result:

Lemma 5.4.

Let the continuous Fourier extension gng_{n} of f(x)=exp(iπωx)f(x)=\exp(i\pi\omega x) have coefficient vector a2n+1a\in\mathbb{C}^{2n+1}. Then

αl2c(T)E(T)n,n<12r(T)ω,\|\alpha\|_{l^{2}}\leq c(T)E(T)^{n},\quad n<\frac{1}{2}r(T)\omega,

for some constant c(T)>0c(T)>0 depending on TT only.

Proof.

By Parseval’s relation, αl2=gnL[T,T]2\|\alpha\|_{l^{2}}=\|g_{n}\|_{L^{2}_{[-T,T]}}. Writing gng_{n} as in Theorem 2.2, we obtain

αl2=gnL[T,T]2k=0n|ak|TkTL[1,1]+k=0n1|bk|UkTL[1,1],\|\alpha\|_{l^{2}}=\|g_{n}\|_{L^{2}_{[-T,T]}}\leq\sum^{n}_{k=0}|a_{k}|\|T^{T}_{k}\|_{L^{\infty}_{[-1,1]}}+\sum^{n-1}_{k=0}|b_{k}|\|U^{T}_{k}\|_{L^{\infty}_{[-1,1]}},

where a=(a0,,an)a=(a_{0},\ldots,a_{n})^{\top}, b=(b0,,bn1)b=(b_{0},\ldots,b_{n-1})^{\top} and aka_{k} and bkb_{k} are given by (9) and (10) respectively. It can be shown that

TkTL[1,1],UkTL[1,1]E(T)k,k,\|T^{T}_{k}\|_{L^{\infty}_{[-1,1]}},\|U^{T}_{k}\|_{L^{\infty}_{[-1,1]}}\sim E(T)^{k},\quad k\rightarrow\infty,

(see [25, Thm. 3.16] for the case T=2T=2 – the extension to general TT is straightforward), and therefore

αl2c(k=0n|ak|E(T)k+k=0n1|bk|E(T)k),\|\alpha\|_{l^{2}}\leq c\left(\sum^{n}_{k=0}|a_{k}|E(T)^{k}+\sum^{n-1}_{k=0}|b_{k}|E(T)^{k}\right),

for some c>0c>0 independent of nn and ω\omega. By the analysis of the previous section, |ak|,|bk|1|a_{k}|,|b_{k}|\sim 1 for k<12r(T)ωk<\frac{1}{2}r(T)\omega, and hence we deduce the result. ∎

This lemma indicates that the coefficient vector of the continuous Fourier extension may well be exponentially large in the unresolved regime. As discussed (see §3), the numerical solution favours representations in the frame with bounded coefficients, since for large nn the system becomes underdetermined and most least squares solvers will seek a solution vector with minimal norm. We therefore conclude that, whilst the resolution power of the frame is bounded by π\pi, the resolution power of all representations in the frame with ‘reasonably small’ coefficients is only 2T2T.

In Fig. 9 we illustrate this discrepancy by comparing the ‘theoretical’ Fourier extension and its counterpart computed in standard precision. In this example, T=4T=4, i.e. E(T)25E(T)\approx 25, and we therefore witness rapid exponential growth of the theoretical coefficient vector, in agreement with Lemma 5.4. In particular, the coefficient vector is of magnitude 105510^{55} at the point at which the function ff begins to be resolved. On the other hand, since for large nn the numerical solver favours small norm coefficient vectors, the computed coefficient vector initially grows exponentially and then levels off around a1015\|a\|\approx 10^{15}. Thus, in finite precision one cannot obtain the coefficient vector required to resolve ff at the point n=12r(T)ωn=\frac{1}{2}r(T)\omega. As commented previously, these arguments can be made rigorous by performing an analysis of the numerical Fourier extension [2].

Refer to caption
(a) Numerical extension
Refer to caption
(b) Theoretical extension
Refer to caption
(c) Numerical extension
Refer to caption
(d) Theoretical extension
Figure 9: The top row shows the error fgnL[1,1]\|f-g_{n}\|_{L^{\infty}_{[-1,1]}} where f(x)=exp(iωπx)f(x)=\exp(i\omega\pi x), ω=202\omega=20\sqrt{2}, and T=4T=4. The bottom row gives the norm of the coefficient vector. The numerical extension was computed in Matlab, the theoretical extension was computed in Mathematica using higher precision arithmetic. The dotted lines indicate the values n=12r(T)ωn=\frac{1}{2}r(T)\omega and n=ωTn=\omega T.

5.4 Fixed TT versus varying TT

Whilst the principal issue highlighted in the previous section, namely, that for large TT we witness r(T)2Tr(T)\approx 2T rather than r(T)πr(T)\approx\pi in practice, is unfortunate, it is mainly the case T1T\approx 1 that is of interest, since this gives the highest resolution power.

However, with TT fixed independently of nn, the resolution constant r(T)>2r(T)>2; in other words, greater than the optimal Fourier resolution constant r=2r=2 (although this difference can be made arbitrarily small). One way to formally obtain the value of 22 is to allow T1+T\rightarrow 1^{+} with increasing nn. For example, if

T=1+cnα,T=1+\frac{c}{n^{\alpha}}, (38)

for some c>0c>0 and 0<α<10<\alpha<1, then we find that

r(T)=2+2cnα+𝒪(n2α)1,n,r(T)=2+\frac{2c}{n^{\alpha}}+{\mathcal{O}}(n^{-2\alpha})\rightarrow 1,\quad n\rightarrow\infty,

thereby giving formally optimal resolution power. The disadvantage of such a choice of TT is that one forfeits exponential convergence. In fact,

E(1+cnα)=1+cπna+𝒪(n2α),E\left(1+\frac{c}{n^{\alpha}}\right)=1+\frac{c\pi}{n^{a}}+{\mathcal{O}}(n^{-2\alpha}),

and therefore

E(1+cnα)nexp(cπn1α),n,E\left(1+\frac{c}{n^{\alpha}}\right)^{-n}\sim\exp(-c\pi n^{1-\alpha}),\quad n\rightarrow\infty, (39)

which indicates subexponential decay of the error (yet still spectral). Larger values of α\alpha lead to slow, algebraic convergence, and are consequently inadvisable.

An alternative way to choose TT is found in the literature on the Kosloff–Tal–Ezer mapping [6, 27] (recall the discussion in §1.3). In [27] the authors suggest choosing the mapping parameter based on some tolerance ϵtol\epsilon_{\mathrm{tol}}. This limits the best achievable accuracy to 𝒪(ϵtol)\mathcal{O}(\epsilon_{\mathrm{tol}}) but gives a formally optimal time-step restriction. We may do the same with the Fourier extension, by solving the equation

E(T)n=ϵtol,E(T)^{-n}=\epsilon_{\mathrm{tol}},

This gives

T=T(n,ϵtol)=π4(arctan(ϵtol)12n)1.T=T(n,\epsilon_{\mathrm{tol}})=\frac{\pi}{4}\left(\arctan\left(\epsilon_{\mathrm{tol}}\right)^{\frac{1}{2n}}\right)^{-1}. (40)

Note that

T(n,ϵtol)1log(ϵtol)πn+𝒪(n2),n,T(n,\epsilon_{\mathrm{tol}})\sim 1-\frac{\log(\epsilon_{\mathrm{tol}})}{\pi n}+\mathcal{O}(n^{-2}),\quad n\rightarrow\infty,

and therefore

r(T(n,ϵtol))22log(ϵtol)πn+𝒪(n2),n.r(T(n,\epsilon_{\mathrm{tol}}))\sim 2-\frac{2\log(\epsilon_{\mathrm{tol}})}{\pi n}+\mathcal{O}(n^{-2}),\quad n\rightarrow\infty.

Hence we expect formally optimal result power with this approach, as well as a best achievable accuracy on the order of ϵtol\epsilon_{\mathrm{tol}}.

As we show in the next section, choosing TT in this manner works particularly well in practice.

5.5 Numerical experiments

In this final section we present numerical results for the Fourier extension applied to the oscillatory functions

f1(x)=(1+x2)cos10xcos100πx,f2(x)=Ai(36x32),\displaystyle f_{1}(x)=(1+x^{2})\cos 10x\cos 100\pi x,\quad f_{2}(x)=\mbox{Ai}(-36x-32), (41)

where Ai(z)\mbox{Ai}(z) is the Airy function. Graphs of these functions are given in the top row of Fig. 10.

Refer to caption
(a) The function f1f_{1}
Refer to caption
(b) The function f2f_{2}
Refer to caption
(c) Fourier extension of f1f_{1}
Refer to caption
(d) Fourier extension of f2f_{2}
Figure 10: The top row shows the functions f1f_{1} and f2f_{2} defined by (41). The bottom row gives the error fgnL[1,1]\|f-g_{n}\|_{L^{\infty}_{[-1,1]}}, where T=1+1nT=1+\frac{1}{n} (squares), T=1+n23T=1+n^{-\frac{2}{3}} (circles), T=1+n12T=1+n^{-\frac{1}{2}} (crosses), T=43T=\frac{4}{3} (diamonds) and a formally optimal TT given by (40) with ϵtol=\epsilon_{\mathrm{tol}}=1e-13 (pluses).

In the bottom row of Fig. 10 we present the error committed by the discrete Fourier extension for various choices of TT. The convergence rates predicted in the previous section are confirmed for these examples. For TT decreasing with nn the oscillations are resolved slightly sooner, but there is slower convergence in the resolved regime. Whether the approximation error is better or worse than that obtained from a fixed value of TT (in this case T=43T=\frac{4}{3}) depends on what level of accuracy is desired. Yet, even when high accuracy is required, α=1/2\alpha=1/2 (i.e. T=1+1n1/2T=1+\frac{1}{n^{1/2}}) appears to be a good choice for these two examples, in spite of the theoretical subexponential convergence rate. The varying value of T=T(n;ϵtol)T=T(n;\epsilon_{\mathrm{tol}}), with tolerance set to ϵtol=\epsilon_{\mathrm{tol}}=1e-13, also yields comparable results. The convergence rate for the case of larger α=2/3\alpha=2/3 is slower, as expected from the estimate (39).

As discussed, one motivation for using Fourier extensions is that, as rigorously proved in this paper, they offer a higher resolution power for small TT than polynomial approximations, for which the resolution constant is π\pi (see §4). In Fig. 11 we compare the behaviour of the Chebyshev expansion and the Fourier extension approximation for the functions (41). Note that the latter resolves the oscillatory behaviour using fewer degrees of freedom, in agreement with the result of §5. Indeed, for the first example function f1f_{1} shown in the left panel, the Chebyshev expansion only begins to converge once nn exceeds 150150 (here, for purposes of comparison, the number of expansion coefficients is 2n2n), in agreement with a resolution constant of π\pi.

The results for the second function f2f_{2} are shown in the right panel of Fig. 11. The differences in resolution are smaller in this case. Yet, the fixed choice T=8/7T=8/7 and varying choices of TT still require significantly fewer degrees of freedom than the Chebyshev expansion to resolve the oscillatory behaviour. As before, a slower asymptotic convergence rate is observed for the case. Note that the oscillations of f2f_{2} are not harmonic, with the frequency increasing towards the right endpoint of the approximation interval. This seems to have the effect of reducing the advantage of schemes with optimized resolution properties.

Refer to caption
(a) Approximation of f1f_{1}
Refer to caption
(b) Approximation of f2f_{2}
Figure 11: The error fgnL[1,1]\|f-g_{n}\|_{L^{\infty}_{[-1,1]}}, where gng_{n} is Chebyshev expansion (squares), or the Fourier extension approximation with T=1+n2/3T=1+n^{-2/3} (circles), T=1+n1/2T=1+n^{-1/2} (crosses), T=87T=\frac{8}{7} (diamonds) and a formally optimal TT given by (40) with ϵtol=\epsilon_{\mathrm{tol}}=1e-13 (pluses).

6 Conclusions and challenges

The purpose of this paper was to describe and analyze the observation that Fourier extensions are eminently well suited for resolving oscillations. To do this, we derived an exact expression for the so-called resolution constant of the continuous/discrete Fourier extension, and offered an explanation as to why this constant need not be realized in the case of large TT.

Although we have touched upon the subject of numerical behaviour of Fourier extensions in this paper, we have not provided a full description. As mentioned, the continuous/discrete Fourier extensions both result in severe ill-conditioning. However, numerical examples in this paper and elsewhere (see [7, 11, 30, 31]) point towards an apparent contradiction. Namely, despite this ill-conditioning, the best achievable accuracy can still be very high (close to machine epsilon for the discrete extension). It transpires that this effect can be quite comprehensively explained, with the main conclusion being the following: the discrete Fourier extension, although the result of solving an ill-conditioned linear system, is in fact numerically stable. A paper on this topic is in preparation [2]. Incidentally, the analysis presented therein also allows one to make rigorous the arguments given in §5.3 on the differing resolution constants of theoretical and numerical Fourier extensions for large TT.

The discrete Fourier extension introduced in this paper uses values of ff on a particular nonequispaced mesh to compute the Fourier extension. As mentioned, one important use of Fourier extensions is to solve PDE’s in complex geometries. In this setting, such meshes are usually infeasible. For this reason, it is of interest to consider the question of how to compute rapidly convergent, numerically stable Fourier extensions with good resolution power from function values on equispaced meshes (as discussed, related methods based on the FE–Gram extension are developed in [3, 10, 32]). It transpires that this can be done, and we will report the details in the upcoming paper [2] (see also [30]).

Finally, we remark that there has been little discussion in this paper on the computational cost of computing Fourier extensions. Instead, we have focused on approximation-theoretic properties, such as resolution. However, M. Lyon has recently introduced a fast algorithm for this purpose [31], which allows for computation of Fourier extensions in 𝒪(n(logn)2){\mathcal{O}}(n(\log n)^{2}) operations.

Acknowledgements

The authors would like to thank Rodrigo Platte for pointing out the relation to the Kosloff–Tal–Ezer mapping. They would also like to thank Jean–François Maitre for useful discussions and comments.

References

  • [1] R. A. Adams. Sobolev Spaces. Academic Press, 1975.
  • [2] B. Adcock, D. Huybrechs, and J. Martín-Vaquero. On the numerical stability of Fourier extensions. In preparation, 2012.
  • [3] N. Albin and O. P. Bruno. A spectral FC solver for the compressible Navier–Stokes equations in general domains I: Explicit time-stepping. J. Comput. Phys., 230(16):6248–6270, 2011.
  • [4] L. Badea and P. Daripa. On a fourier method of embedding domains using an optimal distributed control. Numer. Algorithms, 32(2–4):261–273, 2003.
  • [5] F. P. Bertolotti, T. Herbert, and P. R. Spalart. Linear and nonlinear stability of the blasius boundary layer. J. Fluid Mech., 242:441–474, 1992.
  • [6] J. P. Boyd. Chebyshev and Fourier Spectral Methods. Springer–Verlag, 1989.
  • [7] J. P. Boyd. A comparison of numerical algorithms for Fourier Extension of the first, second, and third kinds. J. Comput. Phys., 178:118–160, 2002.
  • [8] J. P. Boyd. Fourier embedded domain methods: extending a function defined on an irregular region to a rectangle so that the extension is spatially periodic and C{C}^{\infty}. Appl. Math. Comput., 161(2):591–597, 2005.
  • [9] J. P. Boyd and J. R. Ong. Exponentially-convergent strategies for defeating the Runge phenomenon for the approximation of non-periodic functions. I. Single-interval schemes. Commun. Comput. Phys., 5(2–4):484–497, 2009.
  • [10] O. Bruno and M. Lyon. High-order unconditionally stable FC-AD solvers for general smooth domains I. Basic elements. J. Comput. Phys., 229(6):2009–2033, 2010.
  • [11] O. P. Bruno, Y. Han, and M. M. Pohlman. Accurate, high-order representation of complex three-dimensional surfaces via Fourier continuation analysis. J. Comput. Phys., 227(2):1094–1125, 2007.
  • [12] A. Bueno-Orovio. Fourier embedded domain methods: period and cc^{\infty} extension of a function defined on an irregular region to a rectangle via convolution with Gaussian kernels. Appl. Math. Comput., 183(2):813–818, 2006.
  • [13] A. Bueno-Orovio and V. M. Pérez-Garcia. Spectral smoothed boundary methods: The role of external boundary conditions. Numer. Methods Partial Differential Equations, 22(2):435–448, 2006.
  • [14] A. Bueno-Orovio, V. M. Pérez-Garcia, and F. H. Fenton. Spectral methods for partial differential equations on irregular domains: The spectral smoothed boundary method. SIAM J. Sci. Comput., 28(3):886–900, 2006.
  • [15] M. Buffat and L. Le Penven. A spectral fictitious domain method with internal forcing for solving elliptic PDEs. J. Comput. Phys., 230(7):2433–2450, 2011.
  • [16] C. Canuto, M. Y. Hussaini, A. Quarteroni, and T. A. Zang. Spectral methods: Fundamentals in Single Domains. Springer, 2006.
  • [17] O. Christensen. An Introduction to Frames and Riesz Bases. Birkhauser, 2003.
  • [18] K. S. Eckhoff. On a high order numerical method for functions with singularities. Math. Comp., 67(223):1063–1087, 1998.
  • [19] M. Garbey. On some applications of the superposition principle with Fourier basis. SIAM J. Sci. Comput., 22(3):1087–1116, 2000.
  • [20] D. Gottlieb and S. A. Orszag. Numerical Analysis of Spectral Methods: Theory and Applications. Society for Industrial and Applied Mathematics, 1st edition, 1977.
  • [21] D. Gottlieb and C.-W. Shu. Resolution properties of the Fourier method for discontinuous waves. Comput. Methods Appl. Mech. Engrg, 116:27–37, 1994.
  • [22] D. Gottlieb and C.-W. Shu. On the Gibbs’ phenomenon and its resolution. SIAM Rev., 39(4):644–668, 1997.
  • [23] D. Gottlieb, C.-W. Shu, A. Solomonoff, and H. Vandeven. On the Gibbs phenomenon I: Recovering exponential accuracy from the Fourier partial sum of a nonperiodic analytic function. J. Comput. Appl. Math., 43(1–2):91–98, 1992.
  • [24] N. Hale and L. N. Trefethen. New quadrature formulas from conformal maps. SIAM J. Numer. Anal., 46(2):930–948, 2008.
  • [25] D. Huybrechs. On the Fourier extension of non-periodic functions. SIAM J. Numer. Anal., 47(6):4326–4355, 2010.
  • [26] H. H. Juang, S. Hong, and M. Kanamitsu. The NCEP regional model: An update. Mon. Weather Rev., 129:2904–2922, 2001.
  • [27] D. Kosloff and H. Tal-Ezer. A modified Chebyshev pseudospectral method with an 𝒪(N1)\mathcal{O}(N^{-1}) time step restriction. J. Comput. Phys., 104:457–469, 1993.
  • [28] C. L. Lawson and R. J. Hanson. Solving least squares problems. Classics in Applied Mathematics. SIAM, Philadelphia, 1996.
  • [29] S. H. Lui. Spectral domain embedding for elliptic pdes in complex domains. J. Comput. Appl. Math., 225(2):541–557, 2009.
  • [30] M. Lyon. Approximation error in regularized SVD-based Fourier continuations. Preprint, 2011.
  • [31] M. Lyon. A fast algorithm for Fourier continuation. SIAM J. Sci. Comput., 33(6):3241–3260, 2012.
  • [32] M. Lyon and O. Bruno. High-order unconditionally stable FC-AD solvers for general smooth domains II. Elliptic, parabolic and hyperbolic PDEs; theoretical considerations. J. Comput. Phys., 229(9):3358–3381, 2010.
  • [33] S. A. Orszag and M. Israeli. Numerical simulation of incompressible flow. Ann. Revs. Fluid Mech., 6:281–318, 1974. REVIEW.
  • [34] R. Pasquetti and M. Elghaoui. A spectral embedding method applied to the advection–diffusion equation. J. Comput. Phys., 125:464–476, 1996.
  • [35] T. Rivlin. Chebyshev Polynomials: from Approximation Theory to Algebra and Number Theory. Wiley New York, 1990.
  • [36] F. Sabetghadam, S. Sharafatmandjoor, and F. Norouzi. Fourier spectral embedded boundary solution of the Poisson’s and Laplace equations with Dirichlet boundary conditions. J. Comput. Phys., 228(1):55–74, 2009.
  • [37] L. N. Trefethen. Is Gauss quadrature better than Clenshaw-Curtis? SIAM Rev., 50(1):67–87, 2008.