Multidimensional Polynomial Interpolation#
Polynomial interpolation dates back to Newton, Lagrange, and others[1], and its fundamental importance in both mathematics and computing is widely recognized.
Interpolation is rooted in the principle that, in one dimension, there exists exactly one polynomial \(Q_{f, A}\) of degree \(n \in \mathbb{N}\) that interpolates a function \(f : \mathbb{R} \longrightarrow \mathbb{R}\) on \(n+1\) distinct interpolation nodes \(P_A = \{p_i\}_{i \in A} \subseteq \mathbb{R}\) where \(A=\{0, \ldots, n\}\). Specifically, this means
This makes interpolation fundamentally different from approximation, see Fig. 1.

Fig. 1 Interpolation and approximation in one dimension. While the interpolant \(Q_{f, A}\) must match the function \(f\) at the interpolation nodes \(p_0, \ldots, p_n\), an approximation \(Q^*_n\) is not required to coincide with \(f\) at all.#
Weierstrass Approximation Theorem#
The famous Weierstrass Approximation Theorem[2] asserts that any continuous function \(f : \Omega \longrightarrow \mathbb{R}\), defined on a compact domain, such as \(\Omega = [-1,1]^m\), can be uniformly approximated by polynomials[3]. However, the theorem does not require the polynomials to coincide with \(f\) at any given points. In fact, it is possible to have a sequence of multidimensional polynomials \(Q_n^*\), \(n \in \mathbb{N}\), where \(Q_n^*(x) \neq f(x)\) for all \(x \in \Omega\), but still achieve uniform approximation, that is
There are several constructive proofs of the Weierstrass approximation theorem, including the well-known version by Serge Bernstein[4]. The resulting Bernstein approximation scheme is universal and has been shown to reach the optimal (inverse-linear) approximation rate for the absolute value function \(f(x) = |x|\) [5]. However, despite its generality, the Bernstein scheme achieves only slow convergence rates for analytic functions, leading to high computational costs in practical applications.
As a result, extensive research has focused on extending one-dimensional Newton or Lagrange interpolation schemes to multiple dimensions (mD) while preserving their computational efficiency. Any method addressing this challenge must avoid Runge’s phenomenon[6] (overfitting) by ensuring uniform approximation of the interpolation target \(f : \Omega \longrightarrow \mathbb{R}\). Additionally, it resist the curse of dimensionality, achieving highly accurate polynomial approximations for general multidimensional functions with a sub-exponential demand for data samples \(f(p)\,, p \in P_A\), where \(|P_A| \in o(n^m)\).
Lifting the curse of dimensionality#
To address the problem we consider the \(l_p\)-norm \(\|\alpha\|_p = (\alpha_1^p + \cdots +\alpha_m^p)^{1/p}\), \(\alpha = (\alpha_1,\dots,\alpha_m) \in\mathbb{N}^m\), \(m \in \mathbb{N}\) and the lexicographically-ordered and complete multi-index set
This concept generalizes the one-dimensional notion of polynomial degree to multi-dimensional \(l_p\)-degree. Specifically, we consider the polynomial spaces spanned by all monomials with a bounded \(l_p\)-degree, that is
Given \(A=A_{m,n,p}\), we ask for:
Unisolvent interpolation nodes \(P_A\) that uniquely determine the interpolant \(Q_{f,A} \in \Pi_A\) by satisfying \(Q_{f,A}(p_{\alpha}) = f(p_{\alpha})\), \(\forall p_{\alpha} \in P_A\), \(\alpha \in A\).
An interpolation scheme that computes the uniquely determined interpolant \(Q_{f,A} \in \Pi_A\) efficiently and with numerical accuracy (machine precision).
The unisolvent nodes \(P_A\) that scale sub-exponentially with the spatial dimension \(m \in \mathbb{N}\), \(|P_A| \in o(n^m)\) and guarantee uniform approximation of even strongly varying functions (avoiding over fitting), such as as the Runge function \(f_R(x) = 1/(1+\|x\|^2)\), by achieving fast (ideally exponential) approximation rates.
In fact, the results of[7] suggest that the multidimensional DDS resolves issues 1–3 for the so called Trefethen functions by selecting the Euclidian \(l_2\)-degree and Leja ordered Chebyshev-Lobatto unisolvent interpolation nodes. Thus,
scales sub-exponentially with space dimension \(m\) and
converges uniformly and fast (exponentially) on \(\Omega = [-1,1]^m\).
Fig. 2 shows the approximation rates for the classic Runge function[6] in dimension \(m = 4\), which is known to exhibit Runge’s phenomenon (over-fitting) when interpolated naively.

Fig. 2 Approximation errors rates for interpolating the Runge function in dimension \(m = 4\).#
There is an optimal (upper bound) approximation rate
known[8], which we call the Trefethen rate.
In fact, the the DDS scheme numerically reaches the optimal Trefethen rate. In contrast, spline-type interpolation is based on works of Carl de Boor et al.[9][10][11][12] and limited to reach only polynomial approximation rates[13]. Similarly, interpolation using rational functions, such as Floater-Hormann interpolation[14][15] and tensorial Chebyshev interpolation, relying on \(l_{\infty}\)-degree[16], do not achieve optimality.
By combining sub-exponential growth of node (data) counts with exponential approximation rates, the DDS algorithm has the potential to lift the curse of dimensionality for interpolation problems involving regular (Trefethen) functions[7].
References