1 Introduction

We introduce the kernel-based scattered data interpolation problem following [1, 2]. Let and \( {{{\mathcal {X}}}} = \{ {\varvec{x}}_i, \; i = 1, \ldots , n\} \subset \varOmega \) be a set of distinct nodes, \(n\in {\mathbb {N}}\), where \({\varvec{x}}_i=(x_{i,1},\ldots ,x_{i,d})^{\intercal }\), \(i=1,\ldots ,n\). The scattered data interpolation problem consists in recovering an unknown function \(f: \varOmega \longrightarrow {\mathbb {R}}\) given its values at \({{{\mathcal {X}}}}\), i.e. \(\varvec{f}=\varvec{f}_{|\mathcal{X}}=(f(\varvec{x}_1),\ldots ,f(\varvec{x}_n))^{\intercal }=(f_1,\ldots ,f_n)^{\intercal }\). This can be achieved by imposing the interpolation conditions at \({{{\mathcal {X}}}}\). In particular, for kernel-based interpolation, the approximating function assumes the form:

$$\begin{aligned} S_{f,{\mathcal {X}}}(\varvec{x})=\sum _{i=1}^n c_i \kappa _{\varepsilon }(\varvec{x},\varvec{x}_i),\quad \varvec{x}\in \varOmega , \end{aligned}$$

where \(\varvec{c}=(c_1,\ldots ,c_n)^{\intercal }\in {\mathbb {R}}^n\) and \(\kappa _{\varepsilon }:\varOmega \times \varOmega \longrightarrow {\mathbb {R}}\) is a strictly positive definite kernel depending on a shape parameter \(\varepsilon >0\). In addition, \(\kappa _{\varepsilon }\) is supposed to be radial, i.e. there exists a univariate function \(\varphi _{\varepsilon }:{\mathbb {R}}_{\ge 0}\longrightarrow {\mathbb {R}}\) such that \(\kappa _{\varepsilon }(\varvec{x},\varvec{y})=\varphi _{\varepsilon }(r)\), with \(r:=\Vert \varvec{x}-\varvec{y}\Vert _2\). As a consequence, such a setting is commonly referred to as Radial Basis Function (RBF) interpolation. With this clarification and with abuse of notation, we might write simply \(\kappa \) instead of \(\kappa _{\varepsilon }\). The interpolation conditions are satisfied by finding the unique vector \(\varvec{c}\) so that

$$\begin{aligned} {\mathsf {A}}\varvec{c}=\varvec{f}, \end{aligned}$$
(1)

where \({\mathsf {A}}=({\mathsf {A}}_{i,j})=\kappa (\varvec{x}_i,\varvec{x}_j)\), \(i,j=1,\ldots ,n\), is the so-called interpolation (or collocation or simply kernel) matrix. The uniqueness of the solution of (1) is guaranteed as long as \(\kappa \) is strictly positive definite. For a more general formulation of the interpolant that involves conditionally positive definite kernels, and for a complete overview concerning kernel-based approximation, we refer the reader e.g. to [2].

The interpolant \(S_{f,{\mathcal {X}}}\) belongs to the space

$$\begin{aligned} {\mathcal {H}}_{\kappa } = \mathrm {span} \left\{ \kappa (\cdot ,{\varvec{x}}), \; {\varvec{x}} \in \varOmega \right\} , \end{aligned}$$

which, equipped with the bilinear form \((\cdot ,\cdot )_{{\mathcal {H}}_{\kappa }}\), is a pre-Hilbert space with reproducing kernel \(\kappa \). Moreover, the completion of \({\mathcal {H}}_{\kappa }\) with respect to the norm \(\Vert \cdot \Vert _{{\mathcal {H}}_{\kappa }}=\sqrt{(\cdot ,\cdot )_{{\mathcal {H}}_{\kappa }}}\) is the so-called native space \({\mathcal {N}}_{\kappa }\) associated to \(\kappa \). Many upper bounds for the interpolation error, e.g. in terms of the fill-distance [1, Theorem 14.5, p. 121] and using sampling inequalities [3, 4], are available. Here we focus on the following pointwise error bound [1, Theorem 14.2, p. 117]:

$$\begin{aligned} |f(\varvec{x})-S_{f,{\mathcal {X}}}(\varvec{x})|\le P_{\kappa ,{\mathcal {X}}}(\varvec{x})\Vert f \Vert _{{\mathcal {N}}_{\kappa }},\quad f\in {\mathcal {N}}_{\kappa }, \quad \varvec{x}\in \varOmega , \end{aligned}$$
(2)

where \(P_{\kappa ,{\mathcal {X}}}\) is the so-called power function. For the purposes of this paper, we directly define the power function as [5, p. 116, §14]:

$$\begin{aligned} P_{\kappa ,{\mathcal {X}}}(\varvec{x})=\sqrt{\kappa ({\varvec{x}},{\varvec{x}})-\varvec{\kappa }^{\intercal }({\varvec{x}}) {\mathsf {A}}^{-1} \varvec{\kappa }({\varvec{x}})}. \end{aligned}$$
(3)

where

$$\begin{aligned} \varvec{\kappa }(\varvec{x}):=(\kappa (\varvec{x},\varvec{x}_1),\ldots ,\kappa (\varvec{x},\varvec{x}_n))^{\intercal }. \end{aligned}$$

Note that (2) splits the error into two terms. The former, i.e. the power function, only depends on the nodes and on the kernel, while the latter takes into account the function f.

As a consequence, the power function gives information about how the interpolation error relates to the node distributions. Indeed, as for polynomial and spline interpolation, the approximation quality strongly depends on the distribution of the scattered data. In view of this, starting with an initial set of nodes, many adaptive strategies have been studied in order to construct well-behaved interpolation designs, i.e. interpolation sets which provide an accurate reconstruction and, preferably, affordable computational costs. In particular, the so-called greedy approaches match such purposes by iteratively adding new points to the interpolation set. The iterative rule is based on minimizing a pointwise upper bound for the interpolant. Precisely, those strategies rely on the residual of the interpolant (f-greedy) or on the value of the power function (p-greedy); refer to [6,7,8,9] for a general overview. Despite the fact that we only focus on these two schemes, other strategies that combine the two above, known as \(f \cdot p\) and f/p greedy, are available; we refer the reader to [10,11,12]. These methods fall under the context of knot insertion algorithms, which have been studied also in the context of adaptive least-squares approximation [1, §21].

Alternatively, as in our work, one could start by taking a large interpolation set, which provides an accurate approximation, and iteratively remove nodes until the resulting interpolation error does not exceed a fixed tolerance (see e.g. [13] for a general overview). These kinds of knot removal approaches aim to provide reduced models by neglecting as many points as possible from the initial set while preserving a target accuracy. However, they are usually computationally expensive, since many interpolants built upon large sets of nodes need to be constructed [14].

In this paper, in order to overcome the limitations related to the high computational complexity for reduced basis models, we take advantage of the Extension of Rippa’s Algorithm (ERA) recently provided in [15]; see also [16]. More precisely, classical reduced basis iterative strategies are usually quite demanding (about \({{{\mathcal {O}}}}(n^4)\) operations) and our goal is to study an efficient implementation for data removal approaches, namely Efficient Reduced Basis Algorithm (ERBA), which only requires about \({{{\mathcal {O}}}}(n^3)\) operations. Since n might be large, we should call the algorithm more efficient RBA. However, with abuse of language, we simply refer to the algorithm as efficient RBA. For the ERBA scheme, besides the residual-based rule, that leads to the ERBA-r, and that can be derived from the ERA, we provide an efficient scheme for computing the power function as well. The resulting ERBA-p is a reduced model that takes advantage of being accurate, fast and easy to implement.

The paper is organised as follows. In Sect. 2, we fix some notations and we introduce the Reduced Basis Algorithm (RBA) based on knot removal strategies. Its efficient implementation, i.e. the ERBA scheme, and theoretical results are then provided in Sect. 3, where a detailed analysis of the computational complexity of the proposed schemes is also included. Numerical tests that confirm the theoretical findings are presented in Sect. 4. The conclusions are offered in Sect. 5.

2 The Reduced Basis Algorithm (RBA)

In what follows, we introduce our schemes for reduced basis models based on both the residual and on the power function minimization. The latter algorithm is quasi optimal in the sense that, referring to (2), we only take into account the power function, neglecting the term involving the native space norm of the sought function. This should not be seen as a drawback. Indeed, we are able to compute a quasi-optimal subset of points independently of the function values. As a consequence, it might be relevant if one has to deal with many measurements sampled at the same points. We further point out that the residual-based scheme has strong similarities with the knot removal algorithm shown in [1, §21]. We now present the RBA scheme and in the next section we focus on its fast implementation.

Let \( {{{\mathcal {X}}}} = \{ {\varvec{x}}_i, \; i = 1, \ldots , n\} \subset \varOmega \) be the initial set of nodes, let \( {{{\mathcal {X}}}}_{s-1} = \{ {\varvec{x}}_i, \; i = 1, \ldots , n_{s-1}\} \subset \varOmega \) be the reduced set at the \((s-1)\)-th step of the algorithm (\( {{{\mathcal {X}}}}_{0} = {{{\mathcal {X}}}}\)), and let \(\tau \in {\mathbb {R}}_{>0}\) be a fixed tolerance. Choosing \(\rho \in {\mathbb {N}}\), \(\rho <n\), and letting \(\ell = \lfloor n_{s-1} /\rho \rfloor \), the s-th step of the iterative scheme, \(s \ge 1\), is as follows.

  1. 1.

    Partition \({\mathcal {X}}_{s-1}\) into \(\ell \) test sets \({\mathcal {X}}_{s-1}^1,\ldots ,{\mathcal {X}}_{s-1}^{\ell }\), \(|{\mathcal {X}}_{s-1}^j|=\rho ^j\) with \(\rho ^j \in \{\rho ,\ldots ,2\rho -1\}\), \(\rho ^j\le \max (2\rho -1,n_{s-1})\). Moreover, let \(\overline{{\mathcal {X}}}_{s-1}^j:={\mathcal {X}}_{s-1}\setminus {\mathcal {X}}_{s-1}^j\) be the training sets, \(j=1,\ldots ,\ell \).

  2. 2.

    For each \({\mathcal {X}}_{s-1}^j\), \(j=1,\ldots ,\ell \), compute:

    1. 2a.

      for the residual-based scheme:

      $$\begin{aligned} w^j = \dfrac{1}{\sqrt{\rho ^j}}\Vert \varvec{f}^j-\varvec{S}^j \Vert _2, \end{aligned}$$
      (4)

      where \(\varvec{f}^j:=\varvec{f}_{|_{{\mathcal {X}}_{s-1}^j}}\) and \({\varvec{S}}^j :=S_{f,\overline{{\mathcal {X}}}_{s-1}^j}({\mathcal {X}}_{s-1}^j)\) is the evaluation vector on \({\mathcal {X}}_{s-1}^j\) of the interpolant \(S_{f,\overline{{\mathcal {X}}}_{s-1}^j}\). Here, we need to solve an \((n_{s-1}-\rho ^j)\times (n_{s-1}-\rho ^j)\) linear system for each \(j=1,\ldots ,\ell \) (see (1));

    2. 2b.

      for the power-based scheme:

      $$\begin{aligned} w^j = \dfrac{1}{\sqrt{\rho ^j}} \Vert \varvec{P}^j\Vert _2, \end{aligned}$$

      where \(\varvec{P}^j:=P_{\kappa ,\overline{{\mathcal {X}}}_{s-1}^j}({\mathcal {X}}_{s-1}^j)\) is the evaluation vector on \({\mathcal {X}}_{s-1}^j\) of the power function \(P_{\kappa ,\overline{{\mathcal {X}}}_{s-1}^j}\). Here, we need to invert an \((n_{s-1}-\rho ^j)\times (n_{s-1}-\rho ^j)\) matrix for each \(j=1,\ldots ,\ell \) (see (3)).

  3. 3.

    Choose

    $$\begin{aligned} j^{\star }=\mathop {\mathrm {argmin}}\limits _{j\in \{1,\ldots ,\ell \}}{w^j}. \end{aligned}$$
  4. 4.

    Let \(r_{s-1} = w^{j^{\star }}\):

    1. 4a.

      if \(r_{s-1}\le \tau \), define \({\mathcal {X}}_{s} = {\mathcal {X}}_{s-1}\setminus {\mathcal {X}}_{s-1}^{j^{\star }}\) and proceed iteratively with the s-th step;

    2. 4b.

      if \(r_{s-1}>\tau \), \({\mathcal {X}}_{s} = {\mathcal {X}}_{s-1}\) is the final interpolation set obtained by the procedure.

Remark 1

In the presented RBA algorithm, at each iteration, the data set is partitioned randomly. As also confirmed by the numerical experiments carried out in Sect. 4, since \(\rho \ll n\), such randomness does not influence the algorithm in selecting meaningful designs of interpolation node sets according to the chosen approach. Nevertheless, if relevant, particular configurations of subsets can be taken into account (e.g. nearest neighbors).

The convergence of the residual-based rule has been studied in many context and for many basis functions, see e.g. [13] for a general overview. As far as the convergence of the proposed power function-based scheme is concerned, it is a consequence of [7, Lemma 2.3], precisely:

$$\begin{aligned} \Vert P_{\kappa ,{{{\mathcal {X}}}}_{s-1}} \Vert _{L_\infty (\varOmega )} \le \Vert P_{\kappa ,{{{\mathcal {X}}}}_{s}} \Vert _{L_\infty (\varOmega )}, \end{aligned}$$

being \( {\mathcal {X}}_{s} \subset {\mathcal {X}}_{s-1} \) nested sets.

From now on, in analyzing the computational complexity, we assume without loss of generality that \(\rho \) divides \(n_s\) at each step of the algorithm, and thus \(\rho _j\equiv \rho \) holds true. Therefore, at each step, a classical implementation of this method requires to solving \(n_s\) different \((n_s-\rho )\times (n_s-\rho )\) linear systems, in the residual-based case, or the inversion of an \((n_s-\rho )\times (n_s-\rho )\) matrix in the power-based setting. Hence, in both situations, each step is computationally demanding. In [1, §21], a similar residual-based approach has been speeded up by using Rippa’s algorithm in the case \(\rho =1\). In the next section, inspired by ERA [15], which speeds up the computational complexity for any \(\rho \ge 1\), we provide a fast implementation of the proposed algorithm.

3 Efficient Reduced Basis Algorithm (ERBA)

To study the computational complexity of the proposed schemes, we have to focus on the calculation of the residuals (or of the power function) that is performed at each iteration. Then, fixed a certain step s, with \(s=1,2,\ldots \), let us introduce the following notations: \({\mathcal {X}}_s:={\mathcal {X}}\), \(n_s:=n\), and let \(\varvec{p}:=(p_1,\ldots ,p_{\rho })^{\intercal }\) be the vector of \(\rho \) indices related to the elements of the subset \({\mathcal {X}}_s^j:={\mathcal {V}}\), \(|{\mathcal {X}}_s^j|=\rho \). We also adopt the following notations:

$$\begin{aligned} \varvec{f}_{\varvec{p}}:= & {} (f_{i})_{i\in \varvec{p}},\quad \varvec{f}^{\varvec{p}}:=(f_{i})_{i\notin \varvec{p}},\\ {\mathsf {A}}^{\varvec{p},\varvec{p}}:= & {} ({\mathsf {A}}_{i,j})_{i,j\notin \varvec{p}},\quad {\mathsf {A}}_{\varvec{p},\varvec{p}}:=({\mathsf {A}}_{i,j})_{i,j\in \varvec{p}},\quad {\mathsf {A}}_{\varvec{p},:}:=({\mathsf {A}}_{i,j})_{i\in \varvec{p},j\in \{1,\ldots ,n\}},\\ \varvec{\kappa }(\varvec{x}):= & {} (\kappa (\varvec{x},\varvec{x}_1),\ldots ,\kappa (\varvec{x},\varvec{x}_n))^{\intercal }, \\ \varvec{k}_{\varvec{p}}(\varvec{x})= & {} (\varvec{\kappa }(\varvec{x})_{i})_{i\in \varvec{p}},\quad \varvec{k}^{\varvec{p}}(\varvec{x})= (\varvec{\kappa }(\varvec{x})_{i})_{i\notin \varvec{p}}, \end{aligned}$$

In the following we explain our efficient strategy for the implementation of both the residual and power-based schemes, i.e. we speed up Step 2. of the RBA algorithm.

3.1 The ERBA-r

In step 2a. for the computation of \({\varvec{S}}^j\) in Eq. (4), we need to compute the interpolant and hence solve a system that leads to a matrix inversion. Precisely, let us consider

$$\begin{aligned} S_{f,{\overline{{{\mathcal {V}}}}}}(\varvec{x}):=S_{f,{{{\mathcal {X}}}}}^{(\varvec{p})}(\varvec{x})=\sum _{i\notin \varvec{p}} c^{(\varvec{p})}_i \kappa _{\varepsilon }(\varvec{x},\varvec{x}_i),\quad \varvec{x}\in \varOmega ,\; \varvec{x}_i\in {\mathcal {X}}, \end{aligned}$$

where \(\overline{{{\mathcal {V}}}}={\mathcal {X}}\setminus {\mathcal {V}}\) and \(\varvec{c}^{(\varvec{p})}:=\big (c^{(\varvec{p})}_{i}\big )_{i\notin \varvec{p}}\) is determined by solving

$$\begin{aligned} {\mathsf {A}}^{\varvec{p},\varvec{p}}\varvec{c}^{(\varvec{p})}=\varvec{f}^{\varvec{p}}. \end{aligned}$$

We are interested in computing the residual

$$\begin{aligned} \varvec{e}_{\varvec{p}} :=\varvec{f}_{\varvec{p}}-((\varvec{c}^{(\varvec{p})})^{\intercal }\varvec{k}^{\varvec{p}}({\mathcal {V}}))^{\intercal }, \end{aligned}$$
(5)

being \(\varvec{k}^{\varvec{p}}({\mathcal {V}})=(\kappa (\varvec{x}_i,\varvec{x}_j))_{i\in \varvec{p},j\notin \varvec{p}}\) an \((n-\rho )\times \rho \) matrix. Supposing \(n/\rho =\ell \in {\mathbb {N}}\), a classic approach would lead to the resolution of \(\ell \) different \((n-\rho )\times (n-\rho )\) linear systems, and thus, since usually \(n \gg \rho \), to a computational cost of about \({\mathcal {O}}(n^4)\). In case \(\rho =1\) one could use the Rippa’s algorithm reducing the complexity to \({\mathcal {O}}(n^3)\). In case of more folds, i.e. \(\rho >1\), we take advantage of the ERA scheme [15] for which (5) can be computed as

$$\begin{aligned} \varvec{e}_{\varvec{p}} = ({\mathsf {A}}^{-1}_{\varvec{p},\varvec{p}})^{-1}\varvec{c}_{\varvec{p}}. \end{aligned}$$
(6)

Doing in this way, we need to invert an \(n\times n\) matrix and then to solve \(\ell \) different \(\rho \times \rho \) linear systems. Therefore, the computational cost is about \({\mathcal {O}}(n^3)\), leading to a significant saving (cf. [15, §2.2]).

3.2 The ERBA-p

Here, we focus on the computation of the power function vector (see (3) and step 2b. of the algorithm presented in Sect. 2)

$$\begin{aligned} \varvec{P}^{{\mathcal {V}}}=\sqrt{\mathrm {diag}(\varvec{k}_{\varvec{p}}({\mathcal {V}})-\varvec{k}^{\varvec{p}}({\mathcal {V}})^{\intercal }({\mathsf {A}}^{\varvec{p},\varvec{p}})^{-1}\varvec{k}^{\varvec{p}}({\mathcal {V}}))}, \end{aligned}$$
(7)

where \(\mathrm {diag}(\cdot )\) is the diagonal operator. The vector \(\varvec{P}^{{\mathcal {V}}}\) is usually computed by means of a for-loop over the elements of \({\mathcal {V}}\) (see [1, Program 17.1]). Hence, supposing again \(n/\rho =\ell \in {\mathbb {N}}\), we need to invert \(\ell \) matrices of dimension \((n-\rho )\times (n-\rho )\), leading to a computational cost of about \({\mathcal {O}}(\ell (n-\rho )^3)={\mathcal {O}}(n(n-\rho )^3/\rho )\). Since \(n \gg \rho \), the cost for each step of the algorithm is approximately \({\mathcal {O}}(n^4)\).

To provide a fast computation of the power function vector \(\varvec{P}^{{\mathcal {V}}}\), we first recall that, letting \({\mathsf {R}} \in {\mathbb {R}}^{n_1\times n_2}\) and \({\mathsf {S}} \in {\mathbb {R}}^{n_2\times n_1}\), we have that

$$\begin{aligned} \mathrm {diag}({\mathsf {R}}{\mathsf {S}})=\mathrm {sum}({\mathsf {R}}\odot {\mathsf {S}}^{\intercal }), \end{aligned}$$
(8)

where \(\odot \) denotes the Hadamard (or pointwise) product. Then, we obtain the following result.

Theorem 1

The vector (7) can be computed as

$$\begin{aligned} \varvec{P}^{{\mathcal {V}}}=\sqrt{\mathrm {sum}((({\mathsf {A}}^{-1}_{\varvec{p},\varvec{p}})^{-1}{\mathsf {A}}^{-1}_{\varvec{p},:})\odot \varvec{k}({\mathcal {V}})^{\intercal })}, \end{aligned}$$

where \(\mathrm {sum}(\cdot )\) denotes the sum-by-rows operator.

Proof

Putting together (5) and (6), we have that

$$\begin{aligned} ({\mathsf {A}}^{-1}_{\varvec{p},\varvec{p}})^{-1}\varvec{c}_{\varvec{p}}&=\varvec{f}_{\varvec{p}}-\varvec{k}^{\varvec{p}}({\mathcal {V}})^{\intercal }({\mathsf {A}}^{\varvec{p},\varvec{p}})^{-1}\varvec{f}^{\varvec{p}},\\&\varvec{k}^{\varvec{p}}({\mathcal {V}})^{\intercal }({\mathsf {A}}^{\varvec{p},\varvec{p}})^{-1}\varvec{f}^{\varvec{p}}=\varvec{f}_{\varvec{p}}-({\mathsf {A}}^{-1}_{\varvec{p},\varvec{p}})^{-1}\varvec{c}_{\varvec{p}}. \end{aligned}$$

Since \(\varvec{f}\) is arbitrary if the function f is not fixed, we can substitute it with \(\varvec{k}^{\varvec{p}}({\mathcal {V}})\) in the equation. Then,

$$\begin{aligned} \varvec{k}^{\varvec{p}}({\mathcal {V}})^{\intercal }({\mathsf {A}}^{\varvec{p},\varvec{p}})^{-1}\varvec{k}^{\varvec{p}}({\mathcal {V}})=\varvec{k}_{\varvec{p}}({\mathcal {V}})-({\mathsf {A}}^{-1}_{\varvec{p},\varvec{p}})^{-1}{\mathsf {A}}^{-1}_{\varvec{p},:}\varvec{k}({\mathcal {V}}). \end{aligned}$$

Hence, recalling (7), we get

$$\begin{aligned} \varvec{P}^{{\mathcal {V}}}=\sqrt{\mathrm {diag}(({\mathsf {A}}^{-1}_{\varvec{p},\varvec{p}})^{-1}{\mathsf {A}}^{-1}_{\varvec{p},:}\varvec{k}({\mathcal {V}}))}, \end{aligned}$$

Finally, by applying (8) to \(({\mathsf {A}}^{-1}_{\varvec{p},\varvec{p}})^{-1}{\mathsf {A}}^{-1}_{\varvec{p},:}\) and \(\varvec{k}({\mathcal {V}})\), we conclude the proof. \(\square \)

By adopting the scheme proposed in Theorem 1, the computational cost at each step is about \({\mathcal {O}}(n^3)\), indeed we need to invert a unique \(n\times n\) matrix and to perform other minor calculations that are negligible as long as \(n \gg \rho \). The proposed strategy is therefore faster than the classical framework, as confirmed by the experiments carried out in the next section.

4 Numerics

In what follows, we perform some numerical experiments to prove the efficiency of the proposed ERBA. The tests have been carried out in Matlab on a Intel(R) Core(TM) i7-1165G7 [email protected] GHz processor. The software is available for the scientific community at

$$\begin{aligned} \texttt {https://github.com/cesc14/ERBA}\,. \end{aligned}$$

In Sect. 4.1 we compare our efficient implementation with the classical one. In doing so, we take both smooth and non-smooth target functions and different kernel bases. A focus on the CPU times is shown in Sect. 4.2, where we let both n and \(\rho \) vary. In Sect. 4.3, we analyze a stationary setting where the shape parameter varies according to the interpolation set. Finally, we present a test comparing the ERBA method and greedy schemes (which might be thought as forward RBA approaches).

4.1 ERBA versus RBA

Let \(\varOmega =[-1,1]^2\). In this subsection, we take the strictly positive definite kernel

and we set \(\varepsilon = 1\). As interpolation data set \({\mathcal {X}} \subset \varOmega \), we consider an \(n\times n\) grid, with \(n=25\). Moreover, we take an \(m\times m\) evaluation grid \(\varXi \) with \(m=60\).

4.1.1 Regular target function

Let \(f:\varOmega \longrightarrow {\mathbb {R}}\) be defined as

$$\begin{aligned} f(\varvec{x})=\frac{1}{1+(x_1-0.5)^2+(x_2+0.2)^2},\quad \varvec{x}=(x_1,x_2). \end{aligned}$$

The associated RMSE computed via the whole data set \({\mathcal {X}}\) and evaluated on \(\varXi \) is \(e_{{\mathcal {X}}}=9.69\mathrm{E}{-}05\). In Fig. 1, we plot the function f and the pointwise relative error between f and the interpolant \(S_{f,{\mathcal {X}}}\) evaluated on \(\varXi \).

Fig. 1
figure 1

(a) The function f. (b) The pointwise relative error between f and the interpolant constructed via the data set \({\mathcal {X}}\)

For testing the proposed strategy, we set \(\rho =3\) and the tolerance \(\tau _r = 2 {e}_{{\mathcal {X}}}\) for ERBA-r, while for ERBA-p we take \(\tau _p = 2\Vert \varvec{P}^{\varXi } \Vert _{2}/m\), being \(\varvec{P}^{\varXi }\) the power function vector constructed via the nodes \({\mathcal {X}}\) and evaluated on \(\varXi \). Denoting by \({\mathcal {X}}_s \subset {\mathcal {X}}\) the set of resulting reduced interpolation nodes, in Tables 1 and 2 we respectively compare the CPU times of ERBA-r and ERBA-p with their classical implementation. Moreover, the so-constructed reduced data sets, as well as the pointwise relative error between f and the interpolants constructed on \({\mathcal {X}}_s\), are depicted in Fig. 2.

The efficiency of our ERBA scheme is clearly deduced by Tables 1 and 2. Furthermore, as expected, we note that the ERBA-p tends to extract almost uniformly distributed nodes, while ERBA-r cluster the points where the target function has steep gradients. Such a behaviour will be amplified in the next experiment, which is carried out with a discontinuous target function.

Table 1 Results for the ERBA-r scheme with f
Table 2 Results for the ERBA-p scheme with f
Fig. 2
figure 2

Interpolation of f. (a) The reduced ERBA-r data set. (b) The reduced ERBA-p data set. The pointwise relative error between f and the interpolant constructed via the reduced ERBA-r (c) and ERBA-p (d) data set

4.1.2 Discontinuous target function

Let \(g:\varOmega \longrightarrow {\mathbb {R}}\) be defined as

$$\begin{aligned} g(\varvec{x})={\left\{ \begin{array}{ll} x_1+x_2-3 &{} \text {if}\, x_1>0,\\ x_1+x_2-2 &{} \text {if}\, x_1\le 0, \end{array}\right. }\quad \varvec{x}=(x_1,x_2). \end{aligned}$$

The associated RMSE computed on \(\varXi \) is \(e_{{\mathcal {X}}}=1.14\mathrm{E}{-}01\). In Fig. 3, we plot the function g and the pointwise relative error between g and the interpolant \(S_{g,{\mathcal {X}}}\) evaluated on \(\varXi \).

Fig. 3
figure 3

(a) The function g. (b) The pointwise relative error between g and the interpolant constructed via the data set \({\mathcal {X}}\)

Here, we keep \(\rho =3\) and we set the tolerance \(\tau _r = (3/2) {e}_{{\mathcal {X}}}\) for ERBA-r, while for ERBA-p we take \(\tau _p = (3/2)\Vert \varvec{P}^{\varXi } \Vert _{2}/m\), being \(\varvec{P}^{\varXi }\) the power function vector constructed via the nodes \({\mathcal {X}}\) and evaluated on \(\varXi \). As in the previous test, we report the CPU times of ERBA-r and ERBA-p (in Tables 3, 4) and the reduced data sets together with the pointwise relative errors in Fig. 4. From these results with the discontinuous target function, we recover the same patterns of the previous experiments, i.e. the efficiency of the ERBA methods and the clustering effect of the ERBA-r.

Table 3 Results for the ERBA-r scheme with g
Table 4 Results for the ERBA-p scheme with g
Fig. 4
figure 4

Interpolation of g. (a) The reduced ERBA-r data set. (b) The reduced ERBA-p data set. The pointwise relative error between g and the interpolant constructed via the reduced ERBA-r (c) and ERBA-p (d) data set

4.2 Varying n and \(\rho \)

This subsection is dedicated to test how the efficacy of our schemes depends on n and \(\rho \).

4.2.1 A focus on n

As a further experiment, always with the aim of comparing the complexity of ERBA and RBA, we consider the setting of Sect. 4.1 (with target function f) and we compare the CPU times of the algorithms by varying the grid size of the initial data set \({\mathcal {X}}\). Precisely, we take \(n=15+3k\), \(k=0,\ldots ,7\). The results are reported in Fig. 5. We note that the saving in terms of computational complexity is relevant for each choice of n.

Fig. 5
figure 5

The CPU times required by ERBA (dashed blue line) and RBA (solid black line) with f. (a) The residual based case. (b) The power based case

4.2.2 A focus on \(\rho \)

Here, always in the same setting of Sect. 4.1 (but with target function g), our purpose is to show how the CPU times and the number of reduced nodes provided by ERBA vary with the parameter \(\rho \). Thus, we take \(\rho =2k\), \(k=1,\ldots ,7\), and we display the results in Fig. 6. We observe that the number of reduced nodes is almost constant, especially as long as \(\rho \) is small. As a consequence, the CPU times decrease with \(\rho \), indeed less steps of the algorithm are performed.

Fig. 6
figure 6

ERBA-p (dashed blue line) and ERBA-r (solid black line) with g by varying \(\rho \). (a) CPU time. (b) Final number of nodes

4.3 Stationary ERBA

In this subsection, we investigate the role of the shape parameter. We take the kernel

$$\begin{aligned} \varphi (r)= e^{-(\varepsilon r)^2} ,\quad \text {Gaussian}\, C^{\infty }. \end{aligned}$$

As interpolation set, we consider \(n^2\) quasi-random Halton points [17] \({\mathcal {H}}\) in \(\varOmega \), with \(n=37\), while we keep \(\varXi \) as evaluation grid. Here, we focus on a stationary interpolation approach: fixed an initial \(\varepsilon _0>0\), at each iteration of our method, the shape parameter is set as \(\varepsilon =\varepsilon _0/h_{{\mathcal {H}}_{s-1},\varXi }\), where \(h_{{\mathcal {H}}_{s-1},\varXi }\) is the fill-distance at the s-th step of the algorithm, i.e.,

$$\begin{aligned} h_{{\mathcal {H}}_{s-1},\varXi }= \max _{ \varvec{\xi } \in \varXi } \left( \min _{ {\varvec{x}}_i \in {\mathcal {H}}_{s-1}} \left\| \varvec{\xi } - {\varvec{x}}_i \right\| _2 \right) . \end{aligned}$$

In the following, we fix \(\rho =7\) and we consider the target function f introduced in Sect. 4.1. In Fig. 7, we plot the pointwise relative errors between such function and the interpolant constructed on \({\mathcal {H}}\). The associated RMSE computed on \(\varXi \) is \(e_{{\mathcal {H}},f}=1.23\mathrm{E}{-}06\).

Fig. 7
figure 7

The pointwise relative error between f and the interpolant constructed via the data set \({\mathcal {H}}\)

Fig. 8
figure 8

Interpolation of f. (a) The reduced ERBA-r data set. (b) The reduced ERBA-p data set. The pointwise relative error between f and the interpolant constructed via the reduced ERBA-r (c) and ERBA-p (d) data set

We now fix \(\varepsilon _0=0.1\) and we set \(\tau _r = 1\mathrm{E}{+}04 r_{0}\) (for ERBA-r) and \(\tau _p = (6/5) r_{0}\) (for ERBA-p), being \(r_{0}\) the value produced at the first iteration of ERBA scheme as outlined in the algorithm described in Sect. 2. Then, the results of the stationary ERBA implementation are presented in Fig. 8. We point out that, for infinitely smooth kernels, such a stationary strategy becomes essential to partially avoid the typical ill-conditioning of the kernel matrices.

4.4 ERBA versus Greedy

As a final test, we carry out a comparison between our ERBA scheme and a greedy scheme that works in a specular manner with respect to RBA. For the sake of a clearer presentation, we detail it in the Appendix. We take the compactly-supported kernel

$$\begin{aligned} \varphi (r)= \max (0,1-\varepsilon r)^4(4\varepsilon r + 1) ,\quad \text {Wendland}\, C^{2}, \end{aligned}$$

with \(\varepsilon =0.1\). As interpolation data set \({\mathcal {Y}} \subset \varOmega \), we consider an \(n\times n\) grid, with \(n=40\), while we keep \(\varXi \) as evaluation grid. Moreover, we consider the target function \(h:\varOmega \longrightarrow {\mathbb {R}}\) defined as

$$\begin{aligned} h(\varvec{x})=\tan \bigg (\frac{x_1+x_2+3}{5}\bigg ),\quad \varvec{x}=(x_1,x_2). \end{aligned}$$

The associated RMSE computed on \(\varXi \) is \(e_{{\mathcal {Y}}}=3.94\mathrm{E}-06\). In Fig. 9, we plot the function h and the pointwise relative error between h and the interpolant \(S_{h,{\mathcal {Y}}}\) evaluated on \(\varXi \).

Fig. 9
figure 9

(a) The function h. (b) The pointwise relative error between h and the interpolant constructed via the data set \({\mathcal {Y}}\)

For both the ERBA and greedy algorithms we take \(\rho =5\). For ERBA-r and greedy-r, we set \(\tau _r= 1\mathrm{E}{-}06\), while we take \(\tau _p = 5\Vert \varvec{P}^{\varXi } \Vert _{2}/m\), being \(\varvec{P}^{\varXi }\) the power function vector constructed via the nodes \({\mathcal {Y}}\) and evaluated on \(\varXi \), for ERBA-p and greedy-p. The initial set \({\mathcal {X}}_0\) for the greedy schemes is a \(5\times 5\) equispaced grid in \(\varOmega \), which is nested in \({{\mathcal {X}}}\). In Figs. 10 and 11 we report the achieved results.

These experiments point out that in some cases, as when the number of nodes to remove is not too large, the ERBA is faster than the greedy approach. Indeed, the cost of greedy schemes increase when the number of the iterations grows.

Fig. 10
figure 10

Interpolation of h. (a) The reduced ERBA-r data set. (b) The reduced greedy-r data set. The pointwise relative error between g and the interpolant constructed via the reduced ERBA-r (c) and greedy-r (d) data set

Fig. 11
figure 11

Interpolation of h. (a) The reduced ERBA-p data set. (b) The reduced greedy-p data set. The pointwise relative error between h and the interpolant constructed via the reduced ERBA-p (c) and greedy-p (d) data set

5 Discussion and Conclusions

We have investigated a fast computation of knot removal approaches. We have implemented two different strategies: the first one in based on classical residual-based schemes while the second one relies on power function error bounds. The latter is independent of the function values and tends to return quasi-uniform data (cf. [18] and [11, Theorem 15 & 19 & 20]), while the former, as expected, keeps more points where the corresponding function has steep gradients. In both cases we are able to significantly speed up the algorithm thanks to our implementation.

Work in progress consists in extending the proposed tool to other bases, e.g. splines [19] and to variably scaled kernels [20, 21]. Moreover, it can be helpful for validating the shape parameter in RBF interpolation; refer e.g. to [22].