
Neural operator is a newly proposed neural network ar-
chitecture in the field of computational physics [17, 19, 23]
for numerically efficient solvers of partial differential equa-
tions (PDE). Stemming from the operator theory, nueral op-
erators learn mappings between infinite-dimensional func-
tion spaces, which is inherently capable of continuous func-
tion evaluations and has shown promising potentials in var-
ious applications [9,13,27]. Typically, neural operator con-
sists of three components: 1) lifting, 2) iterative kernel in-
tegral, and 3) projection. The kernel integrals operate in the
spatial domain, and thus can explicitly capture the global
relationship constraining the underlying solution function
of the PDE. The attention mechanism in transformers [36]
is a special case of kernel integral where linear transforms
are first exerted to the feature maps prior to the inner prod-
uct operations [17]. Tremendous successes of transformers
in various tasks [6, 22, 34] have shown the importance of
capturing global correlations, and this is also true for SR to
improve performance. [8].
In this paper, we propose the super-resolution neural op-
erator (SRNO), a deep operator learning framework that
can resolve HR images from their LR counterparts at ar-
bitrary scales. As shown in Fig.1, SRNO learns the map-
ping between the corresponding function spaces by treating
the LR-HR image pairs as continuous functions approxi-
mated with different grid sizes. The key characteristics dis-
tinguishing SRNO from prior continuous SR works are: 1)
the kernel integral in each layer is efficiently implemented
via the Galerkin-type attention, which possesses non-local
properties in the spatial dimensions and have been proved
to be comparable to a Petrov-Galerkin projection [3]; and
2) the multilayer attention architecture allows for the dy-
namic latent basis update, which is crucial for SR problems
to “hallucinate” high-frequency information from the LR
image. When employing same encoders to capture features,
our method outperforms previous continuous SR methods
in terms of both reconstruction accuracy and running time.
In summary, our main contributions are as follows:
• We propose the methodology of super-resolution neu-
ral operator that maps between finite-dimensional
function spaces, allowing for continuous and zero-shot
super-resolution irrespective the discretization used on
the input and output spaces.
• We develop an architecture for SRNO that first ex-
plores the common latent basis for the whole training
set and subsequently refines an instance-specific basis
by the Galerkin-type attention mechanism.
• Numerically, we show that the proposed SRNO outper-
forms existing continuous SR methods with less run-
ning time, and even generates better results on the res-
olutions for which the fixed scale SR networks were
trained.
2. Related Work
Deep learning based SR methods. [4, 20, 21, 24, 29,
40, 41] have achieved impressive performances, in multi-
scale scenarios one has to train and store several models
for each scale factor, which is unfeasible when consider-
ing time and memory budgets. In recent years, several
methods [11, 31, 37] are proposed to achieve arbitrary-scale
SR with a single model, but their performances are limited
when dealing with out-of-distribution scaling factors. In-
spired by INF, LIIF [5] takes continuous coordinates and
latent variables as inputs, and employs an MLP to achieve
outstanding performances for both in-distribution and out-
of-distribution factors. In contrast, LTE [18] transforms in-
put coordinates into the Fourier domain and uses the dom-
inant frequencies extracted from latent variables to address
the spectral bias problem [30, 32]. In a nutshell, treat-
ing images as RGB-valued functions and sharing the im-
plicit function space are the keys to the success of LIIF-like
works [5, 18]. Nevertheless, a purely local decoder, like
MLP, is not able to accurately approximate arbitrary im-
ages, although it is rather sensitive to the input coordinates.
Neural Operators. Recently, a novel neural net-
work architecture, Neural Operator (NO), was proposed
for discretization invariant solutions of PDEs via infinite-
dimensional operator learning [10,17,19,23]. Neural opera-
tors only need to be trained once and are capable of transfer-
ring solutions between differently discretized meshes while
keeping a fixed approximation error. A valuable merit of
NO is that it does not require knowledge of the underlying
PDE, which allows us to introduce it by the following ab-
stract form,
(L
a
u)(x) = f(x), x ∈ D,
u(x) = 0, x ∈ ∂D,
(1)
where u : D → R
d
u
is the solution function residing in the
Banach space U, and L : A → L(U; U
∗
) is an operator-
valued functional that maps the coefficient function a ∈ A
of the PDE to f ∈ U
∗
, the dual space of U. As in many
cases the inverse operator of L even does not exist, NO
seeks a feasible operator G : A → U, a 7→ u, directly
mapping the coefficient to the solution within an acceptable
tolerance.
The operator G is numerically approximated by training
a neural network G
θ
: A → U, where θ are the train-
able parameters. Suppose we have N pairs of observa-
tions {a
j
, u
j
}
N
j=1
where the input functions a
j
are sam-
pled from probability measure µ compactly supported on
A, and u
j
= G(a
j
) are used as the supervisory output func-
tions. The infinitely dimensional operator learning problem
G ← G
θ
thus is associated with the empirical-risk mini-
mization problem [35]. In practice, we actually measure
the approximation loss using the sampled observations u
(j)
o
2
评论