ArticlePDF Available

Nonlocal Operators with Applications to Image Processing

Authors:

Abstract

We propose the use of nonlocal operators to define new types of flows and functionals for image processing and elsewhere. A main advantage over classical PDE-based algorithms is the ability to handle better textures and repetitive structures. This topic can be viewed as an extension of spectral graph theory and the diffusion geometry framework to functional analysis and PDE-like evolutions. Some possible applications and numerical examples are given, as is a general framework for approximating Hamilton-Jacobi equations on arbitrary grids in high demensions, e.g., for control theory.
NONLOCAL OPERATORS WITH APPLICATIONS TO IMAGE
PROCESSING
GUY GILBOA AND STANLEY OSHER
Abstract. We propose the use of nonlocal operators to define new types of flows and functionals
for image processing and elsewhere. A main advantage over classical PDE-based algorithms is the
ability to handle better textures and repetitive structures. This topic can be viewed as an extension
of spectral graph theory and the diffusion geometry framework to functional analysis and PDE-like
evolutions. Some possible application and numerical examples are given, as is a general framework
for approximating Hamilton-Jacobi equations on arbitrary grids in high demensions, e.g., for control
theory.
Key words. Nonlocal operators, regularization, total variation, variational methods, spectral
graph theory, Hamilton-Jacobi equations.
AMS subject classifications. 35A15, 68U10, 70H20, 65D25, 35S05, 68R10
1. Introduction.
1.1. Motivation. In this paper our goal is to formalize a systematic and coher-
ent framework for nonlocal image and signal processing. By this we mean that any
point can interact directly with any other point in the image domain (at least in prin-
ciple). In practice, for complexity reasons, the number of interactions is limited only
to the “most relevant” regions (in some sense which is derived from the application).
Our formulation is continuous.
We attempt to extend some known PDE’s and variational techniques to this non-
local framework. The major difference is that classical derivatives are local operators.
However, following ideas from graph theory, and specifically the gradient and diver-
gence operators of Zhou and Sch¨olkopf [58, 59], we observe that many PDE-based
processes, minimizations and computation methods can be generalized to be nonlo-
cal. A main advantage for image processing is the ability to process both structures
(geometrical parts) and textures within the same framework.
We also believe this framework may be useful beyond the scope of image process-
ing, for purposes such as physical modelling of processes with nonlocal behavior. We
outline a method for approximating Hamilton-Jacobi equations in high dimensions in
Section 3, below.
1.2. Short Background. PDE’s have been used very successfully for many
image processing tasks, such as denoising, deconvolution, segmentation, inpainting,
optical-flow and more. For details regarding the theory and the applications see
[1, 43, 42, 17, 50] and the references therein.
Techniques using spectral graph theory [19, 38] were used for image segmentation
[51, 46, 56, 29] and in a more general form for various machine-learning applications
in the diffusion geometry framework [20, 40]. These techniques are based on manipu-
lation of the eigenvalues of the graph-Laplacian. Total variation type regularizations
on graphs were first proposed in [15] and later by [59] and [7]. A related framework
Department of Mathematics, UCLA, Los Angeles, California 90095, tel. (310)-8254952 fax.
(310)-2066673 gilboa@math.ucla.edu.
Department of Mathematics, UCLA, Los Angeles, California 90095, tel. (310)-8251758 fax.
(310)-2066673 sjo@math.ucla.edu. Both authors are supported by grants from the NSF under
contracts ITR ACI-0321917, DMS-0312222, and the NIH under contract P20 MH65166. G.G. is also
supported by NSF DMS-0714087.
1
2G. GILBOA AND S. OSHER
in the context of PDE’s is the Beltrami flow on Riemannian manifolds [52, 31] where
the metric is image-driven and textures can be handled [49]. This framework however
is still local and is based on PDE’s in a classical sense.
For image denoising, nonlocal methods were developed based on gray-level pixel
affinities in the form of the Yoroslavsky filter [57] and the bilateral filters [54]. Deeper
understanding of these filters and their relation to PDE’s were given by Barash and
Elad [5, 6, 26]. Nonlocal denoising based on patch-distances was proposed by Buades
et al in [9]. They have also given in [10] the asymptotic relation of neighborhood
filters to Perona-Malik type PDE’s [47]. The use of patch distances in [9] followed
ideas by Efros and Leung [25] for texture synthesis and completion. We will give
a variational interpretation of this process in this paper. In [53] the filter of [9],
referred to as nonlocal means, was understood as a special case within the diffusion
geometry framework. Other patch distances based on filters were proposed. In [36]
a fast algorithm was designed for computing the fully nonlocal version. The study of
[30] presented a statistical analysis of the problem and suggested to use an adaptive
window approach which minimizes a local risk measure.
The DUDE algorithm [39] denoises data sequences generated by a discrete source
and received over a discrete memoryless channel. DUDE assigns image values using
similarity of neighborhoods based on image statistics. This resembles the construc-
tion of conditional probabilities in Awate and Whitaker [4]. The DUDE approach
is limited to discrete-valued signals as opposed to [4] and our approach, which ad-
dresses continuous- valued signals, such as those associated with grayscale images.
The DUDE algorithm is not very effective in case of additive noise.
Awate and Whitaker’s algorithm [4] can be expressed in our framework (without
our PDE/regularization steps). They use the entropy as a measure of self similarity
and obtain a convolution with weights requiring a convolution. They update their
weights as they proceed in time using their gradient descent approach. Their method
somewhat resembles the approach in [32], Section 3, where g, the function within the
regularizer, involves entropy.
A first variational understanding as a nonconvex minimization was given in [32].
In [28] we proposed an alternative convex quadratic functional, showed the relation to
spectral graph theory, and were able to achieve superior filtering properties, compared
with [9], using an iterative “nonlocal diffusion” process. We also presented a simple
nonlocal supervised segmentation algorithm which follows [51, 35, 29] and analyzed
analytically the step-edge case. In [27] a more general convex framework was proposed
and a method to compute the energy minimizations using graph-cut techniques was
shown. This paper follows and significantly generalizes our previous studies [28, 27].
2. The proposed mathematical framework.
2.1. Basic operators. In the following we use a variant of the gradient and
divergence definitions on graphs given in the context of machine learning [58, 59]. In
our case, the weights are not normalized pointwise and the definitions are continuous.
Recently Bougleux et al. [7] have proposed a regularization framework on graphs
which also uses similar operators. In their study, a family of pLaplace operators
was defined for discrete data and a variational framework was proposed for image and
mesh denoising.
Let Ω Rn,xΩ, u(x) a real function u: Ω R. We extend the notion of
NONLOCAL OPERATORS WITH APPLICATIONS TO IMAGE PROCESSING 3
derivatives to a nonlocal framework by the following definition:
yu(x) := u(y)u(x)
˜
d(x, y), y, x ,
where 0 <˜
d(x, y)≤ ∞ is a positive measure defined between points xand y. To keep
with standard notations related to graphs we define the weights as
w(x, y) = ˜
d2(x, y).
Thus 0 w(x, y)<. In this paper we assume the weights are symmetric, that is
w(x, y) = w(y, x). The nonlocal derivative can be written as
yu(x) := (u(y)u(x))pw(x, y).(2.1)
The nonlocal gradient wu(x) : Ω ×Ω is defined as the vector of all partial
derivatives:
(wu)(x, y) := (u(y)u(x))pw(x, y), x, y .(2.2)
We denote vectors as ~v =v(x, y)×Ω. The standard L2inner product is used
for functions
hu1, u2i:= Z
u1(x)u2(x)dx.
For vectors we define a dot product
(~v1·~v2)(x) := Z
v1(x, y)v2(x, y)dy,
and an inner product
h~v1, ~v2i:= h~v1·~v2,1i=Z×
v1(x, y)v2(x, y)dxdy.
The magnitude of a vector is
|~v|(x) := p~v1·~v1=sZ
v(x, y)2dy.
With the above inner products the nonlocal divergence div w~v(x) : ×Ω is
defined as the adjoint of the nonlocal gradient:
(div w~v)(x) := Z
(v(x, y)v(y, x))pw(x, y)dy. (2.3)
The Laplacian can now be defined by:
wu(x) := 1
2div w(wu(x)) = Z
(u(y)u(x))w(x, y)dy. (2.4)
Note that in order to get the standard Laplacian definition which relates to the graph
Laplacian we need a factor of 1/2.
4G. GILBOA AND S. OSHER
2.2. Some properties. Most of the properties involving a double integral can
be shown by expanding an integral of the form R×f(x, y)dxdy to 1
2R×(f(x, y) +
f(y, x))dxdy, changing the order of integration and using the fact that w(x, y) =
w(y, x). We give an example showing the adjoint relation
h∇wu, ~vi=hu, div w~vi,(2.5)
h∇wu, ~vi=R×(u(y)u(x))pw(x, y)v(x, y)dxdy
=1
2R×h(u(y)u(x))pw(x, y)v(x, y) + (u(x)u(y))pw(y, x)v(y, x)idxdy
=1
2R×[u(y)(v(x, y)v(y, x)) u(x)(v(x, y)v(y, x))] pw(x, y)dxdy
=1
2R×[u(x)(v(y, x)v(x, y)) u(x)(v(x, y)v(y, x))] pw(x, y)dxdy
=Ru(x)R(v(x, y)v(y, x))pw(x, y)dydx.
“Divergence theorem”:
Z
div w~vdx = 0.(2.6)
The Laplacian is self adjoint
hwu, ui=hu, wui(2.7)
and negative semidefinite
hwu, ui=−h∇wu, wui ≤ 0.(2.8)
We can also formulate a nonlocal (mean) curvature:
κw:= div wwu
|∇wu|
=R(u(y)u(x))w(x, y)1
|∇wu|(x)+1
|∇wu|(y)dy, (2.9)
where
|∇wu|(q) := sZ
(u(z)u(q))2w(q, z)dz.
2.3. The Regularizing Functionals. Below we propose two types of regular-
izing nonlocal functionals. The first type is based on the nonlocal gradient. It is
set within the mathematical framework described above. The second type is based
on differences, it appears to be easier to implement, where the minimization can be
accomplished using graph cut techniques, as will be discussed in Section 5. We still
investigate the relations between these functionals and when each of them is preferred.
The gradient-based functional is
J(u) = Rφ(|∇wu|2)dx,
=Rφ(R(u(y)u(x))2w(x, y)dy)dx, (2.10)
where φ(s) is a positive function, convex in swith φ(0) = 0.
NONLOCAL OPERATORS WITH APPLICATIONS TO IMAGE PROCESSING 5
The difference-based functional is
Ja(u) = Z×
φ((u(y)u(x))2w(x, y))dydx. (2.11)
The variation with respect to u(Euler-Lagrange) of (2.10) is
uJ(u) = 2Z
(u(y)u(x))w(x, y)(φ0(|∇wu|2(x)) + φ0(|∇wu|2(y)))dy, (2.12)
where φ0(s) is the derivative of φwith respect to s. This can be written more concisely
as
uJ(u) = 2div ww0(|∇wu|2(x)).
The variation with respect to uof (2.11) is
uJa(u) = 4Z
(u(y)u(x))w(x, y)φ0((u(y)u(x))2w(x, y))dy. (2.13)
Note that for the quadratic case φ(s) = sthe functionals (2.10) and (2.11) coincide
(and naturally so do Eqs. (2.12) and (2.13)).
2.3.1. Relation to isotropic and anisotropic local functionals. The func-
tionals which can be written in the form of Eq. (2.10) correspond in the local case to
isotropic functionals (which have no preferred directionality). The second category,
Eq. (2.11), can be related to anisotropic functionals in the local case. We suggest
later two different methods for efficiently computing each category.
As an example, for total-variation, φ(s) = s, Eq. (2.10) becomes:
JNLT V (u) = Z|∇wu|dx =ZsZ
(u(y)u(x))2w(x, y)dydx (2.14)
whereas Eq. (2.11) becomes
JNLT V a (u) = Z×|u(x)u(y)|pw(x, y)dydx (2.15)
The above functionals correspond in the local two dimensional case to the isotropic
TV
JT V (u) = Z|∇u|dx =Zqu2
x1+u2
x2dx
and to the anisotropic TV
JT V a (u) = Z
(|ux1|+|ux2|)dx.
Following the discussion in Section 3, another analogue to anisotropic TV is:
Z
2
X
i=1 |Z
(u(y)u(x))w(x, y)(yixi)dy|dx.
6G. GILBOA AND S. OSHER
3. Computing Hamilton-Jacobi Equations on Arbitrary Grids in High
Dimensions. In this section we show how this general framework can be used for
computational purposes. Our ultimate goal here is to solve partial differential equa-
tions approximately in high dimensions on irregular grids. We assume to be operating
on a set of isolated data points ΩdRn, with nlarge and Ωdsparse. Note that unlike
the nonlocal models of the next section, here the construction of the weights is based
on different considerations and is not image or signal driven. The calculus however
is similar. A detailed study on the computational aspects with examples will appear
elsewhere.
Using the nonlocal gradient defined in (2.2) enables us to obtain partial derivatives
as follows. We wish to compute partial derivatives of u, i.e.
∂u
∂xiw
, i = 1,...,n
in a consistent way.
Let us first define an approximation of the unit vector in the xidirection as
follows:
xi)w:= w(xi).(3.1)
The corresponding partial derivative estimation is therefore:
∂u
∂xiw
:= w(u)·xi)w=w(u)· ∇w(xi)
=R(u(y)u(x))w(x, y)(yixi)dy,
(3.2)
we remind that dy is shorthand for dy1...dyn. Note that this can be generalized to
any order, e.g. second order derivatives can be estimated by
2u
∂x2
iw
:= ww(u)·xi)w·xi)w.
We construct wsuch that the unit vectors are orthonormal:
xi)w·(ˆxj)w=δij ,
that is:
Z(yjxj)w(x, y)(yixi)dy =δij = 1 if i=j(3.3)
= 0 if i6=j.
A similar framework for approximating partial derivatives for the purpose of strain
localization was found by Chen, Zhang and Belytschko in [18]. Our construction of
monotone schemes for Hamilton-Jacobi equations in high dimensions is new.
A simple and important class is
w(x, y) = w(|xy|) = w(r) (3.4)
normalized so that
Z
r2w(r)dx =n(3.5)
NONLOCAL OPERATORS WITH APPLICATIONS TO IMAGE PROCESSING 7
e.g., if n= 2, we need
Zr3w(r)dr =1
π.(3.6)
We can take w(r) = c/r2,0rR, c > 0 = 0 R < r with R=q2
πc .
Another possibility is
w(Ω) = Cer2, C, σ > 0.
For (3.5) we need σ=q2
πC , e.g., for n= 2.
We compute
∂u
∂xiw
=Z(u(y)u(x))w(x, y)(yixi)dy (3.7)
=uxi(x) + 1
2
n
X
j,k=1 Zuxj,xkw(x, y)(yjxj)(ykxk)(yixi)dy
+···
=uxi(x) + error.
In future work we will estimate the error term and develop a theory for solving
Hamilton-Jacobi equations in high dimensions using relatively few data points. Such
problems arise in control theory and elsewhere. Radial basis functions were used in
[11] to obtain schemes in up to four dimensions. Our present approach seems to be
more flexible. We outline it below. See [45], [44] for classical approaches.
In our framework: We are interested in solving
ut+H(ux1,...,uxn) = 0 (3.8)
u(x, 0) = ϕ(x).
We are interested in finding the unique viscosity solution [21]. We approximate
this by discretization in time, for x
um+1(x)um(x)
t=˜
Hum
∂x1w
,∂um
∂x2w
,... ∂um
∂xnw(3.9)
u0(x) = ϕ(x)
where ˜
His the numerical Hamilton which is consistent with H(definitions will be
given in a future paper)
um(x)u(x, mt).
A scheme is monotone if um+1(x) is a nondecreasing function of the values um(x).
We will take an analogue of the Lax-Friedrichs scheme [45]
um+1(x) = u(x)tH u
∂x1w
,...,∂u
∂xnw(3.10)
+ 2∆tZc(x, y)w(x, y)(u(y)u(x))dy
8G. GILBOA AND S. OSHER
(dropping the superscript m), where c(x, y ) is a nonnegative smooth function, chosen
so that (3.10) gives us a consistent, monotone approximation to the Hamilton-Jacobi
equation, (3.8) (for precise definitions, see [45]).
For this to be monotone, we first require
2c(x, y)w(x, y)X
ν|Hν|w(x, y)|yνxν|>0.(3.11)
So we can take on the support of w(x, y):
2c(x, y)>X
ν|Hν||yνxν|(3.12)
and for consistency c(x, x +h)0 as h0.
Also, we have a time step restriction:
1 + ∆tZ X
ν
Hνw(x, y)(ykxv)2c(x, y)w(x, y)!dy 0 (3.13)
1tZ(w(x, y)2c(x, y)X
ν
Hνw(x, y)(yνxν))dy
so we can take:
14∆tw(x, y)c(x, y).
Just to illustrate how this becomes rather conventional in a simple case, let n= 1
and
w(x, y) = 1
2h2(δ(xyh) + δ(xy+h)), δ the Dirac delta function.(3.14)
Then
∂u
∂x w
=u(x+h)u(xh)
2h
(unsurprisingly) and
um+1(x) = u(x)tH u(x+h)u(xh)
2h(3.15)
+ ∆t(c(x, x +h)(u(x+h)u(x)) c(x, x h)(u(x)u(xh))
h2.
If we take c(x, y) = K|yx|for K > 0 large enough to satisfy (3.12), we have the
conventional Lax-Friedrichs scheme, which is known to converge as h0, if (3.13) is
satisfied.
4. Basic Models. Our proposed nonlocal models are based on the general func-
tional (2.10). The quadratic case, φ(s) = s(resulting in a linear steepest descent), was
investigated in [28] where applications for denoising and segmentation were shown.
Here we will focus on functionals with a TV-type regularizer.
We are interested in the minimizations of the following functionals:
Nonlocal ROF:
JNLT V (u) + λkfuk2
L2,(4.1)
NONLOCAL OPERATORS WITH APPLICATIONS TO IMAGE PROCESSING 9
where JNLT V (u) is defined in (2.14), fis the noisy input image or signal, and
the minimization is over u. We are also interested in the inpainting version of this
functional, following the local TV-inpainting model of [16]:
JNLT V (u) + Z
λ(x)(fu)2dx, (4.2)
with λ(x) = 0 in the inpainting region and λ(x) = cin the rest of the image.
Another very important model following [41, 14] is the extension of T V L1to
a nonlocal version:
JNLT V (u) + λkfukL1.(4.3)
We will show later an interesting application of texture regularization using this mini-
mization. It can both detect and remove anomalies or irregularities from images, and
specifically textures.
We can further generalize Meyer’s G-norm [37] to a nonlocal setting as described
below.
4.1. Generalizing Meyer’s G-norm. Let us define the nonlocal Gspace (the
dual space of nonlocal TV). In the local case this space was considered by Meyer as the
natural space of oscillatory patterns [37]. In our case oscillatory patterns which are
regular and repetitive can be included in the nonlocal TV space, if a proper method
for calculating the weights is used (as seen in our numerical examples). Thus we
anticipate that in this case the nonlocal Gspace will characterize irregularities and
randomness of the signal (and also noise). We have not yet investigated this topic
thoroughly.
Let us define the nonlocal Gspace by
G={vX / gYsuch that v= div w(g)}.(4.4)
The nonlocal Gnorm (if vG) is:
kvkNLG= inf {kgk/ v = div w(g)}(4.5)
where kgk:= supx{|g|(x)}.
We can thus choose to minimize the following alternative to NL-ROF (4.1). We
shall refer to it as NL TV-G:
JNLT V (u) + λkfukN LG.(4.6)
In the experimental section, some examples are given, showing the qualitative char-
acteristics of this regularization.
4.2. Computing the weights. In our examples below we have a simplified
scheme to compute the weights using only binary values (0 or 1) based on smallest
patches distances.
Let us define the patch distance as in [9]:
da(f(x), f (y)) = Z
Ga(t)|f(x+t)f(y+t)|2dt,
where Gais a Gaussian of standard deviation a.
10 G. GILBOA AND S. OSHER
For each point we define the following set Aof area |A|=γ(a parameter) within
a search neighborhood S(x) around x(where AS(x),|A|  |S(x)|):
A(x) := arg min
AZA
da(f(x), f (y))dy, s.t. A ⊂ S(x),|A| =γ.(4.7)
Then the weights are computed as:
w(x, y) = 1,if yA(x) or xA(y)
0,otherwise.(4.8)
This naturally gives the property of symmetric weights w(x, y) = w(y, x). For the
way to discretize the weights see Section 5.1 below. Note that in the following section
we keep with the general case of real-valued non-negative weights, and do not assume
that the values are binary.
5. Computation.
5.1. Basic Discretization. Let uidenote the value of a pixel iin the image
(1 iN), wi,j is the sparsely discrete version of w(x, y). We use the neighbors set
notation j∈ Nidefined as j∈ Ni:= {j:wi,j >0}.
Let wd be the discretization of w:
wd(ui) := (ujui)wi,j , j ∈ Ni(5.1)
Let div wd be the discretization of div w:
div wd(pi,j ) := X
j∈Ni
(pi,j pj,i)wi,j .(5.2)
The discrete inner product for functions is < u, v >:= Pi(uivi) and for vectors we
have the discretized dot product (p·q)i:= Pj(pi,j qi,j ) and inner product < p, q >:=
PiPj(pi,j qi,j ). The vector magnitude is therefore |p|i:= qPj(pi,j)2.
Binary weights:. We use binary weight values of 0 or 1. This way rare features
which also have a very large “patch distance” between them and any other patch in
the image can be regularized as well. In the more common case where the weights are
computed with a Gaussian-like formula, e.g. as in [9],[28],[53],[30], the weights decay
fast for distances above a certain threshold (usually related to the noise variance).
This results in very weak connections (low weight values) for singular regions, thus
such regions are essentially isolated from the rest of the image. This may be a good
property in the case of denoising which avoids blurring of singular patches. However
it is not adequate for the applications presented here, where the purpose is to re-
move irregularities. Note also that with binary weights the “manifold”, as defined by
the values u(x) and the “metric” w(x, y), is not necessarily smooth and can contain
discontinuities or edges (which are handled well by the nonlocal TV regularizer).
The weights are descretized as follows: we take a patch around a pixel i, compute
the distances (da)i,j (a discretization of da(x, y)) to all the patches in the search
window and select the kclosest (with the lowest distance value). The number of
neighbors kis an integer proportional to the area γ. For each selected neighbor jwe
assign the value 1 to wi,j and to wj,i. A maximum of up to m= 2kneighbors for
each pixel is allowed in our implementation. In the examples of Figs. 6.4 - 6.8 we
used 5 ×5 pixel patches, a search window of size 21 ×21 and m= 10.
NONLOCAL OPERATORS WITH APPLICATIONS TO IMAGE PROCESSING 11
5.2. Steepest Descent. In this convex framework, one can resort as usual to a
steepest descent method for computing the solutions. One initializes uat t= 0, e.g.
with the input image: u|t=0 =f, and evolves numerically the flow:
ut=uJduHd(f, u),
where uJdis the discretized version of Eq. (2.12) or Eq. (2.13) and Hd(f , u) is
the discretized fidelity term functional. As in the local case, here also one should
introduce a regularized version of the total variation: φ(s) = s+2(where sis the
square gradient magnitude). Thus the E-L equations are well defined, also for a zero
gradient. When the L1norm is used as fidelity, a similar regularization is needed for
that term also. The time-step restriction (CFL) is proportional to the regularizing
and thus convergence is slow.
5.3. Graph-Cuts. For the difference-based (“anisotropic”) functional, Eq. (2.11),
we can generalize known fast algorithms which use graph-cuts techniques [8, 34]. A
generalization of the algorithm of Darbon and Sigelle [23, 22] can be seen in our CAM
report with Darbon and Chan [27].
5.4. Projections. For the gradient-based (“isotropic”) case graph-cuts tech-
niques cannot be generalized in a straightforward manner. They are restricted to
pairwise interactions between nodes of the graph. When minimizing expressions in-
volving the nonlocal gradient, however, this restriction is not met and one has to
resort to an alternative method.
The projection algorithm of Chambolle [12] generalizes easily in this case. We
show below how to compute the nonlocal ROF and a good approximation of nonlocal
T V L1.
5.4.1. Nonlocal ROF. Chambolle’s projection algorithm [12] for solving ROF
[48] can be extended to solve nonlocal ROF.
A minimizer for the discrete version of Eq. (4.1) can be computed by the following
iterations (fixed point method):
pn+1
i,j =pn
i,j +τ(wd(div wd(pn)2λf ))i,j
1 + τ|(wd(div w d(pn)2λf ))i,j|(5.3)
where p0= 0, and the operators wd and div wd are defined in (5.1) and (5.2),
repectively. The solution is u=f1
2λdiv wd(p).
Theorem 5.1. The algorithm converges to the global minimizer as n→ ∞ for
any 0< τ 1
kdiv wdk2
L2
.
Proof. The proof follows the same lines as the original proof of [12]. One should
replace the numerical gradient, divergence and dot product defined in [12] by the
nonlocal discrete definitions given here (Equations (5.2), (5.1) and the definitions
which follow). Then everything follows in a straightforward manner: Obviously, the
nonlocal TV, Eq. (2.14), is one-homogeneous, that is JNLT V (λu) = λJN LT V (u).
Thus we have a similar “characteristic function” structure of the Legendre-Fenchel
transform (J
NLT V := supu< u, v > J(u)). Solving the projection ,we reach
a similar Euler-Lagrange equation for the constrained problem, resolve the value of
the Lagrange multiplier using the same arguments and reach the above fixed point
iterations. The bound on τ(Th. 3.1 in [12]) follows through in the same manner,
having transformed the operators to their nonlocal counterparts. The only difference
is that in our case kdiv wd k2
L2is not resolved (with the definitions of [12] it is shown
that kdiv k2
L28).
12 G. GILBOA AND S. OSHER
A bound on τ.The bound on τdepends on the operator norm kdiv wd k2which is
a function of the weights wi,j . As the weights are image dependent, so is kdiv wd k2.
We propose below a simple bound which is very straightforward and does not depend
on the image. We assume that the maximal number of neighbors for each pixel is a
fixed parameter (not image dependent) and that the weights are bounded by some
value, typically 1.
Proposition 5.2. Let mbe the maximal number of neighbors of a pixel, m:=
maxi{Pj(sign(wi,j ))}. If the weights are in the range 0wi,j 1i, j, then for
0< τ 1
4mthe algorithm converges.
Proof. We need to show that kdiv wdk24m:
kdiv wd(p)k2=PiPj(pi,j pj,i)wi,j 2
2PiPj(p2
i,j +p2
j,i)Pjwi,j
4 maxiPjwi,j PiPjp2
i,j
4mkpk2.
Remark: . Note that in [12] the discrete local gradient and divergence operators
are not symmetric, thus they do not fall precisely to the framework of this paper. Yet,
the divergence operator of [12] can be viewed as div wd with nonsymmetric weights of
unit value where m= 2. In this sense the original bound kdiv k28 can be viewed
as a special case of the bound presented above.
5.4.2. Nonlocal TV-L1. To solve (4.3) we generalize the algorithm of [3]. We
consider the problem:
inf
u,v JNLT V (u) + 1
2αkfuvk2
L2+λkvkL1(5.4)
The parameter αis small so that we almost have f=u+v, thus (5.4) is a very good
approximation of (4.3). We can solve the discretized version of (5.4) by iterating:
vbeing fixed (we have a nonlocal ROF problem), find uusing the nonlocal
Chambolle’s projection algorithm:
inf
uJNLT V (u) + 1
2αkfuvk2
L2
ubeing fixed, find vwhich satisfies:
inf
v
1
2αkfuvk2
L2+λkvkL1.
The solution for vis given by soft-thresholding fuwith αλ as the threshold [13],
denoted by STαλ (fu), where
STβ(q) :=
qβ, q > β
0,|q| ≤ β
q+β, q < β.
(5.5)
Proposition 5.3. The algorithm converges to the global minimizer as n→ ∞
for any 0< τ 1
kdiv wdk2
L2
.
NONLOCAL OPERATORS WITH APPLICATIONS TO IMAGE PROCESSING 13
Proof. The proof is similar to the one of [3] (which is in the spirit of [2]). Using
similar arguments one can show that having solved the nonlocal ROF problem the it-
erative process presented here converges to the global minimizer. For the convergence
of the NL-ROF part we use Theorem 5.1 above.
5.4.3. Nonlocal TV-G. One can repeat the same arguments for minimizing
nonlocal T V G, Eq. (4.6), and modify the iterative projection algorithm of [2]
to be nonlocal, in a similar manner as the algorithms presented above. We briefly
summarize the idea: Let Xbe the Euclidean space and let Gµbe defined as
Gµ={vNL G/kvkN LGµ}.
We consider the following problem:
inf
(u,v)X×GµJN LT V (u) + 1
2αkfuvk2
L2.(5.6)
To find the minimizer of the discretized functional one can apply the following
simple iterations of NL-ROF minimizations. Initialize u=f , v = 0. Iterate until
convergence:
vbeing fixed (one needs to solve a nonlocal ROF problem), find uusing the
nonlocal projection algorithm (see Section 5.4.1),
inf
uJNLT V (u) + 1
2αkfuvk2
L2
ubeing fixed, find vusing the nonlocal projection algorithm,
inf
˜uJNLT V (˜u) + 1
2µkfu˜uk2
L2,
where v=fu˜u.
In the second step, where uis fixed, we need to solve infvGµkfuvk2
L2. See
the analysis of [2] or Section 3 in [3] showing that the minimization amounts to an
ROF problem. The arguments for our nonlocal version are similar. Note that in
practice one does not need to fully converge for each NL-ROF solution and can use
only a few iterations.
6. Experiments.
6.1. Nonlocal TV Inpainting. Here we show the distinct difference between
local and nonlocal TV-inpainting (see [16] for the local method). We minimize the
functional (4.2). As usual for this problem we assume having the inpainting regions
inp Ω in advance. We would like to fill-in the missing information in a sensible
manner according to the data in the rest of the image Ω/inp. In the toy example of
Fig. 6.1 a textured region has to be inpainted. The local TV chooses a locally smooth
solution (non-oscillatory) which does not fit this data. The nonlocal smoothness,
as defined by our functional (using patch-based distances) fills-in the information
correctly. The notion of smoothness is generalized to regularity. Thus patches with
similar partial data are selected to fill-in the missing information. In Fig. 6.2 two
examples of filling in parts of the Barbara image are shown. One is of regular texture
(knee part) and one is non-texture (face). It is shown that the missing information
is being replaced well by the algorithm, such that it is hard to distinguish visually
14 G. GILBOA AND S. OSHER
Original Inpainting region
Local TV Nonlocal TV
Fig. 6.1.Nonlocal vs. local TV inpainting. Top: original texture (left), the texture with the
inpainting region (in red). Bottom: results of local TV-inpainting [16] (left) and the nonlocal method
using Eq. (4.2). The nonlocal method recovers the texture pattern correctly.
between the original and the inpainted image. The errors of the inpainting results,
with respect to the original image, are presented on the right side (second and fourth
rows of the knes and face, respectively).
In the above cases we performed a single inpainting iteration and the missing
regions are required to be smaller than the patches. For filling-in larger regions, an
iterative process is necessary, where the boundaries have to be filled in first and one
can then recompute the weights deeper into the inpainting region and regularize again.
This process can be viewed as a variational understanding of the process suggested
by Efros and Leung [25]. See also a deterministic approach suggested in [24].
The inpainting regularization can work also with the quadratic regularizer J(u) =
R|∇wu|2dx. In the inpainting problem however, as oppose to denoising, the weights
for pixels to be inpainted are computed based on partial patches, where the central
point of the inpainting region is unknown. Thus there are cases where two types of
completions are possible. The quadratic regularizer will have a weighted averaging
solution whereas the T V -type regularizer will have a weighted median solution, which
is usually sharper and more attractive visually (see illustration in Fig. 6.3).
6.2. Nonlocal T V L1regularization. This model presents a new way of
viewing signal and image variational regularization. It replaces the local notion of
smoothness by the global notion of regularity. Thus features which appear frequently
are preserved.
The local T V L1model is known to remove outliers, such as impulsive noise
[41]. It is also known to keep intact large structures without reducing contrast (as
oppose to the T V L2case) while eliminating the smaller scales [14].
The nonlocal concept of “large” scales replaces the physical size of objects (pixels
with a constant color) with the frequency of their appearance. Thus smaller scales
NONLOCAL OPERATORS WITH APPLICATIONS TO IMAGE PROCESSING 15
Original Inpainting region
Nonlocal TV inpainting Error
Original Inpainting region
Nonlocal TV inpainting Error
Fig. 6.2.Nonlocal TV inpainting, Barbara image. Top: original, knee part (left), inpainting
region (in red). Second row: results of nonlocal TV-inpainting, Eq. (4.2) (left) and errors from
the original image. On the third and fourth rows an example of inpainting the face is shown (non-
textural part). The algorithm (with the same parameters) recovers well both types of regions.
should be interpreted as rare features. We obtain a variational regularization proce-
dure which detects and removes irregularities. This can be very useful for regularizing
textures, as seen in the examples below.
See Section for details regarding the calculation of the weights.
In Fig. 6.4 we give a toy example of small but very repetitive features versus large
but rare ones. We also add some white Gaussian noise (of standard deviation σ= 10).
We can observe that this regularization keeps the textures (small physical scale) and
removes the larger objects (replacing them with texture). The texture itself is also
regularized in the sense that the noise is removed. The residual part fucan be
viewed as an anomaly detector.
Fig. 6.5 depicts an experiment where the search neighborhood S(x) is changed.
S(x) controls the size of the region around each pixel for which similar patches are
examined (see Section 4.2 for details). It should reflect the expected auto-similarity
16 G. GILBOA AND S. OSHER
Fig. 6.3.TV versus quadratic regularization for inpainting. When two filling-in options are
possible, the TV regularizer takes a median solution, whereas the quadratic regularizer takes an
averaging solution.
scale of the image, or how far we should look in order to find repetitive structures. In
this example the patterns are about 30 pixels apart. When S(x) is too small (9 ×9
pixels), the regularization does not take into account the large-scale regularity of the
image. Thus the regularization quality is degraded: corners are eroded and outliers
and scratches are not removed well.
In Figs. 6.6 and 6.7 we try to detect and remove texture irregularities. We
compare the nonlocal T V L1with the local version and a simple 3 ×3 median filter.
We can see that the local T V L1behaves qualitatively in a similar manner to a
median filter (removes small physical features) whereas the nonlocal version keeps well
regular feature, even at the smallest scales. Outliers with a large scale can be removed
while retaining the fine coherence of the textural nature. For the nonlocal and local
T V L1regularization we retained the same residual L1norm: kfukL1. This fair
comparison is harder to obtain in the case of the median filter, but the residual norm
is of a similar value.
In Fig. 6.8 we regularized part of a zebra image with different values of λto show
qualitative the behavior of the nonlocal T V L1“scale-space”. Smaller value of λ
means stronger regularization.
NONLOCAL OPERATORS WITH APPLICATIONS TO IMAGE PROCESSING 17
Original Noisy (σ= 10)
u v
Fig. 6.4.Removal of Anomalies by nonlocal T V L1. Top: original image (left), image with
additive white Gaussian noise. Bottom: result of nonlocal T V L1regularization u(left), vfu.
This type of regularization retains repetitive patterns and removes rare and irregular ones (the light
and dark larger symbols in this case). Note also that a standard removal of the noise is achieved.
18 G. GILBOA AND S. OSHER
Input image
Large similarity scale
u v
Small similarity scale
u v
Fig. 6.5.Auto similarity scale in nonlocal T V L1. The search neighborhood S(x)for comput-
ing the weights (Section 4.2) controls the similarity scale of the regularization. Top: input image.
Middle: nonlocal TV L1with a large search neighborhood (S(x)is a window of size 61 ×61).
Bottom: nonlocal TV L1with a small search neighborhood (S(x)is a window of size 9×9).
NONLOCAL OPERATORS WITH APPLICATIONS TO IMAGE PROCESSING 19
Original f
NL T V L1u v
Local T V L1u v
Median filter u v
Fig. 6.6.Detecting and removing irregularities from textures by different methods. Example 1.
u(left) - regularized texture. v(right) - texture irregularities. The same L1norm of the residual
kvkL1is used for the nonlocal T V L1and the local TV L1.
20 G. GILBOA AND S. OSHER
Original f
NL T V L1u v
Local T V L1u v
Median filter u v
Fig. 6.7.Detecting and removing irregularities from textures by different methods. Example 2.
kvkL1is the same for the nonlocal T V L1and the local T V L1.
NONLOCAL OPERATORS WITH APPLICATIONS TO IMAGE PROCESSING 21
Original
λ= 2, u v
λ= 0.8, u v
λ= 0.3, u v
Fig. 6.8.NL T V L1. Regularization results for different values of λ.
22 G. GILBOA AND S. OSHER
Original f u v
Original f u v
Fig. 6.9.Examples of regularizing images with nonlocal T V G, Eq. (4.6). The filter removes
well random parts of the image, preserving edges and regular patterns.
Original fNL T V G u v
NL T V L1u v
Fig. 6.10.Nonlocal T V Gregularization, Eq. (4.6), compared with nonlocal T V L1, Eq.
(4.3). Nonlocal TV Gremoves random oscillations but keeps outliers.
6.3. Nonlocal T V Gregularization. In Fig. 6.9 two examples of regularizing
images with nonlocal T V G, Eq. (4.6), are shown. The images are taken from the
Kodak collection [33]. The qualitative properties of this regularization are different
from the original functional proposed by Meyer [37], see examples e.g. in [55, 2, 3].
A notable difference is that the regular textural part is preserved and only random
textures are removed from u. Edges, as in the original model, are well preserved
without erosion of contrast. Fig 6.10 compares the nonlocal T V Gminimization
with nonlocal T V L1, showing that the latter is more suited for removing outliers.
NONLOCAL OPERATORS WITH APPLICATIONS TO IMAGE PROCESSING 23
7. Conclusion. A very general framework is presented for processing signals and
images non-locally. Two categories of functionals are suggested: one which is based
on generalized nonlocal gradient and divergence operators. The other is based on
differences. In this paper we focus on the first category, present the general framework
and generalize several projection algorithm for computing the nonlocal versions of
ROF [48], T V L1[41] and T V G[37].
In essence two steps are required for this type of regularization: a first step consists
of finding the weights between pixels. We used patch based similarities following [9].
Other affinity measures between regions and pixels can naturally be proposed. The
second step is choosing the appropriate regularization and functional minimization.
It is shown how nonlocal T V L1can be used to detect and remove irregularities
from textures. In addition we demonstrate that nonlocal TV-inpainting can fill-in
repetitive textures correctly.
Preliminary calculations done elsewhere already indicate that Hamilton-Jacobi
equations in at least five space dimensions can be solved effectively using the approach
outlined in Section 3.
We currently would like to extend the theoretical foundations and also to inves-
tigate additional applications for which this framework can contribute.
REFERENCES
[1] G. Aubert and P. Kornprobst. Mathematical Problems in Image Processing, volume 147 of
Applied Mathematical Sciences. Springer-Verlag, 2002.
[2] J.F. Aujol, G. Aubert, L. Blanc-F´eraud, and A. Chambolle. Image decomposition into a
bounded variation component and an oscillating component. JMIV, 22(1), January 2005.
[3] J.F. Aujol, G. Gilboa, T. Chan, and S. Osher. Structure-texture image decomposition – mod-
eling, algorithms, and parameter selection. International Journal of Computer Vision,
67(1):111–136, 2006.
[4] S. Awate and R. Whitaker. Higher-order image statistics for unsupervised, information-
theoretic, adaptive image filteringg. In Proc. IEEE Int. Conf. Computer Vision and
Pattern Recognition, volume 2, pages 44–51, 2005.
[5] D. Barash. A fundamental relationship between bilateral filtering, adaptive smoothing and
the nonlinear diffusion equation. IEEE Transactions on Pattern Analysis and Machine
Intelligence, 24(6):844–847, 2002.
[6] D. Barash and D. Comaniciu. A common framework for nonlinear diffusion, adaptive smooth-
ing, bilateral filtering and mean shift. Image and Vision Computing, 22(1):73–81, 2004.
[7] S. Bougleux, A. Elmoataz, and M. Melkemi. Discrete regularization on weighted graphs for
image and mesh filtering. In 1st International Conference on Scale Space and Variational
Methods in Computer Vision (SSVM), volume 4485 of Lecture Notes in Computer Science,
pages 128–139, 2007.
[8] Y. Boykov, O. Veksler, and R. Zabih. Fast approximate energy minimization via graph cuts.
IEEE Transactions on Pattern Analysis and Machine Intelligence, 23(11):1222–1239, 2001.
[9] A. Buades, B. Coll, and J-M. Morel. On image denoising methods. SIAM Multiscale Modeling
and Simulation, 4(2):490–530, 2005.
[10] A. Buades, B. Coll, and J-M Morel. Neighborhood filters and PDE’s. Numerische Mathematik,
105(10):1–34, 2006.
[11] T. Cecil, J.L. Qian, and S. Osher. Numerical methods for high dimensional Hamilton-Jacobi
equations using radial bases functions. J. Comput. Phys., 196:327–347, 2004.
[12] A. Chambolle. An algorithm for total variation minimization and applications. JMIV, 20:89–97,
2004.
[13] A. Chambolle, R.A. De Vore, N. Lee, and B.J. Lucier. Nonlinear wavelet image processing:
Variational problems, compression, and noise removal through wavelet shrinkage. IEEE
Transcations on Image Processing, 7(3):319–335, March 1998.
[14] T. Chan and S. Esedoglu. Aspects of total variation regularized l1function approximation,
2004. CAM report 04-07.
[15] T. F. Chan, S. Osher, and J. Shen. The digital TV filter and nonlinear denoising. IEEE Trans.
Image Process., 10(2):231–241, 2001.
24 G. GILBOA AND S. OSHER
[16] T.F. Chan and J. Shen. Mathematical models of local non-texture inpaintings. SIAM J. Appl.
Math., 62(3):1019–1043, 2001.
[17] T.F. Chan and J. Shen. Image Processing and Analysis. SIAM, 2005.
[18] J-S. Chen, X. Zhang, and T. Belytschko. An implicit gradient model by a reproducing ker-
nel strain regularization in strain localization problems. Computer Methods in Applied
Mechanics and Engineering, 193:2827–2844, 2004.
[19] F. Chung. Spectral Graph Theory. Number 92 in CBMS Regional Conference Series in Math-
ematics. American Mathematical Society, 1997.
[20] R.R. Coifman, S. Lafon, A.B. Lee, M. Maggioni, B. Nadler, F. Warner, and S. Zucker. Ge-
ometric diffusion as a tool for harmonic analysis and structure definition of data, part
i: Diffusion maps. Proceedings of the National Academy of Sciences, 102(21):7426–7431,
2005.
[21] M.G. Crandall, H. Ishii, and P.-L. Lions. User’s guide to viscosity solutions of second order
partial differential equations. A.M.S. Bull., 27:1–67, 1992.
[22] J. Darbon and M. Sigelle. Image restoration with discrete constrained total variation part
ii: Levelable functions, convex priors and non-convex cases. Accepted to the Journal of
Mathematical Imaging and Vision, 2005.
[23] J. Darbon and M. Sigelle. Image restoration with discrete constrained total variation part i:
Fast and exact optimization. Journal of Mathematical Imaging and Vision, 2006.
[24] L. Demanet, B. Song, and T. Chan. Image inpainting by correspondence maps: a deterministic
approach. In Proc. VLSM, Nice, 2003. [see also UCLA CAM Report 03-40].
[25] A.A. Efros and T.K. Leung. Texture synthesis by non-parametric sampling. In ICCV (2),
pages 1033–1038, 1999.
[26] M. Elad. On the bilateral filter and ways to improve it. IEEE Transactions On Image Pro-
cessing, 11(10):1141–1151, 2002.
[27] G. Gilboa, J. Darbon, S. Osher, and T.F. Chan. Nonlocal convex functionals for image regu-
larization, 2006. UCLA CAM Report 06-57.
[28] G. Gilboa and S. Osher. Nonlocal linear image regularization and supervised segmentation.
SIAM Multiscale Modeling and Simulation, 6(2):595–630, 2007.
[29] L. Grady. Random walks for image segmentation. to appear in IEEE Trans. on Pattern
Analysis and Machine Intelligence, 2006.
[30] C. Kervrann and J. Boulanger. Unsupervised patch-based image regularization and represen-
tation. In Proc. European Conf. Comp. Vision (ECCV’06), Graz, Austria, 2006.
[31] R. Kimmel, R. Malladi, and N. Sochen. Images as embedding maps and minimal surfaces:
Movies, color, texture, and volumetric medical images. International Journal of Computer
Vision, 39(2):111–129, 2000.
[32] S. Kindermann, S. Osher, and P. Jones. Deblurring and denoising of images by nonlocal
functionals. SIAM Multiscale Modeling and Simulation, 4(4):1091 – 1115, 2005.
[33] Kodak. Kodak image collection, 2002. see http://www.cipr.rpi.edu/resource/stills/kodak.html.
[34] V. Kolmogorov and R. Zabih. What energy functions can be minimized via graph cuts? IEEE
Trans. Pattern Anal. Mach. Intell., 26(2):147–159, 2004.
[35] R.I. Kondor and J.D. Lafferty. Diffusion kernels on graphs and other discrete input spaces. In
ICML, pages 315–322, 2002.
[36] M. Mahmoudi and G. Sapiro. Fast image and video denoising via nonlocal means of similar
neighborhoods. IEEE Signal Processing Letters, 12(12):839–842, 2005.
[37] Y. Meyer. Oscillating patterns in image processing and in some nonlinear evolution equations,
March 2001. The Fifteenth Dean Jacquelines B. Lewis Memorial Lectures.
[38] B. Mohar. The Laplacian spectrum of graphs. In Y. Alavi, G. Chartrand, O. R. Oellermann,
A. J. Schwenk (Eds.), Graph Theory, Combinatorics, and Applications, Wiley, volume 2,
pages 871–898, 1991.
[39] G. Motta, E. Ordentlich, I. Rami’rez, G. Seroussi, , and M.J. Weinberger. The dude framework
for continuous tone image denoising. In Proc. IEEE Int. Conf. on Image Processing,
volume 3, pages 345–348, 2005.
[40] B. Nadler, S. Lafon, R.R. Coifman, and I.G. Kevrekidis. Diffusion maps, spectral clustering,
and the reaction coordinates of dynamical systems. Report, Math. Dept. Yale, Nov. 2004.
To appear in Journal of Applied and Computational Harmonic Analysis.
[41] M. Nikolova. A variational approach to remove outliers and impulse noise. JMIV, 20(1-2):99–
120, 2004.
[42] S. Osher and N. Paragios (Eds.). Geometric Level Set Methods in Imaging, Vision, and Graph-
ics. Springer-Verlag, 2003.
[43] S. Osher and R. Fedkiw. Level Set Methods and Dynamic Implicit Surfaces. Springer, 2002.
[44] S. Osher and J.A. Sethian. Fronts propagating with curvature dependent speed-algorithms
NONLOCAL OPERATORS WITH APPLICATIONS TO IMAGE PROCESSING 25
based on Hamilton-Jacobi formulations. J. Comput. Phys., 79:12–49, 1988.
[45] S. Osher and C.-W. Shu. High order essentially nonoscillatory schemes for Hamilton-Jacobi
equations. SINUM, 28:907–922, 1991.
[46] P. Perona and W.T. Freeman. A factorization approach to grouping. In ECCV, pages 655–670,
1998.
[47] P. Perona and J. Malik. Scale-space and edge detection using anisotropic diffusion. PAMI,
12(7):629–639, 1990.
[48] L. Rudin, S. Osher, and E. Fatemi. Nonlinear total variation based noise removal algorithms.
Physica D, 60:259–268, 1992.
[49] C. Sagiv, N. Sochen, and Y.Y. Zeevi. Integrated active contours for texture segmentation.
IEEE Trans. on Image Processing, 15(6):1633–1646, 2006.
[50] G. Sapiro. Geometric Partial Differential Equations and Image Processing. Cambridge Uni-
versity Press, 2001.
[51] J. Shi and J. Malik. Normalized cuts and image segmentation. IEEE Transactions on Pattern
Analysis and Machine Intelligence, 22(8):888–905, 2000.
[52] N. Sochen, R. Kimmel, and R. Malladi. A general framework for low level vision. IEEE
Transactions on Image Processing, 7:310–318, 1998.
[53] A.D. Szlam, M. Maggioni, Jr. J.C. Bremer, and R.R. Coifman. Diffusion-driven multiscale
analysis on manifolds and graphs: top-down and bottom-up constructions. In SPIE, 2005.
[54] C. Tomasi and R. Manduchi. Bilateral filtering for gray and color images. In ICCV ’98, pages
839–846, 1998.
[55] L. Vese and S. Osher. Modeling textures with total variation minimization and oscillating
patterns in image processing. Journal of Scientific Computing, 19:553–572, 2003.
[56] Y. Weiss. Segmentation using eigenvectors: A unifying view. In International Conference on
Computer Vision, pages 975–982, 1999.
[57] L.P. Yaroslavsky. Digital Picture Processing, an Introduction. Springer-Verlag, Berlin, 1985.
[58] D. Zhou and B. Scholkopf. A regularization framework for learning from graph data. In ICML
Workshop on Statistical Relational Learning and Its Connections to Other Fields, 2004.
[59] D. Zhou and B. Scholkopf. Regularization on discrete spaces. In Pattern Recognition, Proceed-
ings of the 27th DAGM Symposium, Berlin, Germany, pages 361–368, 2005.
... In this regard, mathematical models which are governed by nonlocal operators enrich our modeling spectrum and present useful alternatives as well as supplemental approaches. That is why they appear in a large variety of applications including among others, anomalous or fractional diffusion [10,11,19], peridynamics [25,27,54,64], image processing [31,38,42], cardiology [14], machine learning [44], as well as finance and jump processes [5,6,26,37,59]. Nonlocal operators are integral operators allowing for interactions between two distinct points in space. ...
Article
Full-text available
Shape optimization methods have been proven useful for identifying interfaces in models governed by partial differential equations. Here we consider a class of shape optimization problems constrained by nonlocal equations which involve interface–dependent kernels. We derive a novel shape derivative associated to the nonlocal system model and solve the problem by established numerical techniques. The code for obtaining the results in this paper is published at (https://github.com/schustermatthias/nlshape).
... Nonlocal reaction terms can also describe Darwinian evolution of a structured population density or the behaviors of cancer cells with therapy [5,6]. We refer to [7][8][9] for a wide range of applications. Controllability results for such problems have been considered before by researchers (see, e.g., [10][11][12][13]). ...
Article
Full-text available
In this paper, a one‐dimensional degenerate parabolic equation involving a multiplicative control with a nonlocal space term represented by an integral kernel is considered. After some well‐posedness results, we derive several properties of the control‐to‐state mapping. We establish the existence of at least one optimal control and derive the first and the second conditions for optimality, giving the local uniqueness of an optimal control. Finally, we prove that we may expect the global uniqueness of the optimal control under additional assumptions.
... Digital images, which consist of pixels arranged on a 2D grid, naturally possess a graph structure. For this reason, in recent years, graph-based differential operators ∆ have been introduced to replace standard Euclidean differential operators L in various image processing tasks such as denoising [19], image deblurring [1,8,10,14,42], Computed Tomography (CT) [31], and other applications [5,18,35]. Graph operators showed a general good performance, due to the fact that they can more effectively model the complex structures and textures present in images. ...
Preprint
Full-text available
Deep learning techniques have shown significant potential in many applications through recent years. The achieved results often outperform traditional techniques. However, the quality of a neural network highly depends on the used training data. Noisy, insufficient, or biased training data leads to suboptimal results. We present a hybrid method that combines deep learning with iterated graph Laplacian and show its application in acoustic impedance inversion which is a routine procedure in seismic explorations. A neural network is used to obtain a first approximation of the underlying acoustic impedance and construct a graph Laplacian matrix from this approximation. Afterwards, we use a Tikhonov-like variational method to solve the impedance inversion problem where the regularizer is based on the constructed graph Laplacian. The obtained solution can be shown to be more accurate and stable with respect to noise than the initial guess obtained by the neural network. This process can be iterated several times, each time constructing a new graph Laplacian matrix from the most recent reconstruction. The method converges after only a few iterations returning a much more accurate reconstruction. We demonstrate the potential of our method on two different datasets and under various levels of noise. We use two different neural networks that have been introduced in previous works. The experiments show that our approach improves the reconstruction quality in the presence of noise.
... Drawing from the definition provided in references [32][33][34][35], we proceed to define ∇ u the fractional derivative of order as follows: ...
Article
Full-text available
In this paper, we propose an anisotropic diffusion equation based on fractional‐order tensor diffusion applied on noisy images. We start by investigating the existence and uniqueness results of the proposed model. We also provide a discretization scheme based on the finite difference method. In order to prove the efficacy of our proposed approach, we deliver some simulations, which show robustness and remarkable results against noise while preserving edges and corners, compared to some well‐known models in literature. For a fair comparison, we manage to obtain the optimal parameters of every model thanks to PSNR and SSIM. We demonstrate the robustness efficacy of our model against high level of noise.
... Following Mengesha and Spector [7] (with roots in Bourgain, Brezis, and Mironescu [1], Gilboa and Osher [6], Du, Gunzburger, Lehouck, and Zhou [5]; see also a detailed list of references in [7, p. 254]), we set, for any measurable function u ∈ L 1 loc (R N ), and assuming that the integral below exists, ...
... Compared to traditional PDE models, fractional and nonlocal models preserve sharp boundaries and improve image quality in terms of peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM) [34] for image inpainting [35][36][37]. In addition, researchers have explored the use of fractional and nonlocal operators for other image processing tasks, including image denoising [38][39][40][41][42][43][44], image recovery in handling textures and repetitive structures [45,46], etc. ...
Article
In this study, we introduced a new nonlocal complex derivative operator in fractal dimension based concurrently on the concept of “nonlocal generalized complex backward-forward coordinates” and the “product-like fractal measure”. The quantization of the theory in fractal dimension leads to a higher order Schrödinger equation characterized by a higher order energy operator. As an illustration, we have discussed the cases of infinite quantum well and power-law potentials. Their associated zero-point energies were found to depend on the numerical value of the fractal dimension. For the infinite well, the decrease in zero-point energy with fractal dimension may result in the emission of large wavelengths photons observed experimentally in X-ray laser bursts emitted from the solid.
Article
Aim The accurate reconstruction of cone-beam computed tomography (CBCT) from sparse projections is one of the most important areas for study. The compressed sensing theory has been widely employed in the sparse reconstruction of CBCT. However, the total variation (TV) approach solely uses information from the i-coordinate, j-coordinate, and k-coordinate gradients to reconstruct the CBCT image. Materials and Methods It is well recognized that the CBCT image can be reconstructed more accurately with more gradient information from different directions. Thus, this study introduces a novel approach, named the new multi-gradient direction total variation minimization method. The method uses gradient information from the ij-coordinate, ik-coordinate, and jk-coordinate directions to reconstruct CBCT images, which incorporates nine different types of gradient information from nine directions. Results This study assessed the efficacy of the proposed methodology using under-sampled projections from four different experiments, including two digital phantoms, one patient’s head dataset, and one physical phantom dataset. The results indicated that the proposed method achieved the lowest RMSE index and the highest SSIM index. Meanwhile, we compared the voxel intensity curves of the reconstructed images to assess the edge structure preservation. Among the various methods compared, the curves generated by the proposed method exhibited the highest level of consistency with the gold standard image curves. Conclusion In summary, the proposed method showed significant potential in enhancing the quality and accuracy of CBCT image reconstruction.
Chapter
In this chapter, we study the simplest integro-differential elliptic operator: the square-root of the Laplacian, \(\sqrt {-\varDelta }\). We start by establishing its basic properties, including the harmonic extension representation and the corresponding heat kernel and fundamental solution. We then prove the comparison principle, compute its Poisson kernel in a ball and find the corresponding mean value property, deduce the Harnack inequality, and establish interior regularity estimates. Finally, we construct some explicit solutions and develop the analogous results for the fractional Laplacian \((-\Delta )^s\), with \(s \in (0, 1)\).
Conference Paper
We propose an algorithm for minimizing the total variation of an image, and provide a proof of convergence. We show applications to image denoising, zooming, and the computation of the mean curvature motion of interfaces.
Article
Image sharpening in the presence of noise is formulated as a non-convex variational problem. The energy functional incorporates a gradient-dependent potential, a convex fidelity criterion and a high order convex regularizing term. The first term attains local minima at zero and some high gradient magnitude, thus forming a triple well-shaped potential (in the one-dimensional case). The energy minimization flow results in sharpening of the dominant edges, while most noisy fluctuations are filtered out.
Article
We propose an algorithm for minimizing the total variation of an image, and provide a proof of convergence. We show applications to image denoising, zooming, and the computation of the mean curvature motion of interfaces.
Article
This book provides an introduction to the use of geometric partial differential equations in image processing and computer vision. This research area brings a number of new concepts into the field, providing a very fundamental and formal approach to image processing. State-of-the-art practical results in a large number of real problems are achieved with the techniques described in this book. Applications covered include image segmentation, shape analysis, image enhancement, and tracking. This book will be a useful resource for researchers and practitioners. It is intended to provide information for people investigating new solutions to image processing problems as well as for people searching for existent advanced solutions.