ArticlePDF Available

Feature Matching and Deformation for Texture Synthesis

Authors:

Abstract and Figures

One significant problem in patch-based texture synthesis is the presence of broken features at the boundary of adjacent patches. The reason is that optimization schemes for patch merging may fail when neighborhood search cannot find satisfactory candidates in the sample texture because of an inaccurate similarity measure. In this paper, we consider both curvilinear features and their deformation. We develop a novel algorithm to perform feature matching and alignment by measuring structural similarity. Our technique extracts a feature map from the sample texture, and produces both a new feature map and texture map. Texture synthesis guided by feature maps can significantly reduce the number of feature discontinuities and related artifacts, and gives rise to satisfactory results.
Content may be subject to copyright.
In the ACM SIGGRAPH 2004 conference proceedings
Feature Matching and Deformation for Texture Synthesis
Qing Wu Yizhou Yu
University of Illinois at Urbana-Champaign
{qingwu1,yyz}@uiuc.edu
Abstract
One significant problem in patch-based texturesynthesisisthepres-
ence of broken features at the boundary of adjacent patches. The
reason is that optimization schemes for patch merging may fail
when neighborhood search cannot find satisfactory candidates in
the sample texture because of an inaccurate similarity measure. In
this paper, we consider both curvilinear features and their deforma-
tion. We develop a novel algorithm to perform feature matching
and alignment by measuring structural similarity. Our technique
extracts a feature map from the sample texture, and produces both
a new feature map and texture map. Texture synthesis guided by
feature maps can significantly reduce the number of feature discon-
tinuities and related artifacts, and gives rise to satisfactory results.
CR Categories: I.3.7 [Computer Graphics]: Three-dimensional
Graphics and Realism—color, shading, shadowing, and texture
I.4.3 [Image Processing]: Enhancement—filtering, registration
I.4.6 [Image Processing]: Segmentation—Edge and feature detec-
tion
Keywords: Image Registration, Oriented Features, Texture Warp-
ing, Distance Transforms
1 Introduction
Texture synthesis has been widely recognized as an important re-
search topic in computer graphics. Recently, neighborhood-based
synthesis methods [Efros and Leung 1999; Wei and Levoy 2000;
Ashikhmin 2001; Liang et al. 2001; Hertzmann et al. 2001; Efros
and Freeman 2001; Zhang et al. 2003; Kwatra et al. 2003], espe-
cially patch-based techniques [Liang et al. 2001; Efros and Freeman
2001; Kwatra et al. 2003], have achieved significant progress. Nev-
ertheless, the presence of broken features at the boundary of two
adjacent patches is still a serious problem, though attempts have
been made to alleviate it [Efros and Freeman 2001; Kwatra et al.
2003].
These neighborhood-based algorithms have two common stages:
1) search in a sample texture for neighborhoods most similar to a
context region; 2) merge a patch or a pixel with the (partially) syn-
thesized output texture. Dynamic programming [Efros and Free-
man 2001] and graph cuts [Kwatra et al. 2003] have been used to
optimize the patch merging stage. One problem with these algo-
rithms is that the optimization schemes for the second stage may
fail to find a smooth transition when the first stage cannot find satis-
factory neighborhoods because of an inaccurate similarity measure.
We propose to perform texture synthesis using both salient fea-
tures and their deformation. Note that not every pixel is equally im-
portant when we measure perceptual similarity. A good perceptual
Figure 1: Texture synthesis with feature maps. From left to right:
sample textures (128x128), feature maps of the sample textures,
synthesized feature maps, output textures (256x256). Shown are
FLOOR and FLOWERS
c
1995 MIT VisTex.
measure should account for the fact that the human visual system
is most sensitive to edges, corners, and other high-level features in
textures. We call these high-level features structural information.
Structural similarity should be an important factor during neigh-
borhood search. So far, summed squared differences (SSD) of col-
ors is the most commonly used similarity measure between texture
patches. It is not very good at capturing structural information. A
desirable metric for structural information should estimate the min-
imal accumulated distance between corresponding features.
On the other hand, a small amount of deformation is less no-
ticeable than visual discontinuities. Neighborhood search in tex-
ture synthesis remarkably resembles image registration [Zitova and
Flusser 2003] in computer vision. Rigid template matching is the
simplest of these registration methods. Deformable templates and
elastic matching [Zitova and Flusser 2003] are not uncommon be-
cause object features and shapes may have different levels of dis-
tortion in different images. Deformation has also been considered
for a subclass of textures with near-regular patterns [Liu and Lin
2003]. It is desirable to have deformable models for generic tex-
ture synthesis as well. The visual difference between two texture
neighborhoods should reflect both color differences and shape de-
formations.
Among the large body of image registration techniques, of par-
ticular interest and relevance is chamfer matching [Barrow et al.
1977; Borgefors 1988] which was originally introduced to match
features from two images by means of the minimization of the gen-
eralized distance between them. Inspired by chamfer matching, we
develop a novel feature synthesis algorithm which considers struc-
tural similarity during feature matching and patch deformation dur-
ing feature alignment. We also develop a hybrid method for texture
synthesis by considering both features and colors simultaneously.
In this method, features are used to guide and improve texture syn-
thesis.
1
In the ACM SIGGRAPH 2004 conference proceedings
Context Region
DC
G
A
F
EB
C
C
C
C
1
2
3
0
(a) (b)
Figure 2: (a) The L-shaped context region of a patch ABCD be-
ing inserted into the output feature map. (b) The orientations of
tangents are quantized into four intervals.
2 Curvilinear Feature Matching and
Synthesis
Curvilinear thin features, such as edges and ridges, provide the
overall structural layout of textures. The set of curvilinear features
of a texture can be represented as a binary image, which is called
the feature map of the original texture. In this section, we introduce
a simple but effective algorithm that synthesizes a new feature map
from an existing one. Without loss of generality, we present this al-
gorithm in the context of patch-based synthesis. Curvilinear feature
detection will be addressed in Section 2.3. Examples of input and
synthesized feature maps are given in Fig. 1.
2.1 Curvilinear Feature Matching
We consider two factors when performing feature matching: differ-
ences in both position and tangent orientation. Position is important
since a smaller difference in position between matching features in-
dicates a smaller gap between them. Consistent tangent orientation
is critical to guarantee desirable geometric continuity and, there-
fore, visual smoothness between matching features.
Consider inserting a new patch in the output feature map. This
patch has a causal L-shaped context region in the already synthe-
sized region (Fig. 2(a)). The set of feature pixel locations in the
context region is represented as {f
out
i
}
m
i=1
. We simply translate
the context region to all feasible locations in the input feature map
when searching for a best match. The contextregion has an overlap-
ping region in the input feature map. Its location is specified by the
translation vector, T =(x, y), of the context region. The set
of features in this overlapping region is represented as {f
in
j
}
n
j=1
.
A matching cost between the two sets of aforementioned fea-
tures typically requires the shortest distance between each feature,
f
out
i
, in the first set and all the features in the second set. The fea-
ture in the second set that actually achieves this shortest distance
can be defined as the corresponding feature of f
out
i
. We use a non-
parametric mapping W
f
to represent such a correspondence. Since
our feature matching cost actually considers differences in tangent
vectors in addition to Euclidean distance, directly computing W
f
needs O(mn) time. To avoid such an expensive computation, we
only seek an approximate solution forW
f
through a quantization of
tangent orientation. The discrete quantization levels serve as ”buck-
ets”, making it much faster to search for features with a desired ori-
entation. A distance transform can be performed for features falling
into the same quantization level. Such distance transforms make it
possible to locate the nearest feature with a desired orientation in a
constant amount of time.
The orientations of the tangents at {f
out
i
}
m
i=1
are uniformly
quantized into four intervals (Fig. 2(b)). Each interval involves
two opposite directions. Depending on which interval its tangent
belongs to, we classify a feature pixel into one of four groups,
{C
out
l
}
3
l=0
. We index the groups such that two adjacent groups
3
2
L
L
L
2
1
L
1
L
T
out
out
in
in
in
Figure 3: The edges, L
out
1
and L
out
2
(dashed), in the context region
are being matched against the edges, L
in
1
, L
in
2
and L
in
3
(solid), in
the overlapping region of the input feature map. Pixels on edges
L
out
1
and L
out
2
should be matched to those on L
in
1
and L
in
3
, respec-
tively, because the tangents of L
in
2
are not consistent with those of
L
out
2
even though it is closer to L
out
2
than L
in
3
.
have adjacent orientation intervals. C
out
0
and C
out
3
are consid-
ered adjacent. Similarly, the orientation-based classification for
{f
in
j
}
n
j=1
is denoted as {C
in
l
}
3
l=0
.
Consider a feature f
out
i
which has been classified into C
out
l
i
.To
maintain tangent consistency, only features in C
in
l
i
, C
in
l
i
1
or C
in
l
i
+1
are allowed as candidates for W
f
(f
out
i
). f
out
i
has a nearest feature
in each of these three groups. These nearest features are denoted
as f
in
i
0
, f
in
i
1
and f
in
i
1
, respectively, where 1 i
1
,i
0
,i
1
n. Our
new distance metric between a pair of feature pixels is defined as
gdist(f
out
i
, f
in
j
)=f
out
i
f
in
j
2
+ τ v
out
i
v
in
j
2
(1)
where v
out
i
and v
in
i
represent the tangents at f
out
i
and f
in
i
, respec-
tively, and τ indicates the importance of tangent consistency. In our
approximate solution, W
f
(f
out
i
) satisfies the following condition,
gdist(f
out
i
,W
f
(f
out
i
)) = min
k∈{−1,0,1}
gdist(f
out
i
, f
in
i
k
). Fig. 3
illustrates the necessity of considering tangent consistency. For fea-
ture matching, we simply set τ =0.1 although further fine-tuning
is possible.
To facilitate feature alignment, it is desirable to have a one-
to-one mapping. Therefore, we define another quantity B
f
over
{f
in
j
}
n
j=1
to measure the bijectivity of the mapping W
f
. B
f
(f
in
j
)
represents the number of different features in {f
out
i
}
m
i=1
that are
mapped to the same feature f
in
j
. The matching cost between the
two sets of features is defined to be dependent on the amount of
distortion introduced by W
f
and the bijectivity of W
f
. It is formu-
lated as
1
m
i
gdist(f
out
i
,W
f
(f
out
i
)) + β
1
n
j
|B
f
(f
in
j
) 1| (2)
where β is a positive weight designed to adjust the relative impor-
tance between the two terms. In all experiments, we use β =0.3 if
distance is measured in pixels. The translation vector T that mini-
mizes the above matching costindicates the optimal matchingpatch
in the input feature map. At the end, we enforce a one-to-one map-
ping by removing extraneous corresponding features from the con-
text region.
We precompute four distance transforms for the input feature
map with a distinct transform for features in each ofthe four groups.
Such distance transforms can be efficiently computed using the
level set method [Sethian 1999]. At every pixel of the distance
maps, we also store a pointer which points to the closest feature
pixel. Meanwhile, every feature pixel in the input feature map has
a counter indicating the number of features to which it corresponds
in the current context region. With these data structures, evaluating
the above matching costbetween apair of overlapping regions has a
linear time complexity with respect to the number of feature pixels,
which are often very sparse in the image plane.
2
In the ACM SIGGRAPH 2004 conference proceedings
Input Feature Map Graphcut Quilting Texton Mask
Figure 4: Comparison of our algorithm with Graphcut, Image Quilting and Texton Masks. The second column shows the results from our
method. The graphcut result of the first example is courtesy of the authors of [Kwatra et al. 2003]; the results for the last two were generated
using our own implementation. The quilting results for the first two examples were generated from our implementation of [Efros and Freeman
2001]; the result for the last one is courtesy of Efros and Freeman. The results using texton masks were from our implementation of [Zhang
et al. 2003]. The samples shown are
EGGS, PATTERN(
c
Branko Gr
¨
unbaum and G.C. Shephard) and ROCK
c
1995 MIT VisTex. The
PATTERN
sample shown here is a slightly skewed version of an original periodic tiling. See additional comparisons in the DVD-ROM.
2.2 Feature Alignment Using Deformation
Even with the feature matching cost defined in the previous sec-
tion, feature misalignments may remain especially when the input
texture is aperiodic and contains no exact copies of the context re-
gion. Directly merging the optimal matching patch with the output
feature map would produce discontinuities thatmay not beremoved
by the techniques in [Efros and Freeman 2001; Kwatra et al. 2003].
As an optional step, we explicitly remove feature misalignment by
introducing a small amount of deformation in the image plane. This
can be convenientlyaccomplished by deforming the matching patch
using a smooth warping function.
We first compute a new feature mapping, W
f
, to obtain a sparse
feature correspondence between the context region and the opti-
mal matching patch found in the previous section: (x
i
, x
i
),i =
1, 2, ..., m. This mapping is in general different from the mapping,
W
f
, obtained during feature matching. In the current context, the
mapping W
f
is computed using the same matching cost as in (2) but
with τ in (1) set to 10 to emphasize tangents. A warping function
should smoothly deform the optimal matching patch while moving
each feature x
i
in this patch to the location of its corresponding fea-
ture x
i
in the context region. Note that the optimal matching patch
has four borders. To prevent the accumulation of deformations, we
require that the pixels on the bottom and right borders be fixed dur-
ing warping. These fixed pixels are also considered as constraints
that the warping function should satisfy.
Obtaining the warping function is equivalent to scattered data
interpolation. We apply two commonly used interpolation tech-
niques: thin-plate splines (TPS) [Meinguet 1979; Turk and O’Brien
1999] and Shepard’s method [Hoschek and Lasser 1993]. Thin-
plate splines have the minimal bending energy among all inter-
polants satisfying the warp constraints. However, computing the
thin-plate spline requires solving a linear system which may be oc-
casionally ill-conditioned. Therefore, given a sparse feature cor-
respondence, we first apply thin-plate spline interpolation. If the
resulting warping function cannot satisfy the warp constraints, we
switch to Shepard’s method instead.
2.3 Feature Detection
In this paper, we only consider easy-to-detect features, such as
edges and ridges. For edge detection, we first apply bilateral fil-
tering [Tomasi and Manduchi 1998] to sharpen the edges. In the
bilateral filter, the scale of the closeness function σ
d
is always set
to 2.0, and the scale of the similarity function σ
r
is always set to
10 out of 256 greyscale levels. We then use finite differences along
the two image axes as a simple gradient estimator to obtain an edge
response at every pixel. This is followed by a two-pass classifi-
cation. In the first pass, a global high threshold is used to detect
strong edges which are usually broken into small pieces. In the
second pass, a spatially varying lower threshold is used to detect
weaker edges in the neighborhood of each strong edge. Unlike the
lower threshold in the Canny detector [Canny 1986], it is locally
dependent on the edge responses of the pixels detected in the first
3
In the ACM SIGGRAPH 2004 conference proceedings
pass. In practice, we choose a fixed ratio α. The lower threshold
for a neighborhood surrounding a strong edge with response R is
set to αR. The second pass can effectively connect broken strong
edges. For ridge detection, we apply the Laplacian filter after bilat-
eral filtering. Once there is a filter response at every pixel, the same
two-stage classification for edges is also applied to ridges. For the
results shown in this paper, α is always set to 0.3. The global high
threshold is first estimated automatically using a fixed percentile of
the highest filter response in the entire image. It can then be ad-
justed interactively by the user to improve the results.
Since the detected edges or ridges may have a multi-pixel width,
we further apply a revised thinning algorithm which removes pix-
els with weak filter responses first while preserving the connectiv-
ity of the features. The resulting features always have one-pixel
width. Detailed discussions on thinning algorithms can be found in
[Pavlidis 1982]. At each detected feature pixel, we store its color
and smoothed tangent as its attributes.
3 Feature-Guided Texture Synthesis
Our feature map synthesis is complementary to existing patch-
based texture synthesis algorithms [Liang et al. 2001; Efros and
Freeman 2001; Kwatra et al. 2003]. To incorporate feature maps
into texture synthesis, we designed a hybrid method that generates
a new texture and its feature map simultaneously given an input tex-
ture and its feature map. Every time we need to insert a new patch,
we consider feature matching and color matching simultaneously.
We apply the matching cost in (2) for features, and SSD for col-
ors. The total matching cost is a weighted average between these
two. Once a matching patch is chosen from the sample texture, it is
deformed according to Section 2.2. Before merging the matching
patch with the partial output texture, we apply graph cuts [Kwatra
et al. 2003] to further improve the transition at the patch boundary.
Fig. 1 shows synthesis examples along with feature maps. As
shown in the second row of Fig. 1, stochastic feature maps can
also generate decent synthesis results. A comparison between our
method and three other state-of-the-art algorithms is given in Fig.
4. It demonstrates that our method does very well at maintaining
the continuity of structural features as well as the shapes of individ-
ual objects in the textures without merging them. More synthesis
examples are given in Fig. 5, where the resolution of the output
textures is 256x256. The running time for texture synthesis is less
than two minutes on a 2GHz Pentium IV processor for sample and
output textures at a resolution 128x128 and 256x256, respectively.
The patch size is chosen between 32x32 and 64x64; the relative
weight between feature and SSD color matching costs is always set
to 0.5. For certain sample textures (
FLOOR and FLOWERS in Fig.
1,
LEAVES and WATER in Fig. 5), their rotated or reflected versions
are also presented as input to our program to provide more texture
variations.
4 Discussion
In this paper, we introduced a novel feature-based synthesis method
that extends previous texture synthesis techniques through the use
of local curvilinear feature matching and texture warping. When
such oriented features are absent, our technique reverts back to the
behavior of previously published color-based methods. Unlike tex-
ton mask extraction [Zhang et al. 2003] which needs manual inter-
vention, it is easier to automatically obtain our feature maps. The
feature map of a sample texture can also be improved with user
interaction. Most importantly, new feature maps are synthesized
using a novel matching criterion custom-designed for curvilinear
features. This criterion is capable of guiding texture synthesis to
produce better results.
Figure 5: More texture synthesis results. The smaller images are
the sample textures. Shown are
LEAVES
c
Paul Bourke, A
RABIC
TEXT
, BAMBOO
c
1995 MIT VisTex, and
WATER.
Acknowledgments
We wish to thank the authors of [Kwatra et al. 2003] and [Efros and
Freeman 2001] for sharing their results, Stephen Zelinka for proof-
reading, and the anonymous reviewers for their valuable comments.
This work was funded by NSF (CCR-0132970).
References
ASHIKHMIN, M. 2001. Synthesizing natural textures. In ACM Symposium on Inter-
active 3D Graphics, 217–226.
B
ARROW,H.,TENENBAUM, J., BOLLES, R., AND WOLF, H. 1977. Parametric
correspondence and chamfer matching: Two new techniques for image matching.
In Proc. 5th Intl. Joint Conf. on Art. Intell., 659–663.
B
ORGEFORS, G. 1988. Hierarchical chamfer matching: a parametric edge matching
algorithm. IEEE Trans. Pattern Analysis and Machine Intelligence 10, 849–865.
C
ANNY, J. 1986. A computational approach to edge detection. IEEE Trans. Pat. Anal.
Mach. Intell. 8, 6, 679–698.
E
FROS,A.,AND FREEMAN, W. 2001. Image quilting for texture synthesis and
transfer. In SIGGRAPH’01, 341–346.
E
FROS,A.,AND LEUNG, T. 1999. Texture synthesis by non-parametric sampling. In
Intl. Conf. Computer Vision, 1033–1038.
H
ERTZMANN, A., JACOBS, C., OLIVER, N., CURLESS, B., AND SALESIN, D. 2001.
Image analogies. In SIGGRAPH’01, 327–340.
H
OSCHEK, J., AND LASSER, D. 1993. Fundamentals of Computer Aided Geometric
Design. AK Peters, Ltd.
K
WATRA,V.,SCHODL, A., ESSA, I., TURK,G.,AND BOBICK, A. 2003. Graphcut
textures: Image and video synthesis using graph cuts. In SIGGRAPH’03, 277–286.
L
IANG,L.,LIU, C., XU,Y.,GUO, B., AND SHUM, H.-Y. 2001. Real-time texture
synthesis using patch-based sampling. ACM Trans. Graphics 20, 3, 127–150.
L
IU,Y.,AND LIN, W.-C. 2003. Deformable texture: the irregular-regular-irregular
cycle. In The 3rd intl. workshop on texture analysis and synthesis, 65–70.
M
EINGUET, J. 1979. Multivariate interpolation at arbitrary points made simple. J.
Applied Math. Physics 5, 439–468.
P
AVLIDIS, T. 1982. Algorithms for Graphics and Image Processing. Computer Sci-
ence Press.
S
ETHIAN, J. 1999. Level Set Methods and Fast Marching Methods. Cambridge
University Press.
T
OMASI, C., AND MANDUCHI, R. 1998. Bilateral filtering for gray and color images.
In Proc. Intl. Conf. on Computer Vision, 836–846.
T
URK,G.,AND O’BRIEN, J. 1999. Shape transformation using variational implicit
functions. In SIGGRAPH 99 Conference Proceedings, 335–342.
W
EI, L.-Y., AND LEVOY, M. 2000. Fast texture synthesis using tree-structured vector
quantization. In Proceedings of Siggraph, 479–488.
Z
HANG,J.,ZHOU,K.,VELHO,L.,GUO, B., AND SHUM, H.-Y. 2003. Synthesis of
progressively-variant textures on arbitrary surfaces. In SIGGRAPH’03, 295–302.
Z
ITOVA, B., AND FLUSSER, J. 2003. Image registration methods: a survey. Image
and Vision Computing 21, 977–1000.
4
... Sketch-to-image synthesis generates images from sketches basically which are GAN-based. The various methods proposed earlier for image/video synthesis and editing, include texture synthesis [41][42][43], image inpainting [44][45][46], image stylization [47,48], image deformation [49,50], and so on. Inspite of all these, intelligent synthesis of images is a challenging problem since the traditional methods are based on pixels [41,44], patches [48,50], and low-level image features [43], which are deprived of high level information needed for synthesizing. ...
... The various methods proposed earlier for image/video synthesis and editing, include texture synthesis [41][42][43], image inpainting [44][45][46], image stylization [47,48], image deformation [49,50], and so on. Inspite of all these, intelligent synthesis of images is a challenging problem since the traditional methods are based on pixels [41,44], patches [48,50], and low-level image features [43], which are deprived of high level information needed for synthesizing. ...
Conference Paper
Full-text available
In recent years, Image based (data driven) methodologies like Deep Learning and Computer Vision are more popular and precise in finding the features inside the image and generate realistic images. These give rise to Generative models like Variational Autoencoder and Generative Adversarial Networks (GAN). GANs play a major role in research as they translate images from one domain to another keeping structure aligned with input and generates realistic images as output. They are comprised of generator and discriminator networks trained in competition with each other to learn and generate distributions. There are many GAN algorithms proposed since 2014. GANs with different machine learning algorithms are used in many real-world applications. Very few comprehensive studies on GANs are available on different applications in literature. In this paper we try to give an insight on GANs in terms of theory and algorithms and also emphasize on review of applications of GAN in constrained image synthesis majorly image-to-image, text-to-image and sketch-to-image translation
... Early works of texture generation generates textures in pixel space. Based on exemplary input, texture can be generated pixel-by-pixel [8,9,10] or patch-by-patch [11,12,13,14,15], starting from a small unit and gradually growing to a larger image. These methods, however, suffer from slow generation time [9,12] or limited diversity of generated textures [11,13,16]. ...
... Efros and Leung [9] proposed to synthesize a pixel by randomly choosing from the pixels that have similar neighborhood as the query pixel. Patch-based methods [11,12,13,14,15] usually achieves higher quality than pixel-based methods [8,9,10]. They suffer from two problems. ...
Preprint
An explainable, efficient and lightweight method for texture generation, called TGHop (an acronym of Texture Generation PixelHop), is proposed in this work. Although synthesis of visually pleasant texture can be achieved by deep neural networks, the associated models are large in size, difficult to explain in theory, and computationally expensive in training. In contrast, TGHop is small in its model size, mathematically transparent, efficient in training and inference, and able to generate high quality texture. Given an exemplary texture, TGHop first crops many sample patches out of it to form a collection of sample patches called the source. Then, it analyzes pixel statistics of samples from the source and obtains a sequence of fine-to-coarse subspaces for these patches by using the PixelHop++ framework. To generate texture patches with TGHop, we begin with the coarsest subspace, which is called the core, and attempt to generate samples in each subspace by following the distribution of real samples. Finally, texture patches are stitched to form texture images of a large size. It is demonstrated by experimental results that TGHop can generate texture images of superior quality with a small model size and at a fast speed.
... The pixel from the input texture whose neighborhood has a minimum difference from that of the on-processing pixel from the output texture will be chosen. Another branch of the exemplar-based method is the patch-based method [19,20,21,22], which moves a patch from the examplar per step according to neighborhood-comparing criteria similar to that adopted by pixel-based method. Patch-based methods can be generally more efficient than pixel-based methods, and perform better in preserving large-scale features. ...
Preprint
The emerging Neural Radiance Field (NeRF) shows great potential in representing 3D scenes, which can render photo-realistic images from novel view with only sparse views given. However, utilizing NeRF to reconstruct real-world scenes requires images from different viewpoints, which limits its practical application. This problem can be even more pronounced for large scenes. In this paper, we introduce a new task called NeRF synthesis that utilizes the structural content of a NeRF patch exemplar to construct a new radiance field of large size. We propose a two-phase method for synthesizing new scenes that are continuous in geometry and appearance. We also propose a boundary constraint method to synthesize scenes of arbitrary size without artifacts. Specifically, we control the lighting effects of synthesized scenes using shading guidance instead of decoupling the scene. We have demonstrated that our method can generate high-quality results with consistent geometry and appearance, even for scenes with complex lighting. We can also synthesize new scenes on curved surface with arbitrary lighting effects, which enhances the practicality of our proposed NeRF synthesis approach.
... These methods do not maintain boundaries of the elements in the exemplars while generating textures. A number of patch-based methods that use discrete elements have been proposed [28][29][30][31][32][33][34][35]. These patch-based methods synthesize a texture from exemplars that include discrete elements and their relative distances. ...
Article
We propose a method for modeling nonperiodic aggregates composed of arbitrary elements on a two-dimensional plane. The method effectively generates a dense aggregate from the elements to be arranged. To date, a Poisson disk distribution has been used to generate nonperiodic aggregates; however, the approach creates large gaps among the elements. Dense aggregates of arbitrary elements can be generated with a dart-throwing method; however, this approach is time consuming. To reduce the calculation time, we propose a method that fills in elements in the gaps among the already placed elements. To effectively fill these gaps, our method quantifies the gaps and finds additional positions for the elements. Our experiment showed that our method is more effective than the dart-throwing method. Also, to confirm that our method is generic, examples using two- and three-dimensional elements are shown.
... Galin et al. [22,Sec. 5.1], in their survey of the field, cite three previous example-based terrain synthesis methods that build on texture synthesis: those of Zhou et al. [59] and Tasse et al. [54], based primarily on patchbased texture synthesis method of Wu and Yu [58], and the method by Gain et al. [21], based primarily on pixel-based methods of Lefebvre and Hoppe [31] and Han et al. [26]. ...
Article
We report two studies that investigate the use of subjective believability in the assessment of objective realism of terrain. The first demonstrates that there is a clear subjective feature bias that depends on the types of terrain being evaluated: our participants found certain natural terrains to be more believable than others. This confounding factor means that any comparison experiment must not ask participants to compare terrains with different types of feature. Our second experiment assesses four methods of example-based terrain synthesis, comparing them against each other and against real terrain. Our results show that, while all tested methods can produce terrain that is indistinguishable from reality, all also can produce poor terrain; that there is no one method that is consistently better than the others; and that those who have professional expertise in geology, cartography or image analysis are better able to distinguish real terrain from synthesised terrain than the general population but those who have professional expertise in the visual arts are not.
... Liang's approach with Efros's algorithm decreases the mismatching features around the patch limits by cutting the minimum error boundary [103]. In deformation and feature matchingbased texture synthesis, deformation and curve features are employed to increase the efficiency of identifying the matching patch [104]. New optimization-based approaches have been developed by inter-mating both the pixel-based and patchbased approaches [105], introduced in recent years, combining the advantages of both pixel and patch methods [106]. ...
Chapter
Full-text available
Artificial vision, also known as computer vision or technical vision, is the scientific discipline that includes methods to acquire, process, analyze, and understand images of the real world in order to produce numerical or symbolic information so that they can be processed by a computer. Just as humans use our eyes and brains to understand the world around us, computer vision tries to produce the same effect so that computers can perceive and understand an image or sequence of images and act as appropriate in a given situation. Applications in this discipline show its use in the monitoring and development of the maturity stages for various crops. On the other hand, the cultivation of medicinal cannabis in Colombia is taking on a significant economic and scientific interest, and due to the limited local research around this plant, it is necessary to direct efforts to study its different phenological stages that allow determining the moment of maturity optimal for harvesting. The present work proposes the future contribution of computational vision in the detection of maturity states of medicinal cannabis. In this study, two main topics are discussed: first, a review of the state of the art on the use of computer vision techniques related to the recognition of images applied to monitoring the development of various crops, and second, an elaboration of a methodological proposal for the identification of the optimal harvest time for medicinal cannabis inflorescences and the development of a computational application. Conclusions at the end of the work are raised in which the use of artificial neural networks and the creation of a database for their training stand out.
... This has been used by many "edge-aware" image processing techniques [15] including filtering [51] and denoising [9]. Not surprisingly, they have been also considered essential in texture synthesis: Several methods apply a prior edge detection, such as the Canny edge detector, to improve or guide synthesis [48,53] or to synthesize the edge strokes directly [2,26,34]. Portilla et al. use statistics on wavelets coefficients to generate textures [39], and Wu et al. [49] focus on curvilinear features in art patterns. ...
Article
Full-text available
We introduce an edge-based procedural texture (EBPT), a procedural model for semi-stochastic texture generation. EBPT quickly generates large textures from a small input image. EBPT focuses on edges as the visually salient features extracted from the input image and organizes into groups with clearly established spatial properties. EBPT allows the users to interactively or automatically design new textures by utilizing the edge groups. The output texture can be significantly larger than the input, and EBPT does not need multiple textures to mimic the input. EBPT-based texture synthesis consists of two major steps, input analysis and texture synthesis. The input analysis stage extracts edges, builds the edge groups, and stores procedural properties. The texture synthesis stage distributes edge groups with affine transformation. This step can be done interactively or automatically using the procedural model. Then, it generates the output using edge group-based seamless image cloning. We demonstrate our method on various semi-stochastic inputs. With just a few input parameters defining the final structure, our method can analyze the input size of 512×512512\times {512} in 0.7 s and synthesize the output texture of 2048×20482048\times {2048} pixels in 0.5 s.
Article
A user controlled texture synthesis technique reproduces the characteristic color pattern of a sample image on an output image in accordance with a paint image given by a user. This technique does not specially treat the contour of an object, that is, the boundary between an object and its background on the sample image. Thus, when a user wants to use only the pixels of the object for no background or other backgrounds in the synthesis, it is difficult to reproduce the characteristic contour of the object on an output image in accordance with the contour of a paint region roughly painted by the user. Therefore, a boundary expansion texture synthesis technique was proposed to reproduce the characteristic object contour of a sample image by appropriately expanding the contour of a paint region based on the user controlled texture synthesis. In this paper, we propose a transparent boundary expansion texture synthesis technique based on the previous technique. Our new technique reproduces the finer object contour as well as the characteristic color pattern of a sample image by using opacity values, alphas, given to pixels of sample and paint images.
Article
A user controlled texture synthesis technique was proposed to allow a user to control texture synthesis. The technique creates an output texture (output image) from an input texture (input image) by using a target image painted by a user. When an object and its background exist in the input image, the user controlled texture synthesis technique usually treats the pixels of the object and those of the background equally without treating the contour of the object intentionally. Therefore, if we want to synthesize only the input pixels of the object onto a different background, it is difficult to reconstruct the characteristic contour of the object on an output image by giving a simple contour roughly painted on a target image. In this paper, we propose a texture synthesis technique to reconstruct a characteristic object contour intentionally by expanding the boundary of a painted region on a target image suitably and selecting optimal input pixels to synthesize by the evaluation using boundary pixels and boundary distances.
Article
Full-text available
This paper describes a computational approach to edge detection. The success of the approach depends on the definition of a comprehensive set of goals for the computation of edge points. These goals must be precise enough to delimit the desired behavior of the detector while making minimal assumptions about the form of the solution. We define detection and localization criteria for a class of edges, and present mathematical forms for these criteria as functionals on the operator impulse response. A third criterion is then added to ensure that the detector has only one response to a single edge. We use the criteria in numerical optimization to derive detectors for several common image features, including step edges. On specializing the analysis to step edges, we find that there is a natural uncertainty principle between detection and localization performance, which are the two main goals. With this principle we derive a single operator shape which is optimal at any scale. The optimal detector has a simple approximate implementation in which edges are marked at maxima in gradient magnitude of a Gaussian-smoothed image. We extend this simple detector using operators of several widths to cope with different signal-to-noise ratios in the image. We present a general method, called feature synthesis, for the fine-to-coarse integration of information from operators at different scales. Finally we show that step edge detector performance improves considerably as the operator point spread function is extended along the edge.
Article
Full-text available
This paper aims to present a review of recent as well as classic image registration methods. Image registration is the process of overlaying images (two or more) of the same scene taken at different times, from different viewpoints, and/or by different sensors. The registration geometrically align two images (the reference and sensed images). The reviewed approaches are classified according to their nature (area-based and feature-based) and according to four basic steps of image registration procedure: feature detection, feature matching, mapping function design, and image transformation and resampling. Main contributions, advantages, and drawbacks of the methods are mentioned in the paper. Problematic issues of image registration and outlook for the future research are discussed too. The major goal of the paper is to provide a comprehensive reference source for the researchers involved in image registration, regardless of particular application areas.
Article
Full-text available
In this paper we introduce a new algorithm for image and video texture synthesis. In our approach, patch regions from a sample image or video are transformed and copied to the output and then stitched together along optimal seams to generate a new (and typically larger) output. In contrast to other techniques, the size of the patch is not chosen a-priori, but instead a graph cut technique is used to determine the optimal patch region for any given offset between the input and output texture. Unlike dynamic programming, our graph cut technique for seam optimization is applicable in any dimension. We specifically explore it in 2D and 3D to perform video texture synthesis in addition to regular image synthesis. We present approximative offset search techniques that work well in conjunction with the presented patch size optimization. We show results for synthesizing regular, random, and natural images and videos. We also demonstrate how this method can be used to interactively merge different images to generate new scenes.
Article
Parametric correspondence is a technique for matching images to a three dimensional symbolic reference map. An analytic camera model is used to predict the location and appearance of landmarks in the image, generating a projection for an assumed viewpoint. Correspondence is achieved by adjusting the parameters of the camera model until the appearances of the landmarks optimally match a symbolic description extracted from the image. The matching of image and map features is performed rapidly by a new technique, called "chamfer matching", that compares the shapes of two collections of shape fragments, at a cost proportional to linear dimension, rather than area. These two techniques permit the matching of spatially extensive features on the basis of shape, which reduces the risk of ambiguous matches and the dependence on viewing conditions inherent in conventional image-based correlation matching.
Article
The concrete method of surface spline interpolation is closely connected with the classical problem of minimizing a Sobolev seminorm under interpolatory constraints; the intrinsic structure of surface splines is accordingly that of a multivariate extension of natural splines. The proper abstract setting is a Hilbert function space whose reproducing kernel involves no functions more complicated than logarithms and is easily coded. Convenient representation formulas are given, as also a practical multivariate extension of the Peano kernel theorem. Owing to the numerical stability of Cholesky factorization of positive definite symmetric matrices, the whole construction process of a surface spline can be described as a recursive algorithm, the data relative to the various interpolation points being exploited in sequence.La mthode concrte d'interpolation par surfaces-spline est troitement lie au problme classique de la minimisation d'une semi-norme de Soboleff sous des contraintes d'interpolation; la structure intrinsque des surfaces-spline est ds lors celle d'une extension multivarie des fonctions-spline naturelles. Le cadre abstrait adquat est un espace fonctionnel hilbertien dont le noyau reproduisant ne fait pas intervenir de fonctions plus compliques que des logarithmes et est ais programmer. Des formules commodes de reprsentation sont donnes, ainsi qu'une extension multivarie d'intrt pratique du thorme du noyau de Peano. Grce la stabilit numrique de la factorisation de Cholesky des matrices symtriques dfinies positives, la construction d'une surface-spline peut se faire en exploitant point aprs point les donnes d'interpolation.
Book
A classic reference and text, this book introduces the foundations used to create an accurate computer screen image using mathematical tools. This comprehensive guide is a handbook for students and practitioners and includes an extensive bibliography for further study.
Article
We present an approach for decorating surfaces with progressively-variant textures. Unlike a homogeneous texture, a progressively-variant texture can model local texture variations, including the scale, orientation, color, and shape variations of texture elements. We describe techniques for modeling progressively-variant textures in 2D as well as for synthesizing them over surfaces. For 2D texture modeling, our feature-based warping technique allows the user to control the shape variations of texture elements, making it possible to capture complex texture variations such as those seen in animal coat patterns. In addition, our feature-based blending technique can create a smooth transition between two given homogeneous textures, with progressive changes of both shapes and colors of texture elements. For synthesizing textures over surfaces, the biggest challenge is that the synthesized texture elements tend to break apart as they progressively vary. To address this issue, we propose an algorithm based on texton masks, which mark most prominent texture elements in the 2D texture sample. By leveraging the power of texton masks, our algorithm can maintain the integrity of the synthesized texture elements on the target surface.
Article
We present an algorithm for synthesizing textures from an input sample. This patch-based sampling algorithm is fast and it makes high-quality texture synthesis a real-time process. For generating textures of the same size and comparable quality, patch-based sampling is orders of magnitude faster than existing algorithms. The patch-based sampling algorithm works well for a wide variety of textures ranging from regular to stochastic. By sampling patches according to a nonparametric estimation of the local conditional MRF density function, we avoid mismatching features across patch boundaries. We also experimented with documented cases for which pixel-based nonparametric sampling algorithms cease to be effective but our algorithm continues to work well.