ChapterPDF Available

Abstract and Figures

In this paper, we explore different notions of shape complexity, drawing from established work in mathematics, computer science, and computer vision. Our measures divide naturally into three main categories: skeleton-based, symmetry-based, and those based on boundary sampling. We apply these to an established library of shapes, using k-medoids clustering to understand what aspects of shape complexity are captured by each notion. Our contributions include a new measure of complexity based on the Blum medial axis and the notion of persistent complexity as captured by histograms at multiple scales rather than a single numerical value.
Content may be subject to copyright.
Exploring 2D shape complexity
Erin Chambers, Tegan Emerson, Cindy Grimm, Kathryn Leonard
Abstract
In this paper, we explore different notions of shape complexity, draw-
ing from established work in mathematics, computer science, and com-
puter vision. Our measures divide naturally into three main categories:
skeleton-based, symmetry-based, and those based on boundary sampling.
We apply these to an established library of shapes, using k-medoids clus-
tering to understand what aspects of shape complexity are captured by
each notion. Our contributions include a new measure of complexity based
on the Blum medial axis, and the notion of persistent complexity as cap-
tured by histograms at multiple scales rather than a single numerical
value.
1 Introduction
Quantifying shape complexity and similarity has a rich history in many fields of
mathematics and computer science, including work in fields such as differential
geometry, topology, computational geometry, computer graphics, and computer
vision. This paper draws measures common to several of those fields in order
to compare and contrast their ability to capture complexity of a variety of two-
dimensional shapes, as well as introducing a new approach to complexity based
on measures taken along the Blum medial axis.
1.1 Prior work on shape complexity
Several examples of prior work on shape complexity draw from the information
theoretic framework. For example, in [CS05], the authors consider a notion of
complexity based on computing measures related to the Kolmogorov complexity
of a set of sample points from the boundary of the contour. This method has
the advantage of calculating a single complexity value which in practice does
well at distinguishing between various classes of shapes, in agreement with a
user study. Unfortunately, it is extremely difficult to classify exactly what is
being measured given the loose connection to Kolmogorov complexity, which is
uncomputable.
Similar work in computer vision has taken a different approach, quantifying
complexity by how difficult it is to capture or cover the shape by simpler ones.
For example, in [PA16], the authors measure complexity by attempting to de-
termine how many ellipses are necessary to cover a 2D shape. They compute
1
the medial axis for each edge, prune to reduce noise, then calculate an entropy
in order to determine how many ellipses are necessary to cover the total area as
closely as possible. The basic premise is that a shape requiring more ellipses,
or one resulting in less coverage with more ellipses, will correspond to a more
complex shape.
In [CI84], the authors focus on a classical computational geometric measure
of complexity, namely deciding how quickly they can triangulate a given input
polygon. While of theoretical interest, this measure does not translate to a
richer notion of shape complexity.
Curvature of the boundary appears in several complexity measures. In
[JF05], the authors determine which regions of a particular contour are more
”surprising” in a probabilistic sense, finding that regions of negative curvature
carry more information than those of positive curvature of the same magnitude.
Similarly, [PKS+03] defines probability distributions of curvatures in order to
assign an entropy to each shape. A discretization-independent approach of the
same idea can be found in [RFS05]. In [Leo07], total curvature of the boundary
gives the adaptive codelength for a contour, where one may view codelength as
a proxy for complexity. One drawback to the curvature-based derivations, how-
ever, is their lack of robustness to noise or small-scale detail on the boundary
which can dramatically alter curvature distributions.
It is also worth mentioning work that attempts to classify the complex-
ity of 3-dimensional objects, even though our work focuses on the 2D setting.
In [OFCD02], for example, the authors propose a signature for an object as
a shape distribution, which is a similar idea to how we approach the problem
in the 2D setting, but their shape function primarily measures local geometric
properties of an object. Similarly, in [SPG+06] the authors use curvature-based
and information theoretic measures to classify 3d shapes, although in this work
they focus on shape segmentation and identification of prominent features, while
our work focuses on classifying overall shape complexity.
1.2 Our contribution
The purpose of this work is to lay the foundation for a definition, or collection
of definitions, of shape complexity that captures the full range of its natural
dimensions. To do so, we explore several classical complexity measures and
introduce a skeleton-based measure. Our main contribution is the idea of a
persistent measure of shape complexity, which examines complexity as a shape
is eroded into its least complex approximation. Complexity is then captured by
a collection of values or histogram rather than a single value.
The measures we implement below divide naturally into three categories:
skeleton-based, symmetry-based, and boundary sampling measures. For the
boundary approach, we extract measures as the boundary is downsampled,
which allows us to differentiate persistent from transient sources of complex-
ity.
We begin with Section 2, which establishes the context for our exploration
and provides an initial foray into general principles of shape complexity. Section
2
3 describes the complexity measures we use, with results from clustering based
on those measures displayed in Section 4. Finally, we discuss our results and
future directions to explore in Section 5.
2 Defining complexity
As with many attributes assigned to shapes, the complexity of a given shape
seems intuitively straightforward but computationally elusive. We believe a few
principles should be satisfied in any reasonable definition of 2D shape complex-
ity:
1. A circle should have the minimum complexity.
2. Adding a part to a shape that is different from all existing parts should
increase complexity.
3. A shape with parts that are self-similar should be less complex than a
shape with the same number of parts where the parts are dissimilar.
4. Shapes, together with a complexity measure, should be a partially ordered
set.
The above form an objective measure for shape complexity, independent of
a particular setting. In addition to exploring objective measures we are also
interested in definitions of complexity that are useful in a particular setting.
The next section explores the idea of domain-dependent complexity, and how
that might result in different definitions or representations.
2.1 Qualitative complexity
We outline here two possible qualitative methods of defining what complexity
means, and how those methods might relate to specific tasks or applications.
In this we take the view that defining a complexity measure is not completely
independent of what you are using that measure for. For example, one measure
might capture how difficult a shape is to build, while another might better
capture how difficult it is for a human to recognize or classify a set of shapes.
2.1.1 Construction definition
This definition derives from the idea of procedural modeling [MWZ+14, SW15]
— specifying a sequence of operations that create the shape in question. Com-
plexity is a measure of how deep the tree of operations has to be; conversely,
the further down the tree you go the more complex the object. Obviously, this
depends on the set of available operators. We outline here a plausibly complete
set of operators and a few examples of how a shape could be constructed from
those operators. Note that there may not be a single, unique set of operations
for each shape.
3

Pocket





Figure 1: A possible set of construction operators. This set is complete, but
possibly over-complete.
Potential tasks or applications include shape representation, modeling, func-
tional shape comparison, and part decomposition. This is also a representational
definition, in that the encoding of the shape as a tree of operations means that
the shape can be recovered from the complexity measure. The definition is also
partially semantic, in that the operators form a language for describing shape.
The operators are illustrated in Figure 2.1.1. The assumption is that the
base shape is a circle. The operators are:
Protrusions: Add one (or a pattern of) protrusions/extrusions.
Pockets: Add one (or a pattern of ) pockets or indents
Add handle: Add a handle/join a protrusion to the shape/close off a
pocket.
Bends: Bend or curve a protrusion or pocket to re-shape it.
Sharpen or soften: Change the curvature along the boundary to produce
a feature (such as a corner or edge).
Detail pattern: Add a regular or semi-regular pattern to some (or all) of
the boundary at a small scale (relative to the overall size of the shape).
In Figure 2.1.1 we show an example of applying the operators to produce one
of our test shapes. Note that there is a natural mapping between the operators
and skeleton editing.
4
   

Pocket 

Figure 2: Example construction sequence of two shapes from our data set. Note
that the construction order need not be unique — the heart top, for example,
can be seen as either a pocket or two protrusions.
2.1.2 Predictability definition
This definition is very perceptual or experience based. The essential notion is
that complex shapes both surprise us and require more thought or effort to re-
member well enough to reproduce. For example, although a drawing of a horse
is fairly complex from a total curvature or parts standpoint, it is easily mapped
to a canonical image of a horse in our head. Similarly, a shape might consist
of self-similar or repeating patterns (for example, a Julia set or a symmetric
shape with a pattern added to the boundary) in which case, while the geom-
etry is complex, the mental image of it is not. On the other hand, a random
pattern (such as the blob in the top row of Figure 11) might be fairly simple
geometrically, but complex because it does not map to any template image in
our head.
Essentially, the complexity is a measure of how much the given shape de-
viates from a canonical set of template shapes. Possible applications or tasks
related to this definition are template matching, shape classification, and sta-
bility of shape.
2.2 Evaluating quantitative measures of complexity
In the previous sections we presented both existing quantitative, global measures
of complexity from the literature, and two hypothetical qualitative measures of
complexity. In Section 3 we will define several families of complexity measures
using the notion of sampling and resulting histograms. The questions that link
our qualitative definitions to our quantitative ones are: How do we identify
5
quantitative measures that match our qualitative measures of complexity? How
do measures differ in how they define complexity? In this work, we use clustering
as a first step to answering these questions.
It is likely that complexity is not a 1D, well-defined metric, which means
that simply ranking shapes from low-complexity to high-complexity with one
global measure may not be appropriate, particularly if we want to use multi-
ple measures to capture different aspects of complexity. We therefore perform
unsupervised clustering on a particular measure or set of measures. An effec-
tive measure, or set of measures, is one which consistently groups shapes with
certain recognizable complexity features with a recognizable similarity in com-
plexity level.
In the following section we describe our shape dataset and define our com-
plexity measures. In Section 4 we describe our clustering algorithm and display
the resulting clusters.
3 Measures
Given a particular quantitative measure, we can apply it either to the entire
shape to obtain a global measure, or to the shape at different sampling rates
to determine the scales at which certain values of the measure persist. We
take as global measures four quantities derived from the medial axis represen-
tation [MGM05, Leo07, LCLJ11], which are computed on each medial branch
and three quantities measured on the boundary curve. The measures taken at
different sampling rates, so-called sampling measures, are quantities that cap-
ture the changes in the shape as the resolution of the shape is decreased. We
employ two methods for downsampling the shape: boundary point subsampling
and image-based subsampling.
Although the boundary of the shape and its interior are linked, some mea-
sures focus primarily on measuring how the boundary changes (e.g., curvature)
while others are more concerned with the interior shape (covering the shape
with a set of ellipses). Some, such as the medial axis measures, capture both.
We store the sampling measures and the medial axis measures taken across
branches as histograms (n= 10 bins), rather than computing statistics such as
the mean or measuring entropy [CS05], to preserve the variation in the data.
3.1 Shape data base
We use the MPEG-7 database [Bob01] taking one shape from each of the 70
classes, as well as all shapes in a devices class that allows us to understand how
complexity varies as a very regular shape is perturbed in different ways. All
shapes are represented as a single closed curve, ignoring interior detail. The orig-
inal images are 256 by 256 with the boundary curves extracted for [LMHC16].
This results in around 500 points per shape. We resample the boundary curves
using arclength parameterization and normalize the scale of the shape to fit in
the unit square.
6
To calculate the curvatures we use the formulation in [MS08] which is based
on the angle change over the length. The arclength parameterization and scaling
to the unit square together keep the curvatures in roughly the range -100 to 100.
3.2 Global measures
We include four global measures from the literature: Ratio of perimeter length
to area, total curvature [Leo07], object symmetries [MGM05], and salience
measures on the medial axis such as erosion thickness and shape tubularity
[LCLJ11]. From the domain-independent standpoint, ratio of perimeter length
to area, symmetries, and the medial axis measures capture in various ways how
far a shape has deviated from a circle. The medial axis and symmetry measures
also capture self-similarity of parts. From the domain-dependent standpoint,
total curvature captures the bending and sharpen/soften operations, the ratio
of perimeter length to area captures protrusions, pockets, and handles, and the
medial axis measures capture an aspect of each of the operations. Predictabil-
ity is captured in a simplistic way by object symmetries, but requires a full
probabilistic study of shape occurrence that we do not take up here.
3.2.1 Skeleton measures
We compute the Blum medial axis corresponding to the boundaries of the shapes
at the full sampling rate, and compute several salience measures on the axis.
The Extended Distance Function (EDF) measures the depth of each medial
point within the shape, where depth is defined as the second longest geodesic
distance along the axis from a given point to the boundary curve [LCLJ11].
The Weighted Extended Distance Function (WEDF) measures another kind of
depth, where the depth is the area of the part of the shape subtended by a
particular medial point [LMB+15]. The Erosion Thickness (ET) measures how
blob-like a shape is at a particular point [LCLJ11] and the Shape Tubularity
(ST) measures how tube-like it is [LCLJ11]. The values of these measures vary
continuously across primary branch points, but will exhibit discontinuities mov-
ing from a primary branch to a secondary branch. In this way, the measures
taken together provide a picture of the morphology and relative size of the var-
ious shape parts comprising the shape [LMHC16]. See Figure 3 for an example
of a hand contour and the variation of its ST, ET, WEDF, and EDF values.
We sample the WEDF, EDF, ST, and ET at medial branch points and their
neighbors, as these are the locations of the discontinuities that characterize the
shape parts. We normalize the histograms of ET and ST, but leave the WEDF
and EDF histograms as raw counts in order to preserve information about the
number and types of branches within the shape. Note that when we combine
all measures together, we normalize EDF and WEDF to prevent their values
from dominating the clustering.
7
Figure 3: Heat maps showing values of measures taken from the medial axis.
From left to right: EDF, a length-based depth; WEDF, an area-based depth;
ET, a measure of how blob-like a region is; and ST, a measure of how tube-like
a region is.
8
Figure 4: Downsampling the boundary. Top row, from left to right: Decreasing
the number of points used to represent the boundary (500, 100, 50, 25 and 8).
Bottom row: Shifting the starting point by 4 for the 4th image above (5 shifts
total).
3.3 Sampling-based measures
We employ two sampling approaches, one based on downsampling the boundary
itself and the other downsampling the area enclosed by the boundary. We first
describe the downsampling approaches then the measures we calculate on the
downsampled data. In general, boundary downsampling reduces the complexity
of the boundary by reducing the number of points used to represent it (not
smoothing), while area downsampling uses bigger and bigger blocks to represent
the area coverage of the shape.
3.3.1 Boundary downsampling
To downsample the boundary, we linearly approximate the boundary using a
reduced number of points (5 levels - 500, 100, 50, 25 and 8). For our shapes
this represents the spectrum from roughly full resolution to a convex shape (see
Figure 4). We use arclength sampling, dividing the original curve into equal-
length segments. Because the starting point influences the downsampled shape
we compute multiple downsampled versions, each with a shifted starting point
(by 4 points). We then average (values) or sum (histograms) the measures
as appropriate for each of the downsampled versions at that level. Table 1
summarizes the downsampling values.
When downsampling we also record which segment corresponds to each orig-
inal point in order to calculate the distance from the original points to the
down-sampled curve.
3.3.2 Area downsampling
We begin by scan-converting the original boundary curve into a 256×256 image
Iwith 16 pixels of padding on all sides. We downsample by placing a grid with
an npixel neighborhood (n[2,4,8,16]) on top of I. We store two values per
grid cell: a binary value indicating whether or not the grid cell overlaps Iat all,
9
Boundary
No. Points 500 100 50 25 8
No. Shifts 1 2 3 5 16
Grid
No. Pixels 128X128 64X64 32X32 16X16
No. Shifts 1 2 4 8
Table 1: Sampling rates. The Grid pixel count does not include an extra padding
of 1 pixel around the boundary.
Figure 5: Downsampling using a grid. The largest image is the original 256X256
image. Images are downsampled, with a shift added (all eight images for the
lowest resolution are shown enlarged).
and the percentage of the overlap. As in the boundary case, we shift the grid
starting point (2 pixels in each direction) and combine the results (see Table 1.
Resulting images are shown in Figure 5.
3.3.3 Boundary measures
We normalize the measures, where appropriate, by dividing by the same quantity
associated to the original, fully sampled shape. Continuous values are stored
as histograms. Denote by dithe distance from each of the original points to
the closest point on the corresponding segment of a downsampled boundary.
Three measures use the Voronoi diagram and Delaunay triangulation, which
are computed from the downsampled boundary. More precisely, we calculate:
Length of downsampled boundary / Length of original boundary. This
captures the depth of protrusions and pockets and how many there are.
Area enclosed by the downsampled boundary / Area of original boundary.
This captures the area of protrusions and pockets.
10
L2norm on the approximation error produced by down-sampling: mean
of di. This captures the average distance of the original shape boundary
from successively simpler (more convex) versions of itself.
Hausdorff norm on the approximation error produced by downsampling:
max of di. This captures the maximum deviation of the original shape
boundary from successively simpler (more convex) versions of itself.
Distribution of di(histogram). Bins are evenly spaced between 0 and
0.1, then at 0.5 and the maximum of di. The captures the variability in
distances between the original shape and successively simpler versions of
itself.
Distribution of curvature (histogram). Bins are evenly spaced after taking
arc-tan (so-called Shape Index [KvD92]). This captures the persistence of
high curvature regions as the shape is simplified.
Distribution of edge lengths in the Voronoi diagram. This provides a
measure of the average aspect ratios of shape parts.
Distribution of triangle areas in the Delaunay triangulation. This captures
the percentage of smaller components to larger components in the shape.
Percentage of Voronoi cell centers that lie inside the shape versus outside.
This is a measure of how non-convex a shape remains as it is simplified.
The boundary measures, in general, provide information about the size,
number, shape, and depth of protrusions. By using downsampling we can ad-
ditionally capture how “stable” those protrusions are — i.e., how quickly they
disappear.
3.3.4 Moment measures
In order to capture a coarse self-similarity measure, we compute the seven com-
plex moment invariants described in [MGM05] at the different boundary down-
sampling scales. Taken together, these moments capture all rotational symme-
tries. Note that these will not reflect the addition operations such as protrusions,
as long as they are added along all axes of symmetry; however, any additions
that are not symmetric will change the moment measures quite drastically. In
other words, the moment measures capture an aspect of the self-similarity of
the parts comprising a shape, one of our domain-independent complexity tests.
3.3.5 Area measures
We store the area of the occupied downsampled image over the occupied area in
the original image. Note that this measure tends to increase while the area based
on downsampling the boundary tends to decrease, and is potentially better at
capturing the shape of pockets versus protrusions, and also their alignment with
respect to the overall shape. We also store the distribution of occupied pixels
as a histogram.
11
4 Clustering
In order to interpret what the different complexity measures capture, we perform
k-medoids clustering using each set of measures as feature vectors. K-medoids
is a distance-based, unsupervised learning algorithm that groups the data set
into kclusters. Unlike its sister, the k-means algorithm, k-medoids requires
that the centroid of each cluster (the point to which distances are being com-
puted and compared) be a sample from the data set. In many applications this
forcing of the centroid to be a sample from the data allows for more significant
interpretations.
The standard Euclidean distance is used to determine cluster membership
in the clustering results. For all clusterings illustrated we have let k= 6. This
number was determined based on comparing the average within-cluster distance
sum across clusterings produced for kranging from 1 to 15. When kis equal
to the number of points in the data set the within-cluster distance sum is zero
since each point is its own centroid. Reasonable values of kcan be identified by
looking for an elbow in the graph of average within-cluster distance sum versus
k. For the different feature vectors considered, the elbow occurred between k= 3
and k= 7 in the majority of trials. For the sake of comparing the results across
different feature vectors more smoothly we fixed k= 6. We note that there
are more complex ways of identifying “optimal” cluster numbers that consider
both the within-cluster distance and across-cluster separation. However, for our
purposes the within-cluster measure appears sufficient. In future work a more
thorough analysis may be performed, allowing for different numbers of clusters
for each set of feature vectors.
Results of the clusterings shown herein are based on running 50 trials of
k-medoids for each feature vector set and selecting the one with the smallest
within-cluster distance sum. Figures 8, 12, 14, 10, 6, 16, and 18 show the 6
different medoids chosen for each of the feature vector sets. These medoids
are color-coded to provide cluster membership information in the full clustering
results shown in Figures 9, 13, 15, 11, 7, 17, and 19. In these figures the medoid
of each cluster is the first shape for each new cluster color and has been given a
black border. The subsequent shapes in a given color are ordered from closest
to farthest from the medoid as measured by Euclidean distance.
The different sets of feature vectors considered are produced as follows. We
group the features according to their appearance in Section 3.
Global Measures
– Non-skeleton global feature vectors: The global clusters use a
feature vector that is 12-dimensional. Included in the feature vector
are the seven moment measures computed at the full scale of the im-
age and also boundary length to area ratio, total curvature, curvature
entropy, angular entropy, and distance to original entropy. This set
of feature vectors are based exclusively on the complete sampling of
the shape and complexity values that can be found in existing litera-
12
ture. Figures 6 and 7 show the cluster centroids and entire clustering
based on the global feature vectors.
– Skeleton feature vectors: The skeleton feature vectors are com-
posed of histograms of the ET and ST values at all branch points
and neighbors of branch points. Separate histograms are produced
for ET, ST, and each point type. The ET/ST values are all normal-
ized relative to the max ET/ST for the specific shape and then the
histogram is normalized based on the total number of points (branch
or neighbor points) being considered. EDF and WEDF are also com-
puted for branch points and neighbor points. The values are scaled
relative to the maximum value of EDF/WEDF, but the number of
points are not normalized in order to preserve information about the
number of branches in each bin. This results in 40-dimensional skele-
ton based feature vector. The cluster centroids for the clustering
built on the skeleton feature vectors are shown in Figure 8 and the
full clustering is shown in Figure 9.
Sampling Measures
Boundary feature vectors: Boundary features are all computed at
the five different downsampling scales. Four values are histogrammed
into ten bins for each value as described in Section 3.3.3. Another
four values are computed at each scale, also listed in Section 3.3.3.
Consequently, the boundary feature vectors are 220-dimensional (5
scales x (4 histogrammed values x 10 bins per histogram + 4 non-
histogrammed values)=220). Results of clustering on the boundary
feature vectors and the cluster centroids are contained in Figures 11
and 10 respectively.
– Moment feature vectors: The moment measures considered are
the seven values described in [MGM05]. These seven values are com-
puted at the five different boundary down-sampling scales. This re-
sults in a 35-dimensional feature vector for each shape that we use to
cluster on. Cluster centroids for clustering on the moment measure
feature vectors are shown in Figure 12 and the complete clustering
is shown in Figure 13.
– Coverage feature vectors: The coverage feature vector is 44-
dimensional and consists of one non-histogrammed value (area change)
and one histogrammed value (pixel occupied distribution, 10 bins)
for each of the four down-samplings. Centroids for each cluster and
the full clustering based on the coverage measure feature vectors are
provided in Figures 14 and 15, respectively.
Concatenation of Global and Sampling Measures
With moments: Clusterings based on concatenated sets of feature
vectors are shown in Figures 16, 17, 18, and 19. Figures 16 and
13
Figure 6: Mediods from clustering based on global measures.
17 show the medoids and clustering, respectively, where boundary,
coverage, moment, and normalized skeleton feature vectors are con-
catenated.
– Without moments: Alternatively, the results based on clustering
boundary, coverage, and skeleton feature vectors (leaving out mo-
ment feature vectors) produce the medoids and clusters in Figures
18 and 19. The concatenated feature vectors built by combining
the boundary, moment, coverage, and skeleton feature vectors are
339-dimensional while those without the moment feature vector are
304-dimensional.
5 Discussion
Our measures capture most of the aspects of complexity outlined in Section 2
— except for the predictive approach which we did not explore here. Indeed,
it appears that all of these aspects of complexity are required to fully capture
the range of shape complexity: the most successful clustering appears to be
the concatenated measures without moments shown in Figure 19. Interestingly,
most of the medoids of that cluster are drawn from the symmetric shape class.
In particular, clustering based on boundary feature vectors, see Figure 11, is
the only clustering that manages to group all mostly long, skinny shapes into the
same cluster: watch, bone, knife, hammer, guitar, key, fork, shoe, and phone.
The skeleton feature vectors, see Figure 9, offer several convincing clus-
ters, but fail in two fundamental ways. One, visually simple shapes such
as the rounded square and the heart are grouped with much more complex
shapes. This is because the skeleton generates multiple branches to describe
near-circular regions, making the skeletal description much more complex than
the boundary. The skeleton measures also do not group the hammer in with
the visually most similar shapes: bone, fork, and knife. On the other hand, the
skeleton feature vectors produce the only clustering in which the deer, fly, and
beetle, arguably the most complex shapes, all occur in the same grouping.
Area measures, shown in Figure 15, very successfully group together the large
low-complexity objects, except for the rounded square in a cluster by itself. As
the only shape that takes up almost the entire field, the clustering result makes
sense. Other clusters, such as the one containing shapes with long, pointy parts,
14
Figure 7: Clustering based on global measures.
Figure 8: Mediods from clustering based on skeleton measures.
15
Figure 9: Clustering based on skeleton measures.
Figure 10: Mediods from clustering based on boundary measures.
16
Figure 11: Clustering based on boundary measures.
Figure 12: Mediods from clustering based on moments measures.
17
Figure 13: Clustering based on moments measures.
Figure 14: Mediods from clustering based on area measures.
18
Figure 15: Clustering based on area measures.
19
Figure 16: Mediods from clustering based on concatenated boundary, moment,
coverage, and skeleton feature vectors.
also seem appropriate. On the other hand, the complex beetle is in a cluster
with the much simpler fork instead of in the cluster with the fly and deer.
Boundary measures, meanwhile, cluster the beetle and the deer together but
group the fly with the less complex four-legged animals; see Figure 11. Boundary
measures do, however, cluster simple shapes like the rounded square and hat
together, and also cluster symmetric shapes together despite having no explicit
symmetry measure.
The global measures in Figure 7 cluster the symmetric shapes together as
well, though they include the moment invariants, but also fail to cluster the high-
complexity beetle, fly, and deer together. At first glance, the global feature
vectors appear to cluster all of the simpler animals together, but the dog is
grouped with flowers and with detailed petals. It may be that the moment
invariants in some cases overwhelm the more geometric measures.
The combination of multiple feature vectors via concatenation seems to cap-
ture the best clustering characteristics of the individual feature vector sets. The
moments appear to distract, however, as shown in Figure 17. The single cluster
containing the spiral shape, the deer, and the grouping of some of the simplest
shapes (rounded triangle and semi-circle) with some of the most complex (beetle
and fly) are some of the most obvious inconsistencies.
Combined multiple feature vectors without the moment invariants (bound-
ary, coverage, and normalized skeleton feature vectors) produces the most in-
tuitively correct results; see Figure 19. Simpler shapes with long, slender pro-
trusions (except for the deer) are grouped together, as are the simplest shapes
such as heart and rounded square.
In short, each of the types of measures provides important information about
a shape’s complexity and therefore the aggregate results are the most accurate.
Moreover, our results suggest that measuring complexity measures at multiple
scales captures something important. Finally, we note that while clustering
does not provide an ordering of shapes based on their complexity, we can attach
complexity orders to the clusters based on membership in a cluster of a shape
with a known complexity level.
5.1 Discussion of methodology and future work
Our overall goal is to explore the qualitative effect of possible quantitative mea-
sures when those measures are no longer single numbers (and hence the shapes
20
Figure 17: Clustering based on concatenated boundary, moment, coverage, and
skeleton feature vectors.
21
Figure 18: Mediods from clustering based on concatenated boundary, coverage,
and skeleton feature vectors.
Figure 19: Clustering based on concatenated boundary, coverage, and skeleton
feature vectors.
22
cannot be linearly sorted). Clustering by similarity allows us to visually exam-
ine the measures on actual shapes. We present six different combinations of
measures here, each with six clusters in order to illustrate the approach.
There are many different ways to vary this approach: 1) Use a broader
range of shapes, 2) Use non-linear clustering methods (the Euclidean distance
may not be adequate), 3) Use further analysis to determine which features are
the most important in any given clustering. Further exploratory analysis of this
kind could yield a “visual dictionary” that qualitatively defines the different
measures, and a better understanding of how to mathematically combine them
to produce desired groupings. Additionally, we have not yet analyzed which
features are most significant in determining cluster membership.
It also seems clear that our choice of 6 clusters might not be ideal for each
measure. We chose this number to ensure that the splitting is conservative —
i.e., we may have two clusters that are similar, but we will not have a cluster that
should be split further. This is a possible explanation for why some shapes that
seem similar (such as the devices) end up in two or more clusters. It should be
noted that decreasing the number of clusters will not, in general, simply merge
two clusters. Instead, we plan to explore the optimal number of clusters in each
instance separately in future work.
Finally, we intend to implement an edit-based complexity measure to cap-
ture more explicitly the process displayed in Figure 2.1.1. We may also begin
addressing probabilistic constructions that will allow us to explore the predictive
approach to complexity.
Another key direction to explore is supervised learning, where we are pro-
vided with desired clusters (or relative rankings) and the goal is to find the
combination of measures that best produce those clusters or rankings. The rich
set of measures provided here (and their ability to produce different qualitative
groupings) show promise for supporting this endeavor. For example, one direc-
tion we would like to explore is using user studies (such as the one in [CLH+16])
to produce labeled data, then using techniques such as Rank Support Vector
Machines[Joa06] to determine how to combine the measures in order to produce
similar relative rankings. Note that this will provide the complexity ordering
that our clustering method currently does not.
References
[Bob01] M. Bober. Mpeg-7 visual shape descriptors. IEEE Trans. Cir. and
Sys. for Video Technol., 11(6), June 2001.
[CI84] B. Chazelle and J. Incerpi. Triangulation and shape-complexity.
ACM Trans. Graph., 3(2):135–152, April 1984.
[CLH+16] Axel Carlier, Kathryn Leonard, Stefanie Hahmann, Geraldine
Morin, and Misha Collins. The 2d shape structure dataset: A user
annotated open access database. Computers & Graphics, 58:23–30,
2016.
23
[CS05] Yinpeng Chen and Hari Sundaram. Estimating the complexity of 2d
shapes. In Proc. Multimedia Signal Processing Workshop, October
2005.
[JF05] M.Singh J. Feldman. Information along contours and object bound-
aries. Psychological Reviews, 112(1):243–252, 2005.
[Joa06] Thorsten Joachims. Training linear svms in linear time. In Pro-
ceedings of the 12th ACM SIGKDD International Conference on
Knowledge Discovery and Data Mining, KDD ’06, pages 217–226,
New York, NY, USA, 2006. ACM.
[KvD92] Jan J. Koenderink and Andrea J. van Doorn. Surface shape and
curvature scales. Image Vision Comput., 10:557–565, October 1992.
[LCLJ11] Lu Liu, Erin W Chambers, David Letscher, and Tao Ju. Extended
grassfire transform on medial axes of 2d shapes. Computer-Aided
Design, 43(11):1496–1505, 2011.
[Leo07] Kathryn Leonard. Efficient shape modeling: epsilon-entropy, adap-
tive coding, and boundary curves -vs- blum’s medial axis. Interna-
tional Journal of Computer Vision, 74(2):183–199, 2007.
[LMB+15] Lisa J Larsson, eraldine Morin, Antoine Begault, Rapha¨elle
Chaine, Jeannine Abiva, Evelyne Hubert, Monica Hurdal, Mao Li,
Beatriz Paniagua, Giang Tran, et al. Identifying perceptually salient
features on 2d shapes. In Research in Shape Modeling, pages 129–
153. Springer, 2015.
[LMHC16] K. Leonard, G. Morin, S. Hahmann, and A. Carlier. A 2d shape
structure for decomposition and part similarity. In International
Conference on Pattern Recognition, Dec 2016.
[MGM05] Muharrem Mercimek, Kayhan Gulez, and Tarik Veli Mumcu. Real
object recognition using moment invariants. Sadhana, 30(6):765–
775, 2005.
[MS08] J. McCrae and K. Singh. Sketching piecewise clothoid curves. In
Proceedings of the Fifth Eurographics Conference on Sketch-Based
Interfaces and Modeling, SBM’08, pages 1–8, Aire-la-Ville, Switzer-
land, Switzerland, 2008. Eurographics Association.
[MWZ+14] Niloy J. Mitra, Michael Wand, Hao Zhang, Daniel Cohen-Or,
Vladimir Kim, and Qi-Xing Huang. Structure-aware shape process-
ing. In ACM SIGGRAPH 2014 Courses, SIGGRAPH ’14, pages
13:1–13:21, New York, NY, USA, 2014. ACM.
[OFCD02] Robert Osada, Thomas Funkhouser, Bernard Chazelle, and David
Dobkin. Shape distributions. ACM Trans. Graph., 21(4):807–832,
October 2002.
24
[PA16] Costas Panagiotakis and Antonis Argyros. Parameter-free mod-
elling of 2d shapes with ellipses. Pattern Recognition, 53:259 – 275,
2016.
[PKS+03] D. L. Page, A. F. Koschan, S. R. Sukumar, B. Roui-Abidi, and
M. A. Abidi. Shape analysis algorithm based on information theory.
In International Conference on Image Processing, pages 229–232,
September 2003.
[RFS05] Jaume Rigau, Miquel Feixas, and Mateu Sbert. Shape complexity
based on mutual information. In 2005 International Conference
on Shape Modeling and Applications (SMI 2005), 15-17 June 2005,
Cambridge, MA, USA, pages 357–362, 2005.
[SPG+06] S. Sukumar, D. Page, A. Gribok, A. Koschan, and M. Abidi. Shape
measure for identifying perceptually informative parts of 3d objects.
In 3D Data Processing, Visualization, and Transmission, Third In-
ternational Symposium on, pages 679–686, June 2006.
[SW15] Michael Schwarz and Peter Wonka. Practical grammar-based pro-
cedural modeling of architecture: Siggraph asia 2015 course notes.
In SIGGRAPH Asia 2015 Courses, SA ’15, pages 13:1–13:12, New
York, NY, USA, 2015. ACM.
25
... For the case of two-dimensional shapes, previous definitions for shape complexity used approximations of the Kolmogorov complexity [5] or the entropy in local features [14]. Other work defined strict criteria with circles being the least complex and adding parts that are different from the existing parts of the shape should increase the complexity among others [4] [13]. [4] applied a range of measures grouped into boundary-based, regional, and skeletal methods [4] [1]. ...
... Other work defined strict criteria with circles being the least complex and adding parts that are different from the existing parts of the shape should increase the complexity among others [4] [13]. [4] applied a range of measures grouped into boundary-based, regional, and skeletal methods [4] [1]. These measures are evaluated with a user study [1] and the authors listed three measures capturing most of the complexity information according to the users: boundary, convexity and skeleton. ...
... Other work defined strict criteria with circles being the least complex and adding parts that are different from the existing parts of the shape should increase the complexity among others [4] [13]. [4] applied a range of measures grouped into boundary-based, regional, and skeletal methods [4] [1]. These measures are evaluated with a user study [1] and the authors listed three measures capturing most of the complexity information according to the users: boundary, convexity and skeleton. ...
Preprint
Full-text available
In this paper, we compare methods for estimating the complexity of two-dimensional shapes and introduce a method that exploits reconstruction loss of Variational Autoencoders with different sizes of latent vectors. Although complexity of a shape is not a well defined attribute, different aspects of it can be estimated. We demonstrate that our methods captures some aspects of shape complexity. Code and training details will be publicly available.
... In [10], the authors identify categories of complexification -adding parts to a shape, creating indentations, adding noise to a shape boundary, and disrupting symmetry -and conclude that no single quantitative measure is likely to capture the full range of shape complexity. Instead, they propose an aggregation of measures, and explore an extensive list of possible measures grouped by whether they are local measures on the boundary of the shape, local measures on the region of the shape, measures based on the Blum medial axis of the shape [7], measures that capture self-similarity, or global shape measures. ...
... Contributions: This paper makes three main contributions. First, in Section 4, we apply results from a forced choice user study to identify which of the 282 measures from [10] correlate most strongly with human perception of complexity. We then apply those significant measures to three small datasetes of simulated shapes created to have predetermined complexity levels to see how well the selected measures distinguish between the predetermined levels. ...
... In [10], the authors explore a wide range of measures of shape complexity arising from information theory [13], computer vision [28], computational geometry [12], and curve analysis [15,27,31,20], and introduce new notions of complexity based on measurements taken along Blum medial axis [7] and persistence of certain features under down-sampling. We discuss these in more detail below. ...
Article
Full-text available
The importance of measuring the complexity of shapes can be seen by the wide range of its application such as computer vision, robotics, cognitive studies, eye tracking, and psychology. However, it is very challenging to define an accurate and precise metric to measure the complexity of the shapes. In this paper, we explore different notions of shape complexity, drawing from established work in mathematics, computer science, and computer vision. We integrate results from user studies with quantitative analyses to identify three measures that capture important axes of shape complexity, out of a list of almost 300 measures previously considered in the literature. We then explore the connection between specific measures and the types of complexity that each one can elucidate. Finally, we contribute a dataset of both abstract and meaningful shapes with designated complexity levels both to support our findings and to share with other researchers.
... Recently, developed tools for quantifying shape complexity offer a potential solution, as they do not rely on the placement of homologous landmarks. Shape complexity is distinct from the metrics of shape variation calculated by GMM and can broadly be defined as the number of "simple shapes" required to create a more complex shape, and the selfsimilarity of those composite parts (Chambers et al., 2018;Gardiner et al., 2018). Recent research has deployed shape complexity metrics to investigate a number of biological systems. ...
... This was achieved by comparing perimeters, between EFA shapes of contrasting harmonics. Comparing boundary measures, such as perimeter, is a tenet of many methods of estimating shape complexity (Chambers et al., 2018). The first metric calculated is a low perimeter ratio that here we term "gross complexity." ...
Article
Full-text available
Amblypygi is an arachnid order possessing a unique pair of spined pedipalps: appendages that perform in prey capture, courtship, and contest. Pedipalp length, hypothesized to be under sexual selection, varies markedly across amblypygid species, and pedipalp spination, thought to reflect selection for function in prey capture, also differs interspecifically. Differences in pedipalp shape between species may indicate that the relative strength of selection for prey capture and sexual selection vary across the group. However, interspecific differences in pedipalp shape have not been quantified, due to difficulties in identifying homologous features. For the first time, we quantify trends in amblypygid pedipalp shape complexity. We use elliptical Fourier analysis to quantify 2D complexity in pedipalp outlines across eleven species and six genera. We find that complexity significantly decreases as pedipalp length increases. This appears to be driven by relative spine length, suggesting that a trade‐off exists between pedipalp length and spination. Furthermore, significant female‐biased sexual dimorphism in shape complexity is present in the tibial segment of the amblypygid pedipalp. Our results provide novel insights into the drivers of amblypygid pedipalp evolution and suggest that a functional trade‐off between performance in prey capture and other functions under sexual selection exist in this enigmatic structure. Amblypygids are an arachnid order possessing a unique pair of spined pedipalps: appendages that perform in prey capture, courtship, and contest. We quantify trends in amblypygid pedipalp shape complexity using elliptical Fourier analysis applied to pedipalp outlines. Our results provide novel insights into the drivers of amblypygid pedipalp evolution and suggest that a functional trade‐off between performance in prey capture and other functions under sexual selection exist in this enigmatic structure.
... Recently developed tools for quantifying shape complexity offer a potential solution, as they do not rely on the placement of homologous landmarks. Shape complexity is distinct from the metrics of shape variation calculated by GMM, and can broadly be defined as the number of 'simple shapes' required to create a more complex shape, and the self-similarity of those composite parts (Chambers et al., 2018;Gardiner, Behnsen & Brassey, 2018). ...
... This was achieved by comparing perimeters, either between EFA shapes of contrasting harmonics. Comparing boundary measures, such as perimeter, is a tenet of many methods of estimating shape complexity (Chambers et al., 2018). The first metric calculated is a low perimeter ratio that here we term 'gross complexity'. ...
Preprint
Full-text available
Amblypygids are an arachnid order possessing a unique pair of spined pedipalps: appendages that perform in prey capture, courtship and contest. Pedipalp length, hypothesised to be under sexual selection, varies markedly across amblypygid species, and pedipalp spination, thought to reflect selection for function in prey capture, also differs interspecifically. Differences in pedipalp shape between species may indicate that the relative strength of selection for prey capture and sexual selection vary across the group. However, interspecific differences in pedipalp shape have not been quantified, due to difficulties in identifying homologous features. For the first time, we quantify trends in amblypygid pedipalp shape complexity. We use elliptical Fourier analysis to quantify 2D complexity in pedipalp outlines across eleven species and six genera. We find that complexity significantly decreases as pedipalp length increases. This appears to be driven by relative spine length, suggesting that a trade-off exists between pedipalp length and spination. Furthermore, significant female-biased sexual dimorphism in shape complexity is present in the tibial segment of the amblypygid pedipalp. Our results provide novel insights into the drivers of amblypygid pedipalp evolution, and suggest that a functional trade-off between performance in prey capture and other functions under sexual selection exist in this enigmatic structure.
... Also, it should be noted that our analyses were limited to perceptual features originally included in the L-EFT (i.e., vertical symmetry, closure, number of target lines, and number of continued lines). This set of features is not exhaustive and further studies are needed to explore the influence of other shape features, such as horizontal symmetry, diagonal symmetry, or curvature of boundaries (Chambers et al., 2018), on disembedding performance. ...
Article
Full-text available
Humans naturally perceive visual patterns in a global manner and are remarkably capable of extracting object shapes based on properties such as proximity, closure, symmetry, and good continuation. Notwithstanding the role of these properties in perceptual grouping, studies highlighted differences in disembedding performance across individuals, which are summarized by the field dependence dimension. Evidence suggests that age and educational attainment explain part of this variability, whereas the role of sex is still highly debated. Also, which stimulus features primarily influence inter-individual variations in perceptual grouping has still to be fully determined. Building upon these premises, we assessed the role of age, education level, and sex on performance at the Leuven Embedded Figure Test—a proxy of disembedding abilities—in 391 cisgender individuals. We also investigated to what extent shape symmetry, closure, complexity, and continuation relate to task accuracy. Overall, target asymmetry, closure, and good continuation with the embedding context increase task difficulty. Simpler shapes are more difficult to detect than those with more lines, yet context complexity impairs the recognition of complex targets (i.e., those with 6 lines or more) to a greater extent. Concerning demographic data, we confirm that age and educational attainment are significantly associated with disembedding abilities and reveal a perceptual advantage in males. In summary, our study further highlights the role of shape properties in disembedding performance and unveils sex differences not reported so far.
... The data are usually stored as static images of a script 11 written on paper. These approaches are slightly more conve- 12 nient since it is easier to capture the samples without temporal 13 information. There are cases where it is not possible to obtain 14 online handwriting, making offline data approaches a necessity 15 (for instance, reproducing the writing of a dead person). ...
Article
Today, people rely almost full time on digital texts. It is not surprising that handwriting earned a special status, and solutions to mimic real handwriting became attractive. A particular field called handwriting synthesis generates renderings of text which resemble natural writing but are synthesized from actual handwriting samples. The main idea behind samples’ current solutions is to collect enough samples to capture a given subject’s writing style, and therefore be able to reproduce it in new texts, with natural variability. Nevertheless, the question remains of how much input variability is enough to represent specific handwriting. In this paper, we address sample acquisition for handwriting synthesis. We conducted a study comparing written text similarity between two sets of samples, one using augmented pangrams (with a total of 473 characters) and the other using general texts (with 1586 characters). Our results show that the samples collected with pangrams are statistically equivalent in variation with samples collected using general texts, with many benefits, particularly the shorter time needed to collect the samples. We also made our data collection publicly available, providing a valuable original resource for future research.
Article
Full-text available
Differential entropy, along with fractal dimension, is herein employed to describe and interpret the shape complexity of self-similar organic islands. The islands are imaged with in situ Atomic Force Microscopy, following, step-by-step, the evolution of their shape while deposition proceeds. The fractal dimension shows a linear correlation with the film thickness, whereas the differential entropy presents an exponential plateau. Plotting differential entropy versus fractal dimension, a linear correlation can be found. This analysis enables one to discern the 6T growth on different surfaces, i.e., native SiOx or 6T layer, and suggests a more comprehensive interpretation of the shape evolution. Changes in fractal dimension reflect rougher variations of the island contour, whereas changes in differential entropy correlates with finer contour details. The computation of differential entropy therefore helps to obtain more physical information on the island shape dependence on the substrate, beyond the standard description obtained with the fractal dimension.
Chapter
Spatial complexity is defined here as the difficulty to simplify the structure or form of a 2-and-higher-dimensional surface or object. The study of spatial complexity refers to the geographical space, to mathematically abstract spaces, to physical objects, or to any surface or object, in a n-dimensional space with n equal to two or higher. Spatial complexity should not be confused with “space complexity”, “topological complexity”, “shape complexity” or “complex stystems”. Spatial complexity is scale-dependent (it changes according to the level of generalization at which it is examined and is, under certain conditions, perception-dependent also.
Article
Full-text available
Our goal is to represent a given 2D shape with an automatically determined number of ellipses, so that the total area covered by the ellipses is equal to the area of the original shape without any assumption or prior knowledge about the object structure. To solve this interesting theoretical problem, first we employ the skeleton of the 2D shape which provides important information on the parameters of the ellipses that could approximate the original shape. For a given number of such ellipses, the hard Expectation-Maximization (EM) algorithm is employed to maximise the shape coverage under the equal area constraint. Different models (i.e., solutions involving different numbers of ellipses) are evaluated based on the Akaike Information Criterion (AIC). This considers a novel, entropy-based shape complexity measure that balances the model complexity and the model approximation error. In order to minimise the AIC criterion, two variants are proposed and evaluated: (a) the augmentative method that gradually increases the number of considered ellipses starting from a single one and, (b) the decremental method that decreases the number of ellipses starting from a large, automatically defined set. The obtained quantitative results on more than 4,000 2D shapes included in standard as well as in custom datasets, quantify the performance of the proposed methods and illustrate that their solutions agree with human intuition.
Article
Full-text available
Shape structure is about the arrangement and relations between shape parts. Structure-aware shape processing goes beyond local geometry and low level processing, and analyzes and processes shapes at a high level. It focuses more on the global inter and intra semantic relations among the parts of shape rather than on their local geometry. With recent developments in easy shape acquisition, access to vast repositories of 3D models, and simple-to-use desktop fabrication possibilities, the study of structure in shapes has become a central research topic in shape analysis, editing, and modeling. A whole new line of structure-aware shape processing algorithms has emerged that base their operation on an attempt to understand such structure in shapes. The algorithms broadly consist of two key phases: an analysis phase, which extracts structural information from input data; and a (smart) processing phase, which utilizes the extracted information for exploration, editing, and synthesis of novel shapes. In this course, we will organize, summarize, and present the key concepts and methodological approaches towards efficient structure-aware shape processing. We discuss common models of structure, their implementation in terms of mathematical formalism and algorithms, and explain the key principles in the context of a number of state-of-the-art approaches. Further, we attempt to list the key open problems and challenges, both at the technical and at the conceptual level, to make it easier for new researchers to better explore and contribute to this topic. Our goal is to both give the practitioner an overview of available structure-aware shape processing techniques, as well as identify future research questions in this important, emerging, and fascinating research area.
Article
Full-text available
This paper describes a new method for triangulating a simple n-sided polygon. The algorithm runs in time O(n log s), with s≤n. The quantity s measures the sinuosity of the polygon, that is, the number of times the boundary alternates between complete spirals of opposite orientation. The value of s is in practice a very small constant, even for extremely winding polygons. Our algorithm is the first method whose performance is linear in the number of vertices, up to within a factor that depends only on the shape-complexity of the polygon. Informally, this notion of shape-complexity measures how entangled a polygon is, and is thus highly independent of the number of vertices. A practical advantage of the algorithm is that it does not require sorting or the use of any balanced tree structure. Aside from the notion of sinuosity, we are also able to characterize a large class of polygons for which the algorithm can be proven to run in O(n log log n) time. The algorithm has been implemented, tested, and empirical evidence has confirmed its theoretical claim to efficiency.
Article
Full-text available
Moments and functions of moments have been extensively employed as invariant global features of images in pattern recognition. In this study, a flexible recognition system that can compute the good features for high classification of 3-D real objects is investigated. For object recognition, regardless of orientation, size and position, feature vectors are computed with the help of nonlinear moment invariant functions. Representations of objects using two-dimensional images that are taken from different angles of view are the main features leading us to our objective. After efficient feature extraction, the main focus of this study, the recognition performance of classifiers in conjunction with moment-based feature sets, is introduced
Article
Full-text available
We propose efficiency of representation as a criterion for evaluating shape models, then apply this criterion to compare the boundary curve representation with the medial axis. We estimate the ⋮-entropy of two compact classes of curves. We then construct two adaptive encodings for non-compact classes of shapes, one using the boundary curve and the other using the medial axis, and determine precise conditions for when the medial axis is more efficient. Finally, we apply our results to databases of naturally occurring shapes, determining whether the boundary or medial axis is more efficient. Along the way we construct explicit near-optimal boundary-based approximations for compact classes of shapes, construct an explicit compression scheme for non-compact classes of shapes based on the medial axis, and derive some new results about the medial axis.
Conference Paper
Full-text available
This paper deals with the problem of estimating 2D shape complexity. This has important applications in computer vision as well as in developing efficient shape classification algorithms. We define shape complexity using correlates of Kolmogorov complexity-entropy measures of global distance and local angle, and a measure of shape randomness. We tested our algorithm on synthetic and real world datasets with excellent results. We also conducted user studies that indicate that our measure is highly correlated with human perception. They also reveal an intuitive shape sensitivity curve-simple shapes are easily distinguished by small complexity variations, while complex shapes require significant complexity differences to be differentiated
Article
In this paper we present the 2D Shape Structure database, a public, user-generated dataset of 2D shape decompositions into a hierarchy of shape parts with geometric relationships retained. It is the outcome of a large-scale user study obtained by crowdsourcing, involving over 1200 shapes in 70 shape classes, and 2861 participants. A total of 41953 annotations has been collected with at least 24 annotations per shape. For each shape, user decompositions into main shape, one or more levels of parts, and a level of details are available. This database reinforces a philosophy that understanding shape structure as a whole, rather than in the separated categories of parts decomposition, parts hierarchy, and analysis of relationships between parts, is crucial for full shape understanding. We provide initial statistical explorations of the data to determine representative (" mean ") shape annotations and to determine the number of modes in the annotations. The primary goal of the paper is to make this rich and complex database openly available (through the website http://2dshapesstructure.github.io/index.html), providing the shape community with a ground truth of human perception of holistic shape structure.
Conference Paper
This course provides a comprehensive, in-depth introduction to procedural modeling of architecture using grammar-based approaches. It first presents all necessary fundamentals and discusses the various advanced features of grammar languages in detail. Subsequently, context sensitivity, which is crucial for many practical tasks, and the different forms of support for it are addressed extensively. The course then looks into several further advanced aspects, such as local edits or GPU-based variants, and finally explores related inverse procedural modeling approaches. Elements from a large body of work are covered and presented in a coherent, structured way. The course explores the range of solution approaches, provides examples, and identifies limitations; it also highlights and investigates practical problem cases. The course is useful for practitioners and researchers from many different domains, ranging from urban planning, geographic information systems (GIS) and virtual maps to movies and computer games, with interests ranging from content creation to grammar-based procedural approaches in general. They learn about the arsenal of available techniques and obtain an overview of the field, including more recent developments. The audience benefits from a coherent treatment of ideas, concepts, and techniques scattered across many (sometimes lesser-known) publications and systems. This course helps in developing a realistic understanding of what can be done with current solutions, how difficult and practical that is, and with which tasks existing approaches cannot cope.
Chapter
Maintaining the local style and scale of 2D shape features dur- ing deformation, such as when elongating, compressing, or bending a shape, is essential for interactive shape editing. To achieve this, a necessary first step is to develop a robust classification method able to detect salient shape features, if possible in a hierarchical manner. Our aim is to overcome the limitations of existing techniques, which are not always able to detect what a user immediately identifies as a shape feature. Therefore, we first conduct a user study enabling us to learn how shape features are perceived. We then propose and compare several algorithms, all based on the medial axis trans- form or similar skeletal representations, to identify relevant shape features from this perceptual viewpoint. We discuss the results of each algorithm and compare them with those of the user study, leading to a practical solution for computing hierarchies of salient features on 2D shapes.