Access to this full-text is provided by PLOS.
Content available from PLOS ONE
This content is subject to copyright.
RESEARCH ARTICLE
Materials In Paintings (MIP): An
interdisciplinary dataset for perception, art
history, and computer vision
Mitchell J. P. Van ZuijlenID
1
*, Hubert Lin
2
, Kavita Bala
2
, Sylvia C. Pont
1
, Maarten W.
A. Wijntjes
1
1Perceptual Intelligence Lab, Delft University of Technology, Delft, The Netherlands, 2Computer Science
Department, Cornell University, Ithaca, New York, United States of America
*m.j.p.vanzuijlen@tudelft.nl
Abstract
In this paper, we capture and explore the painterly depictions of materials to enable the
study of depiction and perception of materials through the artists’ eye. We annotated a data-
set of 19k paintings with 200k+ bounding boxes from which polygon segments were auto-
matically extracted. Each bounding box was assigned a coarse material label (e.g., fabric)
and half was also assigned a fine-grained label (e.g., velvety, silky). The dataset in its
entirety is available for browsing and downloading at materialsinpaintings.tudelft.nl. We
demonstrate the cross-disciplinary utility of our dataset by presenting novel findings across
human perception, art history and, computer vision. Our experiments include a demonstra-
tion of how painters create convincing depictions using a stylized approach. We further
provide an analysis of the spatial and probabilistic distributions of materials depicted in
paintings, in which we for example show that strong patterns exists for material presence
and location. Furthermore, we demonstrate how paintings could be used to build more
robust computer vision classifiers by learning a more perceptually relevant feature represen-
tation. Additionally, we demonstrate that training classifiers on paintings could be used to
uncover hidden perceptual cues by visualizing the features used by the classifiers. We con-
clude that our dataset of painterly material depictions is a rich source for gaining insights
into the depiction and perception of materials across multiple disciplines and hope that the
release of this dataset will drive multidisciplinary research.
Introduction
Throughout art history, painters have invented numerous ways to depict the three-dimen-
sional world onto flat surfaces [1–4]. Unlike photographers, painters are not limited to optical
projection [5,6] and therefore paintings have more freedom. This means that a painter can
directly modify and manipulate the 2D image features of the depiction. When doing so, a pain-
ter’s primary concern is not whether a depiction is optically or physically correct. Instead, a
painting is explicitly designed for human viewing [7,8]. The artist does not copy a retinal
PLOS ONE
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 1 / 30
a1111111111
a1111111111
a1111111111
a1111111111
a1111111111
OPEN ACCESS
Citation: Van Zuijlen MJP, Lin H, Bala K, Pont SC,
Wijntjes MWA (2021) Materials In Paintings (MIP):
An interdisciplinary dataset for perception, art
history, and computer vision. PLoS ONE 16(8):
e0255109. https://doi.org/10.1371/journal.
pone.0255109
Editor: Omar Sultan Al-Kadi, University of Jordan,
JORDAN
Received: December 4, 2020
Accepted: July 9, 2021
Published: August 26, 2021
Copyright: ©2021 Van Zuijlen et al. This is an open
access article distributed under the terms of the
Creative Commons Attribution License, which
permits unrestricted use, distribution, and
reproduction in any medium, provided the original
author and source are credited.
Data Availability Statement: The data underlying
this study are available on https://data.4tu.nl/
(https://data.4tu.nl/articles/dataset/Materials_In_
Paintings_MIP_An_interdisciplinary_dataset_for_
perception_art_history_and_computer_vision/
13679200).
Funding: Mitchell van Zuijlen, Maarten Wijntjes,
and Sylvia Pont were financed by the Netherlands
Organization for Scientific Research with the VIDI
project “Visual communication of material
properties”, number 276.54.001. Hubert Lin and
image [9] (which would make the painter effectively a biological camera) but may apply tech-
niques such as iteratively adapting templates until they ‘fit’ perceptual awareness [10].
As a result of this, the depiction contained can deviate from reality [6]. On one hand, this
makes paintings unsuited as ecological stimulus [11]. On the other hand, as Gibson acknowl-
edges, paintings are the result of endless visual experimentation, and therefore, indispensable
for the study of visual perception.
The depiction and perception of pictorial space in paintings [1–5] has historically received
more attention than the depiction and perception of materials. It has previous been found that
human observers are able to visually categorize and identify materials accurately and quickly
for both photos [12–14] and paintings [15]. Furthermore, for these painted materials, we can
perceive distinct material properties such as glossiness, softness, transparency, etc [15–17]. A
single material category (e.g., fabric) can already display a large variety of these material prop-
erties, which demonstrates the enormous variation in visual appearance of materials. This vari-
ation in materials and material properties has received relatively little attention. In fact, the
perceptual knowledge that is captured in the innumerable artworks throughout history can
be thought of as the largest perceptual experiment in human history and it merits detailed
exploration.
A simple taxonomy of image datasets
To explore material depictions within art there is a need for a dataset that relates artworks to
material perception. Therefore, in this study, we create and introduce an accessible collection
of material depictions in paintings, which we call the Materials in Painting (MIP) dataset.
However, the use and creation of art-perception datasets is of broader interest.
We propose a simple taxonomy of three image dataset usages: 1) perceptual, 2) ecological
and, 3) computer vision usage. In the remainder of the introduction below, we will contextual-
ize our dataset within this taxonomy by first discussing existing image and painting datasets as
well as the benefits our MIP dataset can provide for each of these three dataset usages. This
shall be followed by a detailed description of the creation of the MIP dataset in the method sec-
tion. Finally, we perform and discuss several small experiments that exemplify the utility of the
MIP datasets for each of three dataset usages discussed.
Perceptual datasets. To understand the human visual system, stimuli from perceptual
datasets can be used in an attempt to relate the evoked perception to the visual input. We can
roughly categorize three types of stimuli used for visual perception: natural, synthetic and
manipulated.
The first represent ‘normal’ photos of objects, materials and scenes as they can be found in
reality. Experimental design with such stimuli often attempts to relate the evoked perceptions
to natural image statistics within the images or physical characteristics of the contents captured
in the images. Some examples of uses of natural stimuli datasets include, but are not limited to,
the memorability of pictures in general [18] or more specifically the memorability of faces
[19]. In another example, images of natural, but novel objects were used to understand what
underlies the visual classification of objects [20].
The second type, synthetic stimuli, are created artificially, such as digital renderings, draw-
ings and paintings. Synthetic stimuli might represent the real world, but often contain image
statistics that deviate from natural image statistics. Paintings have for example often been used
to study affect and aesthetics [21–23]. In another example [24], used a set of synthetic stimuli
to test for memorability of data visualizations.
Both natural and synthetic images can be manipulated, which leads to the third type of sti-
muli. Manipulated stimuli are often used to investigate the effect of image manipulations by
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 2 / 30
Kavita Bala acknowledge support from NSF (CHS-
1617861 and CHS-1513967), and NSERC (PGS-
D).
Competing interests: The authors have declared
that no competing interests exist.
comparing them to the original (natural or synthetic) image. Here the manipulations function
as the independent variables. For example [25] created a database of images that contain scene
inconsistencies that can be used to study the compositional rules of our visual environment. In
another example, a stimulus set consisting of original and texture (i.e., manipulated) versions
of animals found that perceived animal size is mediated by mid-level image statistics [26].
The advantage of using manipulated or synthetic images is that perceptual judgments can
be compared to some independent variable, which is typically not available for natural images.
Paintings are a special case here. They are a synthetic image of a 3D scene that is rendered
using oils, pigments and artistic instruments. However, the painting is also a mostly flat, physi-
cal object. Retrieving the veridical data is usually impossible for paintings. In other words,
objects or materials depicted in photos can often be measured or interacted with in the real
world but this is rarely possible for paintings. However, the advantage of using paintings is
that it can often be seen, or (historically) inferred, how the painter created the illusory realism.
Even if it cannot be seen with the naked eye, chemical and physical analysis can be performed.
In [27] a perceptually convincing depiction of grapes was recreated using a 17th century rec-
ipe. In this reconstruction, the depiction was recreated by a professional painter one layer at a
time, where each layer represents a separate and perceptually diagnostic image feature that
together lead to the perception of grapes. The physical limitations of painterly depictions rela-
tive to the physical 3D world, such as for example due to luminance compression in paintings
[28–31] may lead to systematically different strategies for material depiction. Despite this [15],
has shown that the perceptions of materials and material properties depicted in paintings are
similar to those previously reported for photographic materials [14].
Therefore, studying paintings in addition to more traditional stimuli like photos or render-
ings, can enrich our understanding of human material perception. It should be noted that in
this paper we focus on the image structure of the painting instead of the physical object. In
other words, we focus on what is depicted within paintings and our data and analysis is limited
to pictorial perception. In the remainder of this paper, when we mention paintings, we mean
images of paintings.
Throughout history, painters have studied how to trigger the perceptual system and create
convincing depictions of complex properties of the world. This resulted in perceptual shortcuts,
i.e., stylized depictions of complex properties of the world that trigger a robust perception.
The steps and painterly techniques applied by a painter to create a perceptual shortcut can be
thought of as a perception-based recipe. Following such a recipe results in a perceptual short-
cut, which is a depiction that gives the visual system the required inputs to trigger a perception.
Many of the successful depictions are now available in museum collections. As such, the crea-
tion of art throughout history can be seen as one massive perceptual experiment. Studying
perceptual shortcuts in art, and understanding the cues, i.e., features required to trigger per-
ceptions, can give insights into the visual system. We will demonstrate this idea by analyzing
highlights in paintings and photos.
Ecological datasets. To understand how the human visual system works it is important to
understand what type of visual input is given by the environment. Visual ecology encompasses
all the visual input and can be subdivided into natural and cultural ecology. Natural ecology
reflects all which is found in the physical world. For example, to understand color-vision and
cone cell sensitivities it is relevant to know the typical spectra of the environment. For this pur-
pose, hyperspectral images [32,33] can be used, in this case to investigate color metamers (per-
ceptually identical colors that originate from different spectra) and illumination variation. In
another example, a dataset of calibrated color images were used to understand color constancy
[34] (the ability to discount for chromatic changes in illumination when inferring object
color). The SYNS database was used to relate image statistics to physical statistics [35]. Another
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 3 / 30
dataset contains photos taken in Botswana [36] in an area that supposedly reflects the environ-
ment of the proto-human and was used to investigate the evolution of the human visual sys-
tem. Spatial statistics of today’s human visual ecology are clearly different from Botswana’s
bushes as most people live in urban areas that are shaped by humans. For example, a dataset
from [37] was used to compute the distribution of spatial orientations of natural scenes [38].
The content depicted within paintings only loosely reflects the natural visual ecology, but
they do strongly represent cultural visual ecology. They have influenced how people see and
depict the world and have influenced visual conventions up to contemporary cinematography
and photography. Both perceptual scientists and art historians have looked for and studied
compositional rules and conventions within art. A good example is the painterly convention
that light tends to originate from the top-left [39,40], which is likely related to the human
light-from-above prior [41–44].
New developments in cultural heritage institutions have made the measurement and study
of paintings much more accessible. In recent years the digitization of cultural heritage has led
to a surge in publicly available digitized art works. Many individual galleries and art institu-
tions have undertaken the admirable task to digitize their entire collection, and have often
make a portion, if not the whole collection digitally available with no or minor copyright
restrictions. The availability of digitized art works, combined with advancements in image
analysis algorithms, has lead to Digital Art History, which concerns itself with the digitized
analysis of artworks by for example analyzing artistic style [45] and beauty [46], or local pat-
tern similarities between artworks [47]. In [48], the authors for example developed a system
that automatically detects and extracts garment color in portraits, which can for example be
used for the digital analysis of historical trends within clothes and fashion.
Crowley and Zisserman [49] pointed out that art historians often have the unenviable task
of finding paintings for study manually. With an extensive dataset of material depictions
within art, this task might become slightly easier for art historians that study the artistic depic-
tion of materials, such as for example stone [50,51]. The ability to easy find fabrics in paintings
and it’s fine-grained subclasses such as velvet, silk and lace could be used for the study of fash-
ion and clothes in paintings in general [52,53] or for paintings from a specific cultural context,
such as Italian [54], English and French [55] or even for the clothes worn by specific artists
[56]. The human body and it’s skin, which clothing covers, is often studied within paintings
[52,57,58]. For example, the Metropolitan Museum, published an essay on anatomy in the
Renaissance, for which artworks depicting the human nude were used [59]. In this work on
anatomy, only items from the Metropolitan Museum were used but with an annotated data-
base of material depictions this could be extended and compared to other museum collections.
Furthermore, through material categories such as food and flora category, the MIP could give
access to typical artistic scenes such as stillives [60,61] and floral scenes [62] respectively. It
should be noted that ‘stuff’ like skin and food might not appear like a stereotypical material,
however in this paper we adhere to the view of Adelson, where each object, or ‘thing’, is con-
sidered to consist of some material, i.e., ‘stuff’ [63]. Within this view non-stereotypical ‘stuff’
such as skin and food can certainly be considered as a material.
Computer vision datasets. Today, the majority of image datasets originate from research
in computer vision. One of the first relatively large datasets representing object categories [64]
has been used to both train and evaluate various computational strategies to solve visual object
recognition. The ImageNet and CIFAR datasets [65,66] are regarded to be standard image rec-
ognition datasets for the last decade of research on deep learning vision systems.
Traditionally much visual research has been concerned with object classification but
recently material perception has received increasing attention [63,67–69]. A notable dataset
that contains material information is OpenSurfaces [67], which contains around 70k crowd-
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 4 / 30
sourced polygon segmentations of materials in photos. The Material In Context database
improved on OpenSurfaces by providing 3 million samples across 23 material classes [68].
To our knowledge, no dataset exists that explicitly provides material information within
paintings.
The majority of image datasets contain photographs, but various datasets exist that contain
artworks. The WikiArt dataset for instance, which is created and maintained by a non-profit
organisation, with the admirable goal “to make world’s art accessible to anyone and anywhere”
[70]. The WikiArt dataset has been widely used for a variety of scientific purposes [45,71–74].
The Painting-91 dataset from [75] consists of around 4000 paintings from 91 artists and was
introduced for the purpose of categorization on style or artist. More recently, Art500k was
released, which contains more than 500k low resolution artworks which were used to automat-
ically identify content and style [76] within paintings.
The visual difference introduced by painterly depiction does not pose any significant diffi-
culties to the human visual system, however it can be challenging for computer vision systems
as a result of the domain shift [77–79]. Differences between painting images and photographic
datasets include for instance composition, textural properties, colors and tone mapping, per-
spective, and style. As for composition, photos in image datasets are often ‘snapshots’, taken
with not too much thought given to composition, and typically intended to quickly capture a
scene or event. In contrast, paintings are artistically composed and are prone to historical style
trends. Therefore, photos often contain much more composition variation relative to paint-
ings. Within paintings, composition can vary greatly between different styles. The human
visual system can distinguish styles—for example, Baroque vs. Impressionism—and also
implicitly judge whether two paintings are stylistically similar. Research in style or artist classi-
fication, as well as neural networks that perform style transfer, attempt to model these stylistic
variations in art [45,80].
Humans can also discount stylistic differences, for example, identifying the same person or
object depicted by different artists. Similarly, work in domain adaptation [77–79] focuses on
understanding objects or ‘stuff’ across different image styles. Models that learn to convert pho-
tographs into painting-like or sketch-like images have been studied extensively for their appli-
cation as a tool for digital artists [80]. Recent work has shown that such neural style transfer
algorithms can also produce images that are useful for training robust neural networks [81].
However, photos that have been converted into a painting-like image are not identical to
paintings; paintings can contain spatial variations of style and statistics that are not present in
photos converted into paintings. Furthermore, painterly convention and composition are not
taken into account by style-transfer algorithms.
Depending on the end goal for a computer vision system, it can be important to learn from
paintings directly. Of course, when the end goal is to detect pedestrians for a self-driving car,
learning from real photos, videos, or renderings of simulations can suffice. However, if the
goal is to simulate general visual intelligence, multi-domain training sets are essential. Further-
more, if the goal is to create computer vision systems with a perception that matches human
vision, training on paintings could be very beneficial. Paintings are explicitly created by and
for human perception and therefor contain all the required cues to trigger robust perceptions.
Therefor, networks trained on paintings are implicitly trained on these perceptual cues.
The multifaceted nature of datasets. While we have distinguished the broad purposes of
datasets and exemplified each with representative datasets, it is important to keep in mind that
these datasets can serve multiple goals across the taxonomy. For example, the Flickr Material
Database [12] was initially created as a perceptual dataset to study how quickly human partici-
pants were capable of recognizing natural materials. However, since then it has also often been
used as a computer vision dataset, including by the original authors themselves [82]. In this
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 5 / 30
study paintings are considered as especially interesting as they can be used for perceptual
experiments, for digital art history, i.e., cultural visual ecology and can furthermore be used
to train and test computer vision networks. The dataset presented in this paper is explicitly
designed with this multidisciplinary nature in mind.
Methods
Here we will first provide a short description of the dataset and the various stages of data col-
lection, followed by an in-depth description of each stage.
Our dataset consists of high-quality images of paintings sourced from international art
institutions and galleries. Within these images, human annotators have created bounding
box around 15 material categories (e.g., fabrics, stone, etc). We further sub-categorized these
material categories into over 50 fine-grained categories (e.g., velvet, etc). Finally, we automati-
cally extract polygon segments for each bounding box. The annotated dataset will be made
publicly available. All paintings, bounding boxes, labels, and metadata are available online.
The data collection was executed in multiple stages. Here we give an itemized overview of
each stage and subsequently we discuss each stage in depth. The first two stages were con-
ducted as part of a previous study [15], but we provide details here for completeness. Partici-
pants were recruited via Amazon Mechanical Turk (AMT). A total of 4451 unique AMT users
participated in this study and gave written consent prior to data collection. Data collection was
approved by the Human Research (ethics) Committee of the Delft University of Technology
and adheres to the Declaration of Helsinki.
• First, we collected a large set of paintings.
• Next, human observers on the AMT platform identified which coarse-grained materials they
perceived to be present in each painting (e.g., “is there wood depicted in this painting?”).
• Then, for paintings identified to contain a specific material, AMT users were tasked with cre-
ating a bounding box of that material in that painting.
• Lastly, AMT users assigned a fine-grained material label to bounding boxes (e.g., processed
wood, natural wood, etc.).
Collecting paintings
We collected 19,325 high-quality digital reproductions of paintings from 9 online, open-access
art galleries. The details of these art galleries are presented in Table 1. Images were downloaded
Table 1. List of galleries. A list of all the galleries, the country in which the gallery is located, and the numberof paint-
ings downloaded from that gallery.
Gallery Name Country Count
The Rijksmuseum Netherlands 4672
The Metropolitan Museum of Art USA 3222
Nationalmuseum Sweden 3077
Cleveland Museum of Art USA 2217
National Gallery of Art USA 2132
Museo Nacional del Prado Spain 2032
The Art Institute of Chicago USA 936
Mauritshuis Netherlands 638
J. Paul Getty Museum USA 399
https://doi.org/10.1371/journal.pone.0255109.t001
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 6 / 30
from the online galleries, either using web scraping or through an API. For the majority of
these paintings we also gathered the following metadata: title of the work, estimated year of
creation and name of the artist.
For 92% of boxes, we also have an estimate of the year of production. These estimates were
made by the galleries from which the paintings were downloaded. The distribution of the year
of production for all paintings are plotted in Fig 1.
Image-level coarse-grained material labels
Next, we collected human annotations to identify material categories within paintings. We cre-
ated a list of 15 material categories: animal, ceramic, fabric, sky, stone, flora, food, gem, wood,
skin, glass, ground, liquid, paper, and metal. Our intention was to create a succinct list, that
would nevertheless allow the majority of ‘stuff’ within a painting to be annotated. Our list is
partially based on material lists used in [12,14], with which our set has 8 materials categories
in common, and partially based on [67], with which our list has 11 material categories in com-
mon. Note however minor difference in the category labels; the lists used in [12,14,67] has
‘water’, which we have named ‘liquid’ instead.
Our working definition of materials here is heavily influenced by [63,83], where material
does not just refer to prototypical materials that are used as construction materials such as
wood and stone, but also to the ‘stuff’ that makes up ‘things’. For example, few people would
consider banana as a material, but nevertheless this object has been made up of some type of
banana-material, which humans are capable of recognizing, distinguishing and estimating
physical properties off. In this rational, we have included some less typical ‘stuff’ categories,
such as for food and animal. Note however that we made an exception for skin, instead of a
more overarching ‘human’ category as one might expect considering based on the previous.
We made this choice because of the scientific interest in the artistic depiction [58], perception
[84,85], and rendering of skin directly [86,87]. Last, we realized that for many paintings a
large portion was dedicated to the depiction of the sky or ground ‘stuff’, neither of which are
considered a prototypical material, but on average both take up a large portion of paintings.
Therefore, in an attempt to more densely annotate the whole region of the painting, we
included sky and ground.
In one AMT task, participants would be presented with 40 paintings at a time and one tar-
get material category. In the task, participants were asked if the painting depicted the target
Fig 1. Histogram of the distribution of paintings over time. Each bin equals 20 years.
https://doi.org/10.1371/journal.pone.0255109.g001
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 7 / 30
material (e.g., does this painting contain wood?). They could reply ‘Yes, the target material is
depicted in this painting.’ by clicking the painting and inversely, by not clicking the painting,
participants would reply with ‘No, the target material is not depicted in this painting.’. Each
painting was presented to at least 5 participants for each of the 15 materials. If at least 80% of
the responses per painting claimed that the material was depicted in the painting, we would
register that material as present for that painting. In total, we collected 1,614,323 human
responses in this stage from 3,233 unique AMT users participating.
Extreme click bounding boxes
In the previous stage, paintings were registered to depict or not to depict a material. However,
that stage does not inform us (1) how often the material is depicted, nor (2) where the material
(s) are within the painting.
We gathered this information on the basis of extreme click bounding boxes. For extreme
click bounding boxes, a participant is asked to click on the 4 extreme positions of the material:
the highest, lowest, most left-, and most right-wards point [88]. See Fig 2 for an example. In
the task, participants were presented with paintings that depicted the target material and
tasked to create up to 5 extreme click bounding boxes for the target material.
To make bounding boxes within the task, the participants would use our interface, which
allows users to zoom in and out, and pan around the image. The interface furthermore allowed
participants to finely adjust the exact location of the extreme points by dragging the points
around. Initially, the tasks were open to all AMT workers, but after around 2000 bounding
boxes were created by 114 AMT users, with manual inspection, we found that the quality of
bounding boxes varied greatly between participants. Therefore, we restricted the work to a
smaller number of manually selected participants who were observed to create good bounding
boxes. After this restriction, new boxes were manually inspected by the authors, and in a few
cases additional participants were restricted due to a deterioration of bounding box quality.
Simultaneously additional participants were granted access to our tasks after passing (paid)
qualification tasks. As a result, the number of manually selected participants varied between 10
and 20 participants. In total, 227,810 bounding boxes were created by participants.
Automatic bounding boxes. While we consider our dataset to be quite larger, it only cov-
ers a small but representative portion of art history. It might be required to access materials in
paintings that are not part of our dataset. To allow for this, we have trained a FasterRCNN [89]
bounding box detector to localize and label material boxes in unlabelled paintings. We use the
publicly available implementation from [90] with a ResNet-50 backbone and feature pyramid
network (R50-FPN). The model is finetuned from a COCO-pretrained model for 100 epochs
using the default COCO hyperparameters from [90]. First we trained the detector on 90% of
annotated paintings in the dataset. In section Automatically detected bounding boxes in the
Results and demonstrations section below, we show our evaluation of the network, which was
performed on the remaining 10% of annotated paintings. While we created this network to be
able to detect paintings outside our dataset, we decided to apply the network on our dataset in
order to more densely annotate our paintings. Therefor, after the evaluation, we ran the detec-
tion network on the entire set of paintings, i.e., training and testing data, in an attempt to
more exhaustively annotate materials within paintings. From the automatic detected bounding
boxes we first removed all boxes that scored <50% confidence (as calculated by FasterRCNN).
Next, we filtered out automatic boxes that were likely already identified by human annotators,
be removing automatic bounding boxes that scored 50% on intersection over union, i.e.,
automatic boxes that shared the majority of it’s content with human boxes. This resulted in an
additional 94k bounding boxes.
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 8 / 30
Fig 2. An example of four extreme clicks (marked in green) made by a user on a piece of fabric.These points correspond to the most left,
most right, highest and lowest points on the annotated item. The red-line displays the resulting bounding box. Samuel Barber Clark, by
James Frothingham, 1810, Cleveland Museum Of Art, image reproduced under a CC0 license.
https://doi.org/10.1371/journal.pone.0255109.g002
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 9 / 30
Fine-grained labels
In this step we supplemented the previously collected material labels with fine-grained mate-
rial labels (see Table 2). For example, a bounding box labelled as fabric could now be labelled
as silk, velvet, fur, etc. We excluded bounding boxes that were too small (e.g., width in pixels ×
height in pixels 5000) and boxes that were labelled as sky, ground or skin for which fine-
grained categorizations were not annotated. We collected fine-grained labels for the remaining
150,693 bounding boxes. Note that this only concerns the bounding boxes created by human
annotations as no automatically detected boxes were assigned a fine-grained material label.
For each of these 150,693 bounding boxes, we gathered responses from at least 5 different par-
ticipants. If the responses reached an agreement of at least 70%, we would assign the agreed
upon label to the bounding box. To guide the workers, we provide a textual description for
each fine-grained category for them to reference during the task. We did not provide visual
exemplars as we did not want to bias the workers into template matching instead of relying on
their own perceptual understanding.
We found that it is non-trivial to define fine-grained labels in such a way that they are con-
cise, uniform and versatile (i.e., useful across different scientific domains) while still being rec-
ognizable and/or categorizable by naive observers. We applied the following reasoning to
select fine-grained labels: first, we tried to divide the materials into an exhaustive list with as
few fine-grained labels as possible. For example, for ‘wood’, each bounding box is either ‘pro-
cessed wood’ or ‘natural wood’. If an exhaustive list would become too long to be useful, we
would include an ‘other’ option. For example, for ‘glass’ we hypothesized that the vast majority
of bounding boxes would be captured by either ‘glass windows’ or ‘glass containers’. However,
to include all possible edge cases such as glass spectacles and glass eyeballs, we included the
‘other’ option.
A possible subset for ‘metal’ we considered was ‘iron’,‘bronze’, ‘copper’, ‘silver’, ‘gold’,
‘other’. However, we feared that naive participants would not be able to consistently categorize
these metals. An alternative would be to subcategorize on object-level, e.g. ‘swords’, ‘nails’, etc.,
but as we are interested in material categorization, we tried to avoid this as much as possible.
Thus, for ‘metal’, and for the same reason ‘ceramic’, we required a different method. We chose
to subcategorize on color, as often the color for these materials are tied to object identity.
Participants are shown one bounding box at a time and are instructed to choose which of
the fine-grained labels they considered most applicable. Additionally, they are able to select a
‘not target material’ option.
We collected over one million responses from 1114 participants. This resulted in a a total of
105,708 boxes assigned with a fine-grained label. See Table 2 for the numbers per category.
Results and demonstrations
We conducted a diverse set of experiments to demonstrate how our annotated art-perception
dataset can drive research across perception, art history, and computer vision. First, we report
simple dataset statistics. Next, we organized our findings under the proposed dataset usage
taxonomy: perceptual demonstrations, cultural visual ecology demonstrations and computer
vision demonstrations.
Dataset statistics
The final dataset contains painterly depictions of materials, with a total of 19,325 paintings.
Participants have created a total of 227,810 bounding boxes and we additionally detected 94k
using a FasterRCNN. Each box has a coarse material label and 105,708 also have been assigned
a fine-grained material label. The total number of instances per material categories (coarse-
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 10 / 30
Table 2. The number of annotated bounding boxes for each coarse- and fine-grained category. Note that not every
bounding box is associated with a fine-grained label since participants were not always able to arrive at a consensus.
See main text for details.
Coarse-grained Fine-grained # Labels
animal 11606
birds 1822
reptiles and amphibians 144
fish and aquatic life 289
mammals 7752
insects 155
other animals 10
ceramic 3641
brown or red 1088
white 381
decorated 289
other ceramic 14
fabric 31557
velvety 261
lace 491
silky/satiny 1354
cotton/wool-like 5712
brocade 96
fur 27
other fabric 12
flora 26693
trees 12851
vegetables 96
fruits 1238
flowers 2515
plants 3699
food 3690
cheese 11
vegetables 107
fruits 1536
meat or poultry 183
bread 127
seafood 183
nuts 8
other 14
gem 10525
pearls 719
gemstones 715
other gems 1
glass 5546
glass window 2243
glass container 1003
other glass 171
ground 2552
liquid 5737
body of water 4583
(Continued)
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 11 / 30
and fine-grained) can be found in Table 2. Further analysis of spatial distribution of categories,
co-occurences, and other related statistics will be discussed in a following section in the con-
text of visual ecology.
Perceptual demonstrations
We believe that one of the benefits of our MIP dataset is that selections of the dataset can be
useful as stimuli for perceptual experiments. We demonstrate this by performing an annota-
tion experiment to study the painterly depiction of highlights on drinking glasses.
Perception-based recipes in painterly depictions. As previously argued, we believe that
painterly techniques are a sort of perception-based recipe. Applying these recipes results in a
stylized depiction that can trigger a robust perception of the world. Studying the image features
in paintings can lead to an understanding of what cues the visual systems needs to trigger a
robust perception.
Here we explore a perceptual shortcut for the perceptions of glass by annotating highlights
in paintings and comparing these with highlights in photos. In paintings, it has previously
been observed that highlights on drinking glasses are typically in the shape of windows, even
in outdoor scenes [91]. This highlight-shape can even often be found in contemporary car-
toons [92,93]. This convention can be considered as a perception-based recipe, where the
result is a window-shaped highlight that appears to be a robust cue that triggers the perception
of gloss for drinking glasses.
We used bounding boxes from our dataset and photographs sourced from COCO [94]. Par-
ticipants for this study included 3 of the authors, and one lab-member naive to the purpose of
this experiment.
Table 2. (Continued)
Coarse-grained Fine-grained # Labels
liquid in container 458
other liquid 172
metal 27708
colorless metal 2933
yellowish metal 4435
brownish or reddish metal 510
multicolored or other colored metal 215
paper 3167
paper book 1380
paper sheets 585
paper scrolls 114
other paper 19
skin 32323
sky 12734
stone 23157
processed stone 9226
natural stone 9429
wood 26953
processed wood 12810
natural wood 10751
https://doi.org/10.1371/journal.pone.0255109.t002
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 12 / 30
Images. We used 110 images of drinking glasses, split equally across paintings and photos.
First, we selected all bounding boxes in the glass, liquid container category in our dataset.
From this set, we manually selected drinking glasses, since this category can also contain items
such as glass flower vases. Next, we removed all glasses that were mostly occluded, were diffi-
cult to parse from the background—for example when multiple glasses were standing behind
each other, and removed images smaller then 300x300 pixels. This resulted in a few hundred
painted drinking glasses.
Next, we downloaded all images containing cups and wineglasses from the COCO [94]
dataset, from which we removed all non-glass cups, occluded glasses, blurry glasses and glasses
that only occupy a small portion of the image, and small images. This left us with 55 photos
of glass cups and wineglass. Next, we randomly selected 55 segmentations from our painted
glass collection. Each image was presented in the task at 650 ×650 pixels, keeping aspect ratio
intact.
During this selection phase, we did not base our decision on the shape of the glass. After the
experiment, as part of the analysis, we divided the glasses into three shapes, namely spherical,
cylindrical, and conical glasses. See Fig 3 for an example of each shape.
Task. Participants annotated highlights on drinking glasses using an annotation interface.
In the annotation interface, users would be presented with an image on which the annotated
geometry was visible. This made it clear which glass should be annotated, in case multiple
glasses were visible in the image. Users were instructed to instruct all visible highlights on that
glass. Once the user started annotating highlights, the geometry would no longer be visible.
Annotations could be made by simply holding down the left-mouse button and drawing on
top of the image. Once a highlight was annotated a user could mark it as finished and continue
with the next highlight, and eventually move to the next image.
Results. To compare the highlights between photos and paintings, we resized each glass to
have the same maximum width and height, and then overlaid each glass on the center. Initially,
we overlaid all images for both types of media (not visualized here) and found the resulting fig-
ure quite noisy. However, when we split the glasses on media and shape, clear patterns emerge
for painted glasses Fig 4.
As can be seen, painters are more likely to depict highlights on glasses adhering to a stylized
pattern, at least for spherical and conical glasses. This pattern of highlights is perceptually con-
vincing and very uniform in comparison with the variation found within reality.
Furthermore, we calculated the agreement between each pair of participants, as the ratio of
pixels annotated by both participants (i.e., overlapping area) divided by the number of pixels
Fig 3. Examples of the three glass shapes. From left to right: spherical, cylindrical and conical. The red geometry
annotations were manually created by the authors, and were used to standardize across glasses for the highlight
analysis. Paintings used, from left to right: Portret van een jongen, zittend in een raamnis en gekleed in een blauw jasje,
by Jean Augustin Daiwaille. 1840. Still Life with a gilded Beer Tankard, by Willem Claesz. Hed, 1634. The White
Tablecloth, by Jean Baptiste Siméon Chardin, 1731. The left and middle images courtesy of The Rijksmuseum. The
right image courtesy of The Art Institute of Chicago. All images reproduced undera CC0 license.
https://doi.org/10.1371/journal.pone.0255109.g003
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 13 / 30
Fig 4. The overlaid highlights created by users, split on media and glass shape. In general, the photographic glass shapes display more
variability and do not display a clear pattern. Note that for photos, no stimuli existed with a conical shape in ourset which leads to a black
image, since there were no highlight-annotations. On the right, for painted glasses, we see clear patterns in the placement of highlights for
each glass shape.
https://doi.org/10.1371/journal.pone.0255109.g004
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 14 / 30
that was an annotated by either participant (i.e., total area). Averaged across participants, the
agreement on paintings (0.33) was around 50% higher relative to the average agreement
between participants on photos (0.21). This means that for our stimuli, highlights in paintings
are less ambiguous when compared to photos.
Cultural visual ecology demonstrations
The ecology displayed within paintings are representative of our visual culture. Our dataset
consists of paintings spanning 500+ years of art history. This provides a unique opportunity to
analyze a specific sub-domain of visual culture, i.e., that of paintings. Here we first analyze the
presence of materials in paintings in the Material presence section and in the next section we
analyse this over time. In The spatial layout of materials, we visualize the spatial distributions
of materials in our dataset. In the last section, we analyze the automatically detected bounding
boxes.
Material presence. Within the 19,325 paintings, participants exhaustively identified the
presence of 123,244 instances of 15 coarse materials. In other words, for each painting, partici-
pants indicated if each material is or is not present. The distribution of unique materials per
painting is normally distributed with an average of 5.7 unique coarse materials present per
painting (std = 2.8 materials). The most frequent materials are skin and fabric. The least fre-
quent are ceramics and food. The relative frequency of each coarse material is presented in
Fig 5. We did not exhaustively identify fine-grained materials within paintings, so we will not
report those statistics here.
Based on prior knowledge of natural ecology, one might assume that some materials, such
as skin and fabric might often be depicted together in paintings. To quantify the extent to
which materials are depicted together, we create a co-occurrence matrix presented in Fig 6,
where each cell is the co-occurrence for each pair of materials as the number of paintings
where both materials are present, divided by the number of paintings where either (but not
both) materials are present. We can see for example, that if skin is depicted, there is a 94%
change to also find fabric in the same painting.
Furthermore, one might expect that the presence of one material can have an influence on
another material. For example, one might expect that gem might almost always be depicted
with skin, but that skin is only sometimes depicted with gem. To quantify these relations, we
calculated the occurrence of a material given that another material is present. We visualize this
in Fig 7. Here we see that if gem is present, then skin is found in 99% of the paintings, but that
Fig 5. The proportion of paintings in our dataset that depict at least one instance of eachmaterial.
https://doi.org/10.1371/journal.pone.0255109.g005
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 15 / 30
if skin is present, then gem is found in only 20% of the paintings. The same relationship is true
for gem and fabric. This implies that gems are almost always depicted with human figures,
however that human figures are not always shown with gems. Another example, when liquid is
present, in 85% of the paintings, wood is also present. One might be reminded of typical naval
scenes, or landscapes with forests and rivers. Inversely, when wood is present, only 34% of the
paintings depict liquid. For food and ceramics, two materials which are present in less then
10% of paintings, we see that if food is present, ceramics has a 53% change to be present as well,
but the inverse is only 33%. This implies that food is served in, or with, ceramic containers half
of the time, but that this is only 1/3rd of what ceramics is used for.
Material presence over time. We have previously shown the distributions of materials in
paintings in Fig 5. When we created similar distributions (not visualised) for temporal cross-
sections, for example for a single century, we found that these distributions were remarkably
similar to the average distribution in Fig 5. We used t-tests, to see if the distribution for any
century was significantly different from the average distribution in Fig 5 and found no signifi-
cant effect. This means that despite the changes in stylistic and artistic techniques over time,
Fig 6. Co-occurrence matrix. Each cell equals the number of paintings where both materials are present divided by the number of
paintings where one or the other material is present.
https://doi.org/10.1371/journal.pone.0255109.g006
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 16 / 30
the distribution of materials (such as in Fig 5 remained remarkably stable over time for the
period covered in our dataset.
The spatial layout of materials. Paintings are carefully constructed scenes and it follows
that a painter would carefully choose the location at which to depict a material. With the
knowledge that spatial conventions exists within paintings (e.g., lighting direction [39,40]),
one can assume that these might extend to materials. The average spatial location and extent of
materials is visualized by taking the (normalized) location of each bounding box for a specific
material and subsequently plotting each box as a semi-transparent rectangle. The result is a
material heatmap, where the brightness of any pixel indicates the likelihood to find a material
at that pixel. In this section, we limit the material heatmaps to only include the bounding
boxes created by human annotators. In the next section, we visualize the material heatmaps for
automated boxes too.
Material heatmaps for the 15 coarse materials are shown in Fig 8. The expected finding that
sky and ground are spatially high and low within images serves as a simple validation or sanity-
Fig 7. Likelihood matrix. This matrix visualizes the influence a material has on the likelihood of finding another material within the
same painting, i.e., if one material on the y-axis is present, then how does this impact the presence of other materials on the x-axis?
Calculated as the number of paintings where both materials are present, divided by the number of paintings that contain only one of the
materials.
https://doi.org/10.1371/journal.pone.0255109.g007
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 17 / 30
check of the data. It is interesting to see how skin and gem are both vertically centered within
the canvas. It appears to suggests a face, with necklaces and jewelry adorning the figure. In gen-
eral, each material heatmap appears to be roughly vertically symmetric. For glass, there does
however appear to be a minor shift towards the top-left. This might be related to an artistic
convention, namely that light in paintings usually comes from a top-left window [39]. When
we look at the heatmaps for the sub-categories for glass in Fig 9, we see that it is indeed glass
windows that show the strongest top-left bias.
Automatically detected bounding boxes. Besides the bounding boxes created by
humans, we also trained a FasterRCNN network to automatically detect bounding boxes with
90% of the data as training data. On the remaining unseen 10% of paintings, the network
detected 90,169 bounding boxes. We removed those with a confidence score below 50%,
which resulted in 24,566 remaining bounding boxes. In the section below, all references to the
automated bounding boxes refer to these 24,566 bounding boxes.
A qualitative sample of detected bounding boxes is given in Fig 10. Our human bounding
boxes are non-spatially exhaustive in nature meaning that not every possible material has been
annotated. As a result, the automatically created bounding boxes cannot always be matched
against our human annotations and thus we cannot use this to evaluate their quality. In order
to validate the automatic bounding box detection, we performed a simple user study to get an
estimate of the accuracy per material class, which is visualized in Fig 11. In the user study, a
total of 50 AMT participants judged a random sample of 1500 bounding boxes. The 1500
bounding box stimuli were divided into 10 sets of 150 stimuli, where each set contain 10 boxes
Fig 8. Material heatmaps, which illustrate the likelihood at any given pixel to find the target material at that pixel. Brighter colors indicate higher likelihoods.
https://doi.org/10.1371/journal.pone.0255109.g008
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 18 / 30
per course material class. Each individual participant only saw one set and each set was seen by
5 unique participants. Therefore, this can be thought of as 10 experiments, each with 5 partici-
pants and 150 stimuli, where participants performed the same task across each set/experiment.
The participants were tasked to rate whether each stimuli is either a good or a bad bounding
box, where a bad bounding box was defined as either 1) having the wrong material label (e.g.,
“I see wood, but the label says stone”) or as having a bad boundary where the edges of the
bounding box were not near the edges of the material. The order of stimuli was randomized
between sets and participants. This leads to a total of 7500 votes, 500 per material classes. The
ratio of good to bad votes per material classes can serve as a measure of accuracy, which has
been visualized in Fig 11.
The participant agreement averaged across bounding boxes was found to be 80%, i.e., on
average 4 out of 5 participants agreed on their rating per bounding box. As a result of the user
study, we found a mean average accuracy of 0.55 across participants. While not high, these
results are somewhat interesting in that they show that a FasterRCNN model is capable of
detecting materials in paintings, without any changes to the network architecture or training
hyperparameters. It is certainly promising to see that an algorithm designed for object
Fig 9. Material heatmaps for glass sub categories. For glass windows, it is interesting to see the clustering in the top-left corner, which is in agreement with the
artistic convention of having light come from the top left.
https://doi.org/10.1371/journal.pone.0255109.g009
Fig 10. Examples of detected materials in unlabeled paintings. Automatically detecting materials can be useful for content retrieval for digital art history and for
filtering online galleries by viewer interests.
https://doi.org/10.1371/journal.pone.0255109.g010
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 19 / 30
localization in natural images can be readily applied to material localization in paintings.
Likely, the accuracy could be further improved by finetuning the network which we have not
done in this paper.
It is interesting to note that the spatial distribution of automatically detected bounding
boxes looks very similar to the spatial distribution of the human annotated bounding boxes.
We have visualized the material heatmap for one material, fabric, for the automated bounding
boxes to show the similarity with the material heatmap for the same material created from
human annotation bounding boxes. This has been visualized in the right side of Fig 11
Computer vision demonstrations
In this section, we will first apply existing segmentation tools designed for natural photographs
to extract polygon segmentations. Next, we perform an experiment to demonstrate the utility
of paintings for automated material classification.
Extracting polygon segmentations. A natural extension of material bounding boxes is
material segments [67–69]. Polygon segmentations are useful for reasoning about boundary
relationships between different semantic regions of an image, as well as the shape of the
regions themselves. However, annotating segmentations is expensive and many modern data-
sets rely on expensive manual annotation methods [67,69,94–96]. Recent work has focused
on more cost effective annotation methods (e.g. [97–100]). One broad family of methods to
relax the difficulty of annotating polygon segmentations is through the use of interactive seg-
mentation methods that transform sparse user inputs into a full polygon masks.
For this dataset, we apply interactive segmentation with the crowdsourced extreme clicks as
input. To evaluate quality, we compared against 4.5k high-quality human annotated segmenta-
tions from [15], which were sourced from the same set of paintings. We find that both image-
based approaches like GrabCut (GC) [101] and modern deep learning approaches such as
DEXTR [98] perform well. Surprisingly, DEXTR transfers quite well to paintings despite being
trained only on natural photographs of objects. The performance is summarized in Table 3.
The performance is summarized using the standard intersection over union (IOU) metric.
IOU is computed as the intersection between a predicted segment and the ground truth seg-
ment divided by the union of both segments. IOU is computed for each class, and mIOU is the
Fig 11. In the bar graph, the accuracy for automatically detected bounding boxes isdisplayed in the same order as in Fig 5.The values were derived from
human quality votes. On the right, we compare the material heatmaps for fabric between the automated and the human annotation bounding boxes. From left to
right, top to bottom: Lake George, by John William Casilear, 1857. Man with a Celestial Globe, by Nicolaes Eliasz Pickenoy, 1624. A Seven-Part Decorative Sequence:
An Interior, by Dirck van Delen, 1631. Thomas Howard, 2nd Earl of Arundel, by Anthony van Dyck, 1620. The Poultry Seller, by Cornelis Jacobsz. Delff, 1631. First
and second digital image courtesy of The Metropolitan Museum of Art. Third and last image courtesy of The Rijksmusuem. Fourth image courty of the Getty’s
Open Content Program. All images reproduced under a CC0 license.
https://doi.org/10.1371/journal.pone.0255109.g011
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 20 / 30
mean IOU over all of the classes. Samples are visualized in Fig 12. Segments produced by these
methods from our crowdsourced extreme points will be released with the dataset.
Learning robust cues for finegrained fabric classification. The task of distinguishing
between images of different semantic content is a standard recognition task for computer
vision systems. Recently, increasing attention is being given to “fine-grained” classification,
where a model is tasked with distinguishing images of the same coarse-grained category (e.g.,
distinguishing different species of birds or different types of flora [102–104]). Classifiers for
material categories can perform reasonably well on coarse-grained classification by relying on
context alone. In comparison, fine-grained classification is more challenging for deep learning
systems as contextual clues are often equal within fine-grained classes. For example, one might
reason that the material of a shirt might be recognized as fabrics partly because of the context,
i.e., being worn by a figure. However, in fine-grained classification context can be held consis-
tent across classes (for example, both velvet shirts and satin shirts are worn). To successfully
Table 3. Segmentations from extreme clicks. Grabcut [101] rectangles use bounding-box only initialization as a reference baseline. Grabcut Extr is based on the improved
GC initialization from [88] with small modifications: (a) we compute the minimum cost boundary with the cost as the negative log probability of a pixel belonging to an
edge; (b) in addition to clamping the morphological skeleton, we also clamp the extreme points centroid as well as the extreme points; (c) we compute the GC directly on
the RGB image. DEXTR [98] Pascal-SBD and COCO are pretrained DEXTR ResNet101 models on the respective datasets. Note that Pascal-SBD and COCO are natural
image datasets of objects, but DEXTR transfers surprisingly well across both visual domains (paintings vs. photos) and annotation categories (materials vs. objects).
mIOU (%)
Grabcut Rectangle Grabcut Extr DEXTR Pascal-SBD DEXTR COCO DEXTR Finetune
44.1 72.4 74.3 76.4 78.4
DEXTR Finetune IOU By Class (%)
Animal Ceramic Fabric Flora Food
76.9 86.8 79.1 77.0 87.5
Gem Glass Ground Liquid Metal
74.4 83.2 69.6 73.0 75.5
Paper Skin Sky Stone Wood
86.1 78.9 78.5 81.7 67.4
https://doi.org/10.1371/journal.pone.0255109.t003
Fig 12. Segmentation visualizations. Left to right: Original Image, Ground Truth Segment, Grabcut Extr Segment,
DEXTR COCO Segment. Both Grabcut and DEXTR use extreme points as input. For evaluation, the extreme points
are generated synthetically from the ground truth segments. The IOU for each segmentation is shown in the bottom
right corner. Top image: Dance before a Fountain, by Nicolas Lancret, 1724, Digital image courtesy of the Getty’s Open
Content Program. Bottom image: Still life with fish, by Pieter van Noort, 1660, Het Rijksmusuem. Images reproduced
under a CC0 license.
https://doi.org/10.1371/journal.pone.0255109.g012
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 21 / 30
distinguish between these two fine-grained classes in a context-controlled setting, a classifier
should use non-contextual features (at least more so relative to uncontrolled settings).
The rational above leads to two interesting possibilities. First, we hypothesize that the
painted depictions of materials can be beneficial for fine-grained classification tasks. Since
artistic depictions focus on salient cues for perception, i.e., paintings are explicitly created for
and by perception, it is possible that a network trained on paintings is able to learn a more
robust feature representation by focusing on these cues.
Second, visualising the features used by a successful fine-grained classifier could potentially
lead to the uncovering of latent perceptual cues. For example, in the Perception-based recipes in
painterly depictions section above, we showed that window-shaped highlights are a robust cue
for the perception of gloss on drinking glasses. However, it is assumed that the visual system
used many such cues which are as of now unknown. Visualising what cues are used by classifi-
ers might lead to the finding of cues used by the perceptual system.
Task. We experimented with the task of classifying cotton/wool versus silk/satin. The latter
can be recognized through local cues such as highlights on the cloth; such cues are carefully
placed by artists in paintings. To understand whether artistic depictions of fabric allow a neu-
ral network to learn better features for classification, we trained a model with either photo-
graphs or paintings. High resolution photographs of cotton/wool and silk/satin fabric and
clothing (dresses, shirts) were downloaded and manually filtered from publicly available pho-
tos licensed under the Creative Commons from Flickr. In total, we downloaded roughly 1K
photos. We sampled cotton/wool and silk/satin samples from our dataset to form a corre-
sponding dataset of 1K paintings. We analyzed the robustness of the classifier trained on paint-
ings versus the classifier trained on photos in two experiments below. Taken together, our
results provide evidence that a classifier trained on paintings can be more robust than a classi-
fier trained on photographs, and that visualizing these features could lead to discovering per-
ceptual cues utilized by the human visual system.
Generalizability of classifiers. Does training with paintings improve the generalizability of
classifiers? To test cross-domain generalization, we test the classifier on types of images that it
has not seen before. A classifier that has learned robust features will outperform a classifier
that has learned features based on more spurious correlations. We tested the trained classifiers
on both photographs and paintings across the two classes using 1000 samples per domain.
In Table 4, the performance of the two classifiers are summarized. We found that both clas-
sifiers perform similarly well on the domain they are trained on. However, when the classifiers
are tested on cross-domain data, we found that the painting-trained classifier performs better
than the photo-trained classifier. This suggests that the classifier trained on paintings has
learned a more generalizable feature representation for this task.
We have reported the confusion matrices in Table 5. The photo classifier applied to paint-
ings is heavily biased towards satin predictions. We hypothesize that this is because the photo
classifier is relying on spurious cues (such as image background or clothing shape) over more
Table 4. Classifier performance across domains. Classifiers are trained to distinguish cotton/wool from silk/satin.
The first column represents the classifier trained on photographs, and the second column represents the classifier
trained on paintings. In the first row, the classifiers are tested on images of the same type they were trained on (i.e.,
trained and tested on photos, and trained and tested on paintings). In the second row, the classifiers are tested on the
other medium, i.e., trained on photos and tested on paintings and vice versa.
Photo !Photo Painting !Painting
MEAN F1 Score 79.6% 80.5%
Photo !Painting Painting !Photo
MEAN F1 Score 49.5% 57.8%
https://doi.org/10.1371/journal.pone.0255109.t004
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 22 / 30
robust cues and thus that the shift from photos to paintings causes its mispredictions. Only
21% of cotton samples are correctly identified as cotton while 79% are identified as satin. This
skew in precision/recall across the classes is also reflected by the F1 scores for each class. On
the other hand, the painting classifier applied to photos is much more balanced in its predic-
tions, with 57-59% of predictions being correct. The precision/recalls are also much better bal-
anced as reflected by the F1 scores.
Human agreement with classifier cues. How indicative are the cues used by each classifier to
humans? We hypothesized that training networks on paintings might lead to the use of more
perceptually relevant image features. If this is true, then the features used by the classifier
trained on paintings should be preferably by humans.
We produced evidence heatmaps with GradCAM [105] from the feature maps in the net-
work before the fully connected classification layer. We extracted high resolution feature maps
from images of size 1024 ×1024 (for a feature map of size 32 ×32). The heatmaps produced by
GradCAM show which regions of an image the classifier uses as evidence for a specific class. If
the cues (i.e., evidence heatmaps, such as in Fig 13) are clearly interpretable, this would imply
the classifier has learned a good representation. For both models, we computed heatmaps for
test images corresponding to their ground truth label. We conducted a user study on Amazon
Mechanical Turk to find which heatmaps are judged as more informative by users. Users were
Table 5. Confusion matrix for the two classifiers. The top represents the classifier trained on photos, tested on paint-
ings which is heavily biased towards satin. The bottom represents the classifier trained on paintings, tested on photos,
which is more balanced in its predictions.
Photo !Painting
Cotton Satin
Cotton 20.83% 79.17%
Satin 9.72% 90.28%
Per class F1 31.91 67.01
Painting !Photo
Cotton Satin
Cotton 58.82% 41.18%
Satin 42.86% 57.14%
Per class F1 55.56 60.00
https://doi.org/10.1371/journal.pone.0255109.t005
Fig 13. Visualization of cues used by classifiers. Left to right: Original Image, Masked Image (Painting Classifier),
Masked Image (Photo Classifier). The unmasked regions represent evidence used by the classifiers for predicting “silk/
satin” in this particular image. See main text for details. Image from Interior of the Laurenskerk at Rotterdam, by
Anthonie De Lorme, with figures attributed to Ludolf de Jongh, 1662. Digital image courtesy of the Getty’s Open
Content Program, reproduced under a CC0 license.
https://doi.org/10.1371/journal.pone.0255109.g013
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 23 / 30
shown images with regions corresponding to heatmap values that are above 1.5 standard devi-
ations above the mean. Fig 13 illustrates an example. Users were instructed to “select the image
that contains the regions that look the most like <material>”, where <material>was either
cotton/wool or silk/satin. We collected responses from 85 participants, 57 of which were ana-
lyzed after quality control. For quality control, we only kept results from participants who
spent over 1 second on average per trial.
Overall, we found that the classifier trained on paintings uses evidence that is better aligned
with evidence preferred by humans (Fig 14). This implies that training on paintings allows
classifiers to learn more perceptually relevant cues, and it shows that this method might be use-
ful to detect previously unknown perceptual cues.
Due to the domain shifts, training and testing a classifier on a single type of images will out-
perform a classifier trained and tested on different kind of images. Based on this, if paintings
do not lead to a more robust feature representation we would expect the painting classifier to
do best on paintings and the photo classifier to perform best on photos. Interestingly however,
this does not hold when testing on photos of the satin/silk category (see last column of Fig 14).
We found that users actually have no preference for the cues from either classifier, i.e., the cues
from the painting classifier appears to be equally informative as the cues from the photo classi-
fier for categorizing silk/satin in photos. This suggests that either (a) the painting classifier has
learned human-interpretable perceptual cues for recognizing satin/silk, or (b) that the photo
classifier has learned to classify satin/silk based on some spurious contextual signals that are
difficult to interpret by humans. We asked users to elucidate their reasoning when choosing
which set of cues they preferred. In general, users noted that they preferred the network which
picks out regions containing the target class. Therefore, it seems that the network trained on
Fig 14. Human preference for classification cues used by each classifier. The y-axis represents how often humans
prefer the cues from a classifier trained on the same domain as the test images. For example, the first bar indicates that
in 73.2% of the cases, humans preferred cues from the classifier trained on paintings when classifying wool/cotton
paintings (and thus, the inverse, that in 26.8% of the cases, humans preferred cues from the photo classifier.)
Interestingly, note the last column—humans equally prefer cues used by both classifiers for classifying silk/satin photos
despite the painting classifier never seeing a photo during training.
https://doi.org/10.1371/journal.pone.0255109.g014
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 24 / 30
paintings has learned better to distinguish fabric through the appearance of such fabrics in the
image over other contextual signals (see Fig 13).
Conclusion
In this paper, we presented the Materials in Paintings (MIP) dataset—a dataset of painterly
depictions of different materials throughout time. The dataset can be visited, browsed and
downloaded at materialsinpaintings.tudelft.nl. The MIP dataset consists of 19,325 high resolu-
tion images of painting. Unlike existing datasets that contains paintings, such as for example
[75,76], the MIP dataset contains exhaustive material labels across 15 categories for all paint-
ings within the set. Additionally, human annotators have created 227,810 bounding boxes and
we automatically identified an additional 94k bounding boxes. Each bounding box also con-
tains a material label and half are additionally assigned with a fine-grained material label.
Although the findings reported in this study are valuable for their own sake, together they
demonstrate the wide utility that a dataset of painterly depictions can serve. We hope that the
MIP dataset can support research in multiple disciplines, as well as promote multidisciplinary
research. We have shown that depictions in paintings are not just of interest for art history, but
that they are also of fundamental interest for perception, as they can illustrate what cues the
visual system may use to construct a perception. We have shown that computer vision algo-
rithms trained on paintings appear to use cues more aligned with the human visual system,
when compared to algorithms trained on photos. The benefits of this might also extend to
learning perceptually robust models for image synthesis.
Our findings support our hope that the MIP dataset will be a valuable addition to the scien-
tific community to drive interdisciplinary research in art history, human perception, and com-
puter vision.
Acknowledgments
We appreciate the work and feedback of AMT participants that participated in our user stud-
ies. We further wish to thank Yuguang Zhao for his help with the design of the website.
Author Contributions
Conceptualization: Mitchell J. P. Van Zuijlen, Hubert Lin, Kavita Bala, Sylvia C. Pont, Maar-
ten W. A. Wijntjes.
Data curation: Mitchell J. P. Van Zuijlen, Hubert Lin.
Formal analysis: Mitchell J. P. Van Zuijlen, Hubert Lin.
Funding acquisition: Kavita Bala, Maarten W. A. Wijntjes.
Investigation: Mitchell J. P. Van Zuijlen.
Methodology: Mitchell J. P. Van Zuijlen, Hubert Lin.
Project administration: Kavita Bala, Sylvia C. Pont.
Software: Mitchell J. P. Van Zuijlen, Hubert Lin, Maarten W. A. Wijntjes.
Supervision: Kavita Bala, Sylvia C. Pont, Maarten W. A. Wijntjes.
Validation: Mitchell J. P. Van Zuijlen.
Visualization: Mitchell J. P. Van Zuijlen.
Writing – original draft: Mitchell J. P. Van Zuijlen, Hubert Lin, Maarten W. A. Wijntjes.
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 25 / 30
Writing – review & editing: Mitchell J. P. Van Zuijlen, Hubert Lin, Kavita Bala, Sylvia C.
Pont, Maarten W. A. Wijntjes.
References
1. Panofsky E. Perspective as symbolic form. Princeton University Press; 2020.
2. White J. The birth and rebirth of pictorial space. London, Faber and Faber; 1957.
3. Kemp M. The Science of Art: Optical themes in western art from Brunelleschi to Seurat. Yale Univer-
sity Press New Haven; 1992.
4. Pirenne MH. Optics, painting & photography. Cambridge University Press; 1970.
5. Willats J. Art and representation: New principles in the analysis of pictures. Princeton University
Press; 1997.
6. Cavanagh P. The artist as neuroscientist. Nature. 2005; 434(7031):301–307. https://doi.org/10.1038/
434301a PMID: 15772645
7. Graham DJ, Field DJ. Statistical regularities of art images and natural scenes: Spectra, sparseness
and nonlinearities. Spatial vision. 2008; 21(1-2):149–164.
8. Graham DJ, Field DJ. Variations in intensity statistics for representational and abstract art, and for art
from the Eastern and Western hemispheres. Perception. 2008; 37(9):1341–1352. https://doi.org/10.
1068/p5971 PMID: 18986061
9. Perdreau F, Cavanagh P. Do artists see their retinas? Frontiers in Human Neuroscience. 2011; 5:171.
https://doi.org/10.3389/fnhum.2011.00171 PMID: 22232584
10. Gombrich EH. Art & Illusion. A study in the psychology of pictorial representation. 5th ed. London:
Phaidon Press Limited; 1960.
11. Gibson JJ. The ecological approach to the visual perception of pictures. Leonardo. 1978; 11(3):227–
235. https://doi.org/10.2307/1574154
12. Sharan L, Rosenholtz R, Adelson EH. Material perception: What can you see in a brief glance? Vision
Sciences Society Annual Meeting Abstract. 2009; 9(8):2009.
13. Sharan L, Rosenholtz R, Adelson EH. Accuracy and speed of material categorization in real-world
images. Journal of vision. 2014; 14(9):1–24. https://doi.org/10.1167/14.9.12 PMID: 25122216
14. Fleming RW, Wiebel C, Gegenfurtner K. Perceptual qualities and material classes. Journal of Vision.
2013; 13(8):1–20. https://doi.org/10.1167/13.8.9 PMID: 23847302
15. van Zuijlen MJP, Pont SC, Wijntjes MWA. Painterly depiction of material properties. Journal of Vision.
2020; 20(7):1–17. https://doi.org/10.1167/jov.20.7.7 PMID: 32634227
16. Cavanagh P, Chao J, Wang D. Reflections in art. Spatial vision. 2008; 21(3-5):261–270. https://doi.
org/10.1163/156856808784532581 PMID: 18534102
17. Sayim B, Cavanagh P. The art of transparency. i-Perception. 2011; 2(7):679–696. https://doi.org/10.
1068/i0459aap PMID: 23145252
18. Isola P, Xiao J, Torralba A, Oliva A. What makes an image memorable? In: Proceedings of the IEEE
Conference on Computer Vision and Pattern Recognition (CVPR); 2011. p. 145–152.
19. Bainbridge WA, Isola P, Oliva A. The intrinsic memorability of face photographs. Journal of Experimen-
tal Psychology: General. 2013; 142(4). PMID: 24246059
20. Horst JS, Hout MC. The Novel Object and Unusual Name (NOUN) Database: A collection of novel
images for use in experimental research. Behavior research methods. 2016; 48(4):1393–1409. https://
doi.org/10.3758/s13428-015-0647-3 PMID: 26424438
21. Machajdik J, Hanbury A. Affective image classification using features inspired by psychology and art
theory. In: Proceedings of the 18th ACM international conference on Multimedia; 2010. p. 83–92.
22. Sartori A, Yan Y, O
¨zbal G, Salah AAA, Salah AA, Sebe N. Looking at Mondrian’s victory Boogie-Woo-
gie: what do I feel? In: Proceedings of the Twenty-Fourth International Joint Conference on Artificial
Intelligence (IJCAI 2015). vol. 1; 2015. p. 1–7.
23. Joshi D, Datta R, Fedorovskaya E, Luong QT, Wang JZ, Li J, et al. Aesthetics and emotions in
images. IEEE Signal Processing Magazine. 2011; 28(5):94–115. https://doi.org/10.1109/MSP.
2011.941851
24. Borkin MA, Vo AA, Bylinskii Z, Isola P, Sunkavalli S, Oliva A, et al. What makes a visualization memo-
rable? IEEE Transactions on Visualization and Computer Graphics. 2013; 19(12):2306–2315. https://
doi.org/10.1109/TVCG.2013.234 PMID: 24051797
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 26 / 30
25. O
¨hlschla
¨ger S, VõMLH. SCEGRAM: An image database for semantic and syntactic inconsistencies in
scenes. Behavior research methods. 2017; 49(5):1780–1791. https://doi.org/10.3758/s13428-016-
0820-3 PMID: 27800578
26. Long B, Yu CP, Konkle T. Mid-level visual features underlie the high-level categorical organization of
the ventral stream. Proceedings of the National Academy of Sciences. 2018; 115(38). https://doi.org/
10.1073/pnas.1719616115 PMID: 30171168
27. Di Cicco F, Wiersma L, Wijntjes MWA, Pont SC. Material properties and image cues for convincing
grapes: the know-how of the 17th-century pictorial recipe by Willem Beurs. Art & Perception. 2020; 1.
28. Graham DJ, Field DJ. Global nonlinear compression of natural luminances in painted art. In: Computer
image analysis in the study of art. vol. 6810. International Society for Optics and Photonics; 2008. p.
68100K.
29. Graham DJ, Friedenberg JD, Rockmore DN. Efficient visual system processing of spatial and lumi-
nance statistics in representational and non-representational art. In: Human Vision and Electronic
Imaging XIV. vol. 7240. International Society for Optics and Photonics; 2009. p. 72401N.
30. Graham DJ. Visual perception: Lightness in a high-dynamic-range world. Current Biology. 2011; 21
(22):R914–R916. https://doi.org/10.1016/j.cub.2011.10.003 PMID: 22115456
31. Graham DJ, Schwarz B, Chatterjee A, Leder H. Preference for luminance histogram regularities in nat-
ural scenes. Vision research. 2016; 120:11–21. https://doi.org/10.1016/j.visres.2015.03.018 PMID:
25872178
32. Foster DH, Amano K, Nascimento SMC, Foster MJ. Frequency of metamerism in natural scenes.
J Opt Soc Am A. 2006; 23(10):2359–2372. https://doi.org/10.1364/JOSAA.23.002359 PMID:
16985522
33. Nascimento SM, Amano K, Foster DH. Spatial distributions of local illumination color in natural
scenes. Vision research. 2016; 120:39–44. https://doi.org/10.1016/j.visres.2015.07.005 PMID:
26291072
34. Ciurea F, Funt B. A large image database for color constancy research. In: Color and Imaging Confer-
ence. vol. 2003. Society for Imaging Science and Technology; 2003. p. 160–164.
35. Adams WJ, Elder JH, Graf EW, Leyland J, Lugtigheid AJ, Muryy A. The southampton-york natural
scenes (SYNS) dataset: Statistics of surface attitude. Scientific reports. 2016; 6. https://doi.org/10.
1038/srep35805 PMID: 27782103
36. Tkačik G, Garrigan P, Ratliff C, Milčinski G, Klein JM, Seyfarth LH, et al. Natural images from the birth-
place of the human eye. PLoS one. 2011; 6(6):1–12. https://doi.org/10.1371/journal.pone.0020409
PMID: 21698187
37. Olmos A, Kingdom FA. A biologically inspired algorithm for the recovery of shading and reflectance
images. Perception. 2004; 33(12):1463–1473. https://doi.org/10.1068/p5321 PMID: 15729913
38. Girshick AR, Landy MS, Simoncelli EP. Cardinal rules: visual orientation perception reflects knowledge
of environmental statistics. Nature neuroscience. 2011; 14(7):926–932. https://doi.org/10.1038/nn.
2831 PMID: 21642976
39. Carbon CC, Pastukhov A. Reliable top-left light convention starts with Early Renaissance: An exten-
sive approach comprising 10k artworks. Frontiers in Psychology. 2018; 9:1–7. https://doi.org/10.3389/
fpsyg.2018.00454 PMID: 29686636
40. Sun J, Perona P. Where is the sun? Nature neuroscience. 1998; 1(3):183–184. https://doi.org/10.
1038/630 PMID: 10195141
41. Ramachandran VS. Perception of shape from shading. Nature. 1988; 331(6152):163–166. https://doi.
org/10.1038/331163a0 PMID: 3340162
42. Gibson JJ. The perception of the visual world. Houghton Mifflin; 1950.
43. Berbaum K, Bever T, Chung CS. Light source position in the perception of object shape. Perception.
1983; 12(4):411–416. https://doi.org/10.1068/p120411 PMID: 6672736
44. Mamassian P, Goutcher R. Prior knowledge on the illumination position. Cognition. 2001; 81(1):1–9.
https://doi.org/10.1016/S0010-0277(01)00116-0 PMID: 11525484
45. Saleh B, Elgammal A. Large-scale Classification of Fine-Art Paintings: Learning The Right Metric on
The Right Feature. International Journal for Digital Art History. 2016;(2). https://doi.org/10.11588/dah.
2016.2.23376
46. De La Rosa J, Sua
´rez JL. A quantitative approach to beauty. Perceived attractiveness of human faces
in world painting. International Journal for Digital Art History. 2015; 1.
47. Shen X, Efros AA, Aubry M. Discovering visual patterns in art collections with spatially-consistent fea-
ture learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recogni-
tion (CVPR); 2019. p. 9270–9279.
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 27 / 30
48. Sarı C, Salah AA, Akdag Salah AA. Automatic detection and visualization of garment color in Western
portrait paintings. Digital Scholarship in the Humanities. 2019; 34(Supplement_1):i156–i171. https://
doi.org/10.1093/llc/fqz055
49. Crowley EJ, Zisserman A. The state of the art: object retrieval in paintings using discriminative regions.
In: British Machine Vision Conference; 2014.
50. Augart I, Saß M, Wenderholm I. Steinformen. Berlin, Boston: De Gruyter; 2018.
51. Dietrich R. Rocks depicted in painting & sculpture. Rocks & Minerals. 1990; 65(3):224–236. https://doi.
org/10.1080/00357529.1990.11761676
52. Hollander A. Seeing through clothes. Univ of California Press; 1993.
53. Hollander A. Fabric of vision: dress and drapery in painting. Bloomsbury Publishing; 2016.
54. Birbari E. Dress in Italian painting, 1460-1500. J. Murray London; 1975.
55. Ribeiro A. The art of dress: fashion in England and France 1750 to 1820. vol. 104. Yale University
Press New Haven; 1995.
56. De Winkel M. Rembrandt’s clothes—Dress and meaning in his self-portraits. In: A Corpus of Rem-
brandt Paintings. Springer; 2005. p. 45–87.
57. Bol M, Lehmann AS. Painting skin and water: towards a material iconography of translucent motifs in
Early Netherlandish painting. In: Symposium for the Study of Underdrawing and Technology in Paint-
ing. Peeters; 2012. p. 215–228.
58. Lehmann AS. Fleshing out the rody: The’colours of the naked’ in workshop practice and art theory,
1400-1600. Nederlands Kunsthistorisch Jaarboek. 2008; 59:86.
59. Bambach C. Anatomy in the renaissance; 2002. Available from: https://www.metmuseum.org/toah/hd/
anat/hd_anat.htm.
60. Grootenboer H. The rhetoric of perspective: Realism and illusionism in seventeenth-century Dutch
still-life painting. University of Chicago Press; 2006.
61. Woodall J. Laying the table: The procedures of still life. Art History. 2012; 35(5):976–1003. https://doi.
org/10.1111/j.1467-8365.2012.00933.x
62. Taylor P. Dutch flower painting, 1600-1720. Yale University Press New Haven; 1995.
63. Adelson EH. On Seeing Stuff: The Perception of Materials by Humans and Machines. Proceedings of
the SPIE. 2001; 4299. https://doi.org/10.1117/12.429489
64. Fei-Fei L, Fergus R, Perona P. Learning generative visual models from few training examples: An
incremental bayesian approach tested on 101 object categories. In: Proceedings of the IEEE Confer-
ence on Computer Vision and Pattern Recognition (CVPR) workshop; 2004. p. 178–178.
65. Deng J, Dong W, Socher R, Li LJ, Li K, Fei-Fei L. Imagenet: A large-scale hierarchical image data-
base. In: 2009 IEEE conference on computer vision and pattern recognition; 2009. p. 248–255.
66. Krizhevsky A. Learning multiple layers of features from tiny images. (Technical Report) University of
Toronto. 2009.
67. Bell S, Upchurch P, Snavely N, Bala K. OpenSurfaces: A richly annotated catalog of surface
appearance. ACM Transactions on Graphics (TOG). 2013; 32. https://doi.org/10.1145/2461912.
2462002
68. Bell S, Upchurch P, Snavely N, Bala K. Material recognition in the wild with the materials in context
database. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
(CVPR); 2015. p. 3479–3487.
69. Caesar H, Uijlings J, Ferrari V. Coco-stuff: Thing and stuff classes in context. In: Proceedings of the
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); 2018. p. 1209–1218.
70. Visual Art Encyclopedia;. Available from: https://www.wikiart.org/en/about.
71. Bar Y, Levy N, Wolf L. Classification of artistic styles using binarized features derived from a deep neu-
ral network. In: European conference on computer vision. Springer; 2014. p. 71–84.
72. Elgammal A, Liu B, Kim D, Elhoseiny M, Mazzone M. The shape of art history in the eyes of the
machine. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol. 32; 2018.
73. Strezoski G, Worring M. Omniart: multi-task deep learning for artistic data analysis. arXiv preprint
arXiv:170800684. 2017.
74. Tan WR, Chan CS, Aguirre HE, Tanaka K. ArtGAN: Artwork synthesis with conditional categorical
GANs. In: IEEE International Conference on Image Processing (ICIP); 2017. p. 3760–3764.
75. Khan FS, Beigpour S, Van de Weijer J, Felsberg M. Painting-91: a large scale database for computa-
tional painting categorization. Machine vision and applications. 2014; 25(6):1385–1397. https://doi.
org/10.1007/s00138-014-0621-6
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 28 / 30
76. Mao H, Cheung M, She J. Deepart: Learning joint representations of visual arts. In: Proceedings of the
25th ACM international conference on Multimedia; 2017. p. 1183–1191.
77. Patel VM, Gopalan R, Li R, Chellappa R. Visual domain adaptation: A survey of recent advances.
IEEE Signal Processing Magazine. 2015; 32(3):53–69. https://doi.org/10.1109/MSP.2014.2347059
78. Wang M, Deng W. Deep visual domain adaptation: A survey. Neurocomputing. 2018; 312:135–153.
https://doi.org/10.1016/j.neucom.2018.05.083
79. Wilson G, Cook DJ. A survey of unsupervised deep domain adaptation. ACM Transactions on Intelli-
gent Systems and Technology (TIST). 2020; 11(5):1–46. https://doi.org/10.1145/3400066 PMID:
34336374
80. Jing Y, Yang Y, Feng Z, Ye J, Yu Y, Song M. Neural atyle transfer: A review. IEEE Transactions on
Visualization and Computer Graphics. 2020; 26(11):3365–3385. https://doi.org/10.1109/TVCG.2019.
2921336 PMID: 31180860
81. Geirhos R, Rubisch P, Michaelis C, Bethge M, Wichmann FA, Brendel W. ImageNet-trained CNNs are
biased towards texture; increasing shape bias improves accuracy androbustness. In: International Confer-
ence on Learning Representations; 2019.Available from: https://openreview.net/forum?id=Bygh9j09KX.
82. Sharan L, Liu C, Rosenholtz R, Adelson EH. Recognizing materials using perceptually inspired fea-
tures. International Journal of Computer Vision. 2013; 103(3):348–371. https://doi.org/10.1007/
s11263-013-0609-0 PMID: 23914070
83. Fleming RW. Material perception. Annual review of vision science. 2017; 3:365–388. https://doi.org/
10.1146/annurev-vision-102016-061429 PMID: 28697677
84. Stephen ID, Coetzee V, Perrett DI. Carotenoid and melanin pigment coloration affect perceived
human health. Evolution and Human Behavior. 2011; 32(3):216–227. https://doi.org/10.1016/j.
evolhumbehav.2010.09.003
85. Matts PJ, Fink B, Grammer K, Burquest M. Color homogeneity and visual perception of age, health,
and attractiveness of female facial skin. Journal of the American Academy of Dermatology. 2007; 57
(6):977–984. https://doi.org/10.1016/j.jaad.2007.07.040 PMID: 17719127
86. Igarashi T, Nishino K, Nayar SK. The appearance of human skin: A survey. vol. 3. Now Publishers
Inc; 2007.
87. Jensen HW, Marschner SR, Levoy M, Hanrahan P. A practical model for subsurface light transport. In:
Proceedings of SIGGRAPH. Los Angeles, Ca, USA; 2001.
88. Papadopoulos DP, Uijlings JRR, Keller F, Ferrari V. Extreme clicking for efficient object annotation.
International Journal of Computer Vision. 2017. https://doi.org/10.1109/ICCV.2017.528
89. Ren S, He K, Girshick R, Sun J. Faster r-cnn: Towards real-time object detection with region proposal
networks. In: Advances in neural information processing systems; 2015.
90. Wu Y, Kirillov A, Massa F, Lo WY, Girshick R. Detectron2; 2019. https://github.com/
facebookresearch/detectron2.
91. Miller J. On reflection. National Gallery Publications London; 1998.
92. Pacanowski R, Granier X, Schlick C, Pierre P. Sketch and paint-based interface for highlight modeling.
In: Fifth Eurographics conference on Sketch-Based Interfaces and Modeling; 2008.
93. Anjyo Ki, Hiramitsu K. Stylized highlights for cartoon rendering and animation. IEEE Computer Graph-
ics and Applications. 2003; 23(4):54–61. https://doi.org/10.1109/MCG.2003.1210865
94. Lin TY, Maire M, Belongie S, Hays J, Perona P, Ramanan D, et al. Microsoft coco: Common objects in
context. In: European conference on computer vision. Springer; 2014. p. 740–755.
95. Zhou B, Zhao H, Puig X, Fidler S, Barriuso A, Torralba A. Scene parsing through ade20k dataset. In:
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR);
2017. p. 633–641.
96. Cordts M, Omran M, Ramos S, Rehfeld T, Enzweiler M, Benenson R, et al. The cityscapes dataset for
semantic urban scene understanding. In: Proceedings of the IEEE/CVF Conference on Computer
Vision and Pattern Recognition (CVPR); 2016. p. 3213–3223.
97. Lin H, Upchurch P, Bala K. Block annotation: Better image annotation with sub-image decomposition.
In: Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV); 2019.
p. 5290–5300.
98. Maninis KK, Caelles S, Pont-Tuset J, Van Gool L. Deep extreme cut: From extreme points to object
segmentation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recog-
nition (CVPR); 2018. p. 616–625.
99. Benenson R, Popov S, Ferrari V. Large-scale interactive object segmentation with human annotators.
In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR);
2019. p. 11700–11709.
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 29 / 30
100. Ling H, Gao J, Kar A, Chen W, Fidler S. Fast interactive object annotation with curve-gcn. In: Proceed-
ings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); 2019.
p. 5257–5266.
101. Rother C, Kolmogorov V, Blake A. “GrabCut” interactive foreground extraction using iterated graph
cuts. ACM transactions on graphics (TOG). 2004; 23(3):309–314. https://doi.org/10.1145/1015706.
1015720
102. Wei XS, Wu J, Cui Q. Deep learning for fine-grained image analysis: A survey. arXiv preprint
arXiv:190703069. 2019.
103. Wah C, Branson S, Welinder P, Perona P, Belongie S. The caltech-ucsd birds-200-2011 dataset.
Computation & Neural Systems Technical Report, CNS-TR California Institute of Technology, Pasa-
dena, CA. 2011.
104. Van Horn G, Mac Aodha O, Song Y, Cui Y, Sun C, Shepard A, et al. The inaturalist species classifica-
tion and detection dataset. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pat-
tern Recognition (CVPR); 2018. p. 8769–8778.
105. Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-cam: Visual explanations
from deep networks via gradient-based localization. In: Proceedings of the IEEE/CVF International
Conference on Computer Vision (ICCV); 2017. p. 618–626.
PLOS ONE
Materials In Paintings
PLOS ONE | https://doi.org/10.1371/journal.pone.0255109 August 26, 2021 30 / 30
Available via license: CC BY 4.0
Content may be subject to copyright.