ArticlePDF Available

Face-to-BMI: Using Computer Vision to Infer Body Mass Index on Social Media

Authors:

Abstract and Figures

A person's weight status can have profound implications on their life, ranging from mental health, to longevity, to financial income. At the societal level, "fat shaming" and other forms of "sizeism" are a growing concern, while increasing obesity rates are linked to ever raising healthcare costs. For these reasons, researchers from a variety of backgrounds are interested in studying obesity from all angles. To obtain data, traditionally, a person would have to accurately self-report their body-mass index (BMI) or would have to see a doctor to have it measured. In this paper, we show how computer vision can be used to infer a person's BMI from social media images. We hope that our tool, which we release, helps to advance the study of social aspects related to body weight.
Content may be subject to copyright.
Face-to-BMI: Using Computer Vision to Infer Body Mass Index on Social Media
Enes Kocabey, Mustafa Camurcu, Ferda Ofli, Yusuf Aytar, Javier Marin, Antonio Torralba, Ingmar Weber
MIT-CSAIL, Northeastern University, Qatar Computing Research Institute, HBKU
{kocabey,yaytar,jmarin,torralba}@mit.edu, camurcu.m@husky.neu.edu, {fofli,iweber}@hbku.edu.qa
Abstract
A person’s weight status can have profound implications on
their life, ranging from mental health, to longevity, to finan-
cial income. At the societal level, “fat shaming” and other
forms of “sizeism” are a growing concern, while increasing
obesity rates are linked to ever raising healthcare costs. For
these reasons, researchers from a variety of backgrounds are
interested in studying obesity from all angles. To obtain data,
traditionally, a person would have to accurately self-report
their body-mass index (BMI) or would have to see a doctor
to have it measured. In this paper, we show how computer
vision can be used to infer a person’s BMI from social me-
dia images. We hope that our tool, which we release, helps to
advance the study of social aspects related to body weight.
Introduction
Together with a person’s gender, age and race, their weight
status is a publicly visible signal that can have profound in-
fluence on many aspects of their life. Most obviously, it can
affect their health as having a larger BMI is linked to an in-
creased risk of both cardio-vascular diseases and diabetes,
though not necessarily in a straight-forward manner (Meigs
and others 2006). However, other aspects of the burden im-
posed by obesity come in the form of “fat shaming” and
other forms of “sizeism”. For example, obesity is related to
a lower income1and part of the reason seems to be due to
weight-based discrimination (Puhl and others 2008). Even
among health professionals “sizeism” is so prevalent that it
has become a health hazard as, when faced with overweight
patients, care providers stop to look for alternative explana-
tions for a medical condition (Chrisler and Barney 2016).
For these reasons, researchers from a variety of back-
grounds are interested in studying obesity from all angles. To
obtain data, traditionally, a person would have to accurately
self-report their body-mass index (BMI) or would have to
see a doctor to have it measured. In this paper, we propose
a new pipeline using state-of-the-art computer vision tech-
niques to infer a person’s BMI from social media images,
This is a preprint of a short paper accepted at ICWSM’17.
Please cite that version instead.
1http://www.forbes.com/sites/
freekvermeulen/2011/03/22/the-price- of-
obesity-how- your-salary- depends-on- your-
weight/
such as their profile picture. We show that the performance
in distinguishing for a given pair the more overweight per-
son is similar to human performance.
Related Work
Being overweight can lead to a range of negative conse-
quences with the most direct ones concerning health. Given
this importance, recent studies in psychology and sociology
investigate how humans perceive health from profile pic-
tures. (Coetzee and others 2009) showed that facial adiposity
(i.e., perception of weight in the face) was a significant pre-
dictor of the perceived health. Furthermore, they showed that
perceived facial adiposity was significantly associated with
cardiovascular health and reported infections, and hence, an
important and valid cue to actual health. In a similar study,
(Henderson and others 2016) explored the effect of a variety
of facial characteristics on humans’ health judgment. They
found that facial features such as skin yellowness, mouth
curvature and shape were correlated positively whereas fa-
cial shape associated with adiposity was correlated nega-
tively with impression of health.
In light of these studies, (Weber and Mejova 2016) took a
crowdsourcing approach to understand health judgments of
humans in a body-weight-inference task from profile pic-
tures. However, since the judgment of whether a picture
“is overweight” is a rather subjective task, their work suf-
fered from the bias of human annotators to falsely equate
“overweight” with “abnormal.” To eliminate such limita-
tions, (Wen and Guo 2013) showed that it is indeed feasible
to some degree to predict BMI from face images automati-
cally using computational techniques. Their approach relied
on detecting a number of fiducial points in each face im-
age and computing hand-crafted geometric facial features to
train a regression model for BMI prediction. Their dataset,
however, comprised exclusively passport-style frontal face
photos with clean background, and hence, the performance
of their BMI prediction model is uncertain for noisy social
media pictures.
Faces with BMI Data
To ensure that our system works with noisy, often low qual-
ity social media pictures, such as profile pictures, we used
arXiv:1703.03156v1 [cs.HC] 9 Mar 2017
the set of annotated images from the VisualBMI project2.
These images are, in turn, collected from Reddit posts that
link to the imgur.com service. Examples of the underly-
ing Reddit posts can be found in the “progresspics” sub-
Reddit3. The VisualBMI dataset comprises a total of 16,483
images containing a pair of “before” and “after” images, an-
notated with gender, height and previous and current body
weights. We manually went through all of the image URLs,
and cropped the faces. We ignored all the images except the
ones with two faces, since we only had previous and cur-
rent body weights. After the manual cleaning process, we
were left with 2103 pairs of faces, with corresponding gen-
der, height and previous and current body weights. Then for
each pair, we computed the previous BMI and current BMI.
The BMI is defined as (body mass in kg) / (body height in
m)2.
This led to a total of 4206 faces with corresponding gen-
der and BMI information. Of these, seven were in the under-
weight range (16 <BMI 18.5), 680 were normal (18.5 <
BMI 25), 1151 were overweight (25 <BMI 30), 941
were moderately obese (30 <BMI 35), 681 were severely
obese (35 <BMI 40) and 746 were very severely obese
(40 <BMI). The dataset contained 2438 males and 1768 fe-
males. Figure 1 shows a selection of the faces that were used
for training and evaluating our system.
Figure 1: Examples of the cleaned images used for model
training. The black bars have been added to respect user pri-
vacy, but the model is learned on the original, public images.
Face-to-BMI System
In this section, we describe our Face-to-BMI system start-
ing with (i) the computer vision architecture used for
building the prediction model, and then (ii) the details of
the evaluation and comparison with human performance.
Our pre-trained models and scripts for using them can be
downloaded for academic research purposes at http://
face2bmi.csail.mit.edu.
2http://www.visualbmi.com/
3https://www.reddit.com/r/progresspics/
Computer Vision Architecture
Many computer vision tasks have greatly benefited from
the recent advances in deep learning (Parkhi and others
2015; Krizhevsky and others 2012; Simonyan and Zisser-
man 2014) and here we also utilize such models for the Face-
to-BMI problem. The features learned in deep convolutional
networks are proven to be transferable and quite effective
when used in other visual recognition tasks (Yosinski and
others 2014; Girshick and others 2014), particularly when
training samples are limited and learning a successful deep
model is not feasible due to overfitting. For instance, (Ozbu-
lak and others 2016) shows the success of this transfer for
age and gender recognition tasks performed on face images.
Considering that we also have limited training examples, we
adopted a transfer learning approach.
Our BMI prediction system is composed of two stages: (i)
deep feature extraction, and (ii) training a regression model.
For feature extraction we use two well-known deep models,
one trained on general object classification (i.e., VGG-Net
(Simonyan and Zisserman 2014)) and the other trained on
a face recognition task (i.e., VGG-Face (Parkhi and others
2015)). Both of these models are deep convolutional mod-
els with millions of parameters, and trained on millions of
images. The features from the fc6 layer are extracted for
each face image in our training set. For the BMI regression,
we use epsilon support vector regression models (Smola and
Vapnik 1997) due to its robust generalization behavior. The
models are trained on the 3368 training images and tested on
838 test images from the VisualBMI dataset. We make sure
that the same individual does not exist in both training and
test sets (e.g., before and after images). The performance of
both of our models are shown in Table 1. As also pointed out
by (Ozbulak and others 2016), the features extracted from a
more relevant model, i.e., VGG-Face, perform better com-
pared to the VGG-Net features. Due to its superiority we
use VGG-Face features in our Face-to-BMI system.
To see if our system could also be used for tracking weight
changes for a single person, rather than comparing across
people, we also defined a different train-test split. We first
randomly selected 838 unique individuals from our dataset.
For each of these individuals, we randomly selected one of
the two corresponding before-after face photos and added
them to a new test set. All the remaining pictures were added
to the new training set. In this way, every person that has a
face image in test set also had a face image in the training
set. We kept the training and test sizes the same to ensure a
fair comparison. Our model achieved 0.68 correlation, com-
pared to 0.65 in the across-people setup, suggesting that our
system benefits from having a history of images to train on
for the individual it is making a prediction for.
Human Evaluation
We conduct a simple experiment to compare our Face-to-
BMI system’s performance to that of humans. Given face
images for two individuals, each “contestant,” i.e., machine
and human, is required to tell which one is more overweight.
Note that our system was not trained for this specific binary
classification task though and a dedicated system might per-
form better.
Model Male Female Overall
Face-to-BMI – VGG-Net 0.58 0.36 0.47
Face-to-BMI – VGG-Face 0.71 0.57 0.65
Table 1: The Pearson rcorrelations on the test set for the
BMI prediction task, broken down by gender. Note that
VGG-Face features yield a much better performance.
For evaluation, we collect a total number of 900 pairs.
This is obtained by using only samples coming from the test
set, chosen such that pairs are equally distributed in gen-
der subcategories (‘male vs. male’, ‘female vs. female’ and
‘female vs. male’) and BMI difference between individuals
within a pair. Concretely, we randomly collect 300 ‘male
vs. male’ pairs consisting in 15 subsets of 20-pairs each,
such that, for each subset Si, with i∈ {0,...,14}, all pairs
(a, b)Sisatisfy:
(0.5 + i)<|BMIaBMIb| ≤ (1.5 + i)
We also collect 300 ‘female vs. female’ and ‘female vs.
male’ pairs following the same strategy. In the latter case,
we also assure that half of the pairs correspond to males be-
ing more overweight and vice-versa. Furthermore, we try to
balance the overall BMI distribution across the whole spec-
trum, from thin to obese.
On the human side, we perform the aforementioned ques-
tionnaire through Amazon Mechanical Turk4. Each question
is shown to three unique users gathering a total of 2700 an-
swers. The human performance is then obtained using all
the answers together and represented as the accuracy. We
did not apply a majority voting approach as we wanted to
evaluate the performance of anindividual human. On the ma-
chine side, for each question we compare the system output
of each individual included in the pair to obtain an answer.
Figure 2 depicts the results of comparison broken down
by the pair’s gender type and the absolute BMI difference
between individuals of each pair. The overall performance
difference between human and machine is less than 2%, but
when looking at different gender subcategories, there is a
bigger gap the ‘male vs. male’ comparisons, 5%. Humans
slightly outperform the machine for small BMI differences,
and there is almost no performance difference for larger BMI
differences.
Discussion
Algorithmic Bias
As demographic groups differ in their BMI distribution, it is
likely that, e.g., the race and the BMI are not independent
attributes in the visualBMI data. This could then mean that
our system perpetuates existing stereotypes. For example,
as African Americans have higher obesity rates in the US
population, an automated system might learn a prior proba-
bility that increases the likelihood of a person to be labeled
as obese simply based on their race.
4http://mturk.com
Figure 2: Human vs. Face-to-BMI comparison broken down
by gender and absolute BMI difference between individuals
of each pair.
To test if our system is biased in outputting a higher BMI
for a picture solely due to the person’s gender or race, we
paired users with a similar BMI, i.e., difference <1.0, but a
different gender or race. Furthermore, these pairs were con-
structed such that, in aggregate, each demographic group
had the same number of (slightly) higher BMIs. For pairs
with such close BMIs, an unbiased tool should pick mem-
bers of either group 50% of the time. Hence we check if our
tool creates a distribution in the output that differs statisti-
cally significantly from 50-50.
For 2000 male-female pairs from the test set the tool pre-
dicted a higher BMI for females in 1037 cases, p=.05.
Though the evidence is inconclusive, it is possible that our
system is slightly biased against females. To test for racial
bias we did not have a sufficient number of pairs in the test
set and hence had to include examples from the training set.
For 2000 White-African American pairs, our tool predicted
a higher BMI for Whites in 1085 cases, p < .05, hinting at
a small bias against Whites.
Ethical Considerations
Historically, 19th century phrenology5studied the potential
link between the shape of skull and moral attributes, apply-
ing a pseudo-scientific methodology to justify racism. Re-
cently, researchers in China have started to predict if a face
belongs to a criminal which, they claim, they can detect with
better-than-random accuracy. However, their work is largely
being viewed both as unethical and as scientifically flawed
(Biddle 2016).
Going beyond moral attributes and the shape of the skull,
psychologists have indeed shown that using only facial in-
formation it is possible for a person to perform better than
random chance at guessing another’s personality (Little and
Perrett 2007), an observation at the heart of physiognomy6.
Over the last couple of years, there has been a growing body
5https://en.wikipedia.org/wiki/Phrenology
6https://en.wikipedia.org/wiki/Physiognomy
of work that successfully applies computer vision techniques
to automatically infer a person’s personality in particular
from images shared on social media (Nie and others 2016;
Dhall and Hoey 2016; Guntuku and others 2015; Liu and
others 2016).
Most of the methods mentioned above work “better than
random guessing” but, when applied to a single individual,
are still highly unreliable. This is partly because concepts
such as personality are inherently vague and partly because
the connection to facial features is weak.
This caveat also largely applies to our tool which, despite
a performance similar to humans, is still noisy at the indi-
vidual level. However, at the population level it can be used
to detect relative trends as long as it is not biased in a sys-
tematic way. In other words, result of the form “the aver-
age BMI for group X is larger than for group Y” are far
more robust than results of the form “this individual from
group X has a higher BMI than this other individual from
group Y”. This distinction also applies to the BMI which
is useful for studying population health but has shortcom-
ings when used as a tool for individual health (Daniels 2009;
Prentice and Jebb 2001).
Conclusions
In this work, we apply the most recent computer vision tech-
niques to obtain a novel Face-to-BMI system. The perfor-
mance of this tool is on par to that of humans for distinguish-
ing the more overweight person when presented with a pair
of profile images. We discuss issues related to algorithmic
bias and ethical considerations when inferring information
from a person’s profile image. To limit the potential of abuse
while allowing others to replicate and build on our results,
we make our pre-trained models only available to academic
researchers after describing the intended use.
In future work, we will apply our method to social media
profile pictures to model population-level obesity rates. Pre-
liminary results show that both regional and demographic
differences in BMI are reflected in large amounts of Insta-
gram profile pictures.
References
[Biddle 2016] Biddle, S. 2016. Troubling study
says artificial intelligence can predict who will
be criminals based on facial features. The Inter-
cept.https://theintercept.com/2016/11/
18/troubling-study-says-artificial-
intelligence-can-predict-who-will-be-
criminals-based-on-facial-features/.
[Chrisler and Barney 2016] Chrisler, J. C., and Barney, A.
2016. Sizeism is a health hazard. Fat Studies 0(0):1–16.
[Coetzee and others 2009] Coetzee, V., et al. 2009. Facial
adiposity: A cue to health? Perception 38(11):1700–1711.
[Daniels 2009] Daniels, S. R. 2009. The use of bmi in the
clinical setting. Pediatrics 124(Supplement 1):S35–S41.
[Dhall and Hoey 2016] Dhall, A., and Hoey, J. 2016. First
impressions - predicting user personality from twitter profile
images. In HBU, 148–158.
[Girshick and others 2014] Girshick, R., et al. 2014. Rich
feature hierarchies for accurate object detection and seman-
tic segmentation. In CVPR, 580–587.
[Guntuku and others 2015] Guntuku, S. C., et al. 2015. Do
others perceive you as you want them to?: Modeling person-
ality based on selfies. In ASM, 21–26.
[Henderson and others 2016] Henderson, A. J., et al. 2016.
Perception of health from facial cues. Philosophical Trans-
actions of the Royal Society of London B: Biological Sci-
ences 371(1693).
[Krizhevsky and others 2012] Krizhevsky, A., et al. 2012.
Imagenet classification with deep convolutional neural net-
works. In NIPS, 1097–1105.
[Little and Perrett 2007] Little, A. C., and Perrett, D. I. 2007.
Using composite images to assess accuracy in personality at-
tribution to faces. British Journal of Psychology 98(1):111–
126.
[Liu and others 2016] Liu, L., et al. 2016. Analyzing person-
ality through social media profile picture choice. In ICWSM,
211–220.
[Meigs and others 2006] Meigs, J. B., et al. 2006. Body mass
index, metabolic syndrome, and risk of type 2 diabetes or
cardiovascular disease. The Journal of Clinical Endocrinol-
ogy & Metabolism 91(8):2906–2912.
[Nie and others 2016] Nie, J., et al. 2016. Social media pro-
filer: Inferring your social media personality from visual at-
tributes in portrait. In PCM, 640–649.
[Ozbulak and others 2016] Ozbulak, G., et al. 2016. How
transferable are cnn-based features for age and gender clas-
sification? In BIOSIG, 1–6.
[Parkhi and others 2015] Parkhi, O. M., et al. 2015. Deep
face recognition. In British Machine Vision Conference.
[Prentice and Jebb 2001] Prentice, A. M., and Jebb, S. A.
2001. Beyond body mass index. Obesity reviews 2(3):141–
147.
[Puhl and others 2008] Puhl, R. M., et al. 2008. Perceptions
of weight discrimination: prevalence and comparison to race
and gender discrimination in america. Int J Obes 32:992–
1000.
[Simonyan and Zisserman 2014] Simonyan, K., and Zisser-
man, A. 2014. Very deep convolutional networks for large-
scale image recognition. arXiv preprint arXiv:1409.1556.
[Smola and Vapnik 1997] Smola, A., and Vapnik, V. 1997.
Support vector regression machines. NIPS 9:155–161.
[Weber and Mejova 2016] Weber, I., and Mejova, Y. 2016.
Crowdsourcing health labels: Inferring body weight from
profile pictures. In DH, 105–109.
[Wen and Guo 2013] Wen, L., and Guo, G. 2013. A compu-
tational approach to body mass index prediction from face
images. Image and Vision Computing 31:392–400.
[Yosinski and others 2014] Yosinski, J., et al. 2014. How
transferable are features in deep neural networks? In NIPS,
3320–3328.
... Recently, body weight and BMI have attracted a lot of attention in applications involving health monitoring and weight loss [7]- [9] as a result of the alarming trend in obesity. BMI is defined as (Body Mass in Kilograms)/(Body Height in Meters) 2 . ...
... To facilitate this, CNN models namely VGG16, ResNet50, Xception, and Efficient-NetB0 for BMI prediction are compared with our proposed PatchBMI-Net model in terms of performance, size, and inference time when deployed on a smartphone. • Experimental investigation of three publicly available datasets namely, VisualBMI [9], IllinoisDOC [20], and FIW-BMI [21] consists of facial images annotated with BMI information. ...
... Kocabey et al. [9] proposed a facial-image-based BMI prediction model based on the Support Vector Regression trained on deep features extracted from the VGG model. Experimental investigations on the VisualBMI dataset suggested a Pearson correlation of 0.71, 0.57, and 0.65 for males, females, and overall, respectively. ...
Conference Paper
Full-text available
Due to an alarming trend related to obesity affecting 93.3 million adults in the United States alone, body mass index (BMI) and body weight have drawn significant interest in various health monitoring applications. Consequently, several studies have proposed self-diagnostic facial image-based BMI prediction methods for healthy weight monitoring. These methods have mostly used convolutional neural network (CNN) based regression baselines, such as VGG19, ResNet50, and Efficient-NetB0, for BMI prediction from facial images. However, the high computational requirement of these heavyweight CNN models limits their deployment to resource-constrained mobile devices, thus deterring weight monitoring using smartphones. This paper aims to develop a lightweight facial patch-based ensemble (PatchBMI-Net) for BMI prediction to facilitate the deployment and weight monitoring using smartphones. Extensive experiments on BMI-annotated facial image datasets suggest that our proposed PatchBMI-Net model can obtain Mean Absolute Error (MAE) in the range [3.58, 6.51] with a size of about 3.3 million parameters. On cross-comparison with heavyweight models, such as ResNet-50 and Xception, trained for BMI prediction from facial images, our proposed PatchBMI-Net obtains equivalent MAE along with the model size reduction of about 5.4× and the average inference time reduction of about 3× when deployed on Apple-14 smartphone. Thus, demonstrating performance efficiency as well as low latency for on-device deployment and weight monitoring using smartphone applications.
... predicting specific features from human face images and converting them into measurable indicators of representation (Kocabey et al., 2017;Jiang et al., 2019;Yousaf et al., 2021;Guo & Mu, 2013;Fu et al., 2014). Advances in these areas have lessened the challenge of automated visual content analysis expanding the frontier of computer capabilities. ...
... 17 The Face-to-BMI prediction system proposed by Kocabey et al. (2017) was trained and tested using 4,206 faces extracted from the VisualBMI dataset, which comprises a total of 16,483 images collected from subreddit progress-pics community on Reddit containing a pair "before" and "after" body transformation images. Each image in the dataset is annotated with gender, height, and previous and current body weights. ...
... However, health records of individual patients are not widely available, not least because of privacy concerns. By contrast, a variety of Web sites have been proven to be a good source of data for public health surveillance: search query logs have been used to forecast the spreading of influenza epidemics [13]; microblogging platforms such as Twitter have been used to monitor public health at scale [14,15] and to estimate the prevalence of a wide range of pathologies, from mental illnesses [16] to obesity [17]; and pictures on social media have been used to estimate values of Body Mass Index (BMI) [18,19]. ...
Preprint
To complement traditional dietary surveys, which are costly and of limited scale, researchers have resorted to digital data to infer the impact of eating habits on people's health. However, online studies are limited in resolution: they are carried out at regional level and do not capture precisely the composition of the food consumed. We study the association between food consumption (derived from the loyalty cards of the main grocery retailer in London) and health outcomes (derived from publicly-available medical prescription records). The scale and granularity of our analysis is unprecedented: we analyze 1.6B food item purchases and 1.1B medical prescriptions for the entire city of London over the course of one year. By studying food consumption down to the level of nutrients, we show that nutrient diversity and amount of calories are the strongest predictors of the prevalence of three diseases related to what is called the "metabolic syndrome": hypertension, high cholesterol, and diabetes. This syndrome is a cluster of symptoms generally associated with obesity, is common across the rich world, and affects one in four adults in the UK. Our linear regression models achieve an R2 of 0.6 when estimating the prevalence of diabetes in nearly 1000 census areas in London, and a classifier can identify (un)healthy areas with up to 91% accuracy. Interestingly, healthy areas are not necessarily well-off (income matters less than what one would expect) and have distinctive features: they tend to systematically eat less carbohydrates and sugar, diversify nutrients, and avoid large quantities. More generally, our study shows that analytics of digital records of grocery purchases can be used as a cheap and scalable tool for health surveillance and, upon these records, different stakeholders from governments to insurance companies to food companies could implement effective prevention strategies.
Preprint
1 Summary Changes in body mass are a key indicator of health and disease in humans and model organisms. Animal body mass is routinely monitored in husbandry and preclinical studies. In rodent studies, the current best method requires manually weighing the animal on a balance which has at least two consequences. First, direct handling of the animal induces stress and can have confounding effects on studies. Second, the acquired mass is static and not amenable to continuous assessment, and rapid mass changes can be missed. A noninvasive and continuous method of monitoring animal mass would have utility in multiple areas of biomedical research. Here, we test the feasibility of determining mouse body mass using video data. We combine computer vision methods with statistical modeling to demonstrate the feasibility of our approach. Our methods determine mouse mass with 4.8% error across highly genetically diverse mouse strains, with varied coat colors and mass. This error is low enough to replace manual weighing with image-based assessment in most mouse studies. We conclude that visual determination of rodent mass using video enables noninvasive and continuous monitoring and can improve animal welfare and preclinical studies.
Chapter
One of the most concerning public health issues the world is currently dealing with is obesity. Body mass index (BMI) is used to measure one’s obesity and is useful for evaluating an individual’s health. BMI is calculated by dividing a person’s weight by the square of their height. Calculating BMI from just a facial image has been a challenge in the field of computer vision, and researchers have used various methodologies. In this paper, various methods used by different researchers have been described and summarized according to the research objectives. Researchers have used two major methods for extracting features from facial images, i.e., extracting geometric features and deep features. According to our research, region-aware global average pooling together with deep features were recommended, and training of dataset based on gender and race was recommended which improves accuracy and gives the best results for estimation of BMI from facial images.
Conference Paper
Full-text available
To use social media for health-related analysis, one key step is the detection of health-related labels for users. But unlike transient conditions like flu, social media users are less vocal about chronic conditions such as obesity, as users might not tweet ``I'm still overweight''. As, however, obesity-related conditions such as diabetes, heart disease, osteoarthritis, and even cancer are on the rise, this obese-or-not label could be one of the most useful for studies in public health. In this paper we investigate the feasibility of using profile pictures to infer if a user is overweight or not. We show that this is indeed possible and further show that the fraction of labeled-as-overweight users is higher in U.S. counties with higher obesity rates. Going from public to individual health analysis, we then find differences both in behavior and social networks, for example finding users labeled as overweight to have fewer followers.
Article
Full-text available
Impressions of health are integral to social interactions, yet poorly understood. A review of the literature reveals multiple facial characteristics that potentially act as cues to health judgements. The cues vary in their stability across time: structural shape cues including symmetry and sexual dimorphism alter slowly across the lifespan and have been found to have weak links to actual health, but show inconsistent effects on perceived health. Facial adiposity changes over a medium time course and is associated with both perceived and actual health. Skin colour alters over a short time and has strong effects on perceived health, yet links to health outcomes have barely been evaluated. Reviewing suggested an additional influence of demeanour as a perceptual cue to health. We, therefore, investigated the association of health judgements with multiple facial cues measured objectively from two-dimensional and three-dimensional facial images. We found evidence for independent contributions of face shape and skin colour cues to perceived health. Our empirical findings: (i) reinforce the role of skin yellowness; (ii) demonstrate the utility of global face shape measures of adiposity; and (iii) emphasize the role of affect in facial images with nominally neutral expression in impressions of health. © 2016 The Author(s) Published by the Royal Society. All rights reserved.
Conference Paper
Full-text available
In this work, selfies (self-portrait images) of users are used to computationally predict and understand their personality. For users to convey a certain impression with selfie, and for the observers to build a certain impression about the users, many visual cues play a significant role. It is interesting to analyse what these cues are and how they influence our understanding of personality profiles. Selfies of users (from a popular microblogging site, Sina Weibo) were annotated with mid-level cues (such as presence of duckface, if the user is alone, emotional positivity etc.) relevant to portraits (especially selfies). Low-level visual features were used to train models to detect these mid-level cues, which are then used to predict users' personality (based on Five Factor Model). The mid-level cue detectors are seen to outperform state-of-the-art features for most traits. Using the trained computational models, we then present several insights on how selfies reflect their owners' personality and how users' are judged by others based on their selfies.
Article
The content of images users post to their social media is driven in part by personality. In this study, we analyze how Twitter profile images vary with the personality of the users posting them. In our main analysis, we use profile images from over 66,000 users whose personality we estimate based on their tweets. To facilitate interpretability, we focus our analysis on aesthetic and facial features and control for demographic variation in image features and personality. Our results show significant differences in profile picture choice between personality traits, and that these can be harnessed to predict personality traits with robust accuracy. For example, agreeable and conscientious users display more positive emotions in their profile pictures, while users high in openness prefer more aesthetic photos.
Conference Paper
Can a large convolutional neural network trained for whole-image classification on ImageNet be coaxed into detecting objects in PASCAL? We show that the answer is yes, and that the resulting system is simple, scalable, and boosts mean average precision, relative to the venerable deformable part model, by more than 40% (achieving a final mAP of 48% on VOC 2007). Our framework combines powerful computer vision techniques for generating bottom-up region proposals with recent advances in learning high-capacity convolutional neural networks. We call the resulting system R-CNN: Regions with CNN features. The same framework is also competitive with state-of-the-art semantic segmentation methods, demonstrating its flexibility. Beyond these results, we execute a battery of experiments that provide insight into what the network learns to represent, revealing a rich hierarchy of discriminative and often semantically meaningful features.
Conference Paper
In this paper, we introduce an interesting but challenging problem: how to infer social media personality from portrait. To address this problem, we jointly consider social media content and behavior information. Specifically, first, we represent social media personality as a reflection in accordance with user behaviors in social media. Second, by means of clustering, people are divided into eight groups and labeled with different personality types. Upon regression analysis, discriminative visual attributes for personality classification are determined. Third, low-level features of selected visual attributes are trained to predict personality from given portrait. To evaluate the proposed method, we collect images of people from the internet and the behaviors of these people from their micro-blog. Comprehensive experiments demonstrate that the proposed method can achieve significant performance gain over the existing method.
Conference Paper
This paper proposes a computer vision based pipeline for inferring the perceived personality of users from their Twitter profile images. We humans make impressions on a daily basis during communication. The perception of personality of a person gives information about the person’s behaviour and is an important attribute in developing rapport. The personality assessment in this paper is referred to as first impressions, which is similar to how humans create a mental image of another person by just looking at their profile pictures. In the proposed automated pipeline, hand crafted (engineered) and learnt feature descriptors are computed on user profile images. The effect of image background is assessed on the perception of the personality from a profile picture. A multivariate regression approach is used to predict the big five personality traits - agreeableness, conscientiousness, extraversion, openness and neuroticism. We study the correlation between the big five personality traits generated from Tweet analysis with the proposed profile image based framework. The experiments show high correlation for scene based first impressions perception. It is interesting to note that the results generated by analysing a profile image uploaded by a user in a particular point in time are in sync with the first impression traits generated by investigating Tweets posted over a longer duration of time.
Article
We trained a large, deep convolutional neural network to classify the 1.2 million high-resolution images in the ImageNet LSVRC-2010 contest into the 1000 dif-ferent classes. On the test data, we achieved top-1 and top-5 error rates of 37.5% and 17.0% which is considerably better than the previous state-of-the-art. The neural network, which has 60 million parameters and 650,000 neurons, consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax. To make train-ing faster, we used non-saturating neurons and a very efficient GPU implemen-tation of the convolution operation. To reduce overfitting in the fully-connected layers we employed a recently-developed regularization method called "dropout" that proved to be very effective. We also entered a variant of this model in the ILSVRC-2012 competition and achieved a winning top-5 test error rate of 15.3%, compared to 26.2% achieved by the second-best entry.
Article
Can a large convolutional neural network trained for whole-image classification on ImageNet be coaxed into detecting objects in PASCAL? We show that the answer is yes, and that the resulting system is simple, scalable, and boosts mean average precision, relative to the venerable deformable part model, by more than 40% (achieving a final mAP of 48% on VOC 2007). Our framework combines powerful computer vision techniques for generating bottom-up region proposals with recent advances in learning high-capacity convolutional neural networks. We call the resulting system R-CNN: Regions with CNN features. The same framework is also competitive with state-of-the-art semantic segmentation methods, demonstrating its flexibility. Beyond these results, we execute a battery of experiments that provide insight into what the network learns to represent, revealing a rich hierarchy of discriminative and often semantically meaningful features.