Available via license: CC BY 4.0
Content may be subject to copyright.
Vol.:(0123456789)
Electronic Markets (2024) 34:14
https://doi.org/10.1007/s12525-024-00699-y
DISCUSSION PAPER
Manipulation bydesign
JanTrzaskowski1
Received: 13 August 2023 / Accepted: 31 January 2024
© The Author(s) 2024
Abstract
Human behaviour is affected by architecture, including how online user interfaces are designed. The purpose of this article
is to provide insights into the regulation of behaviour modification by the design of choice architecture in light of the Euro-
pean Union data protection law (GDPR) and marketing law (UCPD). It has become popular to use the term ‘dark pattern’
(also ‘deceptive practices’) to describe such practices in online environments. The term provides a framework for identify-
ing and discussing ‘problematic’ design practices, but the definitions and descriptions are not sufficient in themselves to
draw the fine line between legitimate (lawful) persuasion and unlawful manipulation, which requires an inquiry into agency,
self-determination, regulation and legal interpretation. The main contribution of this article is to place manipulative design,
including ‘dark patterns’, within the framework of persuasion (marketing), technology (persuasive technology) and law
(privacy and marketing).
Keywords Manipulation· Design· Dark patterns· Marketing· Privacy· Agency
JEL classification K24
Introduction
Human behaviour online can be manipulated by platform
architecture, including how online user interfaces are
designed. We study the need for greater regulation of user
interfaces that platforms design specifically to modify users’
behaviour in ways that most users cannot detect. We study
the regulation of ‘choice manipulation architecture’ in the
context of existing regulations within the European Union,
particularly the EU’s data protection law (GDPR) and mar-
keting law (UCPD). It has become common to use the term
‘dark patterns’ (also ‘deceptive practices’) to describe such
manipulation in online environments. The term provides a
framework for identifying and discussing design practices
that businesses and regulators should consider problematic,
but the definitions and descriptions are not sufficient in them-
selves to draw the delicate distinctions between legitimate
and lawful persuasion and deceptive and unlawful manipula-
tion, which requires an understanding of users’ free choice,
agency and self-determination. It also requires legal inter-
pretation of diverse sets of regulation. The main contribu-
tion of this article is to place manipulative design, including
‘dark patterns’, within the frameworks of persuasion within
marketing, persuasive technology (captology) and the laws
governing privacy and marketing. We advance our under-
standing of online manipulation through design, in order to
better inform regulation and business practices (this article is
based on Trzaskowski, 2021a, and Trzaskowski,2023).
‘Dark patterns’
The term was coined in 2010 by user experience designer
Harry Brignull, and it has entered into legislation (e.g. the
California Consumer Privacy Act, the Colorado Privacy
Act and the Digital Services Act, discussed below), policy
documents (e.g. Forbrukerrådet, 2018; OECD, 2022; BEUC,
2022; EDPB, 2022), legal research (e.g. Luguri & Strahile-
vitz, 2021; Leiser & Caruana, 2021; Jarovsky, 2022) and
mainstream media (Nahai,2015). Dark patterns may be
defined as ‘tricks used in websites and apps that make you
do things that you didn’t mean to, like buying or signing up
Responsible Editor: Maximilian Schreieck
* Jan Trzaskowski
jan@humandignity.ai
1 Copenhagen Business School andAalborg University,
Copenhagen, Denmark
Electronic Markets (2024) 34:14 14 Page 2 of 13
for something’ (www. decep tive. design), or as in the Cali-
fornia Consumer Privacy Act and the Colorado Privacy Act:
‘Dark pattern’ means a user interface designed or
manipulated with the substantial effect of subverting or
impairing user autonomy, decision-making, or choice.
In both laws, an agreement obtained through the use of
dark patterns does not constitute ‘consent’.
In a recent case (Federal Trade Commission v. Amazon.
com, inc., 2023), the US Federal Trade Commission claims
that Amazon has (1) ‘used manipulative, coercive, or decep-
tive user interface designs known as “dark patterns” to trick
consumers into enrolling in automatically-renewing Prime
subscriptions’ (‘Nonconsensual Enrollment’) and (2) ‘know-
ingly complicated the cancellation process for Prime sub-
scribers who sought to end their membership’ (‘the Iliad
Flow’), i.e. ‘complexity resulted from Amazon’s use of dark
patterns—manipulative design elements that trick users into
making decisions they would not otherwise have made’. (See
also CNIL,2022 (fining Microsoft Bing €60M for making it
more difficult to reject than accept cookies) and FTC 2022a
(agreement with Epic Games (Fortnite) on $275M in pen-
alty and $245M in compensation for privacy-invasive default
settings and deceptive interfaces for in-game purchases).)
In research, much focus has been on creating taxonomies
of dark patterns, which may be grouped into eight categories
(and 27 variants) (Luguri & Strahilevitz, 2021, p. 53 with
references): nagging, social proof, obstruction, sneaking,
interface interference, forced action, scarcity and urgency.
Such taxonomies also allow for automated identification and
assessment of such design practices (Mathur etal., 2019).
In the EU, several similar taxonomies have been provided,
in which dark patterns have been categorised as, for instance,
nagging, social proof (endorsements), obstruction, sneaking
(information hiding), interface interference, forced (coerced)
action, urgency (scarcity) and asymmetric choice (European
Commission, 2022; FTC 2022b). In the context of data pro-
tection, words like overloading, skipping, stirring, hindering,
fickle and left in the dark (EDPB 2023) have been used, as
well as categories, including enjoy, seduce, lure, complicate
and ban (CNIL,2019).
The important part in these definitions and descriptions
is the extent to which autonomy is impaired, i.e. you are
‘tricked’ or manipulated into doing ‘things that you didn’t
mean to’.
We use the terms ‘behaviour modification’ (Bandura,
1969) and ‘influence’ interchangeably and understand them
to comprise both persuasion and manipulation. We use
persuasion as the legitimate (lawful) form of influencing
behaviour and manipulation—which includes deception
and coercion—as unlawful influencing (Rushkoff, 1999,
p. 270; Wood, 2014, chapter12). The distinction between
persuasion and manipulation is normative and not always
easy to establish, as discussed below. When a user interface
is designed for unlawful behaviour modification, it consti-
tutes ‘manipulation by design’.
We focus on data protection law where ‘consent’ must
reflect the data subject’s genuine and informed choice and
marketing law where the aim is to ensure that ‘commercial
practices’ do not impair the consumer’s ability to make free
and informed decisions. Manipulation by design—includ-
ing by the use of ‘dark patterns’—is also important in, for
instance, contract law and fraud, which are not dealt with
here.
Markets andtheright toself‑determination
Markets andmarket failures
The overarching idea behind market economies is that effi-
cient competition guided by price signals affected by supply
and demand yields better economic outcomes for consumers.
Neoclassical economics works with three basic assumptions:
People have rational preferences among outcomes that can
be identified and associated with an expected value. Indi-
viduals maximise utility (as consumers) and firms maximise
profit (as producers). People act independently on the basis
of full and relevant information.
The rational choice theory rests on the assumption that
aggregate social behaviour results from the behaviour of
individual actors who make decisions in accordance with
their preferences. This rational agent is expected to take into
account all relevant information, potential costs and benefits
etc. in order to act in accordance with their individual goals,
values and preferences. This theory assumes that consum-
ers’ decisions reveal their individual preferences so as to
maximise their welfare (Akerlof & Shiller, 2015, p. 170).
There are instances where markets cannot in themselves
ensure efficiency. ‘Market failures’ are situations in which
the allocation of goods and services is not efficient such
as when sellers’ (in the following, we use the term ‘trader’
as is common practice in EU law, and which in the GDPR
corresponds to the term ‘data controller’) pursuit of pure
self-interest leads to results that are not efficient. Market
failures may be corrected by means of market regulation,
and the field of consumer protection law seeks to correct
the market failure stemming from an asymmetry in power,
including information.
The market regulation of the European single market (the
internal market)—including consumer protection law, which
is to ensure a high level of consumer protection—usually
has as an additional purpose of ensuring efficiency in the
Electronic Markets (2024) 34:14 Page 3 of 13 14
market by (a) only disturbing it to the extent necessary and
(b) removing barriers to inter-state trade created by differ-
ences in law (harmonisation).
Marketing law
The Unfair Commercial Practices Directive (UCPD; Direc-
tive 2005/29/EC) applies to ‘business-to-consumer commer-
cial practices’, which is a deliberately broad concept (CJEU
cases C-59/12 and C-559/11) encompassing practices car-
ried out before, during and after a commercial transaction
(Article3). The concept (referred to as ‘commercial prac-
tices’) is defined in Article2(1)(d) as:
any act, omission, course of conduct or representa-
tion, commercial communication including advertising
and marketing, by a trader, directly connected with the
promotion, sale or supply of a product to consumers.
Article5(1) prohibits unfair commercial practices, and
Article5(2) sets two cumulative requirements that must
be fulfilled for a commercial practice to be deemed unfair.
Thus, a commercial practice is unfair if:
• it is contrary to the requirements of ‘professional dili-
gence’, and
• it materially distorts or is likely to materially distort the eco-
nomic behaviour of the average consumer with regard to the
product (hereafter referred to as ‘economic distortion’).
The layout of this general prohibition is important in
order to understand the structure of the directive, even
though most cases are likely to be determined by the more
specific prohibitions concerning misleading practices (Arti-
cles 6 and 7), aggressive practices (Articles 8 and 9) and
blacklisted practices (Article 5(5) and Annex I).
In practice, one must, in order to determine whether a
commercial practice is lawful, firstly, consult the items on
the blacklist; secondly, consider whether the practice is mis-
leading and/or aggressive; and thirdly, consider whether the
practice is otherwise contrary to the requirements of profes-
sional diligence.
If the practice is considered to be ‘otherwise contrary to
the requirements of professional diligence’, economic distor-
tion must also be considered. If the practice is found to be
misleading and/or aggressive, it must only be determined
whether that practice causes or is likely to cause the average
consumer to take a transactional decision that he would not
have taken otherwise (‘economic effect’). Such an analysis
of the economic effect/distortion is not required for black-
listed commercial practices.
It is important to emphasise that the UCPD—in order
to seek a generally applied, objective standard—makes no
reference to the trader’s subjective intention behind a com-
mercial practice (CJEU Case C-388/13, paras 47–48).
Data protection law
As data protection law is also important for business-to-
consumer interaction, the General Data Protection Regu-
lation (GDPR; Regulation (EU) 2016/679) must—in addi-
tion to providing democratic safeguards—be perceived as
an important pillar of consumer protection law. In contrast
to marketing law, the respect for privacy and protection of
personal data are fundamental rights rooted in the Charter
of Fundamental Rights in the European Union (the Charter),
Articles 7 and 8, respectively.
The GDPR can be distilled into six overarching princi-
ples that require that the data controller must—subject to the
principle of proportionality—ensure legitimacy, transpar-
ency and security to demonstrate accountability and ensure
empowerment of the data subject (Trzaskowski,2021b). The
principles overlap and may, ultimately, be reduced to a mere
matter of ‘due diligence’.
In this context, we focus on consent—which constitutes
an important part of empowerment—defined in Article
4(11):
“consent” of the data subject means any freely given,
specific, informed and unambiguous indication of the
data subject’s wishes by which he or she, by a state-
ment or by a clear affirmative action, signifies agree-
ment to the processing of personal data relating to him
or her’ (emphasis added).
Interplay betweentheGDPR andtheUCPD
Arguments from consumer protection law may play a role
in data protection law, as the latter requires the process-
ing of personal data to be ‘fair’ and ‘lawful’. The reverse
is also true, as consumer protection law aims at striking a
‘fair balance’ between traders and consumers, a balance that
can be affected by the trader’s processing of personal data
(Calo,2014; Borgesius etal., 2017; Svantesson, 2018; Zar-
sky, 2019).
The non-binding UCPD Guidance by the Commission
suggests that (European Commission 2019, section1.2.10):
[a] violation of the GDPR or of the ePrivacy Direc-
tive will not, in itself, always mean that the practice
is also in breach of the UCPD. However, such privacy
and data protection violations should be considered
when assessing the overall unfairness of commercial
practices under the UCPD, particularly in the situation
where the trader processes consumer data in violation
of privacy and data protection requirements, i.e. for
Electronic Markets (2024) 34:14 14 Page 4 of 13
direct marketing purposes or any other commercial
purposes like profiling, personal pricing or big data
applications.
When it comes to ‘professional diligence’—i.e. the spe-
cial skill and care which a trader may reasonably be expected
to exercise—this is likely to include issues pertaining to
privacy. For instance, the United Nations Guidelines for
Consumer Protection (2016) (UN, 2015), concerning prin-
ciples for good business practices, provides that ‘businesses
should protect consumers’ privacy through a combination
of appropriate control, security, transparency and consent
mechanisms relating to the collection and use of their per-
sonal data’.
It remains fair to say that data protection law provides for
a much tighter, more coherent and more robust framework to
further its aims compared to the field of consumer law. The
UCPD shares the principles of empowerment, proportional-
ity and transparency with the GDPR, but the UCPD does not
contain similar requirements for legitimacy, accountability
and security.
Empowerment
From the dawn of the European Union consumer policy,
the focus was on enabling ‘consumers, as far as possible,
to make better use of their resources, to have a freer choice
between the various products or services offered’ (European
Council 1975a; European Council 1975b, para 8). One of
the main priorities was to ensure protection against ‘forms
of advertising which encroach on the individual freedom of
consumers’ (European Council 1975b, para 30).
Human agency and the right to self-determination are
central concepts in legal theory as well as in consumer pro-
tection law, where the regulatory framework is aimed at
empowering consumers to act in accordance with their pref-
erences (e.g. European Commission 2007). Empowerment
also permeates data protection law, with consent possibly
being the clearest example that also elucidates the interplay
with transparency.
Empowerment in data protection law does not mean that
data subjects have absolute control over what data are being
(or can be) processed about them, nor by whom. The pro-
cessing of personal data must be ‘lawful’, which must require
that the activity also be in compliance with the UCPD.
In short, empowerment must entail (1) human agency
(some sort of free will, see also Kreps & Rowe, 2021), (2)
a sufficient degree of transparency and (3) the absence of
manipulation (Trzaskowski, 2021a). In the following, the
primary focus will be on the absence of manipulation, and
it may be helpful to emphasise that (a) it is widely accepted
that human agency is a limited quantity—as recognised
in, for instance, ‘bounded rationality’ (Kahneman 2003;
Jones 1999) and ‘ego depletion’ (Baumeister & Tierney,
2011)—and (b) information does not equal transparency;
i.e. to determine transparency, both the trader’s encoding
and the user’s reasonable decoding of information must
be determined (Trzaskowski,2021a, chapters6 and 7 with
references).
Behaviour modication
In data-driven business models that rely on personal-
isedadvertising, including targeted advertising, the trader
has an economic incentive to increase (1) the number of
users, (2) their engagement (amount and nature of atten-
tion) and (3) knowledge about individual users (personal
data). As we will discuss below, personal data coupled with
insights from psychology and technology can be used to
(a) increase the value of user experiences (through effec-
tive influence) and (b) increase the amount of attention and
nature of engagement, including by means of creating addic-
tive technology. Artificial intelligence (AI) plays an impor-
tant role in optimising behaviour modification for the most
profitable impact.
Rational decisions andbehavioural insights
As our focus is on the regulation of markets, it is important
to emphasise that economic theories underpinning mar-
kets—in addition to free choice—assume that consumers
are able to make informed/efficient/rational choices. The
economic theory usually applies a thin rationality (‘revealed
preferences’ as introduced above) that disregards value shap-
ing, adaptive preferences and the interest of future genera-
tions (Elster, 2016, p. 36), as well as the fact that people
might prefer to do something other than spending time on
maximising their economic interests (Taleb, 2010, p. 184).
Behavioural economics (see, e.g. Posner, 1997; Jolls
etal., 1998) revolves around the economic consequences
of the continuous stream of studies providing ever more
fine-grained knowledge about human behaviour in general
and human decision-making in particular, with a view to
adjusting neoclassical economics for these insights found
in behavioural sciences.
Market failures that rest on biased demand, generated by
imperfectly rational consumers, have been labelled ‘behav-
ioural market failure’ (Bar-Gill 2012, p. 2 et seq.); and these
may lead to consumer loss. Behavioural sciences, as such, are
not concerned with such losses or market failures, as they are
merely focused on understanding human behaviour. It does not
matter which sciences (psychology, neuroscience, sociology
etc.) the behavioural understanding or models originate from.
Electronic Markets (2024) 34:14 Page 5 of 13 14
Behaviour modification
Insights into human decision-making are used in mar-
keting (Carnegie, 1936; Cialdini, 2021; Rushkoff, 1999;
Godin, 2005a) as well as in regulation, including in the
guise of nudges (Thaler & Sunstein 2021; Sunstein, 2013).
In his classic book, Influence, professor of psychology
and marketing Robert B. Cialdini has identified seven ‘levers
of influence’ (Cialdini, 2021) that are briefly introduced to
elucidate some classical tools in the advertiser’s toolbox.
1. Reciprocation. This is a basic norm in human culture
that requires one person to try to repay what another
person has provided. The rule applies even to uninvited
gifts or favours, and the feeling of indebtedness may lev-
erage substantially larger favours in return. The rule also
works insituations where a request has been declined,
which makes it easier for the requester to successfully
ensure compliance with a smaller favour (‘rejection-
then-retreat’ tactic). Thus, it may be profitable to give
something or ask for a larger favour before asking the
consumer for a favour (Cialdini, 2021, pp. 71–72).
2. Liking. We prefer to say yes to individuals we like. In
addition to physical attractiveness, likeability can be
boosted by compliments and similarity, i.e. to people
whom we believe to be like us, or to those we already
know (even peripherally). Repetition increases familiar-
ity, and thus likeability. Even association with favour-
able events or people will increase likeability (Cialdini,
2021, pp. 124–135; Carnegie, 1936).
3. Social proof. This tactic includes making products
appear popular or trending. Social proof works best
under uncertainty and/or when many people approve of
the product. Liking can be used to further increase this
effect (Cialdini, 2021, pp. 71–72; Ariely, 2008; Thaler
& Sunstein 2021).
4. Authority. Use of an authority—or just symbols such as
titles and uniforms—in marketing may work as a men-
tal shortcut for quality, approval and recommendation
(Cialdini, 2021, pp. 238–240).
5. Scarcity. When something is less available, we lose free-
dom (of choice). Due to loss aversion, we assign more
value to opportunities that are less available. The tactic
is used by stating that availability is limited or by provid-
ing deadlines. We are more susceptible to this tactic if
we have to compete with others, as in ‘two people from
Denmark are also looking at this limited offer’ (Cialdini,
2021, pp. 289–290).
6. Commitment and consistency. We are more willing to
agree to requests when we have given an initial com-
mitment. This is because we want to appear consist-
ent within our words, beliefs, attitudes and deeds. A
reminder may restore and intensify an initial commit-
ment (Cialdini, 2021, pp. 360–362).
7. Unity. This principle is about establishing a ‘we’-ness
in tribes that leads to group solidarity, i.e. increased
agreement with and influence from members of this
tribe. Shared identities may be based on kinship, geog-
raphy, tastes etc. This may also include being ‘friends’
on social media (Cialdini, 2021, pp. 435–436; Godin,
2008).
It may be added that successful traders create partnerships
with their consumers (Godin, 2008; Turow, 2017), including
by offering protection and privilege (Fletcher,1995), to cre-
ate loyalty (Wind & Hays, 2016; Godin, 2018). However, it
has also been observed that loyalty rewards are increasingly
being used for tracking ‘instead of being a straightforward
tit for tat based on frequent visits’ (Turow, 2017, p. 22), and
that it entails that all consumers pay slightly higher prices,
and only members are getting small subsidies from mer-
chants and all other customers (Clemons, 2019, p. 122).
Storytelling andframing effects
The greatest achievement of the human brain may be its
ability to imagine things that do not exist (our capacity for
ideas), which allows us to anticipate the future (Gilbert,
2006, p. 5; Harari, 2015, p. 117). It is hard to overestimate
the role of narratives in this vein. Storytelling is what we
tell ourselves and others, and what makes it easier to live in
a complicated world (Godin, 2005a, 2005b). Creating such
stories is a fallacy that ‘is associated with our vulnerability
to over-interpretation and our predilection for compact sto-
ries over raw truths’ (Taleb, 2010, p. 63).
Frames are the words and images and interactions that
reinforce personal biases (Godin, 2005a, 2005b, p. 51), and
framing effects are closely related to storytelling (Schwartz
& Sharpe, 2010, p. 61; Akerlof & Schiller, 2015, pp. 41–42),
which is important for intersubjectivity (Harari, 2017, p. 150:
‘Sapiens rule the world because only they can weave an inter-
subjective web of meaning: a web of laws, forces, entities and
places that exist purely in their common imagination’.) and
is utilised in both political (Lakoff, 2014) and commercial
marketing (Rushkoff, 1999, p. 181; Godin, 2005b; Akerlof &
Shiller, 2015, pp. 41–42). As observed by Daniel Kahneman
(Kahneman,2003, p. 1459), ‘Framing effects are not a labo-
ratory curiosity, but a ubiquitous reality’. As an example, it
has been proven that when identical options are described in
different terms, people often shift their choices. For example,
if a choice is described in terms of gains, it is often treated
differently than if it is described in terms of losses (Kahne-
man& Tversky, 1986; Jones, 1999). This shift demonstrates
the concept and power of framing.
Electronic Markets (2024) 34:14 14 Page 6 of 13
Manipulation
Manipulation is closely related to agency and the right to
self-determination. Manipulation can be said to steer the
choices of others by morally problematic means, including
‘employing emotional vulnerability or character defects’
(Wood, 2014, chapter12), ‘in order to make us act against—
or, at the very least, without—our better judgment’ (Ruskoff,
1999, p. 270).
In his analysis of manipulation, Cass Sunstein similarly
finds that ‘an action does not count as manipulative merely
because it is an effort to alter people’s behavior’ and that
there is ‘a large difference between persuading people and
manipulating them’ (Sunstein, 2016, p. 215). In defining
manipulation, Sunstein suggests focusing on whether the
effort in question ‘does […] sufficiently engage or appeal
to [… the consumer’s] capacity for reflection and delibera-
tion’ (emphasis added), and where ‘the word “sufficiently”
leaves a degree of ambiguity and openness, and properly so’
(Sunstein, 2016, pp. 215–216).
This is similar to the fine line that must be drawn between
‘legitimate influence’ and ‘unlawful distortion’ of the aver-
age consumer’s behaviour under the UCPD (Trzaskowski,
2011). It could also be argued that to be able to objectively
ascertain whether consent is given, there must be a sufficient
engagement of or appeal to the data subject’s capacity for
reflection and deliberation to assure a ‘genuine choice’.
Sunstein states that ‘no legal system has a general tort
called “exploitation of cognitive biases”’ (Sunstein, 2016,
p. 234), but both the UCPD and the GDPR may be inter-
preted as including a prohibition of manipulation within the
above-mentioned definition, i.e. focusing on the lack of suf-
ficiently engaging or appealing to the consumer’s capacity
for reflection. The UCPD explicitly mentions coercion as an
aggressive practice, and the prohibition of misleading com-
mercial practices could be interpreted to mean the absence
of deception. It may also be argued that deceptive and coer-
cive practices are not likely to result in unambiguous indica-
tions of data subjects’ wishes that signifies agreement to the
processing of personal data, which implies that consent is
not obtained and that the processing is unlawful.
Choice architecture
As discussed above, insights into human decision-making
can be used by traders to influence behaviour.
We don’t make choices in a vacuum. Our physical envi-
ronment is shaped by nature and culture. Our behaviour and
movement are constrained by laws of physics and public
infrastructure (e.g. Jacobs, 1961) (physical architecture).
In commercial contexts, much of our behaviour and many
of our experiences are carefully designed and curated by
businesses (see Underhill, 1999, p. 195, about the design
of brick-and-mortar shops). For instance, milk is usually
shelved in the very back of a grocery store, so as to guide
people through the entire store.
In virtual realities, such as ‘cyberspace’, we are also
subject to constraints stemming from infrastructure (digital
architecture) that define our abilities and shape our expe-
riences. Digital technology can work in tandem with our
imaginative capabilities to create experiences that defy
real-world physics. It is more cumbersome to redesign the
physical architecture than the architecture of virtual realities,
where a few clicks are sufficient to establish connections,
create fora and delete persons.
Activities and experiences in virtual realities impact real
reality—with which it can easily be confused because of its
pervasiveness and human intersubjectivity. ‘Augmented real-
ity’ is a hybrid in which elements of virtual reality are used
to add to or mask parts of real reality. An illustrative exam-
ple is the mobile game Pokémon Go, which allows the user
to interact with virtual creatures, Pokémon (pocket mon-
sters), which appear in the user’s real reality environment
identified by GPS signals. As catching Pokémons requires
you to be at particular geographic locations, it has been used
to drive real visitors to actual McDonald’s restaurants and
other sponsors (Constine, 2017).
Given our widespread usage of and trust in computers
(including smartphones), virtual realities and augmented
realities can be created that may be difficult to distinguish
from real reality. Given the possibilities of influencing (per-
suading as well as manipulating), it is not difficult to imagine
‘abated reality’ in the guise of ‘real virtuality’ where the real
reality decreases in force or intensity.
Digital technology—Distinct advantages
As much of society’s communication takes place online, we
will also take a look at how influence is carried out online.
Three important characteristics of digital technology are that
activities can be automated, scaled and personalised easily.
In addition, technology allows for real-time feedback.
The design of human–computer interaction plays a signif-
icant role in how consumers are influenced in the context of
data-driven business models (Bond etal., 2012; Mik, 2016).
BJ Fogg has in the early 2000s identified the following six
distinct advantages that computers have compared with tra-
ditional media and human persuaders (Fogg, 2003, p. 7):
• be more persistent than human beings;
• offer greater [often perceived (author’s addition)] ano-
nymity;
• manage huge volumes of data;
• use many modalities to influence;
• scale easily; and
Electronic Markets (2024) 34:14 Page 7 of 13 14
• go where humans cannot go or may not be welcome.
Additionally, computers have good memory, as well as
the ability to evoke feelings through social cues without get-
ting tired or requiring reciprocity (Carr, 2010, pp. 202–205).
With digital technology, Cialdini’s principles of persuasion
introduced above can be automated, scaled, personalised and
applied in real time.
Friction
Persuasive technology (captology), introduced by BJ Fogg,
can be distilled into a matter of dispensing ‘friction’ (Fogg,
2003; McNamee, 2019), which relies on (what Daniel Kah-
neman has identified as) the preference for ‘cognitive ease’
(Kahneman, 2011, p. 67). By increasing or reducing friction,
the user can be nudged in a desired direction by designing a
‘path of least resistance’. The importance of friction in the
design of online experiences is hard to overestimate—as, for
instance, in the context of cookie consent pop-ups. Reading,
thinking, clicking, scrolling, writing and paying all consti-
tute friction.
To appreciate the power of friction, it may be helpful to
perceive it as an obstacle to instant gratification, which is
closely related to the bounded willpower introduced above.
Consider, for instance, liking and sharing content on Face-
book/Instagram, absorbing content on TikTok, searching the
world wide web through Google, establishing connection
on LinkedIn, swiping for dates on Tinder and shopping at
Amazon.com. It is all very easy and convenient.
Both Amazon’s recommendations and Google’s spon-
sored links are designed to guide consumers to specific pur-
chases by reducing the effort or ‘friction’ needed to select
these over all others.
Motivation, ability andprompts
Generally speaking, behaviour is affected by motivation,
ability and prompts (Fogg, 2020). A nostalgic view of mar-
keting may assume that creating motivation is key, whereas
today the main focus—especially in digital marketing—is
to increase ability (by removing friction: ‘it’s free!’, ‘click
here!’) and to use prompts (‘act now!’, ‘click here!’). As
expressed by BJ Fogg:
‘prompts are the invisible drivers of our lives’, ‘no
behavior happens without a prompt’, and ‘the prompts
coming from digital technology are harder to man-
age than those from junk mail […] Other than getting
off the grid, we may never find a perfect way to stop
unwanted prompts from companies with business mod-
els that depend on us to click, read, watch, rate, share,
or react. This is a difficult problem that pits our human
frailties against brilliant designers and powerful com-
puter algorithms.’ (Fogg, 2020, pp. 97 and 105–106).
Law, manipulation andchoice architecture
In determining ethical issues with persuasive technology,
BJ Fogg has identified three focal points, (a) the intentions
of the trader, (b) the methods (practices) used to persuade
and (c) the outcomes (effects) of using the technology, and
he suggests as a first step ‘to take technology out of the
picture’, and ‘simply ask yourself, “If a human were using
this strategy to persuade me would it be ethical?”’ (Fogg,
2003, pp. 220–221; similarly Bermejo, 2019, pp. 129–130,
suggests focusing on ‘companies, content, and users […] to
explore some of the consequences of the advertising model
prevalent on the social web’.)
Interpretation ofEU law
Teleological interpretation necessitates that ‘every provision
of [EU] law must be placed in its context and interpreted in
the light of the provisions of [EU] law as a whole, regard
being had to the objectives thereof and to its state of evolu-
tion at the date on which the provision in question is to be
applied’ (CJEU, Case C-283/81, para 20). As noted by the
Court of Justice of the European Union (CJEU) in the con-
text of data protection law, the interpretation of a provision
of EU, the law must take account of (CJEU, Case C-673/17,
para 48 with references):
• its wording;
• the objectives it pursues;
• its legislative context;
• the provisions of EU law as a whole; and
• possibly its origins.
As discussed, there is already a legal framework address-
ing manipulation by design of choice architecture. Both the
GDPR and the UCPD are secondary laws, and as such they
derive their authority from a higher level of the legal hier-
archy, i.e. primary law.
Because privacy, including the protection of personal
data, is protected in the Charter, it is a straightforward
approach to take a fundamental rights perspective in the
interpretation of the ePrivacy Directive (Directive 2002/58/
EC) (privacy) and the GDPR (personal data).
Privacy is a matter of balancing legitimate interests,
including other fundamental rights—such as the freedom
of expression and the right to information—that are also
necessary in a democracy and important for human wel-
fare. Human dignity—which has freedom, privacy and
Electronic Markets (2024) 34:14 14 Page 8 of 13
non-discrimination at its core—may also play a significant
role in the future interpretation of consumer protection
(Trzaskowski, 2021a, chapter10).
Manipulative commercial practices
Under the UCPD, a commercial practice is aggressive if—
in its factual context, taking account of all its features and
circumstances—by harassment; coercion, including the use
of physical force; or undue influence; it is likely to (a) sig-
nificantly impair the average consumer’s freedom of choice
or conduct and (b) cause him to take a transactional decision
that he would not have taken otherwise (Article 8 UCPD).
‘Transactional decision’ is broadly defined, and includes
(potential) decisions concerning whether or not to buy (or
complain about) products and on what terms. This concept
also covers decisions directly related to such purchase deci-
sions, including the consumer’s decision to enter a shop
(CJEU, Case C-281/12, paras 35–36; CJEU, Case C-391/12),
which indicates a relatively low threshold as to the effect or
loss inflicted on the consumer (European Commission, 2021,
“Interplay between the GDPR and the UCPD” section2.4).
When a trader exploits ‘a position of power in relation to
the consumer’ so as to apply pressure—even without physi-
cal force—undue influence exists if this ‘is likely to signifi-
cantly impair the average consumer’s freedom of choice or
conduct’ (Articles 8 and 2(1)(j) UCPD). Account must be
taken of inter alia (a) ‘its timing, location, nature or per-
sistence’ as well as (b) ‘the use of threatening or abusive
language or behaviour’ (emphasis added) (Article 9 UCPD).
From its wording, this list is not exhaustive. It follows from
recital 16 UCPD that ‘the provisions on aggressive com-
mercial practices should cover those practices which signifi-
cantly impair the consumer’s freedom of choice’ (recitals 6
and 14 UCPD).
Traders are—in contrast to consumers—often well-
informed about consumers’ biases and heuristics (CJEU,
Case C-371/20, para 39 with references).
An illustrative example can be found in several airports
where visitors, in order to get to their gates, are forced to
wander through a shop immediately after security—some-
times with a somewhat concealed escape route for people
with allergies. This is a behaviourally informed—and prof-
itable—‘trick’ to make visitors spend more time and money
in the store. This solution allows shoppers, who intend to
buy, to get to their gates faster, but it is not unlikely that at
least some visitors will take a transactional decision that
they would not have taken otherwise.
One could argue that a commercial practice is aggressive,
if it—for instance, by means of choice architecture—under-
mines the user’s capacity for reflection and deliberation, cf.
the above-mentioned definition of manipulation.
Undue influence is not necessarily ‘impermissible influ-
ence’, but that is the case when conducts apply ‘a certain
degree of pressure’ and in the factual context actively entail
‘the forced conditioning of the consumer’s will’ in a way
that is likely to significantly impair the average consumer’s
freedom of choice or conduct (CJEU, Case C-628/17, paras
33–34).
The CJEU has established that it does not constitute an
aggressive commercial practice to ask the consumer to take
his final transactional decision without having time to study,
at his convenience, the documents delivered to him by a
courier if the consumer has been in a position to take cog-
nisance of the standard-form contracts before (CJEU, Case
C-628/17, para 45).
However, the CJEU found that certain additional prac-
tices with the aim of limiting the consumer’s freedom of
choice may lead to the commercial practice being regarded
as aggressive. This includes conducts that ‘put pressure on
the consumer such that his freedom of choice is significantly
impaired’ or establish an attitude that is ‘liable to make
that consumer feel uncomfortable’ such as to ‘confuse his
thinking in relation to the transactional decision to be taken’
(CJEU, Case C-628/17, paras 33, 46–47).
Such pressure may, for instance, be induced by announc-
ing that less favourable conditions are a consequence of
delayed action on the part of the consumer (CJEU, Case
C-628/17, para 48), a means of utilising Cialdini’s scarcity
lever of influence. Information—including in the guise of
storytelling and framing—may also constitute an important
part of manipulative commercial practices, and such prac-
tices may thus fall under both aggressive and misleading
commercial practices under the UCPD.
As discussed in the previous section, the design of
human–computer interaction plays a significant role in how
consumers are influenced in the context of data-driven busi-
ness models. Technology can be automated to both observe
and shape our behaviour at scale (Zuboff, 2019, p. 8; Yeung,
2017).
The CJEU has assumed that the average consumer is
generally aware of how advertising and sales promotions
work in a free market economy; that today’s consumers are
‘much more circumspect and informed’ as a result of their
experiences with marketing; and also that the protection of
consumers is a ‘patronising argument’, which is ‘no longer
convincing’ (Trstenjak, 2010, para 104 and footnote 82
with reference). However, it is not impossible that pervasive
exposure to personalised marketing, including the use of
prompts and the manipulation of emotions, based on online
tracking across platforms coupled with (annoying) cookie
consent boxes, has resulted in additional confusion, cogni-
tive overload and apathy rather than real education (see also
Forbrukerrådet, 2018).
Electronic Markets (2024) 34:14 Page 9 of 13 14
The UCPD Guidance by the Commission suggests that
(European Commission 2021, section4.2.7) ‘In design-
ing their interfaces, traders should follow the principle
that unsubscribing from a service should be as easy as
subscribing to the service’, and that ‘confirmshaming’
should be avoided. As examples of the latter, the Com-
mission suggests avoiding statements like ‘we’re sorry to
see you go’ and ‘here are the benefits you will lose’. The
psychological effect of the first statement may be real, but
unlikely to be considered unfair, and the second statement
may be material information that the consumer needs to
take a transactional decision.
Consent anddata protection law
Consent requires a ‘freely given, specific, informed and
unambiguous indication of the data subject’s wishes’,
which ‘signifies agreement to the processing’ (accord-
ing to the definition in Article 4(1)(11)). Consent under
the ePrivacy Directive (cookies and direct marketing)
should be understood in the same manner as under the
GDPR (CJEU, Case C-673/17, paras 38–43). Generally
speaking, consent must constitute a genuine and informed
choice. (‘Genuine choice’ can be said to comprise the
requirements of ‘freely given’, ‘unambiguous indication’
and ‘signifying agreement’. Similarly, ‘informed choice’
comprises ‘specific’ and ‘informed’ indication.)
To be genuine, consent requires a clear affirmative act,
which may include ‘ticking a box’, but silence, pre-ticked
boxes, and inactivity cannot constitute consent (recital 32
GDPR). The arguments against pre-ticked boxes include
that it would otherwise be ‘impossible in practice to
ascertain objectively’ whether consent is given (CJEU,
Case C-673/17, paras 49, 52 and 55). It seems straight-
forward to extend this principle to situations where the
choice architecture is designed in a way that does not
sufficiently engage or appeal to the user’s capacity for
reflection and deliberation.
The legitimacy principle introduced above comprises
the overarching fairness principle found in Article 5(1)
(a), which requires that ‘personal data shall not be pro-
cessed in a way that is detrimental, discriminatory, unex-
pected or misleading to the data subject’ (EDPB, 2023,
para 9). This fairness principle is corroborated, inter alia,
by the requirements that (i) ‘it shall be as easy to with-
draw as to give consent’ (empowerment); (ii) communi-
cation must be ‘in a concise, transparent, intelligible and
easily accessible form, using clear and plain language’
(transparency); and (iii) ‘data protection by design and
by default’ must be ensured (legitimacy) (Articles 7(3),
12 and 25 GDPR, respectively).
The Digital Services Act
In the recently adopted Digital Services Act (DSA; Regu-
lation (EU) 2022/2065), the design (and organisation) of
the online interface is addressed in Article 25, which pro-
vides that
providers of online platforms shall not design, organ-
ise or operate their online interfaces in a way that
deceives or manipulates the recipients of their ser-
vice or in a way that otherwise materially distorts or
impairs the ability of the recipients of their service to
make free and informed decisions (emphasis added).
This prohibition does not provide clarity as to what is or
should be considered deception and/or manipulation, aside
from clarifying that the design of online interfaces could
constitute such an unlawful practice. In contrast to the
UCPD, the provision is not limited to practices that harm
the economic interests of consumers, as it covers the users’
ability to ‘make free and informed decisions’ in general.
Besides the fact that the provision only applies to online
platforms (defined in Article 3(1)(i))—i.e. excluding all
other information society services—it is particularly
important to notice that the prohibition does not apply to
practices covered by the UCPD and/or the GDPR (Arti-
cle 25(2); recital 67 provides that ‘legitimate practices, for
example in advertising, that are in compliance with Union
law should not in themselves be regarded as constituting
dark patterns’.), which carves a very significant hole in the
scope of application. According to subsection3 “Behaviour
modification”, the Commission may issue guidelines on
how the prohibition applies to specific practices, including:
• giving more prominence to certain choices when asking
the recipient of the service for a decision;
• repeatedly requesting that the recipient of the service
make a choice where that choice has already been
made, especially by presenting pop-ups that interfere
with the user experience; and
• making the procedure for terminating a service more
difficult than subscribing to it.
The term ‘dark pattern’ is explicitly mentioned in
recital 67 DSA, which focuses on practices that ‘materi-
ally distort or impair, either on purpose or in effect, the
ability […] to make autonomous and informed choices
or decisions’ (emphasis added). In contrast to the UCPD,
the trader’s intention may—according only to the wording
of the recital—also play a role, but the trader’s intention
is usually more difficult to determine than the architec-
ture (method) and the typical reaction of the average user
(effect) (Trzaskowski, 2016).
Electronic Markets (2024) 34:14 14 Page 10 of 13
The making of ‘free and informed decisions’ is at the
heart of the provision, as it prohibits the distortion/impair-
ing of ‘decisions’. One particularly important use of ‘dark
patterns’ is a design that increases engagement, including by
exploiting human biases and heuristics (Kahneman, 2011)
or manufacturing addictions (e.g. Schüll, 2012; Eyal, 2019;
Alter, 2017). Examples include newsfeeds, likes, streaks
and the promotion of more engaging content. It is empha-
sised in recital 67 that the regulated practices can be used
to persuade the user to ‘engage in unwanted behaviours or
into undesired decisions which have negative consequences
for them’, which could include addictive behaviour. It is,
however, unclear whether the behaviour manufactured by
these design choices qualifies as a ‘decision’ as used in the
provision.
It follows further from the recital that the trader should
not use ‘exploitative design choices to direct the recipi-
ent to actions that benefit the provider of online platforms,
but which may not be in the recipients’ interests, present-
ing choices in a non-neutral manner, such as giving more
prominence to certain choices through visual, auditory, or
other components, when asking the recipient of the service
for a decision’ (emphasis added). It remains unclear whether
this protection is intended to go beyond the user’s revealed
preferences, as discussed above; i.e. the trader is obliged
to design the choice architecture to support a ‘not negative
outcome’ from the (individual) user’s perspective.
In addition, the recital recognises the use of friction
(making certain choices more difficult, cumbersome or
time-consuming than others) as discussed above. However,
it may be difficult—often impossible—to present choices in
a neutral manner as envisaged in the above quote, as there
may be an inherent preference (bias) for, for instance, click-
ing on the lower/right-hand option. As expressed by Don
Norman, ‘all artificial things are designed’ (Norman, 1988,
p. 4), and the same is true for information, as everything we
say is framed in some way.
Given its relationship to the UCPD and the GDPR—
which are generally interpreted to have a very wide scope
of application (European Commission, 2021, section4.2.7:
‘The UCPD covers the advertising, sales and contract per-
formance stages, including the agreement to the processing
of personal data and the use of personal data for deliver-
ing personalised content, and the termination of a contrac-
tual relationship’.)—the effect of the DSA in this context
depends on the interpretation of the former legislations. It
cannot be ruled out that the DSA will affect the interpreta-
tion of the UCPD and the GDPR, but it does not seem to
have been the intention.
It remains fair to say that recital 67 of the DSA appears
overoptimistic, considering the wording of the correspond-
ing Article 25. Generally, recitals are supposed to provide
the reasons for the main provisions and as such should not
contain normative provisions or political exhortations (Euro-
pean Union, 2015, guideline 10; see, however, CJEU, Case
C-428/11, para 53, making reference to recital 18 UCPD).
It could appear that the scope of the GDPR and the UCPD
were carved out as a result of effective lobbying, thus deflat-
ing the political wish list represented in the recital. However,
we have argued above for how these two legal instruments
can be applied effectively to address ‘dark patterns’.
As a matter of anti-circumvention, Article 13(6) of the
equally recent Digital Markets Act (DMA; Regulation (EU)
2022/1925; see also recital 37) prohibits making the exercise
of particular rights or choices ‘unduly difficult, including
by offering choices to the end-user in a non-neutral manner,
or by subverting end users’ or business users’ autonomy,
decision-making, or free choice via the structure, design,
function or manner of operation of a user interface or a part
thereof’.
Conclusion andperspective
As suggested above, the legality of design practices may be
assessed by considering intention, method and effect, where
intention does not play a role in determining the unfairness
of a commercial practice or whether a data subject has con-
sented to the processing of personal data. The taxonomies
of dark patterns may be helpful in identifying ‘problematic’
design practices (methods), but further analysis of context
and (likely) effect is necessary to determine the lawfulness
of a particular practice.
We have used ‘choice architecture’ as a neutral term, rec-
ognising that choice architecture—online as well as offline—
may be designed in ways that to varying degrees engage or
appeal to the user’s capacity for reflection and deliberation,
cf. our definition of manipulation. We have also shown how
the GDPR and the UCPD may be interpreted to address
such practices, and how we can better understand human
decision-making.
Dark patterns inlaw
Using the term ‘dark pattern’ does not bring us closer to
drawing the fine line between legitimate persuasion and
unlawful manipulation. The two offline examples used above
(design of grocery stores and airports, respectively) can be
said to use ‘behaviour modification through architecture’, i.e.
‘tricks’ that ‘make you do things that you didn’t mean to’, i.e.
walking through the entire store and entering a store, respec-
tively, and thus increasing the likelihood of buying things
you did not intend to buy. Calling the practice a ‘dark pat-
tern’ does not bring us closer to determining its lawfulness.
The ‘darkness’ of a practice may refer to the traders’
intentions or to the user’s ability to identify the practice
Electronic Markets (2024) 34:14 Page 11 of 13 14
(transparency) or its effect. As many ‘dark patterns’ are
clearly visible—even though their implications (effect) for
individual users may not be apparent—and intentions are not
important for the legal analysis, the unlawfulness of these
patterns may not be determined by their darkness.
In 2022, Brignull’s website changed URL from <darkpat-
terns.org> to <deceptive.design> ‘in an effort to be clearer
and more inclusive’. Similarly, in version 2.0 of the guide-
lines on dark patterns in social media platform interfaces, the
European Data Protection Board is using ‘the more inclusive
and descriptive term “deceptive design pattern” instead of
“dark pattern”’ (EDPB, 2023, footnote 3).
The dark pattern taxonomies illustrate patterns of ‘prob-
lematic’ practices, but the existence of ‘patterns’ in commer-
cial conducts, including procedures for obtaining consent, is
not important to determine whether a concrete commercial
practice (method) is unlawful (see for illustration CJEU,
Case C-388/13, para 42). The pattern recognition may, how-
ever, be of relevance for determining the typical reaction of
the average user in a given case, i.e. determining the effect
of a practice.
As long as the taxonomies only describe practices without
providing guidance on how to make the distinction between
legitimate persuasion and unlawful manipulation, it does not
seem like the best idea to create in law a list of such prac-
tices. This was done with the blacklist of the UCPD, and
it has been argued that it—despite the intention to provide
clear prohibitions—is not sufficiently precise because of its
use of vague notions, which is likely to pose important prob-
lems in practice (Stuyck etal., 2006, pp. 131–132).
Education andawareness
‘Dark patterns’ as a term seems to have had a positive effect
in creating a broader (popular) understanding of how tech-
nology is used to design our experiences and behaviour,
including the role of personal data and profiling that allow
for the creation of individualised realities (personalisation)
that may not only undermine our agency, but is also likely
to have negative consequences for social interactions and
democracy as such. The case of ‘dark patterns’ underscores
the importance of having a vocabulary for such practices
with a view to spur conversations that create useful storytell-
ing, which may lead to behavioural change.
Compliance andenforcement
In addition to benefiting awareness, such taxonomies may
be helpful for both law enforces and law abiders, in the lat-
ter instance as a matter of ensuring compliance under legal
uncertainty. However, we argue that so far, taxonomies
have been better in creating an understanding of a problem
rather than creating clarity as to what is lawful and what is
unlawful—hopefully, the CJEU will provide us with fur-
ther building blocks for a model of what constitutes lawful/
unlawful design of choice architecture.
Further research
As historian Melvin Kranzberg pronounced in his first law
of technology, ‘Technology is neither good nor bad; nor is it
neutral’ (Kranzberg, 1986). It should be obvious by now that
choice architecture may be designed for good and for bad,
and that a design that works to the benefit for some users
may be to the detriment of other users. We must also recog-
nise that ‘dark patterns’ are not new and not unique to the
digital environment, let alone online platforms. What is new
online is the scale, scope and precision of data-driven pre-
dictions, which can also be used to design user experiences.
We should be careful not to use the term ‘dark pattern’ to
exclude or overlook the decades of work done and inspired
by the likes of Herbert Simon, Daniel Kahneman, Amos
Tversky, Robert Cialdini and BJ Fogg. The challenge from
a legal perspective is to understand human beings’ ability
to make free and informed decisions, how that ability may
influenced by information and architecture and how design
may help to preserve the user’s right to self-determination.
In designing law in this context, it is important to focus
on the architecture rather than the intent of a trader (see also
Helberger etal., 2022, pp. 195–196). One could also consider
the extent to which legitimacy and accountability, as in the
GDPR, could be introduced or emphasised when business
models rely on harvesting the value of consumer irration-
ality, including consumer inertia. Considering the insights
from behavioural sciences, including persuasive technology,
principles of self-determination by design (market perspec-
tive) and human dignity by design (societal perspective) may
be more challenging—which in itself is a reason for trying.
Funding Open access funding provided by Copenhagen Business
School
Open Access This article is licensed under a Creative Commons Attri-
bution 4.0 International License, which permits use, sharing, adapta-
tion, distribution and reproduction in any medium or format, as long
as you give appropriate credit to the original author(s) and the source,
provide a link to the Creative Commons licence, and indicate if changes
were made. The images or other third party material in this article are
included in the article’s Creative Commons licence, unless indicated
otherwise in a credit line to the material. If material is not included in
the article’s Creative Commons licence and your intended use is not
permitted by statutory regulation or exceeds the permitted use, you will
need to obtain permission directly from the copyright holder. To view a
copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
Electronic Markets (2024) 34:14 14 Page 12 of 13
References
Akerlof, G. A., & Shiller, R. J. (2015). Phishing for phools. Princeton
University Press.
Alter, A. (2017). Irresistible: The rise of addictive technology and the
business of keeping us hooked. Penguin Press.
Ariely, D. (2008). Predictably irrational. Harper Perennial.
Bandura, A. (1969). Principles of behavior modification. Holt.
Bar-Gill, O. (2012). Seduction by contract. Oxford University Press.
Baumeister, R.F. & Tierney, J. (2011). Willpower. Penguin Books.
Bermejo, F. (2019). Online advertising as a shaper of public communi-
cation’. In Rikke Frank Jørgensen (ed.). Human rights in the age
of platforms. The MIT Press.
BEUC (2022). ‘Dark patterns’ and the EU consumer law acquis.
Bond, R. M., etal. (2012). A 61-million-person experiment in social
influence and political mobilization. Nature, 489, 295–298.
https:// doi. org/ 10. 1038/ natur e11421
Borgesius, F. Z., Helberger, N., & Reyna, A. (2017). The perfect
match? a closer look at the relationship between EU consumer
law and data protection law. Common Market Law Review, 54(5),
1427–1465. https:// doi. org/ 10. 54648/ cola2 017118
Calo, R. (2014). Digital market manipulation. George Washington Law
Review, 995–1051.
Carnegie, D. (1936). How to win friends and influence people. Simon
& Schuster.
Carr, N. (2010). The shallows. W. W. Norton & Company.
Cialdini, R. B. (2021). Influence, new and expanded (p. 1984). First
published.
CJEU, Case C-283/81, CILFIT v Ministero della Sanità,
ECLI:EU:C:1982:335.
CJEU, Case C-428/11, Purely creative and others, ECLI:EU:C:
2012:651.
CJEU, Case C-559/11. Pelckmans Turnhout, ECLI:EU:C:2012:615.
CJEU, Case C-59/12. Zentrale zur Bekämpfung unlauteren Wettbew-
erbs, ECLI:EU:C:2013:634.
CJEU, Case C-388/13. UPC Magyarország, ECLI:EU:C:2015:225.
CJEU, Case C-673/17, Planet49, ECLI:EU:C:2019:801.
CJEU, Case C-628/17, Orange Polska, ECLI:EU:C:2019:480.
CJEU, Case C-281/12, Trento Sviluppo and Centrale Adriatica,
ECLI:EU:C:2013:859
CJEU, Case C-371/20, Peek & Cloppenburg, ECLI:EU:C:2021:674.
CJEU, Case C-391/12, RlvS, ECLI:EU:C:2013:669.
Clemons, E. K. (2019). New patterns of power and profit. Springer.
CNIL. (2019). Shaping choices in the digital world. (IP reports innova-
tion and foresight N°06) January 2019.
CNIL. (2022). Decision of 19 December 2022. https:// www. cnil. fr/
en/ cooki es- micro soft- irela nd- opera tions- limit ed- fined- 60- milli
on- euros
Constine, J. (2017). Pokémon GO reveals sponsors like McDonald’s
pay it up to $0.50 per visitor. TechCrunch. 31 May 2017. https://
techc runch. com/ 2017/ 05/ 31/ pokem on- go- spons orship- price/?
gucco unter= 1& guce_ refer rer= aHR0c HM6Ly 93d3c uZ29v Z2xlL
mNvbS 8& guce_ refer rer_ sig= AQAAA JI1RE alkce H66Kf YnrRt-
GnDps WXQhx YPH4r 85AMW 2UUXb PyjpC ejbjZ N3048n_
W8pWh jV7t_ jIQ_ smLYr ZXgLC 8oJgc srGxP xnXuk znQiL ljbf4
p7go8 9O- lSzKf 7EPLP yDV6H QKjCr IN_ 4q7xy ZI1ed 2kXbF
MccG0 ALsjU A7P
EDPB. (2022). Guidelines 3/2022 on dark patterns in social media
platform interfaces: How to recognise and avoid them. https://
edpb. europa. eu/ our- work- tools/ docum ents/ public- consu ltati ons/
2022/ guide lines- 32022- dark- patte rns- social- media_ en
EDPB. (2023). Guidelines 03/2022 on deceptive design patterns in
social media platform interfaces: How to recognise and avoid
them (Version 2.0). https:// edpb. europa. eu/ our- work- tools/
our- docum ents/ guide lines/ guide lines- 032022- decep tive- design-
patte rns- social- media_ en
Elster, J. (1983). 2016. Sour Grapes. Cambridge University Press.
European Commission. (2007). EU Consumer Policy strategy 2007–
2013—Empowering consumers, enhancing their welfare, effec-
tively protecting them. 13 March 2007. COM 99 final. https://
eur- lex. europa. eu/ EN/ legal- conte nt/ summa ry/ eu- consu mer- pol-
icy- strat egy- 2007- 2013. html
European Commission. (2021). Guidance on the interpretation and
application of Directive 2005/29/EC of the European Parliament
and of the Council concerning unfair business-to-consumer com-
mercial practices in the internal market. (Staff Working Docu-
ment) 2021/C 526/01.
European Commission, Directorate-General for Justice and Consum-
ers, Lupiáñez-Villanueva, F., Boluda, A., & Bogliacino, F. (2022).
Behavioural study on unfair commercial practices in the digital
environment – Dark patterns and manipulative personalisation
– Final report. Publications Office of the European Union, 2022.
https:// data. europa. eu/ doi/ 10. 2838/ 859030
European Council. (1975a). Resolution on a preliminary programme
of the European Economic Community for a consumer protection
and information policy. 14 April 1975.
European Council. (1975b). Preliminary programme of the European
Economic Community for a consumer protection and information
policy. 25 April 1975.
European Union - Legal service. (2015). Joint practical guide of the
European Parliament, the Council and the Commission for per-
sons involved in the drafting of European Union legislation, Pub-
lications Office. https:// data. europa. eu/ doi/ 10. 2880/ 5575
Eyal, N. (2019). Hooked: How to build habit-forming products (p.
2014). First published.
Federal Trade Commission v. Amazon.com, inc. (2023) Amended com-
plaint for permanent injunction, v. civil penalties, monetary relief
and other equitable relief. United States District Court, Western
District of Washington. Case 2:23-cv-00932. Filed 21 June 2023.
Fletcher, G. P. (1995). Loyalty: An essay on the morality of relation-
ships. Oxford University Press.
Fogg, B. J. (2003). Persuasive technology. Morgan Kaufmann.
Fogg, B.J. (2020). Tiny habits. Houghton Mifflin.
Forbrukerrådet. (2018). Deceived by design. Report. https:// stora
ge02. forbr ukerr adet. no/ media/ 2018/ 06/ 2018- 06- 27- decei ved-
by- design- final. pdf
FTC. (2022a). Fortnite video game maker epic games to pay more than
half a billion dollars over FTC allegations of privacy violations
and unwanted charges. Federal Trade Commission. Agreement of
19 December 2022 with Epic Games. https:// www. ftc. gov/ news-
events/ news/ press- relea ses/ 2022/ 12/ fortn ite- video- game- maker-
epic- games- pay- more- half- billi on- dolla rs- over- ftc- alleg ations
FTC. (2022b). Bringing dark patterns to light. (Staff Report). Septem-
ber 2022. https:// www. ftc. gov/ sy s tem/ files/ ftc_ go v/ pdf/ P2148 00%
20Dark% 20Pat terns% 20Rep ort% 209. 14. 2022% 20-% 20FIN AL. pdf
Gilbert, D. (2006). Stumbling on happiness. Harper Perennial.
Godin, S. (2005a). All marketers are liars. Portfolio.
Godin, S. (2005b). All marketers are liars. Portfolio. 2009. First
published.
Godin, S. (2008). Tribes. Portfolio.
Godin, S. (2018). This is marketing. Portfolio.
Harari, Y.N. (2015). Sapiens. Harper.
Harari, Y.N. (2017). Homo Deus. Harper.
Helberger, N., Sax, M., Strycharz, J., & Micklitz, H.-W. (2022). Choice
architectures in the digital economy: Towards a new understand-
ing of digital vulnerability. Journal of Consumer Policy, 45, 175–
200. https:// doi. org/ 10. 1007/ s10603- 021- 09500-5
Jacobs, J. (1961). 2011. First published.
Electronic Markets (2024) 34:14 Page 13 of 13 14
Jarovsky, L. (2022). Dark patterns in personal data collection: Defini-
tion, taxonomy and lawfulness. 1 March 2022. https:// ssrn. com/
abstr act= 40485 82
Jolls, C., Sunstein, C. R., & Thaler, R. (1998). Theories and tropes: A
reply to Posner and Kelman. Stanford Law Review, 50(5), 1593–
1608. https:// doi. org/ 10. 2307/ 12293 07
Jones, B. D. (1999). Bounded rationality. Annual Review of Political
Science, 2(1), 297–321. https:// doi. org/ 10. 1146/ annur ev. polis ci.2.
1. 297
Kahneman, D. (2003). Maps of bounded rationality: Psychology for
behavioral economics. The American Economic Review, 93(5),
1449–1475. https:// doi. org/ 10. 1257/ 00028 28033 22655 392
Kahneman, D. (2011). Thinking, fast and slow. Farrar.
Kahneman, D., & Tversky, A. (1986). Rational choice and the framing
of decisions. Journal of Business, 59(4), 251–278.
Kranzberg, M. (1986). Technology and history: “Kranzberg’s Laws”.
Technology and Culture, 544–560. https:// doi. org/ 10. 2307/ 31053
85
Kreps, D., & Rowe, F. (2021). Free will in philosophy and in IS
research: Where do we stand and where do we want to be? ICIS
2021 Proceedings. 5. https:// aisel. aisnet. org/ icis2 021/ adv_ in_
theor ies/ adv_ in_ theor ies/5
Lakoff, G. (2014). [The All New] Don’t think of an elephant (p. 2004).
Chelsea Green Publishing.
Leiser, M. R., & Caruana, M. (2021). Light to be found in Europe’s
consumer protection regime. Journal of European Consumer and
Market Law,10(6), 237–251. https:// hdl. handle. ne t/ 1887/ 32783 62
Luguri, J., & Strahilevitz, L. J. (2021). Shining a light on dark pat-
terns. Journal of Legal Analysis, 13(1), 43–109. https:// doi. org/
10. 1093/ jla/ laaa0 06
Mathur, A., Acar, G., Friedman, M. J., Lucherini, E., Mayer, J., Chetty,
M., & Narayan, A. (2019). Dark patterns at scale: Findings from
a crawl of 11K shopping websites. Proceedings of the ACM on
Human-Computer Interaction, 3(CSCW), 81, 1–32. https:// doi.
org/ 10. 1145/ 33591 83
McNamee, R. (2019). Zucked. Penguin.
Mik, E. (2016). The erosion of autonomy in online consumer transac-
tions. Law, Innovation and Technology, 8(1), 1–38. https:// doi. or g/
10. 1080/ 17579 961. 2016. 11618 93
Nahai, N. (2015). Dark patterns and Ada Lovelace. The
Guardian.https:// www. thegu ardian. com/ techn ology/ audio/ 2015/
oct/ 22/ dark- patte rns- ada- lovel ace- tech- podca st
Norman, D. (1988). 2013. The design of everyday things. Basic Books.
OECD. (2022). Dark commercial patterns (OECD Digital Economy
Papers, No. 336). Paris: OECD Publishing. https:// doi. org/ 10.
1787/ 44f5e 846- en
Posner, R.A. (1997). Rational choice, behavioral economics, and the
law. Stanford Law Review. 1551–1575.
Rushkoff, D. (1999). Coercion, why we listen to what ‘they’ say.
Riverhead.
Schüll, N. D. (2012). Addiction by design. Princeton University Press.
Schwartz, B. & Sharpe, K. (2010). Practical wisdom. Riverhead.
Stuyck, J., Terryn, E., & van Dyck, T. (2006). Confidence through
fairness? The new directive on unfair business-to-business
commercial practices in the internal market. Common Market Law
Review, 43(1), 107–152. https:// doi. org/ 10. 54648/ cola2 005092
Sunstein, C. R. (2013). Simpler. Simon & Schuster.
Sunstein, C. R. (2016). Fifty shades of manipulation. Journal of Mar-
keting Behavior, 1, 214–244. https:// doi. org/ 10. 2139/ ssrn. 25658
92
Svantesson, D. J. B. (2018). Enter the quagmire—The complicated
relationship between data protection law and consumer protection
law. Computer Law & Security Review, 34(1), 25–36. https:// doi.
org/ 10. 1016/j. clsr. 2017. 08. 003
Taleb, N. N. (2010). The black swan (2nd ed., p. 2007). Random House.
Thaler, R. H., & Sunstein, C. R. (2021). Nudge—The final edition(p.
2008). Yale University Press.
Trstenjak. (2010). Opinion of advocate general in case C-540/08.
Mediaprint Zeitungs- und Zeitschriftenverlag. (European Court
Reports 2010 I-10909). ECLI:EU:C:2010:161.
Trzaskowski, J. (2011). Behavioural economics, neuroscience, and
the unfair commercial practices directive. Journal of Consumer
Policy, 34, 377–392. https:// doi. org/ 10. 1007/ s10603- 011- 9169-2
Trzaskowski, J. (2016). Lawful distortion of consumers’ economic
behaviour – Collateral damage under the unfair commercial
practices directive. European Business Law Review, 27(1), 25–49.
https:// doi. org/ 10. 54648/ eulr2 016002
Trzaskowski, J. (2021a). Your privacy is important to us! – Restoring
human dignity in data-driven marketing. Ex Tuto.
Trzaskowski, J. (2021b). GDPR compliant processing of big data
in small business. In Pedersen, C.L., Lindgreen, A., Ritter, T.
& Ringberg, T. (eds). Big data in small business—Data-driven
growth in small and medium-sized enterprises. Edward Elgar.
Trzaskowski, J. (2023). Persuasion, manipulation, choice architecture
and ‘dark patterns’. In Savin, A. & Trzaskowski, J. Research hand-
book on EU Internet Law (2nd edition). Edward Elgar.
Turow, J. (2017). The aisles have eyes. Yale University Press.
UN. (2015). Resolution 70/186 of 22 December 2015, General Assem-
bly. https:// unctad. org/ system/ files/ offic ial- docum ent/ ares7 0d186_
en. pdf
Underhill, P. (1999). 2009. First published.
Wind, Y., & Hays, C. F. (2016). Beyond advertising. Wiley.
Wood, A. W. (2014). The free development of each. Oxford University
Press.
Yeung, K. (2017). “Hypernudge”: Big Data as a mode of regulation by
design. Information, Communication & Society, 20(1), 118–136.
https:// doi. org/ 10. 1080/ 13691 18X. 2016. 11867 13
Zarsky, T. Z. (2019). Privacy and manipulation in the digital age. Theo-
retical Inquiries in Law, 20(1), 157–188.https:// ssrn. com/ abstr
act= 33211 72
Zuboff, S. (2019). Surveillance capitalism. Profile Books.
Publisher's Note Springer Nature remains neutral with regard to
jurisdictional claims in published maps and institutional affiliations.