Nieman Foundation at Harvard
After criticism over “viewpoint diversity,” NPR adds new layers of editorial oversight
ABOUT                    SUBSCRIBE
Oct. 6, 2021, 8:49 a.m.

The rise of dark web design: How sites manipulate you into clicking

While cookie banners are annoying and often dishonest, we need to consider the broader implications of an online ecosystem that is increasingly manipulative by design.

The vast majority of websites you visit now greet you with a pop-up. This annoying impediment to your seamless web browsing is called the “cookie banner,” and it’s there to secure your consent, as per online privacy laws, for websites to retain information about you between browsing sessions.

The cookie banner purports to offer you a choice: consent to only the essential cookies that help maintain your browsing functionality, or accept them all — including cookies that track your browsing history to sell on to targeted advertising firms. Because those additional cookies generate extra revenue for the websites we visit, cookie banners are often designed to trick you into clicking “accept all.” (Whil

The UK’s information commissioner recently urged G7 countries to address this problem, highlighting how fatigued web users are agreeing to share more personal data than they’d like. U.S. users are, presumably, similarly fatigued. But in truth, manipulative cookie banners are just one example of what’s called “dark design” — the practice of creating user interfaces that are intentionally designed to trick or deceive the user.

Dark design has proven to be an incredibly effective way of encouraging web users to part with their time, money and privacy. This in turn has established “dark patterns,” or sets of practices designers know they can use to manipulate web users. They’re difficult to spot, but they’re increasingly prevalent in the websites and apps we use every day, creating products that are manipulative by design, much like the persistent, ever-present pop-ups we’re forced to close when we visit a new website.

Cookie banners remain the most obvious form of dark design. You’ll notice how the “accept all” button is large and cheerfully highlighted, attracting your cursor within a split second of your arrival on a website. Meanwhile, the dowdy, less prominent “confirm choices” or “manage settings” buttons — the ones through which we can protect our privacy — scare us away with more time-consuming clicks.

You’ll know from experience which one you tend to click. Or you can try the Cookie Consent Speed-Run, an online game that exposes how difficult it is to click the right button in the face of dark design.

E-commerce websites also frequently use dark patterns. Say you’ve found a competitively priced product you’d like to buy. You dutifully create an account, select your product specifications, input delivery details, click through to the payment page — and discover the final cost, including delivery, is mysteriously higher than you’d originally thought. These “hidden costs” aren’t accidental: the designer is hoping you’ll just hit “order” rather than spending even more time repeating the same process on another website.

Other elements of dark design are less obvious. Free services such as Facebook and YouTube monetize your attention by placing advertisements in front of you as you scroll, browse or watch. In this “attention economy,” the more you scroll or watch, the more money the companies make. So these platforms are intentionally optimized to command and retain your attention, even if you’d rather close the app and get on with your day. For example, the expertly crafted algorithm behind YouTube’s “Up Next” video suggestions can keep us watching for hours if we let them.

App design

Manipulating users for commercial gain isn’t just used on websites. Currently, more than 95% of Android apps on the Google Play store are free to download and use. Creating these apps is an expensive business, requiring teams of designers, developers, artists, and testers. But designers know that they’ll recoup this investment once we’re hooked on their “free” apps — and they do it using dark design.

In recent research analyzing free app-based games that are popular with teenagers, my colleague and I identified dozens of examples of dark design. Users are forced to watch ads and frequently encounter disguised ads that look like part of the game. They’re prompted to share posts on social media and, as their friends join the game, are prompted to make in-app purchases to differentiate their character from those of their peers.

Some of this psychological manipulation seems inappropriate for younger users. Teenage girls’ susceptibility to peer influence is exploited to encourage them to buy clothes for in-game avatars. Some games promote unhealthy body imagery while others actively demonstrate and encourage bullying through indirect aggression between characters.

There are mechanisms to protect young users from psychological manipulation, such as age rating systems, codes of practice, and guidance that specifically prohibits the use of dark design. But these rely on developers understanding and interpreting this guidance correctly and, in the case of the Google Play store, developers vet their own work and it’s up to users to report any issues. My research indicates that these measures are not yet proving entirely effective.

Shedding light

The problem with dark design is that it’s difficult to spot. And dark patterns, which are established in every developer’s toolbox, spread fast. They’re hard for designers to resist when free apps and websites are competing for our attention, judged on metrics like “time on page” and the “user conversion rate.”

So while cookie banners are annoying and often dishonest, we need to consider the broader implications of an online ecosystem that is increasingly manipulative by design. Dark design is used to influence our decisions about our time, our money, our personal data and our consent. But a critical understanding of how dark patterns work, and what they’re hoping to achieve, can help us detect and overcome their trickery.

Dan Fitton is a professor in user experience design at the University of Central Lancashire. This article is republished from The Conversation under a Creative Commons license.The Conversation

Photo of cookies by The Creative Exchange on Unsplash.

POSTED     Oct. 6, 2021, 8:49 a.m.
Show tags
Join the 60,000 who get the freshest future-of-journalism news in our daily email.
After criticism over “viewpoint diversity,” NPR adds new layers of editorial oversight
“We will all have to adjust to a new workflow. If it is a bottleneck, it will be a failure.”
“Impossible to approach the reporting the way I normally would”: How Rachel Aviv wrote that New Yorker story on Lucy Letby
“So much of the media coverage — and the trial itself — started at the point at which we’ve determined that [Lucy] Letby is an evil murderer; all her texts, notes, and movements are then viewed through that lens.”
Increasingly stress-inducing subject lines helped The Intercept surpass its fundraising goal
“We feel like we really owe it to our readers to be honest about the stakes and to let them know that we truly cannot do this work without them.”