DATA FOR SALE
by
SUSAN FROETSCHEL
______________________________________________________________________
Susan
Froetschel is editor of YaleGlobal Online and the author
of five novels including Fear of Beauty and Allure
of Deceit, both set in Afghanistan.
The scandal
involving Facebook and datamining company Cambridge Analytica
dramatically confirms the old adage of “no free lunch.”
Facebook’s more than 2 billion users are waking up to the
fact that the free online site, part of their daily routine, extracted
a stiff price: valuable personal data. The only way to protect
one’s data is vigilance by all, though that may well be
the equivalent of closing the barn door after the proverbial horse
escaped.
News
reports that Cambridge Analytica swept up details on millions
of Facebook users – then used details for targeted political
advertising in many countries – jolted industry, regulators
and users. Yet users consented to data exchanges, often impatiently,
without reading pages of small print of terms-of-service agreements.
Many companies profit handsomely from knowing the range and length
of users’ phone calls, driving patterns, family history
and genetics, purchase details through credit cards and store
discount cards, insurance claims for health problems, and games
that measure frustration levels or ability to follow rules.
Of course,
gigantic Facebook is not alone. Big-data analysis is big business.
Companies continue to discover new value in cross-industry exchanges,
combining forces to monetize datasets to improve services, reduce
fraud, attract new customers or meet regulatory requirements.
Cambridge Analytica is not alone either. In China, the Shanghai
Data Exchange, started in 2017, offers a platform for trading
all types of consumer information gathered from telecommunications,
credit cards and more with the aim of drawing technology firms
to the city. The exchange anticipates accounting for one third
of China’s data trading volume by 2020. Global revenues
for data analytics are expected to exceed $200 billion in 2020,
according to International Data Corp., which points out the rush
to share data could be prone to conflicts of interest.
Collecting
data to assess target groups is not new. Decades ago, telemarketing
firms relied on typists to go through phone books, cross-listing
names and numbers with other public lists. Librarians understood
the potential privacy pitfalls early on and endorsed policies
to protect confidentiality – before computers became widespread,
libraries stopped using cards listing individuals who had previously
borrowed books. Likewise, the college application process has
long been a data-mining exercise to determine which applicants
are likely to enroll and graduate.
Patients,
borrowers, students who fill out offline application forms are
not exempt from becoming targets. Paper forms are quickly scanned
into computer files. Large community events and fairs offer opportunities
for data gathering. Hundreds of vendors attending large home shows
hold contests to gather potential customer contacts, and job fairs
collect resumes to study the evolving job market and reap new
employer contacts.
Computers
made data collection easy. Any type of data can be packaged and
marketed. Cities already provide data on properties, taxes and
public health as a public service, and the World Economic Forum
suggests communities could do more to distribute data on assets
from traffic to waste collection. Associations offer services
and information on how to research and package data. To improve
efficiency, utilities in India, Europe and the United States rely
on smart meters to monitor and predict patterns of energy and
water use. Committees and policies for monitoring data use and
information governance so far are not keeping up with the growing
numbers of organizations gathering and trading data.
Data
products can be specific, offering details about individuals,
or aggregated to relay broad trends. Laws in the United States
and Europe protect individual health, education or financial information,
but do not ban aggregation as described in privacy policies, terms
of agreement and license agreements. For example, the Common Application
– an online form required for applying to many US colleges
– details its policy: Third parties and contracted researchers
may have access to application and related information which can
then be packaged as “non-personal identifying demographic,
historical, generic, analytical, statistical or aggregate data
obtained from other sources, and/or data.”
Health
is an especially sensitive area, and privacy laws, even the strict
new data protections to be imposed by the European Union in May,
include exceptions. The EU law requires that patient data “be
collected for a specific explicit and legitimate purpose”
but allows that same data to “be re-used for research”
for the public-interest purpose of driving innovative treatments.
The same law limits how long patient data can be stored, “except
for archiving and scientific research purposes.” Explicit
patient consent is not required as long as safeguards masking
identity are in place. Of course, as the Cambridge Analytica case
shows, data collected ostensibly for “academic” research
can be deployed for other operations.
Financial
firms collaborate on data collection to avoid risks. Insurers
form special units for collecting drivers’ data. Digital
strategies fuel growth, explains Boston Consulting Group. Companies
combine online business processes with communications and services
to gather data. App developers respond with entertaining quizzes,
surveys and games designed to entice consumers to hand over more
data. The harvest of Facebook profiles began with a small personality
test, fewer than 300,000 users took part for a tiny sum, and in
the process millions of friends got dragged into the net.
Less
than 20 percent of third-party app developers for Facebook’s
platform are based in the United States. Developers like Elitech
in India provide custom-designed applications or games that assess
target audiences and prioritize user engagement. Developers can
use games to assess user performance and personality with small
tasks from placing an order to solving problems. Facebook encourages
developers around the world with Developer Circles, anticipating
local app development to lead to more local users.
Asia
leads the world with more than 30 percent of 20 million app developers
while Europe and North America each have about 30 percent. Economic
Times reports that India leads the world in Facebook users and
has the second largest base of Facebook developers. Developers
and social media firms worry about new regulations disrupting
the growing industry. Apps and games available from Apple’s
iTunes Store went from a few hundred in 2009 to more than 3 million
in 2017, with downloads in the billions.
Apps
take advantage of the universal desires to play, understand ourselves,
or compare how we perform with others. Experts analyze user choices,
associating interests as detected by searches and clicks with
individual behavior, hunting for patterns and correlations. Some
companies offer discounts to customers deemed as reliable or credit-worthy;
other firms hunt for gullible, impulsive spenders.
Unorganized
data may seem worthless, and Facebook and countless others readily
opened the gates to app developers and advertisers with little
attention to the ultimate goals behind data transfers. Soon after
news emerged about Cambridge Analytica’s use of Facebook
profiles, Mark Zuckerberg issued an apology, admitting that even
social media executives had not realized the full potential of
their platforms and how many insights might be gleaned. He admitted
not imagining in launching Facebook in 2004 that the site could
be accused of changing the course of an election. To Recode, he
admitted to feeling “uncomfortable sitting here in California
at an office, making content policy decisions for people around
the world.” His strategy is for communities to decide their
values and rules for Facebook.
Users
have a choice on what to share and with whom. Like it or not,
big-data analysis influences communities and workplaces, and users
have a responsibility to read lengthy policies with care. A lesson
emerging from the Cambridge Analytica and Facebook debacle is
that those who refuse to surrender data cannot evade the consequences
especially when so many other users do share. Millions of friends
whose data was harvested may not have given specific consent,
but in the end that did not matter.
Reprinted with permission from YaleGlobal Online
www.yaleglobal.yale.edu