When saw a car advertisement pop up on the side of his Facebook newsfeed, it gave him pause. Most of us wouldn鈥檛 notice what seemed like just another ad among many during our daily Internet usage, but Young had a more nuanced perspective on the ad鈥檚 sudden appearance.
As an undergraduate at 一本道无码, he studied ;聽then he went on to get his master鈥檚 degree in information and knowledge strategy at Columbia University, where he鈥檚 also adjunct faculty. After working in analytics for marketing agencies, and now as the global knowledge manager for strategic branding firm Siegel+Gale, he understands why and how online advertisers target their ads, so seeing one for a car sent his mind scrolling back over his Internet activity in the previous weeks.
You see, as a three-year resident of New York City, dependent on public transportation, he had never bought a car, or had a car, or even had a car ad populate on a website before. He knew this meant that something about his recent online behavior suddenly indicated to marketers that he might be interested in buying one. And even more surprising? The ad was right. He really did want to buy a car. This indirectly troubles , an associate professor in both the and departments at 一本道无码鈥攂ut more on that later.
To fully appreciate Young鈥檚 bewilderment at the well-timed car ad, you have to understand the multistep journey every online ad takes before it pops up on your screen. For more than a decade, online marketers have been compiling or purchasing data about you from several sources in order to better target ads. The ads are deployed by software made up of algorithms鈥攃onditional operating instructions鈥攖hat get 鈥渟marter鈥 over time as behavior patterns emerge. These algorithms learn things about your online behavior and about the demographics to which you belong and then use this information to incrementally increase how effective they are at placing an ad on your screen you might actually click.
Sources for the data being collected are varied. They include:
- Clickstreams鈥攐r data collected from 鈥渃ookies鈥 sent by websites that track your browsing and clicking activity on the Internet
- Search engine inquiries
- Online purchases
- Data sold by credit card companies selling data about your offline purchases
- Profile data from social media platforms, what you hashtag, and whom you know
- Location data based on your Internet access
鈥淭here鈥檚 definitely a shift away from using descriptive demographics鈥攑eople who are this race, this age, this sexual preference鈥攁nd instead focusing on online behaviors,鈥 reflects Young. 鈥淚f the person types these search terms and clicks on this thing, they have a demonstrated interest in this topic, so these are the people that we should advertise to.鈥
So for example, if you鈥檝e been Googling flights to Italy, when you then navigate to The New York Times website, you might see advertisements for cheap airfares to Italy. Makes sense; the advertisers are showing you an ad that鈥檚 both relevant and timely.
But a little research into the mechanics of online advertising will reveal something you may not know: while any webpage loads, there is actually a live auction for each spot on the page where an ad could appear. Companies that collect and broker information about your behaviors and preferences鈥攁nd there are many hundreds of them鈥攔each out to marketers who have identified you as their target and essentially ask, 鈥淗ow much is this particular ad worth putting in front of this particular person?鈥 The little spaces on the page that populate with flight suggestions or shoes or graduate programs are from the advertisers who鈥檝e paid the most in that moment for your personal attention. And that all happens in the span of the second or two it takes for a page to load. And algorithms automate it all.
But it鈥檚 not as simple a formula as 鈥減references in, advertisements out,鈥 according to professionals like Young. This means you won鈥檛 see ads only for the exact item or service you were searching for or have previously purchased. Googling flights to Italy might also prompt ads for luggage or travel insurance鈥攐r even for engagement rings if you鈥檙e searching for tickets around Valentine鈥檚 Day and you鈥檝e been 鈥淚n a Relationship鈥 on Facebook for a while.
Not just meeting but anticipating customer needs and wants is the real goal for online advertisers. This means greater convenience for consumers, but it鈥檚 also making it harder to distinguish whether it was really your idea to propose on your trip to Italy, or if it was De Beers using an algorithm to recognize that many consumers before you clicked their way down the same digital paths before popping the question.
And that鈥檚 the chicken-and-egg situation Young found himself in. Had he already been contemplating purchasing a car? Or had the algorithms perfectly anticipated a change in his needs based on his recent online activity? Either way, what information was being collected on him that led the algorithms to such a conclusion?
Datta, the 一本道无码 professor, worries about questions like these. Relatively little research has been done on the subject of online advertising, and even fewer tools exist to examine and analyze the methods and repercussions of these advertising algorithms. So he built one himself.
Datta, along with 一本道无码 PhD student Amit Datta and 一本道无码 alumnus Michael Tschantz, Google鈥檚 ad network. AdFisher can surfing the Internet and then document the types and frequency of ads they are shown as a result of browsing.
One study Datta鈥檚 research team performed has implications in the privacy sector. After simulated users researched sites related to substance abuse, they were served ads for rehab centers. They were relevant, to be sure, but that information can now be used in perpetuity to target users, collected and sold the same way your last Zappos shoe purchase was. The Health Insurance Portability and Accountability Act (HIPAA), passed in 1996, ensures that patients鈥 medical, mental health, and substance abuse diagnoses can鈥檛 be disseminated without consent, but the law, like many others, doesn鈥檛 reach this far into the Wild Wild West of the web.
Relatively little research has been done on the subject of online advertising, and even fewer tools exist to examine and analyze the methods and repercussions of these advertising algorithms. So Anupam Datta built one himself.
Datta stresses that it鈥檚 not just what鈥檚 being done with the information collected that needs oversight, but the collection process itself. 鈥淯nfettered collection of personal data has a chilling effect on individual freedoms,鈥 he explains. Of course, users may alter their behavior and not browse the Internet freely knowing that almost every click is being recorded. However, Datta brings it back to health conditions, positing that a patient may not research a medical issue for fear of that information getting into the hands of a prospective employer or an insurance company that may hike up rates or refuse coverage.
Although many people balk at the veritable encyclopedia of information being collected on everyone who uses the Internet, marketers typically counter by touting the benefits of a customized browsing experience. Wouldn鈥檛 it be irritating, for example, for childless consumers to see advertisements for diapers and preschools when marketers are perfectly capable of knowing they don鈥檛 have children? Wouldn鈥檛 you rather see a product that a sophisticated algorithm has determined you might actually want to purchase?
A few years ago, a spate of stories circulated about and predictions. According to reporting by Forbes and , the retailer looked at purchasing patterns of pregnant women in different trimesters who had already signed up for a baby registry. By sniffing out those same patterns among all its female customers, Target could then assign a 鈥減regnancy prediction score鈥 and estimate due dates to send relevant coupons. When higher-ups at Target realized that customers might find unsolicited coupons congratulating them on their first baby an invasion of privacy, the retailer responded not by stopping the practice, but by 鈥渉iding鈥 the coupons for baby products in booklets of unrelated offers so that customers assumed they were random. And it worked.
Recalling what might have led up to the car ad on Facebook, Young pinpoints a moment on a JetBlue flight when a car advertisement played on the seatback satellite TV. It was a green Kia Soul.
Young immediately liked it. 鈥淚 was like, 鈥楾hat鈥檚 an awesome car. I want that green car!鈥欌
Through the free wi-fi available on the flight via JetBlue鈥檚 web portal, he began researching a few upcoming trips he was planning to take, looking into the cost of flights, which came out to what Young calls an 鈥渙bscene amount of money.鈥
鈥淪o I had those two bits of information,鈥 he recalls. 鈥淚 like this green car, and, separately, looks like I鈥檒l have to spend a lot of money on flights coming up.鈥
When he saw the car advertisement on his Facebook page a few days later, it suddenly occurred to him that it might be more cost effective in the long run to buy a car in order to make these trips rather than spending thousands in airfare and having no equity after his purchases. Somehow, the algorithms had also made the connection that Young was primed to buy a car. They already knew where he lives, that he鈥檚 a renter rather than a homeowner, that he has a dog, that he has a steady job, that he travels frequently. 鈥 Something about his behavior online tipped the scales and made him attractive to car companies for the first time. He began researching cars in earnest.
Online marketing is ubiquitous, and it鈥檚 not going anywhere. If anything, it鈥檚 only becoming more invasive, inventive, and robust. But aside from creeping out customers with how well they can anticipate their wants and needs, what real harm are marketers doing by using these methods?
The research of Datta鈥檚 team offers one possible answer.
Using AdFisher, they ran a simulation of a thousand users鈥500 profiles identified as men and 500 as women鈥攁nd sent them on the same mission of browsing the top 100 job sites to look for employment. Then, these 鈥渕en鈥 and 鈥渨omen鈥 navigated to a third-party website that uses Google Ads鈥攕pecifically, the Times of India鈥攁nd a statistical analysis was run to see whether the ads being served were different depending on the gender of the user.
鈥淭hey were very different,鈥 says Datta. 鈥淭hat already gives us evidence in support of differential treatment.鈥 But that doesn鈥檛, he says, explain whether the reason behind the difference was actually cause for concern, something more problematic than men seeing more clothing ads for men鈥檚 clothing than women鈥檚, for instance.
The next step was figuring out which ads were most influential, that is, which ads appeared most frequently. 鈥淭hat鈥檚 where we got the startling result,鈥 Datta says. The two most influential ads being served to male users were from a career counseling service for 鈥渆xecutives only,鈥 boasting salaries of more than $200,000.
This ad was shown to the male users about 1,800 times. But for the women?
鈥淭here were only about 300,鈥 Datta reveals. 鈥淭hese kinds of ads are a gateway to opportunities for employment. This is where we felt that we started moving from differential treatment toward discriminatory treatment.鈥
The difference is stark, but who鈥檚 to blame for the inequality? Datta says it鈥檚 impossible to know at this point in the research whether the fault lies with marketers who specify 鈥渕en鈥 as a demographic, or whether it was just an unintended bias introduced by Google鈥檚 algorithm as it 鈥渓earned鈥 that men more often click on the ad so it should show the ad to more men鈥攁 kind of self-fulfilling prophecy.
鈥淎t a public policy level, if you look at the physical world, there are protections,鈥 says Datta, referring to protected statuses like race, age, and gender. 鈥淲e will have to think carefully about how those laws can be expanded to cover the online domain.鈥
鈥淯nfettered collection of personal data has a chilling effect on individual freedoms.鈥
But as in his other study involving privacy concerns, the question of who or what is to blame is much more complicated than simply recognizing that these red flags exist. Finding where that flaw lies, he says, is the first step to instituting corrective measures. He鈥檚 currently pursuing a project with Microsoft Research to develop a methodology for assigning responsibility, and then applying it to Microsoft internally, where the researchers will have more visibility into the ecosystem鈥攇etting under the hood, if you will. He hopes other organizations will use tools such as AdFisher to monitor the behavior of their online ad targeting software and that regulatory agencies such as the Federal Trade Commission will use the tool to help spot abuses.
This idea is what he and other researchers call information accountability. As he explains, it鈥檚 not just these algorithms that are making decisions; it鈥檚 a combination of man, machine, and the interactions between them.
鈥淚n modern society, there is a combination of algorithms and human actors who operate inside 鈥榖lack boxes鈥欌濃攎eaning systems with unknown inner workings鈥斺渨ho are making important decisions. And what the field of information accountability is trying to do is to ensure that there are methods for examining these systems.鈥
The goal is to provide global oversight of these ecosystems to detect any deleterious effects like the aforementioned discrimination and privacy violations, or just a lack of transparency.
Or even intentional wrongdoing.
And this is where we loop back to Young again, the recent owner of a diesel Audi A3. That make and model may sound familiar for a very unfortunate reason: Yes, Young bought one of the 11 million cars affected by the recent scandal involving Volkswagen鈥檚 intentionally deceptive 鈥渄efeat device鈥 software. (Audi is owned by Volkswagen.) This software is governed by鈥攜ou guessed it鈥攁lgorithms that told the car to react differently in certain situations, like an emissions test. The car鈥檚 emissions were markedly lower when it was being tested than during everyday use, when in reality pollutant levels were up to 40 times the legal limit. Since the scandal broke in September, Volkswagen has been excoriated in the media and faces up to $18 billion in fines alone, to say nothing of the cost to recall or fix the cars and any damages it may pay to car owners. The car manufacturer recently announced that it would be giving all affected customers a $500 gift card and $500 in credit at dealerships. No word on when actual repairs might begin, but in the meantime, VW stock has plummeted by as much as a third.
Young had debated between a hybrid and a diesel engine before settling on the A3, which he can laugh about now, if a little bitterly. 鈥淚 specifically bought the car so that it was better for the environment because I knew I was going on a lot of road trips!鈥 he says exasperatedly. He notes, with irony, that he also has asthma鈥攁 condition irritated by car emissions.
The algorithms populating Young鈥檚 Facebook feed with ads may have convinced him to buy a car, but the ones lurking under the hood got the last laugh.
Datta reflects on the depth of this seeming betrayal: We expect technology to be fair and impartial in a way that humans inherently cannot, he says. But we鈥檙e forgetting that software and the algorithms that run it are all man-made and can be imbued with man鈥檚 biases and imperfections.
But what are the corrective measures? Datta and Young both agree that a combination of internal awareness and external oversight and accountability is key. Options range from an expansion of tools like AdFisher to regulations that penalize companies found to be in violation.
Citing the example set by the Human Rights Campaign, which every year publishes a Corporate Equality Index as a benchmarking tool for LGBTQ rights in the workplace, Young sees a similar opportunity in the information accountability space. He imagines there could be an organization that stress-tests a company鈥檚 marketing algorithms, patches any problematic holes, and then 鈥渃ertifies鈥 that a company has reached a certain threshold of accountability, similar to how some businesses now become authorized retailers or 鈥渧erified鈥 vendors on PayPal.
On a personal level, though, what can you do when you are sitting in front of your computer screen, which you may be eyeing with increasing suspicion?
Take a page out of Datta鈥檚 and Young鈥檚 book. They both use Internet browser extensions that either inform them when their data are being shuffled off to marketers or block their information from being sent. Datta uses , for example. (If you Google Privacy Badger, be prepared to receive ads for similar browser extensions!) A plug-in called , which Young vouches for, shows users how many separate companies just collected their data when they load any given website. If he sees a long list begin to populate and he doesn鈥檛 absolutely need the information he鈥檚 searching for, he鈥檒l close the site before the search is complete and make a mental note not to revisit.
As for why more people don鈥檛 take actions like this to safeguard their data, Datta says it鈥檚 a combination of the general public lacking a full understanding of potential consequences while also performing a kind of informal cost-benefit analysis鈥攊f you want to opt out of having your data collected, there鈥檚 very little you can do on the Internet. But the decision-making is fundamentally flawed, he warns: 鈥淲ithout understanding the consequences, a cost-benefit analysis can lead you to the wrong conclusion.鈥
Consider yourself informed.