The following story, written by Priya Kumar, originally appeared in Slate’s Future Tense blog.

When Was the Last Time You Read a Privacy Policy?

Tech companies know that everyone skips the fine print. It’s time for them to change approaches.

At one point last fall, I had 16 tabs open on my Web browser, each displaying the privacy policy of a global Internet or telecommunications company. While conducting research for the Ranking Digital Rights Corporate Accountability Index, I read and re-read each policy, trying to figure out what companies said they did with the vast stores of user information that lived on their servers.

The end result? While the policies were lengthy, they frequently glossed over the amount of information being collected and the ways in which it is used. Sure, Google and Facebook each mentioned in their policies that they collected user information through third-party trackers. But that in no way reflects the scope of the third-party trackers: According to UC–Berkeley researcher Ibrahim Altaweel and his team, Google’s tracking mechanisms cover 85 percent of the most popular websites, and Facebook’s tracking reaches 55 percent of the most popular websites, giving both companies extraordinary visibility into the Web browsing habits of millions of people. So even those who actually read company policies don’t get a full picture of what’s going on, making it difficult to protect their personal information.

Altaweel’s was one of many findings discussed at the Federal Trade Commission’s PrivacyCon on Jan. 14 and the Future of Privacy Forum’s Privacy Papers for Policymakers event the day before. Both events highlighted the disconnect between people’s expectations and corporate practices related to the collection and use of personal information, echoing the findings of the Corporate Accountability Index.

Researchers challenged the “notice-and-choice” approach that drives privacy protection in the United States. They emphasized that yes, corporate practices related to user information need additional transparency—but simply understanding what companies do isn’t enough if people lack meaningful alternatives or mechanisms for recourse.

Historically, the United States has approached privacy as a commercial matter while countries in Europe and elsewhere typically view privacy as a fundamental right of their citizens. Consequently, other countries tend to have data protection laws that companies must follow, while American companies protect people’s privacy by providing notification of the company’s privacy practices (like the privacy policy) and giving people a choice about whether to give their data to the company (like the “I agree” button everyone clicks without thinking).

The problems of this notice-and-choice approach are evident to anyone who uses the Internet: No one reads privacy policies, and no one really feels like he or she is exercising a choice when clicking “I agree.” Furthermore, privacy policies are written in a way that satisfies regulators, not regular people. They lack clear explanations of company practices, and companies can change the policies at any time.

As usage of mobile apps, wearable technology, and smart devices for inside and outside the home continues to rise, the notice-and-choice model will become even more obsolete given that user information flows almost constantly between such devices and company servers. Consider that UC–Berkeley researcher Serge Egelman and his team found that applications on an Android phone issue nearly 100,000 requests to access a person’s sensitive data per day. No one wants to receive a notification for each of those requests. Yet 80 percent of the participants in Egelman’s study would have denied at least one of those permission requests, had they been given the opportunity to do so.

The argument that individuals rationally weigh the pros and cons of giving their data to companies doesn’t reflect reality considering the vast amount of data users generate and the number of companies that access that data. Joseph Turow, a professor at the University of Pennsylvania, calls this the “tradeoff fallacy.” He and his team found that more than half of Americans want control over their personal information, but they feel powerless to exert that control.

Instead, people use other shortcuts when deciding whether to interact with a given product or service online. Carnegie Mellon University researcher Ashwini Rao and her team found that people use websites based on their expectations of what the companies do with user information, rather than what the privacy policies say companies actually do. For example, users expect a banking website to collect financial information, not health information. Yet Rao and her team found that Bank of America’s policy says it collects and shares health information from its registered users, meaning people could be using the website under mistaken assumptions about the privacy of certain pieces of information.

In addition, professors Heather Shoenberger from the University of Oregon and Jasmine McNealy from the University of Florida found that people are more likely to accept a website’s terms if the site has a privacy policy—regardless of what the policy actually says—or if it has appealing design aesthetics. These behaviors suggest that people may be using websites under mistaken assumptions of what they believe companies are doing with their information. Additional transparency could ameliorate the situation, but what’s truly needed is for companies to act in ways that respect users’ privacy.

Consider this: Columbia University professor Roxana Geambasu and her team found evidence that an advertising tool used by Gmail until November 2014 targeted ads based on sensitive personal information—something its policies, both those in place in 2014 and those in place now, say it does not do. (Google refused to comment on Geambasu’s research, but a representative said, “We have an extensive set of policies that guide what ads can be shown, and in Gmail we manually review all ads that are shown.”) And while Geambasu and other researchers emphasize that results like these do not imply that such targeting is intentional, such work does highlight the need for better understanding of how systems like Google’s algorithms determine who sees what information.

Users currently lack adequate frameworks to seek redress if they believe a company’s actions violate their privacy. The Corporate Accountability Index found a dearth of clear processes for remedy among companies. India’s Bharti Airtel scored highest in the index among telecommunications companies on remedy, and South Korea’s Kakao scored highest among Internet companies. Both companies are headquartered in countries that maintain certain legal requirements for remedy, which boosted their performance.

Fordham University law professor Joel Reidenberg suggests that a new international treaty will be needed to protect privacy in a global, networked era. Otherwise we risk allowing governments to gain greater insight into the lives of their citizens, while the ways that governments use such information become more opaque. Indeed, while a significant portion of user information lives on company servers, privacy policies typically state that companies may turn information over to the government, though, of course, details about the circumstances in which they would do so are scant.

 What else can be done? Turow encouraged public interest organizations to examine company policies and highlight how well, or not so well, companies perform. Ranking Digital Rights’ Corporate Accountability Index does just that, by evaluating a group of technology companies according to standards related to freedom of expression and privacy. The index found that, overall, company disclosure about the collection, use, sharing, and retention of user information is poor. But companies can take steps in the right direction by, for example, providing a list of the third parties with which they share user information, as Yahoo does, or specifying how long they retain user information when people delete their accounts, as Twitter does. Change won’t happen overnight, but if the robust conversations at events such as PrivacyCon are any sign, many people are working to achieve that change.

Last week, Ranking Digital Rights participated in PrivacyCamp and the Computers, Privacy and Data Protection (CPDP) conference in Brussels. Two issues dominated the discussions: government mass surveillance, especially in light of the Schrems Safe Harbor decision (PDF), and the new EU General Data Protection Regulation. Participants also discussed corporate practices and their impact on privacy.

At CPDP, the panel “Appfail or Appwin” discussed how mobile apps may or may not respect users’ right to privacy. For example, Finn Myrstad of the Norwegian Consumer Council investigated the terms & conditions of apps, which may sometimes change without notice, and sometimes impose perpetual, worldwide and irrevocable licenses on users. They conducted a fun experiment to see how that plays out in the streets of Oslo. The Council will soon release a report about apps’ terms. In turn, Richard Tynan of Privacy International explained that even if apps don’t demand your real identity, your device collects enough data to compile a reliable picture of who you are. An audience member working for the Dutch Data Protection Authority called on Google to require that app developers publish a privacy policy, as well as to add an option for users to temporarily grant permissions for apps to do certain things such as access location only at specific times, but not generally.

Other sessions also touched on how companies should address privacy concerns. We learned that more and more companies are thinking about privacy impact assessments as ways to mitigate risks and ensure legal compliance, even though, according to one speaker, individuals within companies conducting the assessment may lack sufficient awareness about the range of privacy risks people might face when using the company’s services. Also, accountability frameworks are emerging as instruments for companies to go beyond compliance with data protection regulations, to being able to credibly demonstrate their practices to external stakeholders. For example, in order to meet RDR’s standard for corporate policy and practice, companies should be accountable not only to business partners and regulators, but also to affected individuals and the wider public.

In a panel on transparency reporting, Artur Alves of Concordia University recognized that transparency reports have beefed up over the past years, but could gain more in uniformity and transparency on internal processes. Microsoft’s Mark Lange highlighted the company’s Transparency Hub, and further talked about national security related requests. He explained that Microsoft is limited in its disclosure on receiving such requests, due to U.S. legal restrictions. Niels Huijbregts of Dutch ISP XS4ALL, said that customers welcome their transparency reports, but he had to overcome hesitation within XS4ALL’s parent company KPN, where some feared harming government relations. Nate Cardozo talked about EFF’s Who’s Got Your Back report, and said that companies are still deficient when it comes to national security-related reporting, reporting on government requests for terms of service enforcement, as well as reporting on informal processes.

Further discussions at CPDP dealt with anything from interactive toys and the need to have granular control over what kind of data they collect from children, to data minimization, where a European Commission panelist called for privacy compliance as a competitive advantage.

Ranking Digital Rights supports the view that improved privacy practices provide business opportunities for companies. The many conversations at CPDP confirmed that a human rights-centered approach to privacy accountability is necessary to improve companies’ practices.

As we settle into the new year, RDR and its Index continue to make the news.

Rebecca MacKinnon wrote an op-ed for CNN arguing that as technology companies face greater pressure to monitor and censor communication, we risk losing the ability to use such platforms for social justice, activism, and journalism. Priya Kumar wrote the lead article for the January 28th edition of New America Weekly, which describes how RDR’s Index helps foster greater corporate accountability among ICT companies. And in a piece for Slate, Priya drew on the latest academic research and the Index results to highlight what’s lacking from company privacy policies. Nathalie Maréchal wrote a paper about RDR for a special issue of The Fibreculture Journal focused on activism and technology.

PRI-screenshot

Screenshot of Public Radio International’s story on digital rights

Other media outlets and advocacy organizations continued to interview RDR team members and draw on the Index results. Rebecca MacKinnon was quoted in an Atlantic article about American Internet companies doing business in China. The Public Radio International program, “Whose Century Is it?” did a piece on digital rights that highlighted Rebecca’s work in the space, including RDR. An interview with Rebecca (conducted two years ago) was featured in the documentary Facebookistan, released late last year. The film focuses on efforts to get companies such as Facebook to respect users’ rights to freedom of expression and privacy. The title comes from a term Rebecca coined in her book, “Consent of the Networked.” The civic news outlet CivSource quoted Rebecca about an open letter signed by more than 170 organizations and individuals in support of strong encryption. And in a blog post about Internet shutdowns in Egypt and Morocco, AccessNow cited Etisalat’s poor performance in RDR’s Index.

Since its release in early November, the Ranking Digital Rights inaugural Corporate Accountability Index has received worldwide media attention, demonstrating global interest in corporate respect for users’ rights.

In coordination with the launch of the Index on November 3rd, The Guardian newspaper published two in-depth analyses of the Index data and its implications:

Rebecca MacKinnon discussed RDR's Index on the Wall Street Journal's Opinion Journal channel. Screenshot via Opinion Journal

Screenshot of Rebecca MacKinnon discussing RDR’s Index on the Wall Street Journal’s Opinion Journal channel.

The RDR team has been interviewed and invited to write pieces about the Index. Rebecca MacKinnon discussed the findings on the Wall Street Journal’s Opinion Journal channel and with Deutsche Welle. Priya Kumar talked about the Index on the Digital Show on Sirius XM’s Wharton Business Radio channel. Allon Bar wrote about the Index in a guest post for the Measuring Business and Human Rights project, based at the London School of Economics. He also talked about the results on the German Info-radio.

In addition, news about RDR’s Index and its findings appeared in NBC News, Fast Company, Ars Technica, Bloomberg Brief, Fortune, SFGate, TechWeek Europe, Hacked, WNYC, Information Week, CXO Today, IT ProPortal, WebProNews, Beta News, SC Magazine, TechWorm and Digit (India).

An Agence France-Presse story about the Index was syndicated around the world, including The Times of India, The Economic Times (India), Malay Mail Online, The Malaysian Insider, Free Malaysia Today, The Star (Malaysia) The Business Times (Singapore), The Bangkok Post, The Manila Times, Interaksyon (The Philippines) The Peninsula (Qatar), The Citizen (South Africa), France24, and Le Figaro (in French). International outlets also covered the story: la Repubblica (in Italian), Süddeutsche Zeitung and Taz (both in German) and NU.nl and WebWereld (both in Dutch). The Institute for Human Rights and Business, Article 19 and the Committee to Protect Journalists wrote about the Index.

Media are already using RDR’s Index to provide context when discussing digital rights issues. The Pakistani newspaper Dawn referenced the Index results in a story about Facebook’s updated transparency report. ThinkProgress highlighted the Index in an explanation of how the U.S. Congress struggles to craft cybersecurity legislation that respects individual privacy rights. OpenDemocracy underlined RDR’s findings in its argument that technology companies endanger their reputations by ignoring digital rights. Another post on the LSE’s Measuring Business and Human Rights project blog explains how rankings such as RDR’s Index can spur company action.

Stay tuned for more from the RDR team in the coming weeks and months!

YouTube video of 2015 IGF RDR/Internews session on Using the Corporate Accountability Index for Research and Advocacy

YouTube video of 2015 IGF workshop Benchmarking ICT companies on digital rights

After a series of launch events in the U.S., the RDR Corporate Accountability Index took to the global stage earlier this month as members of the RDR team presented the Index’s results at the Internet Governance Forum, an opportunity to engage with civil society advocates, researchers, government officials and private sector representatives from around the world. We received feedback on the project, and brainstormed about future activities with potential partners and collaborators

Reactions to the Index were positive. Several people commented on the painstakingly detailed approach to data collection and analysis. This is data one can stand behind. Most of the critiques could be summarized as “do more of the same… much, much more.” More companies, more services, in more markets… which of course requires more resources than the project currently possesses, but we are working on it…

Another common concern is that the Index measures disclosed policies and practices but does not take the next step to verify whether companies are actually carrying out the policies and practices that they describe. We also lack the resources to do such work, which would require on-the-ground staff in many countries. Instead, we welcome researchers and advocates to develop projects that would carry out the verification work in a globally distributed way: it would be ideal if a multitude of advocacy groups and academic institutions could develop their own approaches to verify and track how companies’ policies and practices are actually experienced by users in different parts of the world.

We also learned more about similar yet complementary projects that approach the same issue — corporate responsibility to respect human rights online — from different angles. If this ecosystem can seem like a messy patchwork at first glance, the conversations that RDR participated in highlighted the importance of collaboration between projects. As Carolina Botero, of Colombia’s Fundación Karisma, shared during our Day 2 workshop, being able to tell representatives from domestic Internet service providers that transparency reporting is now a standard practice that global giants like Google and Facebook routinely engage in was key in changing the tone of the conversation, which had been rather confrontational until then.

Day 0: Brainstorming Session

On “Day 0” of the conference devoted to self-organized events by conference attendees, RDR and Internews held a research and advocacy brainstorming session around the newly released Corporate Accountability Index. What additional research projects could emerge from the Index data, and how can the Index support advocacy?

Following opening remarks by David Kaye, the UN’s Special Rapporteur on Freedom of Expression, and a presentation of the Corporate Accountability Index, the Ranking Digital Rights team engaged the participants in a wide-ranging discussion of the Index, its potential for both research and advocacy, and also its limitations.

Participants rightly pointed out that while this evaluation of 16 companies is a commendable start, future iterations of the Index (ideally on an annual basis) will have greater impact if they include a wider range of companies, including more of the companies’ direct competitors. This, of course, is contingent on a combination of successful fundraising and on forging partnerships with civil society groups around the world. All of Ranking Digital Rights materials’, including our methodology and research guidance, are freely available under Creative Commons licensing, and we encourage our colleagues in the digital rights space to consider applying the RDR criteria to a wider range of companies and subsidiaries, for example by doing regional or country-specific rankings.

A thornier limitation is that because it only evaluates publicly available information (for reasons documented here), the Index cannot address the issue of company practices that diverge from their stated commitments. While there is a clear need to highlight this kind of disparity where it exists, our research has revealed that it is much more common for companies to simply not have a policy in place, or to have a an internal policy that users and the general public have no access to. We believe that RDR can best promote greater corporate accountability for human rights by incentivizing companies to develop and make public policies surrounding commitment to human rights, privacy and freedom and expression. Once companies make public commitments and disclosures, it is then possible for stakeholders to hold them accountable by verifying that they are living up to their commitments and actually adhering to their own stated policies. One project aiming to highlight companies’ actual practices is the recently launched OnlineCensorship.org, where users are able to submit their experiences with content takedown and other online censorship. Lumen (formerly Chilling Effects) allows users to upload information about take-down requests and receive information about their legal rights.

Day 2: Benchmarking ICT companies on digital rights

On Day 2 of the IGF, Rebecca MacKinnon participated in a roundtable workshop on “Benchmarking ICT companies on digital rights,” which I moderated. The participants’ experiences point to the diversity of strategies that can be effective in getting companies to improve their respect for digital rights: from star-based ratings like the EFF’s Who Has Your Back? and Fundación Karisma’s ¿Dónde están mis datos? to the RDR Index’s granular scoring system, there is more than one way to encourage ICT companies to respect human rights. The Transparency Reporting Index maintained by Access Now is formidable resource for activists from a wide variety of subsectors, providing links and brief descriptions for the transparency reports of more than 50 global companies. The Terms of Service & Human Rights project translates the “legalese” that these documents are usually written in into plain language that users can understand, thus empowering them to make informed decisions about the services they use. It is clear that these projects, and others, rely on and support one another. All human rights advocacy is an ecosystem to an extent, but this is all the more true when we’re dealing with multinational corporations whose actions impact users all over the world in different but related ways.

Comments by Cecille Soria of Democracy.Net.PH and Kelly Kim of Open Net Korea reinforced the importance of starting with a low bar and raising it gradually. While both Democracy.Net.PH and Open Net Korea would eventually like to start ranking projects using the RDR criteria, the reality is that many companies in their respective countries do not currently provide the kind of information that the Index’s methodology seeks to surface. Even though the South Korean company Kakao stood out for its relatively robust disclosures in the RDR Index, Kim said Korean companies still have considerable progress to make. Open Net Korea plans to focus on getting companies to publish regular transparency reports before scrutinizing the contents of those reports. This echoed the earlier comments from the EFF’s Jeremy Malcolm, who noted that the Who Has Your Back? report has gradually raised the bar for the companies it evaluates.