The White House has a plan for Big Tech

Donald Trump, Mark Zuckerberg and Joe Biden all agree.

Cory Doctorow
10 min readSep 10, 2022
The logo for the White House, superimposed over a Matrix ‘code waterfall’ effect.

This week, the White House released its long-anticipated plan for addressing monopoly in the tech sector. Fixing Big Tech is important, because a free, fair and open internet is a necessary precondition for organizing all our other fights about human rights, equity, labor, the climate, and racial and gender justice.

https://www.whitehouse.gov/briefing-room/statements-releases/2022/09/08/readout-of-white-house-listening-session-on-tech-platform-accountability/

The White House plan is a mixed bag. They set out six action points, each of them amorphous enough that they could all be summarized as “the devil is in the details” — that is, depending on how these are handled, they could be great, or terrible.

But one point stands out as especially fraught, controversial and dangerous: a vague promise of “fundamental reforms to Section 230,” which is incorrectly characterized as “special legal protections for large tech platforms.”

I’m going to go through all six of the points below and describe how they could go right, or wrong, and in the end I’ll get into more detail on 230 — it’s one of the worst-understood areas of internet law, a favored punching bag of the right and the left, and getting this one wrong could deliver permanent dominance to Big Tech platforms.

I. “Promote competition in the technology sector”: This covers both meat-and-potatoes trustbusting (breakups, merger scrutiny) and modern, tech-specific tactics, like interoperability mandates and bans on self preferencing. This is generally great stuff, but there are three important pitfalls to avoid:

i. Interop mandates that expose users to risk through hasty action. The EU’s Digital Markets Act unwisely kicked off by mandating interop in messaging tools on an unrealistically short timeline. Maintaining the security of encrypted messengers is extremely important; failures in messaging encryption are a source of existential risk to human rights workers, journalists and marginalized people all around the world. Recall that Jamal Khashoggi was lured to his slaughter by the Saudi government after they broke into his peers’ encrypted messages using a cyberweapon produced by the NSO Group.

https://www.eff.org/deeplinks/2022/04/eu-digital-markets-acts-interoperability-rule-addresses-important-need-raises

ii. Must-carry rules that force platforms to carry speech. Big online platforms have become our new public square, except that they aren’t public — they’re private. Their choices about which speech to block and which speech to carry are enormously consequential for our civics and politics. But rules that allow regulators to force providers to carry speech they disagree with set a dangerous precedent. Even if you think that the Biden admin’s compelled speech will be fine (say, a rule requiring warnings alongside vaccine disinformation), imagine how this power will be handled by President Marjorie Taylor Green’s administration.

The platforms’ moderation choices are a danger because the platforms dominate our discourse. Allowing the platforms to corner the market for online speech has profound First Amendment implications:

https://locusmag.com/2020/01/cory-doctorow-inaction-is-a-form-of-action/

But the answer isn’t to turn the platforms into an arm of the state — it’s to make their moderation choices less consequential for all of us, by devolving control over community norms to the communities themselves:

https://www.eff.org/deeplinks/2021/07/right-or-left-you-should-be-worried-about-big-tech-censorship

iii. Self-preferencing bans are very hard to administer. If Apple puts its own weather app at the top of the app-store listings, or if Google shows you an infobox with its weather prediction at the top of a search, that might feel like self-preferencing. But maybe Apple really believes that it has the best weather app. There isn’t an objective standard for “best weather app.” Unless you’ve got a front-row seat for the wall of Plato’s Cave, distinguishing self-preferencing from good-faith curation is often impossible.

That’s not to say that we should tolerate self-preferencing, nor is it to say that we can’t ever detect and punish self-preferencing. Sometimes, tech companies actually document the fact that they’re self-preferencing, as Google did when its engineers emailed their bosses to complain about being forced to put Google’s inferior results ahead of rivals:

https://blog.yelp.com/news/yelp-testifying-in-google-antitrust-hearing/

But we can’t rely on Big Tech tripping over its own dick every time it does a bit of nefarious self-preferencing. The real remedy for self-preferencing is “structural separation”: banning platform operators from competing with platform users. Referees shouldn’t own one of the teams on the field, period.

https://locusmag.com/2022/03/cory-doctorow-vertically-challenged/

II. “Provide robust federal protections for Americans’ privacy.” A no-brainer. The US needs a federal privacy law, with a private right of action that allows individuals (and human rights groups) to sue firms that violate it, rather than waiting for a prosecutor to take up their cause. Do it.

I’m entirely unsympathetic to the argument that “targeted ads” are better than “untargetted ads” because they are “more relevant” to users. Users fucking hate targeted ads. Ad-blockers are the largest boycott in human history. When users are given the change to opt out of targeted ads, they do so in such overwhelming numbers that the holdouts are likely to be people who accidentally clicked the wrong button:

https://arstechnica.com/gadgets/2021/05/96-of-us-users-opt-out-of-app-tracking-in-ios-14-5-analytics-find/

III. “Protect our kids by putting in place even stronger privacy and online protections for them, including prioritizing safety by design standards and practices for online platforms, products, and services.”

Sounds good. As a dad, I like the idea. But there’s so many ways it can go wrong. California’s version of this rule was so vaguely worded that it’s effectively impossible to comply with.

https://www.techdirt.com/2022/09/06/techdirt-podcast-episode-328-the-problems-with-the-california-kids-code/

It’s not just that this could result in kids being banned from using any online service — it’s also that all online services might institute invasive verification procedures (like requiring and storing — and, inevitably, leaking — government IDs to prove that none of their users are kids).

But the difficulties here don’t mean we have to be nihilists. We can demand that platforms that target kids — that market themselves as services for children — eschew advertising, minimize data collection, and take other steps to protect kids from commercial predation.

V. “Increase transparency about platform’s algorithms and content moderation decisions.” Opponents of this one will claim that telling people how you moderate is a gift to trolls and griefers. I’m unsympathetic to the idea that there is “security through obscurity”:

https://doctorow.medium.com/como-is-infosec-307f87004563

There’s a lot of room for debate about how the “civil justice” system of big platforms should operate. One thing is clear: automated judgments about user speech can’t be balanced by human review. The former happens at scale and near-instantaneously. The latter will either be deliberative and too slow to matter, or rapid and too quick to make sense of nuance.

One intriguing idea is to structure content moderation review as a “systemic” matter, which can address “immoderation” (the content that isn’t moderated) as well as moderation. Note that no one has tried this yet, so while it sounds great, it’s also a gamble:

https://pluralistic.net/2022/03/12/move-slow-and-fix-things/#second-wave

VI. “Stop discriminatory algorithmic decision-making.” This one is also maddeningly vague. If they’re talking about ensuring that machine learning classifiers don’t discriminate on the basis of speech, it’s going to be very hard to make work. Remember, algorithmic moderation often operates on the context of speech as much as the content — if a bunch of seemingly coordinated users all post something that seems like harassment all at once, that speech might get labelled or suppressed or deleted. The exact same speech, posted by one person, once, might be left alone.

But there’s another kind of algorithmic discrimination, whose most obvious case is the algorithms that target predatory financial products to Black users, or exclude women and racial minorities from being shown good jobs on employment sites. This is illegal — and we don’t need new laws to prosecute it. But we do need new enforcement powers and resources for existing regulators to tackle it.

All right, that’s the five least controversial points in the White House plan. But I left out point IV: “Remove special legal protections for large tech platforms.”

Here, the White House is talking about Section 230 of the Communications Decency Act, AKA “The 26 Words That Made the Internet.”

https://www.cornellpress.cornell.edu/book/9781501714412/the-twenty-six-words-that-created-the-internet/

CDA230 is a rule that says that if a user’s speech violates federal law, legal responsibility for that speech falls on the speaker, not the intermediary that brought you that speech. It’s a rule that makes hosting user speech possible, period. It’s how we get Facebook and Twitter, sure — but also how we get blog comments, Mastodon instances, and other independent platforms.

It’s also how we get the infrastructure that makes it possible for individuals, nonprofits, private groups and co-ops to create their own speech forums. CDA230 means that a hosting company doesn’t need to review all its customers’ users’ speech before hosting them (imagine if every web-page had to be vetted by your host before you could make it live — and then every change also had to go through legal review).

This is important in a competitive market, but it’s even more important in our current, monopolized world, where getting kicked off of a platform might doom a speech forum (and again, if you’re comfortable with this being used to nuke forums that the politicians you agree with get rid of, imagine which forums President DeSantis will target).

Any gun on the mantlepiece in Act I is sure to go off by Act III. If we hand any aggrieved party the right to remove speech without a trial, we can be sure that this facility will be abused by the worst people in the worst ways.

We know this because we’ve got decades of experience with the “notice-and-takedown” system for copyright enforcement, which allows anyone claiming to be a rightsholder to get almost anything taken down from almost anywhere, irrespective of whether a copyright infringement took place.

To see that in action, check out Eliminalia, which uses fraudulent copyright takedowns to launder the reputations of dictators, torturers, murderers and rapists, getting news articles and personal accounts of their victims and survivors removed from the internet:

https://pluralistic.net/2021/04/23/reputation-laundry/#dark-ops

In Germany, Sony Music is attempting to force Quad9, a public DNS provider, to block the records of websites whose users have allegedly posted links to other websites where infringing copies of Sony’s copyrighted works can be found:

https://quad9.net/news/blog/an-update-to-the-quad9-and-sony-music-german-court-injunction-august-2022/

Sony is a serial abuser of its ability to moderate speech; the company routinely and wantonly deletes independent musicians’ performances of classical compositions by falsely claiming that they violate Sony’s copyrights — in other words, Sony is a music pirate on an unimaginable scale:

https://pluralistic.net/2020/05/22/crisis-for-thee-not-me/#filternet

Our evidence for what a post-CDA230 internet would look like isn’t limited to the copyright wars — for a more recent, more direct look at what happens when you make intermediaries responsible for their users’ speech, look at the aftermath of SESTA/FOSTA.

https://www.eff.org/deeplinks/2018/03/how-congress-censored-internet

SESTA/FOSTA is a (nominal) anti-sex-trafficking rule that creates criminal liability for companies whose services are used in connection with the heinous crime of sex trafficking. The immediate impact of SESTA/FOSTA was the mass, internet-wide removal of sites that sex workers used to keep themselves safe.

SESTA/FOSTA pushed sex workers back onto the streets, deprived them of the forums where they shared information about dangerous clients, and created a renaissance in pimping, as sex workers were forced to turn to third parties for their protection.

https://www.eff.org/deeplinks/2019/02/fosta-already-leading-censorship-we-are-seeking-reinstatement-our-lawsuit

Curbing CDA230 is especially dangerous in light of the calls for a “fairness doctrine” for online platforms. One of the activities that CDA230 protects is moderation, allowing online hosts to remove harassing, hateful, threatening or otherwise odious speech without worrying that this requires that they remove every such instance.

This allows moderators to distinguish between a racist who calls another user by a slur, and a user who says, “Can you believe that racist called me :slur:?” Before 230 was enacted, courts took the position that once a service moderated any speech, it took on the duty to moderate all speech, creating the perverse incentive to ignore bad speech.

Some say that CDA230 protects Big Tech platforms only to the extent that it protects all online speech forums, including independent ones. But this is wrong. CDA230 protects small platforms more than it protects large ones — because large ones are better situated to hire the armies of lawyers and moderators to pore over and comma-fuck everything their users post.

That’s why Mark Zuckerberg supports eliminating CDA230. As he is fond of pointing out, Facebook’s budget for human moderators exceeds Twitter’s total revenue. He understands that if you need to be as big as Facebook to compete with Facebook that:

a) No company will ever compete with Facebook, and

b) No government will ever make Facebook any smaller.

It’s not just Zuck that hates 230 — it’s also Donald Trump. Trump understands that removing legal protections for intermediaries will make them less able to stand up to rich and powerful people who can hire vicious attack lawyers who pride themselves on suppressing speech:

https://www.thedailybeast.com/60-minutes-boss-hired-law-firm-over-metoo-story/

Trump loves the kinds of lawyers who kept #MeToo at bay for decades, not just by threatening the survivors of abuse, but by scaring anyone who might host their testimony into removing it.

The fact that Zuck and Trump think killing CDA230 is a great idea should at least give its progressive opponents a moment’s pause.

https://www.theregister.com/2022/09/09/biden_tech_reform_section230/

Cory Doctorow (craphound.com) is a science fiction author, activist, and blogger. He has a podcast, a newsletter, a Twitter feed, a Mastodon feed, and a Tumblr feed. He was born in Canada, became a British citizen and now lives in Burbank, California. His latest nonfiction book is How to Destroy Surveillance Capitalism. His latest novel for adults is Attack Surface. His latest short story collection is Radicalized. His latest picture book is Poesy the Monster Slayer. His latest YA novel is Pirate Cinema. His latest graphic novel is In Real Life. His forthcoming books include The Shakedown (with Rebecca Giblin), a book about artistic labor market and excessive buyer power; Red Team Blues, a noir thriller about cryptocurrency, corruption and money-laundering (Tor, 2023); and The Lost Cause, a utopian post-GND novel about truth and reconciliation with white nationalist militias (Tor, 2023).

--

--