The European Union plans to beef up its response to, with the Commission saying today it will step up efforts to combat harmful but not illegal content — including by pushing for more minor digital services and adtech companies to sign up to voluntary rules aimed at tackling the spread of this manipulative and often malicious content.
EU lawmakers pointed to risks such as the threat to public health posed by the spread of harmful disinformation about online disinformation on democratic processes are another driver, they said.as driving the need for more brutal action. Concerns about the impacts of
A new, more expansive code of practice on disinformation is now being prepared — and they hope to be finalized in September to be ready for application at the start of. The Commission’s gear change is a fairly public acceptance that the EU’s voluntary code of practice — and approach Brussels has taken since 2018 — has not worked out as hoped. And, well, we did warn them.
Aon board with demonetizing viral disinformation is undoubtedly overdue. The online disinformation problem hasn’t gone away. Some voter manipulation and computational propaganda — has been getting worse in recent years rather than better.
However, getting visibility into the accurate scale of the disinformation problem remains a considerable challenge, given those best placed to know (ad platforms) don’t freely open their systems to external researchers. And that’s something else the Commission would like to change. Signatories to the EU’s current code of practice on disinformation are:
Google, Facebook, Twitter, Microsoft, TikTok, Mozilla, DOT Europe (Former EDiMA), the World Federation of Advertisers (WFA) and its Belgian counterpart, the Union of Belgian Advertisers (UBA); the European Association of Communications Agencies (EACA), and its national members from France, Poland and the— respectively, Association des Agencies Conseils en Communication (AACC), Stowarzyszenie Komunikacji Marketingowej/Ad Artis Art Foundation (SAR), and Associate Komunikacnich Agentur (AKA); the Interactive Advertising Bureau (IAB Europe), creativity & Kommunikation, and Goldbach Audience (Switzerland) AG.
EU lawmakers said they want to broaden participation by getting smaller platforms to join and recruiting all the various players in the adtech space whose tools provide the means for monetizing online disinformation. Commissioners said they want to see the code covering a “whole range” of actors in the online advertising industry (i.e., rather than the current handful).
It’s cerNotably; ital advertising industry body Internet Advertising Bureau is not on that list. (We’ve contacted IAB Europe to ask if it’s planning to join the code and will this report with any response.) In its press release today, the Commission also said it wants platforms and adtech players to exchange information on disinformation ads that have been refused by one of them. So there can be a more coordinated response to shut out .
As for those signed up already, the Commission’s report card on their performance was bleak. Speaking during a press conference, internal market commissioner Thierry Breton said that only one of the five platform signatories to the code has “really” lived up to its commitments — which was presumably a reference to the first five in the above list (aka Google, Facebook, Twitter, Microsoft, and TikTok).
Breton demurred on doing an explicit name-and-shame of the four others — who he said have not “at all” done what was expected of them — saying it’s not the Commission’s place to do that. Instead, he said people should decide which platform giants that signed up to the code have failed to live up to their commitments.
(Signatories since 2018 have pledged to take action to disrupt ad revenues of accounts and websites that spread disinformation; to enhance transparency around political and issue-based ads; tackleand online bots; to empower consumers to report disinformation and access different news sources while improving the visibility and discoverability of authoritative content, and to empower the research community so outside experts can help monitor online disinformation through privacy-compliant access to platform data.)
Safe to say, there’s been a lot more hot air (in the form of selective PR) on the charged topic of disinformation vs. hard accountability from the major social platforms over the past three years. Frankly,to imagine who from the above five tech giants might be meeting the Commission’s bar. (Microsoft, perhaps, because of its relatively modest social activity vs. the others.).’
So it’s perhaps no accident that Facebook chose today to puff up its historical efforts to combat what it refers to as “influence operations” — aka “coordinated efforts to manipulate or corrupt public debate for a strategic goal” — by publishing what it couches as a “threat report” detailing what it’s done in this area between 2017 and 2000.
Influence ops refer to online activity that may be cod by hostile foreign governments or by malicious agents seeking, in this case, to use Facebook’s ad tools as a mass manipulation tool — perhaps to try to skew an election result and influence the shape of looming regulations. And Facebook’s ‘threat report’ states that the tech giant took down and publicly reported only 150 operations over the reporting period.
Yet as we know from Facebook whistleblower Sophie Zhang, the scale of the problem of mass malicious manipulation activity on Facebook’s platform is vast, and its response to it is both under-resourced and PR-led. (A memo by the former Facebook data scientist, covered by BuzzFeed, detailed a lack of institutional support for her work and how takedowns of influence operations could almost immediately respawn — without Facebook doing anything.)
NB: If it’s Facebook’s “broader enforcement against deceptive tactics that do not rise to the level of [Coordinate Inauthentic Behavior]” that you’re looking for, rather than efforts against ‘influence operations’, it has a whole other report for that — the Inauthentic Behavior Report! — because, of course, Facebook gets to mark its homework when tackling fake activity and shapes its level of transparency since there are no legally binding reporting rules on disinformation.
Legally binding rules on handling online disinformation aren’t in the EU’s pipeline either — but commissioners said today that they wanted a beefed-up and “more binding” code. They do have some levers to pull here via a more comprehensive package of digital reforms that are coming (aka the Digital Services Act).
The DSA will bring legally binding rules for how platforms handle illegal content. They intend the more brutal disinformation code to plug into that (as a “co-regulatory backstop for the measures that will be included in the revised and strengthened Code”).
It still won’t be legally binding, but it maycompliant platforms with wider DSA’ credit’. So it looks like disinformation-muck-spreaders’ arms are set to be twisted in a pincer regulatory move by making sure this stuff is looped into the legally binding DSA.
The digital regulation packages the EU has put forward since the 2019 college took up its mandate generally aim to increase transparency, safety, and accountability online, its values and transparency commissioner, Vera Jourova, said today. The risks are that a centralized approach might smell like censorship — and it sounds keen to avoid that charge at all costs. Still, Brussels maintains that it does not want to legislate around disinformation.
Breton also said that now is the “right time” to deepen obligations under the disinformation code — with the DSA incoming — and give the platforms time to adapt (and involve themselves inresponsibilities).
In another exciting remark, he also talked about regulators needing to “be able to audit platforms” — to be able to “check what is happening with the algorithms that push these practices”. Though audit powers can be made to fit with a voluntary, non-legally binding code of practice remains to be seen.
Discussing areas where the current code has fallen short, Jourova pointed to application inconsistencies across different EU Member States and languages. She also said the Commission is keen for the beefed-up code to enable and empower users to act when they see something dodgy online — such as by providing users with tools to flag problem content.
Platforms should also allow users to appeal disinformation content takedowns (to avoid the risk of opinions being incorrectly removed). The focus for the code would be on tackling false “facts, not opinions”, she emphasized, saying the Commission wants platforms to “embed fact-checking into their systems” and for the code to work towards a “decentralized care of facts”.
She said that the current signatories to the code hadn’taccess the Commission would like to see — to support greater transparency into (and accountability around) the disinformation problem.
The code does requires (for COVID-19 disinformation),six-monthly,or yearly reports from signatories (depending on the entity’s size), but what’s being provided so far doesn’t add up to a comprehensive picture of disinformation activity and platform reaction; she said.
She also warned that— while saying the Commission would nonetheless like to see signatories agree on a set of identifiable “problematic techniques” to help speed up responses. She noted that EU lawmakers would have a specific plan for tackling political ads transparency in November.
They are also, in parallel, working on how to respond to the threat posed to European democracies bycyber ops — such as the influence mentioned above operations often found hosted on Facebook’s platform.
The commissioners did not give many details of those plans today. Still, Jourova saidto impose costs on perpetrators,” — suggesting that some interesting possibilities may be considered, such as trade sanctions for state-backed drops (although attribution would be one challenge). Breton said countering foreign influence over the “informational space” is vital work to defend the values of European democracy.
He also said the Commission’s anti-disinformation efforts would focus on support for education to help equip citizens with the necessary critical thinking capabilities to navigate the vast quantities of variable quality information that now surrounds them.