Thursday, June 29, 2023

Transatlantic Dialogue Workshop, Institute for Information Law (IViR), Amsterdam Law School Part 1: Overarching Questions

My apologies, but I’m extremely jetlagged and will not attribute well or capture a lot of nuance.

Chair: João Pedro Quintais

Impulse Statement: Niva Elkin Koren: Déjà vu from 1990s: radical technology change, but the world is different and tech moves in a different direction. Polarization/globalization as a response to isolation of Russia and China. Investment in R&D is not in distributive, generative things that are open to the public but in seeking private domination. Conflicts b/t types of regulation where lots of changes are happening at the same time. Copyright as an example: requiring disclosure of datasets used for training by the initiators of the model doesn’t solve any of the actual problems of how users are using the model by adding new inputs to train it further. Learning from other things regulated in digital ecosystem: Database Directive, GDPR. GDPR has become a gold standard but how much do we know about whether it’s enforced or whether it makes any difference in people’s actual level of privacy? It definitely created a regulatory burden that affected competition, but what became of the big promises? US pushes to provide alternative standards—crossborder privacy regulation as an alternative.

Impulse Statement: Matthias Leistner: You didn’t know generative AI existed and so your regulations don’t cover it. ECJ is getting first action on application of DSA to Zalando’s classification as VLSOP. I’ve never heard of Zalando but it is an online fashion platform; systemic risks, even if they could be identified, are very different for Amazon/retail than for Facebook. Litigation indicates: it isn’t more efficient than standard competition or sector-specific regulation; it just reflects that there wasn’t enough information about what they were regulating.

Sector-specific approach raises fundamental interface issues—DSA sits uncomfortably with GDPR and proposed AI Act. Aggravated if, beyond public law regulation which allows smoothing of inner inconsistencies by nonenforcement, this becomes basis for private law liability. Martin Husovec would say it’s definitely not, but under © liability cases, the CJEU would ask whether a diligent operator has followed all necessary precautions and duties—will be tempting to say that failure to comply w/DSA=© liability. And outside the large platforms we’d really have a problem.

Comment: political pressure to regulate leads to focus on what’s feasible. In an ideal picture, what are the fundamental first principles that are distinct from current platform issues? Internet is multilayer, with infrastructure/nontraditional hosting actors. Can we agree on human involvement in automation processes? Redress mechanisms? Piecemeal approach leads to overlaps and national regime conflicts. Where to focus to fix?

Martin Senftleben: The more complex the system gets, the lighter and less detailed the framework should be—open-ended notice and takedown would be better than DSA. But the DSA offers some room b/c they’ve overdone it so much. There’s so much complexity and inconsistency that we could use that to say there’s no clear answer, so to harmonize the legislation, there is space for us as academics to make sense of it and keep it up to date. DSA is based on three categories of hosting; search engines entered at the last moment, and the DSA underperforms as to entities that do hosting, content generation, and search. This type of content provision is growing enormously, as w/Microsoft’s use of AI. Already outdated.

RT: Jennifer Pahlka’s Recoding America (highly recommended!) documents regulatory failure from a waterfall approach where detail keeps getting added at each stage. Suggestions become requirements and you end up with procedures that require outdated or counterproductive elements because they’re in the regs. Federal/state/local overlaps also affect this. What if anything is different in Europe?

Daphne Keller: There’s a real difference b/t a culture that has real trust in regulators to do things with vague language and one that doesn’t. Europeans don’t think the regulators will be unreasonable. [Tort-based culture can’t be the full explanation because Pahlka documents these problems in places like military contracting where there’s no tort potential.] Data access is an example of openness where researchers are hoping that the process will look very different from what appears to be described in the legislation. #1 ask of civil society: keep us involved in the process; no particular ask other than role in interpretation. Chinese legal culture: The action isn’t in legislation; the action is in administration afterwards. The DSA is more like that. [Is China a culture with real trust in regulators?]

Quintais: DSA gives regulators more power than GDPR did.

Joris van Hoboken: Intermediary liability part of DSA is well-known and there’s a lot of consistency with prior practice—DSA adds a few elements. Separation between liability provisions and all the other stuff. Senftleben could be right about effects on liability of duties, but from a regulatory perspective there’s a separation.

Eric Goldman: what’s our definition of what would qualify as success or failure of the DSA? Might be broken down by topic. Who’s going to hold the powers that be accountable for whether or not this accomplishes the goals that they claimed it would/we think it should?

Bernt Hugenholtz: complexity reflects European tradition of civil law in the books, and other reasons, many already pointed out—desire to deal w/urgent tech developments, etc. But also, this is a regulation—an act—a directly binding instrument which is very different from what we grew up with (directives, which are instructions to member states to do certain things in harmonized fashion). Those could be vaguer/more concise b/c more was left to member states to implement/fill in gaps/national law could continue in the gaps. Regulation requires more specificity b/c you’re the regulator for the entire union. [In the US, we’d talk about preemption here.] DSA might be too early for AI, but on the other hand regulation is always too early in this field, since we never know where the field is going.

Matthias Leistner: Don’t generalize too much—some parts of DSA might be too early; revisions to Ecommerce directive might be too late. The more open the standards are, the less you need to worry about being too early.

US/EU differences: There are always internal European and Union-Member States issues. It is remarkably different in different areas at the European level. GDPR leaves it to member states to specify; DMA/DSA are comprehensive but tend to further centralize, which might have certain advantages (efficiency) and disadvantages (flexibility/power concentrated in Brussels). The data act follows a different approach: seems to provide umbrella regulation w/further specification possible, e.g. open health regime. EU-Member state issues: need to ask new questions—is a regulation v a directive conclusive? Major question is private enforcement. Damages claims of individual users provision in DSA is the first time parliament has considered this kind of enforcement, but creates new questions about tort law and its enforcement. Not clear whether member states could add criminal liability, for example.

Leaving “hot potato” issues to self-regulation as a popular way to avoid them? But that doesn’t work for everything—it is naïve to think that Zalando hasn’t been talking to the Commission for months, but the result is litigation. DSA’s self-audit requirements: can this work? Just delegating it to a further level.

Christophe Geiger: different EU/US traditions around regulatory oversight/verification of compliance with key values v private ordering. DSA is obviously overdoing it, but fundamental approach is very interesting—compare IP: don’t leave rightsholders & platforms alone, someone else has to step in. 27 different coordinators/regimes/traditions of intervention is problematic but the principle of regulatory oversight is at least interesting.

Eleanora Rosati: It is a success to have gotten to the point of adopting direct regulation—DSM Directive has resulted in extremely fragmented transposition and it’s not clear whether they can go as far as they have, including providing new rights not provided for in DSM. The relation b/t different provisions of DSM—link tax, data mining, “value gap,” etc—is unclear. So we have at least skipped that frustration created by directives of this type, including Ecommerce Directive. Vagueness of the safe harbors in that directive is no indicator of success. Didn’t create a level playing field.

Senftleben: What would DSA success look like?: a great question. We used to navigate cities ourselves, didn’t leave it to the machine. We knew where the different parts of the city were. Now we don’t. We’re lazy. We don’t do our own content filtering. The DSA puts burdens on citizens to be active. Transparency information: you can see how things function; the burden is on the citizen to do something (including creating an NGO or using the redress mechanisms). Fear: huge failure b/c we are just too lazy to use that. We’ll just behave like we do with Google Maps and follow what the automated systems tell us. Success=people are empowered in a real way and track how information flows/reaches them.

Pam Samuelson: How do I teach people to follow this? It is too complicated/requires a kind of perspective about regulation as a good that isn’t part of the current US regulatory culture. One thing US might see in a Brussels effect: larger platforms adapt so US doesn’t have to do anything legislatively, whether different or the same. But is that really a good thing? Do you care about barriers to entry? If you do, then maybe this isn’t the optimal strategy. [Goldman: that’s one question for the success/failure metrics.] I’d tell a startup to come to the US and only think of entering the European market after a certain amount of success.

Elkin-Koren: Common market creation is a measure of success, and digital platforms are likely to comply. What are the other consequences? Intended or not? Not intended to lower competition; tries to look at mice and elephants differently, but may turn out the opposite.

Giancarlo Frosio: Fundamental rights as a missing concept! That’s the great achievement of the DSA, meant to be interpreted and applied w/reference to fundamental rights recognized by Charter, to achieve a fair balance of conflicting fundamental rights.

Sebastian Schwemer: Goal is to regulate recommendation on the internet generally, focusing on process and not content. Would advise startups to start in Europe b/c as long as you’re in compliance everything is simple [I think I misunderstood this]. We focus too much on VLOPs.  DSA is overloaded, to be sure, but there’s important stuff. VLOPs provisions are different—competences aren’t clear where Commission has so much power not just over process but over content, setting rules and enforcing them both. 2021: Denmark proposed a social media law doing DSA+; it was shut down, but now they’re trying to introduce age verification on all platforms. DSA allows more leeway, again by focusing on process. Problem is two-tier system with regular platforms and VLOPs, which are understudied.

Justin Hughes: Some topics here are premature, others old hat (trusted flaggers). Opposite of the precautionary principle: AI Act seems like precautionary principle issue. Take on the known unknowns at least. 1201 looked like a failure, now it looks like a success. [????] Maybe the self-audits will fail, and they’ll never be eliminated b/c people are afraid to eliminated them. Maybe a few national authorities will dominate rather than 27. In the US: Maybe [big platforms] will take their lobbying energy elsewhere and not oppose new US regulations, but maybe they’ll continue to oppose them on principle.

Leistner: might compare amount of litigation under different regulations, directives. DMA: goal is to have more traders on platforms/increase diversity—could just check whether this happens five years from now. That would be clear-cut. Much more difficult w/r/t DSA b/c we don’t know what the DSA actually wants; there are a number of theories (nerd harder?). Could look at share of platform resources devoted to content moderation over time, strength of European democracy, whether there is European brain drain/smart Indian innovators come to California or to Munich, or other things. What kind of data would we need to have some plausible natural experiment?

Giancarlo Frosio: Again, protection of fundamental rights is a key measure—tools to force platforms to set up algorithms so they don’t limit fundamental rights. [This is why I don’t get why Wikipedia and the Internet Archive are even covered, since they don’t have the algorithmic problems at which the DSA is supposedly aimed.]

Keller: one key issue is that VLOPs have the ability to shape the rules for those who come up after them—codes of conduct negotiated by older companies that work for them but may not work for new entrants. This comes up with questions of what risk mitigation looks like.

Geiger: fundamental rights is a main pillar of DSA. But here I think we need to push for academic community’s permission to do homework/coordinate efforts to ensure this is actually happening. Most regulatory authorities are political appointments w/no specialization in IP or fundamental rights. Some will be strongly captured by IP claimants.

No comments: