Congratulations!

[Valid Atom 1.0] This is a valid Atom 1.0 feed.

Recommendations

This feed is valid, but interoperability with the widest range of feed readers could be improved by implementing the following recommendations.

Source: http://www.mnot.net/blog/index.atom

  1. <?xml version="1.0" encoding="utf-8"?>
  2. <feed xmlns="http://www.w3.org/2005/Atom">
  3.  <title>mark nottingham</title>
  4.  <link rel="alternate" type="text/html" href="https://www.mnot.net/blog/" />
  5.  <link rel="self" type="application/atom+xml" href="https://www.mnot.net/blog/index.atom" />
  6.  <id>tag:www.mnot.net,2010-11-11:/blog//1</id>
  7.  <updated>2024-04-22T06:36:08Z</updated>
  8.  <subtitle></subtitle>
  9.  
  10.  <entry>
  11.    <title>Considerations for AI Opt-Out</title>
  12.    <link rel="alternate" type="text/html" href="https://www.mnot.net/blog/2024/04/21/ai-control" />
  13.    <id>https://www.mnot.net/blog/2024/04/21/ai-control</id>
  14.    <updated>2024-04-21T00:00:00Z</updated>
  15.    <author>
  16.        <name>Mark Nottingham</name>
  17.        <uri>https://www.mnot.net/personal/</uri>
  18.    </author>
  19.    <summary>Creating a Large Language Model (LLM) requires a lot of content – as implied by the name, LLMs need voluminous input data to be able to function well. Much of that content comes from the Internet, and early models have been seeded by crawling the whole Web.</summary>
  20.    
  21. <category term="Tech Regulation" />
  22.    
  23.    <content type="html" xml:lang="en" xml:base="https://www.mnot.net/blog/2024/04/21/ai-control">
  24.    <![CDATA[<p class="intro">Creating a Large Language Model (LLM) requires a <em>lot</em> of content – as implied by the name, LLMs need <a href="https://www.nytimes.com/2024/04/06/technology/tech-giants-harvest-data-artificial-intelligence.html">voluminous input data to be able to function well</a>. Much of that content comes from the Internet, and early models have been seeded by crawling the whole Web.</p>
  25.  
  26. <p>This now widespread practice of ingestion without consent is contentious, to put it mildly. Content creators feel that they should be compensated or at least have a choice about how their content is used; AI advocates caution that without easy access to input data, their ability to innovate will be severely limited, thereby curtailing the promised benefits of AI.</p>
  27.  
  28. <h3 id="the-policy-context">The Policy Context</h3>
  29.  
  30. <p>In the US, the Copyright Office has launched <a href="https://copyright.gov/ai">an initiative</a> to examine this and other issues surrounding copyright and AI. So far, they have avoided addressing the ingestion issue, but nevertheless it has come up repeatedly in their <a href="https://copyright.gov/ai/listening-sessions.html">public proceedings</a>:</p>
  31. <blockquote>
  32.  <p>“The interests of those using copyrighted materials for AI ingestion purposes must not be prioritized over the rights and interests of creators and copyright owners.” – <em>Keith Kupferschmid, Copyright Alliance</em></p>
  33. </blockquote>
  34.  
  35. <blockquote>
  36.  <p>“Training of AI language models begins with copying, which we believe has infringed our copyrights and has already deprived us of hundreds of millions of dollars in rightful revenues.  The additional violation of our moral right of attribution makes it impossible to tell which of our works have been copied to train AI and thus frustrates redress for either the economic infringement or the violation of our moral right to object to use of our work to train AI to generate prejudicial content. […] OpenAI, for example, has received a billion dollars in venture capital, none of which has been passed on to the authors of the training corpus even though, without that training corpus, chatGPT would be worthless.” – <em>Edward Hasbrouck, National Writers Union</em></p>
  37. </blockquote>
  38.  
  39. <p>It’s uncertain when (or if) the Copyright Office will provide more clarity on this issue. Also relevant in the US are the outcomes of cases like <a href="https://www.courtlistener.com/docket/66788385/getty-images-us-inc-v-stability-ai-inc/">Getty Images (US), Inc. v. Stability AI, Inc.</a></p>
  40.  
  41. <p>However, Europe has been more definitive about the ingestion issue. <a href="https://eur-lex.europa.eu/eli/dir/2019/790/oj">Directive 2019/790</a> says:</p>
  42. <blockquote>
  43.  <p>The [exception for copyright] shall apply on condition that the use of works and other subject matter referred to in that paragraph has not been expressly reserved by their rightholders in an appropriate manner, such as machine-readable means in the case of content made publicly available online.<sup id="fnref:1" role="doc-noteref"><a href="#fn:1" class="footnote" rel="footnote">1</a></sup></p>
  44. </blockquote>
  45.  
  46. <p>This is reinforced by the <a href="https://www.europarl.europa.eu/news/en/press-room/20240308IPR19015/artificial-intelligence-act-meps-adopt-landmark-law">recently</a> adopted <a href="https://www.europarl.europa.eu/doceo/document/A-9-2023-0188-AM-808-808_EN.pdf">AI Act</a>:</p>
  47. <blockquote>
  48.  <p>Any use of copyright protected content requires the authorisation of the rightsholder concerned unless relevant copyright exceptions and limitations apply. Directive (EU) 2019/790 introduced exceptions and limitations allowing reproductions and extractions of works or other subject matter, for the purpose of text and data mining, under certain conditions. Under these rules, rightsholders may choose to reserve their rights over their works or other subject matter to prevent text and data mining, unless this is done for the purposes of scientific research. Where the rights to opt out has been expressly reserved in an appropriate manner, providers of general-purpose AI models need to obtain an authorisation from rightsholders if they want to carry out text and data mining over such works.</p>
  49. </blockquote>
  50.  
  51. <p>In other words, European law is about to require commercial AI crawlers to support an opt-out. However, it does not specify a particular mechanism: it only says that consent must be ‘expressly reserved in an appropriate manner.’</p>
  52.  
  53. <p>So, what might that opt-out signal look like?</p>
  54.  
  55. <h3 id="robotstxt-as-an-opt-out">Robots.txt as an Opt-Out</h3>
  56.  
  57. <p>Since most of the publicly available content on the Internet is accessed over the Web, it makes sense to consider how an opt-out might be expressed there as a primary mechanism. The Web already has a way for sites to opt-out of automated crawling: the <code class="language-plaintext highlighter-rouge">robots.txt</code> file, now specified by an <a href="https://www.rfc-editor.org/rfc/rfc9309.html">IETF Standards-Track RFC</a>.</p>
  58.  
  59. <p>At first glance, robots.txt intuitively maps to what’s required: a way to instruct automated crawlers on how to treat a site with some amount of granularity, including opting out of crawling altogether. Some LLMs have latched onto this it already; for example, OpenAI <a href="https://platform.openai.com/docs/gptbot">allows their crawler to be controlled by it</a>.</p>
  60.  
  61. <p>There are a lot of similarities between gathering Web content for search and gathering it for an LLM: the actual crawler software is very similar (if not identical), crawling the whole Web requires significant resources, and both uses create enormous potential value not only for the operators of the crawlers, but also for society.</p>
  62.  
  63. <p>However, it is questionable whether merely reusing to robots.txt as the opt-out mechanism is sufficient to allow rightsholders to fully express their reservation. Despite the similarities listed above, it is hard to ignore the ways that LLM ingest is different.</p>
  64.  
  65. <p>That’s because Web search can be seen as a service to sites; it makes them more discoverable on the Web, and is thus symbiotic – both parties benefit. LLM crawling, on the other hand, doesn’t have any benefits to the content owner, and may be perceived as harming them.</p>
  66.  
  67. <p>Through the lenses of those different purposes and their associated power dynamics, a few issues become apparent.</p>
  68.  
  69. <h3 id="1-usability-and-ecosystem-impact">1. Usability and Ecosystem Impact</h3>
  70.  
  71. <p>Robots.txt allows sites to target directives to bots in two different ways: by path on the site (e.g., <code class="language-plaintext highlighter-rouge">/images</code> vs. <code class="language-plaintext highlighter-rouge">/users</code>) and by User-Agent. The User-Agent identifies the bot, allowing sites to specify things like “I allow Google to crawl my site, but not Bing.” Or, “I don’t allow any bots.”</p>
  72.  
  73. <p>That might be adequate for controlling how your site appears in search engines, but problematic when applied to AI. Let’s look at an example.</p>
  74.  
  75. <p>To stop OpenAI from crawling your site, you can add:</p>
  76.  
  77. <div class="language-plaintext highlighter-rouge"><div class="highlight"><pre class="highlight"><code>User-Agent: GPTBot
  78. Disallow: /
  79. </code></pre></div></div>
  80.  
  81. <p>However, that directive doesn’t apply to Google, Mistral, or any other LLM-in-waiting out there; you’d have to target each individual one (and some folks are <a href="https://www.20i.com/blog/how-to-prevent-ai-from-scraping-your-website/">already advising on how to do that</a>).</p>
  82.  
  83. <p>If you miss one, that’s your fault, and it’ll be in that model forever, so careful (or just frustrated) people might decide to just ban everything:</p>
  84.  
  85. <div class="language-plaintext highlighter-rouge"><div class="highlight"><pre class="highlight"><code>User-Agent: *
  86. Disallow /
  87. </code></pre></div></div>
  88.  
  89. <p>But that has the downside of disallowing AI <em>and</em> search crawlers – even though presence in search engines is often critical to sites. To avoid that, you would have to enumerate all of the search engines and other bots that you want to allow, creating more work.</p>
  90.  
  91. <p>Significantly, doing so could also have a negative effect on the Web ecosystem: if sites have a stronger incentive to disallow unknown bots thanks to AI, it would be much harder to responsibly introduce new crawler-based services to the Web. That would tilt the table even further in the favour of already established ‘big tech’ actors.</p>
  92.  
  93. <p>There are two easy ways to fix these issues. One would be to define a special User-Agent that applies to <em>all</em> AI crawlers. For example:</p>
  94.  
  95. <div class="language-plaintext highlighter-rouge"><div class="highlight"><pre class="highlight"><code>User-Agent: AI-Ingest
  96. Disallow: /
  97. </code></pre></div></div>
  98.  
  99. <p>The other approach would be to create a new <a href="https://www.rfc-editor.org/rfc/rfc8615.html">well-known location</a> just for AI – for example <code class="language-plaintext highlighter-rouge">/.well-known/ai.txt</code>. That file might have the same syntax as <code class="language-plaintext highlighter-rouge">robots.txt</code>, or its notoriously quirky syntax could be ditched for something more modern.</p>
  100.  
  101. <p>Either solution above would make it easy for a site to opt-out of AI crawling of any sort without enumerating all of the potential AI crawlers in the world, and without impacting their search engine coverage or creating ecosystem risk.</p>
  102.  
  103. <p>I suspect that many have been assuming that one of these things will happen; they’re fairly obvious evolutions of existing practice. However, at least two more issues are still unaddressed.</p>
  104.  
  105. <h3 id="2-previously-crawled-content">2. Previously Crawled Content</h3>
  106.  
  107. <p>Web search and LLMs also differ in how they relate to time.</p>
  108.  
  109. <p>A search engine crawler has a strong interest in assuring that its index reflects the <em>current</em> Web. LLM crawlers, on the other hand, are ravenous without regard to its age or current availability on the Web. Once ingested content forms part of a model, they add value to that model for the lifetime of its use – and the model often persists for months or even years after the ingested content was obtained. Furthermore, that content might be reused to create future models, indefinitely.</p>
  110.  
  111. <p>That means that a content owner who isn’t aware of the LLM crawler <em>at crawl time</em> doesn’t have any recourse. From the Copyright Office sessions:</p>
  112.  
  113. <blockquote>
  114.  <p>We believe that writers should be compensated also for past training since it appears that the massive training that has already occurred for GPT and Bard to teach the engines to think and to write has already occurred[.] – <em>Mary Rasenberger, The Authors Guild</em></p>
  115. </blockquote>
  116.  
  117. <p>This shortcoming could be addressed by a relatively simple measure: stating that the policy for a given URL applies to any use of content obtained from that URL at model creation time, <em>regardless of when it was obtained</em>.</p>
  118.  
  119. <p>A significant amount of detail would need to be specified to make this work, of course. It would also likely necessitate some sort of grandfathering or transition period for existing models.</p>
  120.  
  121. <p>Needless to say, the impact of this kind of change <em>could</em> be massive: if 90% of the sites in the world opt out in this fashion (a la <a href="https://www.theverge.com/2021/10/31/22756135/apple-app-tracking-transparency-policy-snapchat-facebook-twitter-youtube-lose-10-billion">App Tracking Transparency</a>), it would be difficult to legally construct a new model (or at least market or use such a model in Europe, under the forthcoming rules).</p>
  122.  
  123. <p>On the other hand, if that many people don’t want to allow LLMs to use their content when offered a genuine chance to control it, shouldn’t their rights be honoured? Ultimately, if that’s the outcome, society will need to go back to the drawing board and figure out what it values more: copyright interests or the development of LLMs.</p>
  124.  
  125. <h3 id="3-control-of-metadata">3. Control of Metadata</h3>
  126.  
  127. <p>Another issue with reusing robots.txt is how that file itself is controlled. As a site-wide metadata mechanism, there is only one controller for robots.txt: the site administrator.</p>
  128.  
  129. <p>That means that on Facebook, Meta will decide whether your photos can be used to feed AI (theirs or others’), not you. On GitHub, Microsoft will decide how your repositories will be treated. And so on.</p>
  130.  
  131. <p>While robots.txt is great for single-owner sites (like this one), it doesn’t meet the needs of a concentrated world – it leverages the power that accrues to a small number of platform owners to decide policy for all of their users.</p>
  132.  
  133. <p>Avoiding that outcome means that users need to be able express their preference in the content itself, so that it persists no matter where it ends up. That means it’s necessary to be able to embed policy in things like images, videos, audio files, document formats like PDF, Office, and ePub, containers like ZIP files, file system paths for things like git repos, and so on. Assuming that a robots.txt-like approach is also defined, their relative precedence will also need to be specified.</p>
  134.  
  135. <p>Luckily, this is not a new requirement – our industry has considerable experience in embedding such metadata into file formats, for use cases like <a href="https://c2pa.org/specifications/specifications/2.0/specs/C2PA_Specification.html">content provenance</a>. It just needs to be specified for AI control.</p>
  136.  
  137. <h3 id="whats-next">What’s Next?</h3>
  138.  
  139. <p>Policy decisions like that just made by Europe might be the drivers of change in LLM ingest practices, but I hope I’ve shown that the technical details of that ‘appropriate manner’ of opting out can significantly steer power between AI companies and content owners.</p>
  140.  
  141. <p>Notably, while the worldwide copyright regime is explicitly opt-in (i.e., you have to explicitly offer a license for someone to legally use your material, unless fair use applies), the European legislation changes this to opt-out for AI.<sup id="fnref:2" role="doc-noteref"><a href="#fn:2" class="footnote" rel="footnote">2</a></sup> Given that, offering content owners a genuine opportunity to do so is important, in my opinion.</p>
  142.  
  143. <p>I’ve touched on a few aspects that influence that opportunity above; I’m sure there are more.<sup id="fnref:3" role="doc-noteref"><a href="#fn:3" class="footnote" rel="footnote">3</a></sup> As I implied at the start, getting the balance right is going to take careful consideration and perhaps most importantly, sunlight.</p>
  144.  
  145. <p>However, It’s not yet clear where or how this work will happen. Notably, the <a href="https://ec.europa.eu/transparency/documents-register/detail?ref=C(2023)3215&amp;lang=en">standardisation request to the European Standardisation Organisations in support of safe and trustworthy artificial intelligence</a> does not mention copyright at all. Personally, I think that’s a good thing – worldwide standards need to be in open international standards bodies like the IETF, not regionally fragmented.</p>
  146.  
  147. <p>In that spirit, the IETF has recently created a <a href="https://www.ietf.org/mailman/listinfo/ai-control">mailing list to discuss AI control</a>. That’s likely the best place to follow up if you’re interested in discussing these topics.</p>
  148.  
  149. <div class="footnotes" role="doc-endnotes">
  150.  <ol>
  151.    <li id="fn:1" role="doc-endnote">
  152.      <p>See also <a href="https://eur-lex.europa.eu/eli/dir/2019/790/oj#rct_18">Recital 18</a>. <a href="#fnref:1" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
  153.    </li>
  154.    <li id="fn:2" role="doc-endnote">
  155.      <p>And I suspect other jurisdictions might follow the same approach; time will tell. <a href="#fnref:2" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
  156.    </li>
  157.    <li id="fn:3" role="doc-endnote">
  158.      <p>For example, some of the input to the Copyright Office mentioned group licensing regimes. An opt-out mechanism could be adapted to support that. <a href="#fnref:3" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
  159.    </li>
  160.  </ol>
  161. </div>]]>
  162.    </content>
  163.  </entry>
  164.  
  165.  <entry>
  166.    <title>There Are No Standards Police</title>
  167.    <link rel="alternate" type="text/html" href="https://www.mnot.net/blog/2024/03/13/voluntary" />
  168.    <id>https://www.mnot.net/blog/2024/03/13/voluntary</id>
  169.    <updated>2024-03-13T00:00:00Z</updated>
  170.    <author>
  171.        <name>Mark Nottingham</name>
  172.        <uri>https://www.mnot.net/personal/</uri>
  173.    </author>
  174.    <summary>It happens fairly often. Someone brings a proposal to a technical standards body like the IETF and expects that just because it becomes an RFC, people will adopt it. Or they’ll come across a requirement in an RFC and expect it to be enforced, perhaps with some kind of punishment. Or they’ll get angry that people don’t pay attention to an existing standard and do their own thing. This is so common that there’s a ready response widely used by IETF people in these situations:</summary>
  175.    
  176. <category term="Standards" />
  177.    
  178. <category term="Tech Regulation" />
  179.    
  180.    <content type="html" xml:lang="en" xml:base="https://www.mnot.net/blog/2024/03/13/voluntary">
  181.    <![CDATA[<p class="intro">It happens fairly often. Someone brings a proposal to a technical standards body like the IETF and expects that just because it becomes an RFC, people will adopt it. Or they’ll come across a requirement in an RFC and expect it to be enforced, perhaps with some kind of punishment. Or they’ll get angry that people don’t pay attention to an existing standard and do their own thing. This is so common that there’s a ready response widely used by IETF people in these situations:</p>
  182.  
  183. <p class="intro">“There are no standards police.”</p>
  184.  
  185. <p>In other words, even if you do consider Internet standards to be <a href="https://www.mnot.net/blog/2023/11/01/regulators">a regulatory force</a>, there is no <em>enforcement mechanism</em>. One of their key characteristics is that they’re <strong>voluntary</strong>. No one forces you to adopt them. No one can penalise you for violating a MUST; you have to want to conform.</p>
  186.  
  187. <p>Of course, you can still <em>feel</em> compelled to do so. If an interoperability standard gets broad adoption and everyone you want to communicate with expects you to honour it, you don’t have many options. For example, if you want to have a Web site, you need to interoperate with browsers; most of the time, they write down what they do in standards documents, and so you’ll need to conform to them.</p>
  188.  
  189. <p>But that’s the successful path. For every HTTP or HTML or TCP, there are hundreds of IETF RFCs, W3C Recommendations, and other standards documents that haven’t caught on – presumably much to their authors’ dismay. Adopting and using those documents was optional, and the market spoke: there wasn’t interest.</p>
  190.  
  191. <p class="hero">This aspect of the Internet’s standards has been critical to its success. If people were forced to adopt a specification just because some body had blessed it, it would place immense pressure on whatever process was used to create it. The stakes would be high because the future of the Internet would be on the line: businesses would play dirty; trolls would try to subvert the outcomes; governments would try to steer the results.</p>
  192.  
  193. <p>Of course, all of those things already happen in Internet standards; it’s just that the stakes are much lower.</p>
  194.  
  195. <p>So, voluntary adoption is a <em>proving function</em> – it means that not all of the weight of getting things right is on the standardisation process, and that process can be lighter than, for example, that used by the governments or the United Nations (I’ll get back to that in a minute). That’s important, because it turns out that it’s already incredibly difficult to create useful, successful, secure, private, performant, scalable, architecturally aligned technical specifications that change how the Internet works within all of the other natural constraints encountered; it’s threading-the-needle kind of stuff. And we need to be able to fail.</p>
  196.  
  197. <p class="hero">Historically, voluntary standards have been encouraged by governments in their purchasing and competition policies - for example,  <a href="https://www.federalregister.gov/documents/2016/01/27/2016-01606/revision-of-omb-circular-no-a-119-federal-participation-in-the-development-and-use-of-voluntary">OMB Circular A-119</a>, <a href="https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX:32012R1025">EU Regulation 1025/2012</a>, and the <a href="https://competition-policy.ec.europa.eu/system/files/2023-07/2023_revised_horizontal_guidelines_en.pdf">EC guidelines on horizontal agreements</a>. Standards bodies are a ‘safe space’ where competitors can cooperate without risking competition enforcement, so long as they follow a set of rules – and one of the biggest rules is that adoption should be voluntary, not mandatory or coerced (at least by those setting the standard).</p>
  198.  
  199. <p>But it’s no secret that the policy landscape for the Internet has changed drastically. Now, there is increasing interest in using interoperability standards as a mechanism to steer the Internet. Academics are <a href="https://hackcur.io/whats-wrong-with-loud-men-talking-loudly-the-ietfs-culture-wars/">diving deep into the cultures and mechanisms of technical standards</a>. Civil society folks are coming to technical standards bodies and <a href="https://datatracker.ietf.org/rg/hrpc/about/">trying to figure out how to incorporate human rights goals</a>.  <a href="https://themarket.ch/interview/tech-is-becoming-a-regulated-industry-ld.7520">Regulation is coming</a>, and policy experts are <a href="https://datatracker.ietf.org/doc/draft-hoffmann-gendispatch-policy-stakeholders/">trying to figure out how to get involved too</a>.</p>
  200.  
  201. <p>This influx has caused concern that that these relative newcomers are mistakenly focusing on standards as a locus of power when, in fact, the power is expressed in the <em>adoption</em> of a standardised technology. For example, Geoff Huston recently wrote an <a href="https://blog.apnic.net/2024/03/06/opinion-digital-sovereignty-and-standards/">opinion piece</a> along these lines.</p>
  202.  
  203. <p>I have no doubt that some still come to the IETF and similar bodies with such misapprehensions; we still have to remind people that ‘there are no standards police’ on a regular basis. However, I suspect that at least the policy people (including regulators) largely understand that it’s not that simple.</p>
  204.  
  205. <p>That’s because modern regulators are very aware that there are many influences on a regulatory space. They want to learn about the other forces acting on their target, as well as <a href="http://johnbraithwaite.com/responsive-regulation/">persuade and inform</a>. Similarly, those who are involved in policymaking are intensely aware of the diffuse nature of power. In short, their world view is more sophisticated than people give them credit for.</p>
  206.  
  207. <p>(All that said, I’m still interested and a bit nervous to see what <a href="https://www.un.org/techenvoy/global-digital-compact">Global Digital Compact</a> contains when it becomes public.)</p>
  208.  
  209. <p class="hero">Another concern is that governments might try to influence Internet standards to suit their purposes, and then exert pressure to make the results mandatory – short circuiting the proving function of voluntary standards.</p>
  210.  
  211. <p>Avoiding that requires separating the legal requirement from the standards effort, to give the latter a chance to fail. For example, <a href="https://datatracker.ietf.org/group/mimi/about/">MIMI</a> may or may not succeed in satisfying the DMA requirement for messaging interop. It is an attempt to establish voluntary standards that, if successful in the market, could satisfy legal regulatory requirements without using a preselecting standards venue.</p>
  212.  
  213. <p>Of course, that pattern is not new – for example, <a href="https://www.w3.org/WAI/">accessibility work in the W3C</a> is the basis of many regulatory requirements now, but wasn’t considered (AFAIK) by regulators until many years after its establishment.</p>
  214.  
  215. <p>Because of the newly intense focus on regulating technology, there’s likely to be increasing pressure on such efforts: both the pace and volume of standardisation will need to increase to meet the requirements that the standards bodies want to attempt to address. I suspect aligning the timelines and risk appetites of standards bodies and regulators are going to be some of the biggest challenges we’ll face if we want more successes.</p>
  216.  
  217. <p>So right now I believe the best way forward is to create ‘rails’ for interactions with legal regulators – e.g., improved communication, aligned expectations, and ways for an effort to be declined or to fail without disastrous consequences. Doing that will require some capacity building on the parts of standards bodies, but no fundamental changes to their models or decision-making processes.</p>
  218.  
  219. <p>This approach will not address everything. There are some areas where at least some regulators and the Internet standards community are unlikely to agree. Standards-based interoperability may not be realistically achievable in some instances, because of how entrenched a proprietary solution is. Decentralising a proprietary solution can face <a href="https://www.rfc-editor.org/rfc/rfc9518.html">many pitfalls</a>, and may be completely at odds with a centralized solution that already has broad adoption. And, most fundamentally, parties that are not inclined to cooperate can easily subvert a voluntary consensus process.</p>
  220.  
  221. <p>However, if things are arranged so that when conforming to a voluntary consensus standard that has seen wide review and market adoption is considered to be <em>prima facie</em> evidence of conformance to a regulatory requirement, perhaps we <em>do</em> sometimes have standards police, in the sense that legal requirements can be used to help kickstart standards-based interoperability where it otherwise wouldn’t get a chance to form.</p>]]>
  222.    </content>
  223.  </entry>
  224.  
  225.  <entry>
  226.    <title>RFC 9518 - What Can Internet Standards Do About Centralisation?</title>
  227.    <link rel="alternate" type="text/html" href="https://www.mnot.net/blog/2023/12/19/standards-and-centralization" />
  228.    <id>https://www.mnot.net/blog/2023/12/19/standards-and-centralization</id>
  229.    <updated>2023-12-19T00:00:00Z</updated>
  230.    <author>
  231.        <name>Mark Nottingham</name>
  232.        <uri>https://www.mnot.net/personal/</uri>
  233.    </author>
  234.    <summary>RFC 9518: Centralization, Decentralization, and Internet Standards has been published, after more than two years of review, discussion, and revision.</summary>
  235.    
  236. <category term="Internet and Web" />
  237.    
  238. <category term="Standards" />
  239.    
  240. <category term="Tech Regulation" />
  241.    
  242.    <content type="html" xml:lang="en" xml:base="https://www.mnot.net/blog/2023/12/19/standards-and-centralization">
  243.    <![CDATA[<p><a href="https://www.rfc-editor.org/rfc/rfc9518.html">RFC 9518: Centralization, Decentralization, and Internet Standards</a> has been published, after more than two years of review, discussion, and revision.</p>
  244.  
  245. <p>It’s no secret that most people have been increasingly concerned about Internet centralization over the last decade or so. Having one party (or a small number of them) with a choke hold over any important part of the Internet is counter to its nature: as a ‘network of networks’, the Internet is about fostering relationships between <em>peers</em>, not allowing power to accrue to a few.</p>
  246.  
  247. <p>As I’ve <a href="/blog/2023/11/01/regulators">discussed previously</a>, Internet standards bodies (like the <a href="https://www.ietf.org/">IETF</a> and <a href="https://www.w3.org/">W3C</a>) can be seen as a kind of regulator, in that they constrain the behaviour of others. So it’s natural to wonder whether they can help avoid or mitigate Internet centralization.</p>
  248.  
  249. <p>I started drafting a document that explored these issues when I was a member of the <a href="https://iab.org/">Internet Architecture Board</a>. That eventually became <a href="https://datatracker.ietf.org/doc/draft-nottingham-avoiding-internet-centralization/">draft-nottingham-avoiding-internet-centralization</a>, which became an Independent Stream RFC today.</p>
  250.  
  251. <p>But it was a long journey. I started this work optimistic that standards could make a difference, in part because Internet standards bodies are (among many things) communities of people who are deeply invested in the success of the Internet, with a set of shared <a href="https://www.rfc-editor.org/rfc/rfc8890.html">end user-focused</a> values.</p>
  252.  
  253. <p>That optimism was quickly tempered. After digging into the mechanisms that we have available, the way that the markets work, and the incentives on the various actors, it became apparent that it was unrealistic to expect that standards documents – which of course don’t have any intrinsic power or authority if no one implements them – are up to the task of controlling centralization.</p>
  254.  
  255. <p>Furthermore, centralization is inherently difficult to eradicate: while you can reduce or remove some forms of it, it has a habit of popping up elsewhere.</p>
  256.  
  257. <p>That doesn’t mean that standards bodies should ignore centralization, or that there isn’t anything they can do to improve the state of the world regarding it (the RFC explores several); rather, that we should not expect standards to be sufficient to effectively address it on their own.</p>
  258.  
  259. <p>You can read <a href="https://www.rfc-editor.org/rfc/rfc9518.html">the RFC</a> for the full details. It covers what centralization is, how it can be both beneficial and harmful, the decentralization strategies we typically use to control it, and finally what Internet standards bodies can do in relation to it.</p>
  260.  
  261. <p>One final note: I’d be much less satisfied with the result if I hadn’t had the excellent reviews that Eliot Lear (the Independent Submissions Editor) sourced from <a href="https://www.apnic.net/about-apnic/team/geoff-huston/">Geoff Huston</a> and <a href="https://www.internetgovernance.org/people/milton-mueller/">Milton Mueller</a>. Many thanks to them and everyone else who contributed.</p>]]>
  262.    </content>
  263.  </entry>
  264.  
  265.  <entry>
  266.    <title>How to Run an Australian Web Site in 2024</title>
  267.    <link rel="alternate" type="text/html" href="https://www.mnot.net/blog/2023/11/27/esafety-industry-standards" />
  268.    <id>https://www.mnot.net/blog/2023/11/27/esafety-industry-standards</id>
  269.    <updated>2023-11-27T00:00:00Z</updated>
  270.    <author>
  271.        <name>Mark Nottingham</name>
  272.        <uri>https://www.mnot.net/personal/</uri>
  273.    </author>
  274.    <summary>A while back, the eSafety Commissioner declined to register the proposed Industry Codes that I’ve previously written about. Now, they’ve announced a set of Industry Standards that, after a comment period, will likely be law.</summary>
  275.    
  276. <category term="Australia" />
  277.    
  278.    <content type="html" xml:lang="en" xml:base="https://www.mnot.net/blog/2023/11/27/esafety-industry-standards">
  279.    <![CDATA[<p class="intro">A while back, the eSafety Commissioner declined to register the proposed Industry Codes that <a href="https://www.mnot.net/blog/2022/09/11/esafety-industry-codes">I’ve previously written about</a>. Now, they’ve announced a set of <a href="https://www.esafety.gov.au/industry/codes/standards-consultation">Industry Standards</a> that, after a comment period, will likely be law.</p>
  280.  
  281. <p>If you run an online service that’s accessible to Australians, these Standards will apply to you. Of course, if you don’t live here, don’t do business here, and don’t want to come here, you can <em>probably</em> ignore them.</p>
  282.  
  283. <p>Assuming you do fall into one of those buckets, this post tries to walk through the implications, as a list of questions you’ll need to ask yourself.</p>
  284.  
  285. <p>I’m going to try to focus on the practical implications, rather than “showing my work” by deep-diving into the text of the standards and <a href="https://www.legislation.gov.au/Details/C2022C00052">supporting legislation</a>. This is based only upon my reading of the documents and a miniscule dollop of legal education; if there are things that I get wrong, corrections and suggestions are gladly taken. Note that this is not legal advice, and the Standards might change before they’re registered.</p>
  286.  
  287. <h3 id="does-the-standard-apply-to-your-service">Does the Standard Apply to Your Service?</h3>
  288. <p>The first question to answer is whether your service is covered by the <a href="https://www.esafety.gov.au/sites/default/files/2023-11/Draft%20Online%20Safety%20%28Designated%20Internet%20Services-Class%201A%20and%20Class%201B%20Material%29%20Industry%20Standard%202024.pdf">Online Safety (Designated Internet Services – Class 1A and Class 1B Material) Industry Standards 2024</a>.</p>
  289.  
  290. <p>The short answer is “yes, even <em>that</em> one.”</p>
  291.  
  292. <p>A Designated Internet Service (DIS) is one that allows “end-users to access material using an Internet carriage service.” This is a very broad definition that explicitly applies to Web sites. For simplicity, the remainder of this article will assume your service is a Web site, even though other information services can be a DIS.</p>
  293.  
  294. <p>In a nutshell, if “none of the material on the service is accessible to, or delivered to, one or more end-users in Australia”, your site is exempt. Otherwise, it’s covered (unless one of the other Codes or Standards takes precedence; see below).</p>
  295.  
  296. <p>So whether you’re Elon Musk or you have a personal Web site with no traffic, this standard applies to you, so long as it’s available to one Australian person – even if none actually visit. Don’t be fooled by “Industry” in the title. That default page that your Web server comes up with when your new Linux box boots for the first time? Covered. Note that it doesn’t even need to be on the <em>public</em> Internet; things like corporate Intranet sites are covered, as are content-free static sites like those used to park domains.</p>
  297.  
  298. <p>Given how broadly the legislation and standard are written, combined with how prevalent HTTP and similar protocols are on today’s Internet, it’s also reasonable to say that APIs are covered; there’s no inherent restrictions on formats or protocols in the eSafety standards – in fact, the definition of <em>material</em> in the Act includes “data”.</p>
  299.  
  300. <p>So, to be safe, <em>any</em> server available on the Internet is covered by the eSafety scheme, so long as it <em>can</em> be accessed by Australians.</p>
  301.  
  302. <h3 id="do-you-need-a-risk-assessment">Do You Need a Risk Assessment?</h3>
  303. <p>Assuming that your site is covered by the Standard, your next step is to figure out whether you need to perform a risk assessment.</p>
  304.  
  305. <p>Assuming that you’re not running a large commercial web site, a (ahem) “high impact” service (i.e., one that specialises in porn, violent content, and similar), or an AI-flavoured service, there are two interesting categorise that might get you out of performing a risk assessment.</p>
  306.  
  307. <p>The first is a “pre-assessed general purpose DIS.” You can qualify for this if you don’t allow users in Australia to post any material (including comments), or if posting is “to review or provide information on products, services, or physical points of interest or locations made available on the service.” It’s also OK if they are “sharing […] with other end-users for a business, informational, or government service or support purpose.”<sup id="fnref:1" role="doc-noteref"><a href="#fn:1" class="footnote" rel="footnote">1</a></sup></p>
  308.  
  309. <p>Does it seem like your site qualifies? Not so fast; that only covers “pre-assessment.” A <em>general purpose DIS</em> is a</p>
  310.  
  311. <blockquote>
  312.  <p>website or application that […] primarily provides information for business, commerce, charitable, professional, health, reporting news, scientific, educational, academic research, health, reporting news, scientific, educational, academic research, government, public service, emergency, or counselling and support service purposes.</p>
  313. </blockquote>
  314.  
  315. <p>Unless your site falls cleanly into one of those categories, you don’t have a general purpose DIS.<sup id="fnref:2" role="doc-noteref"><a href="#fn:2" class="footnote" rel="footnote">2</a></sup></p>
  316.  
  317. <p>The second is an “enterprise DIS.” This is a site where “the account holder […] is an organisation (and not an individual).” Basically, if your users are companies or other organisations and not individual people, you don’t have to do an assessment.</p>
  318.  
  319. <h3 id="what-does-your-risk-assessment-contain">What Does Your Risk Assessment Contain?</h3>
  320. <p>Assuming you need a risk assessment (spoiler: you probably do, to be safe), you</p>
  321.  
  322. <blockquote>
  323.  <p> must formulate in writing a plan, and a methodology, for carrying out the assessment that ensure that the risks mentioned in subsection 8(1) in relation to the service are accurately assessed.</p>
  324. </blockquote>
  325.  
  326. <p>The risk referred to is that class 1A or class 1B material will be “generated or accessed by, or distributed by or to, end-users in Australia using the service.” Storage of such material is also included (even if it isn’t accessed).</p>
  327.  
  328. <p>To answer your next question, class 1A material is “child sexual exploitation material”, “pro-terror material”, or “extreme crime and violence material.” class 1B material is “crime and violence material” and “drug-related material.” There are long definitions of each of these kinds of material in the standard; I won’t repeat them here.</p>
  329.  
  330. <p>Your risk assessment must “undertake a forward-looking analysis” of what’s likely to change both inside and outside of your service, along with the impact of those changes. It’s also required to “specify the principle matters to be taken into account”, including eleven factors such as “the ages of end-users and likely end-users”, “safety by design guidance”, AI risks, terms of use, and so forth.</p>
  331.  
  332. <p>Your risk assessment has to be written down in detail. You must also “ensure that [it] is carried out by persons with the relevant skills, experience, and expertise” – although it’s not yet clear what that means in practice or how it will be enforced.<sup id="fnref:3" role="doc-noteref"><a href="#fn:3" class="footnote" rel="footnote">3</a></sup></p>
  333.  
  334. <h3 id="whats-your-risk-profile">What’s Your Risk Profile?</h3>
  335. <p>Once you’ve done a risk assessment, you’ll have a risk profile – one of Tier 1, Tier 2, or Tier 3.</p>
  336.  
  337. <p>Let’s assume your site has no user-generated content, and you only upload very… normal… content– like this site.<sup id="fnref:4" role="doc-noteref"><a href="#fn:4" class="footnote" rel="footnote">4</a></sup> You’re likely to be Tier 3.</p>
  338.  
  339. <p>If so, congratulations! Your work is just about done. Sections 34, 40, and 41 of the Standard apply to you – basically, the eSafety Commissioner can demand that you provide them with your risk assessment and how you arrived at it. You also have to investigate complaints, and keep records.</p>
  340.  
  341. <p>If you’re not Tier 3 – for example, you blog about drugs or crime, or you allow user uploads or comments, there are a whole slew of requirements you’ll need to conform to, which are well out of scope for this blog entry (since I’m mostly interested in the impact of regulation on small, non-commercial sites). Tip: get some professional help, quickly.</p>
  342.  
  343. <h3 id="what-other-standards-will-apply">What Other Standards Will Apply?</h3>
  344. <p>Keep in mind that we’ve gone through just one of the proposed Standards above. The other one is about <a href="https://www.esafety.gov.au/sites/default/files/2023-11/Draft%20Online%20Safety%20%28Relevant%20Electronic%20Services%20-%20Class%201A%20and%20Class%201B%20Material%29%20Industry%20Standard%202024%20_0.pdf">e-mail and chat services</a>, so if you run a mail server (of any flavour – maybe even on your infrastructure?), a chat server (e.g., Prosody, jabberd), or Mastodon server, buckle up.</p>
  345.  
  346. <p>There are also another set of <a href="https://www.esafety.gov.au/industry/codes/register-online-industry-codes-standards">Industry Codes</a> that cover things like hosting services, app stores, social media, search engines, and operating systems, if you happen to provide one of those.</p>
  347.  
  348. <p>Keep in mind that if you change anything on your site that impacts risk (e.g., adding a comment form), you’ll need to re-assess your risk (and likely conform to new requirements for reporting, etc.).</p>
  349.  
  350. <h3 id="what-does-enforcement-look-like">What Does Enforcement Look Like?</h3>
  351. <p>There are a <em>lot</em> of small Internet services out there – there are a lot of IP addresses and ports, after all.  I suspect many people running them will ignore these requirements – either because they don’t know about them, they think they’re too small, that the eSafety Commissioner won’t care about their site, or they’re willing to run the risk.</p>
  352.  
  353. <p>What <em>is</em> the risk, though?</p>
  354.  
  355. <p>Section 146 of the Online Safety Act 2021 sets the penalty for not complying with an Industry Standard at 500 penalty units – currently, AU$156,500 (a bit more than US$100,000).</p>
  356.  
  357. <p>In practice, the eSafety Commissioner is unlikely to come after any site if its content isn’t problematic in their eyes. Whether you want to rely upon that is up to you. Because the legislation and standard don’t have any exemptions for small services – even with limited audiences – you are relying upon their discretion if you don’t have a risk assessment ready for them.</p>
  358.  
  359. <h3 id="what-do-you-really-think">What Do You Really Think?</h3>
  360. <p>Improving online safety is an important task that needs more focus from society, and I’m proud that Australia is trying to improve things in this area. I’m critical of the eSafety Industry Codes and now Standards not because of their objective, but because of their unintended side effects.</p>
  361.  
  362. <p>Both the enabling instrument and this delegated legislation are written without consideration for the chilling effects and regulatory burden they create on parties that are arguably not its target. Requiring professional risk assessment raises costs for everyone, and creates incentives to just use big tech commercial services, rather than self host – leaning us further into things being run by a few, big companies.</p>
  363.  
  364. <p>Moreover, if a small personal site is distributing child porn or inciting terrorism, they’re not going to be caught because it doesn’t have a properly considered risk assessment ready to produce on demand – the eSafety Commissioner already has a range of other powers they can use in that case. They don’t have the resources to go after the countless small services out there for compliance issues, so all that will remain is the lingering chilling effects of these pointless requirements.</p>
  365.  
  366. <p>I get that most people will ignore these requirements, and the eSafety Commissioner is presumably relying upon that to give them the leeway to go after the people they need to target. I just think that creating laws that can be applied with so much discretion – where technically everyone is in violation, and the regulator can pick who they prosecute – is a shitty way to run a democracy.</p>
  367.  
  368. <div class="footnotes" role="doc-endnotes">
  369.  <ol>
  370.    <li id="fn:1" role="doc-endnote">
  371.      <p>Is it just me, or is “informational” a hole big enough to drive a truck through here? <a href="#fnref:1" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
  372.    </li>
  373.    <li id="fn:2" role="doc-endnote">
  374.      <p>Notably, the site you’re reading this on doesn’t clearly qualify for any of them, and so when these codes are registered, I’ll likely be doing a risk assessment (and posting it), even though it doesn’t allow comments any more (because, spam). <a href="#fnref:2" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
  375.    </li>
  376.    <li id="fn:3" role="doc-endnote">
  377.      <p>This seems to foretell the establishment of a new industry. <a href="#fnref:3" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
  378.    </li>
  379.    <li id="fn:4" role="doc-endnote">
  380.      <p>Although it’s always tempting to write a blog entry that <em>depicts, expresses or otherwise deals with matters of drug misuse or addiction in such a way that the material offends against the standards of morality, decency and propriety generally accepted by reasonable adults to the extent that the material should be classified RC</em>. <a href="#fnref:4" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
  381.    </li>
  382.  </ol>
  383. </div>]]>
  384.    </content>
  385.  </entry>
  386.  
  387.  <entry>
  388.    <title>Technical Standards Bodies are Regulators</title>
  389.    <link rel="alternate" type="text/html" href="https://www.mnot.net/blog/2023/11/01/regulators" />
  390.    <id>https://www.mnot.net/blog/2023/11/01/regulators</id>
  391.    <updated>2023-11-01T00:00:00Z</updated>
  392.    <author>
  393.        <name>Mark Nottingham</name>
  394.        <uri>https://www.mnot.net/personal/</uri>
  395.    </author>
  396.    <summary>There are lots of ways to view what Internet standards bodies like the IETF and W3C do. They are horizontal agreements between competitors as well as mission-driven public-good charities. One might believe they’re the homes of innovation that brought us the Internet and the Web, or that they’re boring, ineffective and slow committee talking shops. Some see them as vibrant, diverse communities, while others believe they’re completely dominated by big tech.</summary>
  397.    
  398. <category term="Tech Regulation" />
  399.    
  400. <category term="Standards" />
  401.    
  402.    <content type="html" xml:lang="en" xml:base="https://www.mnot.net/blog/2023/11/01/regulators">
  403.    <![CDATA[<p>There are lots of ways to view what Internet standards bodies like the IETF and W3C do. They are horizontal agreements between competitors as well as mission-driven public-good charities. One might believe they’re the homes of innovation that brought us the Internet and the Web, or that they’re boring, ineffective and slow committee talking shops. Some see them as vibrant, diverse communities, while others believe they’re completely dominated by big tech.</p>
  404.  
  405. <p>My preferred way of thinking of them these days, however, is as <em>regulators</em>. Just like the <a href="https://www.ftc.gov">FTC</a> in the US, the <a href="https://www.esafety.gov.au">eSafety Commissioner</a> in Australia, or the <a href="https://ico.org.uk">ICO</a> in the UK, Standards Developing Organizations (SDOs) have a fundamentally regulatory aspect to them, and considering them in this way clarifies how they relate to Internet governance.</p>
  406.  
  407. <p>In particular, it helps to understand what kind of regulator they are, what tools they use, and the nature of the regime they operate within.</p>
  408.  
  409. <h3 id="a-specific-kind-of-regulator">A specific kind of regulator</h3>
  410.  
  411. <p>When most people think of a regulator, they assume it’s always state-backed; sovereign power (and hopefully a democratic mandate) imbues the regulator with legitimacy. As <a href="https://eprints.lse.ac.uk/35985/1/Disspaper4-1.pdf">Julia Black put it back in 2002</a>:</p>
  412.  
  413. <blockquote>
  414.  <p>The core understanding that many have of ‘regulation’ is some form of ‘command and control’ (CAC) regulation: regulation by the state through the use of legal rules backed by (often criminal) sanctions. ‘CAC’ has also however become shorthand to denote all that can be bad about regulation, including poorly targeted rules, rigidity, ossification, under- or over- enforcement, and unintended consequences.</p>
  415. </blockquote>
  416.  
  417. <p>Modern conceptions of regulation are much more expansive (or ‘decentered’), encompassing not only public (government) regulation but also regulation by private actors. For example, <em>lex mercatoria</em> – commercial law and customs followed by merchants – goes back to at least medieval times, and is now considered a kind of regulation. States regularly defer to such ‘soft law’, and while it can always be overridden in a single jurisdiction by legal power, policymakers have strong motivations to avoid over-regulating areas that are capable of self-regulation.</p>
  418.  
  419. <p>Further complicating Internet regulation is its global span, which means that more than one state is involved. <em>Transnational Private Regulators</em> (TPRs) are non-government regulators who work across national boundaries.</p>
  420.  
  421. <p>Internet SDOs are often used as examples of TPRs. Other common examples include organisations like the <a href="https://fsc.org/en">Forestry Stewardship Council</a>, the <a href="https://www.fairtrade.org.uk">Fairtrade Foundation</a>, the <a href="https://www.ifrs.org/groups/international-accounting-standards-board/">International Accounting Standards Board</a>, and the <a href="https://www.isealalliance.org">ISEAL Alliance</a>.</p>
  422.  
  423. <p><a href="https://www.oecd-ilibrary.org/governance/international-regulatory-co-operation-case-studies-vol-3/transnational-private-regulation_9789264200524-3-en">Caffagi identified</a> a few factors that have “caused and helped to accelerate the emergence of TPRs”:</p>
  424. <ol>
  425.  <li>Because “many goods and services today transcend national boundaries [they] can hardly be regulated by national regulations. This is particularly the case with global public goods […] for which international regulatory co-operation is substantially needed to avoid a ‘race to the bottom’ between domestic regulations.” This is very much the case for the Internet.</li>
  426.  <li>“There are markets that exhibit fast-changing dynamics [that are] difficult for public policy makers to try to regulate[.] In particular, this is the case of high-tech and knowledge-intensive markets [which] effectively leads policymakers to rely on private parties, at least for the definition of implementing measures and technical specifications.”</li>
  427.  <li>Finally, “there are policy problems that inevitably require heavy reliance on the expertise of private actors, [who] are the most informed parties, or the players in the best position to deal with a given failure, or simply the only parties holding control over central essential resources.”</li>
  428. </ol>
  429.  
  430. <p>Importantly, the legitimacy (and therefore authority) of a TPR isn’t based on democracy – inherently they have no <em>demos</em> so they cannot be democratic in the sense that a state is. Instead, they draw on other sources of legitimacy, including their input (who participates), their output (what impact they have), and their throughput (what processes they use to assure fair and good outcomes).</p>
  431.  
  432. <h3 id="with-unique-regulatory-tools">With unique regulatory tools</h3>
  433.  
  434. <p>The regulatory tools available to Internet SDOs are specific and limited – they write down technical specifications that, on a good day, get reflected in code.</p>
  435.  
  436. <p>This is ‘architectural regulation’, according to <a href="https://www.journals.uchicago.edu/doi/10.1086/468039">Lessig</a>. It sits alongside other modalities of regulation like law, norms, and markets. Where the FTC uses law, the IETF uses architecture – shaping behaviour by limiting what is possible in the world, rather than imposing <em>ex post</em> consequences.</p>
  437.  
  438. <p>While much of regulatory theory and practice is taken up with issues like monitoring and enforcement, architectural regulation doesn’t need those tasks to be performed;  the best approximation is conformance testing (which the IETF and W3C don’t formally do anyway; they certainly don’t attempt certification).</p>
  439.  
  440. <p>Another interesting aspect of this form of regulation is its <em>quasi-voluntary</em> nature. Internet standards are optional to adopt and implement; no one is forcing you to do so. However, if they’re successful and widely adopted, they do constrain your behaviour while you’re on the Internet, because <em>everyone else is following them</em>. In that sense, they are mandatory.</p>
  441.  
  442. <p>Architectural regulation of the Internet is also constrained in how it can introduce change. While a law can be repealed or overridden by a newer law, Internet protocol standards have to consider the dependencies that people already have on infrastructure; we can’t have a ‘flag day’ where we change how the Internet works. Instead, we have to carefully extend and evolve it, working within the constraints of what people already do, because once code is deployed, we lose control.</p>
  443.  
  444. <p>These features provide interesting advantages to SDOs as regulators. While one might see a non-state regulator without an enforcement problem as too powerful, standards’ lack of binding force means that an SDO can’t just impose its will; its product has to be proven by market adoption. A successful, widely adopted standard is (qualified) proof of cooperation, and thus has gained legitimacy at the same time it becomes binding.</p>
  445.  
  446. <h3 id="in-a-large-regulatory-space">In a large regulatory space</h3>
  447.  
  448. <p>If we step back from this, we can now consider the context of this regulation - Internet Governance overall. Plenty has been written about this that I won’t attempt to summarise, but there are a couple of aspects that I’d like to point out.</p>
  449.  
  450. <p>First of all, there are (obviously) other regulators present too – legal regulators especially (from various governments around the world), but also others using various combinations of the regulatory modalities.</p>
  451.  
  452. <p>Second, Internet Governance is <em>polycentric</em> (also referred to as ‘regulatory pluralism’) - there is no hierarchy and no regulator can tell another what to do. There are many sources of power (of various natures) that interact in different ways – sometimes reinforcing each other, occasionally conflicting.</p>
  453.  
  454. <p>Lessig <a href="https://lessig.org/product/codev2/">talks about this</a> (with ‘constraints’ being a synonym for ‘regulators’):</p>
  455.  
  456. <blockquote>
  457.  <p>The constraints are distinct, yet they are plainly interdependent. Each can support or oppose the others. Technologies can undermine norms and laws; they can also support them. Some constraints make others possible; others make some impossible. Constraints work together, though they function differently and the effect of each is distinct. Norms constrain through the stigma that a community imposes; markets constrain through the price that they exact; architectures constrain through the physical burdens they impose; and law constrains through the punishment it threatens.</p>
  458. </blockquote>
  459.  
  460. <p>Third, the regulatory space is also <em>fragmented</em>. Information, authority, responsibility, and capacity to regulate are dispersed unevenly across multiple regulators. As Scott <a href="https://researchrepository.ucd.ie/server/api/core/bitstreams/af2aa3c2-16ed-46f7-9961-7e8a66670bb0/content">points out</a>, ‘[r]elations can be characterized as complex, dynamic horizontal relations of negotiated interdependence.’</p>
  461.  
  462. <p>This means that <em>no</em> regulator in the space is truly independent. Standards have to operate in the legal contexts where they’re deployed; laws need to take the reality of the deployed Internet into account. Each party can act unilaterally, and might even meet their immediate goals, but the reaction to imprudent actions might be worse than the original issue they were trying to address.</p>
  463.  
  464. <p>Overall, this is healthy. Power is not concentrated in any one institution. States are able to claim sovereignty over what happens inside their borders, but if they differ too much from the global norm, they put at risk the economic and cultural benefits of being part of the global Internet.</p>
  465.  
  466. <h3 id="what-does-this-mean-for-the-ietf-and-w3c">What does this mean for the IETF and W3C?</h3>
  467.  
  468. <p>Accepting the regulatory nature of SDOs leads to a few conclusions.</p>
  469.  
  470. <p>First, the IETF and W3C need to coordinate more closely with other regulators – especially national regulators who have their sights set on taming particular aspects of the Internet.</p>
  471.  
  472. <p>That doesn’t mean that SDOs should defer to national regulators – far from it. I’ve heard more than a few conversations where technical people think they need to implement the law in protocols. This is not the case, because laws are generally limited to a specific territory; countries can’t regulate the entire Internet by themselves. Furthermore, laws typically don’t apply to the standards themselves; instead, they apply to their <em>use</em>.</p>
  473.  
  474. <p>It doesn’t even mean that standards work should block on getting input from policymakers (just as policymakers don’t block lawmaking on feedback from SDOs!); doing so would introduce problematic incentives, muddy the technical decision-making process, and remove many of the advantages of private regulation.</p>
  475.  
  476. <p>It <em>does</em> mean that technical discussions should be informed by ‘policy considerations’, even if they’re ultimately dismissed. Understanding how legal regulators see the Internet, what their goals are, and how they attempt to use the regulatory tools in their hands helps technical regulators evaluate what additional constraints are likely to be layered onto the Internet. That might result in alignment between technical regulation and legal regulation, but this is emphatically not a requirement – in some cases, they might conflict.</p>
  477.  
  478. <p>Those conflicts should be avoided when they’re unnecessary, so SDOs need to do their part to inform legal regulators as well, particularly when their proposals have impact on the architecture.</p>
  479.  
  480. <p>This is not a new perspective – there has been considerable discussion in both the IETF and the W3C recently about ‘policy engagement.’ What’s different here is the emphasis on being a peer of other regulators, rather than automatically subject to them. That is fundamentally different than the relationship that most corporate policy units have with regulators, for example.</p>
  481.  
  482. <p>Second, this view reinforces the notion that regulation by technical standards bodies has very specific sources of legitimacy – the technical expertise that it embodies, and the demonstrated success of its output. That legitimacy might be enhanced by the unique global scope of these bodies – unlike national regulators, they are responsible for the <em>entire</em> Web and Internet.</p>
  483.  
  484. <p>That suggests the positions taken by these bodies need to be focused on their areas of expertise, rather than trying to draw on other sources of legitimacy (for example, pseudo-democratic ones, or notions of openness, although the latter does enhance their legitimacy). This is well-recognised in the IETF, where arguments like <a href="https://datatracker.ietf.org/doc/html/rfc7258">Pervasive Monitoring is an Attack</a> are couched in technical terms, not value-driven ones.</p>
  485.  
  486. <p>Third, the polycentric and fragmented nature of the regulatory space suggests that it’s entirely appropriate for architectural regulators like SDOs to focus on areas where their tools are most effective.</p>
  487.  
  488. <p>For example, the HTTP Cookie specification has been <a href="https://httpwg.org/http-extensions/draft-ietf-httpbis-rfc6265bis.html#name-privacy-considerations">working towards eradicating third-party cookies</a> for some time, because they’re horrible for privacy. Some point out that this doesn’t address the privacy issues with first-party cookies - a site you’re interacting with can still track your activity, profile you, and so on.</p>
  489.  
  490. <p>That doesn’t mean that we should back away from regulating third-party cookies with architecture; they’re extremely amenable to this form of regulation (because of the user agency of the browser), and legal regulation of third-party cookies has proven difficult. On the other hand, regulating first-party privacy abuses on the Web with architecture is <em>hard</em> – if you interact with someone, you’re giving them your data – but legal regulation of how entities handle first-party data is on much firmer ground (provided there is a political will to do so).</p>]]>
  491.    </content>
  492.  </entry>
  493.  
  494. </feed>
  495.  

If you would like to create a banner that links to this page (i.e. this validation result), do the following:

  1. Download the "valid Atom 1.0" banner.

  2. Upload the image to your own server. (This step is important. Please do not link directly to the image on this server.)

  3. Add this HTML to your page (change the image src attribute if necessary):

If you would like to create a text link instead, here is the URL you can use:

http://www.feedvalidator.org/check.cgi?url=http%3A//www.mnot.net/blog/index.atom

Copyright © 2002-9 Sam Ruby, Mark Pilgrim, Joseph Walton, and Phil Ringnalda