This article was originally published on The Dispatch - Policy. You can read the original article HERE
Welcome back to Techne! Reading H.G. Wells as a kid got me obsessed with the idea of terraforming Mars. Although it has abundant water, the planet is currently too cold to sustain life. For some time, scientists have thought that Mars could be warmed using greenhouse gasses, but that would require a large mass of ingredients that are rare on the planet. A new paper suggests that warming might be accomplished through Martian dust, which is rich in iron and aluminum, giving the planet its characteristic color.
California’s SB 1047 Moves Closer to Changing the AI Landscape
Last week, the California State Assembly passed SB 1047, a controversial AI safety bill that supporters contend would regulate advanced AI models to reduce the possibility of AI going haywire and posing a serious threat to people. Formally titled the Safe and Secure Innovation for Frontier Artificial Intelligence Models Act, SB 1047 now heads to Gov. Gavin Newsom’s desk, where it faces an uncertain future.
I first wrote about SB 1047 back in May, when the legislative debate was heating up. Then in August, I wrote about the four fault lines in AI policy, based on my discussions with people who were actively campaigning for the bill. Since then, I have only become more convinced of the first fault line:
AI policy often echoes the misunderstood Kipling line: “Oh, East is East, and West is West, and never the twain shall meet.” In the East—in Washington, D.C., statehouses, and other centers of political power—AI is driven by questions of regulatory scope, legislative action, law, and litigation. And in the West—in Silicon Valley, Palo Alto, and other tech hubs—AI is driven by questions of safety, risk, and alignment. D.C. and San Francisco inhabit two different AI cultures.
I stand by what I said then. In fact, I am even more convinced that,
There is a common trope that policymakers don’t understand tech. But the obverse is even more true: Those in tech aren’t often legally conversant. Only once in those dozen or so conversations did the other person know about, for example, the First Amendment problems with all AI regulation, and that’s because he read my work on the topic.
To understand the problems with SB 1047, it needs to be viewed through a legal lens: from the viewpoint of an overly compliant legal counsel for a company, of a judge that has to rule on its legality, and of an administrator who wants to push the bounds of the law.
But from a more philosophical position, I’m just not a huge fan of the approach taken by SB 1047. The bill regulates a class of advanced AI models, called frontier models, that are only just now being developed, and imposes a series of safety protocols that lack consensus among experts and haven’t been fully fleshed out yet. The entire framework is built on the premise that these advanced models will pose a threat, which is an assumption that remains highly contested. And to top it off, if these AI models are truly dangerous in the way that some claim, then California shouldn’t be regulating them anyway—it should be the purview of the federal government.
Oh, and SB 1047 likely runs afoul of the First Amendment and the Stored Communications Act. It’s deeply concerning that the bill’s supporters are just glossing over these significant legal issues but that seems to be the state of the discourse.
The bill’s provisions.
SB 1047 went through 10 major revisions, but the core of the bill remains the same. A class of the most advanced AI models will be designated as “covered AI models” in California and then compelled to adhere to a range of requirements, including safety assessments, testing, shutdown mechanisms, certification, and safety incident reporting.
The “covered” designation comes from a technical definition, as I explained in a previous edition of Techne:
Covered AI models under SB 1047 are partially defined by the amount of computing power needed to train the model. The industry typically couches AI models in petaFLOPS, which are 10^15 floating-point operations. OpenAI’s GPT-4 is estimated to have taken 21 billion petaFLOPS to train, while Google’s Gemini Ultra probably took 50 billion petaFLOPs. Similar to the standard set by President Joe Biden’s executive order on AI, SB 1047 would apply to models with greater than 10^26 floating-point operations, which amounts to 100 billion petaFLOPS. So the current frontier models are just below the covered AI model threshold, but the next generation of models—including GPT-5—should probably hit that regulation mark.
Stripped from the bill was a provision that regulated any models that could achieve similar benchmarks to those of 10^26 floating-point operations. In its place is the requirement that the model must also cost $100 million to train to be covered. There was also a provision added that gives California’s Government Operations Agency the ability to designate any model that cost $100 million as being covered by the law as well.
What hasn’t substantially changed are the requirements for covered AI models. Among others, covered models will have to:
- “Implement reasonable administrative, technical, and physical cybersecurity protections”;
- Build in a killswitch;
- Implement a detailed safety and security protocol that is certified by the company;
- Conduct annual reviews of the safety procedures; and
- “Take reasonable care to implement other appropriate measures to prevent covered models and covered model derivatives from posing unreasonable risks of causing or materially enabling critical harms.”
SB 1047 is built on reasonableness standards, which are notoriously tricky to define in the law. Indeed, an astute commenter on the blog Astral Codex Ten explained what it might mean if developers were to take seriously the reasonableness requirements:
Under the traditional Learned Hand formula, you are obligated to take a precaution if the burden of the precaution (B) is less than the probability of the accident it would prevent (P) multiplied by the magnitude of the loss resulting from the accident (L). B < P*L. Given that the “loss” threatened by advanced AI is complete and total destruction of all value on earth, the right side of the equation is infinite, and reasonableness requires spending an unlimited amount on precautions and taking every single one within your power. Even if we just cap the L at $100T, the estimated value of the global economy, a p(doom) of even 10% would mean that any precaution up to $10T was justified. Now presumably there are smaller-scale cyber attacks and things of that nature that would be what actually happens and prompts a negligence suit, if DOOM happens nobody’s around to sue, so this isn’t gonna come up in court this way, but as a way to think about legal “reasonableness” that’s what seems to be required.
Yes, it is absurd, but that’s what happens when you start trying to mandate these ideas into law.
While I’m skeptical of the bill, it has garnered the support of influential online writers like Zvi Mowshowitz and Scott Alexander. I hold both in high regard and have learned much from them. Still, I see their analysis as being fundamentally flawed because they are grounded in rationalism rather than legality. Alexander previously ran the Slate Star Codex blog, which helped to foster the rationalist community. Mowshowitz recently discussed regulating frontier AI models in a spate of articles. What they evince, to me, is a naivete about legal processes and history.
For example, when Mowshowitz wrote about Section 22605, which eventually was removed from the bill, he pointed out that this part of the bill “requires sellers of inference or a computing cluster to provide a transparent, uniform, publicly available price schedule, banning price discrimination, and bans ‘unlawful discrimination or noncompetitive activity in determining price or access.’” He continues, “I always wonder about laws that say ‘you cannot do things that are already illegal,’ I mean I thought that was the whole point of them already being illegal.” But the entire point of Section 22605 was to create rate regulation. When I read this part of the bill, I thought about the decades-long fight in telecom over total element long-run incremental cost.
Similarly, Alexander seems to underweight legal review and administrative process when he wrote:
Finally – last week discussed Richard Hanania’s The Origin Of Woke, which claimed that although the original Civil Rights Act was good and well-bounded and included nothing objectionable, courts gradually re-interpreted it to mean various things much stronger than anyone wanted at the time. … But Hanania’s book, and the process of reading this bill, highlight how vague and complicated all laws can be. The same bill could be excellent or terrible, depending on whether it’s interpreted effectively by well-intentioned people, or poorly by idiots. That’s true here too.
Administrative law doesn’t simply depend “on whether it’s interpreted effectively by well-intentioned people, or poorly by idiots.” An aggressive, well-intentioned agency can push the bounds of its authority. The Federal Communications Commission has been in and of court for two decades because of how it interprets its authority. The Food and Drug Administration was challenged for its regulation of tobacco using statutory authority for “drugs” and “devices.” The list of agencies using their authority in one arena for a way that it wasn’t intended is extensive. Admittedly, some of this has been curtailed by recent Supreme Court decisions, but SB 1047 gives a state government a lot of power to determine what is considered safe.
But more importantly, mandating a killswitch inherently involves the First Amendment. I’ve touched on this point before, building on a commentary by John Villasenor of the Brookings Institution. But SB 1047’s backers don’t appear to have given these legal concerns the attention they deserve.
Newsom has until September 30 to sign the bill or veto it. Since the governor has been largely silent about which way he’ll go, both supporters and detractors of the legislation have been inundating his office with letters meant to persuade him.
Whatever happens, it is hard not to read SB 1047’s passage as part of a larger normalization of relations between AI developers and the government. Only just last week, the U.S. Artificial Intelligence Safety Institute announced agreements that enable the agency to access major new models from OpenAI and Anthropic before they go public. While this route has its own problems, it is far better than SB 1047.
A lot changed this summer. AI developers, once largely independent of governmental influence, are now establishing deeper institutional ties to navigate regulatory challenges. I can’t imagine this latest development is a good thing.
Until next week,
Notes and Quotes
- Techne is anti-COVID but pro-Corvid. New evidence suggests that the songbird family that includes crows are actually more intelligent than previously thought.
- On Monday, Brazil’s Supreme Court confirmed the nationwide ban on Elon Musk’s X platform after the social media company failed to follow orders issued by the country’s top judge, Alexandre de Moraes, to name a legal representative. In previous comments about misinformation on social media, Moraes has said, “Freedom of expression doesn’t mean freedom of aggression.”
- The 3rd U.S. Circuit Court of Appeals ruled last week that TikTok can be sued for recommending the “blackout challenge” video that led to a young girl’s death. The ruling potentially sets up another fight over Section 230 at the Supreme Court. Legal scholar Daphne Keller had this to say on X: “The 3rd Circuit is engaging in the absurd pretense that the [Supreme] Court actually decided this issue in Moody v. NetChoice, because it said algorithmic ranking that advances the platform’s content moderation goals is the platform’s 1st Am protected speech.”
- Worth checking out: Works in Progress is a magazine that explores fresh, overlooked ideas aimed at making the world a better place. Issue 16 just dropped and it includes articles on lab-grown diamonds, advanced market commitments, and how pour-over coffee got good.
- The Boeing Starliner saga keeps getting weirder. Last week in Techne, I told you about how SpaceX will shuttle the astronauts back. Now, strange sounds are coming from the Starliner capsule that the crew describes as “almost like a sonar ping.”
- Can the cost of solar panels keep dropping? Tomas Pueyo investigates.
- Humans have killed off countless species, but scientists are inching closer to reviving some of them. “De-extinction” science has advanced dramatically in the past two decades, such that scientists are close to constructing the full genome of wooly mammoths, dodos, and Tasmanian tigers.
- I recently ran into my old colleague, Collin Hitt of St. Louis University, who helped organize a new poll on what Missouri voters think about banning cell phones in schools. The big takeaway is that nearly 4 of 5 (79 percent) likely voters in Missouri support prohibiting students from accessing their phones during class. The accompanying analysis offers a solid overview of the degree of support for banning cell phones in school.
- There have been a lot of advancements in cancer treatments in recent years, but pancreatic cancer remains underfunded relative to the others. A long read in the Financial Times explains why: “Pancreatic cancer does not have patient advocates because they all die,” said Julie Fleshman, chief executive of the Pancreatic Cancer Action Network.
- Jonathan Kay’s review on Quillette of David Alff’s The Northeast Corridor: The Trains, the People, the History, the Region, makes me want to pick up the book: “Alff has produced a proper history of the eponymous rail line from Boston to Washington, D.C. that became early America’s infrastructural backbone. And somehow, he managed to pack it into a book that’s scarcely more than 250 pages, despite also providing an abundance of memorable digressions into the arts of rail-station architecture, bridge construction, and tunnel blasting.”
AI Roundup
This article was originally published by The Dispatch - Policy. We only curate news from sources that align with the core values of our intended conservative audience. If you like the news you read here we encourage you to utilize the original sources for even more great news and opinions you can trust!
Comments