site banner

Culture War Roundup for the week of November 20, 2023

This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.

Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.

We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:

  • Shaming.

  • Attempting to 'build consensus' or enforce ideological conformity.

  • Making sweeping generalizations to vilify a group you dislike.

  • Recruiting for a cause.

  • Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.

In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:

  • Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.

  • Be as precise and charitable as you can. Don't paraphrase unflatteringly.

  • Don't imply that someone said something they did not say, even if you think it follows from what they said.

  • Write like everyone is reading and you want them to be included in the discussion.

On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.

7
Jump in the discussion.

No email address required.

Senior OpenAI leaders, including Mr. Sutskever, who is deeply concerned that A.I. could one day destroy humanity

These people are so weird. They are obsessed with making something they think will destroy the world. Just build AI or don't. If AI is so dangerous, nothing you can do will really stop it in the long run. Eventually, over hundreds or even thousands of years, the inevitable will happen. I feel like this is almost a religion for them and they are the absolute worst people to be in charge or this kind of technology.

  1. They think AGI is inevitable, and if it's not made by people with safety and alignment in mind, the odds of it being misaligned skyrocket.

  2. Control of an aligned Superintelligent AGI is equivalent to having the keys to the lightcone, if you make it through the gauntlet of it not killing you and it listens to what you tell it, then you have the means to dominate everyone else, including others who make misaligned AGI, if yours is capable of squashing them at birth, or at the very least capable of panopticon surveillance to prevent anyone from building one in the first place.

  3. Even prior to that, being on the cutting edge of AI research gives you a voice, people can dismiss Eliezer Yudkowsky as a crank, far harder to do that to Geoffrey Hinton or Ilya Sutskever. You have far more power to get governments to regulate things, or develop industry best standards that reduce the risk of more laissez-faire competitors YOLOing the whole thing.

AI is dangerous, as is anything much smarter than you that potentially doesn't share your goals, but if you see it as inevitable, then your best bet is making sure it comes out with goals you share or control.

My own p(doom) from AI has dropped to about 30% from a high of 70%, when RLHF and other techniques showed that it was possible to ~mostly align the best AI today, in the form of LLMs, which are the frontrunners for the best AIs of tomorrow. Just because I currently do not think AI will probably kill us in a decade doesn't mean I don't think it's a serious risk, and there are few things on the planet more worth being passionate about.

It is not remotely as simple as build AI or don't, even if everyone associated with the AI X-risk community died overnight, the genie is out of the bottle, and others will pursue it with billions of dollars. OAI had a value, before this debacle, of about $80 billion, with operating expenses on the order of 500 million p/a.

If you can't undo that, or enforce a "pause", then you grudgingly find a way to stay ahead of the competition while doing your best not to be the reason why it all went to hell.

Control of an aligned Superintelligent AGI is equivalent to having the keys to the lightcone

I think this sort of argument consistently relies on assumptions regarding the possibilities of various highly-powerful technologies being physically possible but not yet discovered due to humans not being smart enough that I find insufficiently supported. It's always "AI will give us X" — Drexlerian nanotech, space elevators, "energy too cheap to meter," whatever — without considering the outcome where it turns out none of these is actually possible, no matter how smart you are. To quote from the Futurama episode "A Clone of My Own":

Professor Hubert Farnsworth: Nothing is impossible. Not if you can imagine it. That's what being is a scientist is all about.

Cubert J. Farnsworth: No, that's what being a magical elf is all about.

I disagree, it's largely Yudkowsky who vocally claims that a SAGI will rely on things like "diamondoid bacteria" and other nanotech to get an advantage.

For me, and many others, subversion of existing human infrastructure through social engineering to do things like launch nukes, engineering hyper-lethal and virulent pathogens and the like are all feasible for something modestly above human, without relying on anything that doesn't exist. The AI will need robust automation to replace humans, but we're already doing that ourselves, so..

We could already have had energy too cheap to meter if we went full send on nuclear, for one. It would certainly be dirt cheap compared to today's rates.

subversion of existing human infrastructure through social engineering

I think this is overrated, too — though that might be due to reading too many "unboxing" arguments predicated on the assumption that absolutely anyone can be convinced to do absolutely anything, if only you're smart enough to figure out the particular individually-tailored set of Magic Words.

I have never claimed it can convince anyone of literally anything. We've already had plenty of nuclear close-calls simply because of the fog of war or human/technical error.

Similarly, there are already >0 misanthropically omnicidal people around and kicking, and an AI could empower them to pursue their goals, or they might choose to adopt the AI for that purpose.

Mere humans, or human-run orgs like the CIA have long engineered regime change, it seems to me incredibly unlikely, to the point it can be outright dismissed from consideration, that an AGI only modestly higher in intelligence couldn't do the same, and even independently play multiple sides against each other until they all make terrible decisions.

Besides, it's clear that nobody even tries the Yudkowskian boxing approach these days. ARC evals, red-teaming and the like are nowhere close to the maximally paranoid approach, not even for SOTA models.

A group of say, 160 IQ humans with laser-focus and an elimination of many/most of the coordination and trust bottlenecks we face could well become an existential threat. Even a modestly superintelligent or merely genius level AGI can do that and more.

and an AI could empower them to pursue their goals, or they might choose to adopt the AI for that purpose.

Empower them how, exactly? What is it that they aren't able to do now only because they're not smart enough, that more intelligence alone can solve? Intelligence isn't magic.

that an AGI only modestly higher in intelligence couldn't do the same, and even independently play multiple sides against each other until they all make terrible decisions.

Perhaps, but what's your proof that it could do this so much better than the CIA or anyone else, just because it's smarter? Intelligence isn't magic.

A group of say, 160 IQ humans with laser-focus and an elimination of many/most of the coordination and trust bottlenecks we face could well become an existential threat.

Actually, as a 151 IQ human, I mostly disagree with this, so that's part of it right there.

Even a modestly superintelligent or merely genius level AGI can do that and more.

What's your proof of the part I just emphasized? You appear to simply assume it.

I think you might be a uniquely ineffective 151 IQ human if it doesn't seem plausible to you that a group of very smart humans could do extreme and perhaps existential harm. To me, the main thing preventing that seems to be not the inherent hardness or weakness of, say, COVID-Omicron-Ebola, but the resistance of an overwhelming majority of other humans (including both very smart ones and mediocre but well-organized ones).

As for what a superintelligent AI changes? Well for one thing, it eliminates the need to find a bunch of peers. And, with robots, the need for lab assistants.

And I have like 3% P(AI Doom).

This is an excellent answer. One small quibble:

Control of an aligned Superintelligent AGI is equivalent to having the keys to the lightcone, if you make it through the gauntlet of it not killing you and it listens to what you tell it, then you have the means to dominate everyone else, including others who make misaligned AGI, if yours is capable of squashing them at birth, or at the very least capable of panopticon surveillance to prevent anyone from building one in the first place.

For the record I think Yudkowsky and friends are wrong about this one. Control of the only superintelligent AGI, if that AGI is a single coherent entity, might be the keys to the lightcone, but so far it looks to me like AGI scales horizontally much better than it scales vertically.

This, if anything, makes things more dangerous rather than less, because it means there is no permanent win condition, only the deferral of the failure condition for a bit longer.

Thanks!

but so far it looks to me like AGI scales horizontally much better than it scales vertically.

This particular concern hinges on recursive self-improvement, and I agree that we haven't seen much evidence of that, yet, but it's still the early days.

I think that the intelligence of LLMs needs to at least reach that of the average ML researcher capable of producing novel research and breakthroughs before we can call it one way or another, and we're not there yet, at least in terms of released models, not that I expect Gemini or GPT-5 to be that smart yet. The closest I can think of is training LLMs on synthetic data curated by other models, or something like Nvidia using ML models to optimize their hardware, but that's still weaksauce.

If it turns out to be feasible, it still remains to be seen whether we have a hard take-off with a Singleton or a slow (yet fast on human timescales, just months or years) takeoff which might allow for multipolarity. I remain agnostic yet gravely concerned myself.

This particular concern hinges on recursive self-improvement

And most of the talk on that issue assumes that the point where said self-improvement hits steep diminishing returns must necessarily be somewhere far above human intelligence — again, apparently based on nothing beyond it being more conducive to one's preferred outcomes than the alternative.

Diminishing returns != no or negative returns. Intelligence is the closest thing we have to an unalloyed good, and the difference in capabilities between people with just 20 or 30 IQ points is staggering enough.

Nothing at all suggests that the range of IQ/intelligence seen in unmodified humans constrained by a 1.4 kg brain in a small cranium applies at all to an entity that spans data-centers, especially those that can self-modify and fork themselves on demand. You don't need a bazillion IQ points to be immensely dangerous, human scientists with maybe 160 or 170 invented nukes.

We have AI that already matches human intelligence on many or even most cognitive tasks, the scaling laws still hold, and companies and nations can easily afford to throw several OOMs more money at the problem.

Humanity itself has seen exponential or even super-exponential advancement, and we've barely gained a handful of IQ points from the Flynn effect, most of it was merely technological compounding.

Since the theoretical or practical upper limits on the size and speed of an AGI are massive, I wish to see what reason anyone has to claim they'll bottom out within spitting distance of the smartest humans. That is ludicrous prima facie, even if we don't know how fast further progression will be.

You don't need a bazillion IQ points to be immensely dangerous, human scientists with maybe 160 or 170 invented nukes.

Yes, but you're assuming there's a lot more even more dangerous things "out there" for a smarter entity to discover.

What is intelligence for? That is, what is its use? Primarily,

  1. modeling reality
  2. modeling other agents.

Our first day of Physics lab classes at Caltech, the instructor told us that it doesn't matter how many digits of pi we'd all memorized (quite a bunch), just use 3.14, or a scientific calculator's pi key, whichever was faster, because any rounding error would be swamped out by the measurement error in our instruments.

When it comes to modeling the physical world, sure, going from knowing, say, Planck's constant to two decimal places to knowing it to three decimal places will probably net you a bunch of improvements. But then going from, say, ten decimal places to eleven, or even ten decimal places to twenty, almost certainly won't net the same level of improvement.

When modeling other minds, particularly modeling other minds modeling you modeling… — the whole "I know that you know that I know…" thing — well, that sort of recursion provides great returns on added depth… in certain games, like chess. But AIUI, in most other situations, that kind of thing quickly converges to one or another game-theoretic equilibrium, and thus the further recursion allowed by greater intelligence provides little additional benefit.

I'm not saying we can't produce an intelligence "that spans data-centers" much smarter than us, and I'm not saying it's impossible that there are dangerous and powerful things such an intelligence might figure out, I'm just saying it can't just be assumed, or treated as highly likely by default. That it's unsupported extrapolation to reason 'smart=nukes, therefore super-smart=super-nukes and mega-smart=mega-nukes.' I'm not saying that machine intelligence will "bottom out within spitting distance of the smartest humans," I'm saying that it's possible that the practical benefits of such intelligence, no matter how much vaster than our own, may "bottom out" well below the dreams of techno-optimists like yourself, and you can't just rule that out a priori on an unsubstantiated faith that there's vast undiscovered realities beyond our limited comprehension just waiting for a smarter being to uncover.

I want you to at least consider, just for a moment, the idea that maybe we humans, with our "1.4 kg brain[s] in a small cranium," may have a good enough understanding of reality, and of each other, that a being with "900 more IQ points" won't find much room to improve on it.

I'm not saying "a machine can never be smarter than a man!" I'm saying "what if a machine a thousand times smarter than us says, 'yeah, you already had it mostly figured out, the rest is piddly details, no big revelations here'?"

Yes, but you're assuming there's a lot more even more dangerous things "out there" for a smarter entity to discover.

I repeat that, while I think this is true, it's still not necessary for a genius AI to be an existential risk. I've already explained why multiple times.

Nukes? They exist.

Pandemics? They exist. Can they be made more dangerous? Yes. Are humans already making them more dangerous for no good reason? Yes.

Automation? Well underway.

Our first day of Physics lab classes at Caltech, the instructor told us that it doesn't matter how many digits of pi we'd all memorized (quite a bunch), just use 3.14, or a scientific calculator's pi key, whichever was faster, because any rounding error would be swamped out by the measurement error in our instruments.

When it comes to modeling the physical world, sure, going from knowing, say, Planck's constant to two decimal places to knowing it to three decimal places will probably net you a bunch of improvements. But then going from, say, ten decimal places to eleven, or even ten decimal places to twenty, almost certainly won't net the same level of improvement.

I do not think that the benefits of additional intelligence as seen even in human physicists is well addressed by this analogy. The relevant one would be comparing Newtonian physics to GR, and then QM. In the domains where such nuance becomes relevant, the benefits are grossly superior.

For starters, while the Standard Model is great, it still isn't capable of conclusively explaining most of the mass or energy in the universe. Not to mention that even if we have equations for the fundamental processes, there are bazillions of higher-order concerns that are intractable to simulate from first-principles.

AlphaFold didn't massively outpace SOTA on protein folding by using QM on a molecule by molecule basis. It found smarter heuristics, that's also something intelligence is indispensable for. I see no reason why a human can't be perfectly modeled using QM, it is simply a computationally intractable problem even for a single cell within us.

In other words, knowing the underling rules of a complex system != knowing all the potential implications or applications. You can't just memorize the rules of Chess and then declare it's a solved problem.

That it's unsupported extrapolation to reason 'smart=nukes, therefore super-smart=super-nukes and mega-smart=mega-nukes.'

I'm sure there people who might make such a claim. I'm not one of them, and like I said, it's not load bearing. Just nukes is sufficient really. Certainly in combination with automation so the absence of those pesky humans running the machines isn't a problem.

I want you to at least consider, just for a moment, the idea that maybe we humans, with our "1.4 kg brain[s] in a small cranium," may have a good enough understanding of reality, and of each other, that a being with "900 more IQ points" won't find much room to improve on it.

I have considered it, at least to my satisfaction, and I consider it to be exceedingly unlikely. Increases in intelligence, even within the minuscule absolute variation seen within humans, is enormously powerful. There seems to be little to nothing in the way of further scaling in the context of inhuman entities that are not constrained by the same biological limitations in size, volume, speed or energy. They already match or exceed the average human in most cognitive tasks, and even if returns from further increases in intelligence diminish grossly or become asymptotic, I am the furthest from convinced that stage will be reached within spitting distance of the best of humanity, or that such an entity won't be enormously powerful and capable of exterminating us if it wishes to do so.

Nukes? They exist.

Pandemics? They exist. Can they be made more dangerous? Yes. Are humans already making them more dangerous for no good reason? Yes.

But I don't see why super-intelligent AI will somehow make these vastly more dangerous, simply by being vastly smarter.

I consider it to be exceedingly unlikely.

Based on what evidence?

There seems to be little to nothing in the way of further scaling in the context of inhuman entities that are not constrained by the same biological limitations in size, volume, speed or energy.

Again, I agree, but again, further scaling in intelligence≠further scaling in power.

They already match or exceed the average human in most cognitive tasks

Again, so what? What part of "greater ability in cognitive tasks"≠"greater power over the material world" are you not getting — beyond, apparently, your need for it to be so based on you tying so much of your own ego to your own higher-than average intelligence?

I am the furthest from convinced that stage will be reached within spitting distance of the best of humanity.

Based on what evidence?

The relevant one would be comparing Newtonian physics to GR, and then QM. In the domains where such nuance becomes relevant, the benefits are grossly superior.

My degree is in physics. Yes, there's problems with the Standard Model. But there's far from any guarantees that whatever might replace it will provide anything like as revolutionary as those previous changes, particularly when it comes to practical effects.

Maybe I've just listened to Eric Weinstein go on about needing to put vast amounts of funding into physics too many times, because he never stops to consider that the "revolutionary new physics" we "need" to become "interplanetary" just aren't there. And then what?

More comments

The theory is that a) there's so much power and money on the table here that someone doing something world-changing is unavoidable, and b) that the early design stages of such a world-changing system will have massive impact on whether. There are ways to argue against either or both of these assumptions, or to point out separate issues.

((Some of which I agree with: even accepting those propositions, these guys are demonstrably putzes when it comes to actually persuading or planning.))

But 'it's like a religion' isn't a very meaningful claim.

Yeah other people have come to a similar conclusion: What OpenAI shares with Scientology