site banner

Culture War Roundup for the week of September 23, 2024

This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.

Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.

We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:

  • Shaming.

  • Attempting to 'build consensus' or enforce ideological conformity.

  • Making sweeping generalizations to vilify a group you dislike.

  • Recruiting for a cause.

  • Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.

In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:

  • Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.

  • Be as precise and charitable as you can. Don't paraphrase unflatteringly.

  • Don't imply that someone said something they did not say, even if you think it follows from what they said.

  • Write like everyone is reading and you want them to be included in the discussion.

On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.

5
Jump in the discussion.

No email address required.

Sam is going to get us all killed; that he's entirely misanthropic and sincerely believes that humanity should die out giving birth to machine intelligence.

...Fine, I'll bite. How much of this impression of Sam is uncharitable doomer dressing around something more mundane like "does not believe AI = extinction and thus has no reason to care", or even just same old "disregard ethics, acquire profit"?

I have no love for Altman (something I have to state awfully often as of late) but the chosen framing strikes me as highly overdramatic, besides giving him more competence/credit than he deserves. As a sanity check, how -pilled would you say that friend of yours is in general on the AI question? How many years before inevitable extinction are we talking here?

You are making an "argument from incredulity", i.e. the beliefs of Sam Altman are so crazy that they can’t be real. I don't think this is the case. Many powerful people in Silicon Valley have beliefs that are far outside the Overton Window.

Say what you will about Elon Musk, he is at least pro-human. This is not at all the case for many of his peers. For example, Larry Page and Elon Musk broke up as friends over Musk's "speciesist" belief that humanity should remain dominant over god-like AI's.

The idea that Sam Altman would literally want to destroy humanity to birth in a superior AI life form might sound ridiculous to you. But you don't know these people.

There's a good chance (not 100%, but not 0% either) that we're going to build superintelligence while the "adults in the room" argue about GDP numbers or whatever. If this happens it could make some people (perhaps a single person) more powerful than anyone in history. Do you want Sam Altman to be that person? Because I sure as hell don't.

You are making an "argument from incredulity", i.e. the beliefs of Sam Altman are so crazy that they can’t be real. I don't think this is the case.

The idea that Sam Altman would literally want to destroy humanity to birth in a superior AI life form might sound ridiculous to you. But you don't know these people.

Besides this being a gossip thread, your argument likewise seems to boil down to "but the beliefs might be real, you don't know". I don't know what to answer other than reiterate that they also might not, and you don't know either. No point in back-and-forth I suppose.

There's a good chance (not 100%, but not 0% either) that we're going to build superintelligence while the "adults in the room" argue about GDP numbers or whatever. If this happens it could make some people (perhaps a single person) more powerful than anyone in history. Do you want Sam Altman to be that person? Because I sure as hell don't.

At least the real load-bearing assumption came out. I've given up on reassuring doomers or harping on the wisdom of Pascal's mugging, so I'll simply grunt my isolated agreement that Altman is not the guy I'd like to see in charge of... anything really. If it's any consolation I doubt OpenAI is going to get that far ahead in the foreseeable future. I already mentioned my doubts on the latest o1, and coupled with the vacuous strawberry hype and Sam's antics apparently scaring a lot of actually competent people out of the company, I don't believe Sam is gonna be our shiny metal AI overlord even if I grant the eventual existence of one.

Since this is a gossip thread...

I have a couple friends who genuinely want the extinction of the human race. Not in a mass murder sense as they conceptualize it, but in a create a successor species, give a good life to the remaining humans, maybe offer them the chance for brain uploads, sense. Details and red lines vary between them, but they'd broadly agree that this is a fair characterization of their goals and desires.

Where do they work? OAI, Anthropic, GDM.

I have a fair amount of sympathy for their viewpoints, but it's still genuinely shocking. It's as if you suddenly found out that every government official was secretly a Hare Krishna or part of the People's Temple, and then when you point it out, everyone thinks the accusation is too absurd to be real.

In their defense: why do we care so much about the survival of homo sapiens qua sapiens? We're different from how we were 50,000 years ago, and we'll be more different still in 5,000, and maybe even 500. So what? So long as we have continuity of culture and memory, does it matter if we engineer ourselves into immortal cyborgs or whatever is coming? What's so special about the biped mammal vessel for a mind?

What's so special about the biped mammal vessel for a mind?

The biped mammal vessel. An immortal cyborg is a qualitatively different existence and so it will have a correspondingly different mind.

A 6'7 NBA player has a qualitatively different experience from a 5'1 ballerina, but they're both humans with minds.

if we engineer ourselves into immortal cyborgs

Hubris of the highest order.

We don't let humans so much as stitch up some skin unless they've gone through a decade of training. We don't let new engineers commit new code, unless they've spent time understanding the base architecture. What makes you think we know enough about what it means to be homo sapiens that we can go replacing entire parts wholesale ?

Just look at the last few decades. We put a whole generation of women on pills that accidentally change that characteristics of which men they're attracted to. The last-gen painkillers caused the biggest drug epidemic in the country. The primary stimulant of the century (cigarettes) was causing early death enmasse. We don't know why there is a detectable difference in immunity between c-section vs natural deliveries, and this is a difference of a few seconds. That's how little we know about these flesh-suits of ours. We have no clue what we're doing.

What's so special about the biped mammal vessel for a mind?

Don't take this the wrong way. What I'm about to say is definitely stereotyping a certain type of person.

But, I only ever see internet neuro-divergents ask these sort of questions. To normies, your question sounds like the equivalent of ,"What's so great about fries?". You'd only ever ask the question if you've never enjoyed a good pack of fries or a equivalent food that makes you feel that special thing. It reveals the absence of a fundamental human experience. To a degree, it reveals that you're less human or at least 'dis-abled'.

I'm entitled. I don't think I need to explain what makes some things special. The first day of the monsoon, petting a puppy, making faces at a toddler, a warm hug, the top of a mountain, soul food, soul music, the first time you hold your child, the last time you hold your parent, the first time a sibling defeats you at a game.

In a way, these unspoken common traits are what makes all of us human. I care about the survival of these consistent 300k-old traits, because I cherish these things. And I believe that a non-human would not be able to. Because we aren't taught to cherish these things. We just do. I don't expect everyone to have experienced all of these, in the same way. Civilizational differences mean that specifics differ. But, the patterns are undeniable.

Why do I care about the authentic experiences of my imperfect body and imperfect mind ? Because that is what it means to be human.

P.S: and I am every bit an atheist. Do I have to believe in divinity to believe in beauty ?

Not gonna make an argument here because I don't think there would be a point, but I'll mention that you're doing a great job demonstrating my concerns about atheists.

Well, leaving it at that would be a cheap shot, so,

I don't think I'm my mind any more than I'm my body. Which is to say, yes to both, but there's more going on than that. Also, human beings are uniquely divine, and God is a man in heaven. Human existence and experience are uniquely important, and uniquely destined.

Believe it or not I'm open to the idea that at some point 'we' make the transition to non-organic substrates. I just don't know enough about what actually matters to rule that out. But when people are eager to make the jump to artificial bodies and minds (not that you actually advocated for this), they strike me as dangerously naive in terms of their assumptions.

How sure are you that what we are can be digitized? What, specifically, is valuable to you, and worthy of cultivation? In symbolic terms, which gods do you actually serve?

So you're arguing for qualia and souls, yes? I believe I am my mind, that the mind is computation, and that its computational substrate is irrelevant. I'm honestly baffled by people who hold otherwise --- I want to be charitable, but I'm having a hard time seeing past opposition being ultimately a product of personal incredulity regarding our conscious experience being a worldly, temporal information processing phenomenon.

Our minds are worldly, temporal information processing phenomena, yes. At least mostly, as we experience them. No disagreement there. The question is whether, if and when our minds die, there is anything of us left. I think so.

We have no idea what consciousness is, how it happens, or even why it should ever arise in the first place. Until that's sorted there's a ton of room for other perspectives. Soul of the gaps, sure. That accusation wouldn't trouble me.

Perhaps I could say that I think our minds are so loud in our conscious experience that we fall into the mistaken assumption that everything occurring in our consciousness is our minds. The only way to find out is to die. In the meantime I'm not in a rush to create perfect, immortal copies of my mind which have no internal conscious experience, let the last bio-humans die off, and call it a day.

But I want to repeat the question:

How sure are you that what we are can be digitized? What, specifically, is valuable to you, and worthy of cultivation? In symbolic terms, which gods do you actually serve?

Your position is fundamentally religious, isn't it? We feel that existing, thinking, being are so profound that they must continue after death. But what if they aren't? I've never seen evidence that they are. If you'd like to adopt a religiously flavored epistemology, that's fine, but having done so, you've departed from the realm of logical argumentation.

There are several parts of your comment to which I could respond, but we're fundamentally coming at this in different reference frames and it would take an entire overhaul to communicate the things I want to. All I can say is that I used to see it your way, and now I don't, and basically feel about your perspective the way you feel about mine. Also that we've had different experiences, and mine have me as entirely convinced as I can be, I think.

It would be rude to throw out a bunch of examples knowing full well that I don't intend to try to discuss them since, as I just said, that'd be pretty fruitless. So I won't.

But I do want to ask again, and let me rephrase here. Gonna ramble a bit.

Let's talk about concepts. Patterns. From a strictly reductionist standpoint, all that exists is the quantum waveform, and no part of it can be severed from the whole. It's normal to regard concepts like 'justice' or 'joy' as abstract, somehow qualitatively different from concepts like 'elk' or 'door'. But again, strictly speaking from a reductionist standpoint, none of these things actually exist. Or if you don't care for the quantum thing, let's say that all that exists is fundamental particles and energy. You can take what we call a door (or an elk) apart particle by particle -- at what point does it stop being that thing? Was it ever that thing, or are concepts like 'door' and 'elk' entirely artifacts of conscious minds? Does our categorization of a thing make it into what it is? If there were no conscious life in our universe, would there be planets? Or only the configurations of matter that we call planets? If there were humans (but without consciousness), would they still be humans?

At its root, theism is the perspective that these concepts, these 'gods', exist independently of our perception of them. An easy example which people often go to for other reasons is, do numbers exist independently of us, or is their existence intrinsic to (perhaps emergent from) the universe?

I'm throwing all of this out there to hopefully help you see what I'm asking when I ask you: Which gods do you serve? Are there any patterns, gods, that you think should exist; which deserve to be prioritized over others? What do you value? And why?

All I have from you is that you think it's good (personally preferable?) for cultural continuity and memory to be cultivated, presumably until it's all wiped away in some kind of cosmic apocalypse beyond our ability to control. And no, I'm not planning on going anywhere with this. There's no 'gotcha' waiting for you around the corner. I'm just having a hard time seeing why you're saying what you're saying.

Regarding the realm of logical argumentation, what's your view on Determinism and the free-will problem?

More comments

I'm sure the Neanderthals' last thoughts included "so what, those skinny folks with the funny heads will survive even after they've wiped us out. We shall go gently into that good night."

We're homo sapiens. If we take AI true believers seriously, this isn't hundreds of years in someone else's lifetime; it could be less than ten years before an amoral sociopath unleashes something beyond our control. I plan on being alive in ten years.

I do not happen to think AI (from the LLM model) is likely to be an extinction-level threat (that's a specific phrasing). I do think Sam Altman is a skilled amoral sociopath who shouldn't be trusted with so much as kiddy scissors, and it should haunt Paul Graham that he didn't smother Altman's career when he had a chance.

We're also part Neanderthal. (Most people reading this message in 2024 are, anyway.) Their legacy got folded into ours. Why does their story have a sad ending?

Agreed on jitters about Altman. I'm just pointing out that the AI successor species people kind of have a point.

The companies being a cult is a big part of their strategy.

Information secrecy is top notch, everyone willingly works insane hours and you can get engineers to do borderline illegal things (around data privacy and ownership) without being questioned.

I know a few people at Facebook AI research, MSR and (old) Google Brain. They seem normal. But folks at OpenAI, Anthropic & Deep mind are well known to be ..... peculiar (and admittedly smarter than I am).

There's peculiar people at every part of every company. IME people at deep mind are not more peculiar than those working at other parts of Goog, and I certainly wouldn't describe them as cultists. Can't speak for the other labs.

On further thought, I have met more cultists at some of these companies, but a majority (50%+) have been normal. Also, can't exactly scale those anecdotes up.

With that reflection, I'll take back my earlier comment.

I imagine many people of the more materialist bend are both more likely to be excited by AI and more likely to not believe uploading is extinction (in a way that matters).

Totally in line though with stories about other Silicon Valley leaders.

https://www.astralcodexten.com/p/should-the-future-be-human

Business Insider: Larry Page Once Called Elon Musk A “Specieist”:

Tesla CEO Elon Musk and Google cofounder Larry Page disagree so severely about the dangers of AI it apparently ended their friendship.

At Musk's 44th birthday celebration in 2015, Page accused Musk of being a "specieist" who preferred humans over future digital life forms [...] Musk said to Page at the time, "Well, yes, I am pro-human, I fucking like humanity, dude."

Imagine the mind set where this is not a pot fueled friendly banter, but actually a more and more heated argument. Maybe it was blown out of proportion? When Page bought DeepMind, Musk approached DeepMind's founder Demis Hassabis to convince him not to take the offer. "The future of AI should not be controlled by Larry," Musk told Hassabis.

(I don’t quote this to praise Musk, him being humanities champion frightens me a bit, but the misanthropic outlook Effective Accelerationists have.)