site banner

Culture War Roundup for the week of December 4, 2023

This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.

Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.

We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:

  • Shaming.

  • Attempting to 'build consensus' or enforce ideological conformity.

  • Making sweeping generalizations to vilify a group you dislike.

  • Recruiting for a cause.

  • Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.

In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:

  • Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.

  • Be as precise and charitable as you can. Don't paraphrase unflatteringly.

  • Don't imply that someone said something they did not say, even if you think it follows from what they said.

  • Write like everyone is reading and you want them to be included in the discussion.

On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.

5
Jump in the discussion.

No email address required.

Your post puts my pause about supporting EA into better words than I possibly can write. I've always found it... cheating, kinda?... that the entire premise of EA seems to just be brute-forcing morality and ethics by shoving as many zeroes into a number as possible. And that's how we get what you have described here, where it's easy to say that you've saved 200,000 or however many lives, but then people don't interrogate that result beyond that. People just see the number and go "wow that's a lot of zeroes so it must be good".

I guess what I'm wondering is if there's much focus put in to long-term solutions (and I don't mean "longtermism" like figuring out how to get humans to colonize the stars or whatever to maximize the number of future lives) rather than just whatever saves the most lives in the short term. For example, I was always under the impression that you can't just brute-force solving world hunger by confiscating all the world's billionaires' wealth (ignoring the fact that much of it isn't liquid and actually kinda doesn't exist and if you confiscated it then most of the wealth would just go away) and funneling it into programs to distribute food to starving populaces (ignoring the fact that this would outcompete and devastate local markets, etc.). Sooner or later, their governments would stop you, because it turns out that the reason they're starving in the first place is because their government wants them to, and there's plenty of things the government can do to get their country in a place to feed them, but they don't for various reasons. So there's a good short-term solution by just distributing as much as you can, but an actual long-term solution requires some change to the government, and a lot of focus seems to be put on the short-term brute-force way of doing things.

Downthread, @FirmWeird posits a similar scenario where the population is way beyond carrying capacity. What do you do? Feeding them makes the line on a graph go up, if you ignore that this means you'll need even more in the future (induced demand). Not feeding them makes the line go down but it results in a more stable equilibrium. "Just shove as many zeroes in as you can" ignores plenty of side effects that may or may not be desirable, almost like a paperclip maximizer.

People just see the number and go "wow that's a lot of zeroes so it must be good".

In theory you can fix this by counting QALYs instead of lives saved.

Of course, counting QALYs meaningfully isn't easy, but it is easy to come up with bad ways to count them and hard to prove them bad.

Your post puts my pause about supporting EA into better words than I possibly can write. I've always found it... cheating, kinda?... that the entire premise of EA seems to just be brute-forcing morality and ethics by shoving as many zeroes into a number as possible. And that's how we get what you have described here, where it's easy to say that you've saved 200,000 or however many lives, but then people don't interrogate that result beyond that. People just see the number and go "wow that's a lot of zeroes so it must be good".

I have the exact opposite intuition about "cheating." To me, regular charities seem like cheating, since all most people do is give money, get warm and fuzzy feelings, decide that that means they did good, and then profusely refuse to actually verify how much good they did. It's essentially stolen valor - getting all the personal benefits of appearing altruistic without doing any of the difficult work of helping anyone out which is normally implied by things like "charity" and "altruism." Which looks a lot like cheating. EA at least seems to make a gesture at having some referee system in place to detect cheaters.

The thing is, as we see in international sports all the time like in FIFA or Olympics, the best way to cheat is to set the entire system of judgment itself as corrupted in your favor, so one could probably make a good argument that EA is performing this "meta-cheating" by claiming to actually be setting up objective standards for effectiveness while actually setting up corrupted standards that lead to [charity I like] being [effective]. The tough thing for EA is, every single person, down to the individual, involved in the EA movement could be perfectly transparent and honest with perfectly good intentions, and overall EA could still be engaging in this "meta-cheating" due to the biases that all people are susceptible to, and so they have some responsibility to set up the structure in such a way as to counter and negate these biases. I think they may be failing to do this properly.

But in terms of their attempt at brute-forcing morality, this seems to me the correct way to counter the massive "cheating" that's happening in basically all realms of altruism in our lives, even down to individual relationships, where most people don't bother checking and just get the beneficial warm and fuzzy feelings through "cheating" without doing any of the work that is supposed to make someone actually deserving of those warm and fuzzy feelings.