site banner

Culture War Roundup for the week of September 30, 2024

This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.

Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.

We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:

  • Shaming.

  • Attempting to 'build consensus' or enforce ideological conformity.

  • Making sweeping generalizations to vilify a group you dislike.

  • Recruiting for a cause.

  • Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.

In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:

  • Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.

  • Be as precise and charitable as you can. Don't paraphrase unflatteringly.

  • Don't imply that someone said something they did not say, even if you think it follows from what they said.

  • Write like everyone is reading and you want them to be included in the discussion.

On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.

3
Jump in the discussion.

No email address required.

This is an unfalsifiable theory. If there is Russian interference, hey, wow, I was right. If there's not, well, whatever, I was just being careful, and it's always good to be careful.

Russian social media campaigns being in any way influential is extremely implausible. Whatever they might be spending would be a drop in the bucket relative to what Americans spend on social media all the time. That has been the case every time a number is attached to whatever Russia is supposedly spending.

Did he claim they were influential, or was he claiming a style?

If he's claiming a style, then that would actually be falsifiable, by establishing a different style is what is actually pursued.

When the style claimed is "increases discord", it's indistinguishable from internal partisans who are unhappy with the current state of affairs, and post their (discordant) opinions on social media.

I guess this is falsifiable if you found some russian operatives posting so as to... increase harmony, but this seems unlikely, and I can't really visualize what "increase discord" looks like on the other end. "Here's some rubles, go stir the shit on twitter"? Government propaganda campaigns always have some sort of goal in mind IME -- it used to be "promote global communism", but what is it now?

When the style claimed is "increases discord", it's indistinguishable from internal partisans who are unhappy with the current state of affairs, and post their (discordant) opinions on social media.

Absolutely. Or at least, almost indistinguishable. There are occasionally tells- for example, intermixing the awkward fixing of things an internal partisan wouldn't care about that happens to align with a foreign propaganda interest (plenty of Americans don't like the idea of fighting China over Taiwan, but only a minute number do so on grounds of appeals to the Century of Humiliation narrative)- but often it is indistiguishable.

This is why I'm fully sympathetic to people whose ideological immune system is flaring in suspicion.

I guess this is falsifiable if you found some russian operatives posting so as to... increase harmony, but this seems unlikely, and I can't really visualize what "increase discord" looks like on the other end. "Here's some rubles, go stir the shit on twitter"?

Unironically pretty close to that.

One of the origins of the modern Russian troll factory is that one of the more notorious- the Information Research Agency- was founded by Yevgeny Prigozhin. Yes, the Wagner Mercenary guy. Prigozhin was basically somewhere between a front, a fence, and a semi-autonomous vassal of Putin's security establishment. The distinction is that not only did he do what he was told, but he had a degree of freedom to try initiatives on his own. This was/is part of Putin's power structure, where inner-circle elites compete for power and influence and attention... and one of the ways is to do something impressive. Or, in Prigozhin's case, something that appeals to Putin's spy-mentality, while also serving as an excuse to charge the Russian government for services rendered. Other elites began copycatting later, and the American reaction probably justified the investment in Russian views, but IRA was the first (until it's dismantling / repurposing after the Wagner Coup and Prigozhin's assassination).

The IRA began in 2013, and by 2015 it had a reported ~1000 people working in a single building. One of its earlier claims to notice, before the 2016 election and compromise of American political discourse on that front, was back in 2014 when Russia was trying to recalibrate international opinion on its post-Euromaidan invasion of Ukraine. Buzzfeed published some leaked/stolen IRA documents, including a description of daily duties.

To quote-

"Foreign media are currently actively forming a negative image of the Russian Federation in the eyes of the global community," one of the project's team members, Svetlana Boiko, wrote in a strategy document. "Additionally, the discussions formed by comments to those articles are also negative in tone.

"Like any brand formed by popular opinion, Russia has its supporters ('brand advocates') and its opponents. The main problem is that in the foreign internet community, the ratio of supporters and opponents of Russia is about 20/80 respectively."

So how does one counter that narrative mismatch?

The documents show instructions provided to the commenters that detail the workload expected of them. On an average working day, the Russians are to post on news articles 50 times. Each blogger is to maintain six Facebook accounts publishing at least three posts a day and discussing the news in groups at least twice a day. By the end of the first month, they are expected to have won 500 subscribers and get at least five posts on each item a day. On Twitter, the bloggers are expected to manage 10 accounts with up to 2,000 followers and tweet 50 times a day.

...

The trolls appear to have taken pains to learn the sites' different commenting systems. A report on initial efforts to post comments discusses the types of profanity and abuse that are allowed on some sites, but not others. "Direct offense of Americans as a race are not published ('Your nation is a nation of complete idiots')," the author wrote of fringe conspiracy site WorldNetDaily, "nor are vulgar reactions to the political work of Barack Obama ('Obama did shit his pants while talking about foreign affairs, how you can feel yourself psychologically comfortable with pants full of shit?')." Another suggested creating "up to 100" fake accounts on the Huffington Post to master the site's complicated commenting system.

And how does one fund that?

The trolling project's finances are appropriately lavish for its considerable scale. A budget for April 2014, its first month, lists costs for 25 employees and expenses that together total over $75,000. The Internet Research Agency itself, founded last summer, now employs over 600 people and, if spending levels from December 2013 to April continue, is set to budget for over $10 million in 2014, according to the documents. Half of its budget is earmarked to be paid in cash.

So, yes. "Here's some rubles, go stir the shit on twitter" is unironically close to what happened. Reportedly.

And this was back in 2014, when it was still very new and immature as an institution. As internet social media technologies evolved, so did the Russian technical infrastructure and incorporation into information warfare theory, which itself evolved. Note that IRA in the early days functioned as a more message-focused concept (a russian position). However, other parts of the Russian information-proxy sphere were decentralized and took other, even contradictory stances- most notable to western observors in the pro-wagner vs pro-MOD narrative wars before the Wagner Coup.

If you'll forgive an unrepentantly NATO-based analysis, the Irregular Warfare Center has a pretty comprehensive analysis of how the Russian information efforts has evolved over time.

Government propaganda campaigns always have some sort of goal in mind IME -- it used to be "promote global communism", but what is it now?

Other models of propaganda include making you want to buy something (advertisement), go to a specific church (missionary work), think favorably of a specific cause or subject (advocacy), think worse of a specific cause (defamation),undercut a subject's moral authority (deligitmization), spread a cultural viewpoint (normalization), and so on.

For a more typical model, China's propaganda apparatus is much more focused on specific topics where it wants you have a specific position, such as a good view on Xi, the CPC, multipolarism, etc, while having no particular stance and spending no particular effort on others. Arguing both sides of an argument is rarely done, because point of propaganda is seen as to persuade / push to a certain perspective, and playing both sides at the same time is generally seen as information fratricde countering your own efforts. When confusion is the point, it can be pursued, but these are shorter-term and generally the exception rather than the norm. To a degree this is itself a measure of centralization- the Chinese government has a stronger message control over its directly employed propagandists than the Russians imposed on their associated blogosphere and elite-owned influencer networks.

A general 'increase discord by truth and fiction on any topic any time' motive is relatively rare as a result. Not only does that lead to contradictory themes, but doing so is a success on its own standing. Note how Russian sources fed both a source of anti-Trump narratives (the Steelle Dossier), and in anti-anti-trump narratives (social media boosting), or how in the Ukraine context Ukraine was simultaneously a NATO puppet controlled from abroad (attempting to generate nationalist resistance to foreign meddling against European liberalism) and a Nazi regime suppressing locals (a justification for foreign intervention to prevent an antithesis of European liberalism) . If the goal of propaganda was to actually enable a favored manchurian candidate or promote a foreign (Russian) intervention, this would be self-defeating, since you'd still be having primary state-propaganda persuasion of the classical model, but be actively undercutting it with more contradictory messaging.

An implication of this sort of model is not only is it cause-agnostic, but it can take both sides of the same argument at the same time- support Tribe A with social media via venue C, and Tribe B on the other stance with different media via venue D. (In a non-single-nation context, if you ever get the chance, look up the global conspiracy variations of 'who is to blame for COVID.' The US and China are not the only candidates claimed.) I've long since lost the articles, but a personal pet peeve back in the early Trump administration when the disinformation craze was at it's peak was how much of the coverage of 'Russian interference' in US politics didn't actually identify relative partisan themes being boosted.... because it was both Republican and Democratic themes.

Which, as you say, can be indistinguishable from partisan propaganda, even though it has a different intent.

and I can't really visualize what "increase discord" looks like on the other end. "Here's some rubles, go stir the shit on twitter"?

If you love what you do, you’ll never work a day in your life.

That would be even emptier. Be careful about what you see on social media, because it could have the same effect as Russian disinformation. That parses to something like: Look both ways before you cross the street, because a plane could fall on you.

Counter-point, "Remind yourself that overconfidence is a slow and insidious killer."

Which has the merit and utility of being actually useful advice. Overconfidence is a risk factor, and it can take a long time to take detrimental effect. You could dismiss the warning on the same grounds of falsifiability- if overconfidence does get you killed here then you were right and if it doesn't you're just being careful and careful is good- but this ignores that sustaining carefulness is an enduring good in and of itself.

This is a relatively common form of warning for harms that can come with unclear immediate impacts. Don't just eat mushrooms you find in a forest, they may be poisonous. Walk slower on just-mopped floors, they may be slippery. Don't trust strangers on the internet, they might be bad. The fact that these warnings don't have to come in a context where the element of danger is immediate or guaranteed doesn't make them non-falsifiable, and their value can come because the warned against function is rare. When an element of danger is rare, it's easy to ignore the possibility of something that could be prevented with diligence.

By contrast, 'look both ways because a plane could fall on you' has no link between cause of warning and effect of warning. Looking both ways does nothing to warn you of the danger that comes with 'up,' so there's no merit of dilligent reminder. It also an argument of a specific instance (planes crashing into crosswalks is so singular that it can't really be claimed as a trend) as opposed to a trend-consequence of mounting risks (overconfidence may not get you killed this time, but the reoccuring and persistent nature can lead the threat to grow over time).

Which simile is better for "the danger of the Russian style of disinformation" is up for debate, but I'd wager (and right) on the comparison to overconfidence than to airplanes-on-crosswalks.