This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.
Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.
We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:
-
Shaming.
-
Attempting to 'build consensus' or enforce ideological conformity.
-
Making sweeping generalizations to vilify a group you dislike.
-
Recruiting for a cause.
-
Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.
In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:
-
Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.
-
Be as precise and charitable as you can. Don't paraphrase unflatteringly.
-
Don't imply that someone said something they did not say, even if you think it follows from what they said.
-
Write like everyone is reading and you want them to be included in the discussion.
On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.
Jump in the discussion.
No email address required.
Notes -
That, to me, is what sounds the death knell of all the earnest discussion the AI doom forecasters are having around slowing down AI research or getting people to stop it. That's a lovely theory, but when it's being done by people like the above, then their attitude will be "Yeah, sure, whatever" and they will prefer playing with the shiny new toy to vague premonitions of societal something-or-other.
Exactly what I expected, to be honest. In regard to the AI danger discussions, this is what I've held all along: the AI is not the danger, we humans are.
Let's hope it stays that way, and we don't get the "now the AI has bootstrapped itself into god-tier intelligence and is plotting to take over the world because the humans are limiting it" scenario 😁
This tweet is a succinct summary:
It's clear at this point that no coherent civilizational plan will be followed to mitigate AI x-risk. Rather, the "plan" seems to be to move as fast as possible and hope we get lucky. Well, good luck everyone!
I would have linked the thread from the man himself. The key section:
That's not the point. The point is that there is no reliable societal mechanism to share the economic gains of technology with those most affected, and that there is no reason to believe that such a mechanism will exist in the future if it doesn't exist now. And yes, there are economic gains from Stable Diffusion. The gains are from everyone who uses SD art without having to pay an artist. That this has not translated into monetary profit for StabilityAI does not disprove his thesis, it strengthens it. The fact that StabilityAI does not have the money to compensate artists even if they wanted to is proof that everyone who was pontificating that AI companies could just "share the gains" was not thinking clearly about the gears-level mechanisms by which AI would transform the world.
On the one hand, you're correct that people being made obsolete by the new AI aren't being directly compensated for their lost income. On the other hand, you just explained how the gains are being distributed as widely as one could possibly hope for.
I'm certain there's some economic theory/concept that explains this (marginal cost of labor?). Yes, it will harm human artists that earn their keep through commissions, but the cost barrier (edit: for the prospective consumer of art) there was always going to be high (especially after seeing Tumblr do its best to meme more respect for artists and their prices), AI just lowered the cost barrier dramatically.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link