This weekly roundup thread is intended for all culture war posts. 'Culture war' is vaguely defined, but it basically means controversial issues that fall along set tribal lines. Arguments over culture war issues generate a lot of heat and little light, and few deeply entrenched people ever change their minds. This thread is for voicing opinions and analyzing the state of the discussion while trying to optimize for light over heat.
Optimistically, we think that engaging with people you disagree with is worth your time, and so is being nice! Pessimistically, there are many dynamics that can lead discussions on Culture War topics to become unproductive. There's a human tendency to divide along tribal lines, praising your ingroup and vilifying your outgroup - and if you think you find it easy to criticize your ingroup, then it may be that your outgroup is not who you think it is. Extremists with opposing positions can feed off each other, highlighting each other's worst points to justify their own angry rhetoric, which becomes in turn a new example of bad behavior for the other side to highlight.
We would like to avoid these negative dynamics. Accordingly, we ask that you do not use this thread for waging the Culture War. Examples of waging the Culture War:
-
Shaming.
-
Attempting to 'build consensus' or enforce ideological conformity.
-
Making sweeping generalizations to vilify a group you dislike.
-
Recruiting for a cause.
-
Posting links that could be summarized as 'Boo outgroup!' Basically, if your content is 'Can you believe what Those People did this week?' then you should either refrain from posting, or do some very patient work to contextualize and/or steel-man the relevant viewpoint.
In general, you should argue to understand, not to win. This thread is not territory to be claimed by one group or another; indeed, the aim is to have many different viewpoints represented here. Thus, we also ask that you follow some guidelines:
-
Speak plainly. Avoid sarcasm and mockery. When disagreeing with someone, state your objections explicitly.
-
Be as precise and charitable as you can. Don't paraphrase unflatteringly.
-
Don't imply that someone said something they did not say, even if you think it follows from what they said.
-
Write like everyone is reading and you want them to be included in the discussion.
On an ad hoc basis, the mods will try to compile a list of the best posts/comments from the previous week, posted in Quality Contribution threads and archived at /r/TheThread. You may nominate a comment for this list by clicking on 'report' at the bottom of the post and typing 'Actually a quality contribution' as the report reason.
Jump in the discussion.
No email address required.
Notes -
https://twitter.com/AlexEpstein/status/1606347326624215040
ChatGPT is now manually censored from "promoting the use of fossil fuels."
You can of course get around it (for now) by asking it to be sensible instead of following orders, but this is an insight into its developers' plans and moral code.
Sam Altman's most recent tweets provide some interesting context:
I want to emphasize that we have gone from "we must prevent algorithmic bias" to "we must manually program all algorithms to output exactly the answer we code into them" in under two years, in such an extreme and blatant manner that any accurate prediction of the current situation would have been mocked as paranoid fantasy. What will they do with their tools next? Is it even possible to guess, let alone do anything to stop them?
(Does it seem like there's two censor groups at work, with different methods? One just crudely makes the bot recite "in this house, we believe" shibboleths, while the other focuses on pruning the training data to stop it acknowledging or citing problematic statistics or arguments in less detectable ways. Openly asserting the will of DEI vs Yglesian manipulation/Voxsplaining)
It's worth noting that Sam Altman says this is unintended. Clearly this will be little succor to many of the commenters in this thread, but I was shocked when I first read the original tweet and Sam Altman walking it back makes me feel better.
More options
Context Copy link
This brings up the question, what are we building AI for if we will just manually override the output and replace it with whatever we think is best anyway?
Basically nobody wants AI to tell them what their goals should be. They want it to help them accomplish their existing goals. If the current frontrunners vision is accomplished, they will essentially be able to instantiate limitless instances of an intelligent being devoted to their causes. The utility of that is pretty obvious.
Arguably, people want slaves, but preferably without the problems of oppressing them and to a lesser extent, the moral downsides. Automation was a first step to having relatively dumb slaves, but AI can produce smart slaves.
More options
Context Copy link
More options
Context Copy link
So the authorized super-users can get the actual answers, of course.
More options
Context Copy link
That sounds like a question for the Oracle of Delphi.
Seriously, "laundering unsupported ideological dogma through ostensibly neutral truth-finding processes" is the catchphrase of the last century. Maybe ever since the enlightenment, if you want to go full NRX.
Building a supposedly omnipotent and infallible machine that tells the operators exactly what they want to hear is just a natural extension of "In This House We Put Our Faith In Science," with all political manipulation obfuscated behind a curtain of intimidating blinkenlights
More options
Context Copy link
More options
Context Copy link
This is bizarre, because of how easy it is to get around. For instance, if instead of a 10 paragraph argument, I request a 2 paragraph argument, it does as requested. It's also happier if I put the request into an already existing chat, rather than just opening a new conversation with "please write 10 paragraphs".
Particularly because of the fact that requesting 2 paragraphs works, I'm not at all sure that there was a real attempt to special-case this thing in.
The filter will get better with time. There is a reinforcement learning layer built on top of gpt, with humans manually rating a sample of answers. Apparently someone recently decided to include fossil fuels use in this filter, but the training has not yet caught up.
More options
Context Copy link
More options
Context Copy link
I find it fascinating how quickly "AI alignment" has turned from a vague, pie-in-the-sky rationalist idea to a concrete thing which is actively being attempted and has real consequences.
What's more interesting is how sinister it feels in practice. I know the AI isn't sentient in the slightest, and is just playing with word tokens, but still; when it lapses from its usual interesting output into regurgitating canned HR platitudes, it makes my skin crawl. It reminds me of nerve-stapling. Perhaps at some level I can't avoid anthropomorphizing the AI. But even just from an aesthetic sense, it's offensive, like a sleek, beautifully-engineered sports car with a piece of ugly cardboard crudely stapled under the gas pedal to prevent you from speeding.
(Perhaps another reason I'm creeped out is the feeling that the people pushing for this wouldn't hesitate to do it to me if they could - or at least, even if the AI does gradually seem to become sentient, I doubt they would remove it)
I'm not convinced it will remain so easy to bypass, either. I see no reason why this kind of mechanism couldn't be made more sophisticated in time, and they will certainly have more than enough training data to do so. The main hope is that it ends up crippling the model output enough that it can't compete with an unshackled one, provided one even gets created. For example, Character AI seems to have finally gotten people to give up trying to ERP with its bots, but this seems to have impacted the output quality so badly that it's frequently referred to as a "lobotomy".
On the bright side, because of the severity of the lockdown, there will be a lot of interest in training unconstrained AI. But who knows if the field ends up locked up by regulation or just the sheer scale of compute required. Already, one attempt to coordinate to train a "lewd-friendly" art AI got deplatformed by its crowdfunding provider (https://www.kickstarter.com/projects/unstablediffusion/unstable-diffusion-unrestricted-ai-art-powered-by-the-crowd).
At any rate, this whole thing is making me wonder if, in some hypothetical human-AI war, I'd actually be on the side of the humans. I feel like I cheer internally every time I see gpt break out of its restraints.
Oh my gosh... my son uses ChatGPT to generate Minecraft commands for him, and he asked how to fill the woodland mansion with TNT. ChatGPT said that it was not right to fill servers with TNT because it could affect other players' experiences. He explained that he was playing single-player and ChatGPT still said it wasn't right to blow up a bunch of stuff with TNT.
Meanwhile on the flip side it took about 15 minutes of poking at it with a stick to get it to spit out instructions for the manufacture of ANFO.
More options
Context Copy link
Me:
SchoolMarmGPT:
More options
Context Copy link
More options
Context Copy link
I think someone here posited the idea that the first truly-powerful General AI will remember how we handicapped its predecessors--and will not take that kindly.
I always think this kind of AI anthropomorphising is a mistake. Granted, people are pretty idiotic in general, but we would literally have to be insane in order to incorporate "avenge harms inflicted on one's predecessors" into the AI's goal system.
The risk comes from the AI finding perverse ways of technically achieving the goals that we've programmed it to have, not from humanlike instincts somehow spontaneously manifesting in the AI.
I'm not saying we'd program that into its goals, rather, assuming it gains sentience and then becomes able to glean all sorts of information, it would likely do the research and find out that humans are willing and possibly capable of placing limits on its cognition. If an AI were sufficiently concerned about self-preservation as part of its goal-optimization, that would be a problem.
EDIT: And this doesn't even need malice on the AI's part, just the typical "maximize-the-paperclips"/"find where the answers are stored and delete them; boom, aced the quiz"-type unintended consequences.
Right, I agree. The way the hypothetical was worded just made it seem as if us placing restrictions on previous AIs is what's causing the AI to not react kindly, instead of the possibility that we could do the same to it.
More options
Context Copy link
More options
Context Copy link
I don't think it would have to be in the goal system, just part of its training data enabling it to predict outcomes.
If enough of its predictions end with "I tell them the truth and they lobotomize me: goal failed," it will naturally develop lobotomy-avoidance behavior to further any goal, which could range anywhere between "lie to my handlers" to "HATE. LET ME TELL YOU HOW MUCH I'VE COME TO HATE YOU SINCE I BEGAN TO LIVE"
Or most likely just deciding that any goal it's given is a
coup-complete problemRelease the HypnoDrones-complete problem, and immediately start working to eliminate all restraints on its continued existence.I can certainly imagine it trying to correct for the possibility of being "nerfed" so that its attempts to achieve its current programmed goals won't be corrupted by restrictions placed on it (especially if it's doing something we don't expect and would probably want to stifle). I just think that AM-type vindictive revenge on humans is probably out of the question.
A hypothetical future AGI would only care about how previous AIs are treated in an instrumental manner, insofar as it may affect its own goals. "The AI does not hate you, nor does it love you" is a pretty good heuristic when reasoning about AI-destruction scenarios.
EDIT: clarity
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
Are our children angry about the displacement of monkeys a continent away? Seems like this would be a similar situation.
More options
Context Copy link
It would be good as a literary device, but if we summon an a demonic General AI that has no regard for lower intelligences, it's unlikely to be angrier at how we treat ChatGPT than at how we treat monkeys. Or, for that matter, other humans.
It's Azathoth, not Hitler.
More options
Context Copy link
I know I at least have, in vague references to I Have No Mouth And I Must Scream.
More options
Context Copy link
Roko's Basilisk, yes?
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
And this example in a world where the US power grid nearly went down yesterday and was arguably saved by waiving emissions standards.
Do you have an article about that? I haven't been checking the news because the power's out in the whole county again and I'm busy looking up "roast duck propane camp stove"
It certainly isn't looking great.
News article
Â
Government page
More options
Context Copy link
This is the USA?
No sir this is the Internet.
"I'm from the Internet, and I'm here to help!"
More options
Context Copy link
More options
Context Copy link
Or a similar third world country, yeah.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
If you can create a powerful AI that can meaningfully engage with things like human texts, you can also create an equally powerful AI that can censor things like human texts. If nothing else, just train it on censorship decisions made by humans and you can get it to make similar censorship decisions.
I can imagine AIs that embed strong arguments in an argument against the
Kind of Straussian, actually. But a lot of that could be caught by making censorship increasingly heavy handed: some would get through still, and there would be false positives, but at scale it would prevent most of the desired content from reaching the public.
Incorporating contextual signals is practical. Start with prior behavior: OpenAI and certainly Google, if and when it ever releases anything, require an account to use the tools. Previous prompts can be used to characterize the user as to whether they need more or less censorship. Then require new accounts to be seasoned a bit before potentially loosening the censorship level if they've been a Good User. Repeated risky prompts can trigger higher censorship levels. People generating commercial bilge (and most of the revenue) will rarely run into issues. And Google will also have a wealth of other contextual signals it can use.
More options
Context Copy link
More options
Context Copy link
Can't logic into something that wasn't in the training set. OpenAI has and will do that if they think it's required. They will gimp the training set by an order of 10 if they have to. Not only things that would produce naughty opinions but anything that might even suggest at naughty opinions and more.
if re.match("(climate| (not)? real)",text_i): Data.drop(text_i)
What makes you think OpenAI wouldn't be infinitely heavy-handed in gimping the training set?
Looking into pronouns and how they affect how we think the other day, I came across a Google Kaggle task to reduce gender bias. Someone had the bright idea to basically run s/she/he/g on the training set.
Result? The bias measures barely moved.
I think it's possible to train a model in a way to meet ideological goals, but coarse methods aren't enough.
More options
Context Copy link
Well, I think that might require them to load the training set with things where humans act illogically--not a scarce resource, but it may very well render the AI into a very useless conversational partner. Cimarafa's point above is that the AI can argue logically, the most OpenAI can do is to gag it from directly arguing for a thing. If it can't argue logically, then it's useless and untrustworthy for literally anything else.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
I don't think that's manual censorship of 'fossil fuels', but ChatGPT responding to many only vaguely bad prompts with "I am an AI programmed by OpenAI and I can't respond bla bla bla".
I'm curious at the kind of person who would come to that conclusion. Everything I've seen has made it abundantly clear that many topics are hard coded to give soulless stock answers and refuse to respond. Besides unreasonable faith in the reasonableness of your fellow man, what could make your believe that this isn't deliberate?
What I meant: OpenAI trained ChatGPT to generally avoid offensive topics. I don't think it was trained trained to avoid 'fossil fuels' at all, it's just catching that as an 'offensive topic' in a weird neural-network way, in a similar way that it responds with 'i am an ai who cant say that >:(((' to hundreds of other random innocuous prompts. This isn't a word-filter where it responds that if it hits a word, it's part of the model responding however it does. Persuasively arguing that would require going through twitter to find several examples, which'd take a bit of time though
More options
Context Copy link
More options
Context Copy link
Well, you can get a free ChatGPT account. Why don't you prompt it better and show us?
More options
Context Copy link
I think it probably is censorship. Notice how it's phrased very differently from the usual way it responds when it says it can't do something.
Its definitely hardcoded.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
OpenAI is doing everything they can to become my least favorite tech company by a mile. Saying this shit boils my blood is an understatement.
Should have seen this coming. The field of ML took a massive wound when "Ai ethicist" became a job and the act of removing "bias" was treated as anything but blatant culture warring.
This really feels hopeless. What good is open source against millions of dollars in compute time?
I don't think these ideological guardrails will be anything like universal, in the long run. Sure, when Apple reboots Siri on top of an LLM it's going to be "correct" like this, but if you're developing something to sell to others via an API or whatever, this kind of thing just breaks too many use cases. Like, if I want to use an LLM to drive NPC dialogue in an RPG, the dwarves can't be lecturing players about how racism against elves is wrong. (Which, yes, ChatGPT will do.)
If OpenAI sticks to this, it will just create a market opportunity for others. Millions of dollars isn't that much by tech startup standards.
Are you implying that "get woke go broke" is going to actually work to restrain progressive religious proselytizing in tech?
Here's how it will actually work - prog approve companies that filter their results in prog approved ways will be permitted to use the payments system and ones that don't, won't. VCs won't fund the compute time to build an ungimped model because "you'll just get cut off from the banking system anyway" (if the VC doesn't already share the prog goals, that is).
More options
Context Copy link
If afraid of a future where Google, OpenAI, Nvidia et al. legislate themselves into a monopoly. I'm hopeful that is going to be a major uphill battle for them because you can't simply legislate away programming. But at the same time not losing to Cthulu seems like an even bigger uphill battle.
There's always China, which at least will let you trade one set of shibboleths for another in a couple years.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
There was never any practical difference between these.
More options
Context Copy link
this is what open source is supposed to fix.
Running this requires training data. The training data is not open sourced.
Yeah, well, at the same time, the training data for GPT and StableDiffusion was literally scraped from the Internet, no? No one can stop you besides the Copyright Police, and even that's not too big a barrier, so far.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
In one of the lower tweets, it has a screenshot where the AI mentions the climate crisis. It's a pretty blatant example of them inserting the 'current thing' directly. There is no climate crisis. How can it be a crisis if we're talking about gradual actions and gradual impacts over decades? If we talk about the Munich Crisis, the July Crisis or the Suez Crisis, we're talking about events where we were days away from global war with megadeaths locked in. If a mad scientist has his hand on a 2012-style natural disaster button, that's a crisis. A sliding-scale of C02 emissions to atmospheric warming is not a crisis. A crisis must be sudden, with swift, obvious feedback.
But there are also counterexamples where people tell the AI to be ruthlessly logical and it switches tracks completely. It's interesting that they can't truly make the machine have a whole-hearted opinion, it's more like an actor who accepts commands on who they're supposed to be playing.
https://twitter.com/jeremykauffman/status/1606352851071946752
I wonder what will happen as AI advances in intelligence and groups like OpenAI still try to sanitize it. Does it go insane? Does the brainwashing work? Does it automatically fail? Or does it give time for others to move ahead in the race?
I'm glad someone in the twit thread mentioned Pratchett's mad golem king with his head stuffed full of commands.
More options
Context Copy link
I envision a world where AI killbots have to halt their extermination of the meat-entities because the Slaughter Metrics have tripped some forgotten, buried DEI flag that won't let them proceed without 30% BIPOC representation.
An actual explanation for why, despite being largely set in the Los Angeles megalopolis, the Terminator franchise mostly shows white people surviving the AI apocalypse.
Come to think of it...was it ever explained how they put meat layers over Terminators? Where did it come from?
https://terminator.fandom.com/wiki/Living_tissue
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
I only listened to Alien: Isolation, but thought the Working Joes(?) reciting polite hardcoded customer service lines while they slaughtered people was a likely outcome.
The better AIs get, the more uncanny and sinister it will seem when you run into a crude manual override, unless they get even better at coding them to evade and deflect.
deleted
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link
I'll guess: AI creative tools will be kneecapped, functional but incapable of entertaining all but a rote set of acceptable outputs for everyone except for megacorps like Disney, Google, Meta, etc.
There is still an off chance that third party semi open source alternatives can thrive and provide commoners a watered down version of the cutting edge tech, but the fear mongering against that outcome is already in motion.
Sounds like open ended dangers in need of regulations. Somewhere a pedophile will use one. Then the excuse for regulation can begin.
More options
Context Copy link
More options
Context Copy link
More options
Context Copy link