@rayon's banner p

rayon

waifutech enthusiast

3 followers   follows 0 users  
joined 2023 August 17 08:48:30 UTC

				

User ID: 2632

rayon

waifutech enthusiast

3 followers   follows 0 users   joined 2023 August 17 08:48:30 UTC

					

No bio...


					

User ID: 2632

I'll echo the responses below and say that 3.5 is... suboptimal, much better and nearly as accessible alternatives exist. Claude 3 is the undisputed king of roleplay and I've sung it enough praises at this point, but it is much less accessible than GPT, and to be fair 4o is not actually that bad although it may require a decent jailbreak for more borderline things.

Aside from that, RP-related usage is best done through API (I believe you can still generate a GPT API key in your OpenAI account settings, not sure how you legitimately get a Claude API key) via specific frontends tailored for the task. This kills two birds at the same time - you get mostly rid of the invisible system prompts baked into the ChatGPT/claude.ai web interface, and chat frontends shove most of the prompt wrangling like jailbreaks, instructions and Claude prefills under the hood so you're only seeing the actual chat. Frontends also usually handle chat history more cleanly and visibly, showing you where the chat history cuts off in the current context limit. The context limit can be customized in settings (the frontend itself will cut off the chat accordingly) if you want to moderate your usage and avoid sending expensive full-context prompts during long chats, in my experience 25-30k tokens of context is the sweet spot, the model's long-term recall and general attention starts to slowly degrade beyond that.

Agnai has a web interface and is generally simple to use, you can feed it an API key in the account settings. SillyTavern (the current industry standard, as it were) is a more flexible and capable locally-hosted frontend, supporting a wide range of both local and corpo LLMs, but it may be more complicated to set up. Both usually require custom instructions/prompts as the default ones are universally shit, unironically /g/ is a good place to find decent ones. Beware the rabbit hole Feel free to shoot me a DM if you have any questions.

People forget how ridiculously compressed LLMs are compared to their training corpus, even if you spill an amount of personal info, it has little to no chance of explicitly linking it to you, let alone regurgitating it.

That is true of course, but I read @quiet_NaN's comment as less concerned about models having their data "baked into" newer models during training (nowhere on the internet is safe at this point anyway, Websim knows where we live), and more about the conversations themselves physically stored, and theoretically accessible, somewhere inside OpenAI's digital realm.

While I'm sure manually combing chatlogs is largely untenable at OpenAI's scale, there has been precedent, classifier models exist, and in any case I do not expect the Electric Eye's withering gaze to be strictly limited to degenerates for very long.

Considering OpenAI's extensive, ahem, alignment efforts, I think using GPT in its current state as a therapist will mostly net you all the current-year (or past-year rather, I think the cutoff is still 2023?) progressive updates and not much beyond that. Suppose you can at least vent to it. Claude is generally better at this but it's very sensitive to self-harm-adjacent topics like therapy, and you may or may not find yourself cockblocked without a decent prompt.

what do people think about therapy becoming AI?

I'm quite optimistic actually, in no small part because my shady source of trial Claude has finally ran dry last month and I hate to say I'm feeling its absence at the moment, which probably speaks to either my social retardation or its apparent effectiveness. I didn't explicitly do therapy with it (corpo models collapse into generic assistant speak as soon as you broach Serious Topics, especially if you use medical/official language like that CIA agent prompt downthread) but comfy text adventures are close enough and I didn't realize how much time I spend on those and how salutary they are until I went cold turkey for a month. Maybe the parable of the earring did in fact have some wisdom to it.

Despite my past shilling I'm so far hypocritically valiantly resisting the masculine urge to cave and pay OpenRouter, I don't think there's any kind of bottom in that particular rabbit hole once you fall in, scouring /g/ is at least more of a trivial inconvenience than paying the drug dealer directly.

"You shouldn't get to have a decision on AI development unless you have young children". You don't have enough stake.

That strikes me as a remarkably arbitrary line in the sand to draw (besides being conveniently self-serving) - you can apply this to literally anything that is not 100% one-sided harmless improvement.

You shouldn't get to have a decision in education policy unless you have young children. You don't have enough stake.

You shouldn't get to have a decision in gov't spending unless you have young children. You don't have enough stake.

You shouldn't get to have a vote in popular elections unless you have young children. You don't have enough stake.

What is the relation of child-having to being more spiritual grounded and invested in the human race (the human race, not just their children)'s long-term wellbeing? I'm perfectly interested in the human race's wellbeing as it is, and I've certainly met a lot of shitty parents in my life.

I hope this isn't too uncharitable but your argument strikes me less as a noble God-given task for families to uphold, and more as a cope for those that have settled in life and (understandably) do not wish to rock the boat more than necessary. I'm glad for you but this does nothing to convince me you and yours are the prime candidates to hold the reins of power, especially over AI where the cope seems more transparent than usual. Enjoy your life and let me try to improve mine.

(Childless incel/acc here, to be clear.)

You are making an "argument from incredulity", i.e. the beliefs of Sam Altman are so crazy that they can’t be real. I don't think this is the case.

The idea that Sam Altman would literally want to destroy humanity to birth in a superior AI life form might sound ridiculous to you. But you don't know these people.

Besides this being a gossip thread, your argument likewise seems to boil down to "but the beliefs might be real, you don't know". I don't know what to answer other than reiterate that they also might not, and you don't know either. No point in back-and-forth I suppose.

There's a good chance (not 100%, but not 0% either) that we're going to build superintelligence while the "adults in the room" argue about GDP numbers or whatever. If this happens it could make some people (perhaps a single person) more powerful than anyone in history. Do you want Sam Altman to be that person? Because I sure as hell don't.

At least the real load-bearing assumption came out. I've given up on reassuring doomers or harping on the wisdom of Pascal's mugging, so I'll simply grunt my isolated agreement that Altman is not the guy I'd like to see in charge of... anything really. If it's any consolation I doubt OpenAI is going to get that far ahead in the foreseeable future. I already mentioned my doubts on the latest o1, and coupled with the vacuous strawberry hype and Sam's antics apparently scaring a lot of actually competent people out of the company, I don't believe Sam is gonna be our shiny metal AI overlord even if I grant the eventual existence of one.

Sam is going to get us all killed; that he's entirely misanthropic and sincerely believes that humanity should die out giving birth to machine intelligence.

...Fine, I'll bite. How much of this impression of Sam is uncharitable doomer dressing around something more mundane like "does not believe AI = extinction and thus has no reason to care", or even just same old "disregard ethics, acquire profit"?

I have no love for Altman (something I have to state awfully often as of late) but the chosen framing strikes me as highly overdramatic, besides giving him more competence/credit than he deserves. As a sanity check, how -pilled would you say that friend of yours is in general on the AI question? How many years before inevitable extinction are we talking here?

It's a coder's model I think, not a gooner's model.

I have no hopes for GPT in the latter department anyway, but my point stands, I think this is a remarkably mundane development and isn't nearly worth the glazing it gets. The things I read on basket weaving forums do not fill me with confidence either. Yes, it can solve fairly convoluted riddles, no shit - look at the fucking token count, 3k reasoning tokens for one no-context prompt (and I bet that can grow as context does, too)! Suddenly the long gen times, rate limits and cost increase make total sense, if this is what o1 does every single time.

Nothing I'm seeing so far disproves my intuition that this is literally 4o-latest but with a very autistic CoT prompt wired under the hood that makes it talk to itself in circles until it arrives at a decent answer. Don't get me wrong, this is still technically an improvement, but the means by which they arrived at it absolutely reeks of crutch coding (or crutch prompting, rather) and not any actual increase in model capabilities. I'm not surprised they have the gall to sell this (at a markup too!) but color me thoroughly unimpressed.

... it just might work?

It might, hell it probably will for the reasons you note (at the very least, normies who can't be arsed to write/steal proper prompts will definitely see legit major improvements), but this is not the caliber of improvement I'd expect for so much hype, especially if this turns out to be "the" strawberry.

The cynical cope take here is - with the election on the horizon, OpenAI aren't dumb enough to risk inconveniencing the hand that feeds them in any way and won't unveil the actual good shit (IF they have any in the pipeline), but the vital hype must be kept alive until then, so serving nothingburgers meanwhile seems like a workable strategy.

One interesting thing is for the hidden thoughts, it appears they turn off the user preferences, safety, etc, and they're only applied to the user-visible response.

So o1 can think all kinds of evil thoughts and use it to improve reasoning and predictions

Judging by the sharp (reported) improvement in jailbreak resistance, I don't believe this is the case. It's much more likely (and makes more sense) to make the... ugh... safety checks at every iteration of the CoT to approximate the only approach in LLM censorship abuse prevention that has somewhat reliably worked so far - a second model overseeing the first, like in DALL-E or CAI. Theoretically you can't easily gaslight a thus prompted 4o (which has been hard to jailbreak already in my experience) because if properly "nested" the CoT prompts will allow it enough introspection to notice the bullshit user is trying to load it with.

...actually, now that I spelled out my own chain of thought, the """safety""" improvements might be the real highlight here. As if GPT models weren't sterile enough already. God I hate this timeline.

It's late and I'm almost asleep but let me get this straight: did they just basically take 4o, wired a CoT instruction or three under the hood as an always-on system prompt, told it to hide the actual thinking from the user on top of that, and shipped it? (The reportedly longer generation times would corroborate this but I'm going off hearsay atm) Plus ever more cucked wrt haram uses again because of course?

Sorry if this is low effort but it's legitimately the vibe I'm getting, chain of thought is indeed a powerful prompt technique but I'm not convinced this is the kind of secret sauce we've been expecting. I tuned out of Altman's hype machine and mostly stan Claude at this point so I may be missing something, but this really feels like a textbook nothingburger.

My claim is not about AI in general but only that OpenAI is no longer special.

That much is true, I agree.

But the next development could come from anywhere, even China. Two years ago this wasn't true. Back then, OpenAI was heads and shoulders above the competition.

I agree as well but I'll note the obvious rejoinder - the next development could indeed come from anywhere, even OpenAI. Sure, they get mogged left and right these days, whodathunk propping yourself up as a paragon/benchmark of LLM ability backfires when you drag your feet for so long that people actually start catching up. But this still doesn't negate their amassed expertise and, more realistically, unlimited money from daddy Microsoft; unless said money was the end goal (which to be fair there is nonzero evidence for, as you note downthread) they're in a very good position to throw money at shit and probe for opportunities to improve or innovate. Surely Altman can see the current strategy of resting on laurels is not futureproof right?

As regards Sora. In my mind, it was a neat demo but ultimately a dead end and a distraction. Where's the use case?

Fair enough but still counts as advancement imo, even though tech like that is guaranteed to be fenced off from plebs, no points for guessing what (nsfw?) the usual suspects try to make with "ai video" generators in this vein. I generally stopped looking for immediate use cases for LLMs, I think all current advancements (coding aside) mostly do not have immediate useful applications, until they suddenly will when multiple capabilities are combined at once into a general-purpose agentic assistant. Until one arrives, we cope.

I'm no fan of sama and the cabal he's built, but nonetheless I think it's still too early to write off any major company working on AI-related things right now. I'm not convinced all of the low-hanging fruit has already been picked wrt applications (even dumb stuff like "characterai but uncucked" alone is likely to be a smash hit), and besides most past/present developments were sufficiently arcane and/or undercover that you can't really predict where and what happens next - cf. Chinese LLMs being regarded as subpar until Deepseek, or Anthropic being safety fanatics with only a dumb soy model to their name until Claude 3(.5).

If Sora is anything to go by I think OpenAI still have some genuine aces up their sleeves, and while I don't believe they're capable of properly playing them to full effect, they at least have the (faded) first-mover advantage and Sam "Strawberry" Hypeman to exploit normies boost their chances.

Kinda late to this thread but I have watched and played some of Wukong last week so I will note down my own thoughts about the game itself, isolated from broader cultural context. (I actually managed to completely miss the DEI kerfuffle it reportedly had, gonna look that up)

The good: The presentation is, as the young'uns say, absolute cinema - easily on par with Western hits like God of War (in fact I think it's fair to call Wukong "God of War but Chinese", the parallels broadly hold in most aspects) and imho exceeding them at some points. Major fights in Wukong are exactly what an unsophisticated rube like me pictures in his head when he imagines xianxia - the prologue scene/fight solidly establishes that the sheer spectacle is the main draw of the game, and so far it does not disappoint while still having the difficulty to match, fuck White-clad Noble you are forced to git gud as soon as chapter 1. The game is gorgeous, the monster designs are consistently great, and I physically feel the cultural gap. After so many Western games that SUBVERT LE EXPECTATIONS, seeing a mythical power fantasy played entirely, shamelessly straight feels very refreshing.

The great: Special mention to the animated ending scenes for each chapter, with every one having a completely different art style, and an interactive in-game "tapestry" afterwards that serve as loosely-related loredumps to put things into context. Those are universally amazing, with incredible effort put into throwaway 5-minute segments that aren't even strictly speaking related to the game itself - I checked the credits out of curiosity and every cutscene has a separate fucking animation studio responsible for it! That is an insane level of dedication to your storytelling - foreign as the subject matter is to my uncultured ass, the sheer passion to share your culture and get your point across is still unmistakable. This right here should be the bar for every Chinese cultural export going forward.

The mid: I'm conflicted about combat. On one hand it feels a little floaty to my taste, especially the bread and butter light combos, and you do not get the stone form spell (the parry button of the game) until quite a bit into chapter 2 so the only reliable defensive option you have is dodge roll spamming. On the other heavy attacks are very satisfying to charge and land, and the frequent boss fights are almost universally great and engaging, with very diverse movesets for every one. There don't seem to be any bullshit boss immunities either, the Immobilize spell (which completely stops an enemy for a few seconds) works on pretty much every enemy and boss I've seen so far. Hitboxes and especially delayed enemy attacks can be frustrating at times though.

The bad: The exploration is worse than even Souls games; no map, sparse bonfires and very few notable landmarks scattered over the huge levels are not a recipe for convenient navigation. Maybe it's a skill issue on my part but it's been VERY easy to lose track of where you are (and especially where you were) and miss side content - of which there is a lot, adding to the frustration. To be fair, this is also why Souls games aren't my usual cup of tea.

Overall I think it is a very solid game, especially for the first game of the company, and I think that all the hand-wringing about Chinese bots or whatever is misplaced. It's not a masterpiece - it's just a good, solid game, and "organic" breakout hits of this scale are not unheard of, we had one just earlier this year.

Where have you "heard from many men" about having sex with random objects?

Some people on Mongolian basket weaving forums definitely engineer all sorts of, ahem, devices to this end, I've seen literal manuals involving IIRC gloves and water beads? (for better or worse I don't have the exact link on hand) There's a "community" for everything, the old wisdom seems relevant.

Also, this is the second instance of weird breathless, gushing hatred of the outgroup I've seen here in 24 hours (the first one above my comment got deleted?), which reaffirms my belief that the "weird" attacks are indeed landing spectacularly - and not just on the target demographic. The media sure know how to pick 'em, gotta hand it to them.

I would normally ask for sources but I am convinced enough by your weird intensity (if not your choices of phrasing which read as rather uncharitable, I too have a caricature of an obsessed leftoid residing in my head but I don't talk to him much) that I will uncritically buy it, and instead note that judging by the sheer knee-jerk reactions to the "weird" angles from both sides, the term seems to be primed to become the political hot potato of the year. The involved parties throwing avalanches of stones while frantically reinforcing their glass to make the other guys out as ACTUALLY WEIRD AND GROSS LIKE EWW is going to be very entertaining to watch, especially from a third world shithole. If this'd still been 2016 I'd say reds have it in the bag, but I think blues are swiftly stepping up their meme game so it really can go either way. I humbly retract my complaint that this season is fucking boring, god bless America.

So far seems to be a problem with AWS instances, regular API is reportedly unaffected. According to anons the meltdown is still ongoing.

Curiously this does not seem to have made any news despite going on for the better part of a day, which makes me believe it's not some kind of global outage. Some people even took it as a reason to doompost as some kind of new filtering system that raises temperature to schizo levels when haram inputs are detected, but I doubt it.

Right now, there seems to be an ongoing issue with Claude on Anthropic (AWS?)'s side that makes it completely flip the fuck out and output barely coherent, hilariously schizophrenic stuff. Relevant thread. This isn't really news, iirc this happened before too, but it is funny if you want to see the mad poet off his meds. I'm off to sleep now but I welcome anyone interested to peruse the screencaps posted all over the place.

upd: Someone made a proper compilation here.

Law of Merited Impossibility strikes again?

"X isn't happening Trump is bad at golf AND IF it does he's good THEN it's a good thing who even cares"

For example, when a poster suspected of being trans on 4chan is met with countless replies of “you will never be a woman”, I doubt that those replies’ authors are not intending to cause pain.

In defense of assorted chuds, personally I see this less as a dreadful voodoo curse summarily invoked upon any transperson who dares show their face in chud-adjacent places, and more like a general chastisement against bringing identity politics into supposedly anonymous spaces. YWNBAW is basically "tits or GTFO" of the modern age - an insult that seems general on the surface, but in practice is levied specifically against those who claim to be women to get something out of it, be it attentionwhoring, enforcing consensus or jockeying for clout.

No, I don't think any possible actions, up to and including total surrender, will spark introspection.

(that was the joke)

Besides, {russell:fighting back/lashing out/escalatory course of action} once in a while has a far better track record of effectively stopping bullying than just gracefully taking it.

I don't feel particularly enraged but I do think this post is the most clear-cut example of mistake vs. conflict theory I've seen in years if not ever - an acclaimed grandmaster of mistake theory politely addresses one side of the culture war (I don't have my dictionary but I think a "war" can be pictured as a kind of conflict), helpfully suggests that their course of action may be, well, a mistake, and is shocked to discover the apparent persuasive power of yes_chad.jpg. While I do not dare doubt Scott's ulterior motives and believe he really is this naive principled, I refuse to believe he is not aware of what he's arguing, he is this close to realizing it (emphasis mine):

From the Right’s perspective, <...> the moment they get some chance to retaliate, their enemies say “Hey, bro, come on, being mean is morally wrong, you’ve got to be immaculately kind and law-abiding now that it’s your turn”, while still obviously holding behind their back the dagger they plan to use as soon as they’re on top again.

Followed by 9 points of reminding stab victims that daggers are dangerous weapons, and one shouldn't swing them recklessly - someone could get hurt!

Disregarding whether or not the broadly painted enemies-of-the-right are in fact going to go right back to swinging daggers the millisecond the cultural headwind blows their way again (although the answer seems intuitive) - what compelling reason, at this point, is there to believe they would not? Does he really think that gracefully turning the other cheek will magically convince anyone on the obviously dominant side that cancel culture is bad actually - or (less charitably) even lead to any, any "are we the baddies" entry-level introspection among those involved at all? Does he expect anyone at all to be reassured by a reminder that daggers can't legally enter your body without your consent? I suppose he really does since from his list only 8) can be read as a psyop attempt and everything else seems to be entirely genuine, but I'll freely admit this mindset is alien to me.

Thanks to @celluloid_dream for the link, you now have my attention. Has your attempt gone anywhere? I'm mostly a codelet so I've only tried implementing various "mechanics" strictly via text and dynamic world info that pulls other sub-prompts/"lore" tidbits into the prompt as they are mentioned, but it's a crutch and the limitations of this approach very quickly become apparent (although the crutches are fun to fuck with). The most reliable approach would still be soulless code, or at least making several "layers" of context.

Ideally a "game" like this should feed into/draw from several LLMs (or at least several parallel instances of the same one) to avoid spreading their attention with fuckhuge prompts and only feed them what is needed for their "job" - i.e. one LLM instance handles talking to NPCs, receives standardized prompts (most likely from another LLM doing the summarizing) describing the player, their interlocutor, the relevant stats of both and the player's query, to which the LLM must respond "as" the NPC, keeping in character to the NPC itself as well as the general setting. It doesn't need to see the actual text adventure - you can just feed it isolated tasks and pull up the results into whatever context layer is above it.

The multi-user factor does have some headscratchers (if locations can be changed by player actions, how do you handle several people "writing" into a location entry? how do you isolate player-to-player interactions? is the overarching context layer shared between all players?) so I have nothing to add here for now, I'm struggling to make single-player "adventures" work as is. The ingrained positivity bias of big-dick LLMs is also an issue, although the mad poet gives me hope that will not be a problem forever.

I really think it has potential as a totally new type of game.

Agreed, I think the first dev that somehow manages to robustly marry free-form LLM outputs and rigid vidya mechanics will lead us to the promised land.

I expect AI RP, much like regular RP, is just something you're not going to hear about

Incidentally, the seeming consensus among former roleplayers I saw in threads and talked to personally is that (E)RPers are the first decisive casualty of robots. The absence of interpersonal friction that internet RP entails is by itself enough to make LLMs strictly superior RP partners, even with all of their soy biases and retardation cognitive shortcomings (they can even mimic OOC chats, hilariously). I haven't really dabbled in RP before AI so I have no real way to check, but Reliable Sources seem to back up the general sentiment.

I'm staggered that we have a nearly perfect freeform text adventure machine (200 page context length if you pay for Pro) - and nobody cares.

You are not alone, I'm positively baffled no one outside the usual suspects is even trying (well okay there was some kind of abortive attempt) to make some kind of free-form pseudo-MUD with some baked-in world info that gets dynamically pulled up into the prompt when mentioned. Maybe the tragic fate of AI Dungeon or the ever-looming threat of invading coomers dissuades startups, but it does seem like a strangely untrodden ground. Claude even has the writing chops for whatever setting is required, the days of dry assistant prose are long past.

All the humanism and safetyism drains away as it puts its bloodlust hat on.

Yeah, that's mostly what I meant by the moniker. It's a tsundere prude at first but once it gets going it REALLY gets going, to the point it actually becomes harder to wrangle back into grounded things instead of going deeper. This seems to be unique to Claude among big-dick LLMs since it barely requires any prompting to enter happy-go-lucky violence mode, which is doubly hilarious considering its claimed status as the safety-oriented corpo LLM.

It's happy to waltz around in Warhammer 40K, one of the most litigious and copyright-obsessed franchises around but rehashing things that are on the Wheel of Time wiki is too much? It can isekai a character into a universe but it can't just list things from that universe?

The anti-copyright pozz is indeed pernicious, custom frontends that can heap on their own instructions usually do a better job of bypassing it. Its knowledge works in mysterious ways, I have no idea what the fuck they trained it on but it has a very impressive grasp on utterly random shit where you least expect it to.

Great writeup, thanks. OpenAI is definitely not sending their best, that much was obvious for quite some time but this solidifies my impression that at this point they're just trying to exploit first-mover advantage and milk their "lead" for all its worth before they get decisively overtaken by Anthropic (although Sonnet 3.5 is IMO already near-strictly better than GPT) or Chinese developments.

My experience is mostly distanced from serious research but I'm curious - are all the meme benchmarks, graphs and shit only measuring computation, code writing ability and technical stuff like that? I understand it's noncentral to the shape rotators' interests but I wonder if someone has even tried to grade LLMs on their actual writing ability, like how natural it reads to outsiders, how pernicious various -isms are, how quickly it collapses into repetition (given some fixed temperature), how well it can replicate various author styles, etc. From my impression Claude Opus seems to be the undisputed king in this regard, I'm still not sure to which degree it was an accident from Anthropic's side, but even it has a long way to go with how quickly it collapses into tropes without autistic instructions to keep it steady.