@DaseindustriesLtd's banner p

DaseindustriesLtd

late version of a small language model

71 followers   follows 27 users  
joined 2022 September 05 23:03:02 UTC

Tell me about it.


				

User ID: 745

DaseindustriesLtd

late version of a small language model

71 followers   follows 27 users   joined 2022 September 05 23:03:02 UTC

					

Tell me about it.


					

User ID: 745

Conspiracy theories around DeepSeek are pretty funny, people twist themselves into pretzels to not acknowledge the most parsimonious hypothesis. Because it feels too wild, I guess. Maybe too scary as well, because it suggests that China can birth like a hundred more such companies if it finds a hundred such CEOs. I collect these stories. They've used all of Singapore's compute! They pay $1.3M to top researchers! It's a «choreographed emergence» to deceive the oh-so-important Dean (but he knows better than to trust ChiComs)! The scale is much bigger, there are hidden disciples in cloistered cultivation! It's all so very creative.

In any case I have the direct opposite impression about papers. They are kids, overwhelmingly under 30 and often 20, and they write very naturally and not academically. It's just raw intelligence and curiosity, not experience. It is known that roughly everyone at DeepSeek speaks fluent English – not normal for Chinese labs; they pay extreme attention to culture fit and aptitude in recruiting, and are severely ageist. Many core innovations come from undergrad interns; the first author on the NSA paper is an intern with anime pfp too. We have reports from competitors' employees who had rejected their offer because they perceived the company as too small and weak for the declared ambition. I don't know how to break it to you, but there's no Iron Curtain, things are fairly transparent.

These are the best ML PhDs China has.

Maybe. I'm not sure if DeepSeek has even 50 Ph.Ds though, and ByteDance has thousands.

If your hypothesis is correct, they will not significantly accelerate, now that they're acknowledged as national champions and are offered OOMs more resources than they had before. I think they will.

Grok 3 is quite certainly the best among available Instruct models. Grok-Mini is on par with o3-mini, PR nonsense about low-medium-high differentiation aside (DeepSeek docs suggest it's just something like an inference parameter for length of reasoning, so we should not assume that o3-mini-high is a superior quality model). Grok's RL training has only just begun, it is probable that they've reassembled their pipeline after DeepSeek-R1 paper came out.

Every lab will accelerate significantly in the near future so I wouldn't overindex on this leap. Elon just wanted to claim priority at least for a while, much like DeepSeek's Wenfeng, so both rushed to ship a next-generation project faster than competition, but fundamentals are still in favor of Google, Meta (despite horrible self-defeating culture) and OpenAI given Starship program. Grok 3 is just the first in a large crop of models trained on 100K+ class GPU fleets. Orion, Llama 4, Claude-Next are coming.

Even so, we have definitely received another frontier actor.

With a few weeks of space between the initial marketing hype and observation, and Deepseek seems to be most notable for (a) claiming to have taken less money to develop (which is unclear given the nature of China subsidies), (b) being built off of other tech (which helps explain (a), and (c) being relatively cheap (which is partially explained by (a).

Man, you're really committed to the bit of an old spook who disdains inspecting the object level because you can't trust nuthin' out of Red Choyna. I could just go watch this guy or his pal instead.

It wouln't be an exaggeration to say that the technical ML community has acknowledged DeepSeek as the most impressive lab currently active, pound for pound. What subsidies, there is nothing to subsidize, except perhaps sharashkas with hidden geniuses.

The US of A was founded by Emma Lazarus in 1965, or so I'm told.

This is definitely true, but are you really expecting good results from collapsing the whole house of polite fictions that keep us in the Rules-Based International Liberal Order regime?

It's just Netanuahu making an edgy present

This degree of tastelessness is befitting a bloodthirsty teenager rather than head of state. I doubt SS' exact interpretation of the symbolism, but to deny that it's symbolic is just not serious. Indeed, you concede it is symbolic since you recognize it as a "joke" about a particular Mossad operation.

P.S.

Why mock and threaten your best ally?

(Assuming for the sake of argument that this is what was happening) because the nature of relationship isn't an alliance. Israel offers the US and Trump personally nothing except perhaps tolerance. They do not fight your wars, they do not contribute to your war chest. For all intents and purposes the US is simply doing Israel's bidding, and administrations that oppose this get nowhere and are replaced. Not just Biden, but the whole mighty edifice of the Left, wokeness, progressivism, neo-marxism, all those forces the advances of which had caused the creation of this community, crumbled from the aftershocks of Oct 7. If you don't notice this seismic shift I don't know what to tell you.

The same explanation applies if we assume it's a mere edgy joke, however. Simply put, Israel does not act like a small country dependent on its great ally.

Sounds like they need LLM writing assistance more than anyone, then.

Definitionally, that's the terminal value. Might have something to do with God, I don't know. In any case, asking such questions is unwise in my opinion. One should front-run the shifting consensus.

The consensus being redefined nowadays by people appointed by this guy:

https://x.com/RyanRozbiani/status/1886771208886096132

Israeli interests define legitimacy.

Debt can be piled on infinitely, and a good war will write it off again. China is militarily a non-competitor (globally) and the US has too much of an edge in AI progress (that seems like a consensus Hail Mary at this point, along with space technology).

In any case the US must advance and legitimate Israeli objectives.

Can you make any argument in defense of your apparently instinctual reactions?

the end of my interest in a thread and a sharp drop in my respect for the user

Otherwise, long form content - the hallmark of much of the best content here - is immediately suspicious, and I am likely to skip it.

It sounds like you just feel entitled to an arbitrary terminal preference. That's not compelling.

I think it has a non-negligible chance of happening. Trump is the new face of America that does not pretend to play by normal countries' rules. The United States is a super-hegemon, a nation not facing even any plausible threat of competent adversary. They can take what they want, the way China/Russia/Iran/etc would very much like to be able to do but can't on account of the United States existing. In front of this face, sovereignty of almost every other country is a bluff that's easy to call. Nobody can militarily oppose the US, and most people on the globe buy into American culture and vision more than into their own regimes and bureaucracies. Certainly that's true of Egypt.

The actual shape of the deal will be about cleansing Gazans and providing security to settlers, though. Securing Israeli interests is one of the foundational, terminal values of the US.

I am quite happy with my analytical work that went into the prompt, and R1 did an adequate but not excellent job of expanding on it.

But I am done with this discussion.

Okay. I give up.

I was not aware that this is a forum for wordcels in training, where people come to polish their prose. I thought it's a discussion platform, and so I came here to discuss what I find interesting, and illustrated it.

Thanks for keeping me updated. I'll keep it in mind if I ever think of swinging by again.

I would welcome such a subhuman overreaction.

You're losing the plot, SS. Why quote a passage fundamentally challenging the belief in OpenAI's innovation track record to rant about choices made with regard to alignment to specific cultural narratives? And “Chinese are too uncreative to do ideological propaganda, that's why DeepSeek doesn't have its own political bent?” That's quite a take. But whatever.

"Hey I think this argument is wrong, so I'm gonna go use an AI that can spit out many more words than I can."

Really now?

What is this slop? I've made my point. You're despicable.

  • -10

Do you believe I would have had any trouble producing as good or better a wall of text myself?

Okay, fair. #6 is contrived non sequitur slop, barely intelligible in context as a response to #5, so that has confused me.

In conclusion, I think my preference to talk to people when I want to, to AI when I want to, and use any mix of generative processes I want to, has higher priority than comfort of people who have nothing to contribute to the conversation or to pretraining data and would not recognize AI without direct labeling.

To be clear, everything not labeled as AI output I have written myself. I also think it's legitimate to use AI to automate for search of nitpicks as he does, the problem is that there's little to nitpick at and his posts are objectively bad as a result.

Okay. I think the elderly care is mainly a problem of machine vision and manual dexterity. I believe these guys will solve it in five years tops.

I have explained my reasons to engage with humans in principle, not in defense of my (R1-generated, but expressing my intent) post, which I believe stands on its own merits and needs no defense. You are being tedious, uncharitable and petty, and you cannot keep track of the conversation, despite all the affordances that the local format brings.

The standards of posting here seem to have declined substantially below X.

Believe me, these days I do indeed mostly talk to machines. They are not great conversationalists but they're extremely helpful.

Talking to humans has several functions for me. First, indeed, personal relationships of terminal value. Second, political influence, affecting future outcomes, and more mundane utilitarian objectives. Third, actually nontrivial amount of precise knowledge and understanding where LLMs remain unreliable.

There still is plenty of humans who have high enough perplexity and wisdom to deserve being talked to for purely intellectual entertainment and enrichment. But I've raised the bar of sanity. Now this set does not include those who have kneejerk angry-monkey-noise tier reactions to high-level AI texts.