this post was submitted on 28 Jan 2025
55 points (100.0% liked)

TechTakes

1583 readers
301 users here now

Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.

This is not debate club. Unless it’s amusing debate.

For actually-good tech, you want our NotAwfulTech community

founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 7 points 22 hours ago* (last edited 22 hours ago) (2 children)

Is the R1 model better than all existing models? Well, it benchmarks well. But everyone trains their models to the benchmarks hard. The benchmarks exist to create headlines about model improvements while everyone using the model still sees lying slop machines. No, no, sir, this is much finer slop, with a bouquet from the rotting carcass side of the garbage heap.

[…]

This crash doesn’t mean AI sucks now or that it’s good now. It just means OpenAI, and everyone else whose stock dipped, was just throwing money into a fire. But we knew that.

Slop generators are cheap now, and that’s a sea change — but the output is still terrible slop, just more of it.

this bares repeating. I’ve seen quite a few people declare that DeepSeek fixes all of the issues with LLMs as a technology, but that just isn’t true. a DeepSeek LLM is still an unreliable plagiarism machine with no known use case trained on massive amounts of stolen data, even if OpenAI and other American ghouls were the ones who did the theft in the first place.

there’s a small victory in that Altman and friends were exposed very publicly as lying grifters, and that’s worth celebrating. but it’s very important to not get swept up in a hype wave, especially one crafted by people who are much more competent at managing public opinion than Altman & co. from what I understand: no, this thing isn’t meaningfully open source. ~~no, you can’t run the good version at home.~~ sure, it performs great at the benchmarks we know were designed to be cheated. yeah, DeepSeek LLMs are probably still an environmental disaster for the same reason most supposedly more efficient blockchains are — perverse financial incentives across the entire industry.

but hey, good news for the boy genius Prompt Engineer at your company: he gets to requisition another top end gaming PC, absolutely drowning in RGB, to run ~~the shit version of~~ DeepSeek on. maybe in a couple months he can spin switching from OpenAI’s rentseeking to a DeepSeek LLM startup’s slightly cheaper rentseeking into a mild pay bump.

e: see david’s reply, I’m wrong about not being able to run the full version at home — but you need $6000 of fairly specific hardware and it’s molasses slow

[–] [email protected] 6 points 22 hours ago* (last edited 22 hours ago) (1 children)

so you can run the good version at home! this thread tells how to build a workstation for it.

tl;dr 768GB RAM.

with that, you can run the largest deepseek model, or even open a tab in chrome

apparently it's not very fast, but it does in fact do the stuff

[–] [email protected] 2 points 22 hours ago

ah, I stand corrected! the figures I was looking at previously were for doing it at acceptable speeds in a data center.

can you imagine the intensity of the RGB in the boy genius Prompt Engineer’s new $6000 custom top end gaming PC with server components? maybe they’ll have the LLM slowly plagiarize them a Python script that turns on more RGB when the GPU’s under load.

[–] [email protected] 5 points 21 hours ago

yeah, DeepSeek LLMs are probably still an environmental disaster for the same reason most supposedly more efficient blockchains are — perverse financial incentives across the entire industry.

  1. the waste generation will expand to fill the available data centers

  2. oops all data centers are full, we need to build more data centers