this post was submitted on 07 Nov 2025
22 points (100.0% liked)

TechTakes

2282 readers
40 users here now

Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.

This is not debate club. Unless it’s amusing debate.

For actually-good tech, you want our NotAwfulTech community

founded 2 years ago
MODERATORS
top 10 comments
sorted by: hot top controversial new old
[–] Architeuthis@awful.systems 3 points 20 hours ago (1 children)

So if a company does want to use LLM, it is best done using local servers, such as Mac Studios or Nvidia DGX Sparks: relatively low-cost systems with lots of memory and accelerators optimized for processing ML tasks.

Eh, Local LLMs don't really scale, you can't do much better than one person per one computer, unless it's really sparse usage, and buying everyone a top-of-the-line GPU only works if they aren't currently on work laptops and VMs.

Sparks type machines will do better eventually but for now they're supposedly geared more towards training than inference, it says here that running a 70b model there returns around one word per second (three tokens) which is snail's pace.

[–] dgerard@awful.systems 3 points 19 hours ago (1 children)

yeah. LLMs are fat. Lesser ML works great tho.

[–] pikesley@mastodon.me.uk 3 points 18 hours ago
[–] Doomsider@lemmy.world 2 points 19 hours ago

There is no future. They will be outdated by the time they are finished and the most expensive part wears out quickly and has to be replaced. Literally DOA.

[–] zbyte64@awful.systems 5 points 1 day ago (1 children)

Let me see if I got this right: Because use cases for LLMs have to be resilient to hallucinations, large data centers will fall out of favor for smaller, cheaper deployments at the cost of accuracy. And once you have a business that is categorizing relevant data, you will gradually move away from black box LLMs and towards ML on the edge to cut costs and also at the cost of accuracy.

[–] sleepundertheleaves 4 points 1 day ago* (last edited 9 hours ago)

I read it this way: because LLMs inevitably hallucinate, no matter how resource intensive the LLM is, it makes economic sense to deploy smaller, cheaper LLMs that hallucinate a little more. The tradeoff isn't "hallucinations vs no hallucinations", it's "more hallucinations vs fewer hallucinations", and the slight gain in accuracy from using the big data center isn't worth the huge expense of using those big data centers.

[–] o7___o7@awful.systems 2 points 1 day ago* (last edited 1 day ago)

A++ episode, you're a great interviewer

[–] Soyweiser@awful.systems 6 points 1 day ago (1 children)

Like how this is an explainer for laymen but still just casually drops an 'on the edge' reference. The meaning of which might not be clear to laymen (the context explains it however, so it isn't bad, just talking about how much jargon we all use).

[–] jonhendry@iosdev.space 4 points 1 day ago (1 children)

@Soyweiser

I hate "the edge". Such a vague expression.

[–] Soyweiser@awful.systems 4 points 1 day ago

Think that is in part intentional so people don't start squabbling over what does and doesn't count as 'the edge' in edge cases, as it also quite depends on the setup of the organization/people you are talking about. But yeah it is badly defined, which is also why I noticed it.