this post was submitted on 03 Aug 2025
410 points (86.7% liked)

Fuck AI

4293 readers
798 users here now

"We did it, Patrick! We made a technological breakthrough!"

A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.

founded 2 years ago
MODERATORS
 

Source (Bluesky)

(page 4) 36 comments
sorted by: hot top controversial new old
[–] HalfSalesman@lemmy.world 0 points 2 months ago

I use LLMs in a way that reduces social anxiety from my autism, I give it details of a strange social interaction that I could not parse on my own and ask if I should worry about it, or if I should make any kind of amends or inquiries, or if I'm over thinking something and leave it alone.

I use LLMs to bounce my own ideas off of that I'm not comfortable bouncing off someone I know IRL.

I use LLM's to role play. (all kinds)

I use LLM's to find things that I can't find via conventional research methods.

And you know what, my perspective on using it for "productive/generative" usage is nuanced. I get why artists and writers are upset, however there is nothing magical about human's and their artistic abilities and in terms of material economic impacts automation of various kinds has screwed working people in the past and generally I've seen a lot less push back.

I do think that generated images and writing is pretty bland and near worthless though without a ton of human done work atm anyway. Like, sure I could generate a video of a cat dancing on a moving bus while a nuclear bomb is going off in the background or whatever wacky shit with a simple prompt but what exactly am I even going to do with that?

Highly directed AI content that includes a lot of human work tends to actually be pretty amazing IMO.

And even though all the outrage pertains to intellectual work, this technology is going to likely result in a lot of blue collar work being automated via "embodied" neural network AI's. In fact, it may be that it was needed for this kind of automation to really take off at all. Its not just white collar work. We aren't just automating slop content and corporate purposed art. The day is coming when stuff like laundry, factory/warehouse work, and kitchen work, etc. is also all being done by robots.

[–] TeraByteMarx@lemmy.dbzer0.com 0 points 2 months ago

Sex bots are taking work away from sex workers. Turning genuine sensuality into some kind of horrible mimicry of a genuine connection. How am I supposed to pay my bills

[–] Evotech@lemmy.world -1 points 2 months ago (7 children)

Sorry but you can deny and hate all you want, it’s not going anywhere

[–] korazail@lemmy.myserv.one 0 points 2 months ago

sadly. I don't have enough money to turn this shit-hose off.

Gen AI is neat, and I use it for personal processes including code, image gen, llm/chat; but it is sooooo faaaar awaaaay from being a real game changer - while all the people poised to profit off it claim it is - that it's just insane to claim it's the next wave. evidence: all the creative (photo/art/code/etc) people who are adamantly against it and have espoused reasoning.

There's another story on my feed about a 10-year-old refactoring a code base with a LLM. Go look at the comments from actual experts that take into account things like unit tests, readability, manageability, security. Humans have more context than any AI will.

LLMs are not intelligent. They are patently not. They make shit up constantly, since that is exactly what they do. Sometimes, maybe even most of the time, the shit they make up is mostly accurate... but do you want to rely on them?

When a doctor prescribes you the wrong drug, you can sue them as a recourse. When a software company has a data breach, there is often a class-action (better than nothing) as a recourse. When an AI tells you to put glue on your pizza to hold the toppings, there is no recourse, since the AI is not a legal thing and the company disclaims all liability for its output. When an AI denies your health insurance claim because of inscrutable reasons, there is no recourse.

In the first two, there is a penalty for being wrong, which is in effect an incentive to be correct -- to be accurate, to be responsible.

In the last, as an AI llm/agent/fuckingbuzzword, there is no penalty and no incentive. The AI just is as good as its input, and half the world is fucking stupid, so if we average out all the world's input, we get "barely getting by" as a result. A coding AI is at least partially trained on random stackoverflow posts asking for help. The original code there is wrong!

Sadly, it's not going anywhere. But people who rely on it will find short-term success for long-term failure. And a society relying on it is doomed. AI relies on the creative works that already exist. If we don't make any new things, AI will stagnate and die. Where will we be then?

There are places AI/LLM/Machine-Learning can be used successfully and helpfully, but they are niche. The AI bros need to be figuring out how to quickly meet a specific need instead of trying to meet all needs at the same time. Think the early 2000-s Folding at Home, how to convince republicans to wear a fucking mask during covid, why we shouldn't just eat the billionaires*.

*Hermes-3 says cannibalism is "barbaric" in most cultures, but otherwise doesn't give convincing arguments.

load more comments (6 replies)
[–] TeraByteMarx@lemmy.dbzer0.com -2 points 2 months ago

Marx would have been pro AI and anti capitalist

[–] nonentity@sh.itjust.works -2 points 2 months ago* (last edited 2 months ago)

Slopvangelical LLM thumpers.

Edit: is this being misinterpreted as a pro-AI statement?

load more comments
view more: ‹ prev next ›