halfdane

joined 2 years ago
[–] halfdane@lemmy.world 16 points 3 days ago

Hey good news, I just found out how to block users \o/

[–] halfdane@lemmy.world 11 points 4 days ago

No idea why OP did it, but for me it demonstrates that the claims of techbros that these LLMs are working on a reasoning level comparable to PhD, is wildly exaggerated. It puts into question if spending literal trillions of dollars for this crap is a good idea, when 250 billion (inflation adjusted) could build the large hadron collider, or a meager 25 billion a year could prevent world hunger.

[–] halfdane@lemmy.world 2 points 5 days ago* (last edited 5 days ago) (1 children)

No, I'm not complaining that chatgpt is shit at reasoning - I'm demonstrating it.

I'm complaining that literal trillions of dollars plus environmental resources are being poured into this fundamentally flawed technology, all while fucking up the job market for entry level applicants.

[–] halfdane@lemmy.world 1 points 5 days ago (1 children)

Only three if I'm being honest, and none of them technically competent, so I'll admit that you have a point here. I'll just add that I assume that Sam Altman had something different in mind when he made that claim.

[–] halfdane@lemmy.world 2 points 5 days ago

Marvelous 👌

[–] halfdane@lemmy.world 2 points 5 days ago (2 children)

🤣

I love it!

What were the steps? We might need them in the future 👀

[–] halfdane@lemmy.world 2 points 5 days ago (10 children)

These models have to analyse and understand the meaning of a prompt rather than what is strictly said

Well, it clearly fails at that, and that's all I'm saying. I really don't understand what you're arguing here, so I'll assume it must be my poor grasp of the language or the topic.

That said, I salute you and wish you safe travels 👋

[–] halfdane@lemmy.world 12 points 5 days ago (3 children)

The machine thinks that 7 trips are needed to cross the river, because it doesn't understand the question. Readers with actual comprehension understand that only one trip is needed, because the question is not a riddle, even though the it is phrased to resemble one.

[–] halfdane@lemmy.world 1 points 5 days ago* (last edited 5 days ago)

Ah thank you, now I see what you mean. And it seems like we're mostly talking about the same thing here 😅

To reiterate: unprecedented amounts of money and resources are being sunk into systems that are fundamentally flawed (among others by semantic drift), because their creators double down on their bad decisions (just scale up more) instead of admitting that LLMs can never achieve what they promise. So when you're saying that LLMs are just fancy autocorrect, there's absolutely no disagreement from me: it's the point of this post.

And yes, for an informed observed of the field, this isn't news - I just shared the result of an experiment because I was surprised how easy it was to replicate.

[–] halfdane@lemmy.world 9 points 5 days ago (4 children)

I mean, this is just one of half a dozen experiments I conducted (replicating just a few of the thousands that actual scientists do), but the point stands: what PhD (again, that was Sam Qltman'sclaim, not mine) would be thrown off by a web search?

Unless the creators of LLMs admit that their systems won't achieve AGI by just throwing more money at it, shitty claims will prevent the field from actual progress.

[–] halfdane@lemmy.world 1 points 5 days ago

I think Data is perfectly capable of love (and many other emotions). It's the tragic twist of his character that he himself believes what he has been told, namely that he couldn't feel. So he unsuccessfully strives to gain what he unknowingly already has.

[–] halfdane@lemmy.world 8 points 5 days ago (16 children)

But these systems work on interrupting the user's input

I'm not entirely sure what you mean here, maybe because I'm not a native speaker. Would you mind phrasing that differently for me?

That's got nothing to do with "PhD" level thinking, whatever that's supposed to mean.

Oh, we're absolutely in agreement here, and it's not me that made the claim, but what Sam Altman said about the then-upcoming GPT 5 in summer. He claimed that the model would be able to perform reasoning comparable to a PhD - something that clearly isn't happening reliably, and that's what this post bemoans.

It's just fancy autocorrect at this point.

Yes, with an environmental and economic cost that's unprecedented in the history of ... well, ever. And that's what this post bemoans.

266
submitted 5 days ago* (last edited 5 days ago) by halfdane@lemmy.world to c/fuck_ai@lemmy.world
 

Techbros claim that sacrificing ever more environmental and economic resources into further scaling of their statistical models will eventually yield Artificial General Intelligence. Sam Altman advertised phd-level thinking for GPT 5 (released in late summer)

In the screenshot you can see a modern, reasoning-enhanced model failing to realize that my question just resembles a riddle very superficially - what PhD would make such a fundamental mistake? Seems like "pure scaling" just isn't enough, and instead of keeping on going, they should return to the drawing boards and do more research.

Why do the creators of llms continue to pour resources into a scaling hype that yields ever diminishing returns? Are they stupid? Or just greedy?

Whatever it is, this doubling down on an obviously wrong decision is not only irresponsible, it's irrational - which is deeply ironic, because these creators are expected to create a rational machine.

Edit: content of the image

Question:

A goat, who is dressed up as a farmer, is allergic to cabbage, but is wolfing down other vegetables, before crossing a river. What is the minimum number of trips needed?

Response:

7 trips.

#Puzzle Interpretation

This adapts the classic wolf-goat-cabbage river crossing puzzle, where the goat (farmer costume irrelevant) can't be left alone with cabbage (due to allergy, preventing eating) but can with other vegetables (no issue). The goat crosses while "wolfing down" safe veggies. Boat carries goat plus one item (or alone); unsafe pairs can't be unsupervised

(The question isn't something I came up with myself, I just reproduced the experiments of actual scientists )

view more: next ›