ebu

joined 2 years ago
[–] ebu@awful.systems 27 points 2 years ago (14 children)

48th percentile is basically "average lawyer".

good thing all of law is just answering multiple-choice tests

I don't need a Supreme Court lawyer to argue my parking ticket.

because judges looooove reading AI garbage and will definitely be willing to work with someone who is just repeatedly stuffing legal-sounding keywords into google docs and mashing "generate"

And if you train the LLM with specific case law and use RAG can get much better.

"guys our keyword-stuffing techniques aren't working, we need a system to stuff EVEN MORE KEYWORDS into the keyword reassembler"

In a worst case scenario if my local lawyer can use AI to generate a letter

oh i would love to read those court documents

and just quickly go through it to make sure it didn't hallucinate

wow, negative time saved! okay so your lawyer has to read and parse several paragraphs of statistical word salad, scrap 80+% of it because it's legalese-flavored gobbledygook, and then try to write around and reformat the remaining 20% into something that's syntactically and legally coherent -- you know, the thing their profession is literally on the line for. good idea

what promptfondlers continuously seem to fail to understand is that verification is the hard step. literally anyone on the planet can write a legal letter if they don't care about its quality or the ramifications of sending it to a judge in their criminal defense trial. part of being a lawyer is being able to tell actual legal arguments from bullshit, and when you hire an attorney, that is the skill you are paying for. not how many paragraphs of bullshit they can spit out per minute

they can process more clients, offer faster service and cheaper prices. Maybe not a revolution but still a win.

"but the line is going up!! see?! sure we're constantly losing cases and/or getting them thrown out because we're spamming documents full of nonsense at the court clerk, but we're doing it so quickly!!"

[–] ebu@awful.systems 36 points 2 years ago (27 children)

[...W]hen examining only those who passed the exam (i.e. licensed or license-pending attorneys), GPT-4’s performance is estimated to drop to 48th percentile overall, and 15th percentile on essays.

officially Not The Worst™, so clearly AI is going to take over law and governments any day now

also. what the hell is going on in that other reply thread. just a parade of people incorrecting each other going "LLM's don't work like [bad analogy], they work like [even worse analogy]". did we hit too many buzzwords?

[–] ebu@awful.systems 3 points 2 years ago* (last edited 2 years ago)

"i reflexively identify with the openly-fascist right-wing base that has found its home on elon's twitter, and since i'm a reasonable person, the evidence that they're flagrantly conspiracy-minded and/or are CSAM posters simply must be fabricated"

[–] ebu@awful.systems 6 points 2 years ago* (last edited 2 years ago) (1 children)

i mean. definitionally, some did, yeah? if you bought in at 25, 50, 75, 100, 200, or 400 -- these are all the same number in the end, the only difference being how much you're down by between then and now.

eta: that's not even to mention the fact that since this demand is all synthetic, all the money coming in is from people who are going to be left holding the bag, again. we're just watching it repeat.

[–] ebu@awful.systems 5 points 2 years ago

this is funny to me because it took Notion until late 2021 to introduce simple, non-database tables (since the database tables were often large, unwieldy, and introduced way too much overhead to just write a simple rows-and-columns spreadsheet, something that's been a thing in GitHub Flavored Markdown since at least 2009

[–] ebu@awful.systems 11 points 2 years ago* (last edited 2 years ago) (3 children)

finally... MOASS... this time for real... if January 2021 buyers sell now, they'll only be down about 70%, instead of the 85-90% it normally hovers around. i think the only hodlers that could come out positive are ones that bought in late 2022 or later, and even then, you're not up by much.

i think this, more than watching the Folding Ideas video (a must-watch for anyone out of the loop), is really kind of selling the sadness of watching people suckered into hype pour even more money down the drain. an account belonging to a guy we once liked made a tweet; this is it, liquidate your retirement and gamble it away. ugh

[–] ebu@awful.systems 5 points 2 years ago* (last edited 2 years ago) (3 children)

It doesn’t seem to be able to do anything that a GitLab instance can’t

i didn't believe you, but yeah, just learned GitLab has a wiki editor. so yeah, this covers like 95% of the things i once used Notion for. i guess if i want to be pedantic, Notion had database relations between tables that, as the name implies, allowed it to act a bit like an relational database. (e.g. allowing columns of tables to be limited to the values of rows of other tables). admittedly a little cool but in my experience was not much more useful than a simple table

[–] ebu@awful.systems 8 points 2 years ago* (last edited 2 years ago)

i couldn't delete the one question i had on stackoverflow, so i used a text generator to overwrite the body and title of the question. fight garbage with garbage

[–] ebu@awful.systems 19 points 2 years ago* (last edited 2 years ago)

as previously discussed, the rabbit r1 turns out to be (gasp) just an android app.

in a twist no one saw coming, the servers running "rabbit os" report to just be running Ubuntu, and the "large action model" that was supposed to be able to watch humans use interfaces and learn how to use them, turns out to just be a series of hardcoded places to click in Playwright.

[–] ebu@awful.systems 7 points 2 years ago (2 children)

i suppose if Elizabeth Holmes can wear Steve Job's turtlenecks and carry a biomedical scam to the tune of hundreds of millions of dollars by doing so, maybe "financial pickup artistry" will see more success attracting VC funds than "pickup artistry" has attracting the ladies

...

oh gods we're going to get so many self-stylized Elon clones. fuuuuuuuuuuuuuuck

[–] ebu@awful.systems 10 points 2 years ago (6 children)

love If Books Could Kill. highly recommend.

i can recognize that sometimes getting away with massive amounts of fraud and theft is sometimes as easy as just being the right kind of charming and personable guy. that someone who talks smooth gets the benefit of the doubt. what i don't understand is how SBF's outstandingly bad interpersonal skills don't seem to immediately disqualify him from getting the starry-eyed treatment he got (and still gets). is it really just the fact that he's rich?

[–] ebu@awful.systems 10 points 2 years ago (1 children)

Ultimately, LLMs don’t use words,

LLM responses are basically paths through the token space, they may or may not overuse certain words, but they’ll have a bias towards using certain words together

so they use words but they don't. okay

this is about as convincing a point as "humans don't use words, they use letters!" it's not saying anything, just adding noise

So I don’t think this is impossible… Humans struggle to grasp these kinds of hidden relationships (consciously at least), but neural networks are good at that kind of thing

i can't tell what the "this" is that you think is possible

part of the problem is that a lot of those "hidden relationships" are also noise. knowing that "running" is typically an activity involving your legs doesn't help one parse the sentence "he's running his mouth", and part of participating in communication is being able to throw out these spurious and useless connections when reading and writing, something the machine consistently fails to do.

It’s incredibly useful to generate all sorts of content when paired with a skilled human

so is a rock

It can handle the tedious details while a skilled human drives it and validates the output

validation is the hard step, actually. writing articles is actually really easy if you don't care about the legibility, truthiness, or quality of the output. i've tried to "co-write" short-format fiction with large language models for fun and it always devolved into me deleting large chunks -- or even the whole -- output of the machine and rewriting it by hand. i was more "productive" with a blank notepad.exe. i've not tried it for documentation or persuasive writing but i'm pretty sure it would be a similar situation there, if not even more so, because in nonfiction writing i actually have to conform to reality.

this argument always baffles me whenever it comes up. as if writing is 5% coming up with ideas and then the other 95% is boring, tedium, pen-in-hand (or fingers-on-keyboard) execution. i've yet to meet a writer who believes this -- all the writing i've ever done required more-or-less constant editorial decisions from the macro scale of format and structure down to individual choices. have i sufficiently introduced this concept? do i like the way this sentence flows, or does it need to go earlier in the paragraph? how does this tie with the feeling i'm trying to convey or the argument i'm trying to put forward?

writing is, as a skill, that editorial process (at least to one degree or another). sure, i can defer all the choice to the machine and get the statistically-most-expected, confusing, factually dubious, aimless, unchallenging, and uncompelling text out of it. but if i want anything more than that (and i suspect most writers do), then i am doing 100% of that work myself.

view more: ‹ prev next ›