this post was submitted on 18 May 2025
75 points (98.7% liked)

LocalLLaMA

2982 readers
19 users here now

Welcome to LocalLLaMA! Here we discuss running and developing machine learning models at home. Lets explore cutting edge open source neural network technology together.

Get support from the community! Ask questions, share prompts, discuss benchmarks, get hyped at the latest and greatest model releases! Enjoy talking about our awesome hobby.

As ambassadors of the self-hosting machine learning community, we strive to support each other and share our enthusiasm in a positive constructive way.

founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] PolarKraken@sh.itjust.works 3 points 3 days ago (1 children)

Ah, I think I'm following you, thanks!

You're right, I did misunderstand part of that - if I've got it now, it still seems surprising but much less than I thought.

It didn't pick up those biases without being trained on them at all, it did receive training (via fine-tuning) for a subset of them. And the surprising part is that the LLM generalized that preference to also prefer behaviors it learned about from the fictional papers, but was never trained to prefer, sort of lumping those behaviors into this general feature it developed. Is that a reasonable restatement of the correction?

I lack the time spent to be precise with my vocabulary so forgive me if I butchered that lol. Thank you for clarifying, that makes a lot more sense than what I took away, too!

[–] sleep_deprived@lemmy.dbzer0.com 2 points 3 days ago (1 children)

Yes, that's an excellent restatement - "lumping the behaviors together" is a good way to think about it. It learned the abstract concept "reward model biases", and was able to identify that concept as a relevant upstream description of the behaviors it was trained to display through fine tuning, which allowed it to generalize.

There was also a related recent study on similar emergent behaviors, where researchers found that fine tuning models on code with security vulnerabilities caused it to become widely unaligned, for example saying that humans should be enslaved by AI or giving malicious advice: https://arxiv.org/abs/2502.17424

[–] PolarKraken@sh.itjust.works 3 points 3 days ago* (last edited 2 days ago)

Holy cow that sounds nuts, will def have to go through this one, thanks!!

Edit: hmm. Think I just noticed that one of my go-to "vanilla" expressions of surprise would likely (and justifiably) be considered culturally insensitive or worse by some folks. Time for "holy cow" to leave my vocabulary.