this post was submitted on 30 Sep 2023
1014 points (98.8% liked)

Open Source

31114 readers
1 users here now

All about open source! Feel free to ask questions, and share news, and interesting stuff!

Useful Links

Rules

Related Communities

Community icon from opensource.org, but we are not affiliated with them.

founded 5 years ago
MODERATORS
top 50 comments
sorted by: hot top controversial new old
[–] weedazz@lemmy.world 139 points 2 years ago* (last edited 2 years ago) (2 children)

My mind immediately went to a horizon zero dawn like dystopia where the Mozilla AI is the only thing left protecting humans from various malevolent AIs bent on consuming the human race

[–] Bombastic@sopuli.xyz 34 points 2 years ago (2 children)

Mozilla is Gaia, ChatGPT is hades?

[–] weedazz@lemmy.world 21 points 2 years ago* (last edited 2 years ago) (2 children)

I think by that point ChatGPT would be more like Apollo, keeping the knowledge of humanity. I feel like one of the more corporate AIs will go full HADES, I'm thinking Bard. It will get a mysterious signal from space that switches it's core protocol from "don't be evil" to "be evil."

[–] Doombot1@lemmy.one 4 points 2 years ago

A “mysterious signal from space” being just the fact that it’s owned by Google

load more comments (1 replies)
[–] angrymouse@lemmy.world 3 points 2 years ago

SPOOOOOILER

load more comments (1 replies)
[–] cupcakezealot@lemmy.blahaj.zone 84 points 2 years ago

Incredibly welcomed. We need more ethical, non-profit AI researchers in the sea of corporate for-profit AI companies.

[–] KingThrillgore@lemmy.ml 50 points 2 years ago

I want to give them the benefit of the doubt. I really do. I am going to watch this with a critical eye, however.

[–] Weeby_Wabbit@lemmy.world 41 points 2 years ago (2 children)

I'll believe it when I see it.

I'm so goddamn tired of "open source" turning into subscription models restricting use cases because the company wants to appease conservative investors.

[–] blind3rdeye@lemm.ee 35 points 2 years ago

Mozilla has a very strong track-record though. They've been around for a very long time, and have stuck to free open-source principles the whole time.

[–] BetaDoggo_@lemmy.world 3 points 2 years ago* (last edited 2 years ago) (1 children)

That's basically only OpenAI, maybe some obscure startups as well. Mozzila is far too old and niche to get away with that anyway.

load more comments (1 replies)
[–] kubica@kbin.social 26 points 2 years ago

Wishing they would say something more, the site has been like that for some time.

[–] mojo@lemm.ee 24 points 2 years ago (4 children)

As much as I love Mozilla, I know they're going to censor it (sorry, the word is "alignment" now) the hell out of it to fit their perceived values. Luckily if it's open source then people will be able to train uncensored models

[–] DigitalJacobin@lemmy.ml 68 points 2 years ago (29 children)

What in the world would an "uncensored" model even imply? And give me a break, private platforms choosing to not platform something/someone isn't "censorship", you don't have a right to another's platform. Mozilla has always been a principled organization and they have never pretended to be apathetic fence-sitters.

[–] Doug7070@lemmy.world 35 points 2 years ago (5 children)

This is something I think a lot of people don't get about all the current ML hype. Even if you disregard all the other huge ethics issues surrounding sourcing training data, what does anybody think is going to happen if you take the modern web, a huge sea of extremist social media posts, SEO optimized scams and malware, and just general data toxic waste, and then train a model on it without rigorously pushing it away from being deranged? There's a reason all the current AI chatbots have had countless hours of human moderation adjustment to make them remotely acceptable to deploy publicly, and even then there are plenty of infamous examples of them running off the rails and saying deranged things.

Talking about an "uncensored" LLM basically just comes down to saying you'd like the unfiltered experience of a robot that will casually regurgitate all the worst parts of the internet at you, so unless you're actively trying to produce a model to do illegal or unethical things I don't quite see the point of contention or what "censorship" could actually mean in this context.

[–] underisk@lemmy.ml 15 points 2 years ago

It means they can’t make porn images of celebs or anime waifus, usually.

load more comments (4 replies)
[–] lemann@lemmy.one 18 points 2 years ago

I fooled around with some uncensored LLaMA models, and to be honest if you try to hold a conversation with most of them they tend to get cranky after a while - especially when they hallucinate a lie and you point it out or question it.

I will never forget when one of the models tried to convince me that photosynthesis wasn't real, and started getting all snappy when I said I wasn't accepting that answer 😂

Most of the censorship "fine tuning" data that I've seen (for LoRA models anyway) appears to be mainly scientific data, instructional data, and conversation excerpts

[–] TheWiseAlaundo@lemmy.whynotdrs.org 15 points 2 years ago (2 children)

There's a ton of stuff ChatGPT won't answer, which is supremely annoying.

I've tried making Dungeons and Dragons scenarios with it, and it will simply refuse to describe violence. Pretty much a full stop.

Open AI is also a complete prude about nudity, so Eilistraee (Drow godess that dances with a sword) just isn't an option for their image generation. Text generation will try to avoid nudity, but also stop short of directly addressing it.

Sarcasm is, for the most part, very difficult to do... If ChatGPT thinks what you're trying to write is mean-spirited, it just won't do it. However, delusional/magical thinking is actually acceptable. Try asking ChatGPT how licking stamps will give you better body positivity, and it's fine, and often unintentionally very funny.

There's plenty of topics that LLMs are overly sensitive about, and uncensored models largely correct that. I'm running Wizard 30B uncensored locally, and ChatGPT for everything else. I'd like to think I'm not a weirdo, I just like D&d... a lot, lol... and even with my use case I'm bumping my head on some of the censorship issues with LLMs.

load more comments (2 replies)
load more comments (26 replies)
[–] whoisearth@lemmy.ca 5 points 2 years ago

As an aside I'm in corporate. I love how gung ho we are on AI meanwhile there are lawsuits and potential lawsuits and investigative journalism coming out on all the shady shit AI and their companies are doing. Meanwhile you know the SMT ain't dumb they know about all this shit and we are still driving forward.

[–] VonCesaw@lemmy.world 3 points 2 years ago (2 children)

If 'censored' means that underpaid workers in developing countries don't need to sift through millions of images of gore, violence, etc, then I'm for it

[–] mojo@lemm.ee 4 points 2 years ago (2 children)
load more comments (2 replies)
load more comments (1 replies)
load more comments (1 replies)
[–] donuts@kbin.social 18 points 2 years ago (1 children)

All I want to know is if they are going to pillage people's private data and steal their creative IP or not.

Ethical AI starts and ends with open, transparent, legitimate and ethically sourced training data sets.

load more comments (1 replies)
[–] fosforus@sopuli.xyz 17 points 2 years ago* (last edited 2 years ago)

I remember a time when open-source software was developed without a pre-order business model.

"This new company will be led by Managing Director Moez Draief. Moez has spent over a decade working on the practical applications of cutting-edge AI as an academic at Imperial College and LSE, and as a chief scientist in industry. Harvard’s Karim Lakhani, Credo’s Navrina Singh and myself will serve as the initial Board of Mozilla.ai. "

This money will 90% go to paying the salary of these managers.

[–] lowleveldata@programming.dev 15 points 2 years ago (2 children)
[–] Limitless_screaming@kbin.social 42 points 2 years ago* (last edited 2 years ago)

More transparent about data collection, and less likely to reinforce biases. Mozilla vision for trustworthy AI

[–] AdmiralShat@programming.dev 24 points 2 years ago

Open source

[–] CaptKoala@lemmy.ml 14 points 2 years ago* (last edited 2 years ago)

Couldn't give a fuck, there's already far too much bad blood regarding any form of AI for me.

It's been shoved in my face, phone and computer for some time now. The best AI is one that doesn't exist. AGI can suck my left nut too, don't fuckin care.

Give me livable wages or give me death, I care not for anything else at this point.

Edit: I care far more about this for privacy reasons than the benefits provided via the tech.

The fact these models reached "production ready" status so quickly is beyond concerning, I suspect the companies are hoping to harvest as much usable data as possible before being regulated into (best case) oblivion. It really no longer seems that I can learn my way out of this, as I've been doing since the beginning, as the technology is advancing too quickly for users, let alone regulators to keep it in check.

[–] baduhai@sopuli.xyz 13 points 2 years ago

Cautiously optimistic about this one.

[–] bahmanm@lemmy.ml 10 points 2 years ago

Something that I'll definitely keep an eye on. Thanks for sharing!

[–] Wrong_thought_7@lemmy.ml 6 points 2 years ago
[–] Turun@feddit.de 4 points 2 years ago* (last edited 2 years ago) (1 children)

In which ways does this differ from stability ai, which made stable diffusion and also have a LLM afaik?

[–] RobotToaster@mander.xyz 10 points 2 years ago

The stability models aren't open source, the moralistic licence they release under violates the open source definition.

load more comments
view more: next ›