top | item 43838084

(no title)

amusingimpala75 | 10 months ago

Meta needs to stop open-washing their product. It simply is not open-source. The license for their precompiled binary blob (ie model) should not be considered open-source, and the source code (ie training process / data) isn’t available.

discuss

order

michaelt|10 months ago

> the source code (ie training process / data) isn’t available

The training data is all scraped from the internet, ebooks from libgen, papers from Sci-Hub, and suchlike.

They don't have the right to redistribute it.

observationist|10 months ago

They've painted themselves into a corner - the second people see the announcement that they've enforced the license on someone, people will switch to actual open source licensed models and Meta's reputation will take a hit.

It's ironic that China is acting as a better good faith participant in open source than Meta. I'm sure their stakeholders don't really care right now, but Meta should switch to Apache or MIT. The longer they wait the more invested people will be and the more intense the outrage when things go wrong.

piperswe|10 months ago

Applying Apache or MIT to a binary blob doesn't make it open source either

bbayer|10 months ago

This is actually my first impression while I am reading the post. Mentions "open source" everywhere but dude how the earth it is open source without training data.

ronsor|10 months ago

Almost no company is going to release training data because they don't want to waste time with lawsuits. That's why it doesn't happen. Until governments fix that issue, I don't even think the "it's not really open without training data!!!" argument is worth any time. It's more worth focusing on the various restrictions in the LLaMA license, or even better, questioning whether model weights can be licensed at all.

NitpickLawyer|10 months ago

> their precompiled binary blob (ie model)

I agree with you that their license is not open source, but model weights are not binary blobs! Please stop spreading this misconception.

aprilthird2021|10 months ago

I get the argument completely, but isn't the open-washing a little acceptable if they're the only big company releasing open-weights models?

lern_too_spel|10 months ago

What is the point of considering this hypothetical? Google, Microsoft, Nvidia, Apple, and many Chinese big-tech companies also release open weights models, most with fewer restrictions.

CaptainFever|10 months ago

My issue with Meta's open-washing is that it is also not open-weight, given the license restrictions. It's "weight-available", I suppose. Try OLMo instead.