marmaduke
a year ago
> Unlike most prior efforts, we release not only model weights and inference code, but also the reproducible training data, complete data processing pipeline, rigorous experimental ablation results, and detailed training protocols for open scientific research.
Regardless of the specific performance of this model versus another model, I think it’s good to keep in mind that everyone benefits from this kind of work
WiSaGaN
a year ago
This is probably the closest you can get as an "open source" model.
perching_aix
a year ago
Is this not what's referred to as an open-data (& open-source) solution?
lordswork
a year ago
When it comes to how "open" LLMs are, there are basically three categories:
- Open Source: The architecture of the model is available, so you can bring your own data and compute to train a similar model.
- Open Weights: The trained model itself is available for you to use. You can download it and run it on your own hardware without needing to train anything from scratch.
- Open Data: You get access to the data that was used to train the model. With this, you can completely reproduce the training process, which is super helpful if you want to fine-tune the model or see exactly how it was trained.
There are some other nuances around hyperparams, training methodology, etc. but that's basically the 3 main categories.
szundi
a year ago
The data is a kind of “source” in the original meaning when source code was named.
I think open-weight is somewhere between open source and binary.
Reason: cannot be reproduced or practically modified without the source data collection.
mistrial9
a year ago
agree but be sure to check for microphones in the flowerpot