Office space meme:
“If y’all could stop calling an LLM “open source” just because they published the weights… that would be great.”
Office space meme:
“If y’all could stop calling an LLM “open source” just because they published the weights… that would be great.”
It’s not just the weights though is it? You can download the training data they used, and run your own instance of the model completely separate from their servers.
You don’t download the training data when running an LLM locally. You are downloading the already baked model.
Did “they” publish the training data? And the hyperparameters?
I mean, I downloaded it from the repo.
You downloaded the weights. That’s something different.
I may misunderstand, but are the weights typically several hundred gigabytes large?
Yes. The training data is probably a few hundred petabytes.
Oh wow that’s fuckin huge
Yeah, some models are trained on pretty much the entire content of the publicly accessible Internet.