- cross-posted to:
- [email protected]
- cross-posted to:
- [email protected]
Fully open and accessible: Fully open-source release of model weights, training hyperparameters, datasets, and code, fostering innovation and collaboration within the AI community.
That’s actually pretty good. Seems to be open source as the OSI defines it, rather than the much more common “this model is open source, but the dataset is a secret”.
I need to catch up on training. I need an LLM that I can train on all my ebooks and digitized music, and can answer questions “what’s that book where the girl goes to the thing and does that deed?”
You probably could use RAG for this instead of actually training a model.
Existing implementations can probably do that already.
I’m sure; I just don’t know how. I need to set aside some time and educate myself.
Frankly, this generation of AI I find rather dull. It won’t directly lead a AGI, although I’m sure it’ll be a component, but I think that’ll be another 10-20 years before the next breakthrough. I personally don’t think it’s as interesting as the symbolic, knowledge-based systems of the mid-80’s; at least those were reasoning systems. LLMs look impressive to lay people (including myself - I understand the general concepts, but have no experience with the programming or training, so I’m just another lay user), but there’s no reasoning or understanding behind it, and if what it produces is truthful or accurate, it’s largely on accident. So I’ve had trouble getting excited about it.
I mean literally just go to chatgpt or whatever and ask it “what’s that movie with Morgan Freeman playing God” and it’ll give a few guesses. For common info, it’s usually pretty good.
I don’t want ChatGPT to remember that I was searching for Cafe Flesh by description.
I see all these graphs about how much better this LLM is than another, but do those graphs actually translate to real world usefulness?
I think more of the issue is what constitutes actual open source. This is actually open source, and it performs well. If you’re familiar with the space, then it’s a big deal.
Is it really or is it just a binary release like everything else?
Everything is explained and linked in the project, so…
I have yet to see a 3B model that’s not dumb.
Got it up and running on a Debian distrobox… now I need to figure out how to train it. Will be my first steps into this type of thing – so prob will take me a bit to figure out how it all works
Smart people, I beg of thee, explain! What can it do?
Edit: looks to be another text based one, not image generation right?
It’s language only, hence, LM
To be fair, I didn’t know if that language included programming language, and thus maybe still consider image based AI to be included in LLM. Is there a different designation for the type of AI that does image generation?
The problem is… How do we run it if rocm is still a mess for most of their gpus? Cpu time?
There are ROCm versions of llama.cpp, ollama, and kobold.cpp that work well, although they’ll have to add support for this model before they could run it.