• TheGrandNagus@lemmy.world
    link
    fedilink
    English
    arrow-up
    6
    ·
    2 months ago

    Fully open and accessible: Fully open-source release of model weights, training hyperparameters, datasets, and code, fostering innovation and collaboration within the AI community.

    That’s actually pretty good. Seems to be open source as the OSI defines it, rather than the much more common “this model is open source, but the dataset is a secret”.

      • I’m sure; I just don’t know how. I need to set aside some time and educate myself.

        Frankly, this generation of AI I find rather dull. It won’t directly lead a AGI, although I’m sure it’ll be a component, but I think that’ll be another 10-20 years before the next breakthrough. I personally don’t think it’s as interesting as the symbolic, knowledge-based systems of the mid-80’s; at least those were reasoning systems. LLMs look impressive to lay people (including myself - I understand the general concepts, but have no experience with the programming or training, so I’m just another lay user), but there’s no reasoning or understanding behind it, and if what it produces is truthful or accurate, it’s largely on accident. So I’ve had trouble getting excited about it.

  • HappyFrog@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    1
    ·
    2 months ago

    I see all these graphs about how much better this LLM is than another, but do those graphs actually translate to real world usefulness?

  • Rando@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    1
    ·
    2 months ago

    Got it up and running on a Debian distrobox… now I need to figure out how to train it. Will be my first steps into this type of thing – so prob will take me a bit to figure out how it all works

  • GaMEChld@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    2 months ago

    Smart people, I beg of thee, explain! What can it do?

    Edit: looks to be another text based one, not image generation right?

  • brokenlcd@feddit.it
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    The problem is… How do we run it if rocm is still a mess for most of their gpus? Cpu time?

    • swelter_spark@reddthat.com
      link
      fedilink
      English
      arrow-up
      1
      ·
      4 days ago

      There are ROCm versions of llama.cpp, ollama, and kobold.cpp that work well, although they’ll have to add support for this model before they could run it.