• astro_ray@lemdro.id
    link
    fedilink
    English
    arrow-up
    14
    arrow-down
    1
    ·
    5 months ago

    I feel like one of thr problem is LLMs hijacked the definition of AI. Like another comment said, the way they trained on copyrighted material, it’s probably not possible. But imagine there was another model (not necessarily LLM) and it was trained with completely public domain material. For example maybe something trained to find genetic diseases from genetic samples of a person, or detecting asteroids from telescope images. Those could become open source. Now, I am not an expert, but do we consider those AI?

    • chebra@mstdn.io
      link
      fedilink
      arrow-up
      5
      arrow-down
      2
      ·
      5 months ago

      @astro_ray @marvelous_coyote It seems you have the incorrect idea about what open-source means, which is quite sad here in the open-source lemmy community. Being trained on public domain material does NOT make the model open-source. It’s about the license - what the recipients of the model are allowed to do with it - open-source must allow derivative works and commercial use, on top of seeing the code, but for LLM models the “code” is just a bunch of float numbers, nothing interesting to see.