With the OSI publishing their abysmal - explicitly not open source - “Open Source AI” definition I thought I’d post my argument, why it is bad and why “Open Source AI” currently probably does not exist.

  • bitofhope@awful.systems
    link
    fedilink
    English
    arrow-up
    23
    ·
    2 months ago

    The stretching is just so blatant. People who train neural networks do not write a bunch of tokens and weights. They take a corpus of training data and run a training program to generate the weights. That’s why it is the training program and the corpus that should be considered the source form of the program. If either of these can’t be made available in a way that allows redistribution of verbatim and modified versions, it can’t be open source. Even if I have a powerful server farm and a list of data sources for Llama 3, I can’t replicate the model myself without committing copyright infringement (neither could Facebook for that matter, and that’s not an entirely separate issue).

    There are large collections of freely licensed and public domain media that could theoretically be used to train a model, but that model surely wouldn’t be as big as the proprietary ones. In some sense truly open source AI does exist and has for a long time, but that’s not the exciting thing OSI is lusting after, is it?

    • BlueMonday1984@awful.systems
      link
      fedilink
      English
      arrow-up
      9
      ·
      2 months ago

      I’ve already talked about the indirect damage AI’s causing to open source in this thread, but this hyper-stretched definition’s probably doing some direct damage as well.

      Considering that this “Open Source AI” definition is (almost certainly by design) going to openwash the shit out of blatant large-scale theft, I expect it’ll heavily tar the public image of open-source, especially when these “Open Source AIs” start getting sued for copyright infringement.

    • JFranek@awful.systems
      link
      fedilink
      English
      arrow-up
      7
      ·
      2 months ago

      Yeah, neural network training is notoriously easy to reproduce /s.

      Just few things can affect results: source data, data labels, network structure, training parameters, version of training script, versions of libraries, seed for random number generator, hardware, operating system.

      Also, deployment is another can of worms.

      Also, even if you have open source script, data and labels, there’s no guarantee you’ll have useful documentation for either of these.

    • V0ldek@awful.systems
      link
      fedilink
      English
      arrow-up
      3
      ·
      edit-2
      2 months ago

      People who train neural networks do not write a bunch of tokens and weights.

      Reading this made me think of an analogy of generated code. This is basically exactly the same thing as distributing the code of your program but not in the source language, rather the assembly listing of the final binary, and calling it open source. You can turn any defense of the AI model of “open-source” into a defense of that model of distributing code. You can run my AI/code (if you have a powerful/similar enough machine), you can inspect it (it’s just not going to tell you anything), you can modify it (lol), so it’s open source!

      Edit: The more I think about it the more I come to the realisation that the assembly listing is actually still vastly more useful than the AI models. Like at least a very dedicated and insane enough programmer could technically track down a bug in the assembly and correct it if given enough coffee.