Just to be precise on llama, what I proposed was to include the port of Facebook code to CPP, (llama.cpp, see ticket 62443 on guix-patches), which itself has a license. The weight themselves indeed d
Hi there FSF Licensing! (CC: Guix devel, Nicholas Graves) This morning I read through the FSDG to see if it gives any guidance on when machinelearning model weights are appropriate for inclusion in
Hi, Well, we already discussed in GWL context where to put “large” data set without reaching a conclusion. Having “large” data set inside the store is probably not a good idea. But maybe thes
Hi Guix! I've recently contributed a few tools that make a few OSS machinelearning programs usable for Guix, namely nerd-dictation for dictation and llama-cpp as a converstional bot. In the first ca
Hi, Hum, I am probably not this Someone™ but here the result of my looks. :-) First, please note that the Debian thread [1] is about, Concerns to software freedom when packaging deep-learning based
re-learning because how to draw the line between biased weights, mistakes on their side, mistakes on our side, etc. and it requires a high level of expertise to complete a full re-learning. This str
Hi, Feel free to pick real-world model using 15 billions of parameters and then to train it again. And if you succeed, feel free to train it again to have bit-to-bit reproducibility. Bah the cost (CP
Hi, Years ago, I asked to FSF and Stallman how to deal with that and I had never got an answer back. Anyway! :-) Debian folks discussed such topic [1,2] but I do not know if they have an “official
Hi, Thanks for pointing this article! And some non-mathematical part of the original article [1] are also worth to give a look. :-) First please note that we are somehow in the case “The Open Box
For a more concrete example, with facial reconition in particular, many models are quite good at recognition of faces of people of predominantly white european descent, and not very good with people
I feel like, although this might (arguably) not be the case for leela-zero nor Lc0 specifically, for certain machinelearning projects, a pretrained network can affect the program’s behavior so de
Hi, Someone™ has to invest time in studying this specific case, look at what others like Debian are doing, and seek consensus on a way forward. Based on that, perhaps Someone™ can generalize that
Hi Ludo, Your concern in this thread was: My point is about whether these trained neural network data are something that we could distribute per the FSDG. https://issues.guix.gnu.org/36071#3-lineno21
I feel like it’s important to have a guideline for this, at least if the issue becomes recurrent too frequently. To me, a sensible *base criterion* is whether the user is able to practically produ
Hello, We discussed it in 2019: https://issues.guix.gnu.org/36071 This LWN article on the debate that then took place in Debian is insightful: https://lwn.net/Articles/760142/ To me, there is no doub
Hi, Well, I do not know if we have reached a conclusion. From my point of view, both can be included *if* their licenses are compatible with Free Software – included the weights (pre-trained model)
Hello, zamfofex submited a package 'lc0', Leela Chess Zero” (a chess engine) with ML model, also it turn out that we already had 'stockfish' a similiar one with pre-trained model packaged. Does we
Yeah, I didn't mean to give the impression that I thought bit-reproducibility was the silver bullet for AI backdoors with that analogy. I guess my argument is this: if they release the training info,