From: Ihor Radchenko <yantar92@posteo.net>
To: chad <yandros@gmail.com>
Cc: emacs-tangents@gnu.org, Jim Porter <jporterbugs@gmail.com>,
ahyatt@gmail.com, rms@gnu.org
Subject: Re: [NonGNU ELPA] New package: llm
Date: Fri, 01 Sep 2023 09:53:22 +0000 [thread overview]
Message-ID: <87pm32rzhp.fsf@localhost> (raw)
In-Reply-To: <CAO2hHWZ9f2fwr9NbtABFPuUv+Qsk6dmzhavwKV3BJ4P-YNEsOw@mail.gmail.com>
chad <yandros@gmail.com> writes:
> For large AI models specifically: there are many users for whom it is not
> practical to _actually_ recreate the model from scratch everywhere they
> might want to use it. It is important for computing freedom that such
> recreations be *possible*, but it will be very limiting to insist that
> everyone who wants to use such services actually do so, in a manner that
> seems to me to be very similar to not insisting that every potential emacs
> user actually compile their own. In this case there's the extra wrinkle
> that the actual details of recreating the currently-most-interesting large
> language models involves both _gigantic_ amounts of resources and also a
> fairly large amount of not-directly-reproducible randomness involved. It
> might be worth further consideration.
Let me refer to another message by RMS:
>> > While I certainly appreciate the effort people are making to produce
>> > LLMs that are more open than OpenAI (a low bar), I'm not sure if
>> > providing several gigabytes of model weights in binary format is really
>> > providing the *source*. It's true that you can still edit these models
>> > in a sense by fine-tuning them, but you could say the same thing about a
>> > project that only provided the generated output from GNU Bison, instead
>> > of the original input to Bison.
>>
>> I don't think that is valid.
>> Bison processing is very different from training a neural net.
>> Incremental retraining of a trained neural net
>> is the same kind of processing as the original training -- except
>> that you use other data and it produces a neural net
>> that is trained differently.
>>
>> My conclusiuon is that the trained neural net is effectively a kind of
>> source code. So we don't need to demand the "original training data"
>> as part of a package's source code. That data does not have to be
>> free, published, or available.
--
Ihor Radchenko // yantar92,
Org mode contributor,
Learn more about Org mode at <https://orgmode.org/>.
Support Org development at <https://liberapay.com/org-mode>,
or support my work at <https://liberapay.com/yantar92>
next prev parent reply other threads:[~2023-09-01 9:53 UTC|newest]
Thread overview: 13+ messages / expand[flat|nested] mbox.gz Atom feed top
[not found] <CAM6wYYJHa+tCUKO_SsnT77g-4MUM0x4FrkoCekr=T9-UF1ADDA@mail.gmail.com>
[not found] ` <E1qTaA2-00038O-UA@fencepost.gnu.org>
[not found] ` <CAM6wYY+E=z5VqV2xXMbhbpN7vn+-tyzfOGKFAuG0s+croRmEPA@mail.gmail.com>
[not found] ` <E1qV08g-0001mb-11@fencepost.gnu.org>
[not found] ` <CAM6wYYLZ26E4rpo2Ae2PyxKSBYQKAXQ6U5_QGMoGx5SQy7AMSA@mail.gmail.com>
[not found] ` <87v8d0iqa5.fsf@posteo.net>
[not found] ` <E1qaR6l-00012I-VP@fencepost.gnu.org>
[not found] ` <CAM6wYYLYrQL9+3cgUELYavUdHQg5m0bqdW89_qJFvk050-sGNQ@mail.gmail.com>
[not found] ` <fd98dcaf-5016-1a84-f281-36ef6eb108c5@gmail.com>
[not found] ` <E1qbX8C-0004EP-3M@fencepost.gnu.org>
[not found] ` <87cyz3vaws.fsf@localhost>
2023-08-31 16:29 ` [NonGNU ELPA] New package: llm chad
2023-09-01 9:53 ` Ihor Radchenko [this message]
[not found] ` <E1qcyN3-0001al-5t@fencepost.gnu.org>
2023-09-06 12:51 ` Is ChatGTP SaaSS? (was: [NonGNU ELPA] New package: llm) Ihor Radchenko
2023-09-06 16:59 ` Andrew Hyatt
2023-09-09 0:37 ` Richard Stallman
2023-09-06 22:52 ` Emanuel Berg
2023-09-07 7:28 ` Lucien Cartier-Tilet
2023-09-07 7:57 ` Emanuel Berg
2023-09-09 0:38 ` Richard Stallman
2023-09-09 10:28 ` Collaborative training of Libre LLMs (was: Is ChatGTP SaaSS? (was: [NonGNU ELPA] New package: llm)) Ihor Radchenko
2023-09-09 11:19 ` Jean Louis
2023-09-10 0:22 ` Richard Stallman
2023-09-10 2:18 ` Debanjum Singh Solanky
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
List information: https://www.gnu.org/software/emacs/
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=87pm32rzhp.fsf@localhost \
--to=yantar92@posteo.net \
--cc=ahyatt@gmail.com \
--cc=emacs-tangents@gnu.org \
--cc=jporterbugs@gmail.com \
--cc=rms@gnu.org \
--cc=yandros@gmail.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for read-only IMAP folder(s) and NNTP newsgroup(s).