unofficial mirror of emacs-tangents@gnu.org
 help / color / mirror / Atom feed
From: Ihor Radchenko <yantar92@posteo.net>
To: chad <yandros@gmail.com>
Cc: emacs-tangents@gnu.org, Jim Porter <jporterbugs@gmail.com>,
	ahyatt@gmail.com, rms@gnu.org
Subject: Re: [NonGNU ELPA] New package: llm
Date: Fri, 01 Sep 2023 09:53:22 +0000	[thread overview]
Message-ID: <87pm32rzhp.fsf@localhost> (raw)
In-Reply-To: <CAO2hHWZ9f2fwr9NbtABFPuUv+Qsk6dmzhavwKV3BJ4P-YNEsOw@mail.gmail.com>

chad <yandros@gmail.com> writes:

> For large AI models specifically: there are many users for whom it is not
> practical to _actually_ recreate the model from scratch everywhere they
> might want to use it. It is important for computing freedom that such
> recreations be *possible*, but it will be very limiting to insist that
> everyone who wants to use such services actually do so, in a manner that
> seems to me to be very similar to not insisting that every potential emacs
> user actually compile their own. In this case there's the extra wrinkle
> that the actual details of recreating the currently-most-interesting large
> language models involves both _gigantic_ amounts of resources and also a
> fairly large amount of not-directly-reproducible randomness involved. It
> might be worth further consideration.

Let me refer to another message by RMS:

    >>   > While I certainly appreciate the effort people are making to produce 
    >>   > LLMs that are more open than OpenAI (a low bar), I'm not sure if 
    >>   > providing several gigabytes of model weights in binary format is really 
    >>   > providing the *source*. It's true that you can still edit these models 
    >>   > in a sense by fine-tuning them, but you could say the same thing about a 
    >>   > project that only provided the generated output from GNU Bison, instead 
    >>   > of the original input to Bison.
    >> 
    >> I don't think that is valid.
    >> Bison processing is very different from training a neural net.
    >> Incremental retraining of a trained neural net
    >> is the same kind of processing as the original training -- except
    >> that you use other data and it produces a neural net
    >> that is trained differently.
    >> 
    >> My conclusiuon is that the trained neural net is effectively a kind of
    >> source code.  So we don't need to demand the "original training data"
    >> as part of a package's source code.  That data does not have to be
    >> free, published, or available.

-- 
Ihor Radchenko // yantar92,
Org mode contributor,
Learn more about Org mode at <https://orgmode.org/>.
Support Org development at <https://liberapay.com/org-mode>,
or support my work at <https://liberapay.com/yantar92>



  reply	other threads:[~2023-09-01  9:53 UTC|newest]

Thread overview: 13+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
     [not found] <CAM6wYYJHa+tCUKO_SsnT77g-4MUM0x4FrkoCekr=T9-UF1ADDA@mail.gmail.com>
     [not found] ` <E1qTaA2-00038O-UA@fencepost.gnu.org>
     [not found]   ` <CAM6wYY+E=z5VqV2xXMbhbpN7vn+-tyzfOGKFAuG0s+croRmEPA@mail.gmail.com>
     [not found]     ` <E1qV08g-0001mb-11@fencepost.gnu.org>
     [not found]       ` <CAM6wYYLZ26E4rpo2Ae2PyxKSBYQKAXQ6U5_QGMoGx5SQy7AMSA@mail.gmail.com>
     [not found]         ` <87v8d0iqa5.fsf@posteo.net>
     [not found]           ` <E1qaR6l-00012I-VP@fencepost.gnu.org>
     [not found]             ` <CAM6wYYLYrQL9+3cgUELYavUdHQg5m0bqdW89_qJFvk050-sGNQ@mail.gmail.com>
     [not found]               ` <fd98dcaf-5016-1a84-f281-36ef6eb108c5@gmail.com>
     [not found]                 ` <E1qbX8C-0004EP-3M@fencepost.gnu.org>
     [not found]                   ` <87cyz3vaws.fsf@localhost>
2023-08-31 16:29                     ` [NonGNU ELPA] New package: llm chad
2023-09-01  9:53                       ` Ihor Radchenko [this message]
     [not found]                     ` <E1qcyN3-0001al-5t@fencepost.gnu.org>
2023-09-06 12:51                       ` Is ChatGTP SaaSS? (was: [NonGNU ELPA] New package: llm) Ihor Radchenko
2023-09-06 16:59                         ` Andrew Hyatt
2023-09-09  0:37                           ` Richard Stallman
2023-09-06 22:52                         ` Emanuel Berg
2023-09-07  7:28                           ` Lucien Cartier-Tilet
2023-09-07  7:57                             ` Emanuel Berg
2023-09-09  0:38                         ` Richard Stallman
2023-09-09 10:28                           ` Collaborative training of Libre LLMs (was: Is ChatGTP SaaSS? (was: [NonGNU ELPA] New package: llm)) Ihor Radchenko
2023-09-09 11:19                             ` Jean Louis
2023-09-10  0:22                             ` Richard Stallman
2023-09-10  2:18                               ` Debanjum Singh Solanky

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

  List information: https://www.gnu.org/software/emacs/

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=87pm32rzhp.fsf@localhost \
    --to=yantar92@posteo.net \
    --cc=ahyatt@gmail.com \
    --cc=emacs-tangents@gnu.org \
    --cc=jporterbugs@gmail.com \
    --cc=rms@gnu.org \
    --cc=yandros@gmail.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for read-only IMAP folder(s) and NNTP newsgroup(s).