From mboxrd@z Thu Jan 1 00:00:00 1970 Path: news.gmane.io!.POSTED.blaine.gmane.org!not-for-mail From: Jean Louis Newsgroups: gmane.emacs.tangents Subject: Re: Help building Pen.el (GPT for emacs) Date: Mon, 19 Jul 2021 20:00:47 +0300 Message-ID: References: <83im1948mj.fsf@gnu.org> <83lf642jeh.fsf@gnu.org> Mime-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit Injection-Info: ciao.gmane.io; posting-host="blaine.gmane.org:116.202.254.214"; logging-data="36989"; mail-complaints-to="usenet@ciao.gmane.io" User-Agent: Mutt/2.0.7+183 (3d24855) (2021-05-28) Cc: Eli Zaretskii , emacs-tangents@gnu.org, Stefan Kangas , rms@gnu.org To: Shane Mulligan Original-X-From: emacs-tangents-bounces+get-emacs-tangents=m.gmane-mx.org@gnu.org Mon Jul 19 19:05:20 2021 Return-path: Envelope-to: get-emacs-tangents@m.gmane-mx.org Original-Received: from lists.gnu.org ([209.51.188.17]) by ciao.gmane.io with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.92) (envelope-from ) id 1m5Whk-0009Qa-An for get-emacs-tangents@m.gmane-mx.org; Mon, 19 Jul 2021 19:05:20 +0200 Original-Received: from localhost ([::1]:39578 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1m5Whj-0000Xl-8t for get-emacs-tangents@m.gmane-mx.org; Mon, 19 Jul 2021 13:05:19 -0400 Original-Received: from eggs.gnu.org ([2001:470:142:3::10]:39854) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1m5Wgw-0000WU-Pf for emacs-tangents@gnu.org; Mon, 19 Jul 2021 13:04:32 -0400 Original-Received: from stw1.rcdrun.com ([217.170.207.13]:51419) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1m5Wgr-00054s-11; Mon, 19 Jul 2021 13:04:28 -0400 Original-Received: from localhost ([::ffff:197.157.0.54]) (AUTH: PLAIN admin, TLS: TLS1.3,256bits,ECDHE_RSA_AES_256_GCM_SHA384) by stw1.rcdrun.com with ESMTPSA id 0000000000057F2A.0000000060F5B095.00005793; Mon, 19 Jul 2021 10:04:20 -0700 Content-Disposition: inline In-Reply-To: Received-SPF: pass client-ip=217.170.207.13; envelope-from=bugs@gnu.support; helo=stw1.rcdrun.com X-Spam_score_int: 29 X-Spam_score: 2.9 X-Spam_bar: ++ X-Spam_report: (2.9 / 5.0 requ) BAYES_00=-1.9, RCVD_IN_SBL_CSS=3.335, RCVD_IN_SORBS_WEB=1.5, SPF_HELO_PASS=-0.001, SPF_PASS=-0.001 autolearn=no autolearn_force=no X-Spam_action: no action X-BeenThere: emacs-tangents@gnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: Emacs news and miscellaneous discussions outside the scope of other Emacs mailing lists List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: emacs-tangents-bounces+get-emacs-tangents=m.gmane-mx.org@gnu.org Original-Sender: "Emacs-tangents" Xref: news.gmane.io gmane.emacs.tangents:660 Archived-At: * Shane Mulligan [2021-07-18 11:01]: > Pen.el stands for Prompt Engineering in emacs. > Prompt Engineering is the art of describing what you would > like a language model (transformer) to do. It is a new type of programming, > example oriented; > like literate programming, but manifested automatically. Sounds like a replacement for a programmer's mind. > A transformer takes some text (called a prompt) and continues > it. However, the continuation is the superset of all NLP tasks, Where is definition of the abbreviation NLP? > as the generation can also be a classification, for instance. Those > NLP tasks extend beyond world languages and into programming > languages (whatever has been 'indexed' or 'learned') from these > large LMs. What is definition of the abbreviation LM? > Pen.el is an editing environment for designing 'prompts' to LMs. It > is better than anything that exists, even at OpenAI or at > Microsoft. I have been working on it and preparing for this for a > long time. Good, but is there a video to show what it really does? > These prompts are example- based tasks. There are a number of design > patterns which Pen.el is seeking to encode into a domain-specific > language called 'examplary' for example- oriented programming. Do you mean "exemplary" or "examplary", is it spelling mistake? I have to ask as your description is still pretty abstract without particular example. > Pen.el creates functions 1:1 for a prompt to an emacs lisp function. The above does not tell me anything. > Emacs is Grammarly, Google Translate, Copilot, Stackoveflow and > infinitely many other services all rolled into one and allows you to > have a private parallel to all these services that is completely > private and open source -- that is if you have downloaded the > EleutherAI model locally. I understand that it is kind of fetching information, but that does not solve licensing issues, it sounds like licensing hell. > ** Response to Jean Louis > - And I do not think it should be in GNU ELPA due to above reasons. > > I am glad I have forewarned you guys. This is my current goal. Help > in my project would be appreciated. I cannot do it alone and I > cannot convince all of you. Why don't you tell about licensing issues? Taking code without proper licensing compliance is IMHO, not an option. It sounds as problem generator. > > Why don't you simply make an Emacs package as .tar as described in Emacs > Lisp manual? > Thank you for taking a look at my emacs package. It's not ready net > for Melpa merge. I hope that I will be able to find some help in > order to prepare it, but the rules are very strict and this may not > happen. I did not say to put it in Melpa. Package you can make for yourself and users so that users can M-x package-install-file That is really not related to any online Emacs package repository. It is way how to install Emacs packages no matter where one gets it. > > How does that solves the licensing problems? > The current EleutherAI model which competes with GPT-3 is GPT-Neo. > It is MIT licensed. That is good. But the code that is generated and injected requires proper contribution. > Also the data it has been trained on is MIT licensed. Yes, and then the program should also solve the proper contributions automatically. You cannot just say "MIT licensed", this has to be proven, source has to be found and proper attributions applied. Why don't you implement proper licensing? Please find ONE license that you are using from code that is being used as database for generation of future code and provide link to it. Then show how is license complied to. > The current EleutherAI model which competes with Codex is GPT-j. > It is licensed with Apache-2.0 License That is good, but I am referring to the generated code. -- Jean Take action in Free Software Foundation campaigns: https://www.fsf.org/campaigns In support of Richard M. Stallman https://stallmansupport.org/