sanxiyn
sanxiyn t1_jcw2yoz wrote
Reply to comment by clueless1245 in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
On the other hand, commercial use restriction is not compatible with generally accepted definition of open source, for example The Open Source Definition.
> 6) No Discrimination Against Fields of Endeavor. The license must not restrict anyone from making use of the program in a specific field of endeavor. For example, it may not restrict the program from being used in a business, or from being used for genetic research.
sanxiyn t1_jc598b3 wrote
Reply to comment by modeless in [R] Stanford-Alpaca 7B model (an instruction tuned version of LLaMA) performs as well as text-davinci-003 by dojoteef
Eh, authors do claim they performed blind comparison and "Alpaca wins 90 versus 89 comparisons against text-davinci-003". They also released evaluation set used.
sanxiyn t1_jd68827 wrote
Reply to [D] Running an LLM on "low" compute power machines? by Qwillbehr
You don't need leaked LLaMA weight. ChatGLM-6B weight is being distributed by the first party.