Bulky_Highlight_3352
Bulky_Highlight_3352 t1_jc34398 wrote
Reply to comment by cyvr_com in [R] Stanford-Alpaca 7B model (an instruction tuned version of LLaMA) performs as well as text-davinci-003 by dojoteef
nada, last commit last week
Bulky_Highlight_3352 t1_jc33l11 wrote
Reply to comment by cyvr_com in [R] Stanford-Alpaca 7B model (an instruction tuned version of LLaMA) performs as well as text-davinci-003 by dojoteef
source?
Bulky_Highlight_3352 t1_jc31exq wrote
Reply to [R] Stanford-Alpaca 7B model (an instruction tuned version of LLaMA) performs as well as text-davinci-003 by dojoteef
really nice, thanks for sharing.
The license is still limited to non-commercial use due to model being fine-tuned LLaMA.
>We emphasize that Alpaca is intended only for academic research and any commercial use is prohibited. There are three factors in this decision: First, Alpaca is based on LLaMA, which has a non-commercial license, so we necessarily inherit this decision. Second, the instruction data is based OpenAI's text-davinci-003, whose terms of use prohibit developing models that compete with OpenAI. Finally, we have not designed adequate safety measures, so Alpaca is not ready to be deployed for general use.
Bulky_Highlight_3352 t1_jc0bp3s wrote
"Hey, we made this commercial tool that is better than open source!"
Bulky_Highlight_3352 t1_jbyxv8s wrote
I've tried Aleph's playground and mostly saw it generate complete garbage. Not sure how they will solve any of the ChatGPT's problems.
Bulky_Highlight_3352 t1_j3hsmp7 wrote
Reply to comment by mandogbeer in [Project] Major drawback/limitation of GPT-3 by trafalgar28
I believe so, I think you can experiment with different summarization prompts too. For me it is still trial and error when dealing with large context windows.
Bulky_Highlight_3352 t1_j3fqhij wrote
There are tools to work around this limitation such as LangChain with its support of summarization of previous context https://github.com/hwchase17/langchain
Bulky_Highlight_3352 t1_iujz3ty wrote
Reply to [News] The Stack: 3 TB of permissively licensed source code - Hugging Face and ServiceNow Research Denis Kocetkov et al 2022 by Singularian2501
impressive, bash automation, here I come
Bulky_Highlight_3352 t1_jc4ajcf wrote
Reply to comment by lxe in [R] Stanford-Alpaca 7B model (an instruction tuned version of LLaMA) performs as well as text-davinci-003 by dojoteef
inference code is, the model weights are under a separate non-commercial license