big_ol_tender
big_ol_tender t1_jdy0c6t wrote
Reply to comment by sad_dad_is_a_mad_lad in [D] Instruct Datasets for Commercial Use by JohnyWalkerRed
100% agree but for those of us working for a company I can’t knowingly open us up to that risk even if the probability is 1%
big_ol_tender t1_jdvu92g wrote
Thank you for posting this. I’ve raised this issue on a number of threads and even opened an issue on the alpaca repo. Everyone seems to ignore this and I’m worried about downstream issues with these models, and would love an open source alternative ( have been exploring making one myself).
big_ol_tender t1_jdsvods wrote
Reply to [D] Build a ChatGPT from zero by manuelfraile
Lol
big_ol_tender t1_jdoe9k1 wrote
Alpaca dataset is not open source so alpaca-lora is not open source.
big_ol_tender t1_jdjtwdk wrote
Reply to comment by danielbln in [R] Hello Dolly: Democratizing the magic of ChatGPT with open models by austintackaberry
Pls do! I believe in u
big_ol_tender t1_jdjl1wx wrote
Reply to comment by Colecoman1982 in [R] Hello Dolly: Democratizing the magic of ChatGPT with open models by austintackaberry
I opened an issue on GitHub specifically about the data license and linked to the data bricks release :)
big_ol_tender t1_jdjcfc8 wrote
The alpaca dataset has a no commercial license so idk what they are doing.. I’ve asked Stanford to change it but heard nothing back
big_ol_tender t1_jdefnc7 wrote
Reply to comment by Difficult_Bid_9828 in [P] ChatLLaMA - A ChatGPT style chatbot for Facebook's LLaMA by imgonnarelph
thanks for this- very cool project, this indeed solves the issue with LLama weights but nfortunately the issue remains with the alpaca dataset license itself being non commercial:
https://github.com/tatsu-lab/stanford_alpaca/blob/main/DATA_LICENSE
big_ol_tender t1_jdct16f wrote
I’d love to try this out but isn’t there an issue with licensing? OpenAI said you can’t use their model output to train competitors to chatgpt (which is total BS) and the alpaca dataset is all davinci output. I’m desperately trying to find some open source alternative that I can use for some experiments at work because I don’t want to give closedai any $.
big_ol_tender t1_jc4lrqf wrote
Reply to [D] ChatGPT without text limits. by spiritus_dei
This makes me depressed because I’ve been working with the llama-index project and I feel like these huge companies are going to take my ball away 😢. They just have too many resources to build stuff.
big_ol_tender t1_jbm8s9r wrote
Literally none of them
big_ol_tender t1_jatvx64 wrote
If you have tabular data just use xgboost, forget the nn
big_ol_tender t1_jakmlmc wrote
Reply to comment by caedin8 in [D] OpenAI introduces ChatGPT and Whisper APIs (ChatGPT API is 1/10th the cost of GPT-3 API) by minimaxir
-totally not chatgpt
big_ol_tender t1_j89ruh6 wrote
If you haven’t already, I’d suggest the 3blue1brown series on neural networks on YouTube. It is the easiest introduction I’ve come across.
big_ol_tender t1_j89k4f7 wrote
Thanks for putting this together. I’d add deepsparse and sparsezoo for training/deploying sparse models. Also I can’t vouch for it because I haven’t used it (yet) but DVC (data version control) for ML Dev
big_ol_tender t1_j7vx88w wrote
Reply to [D]Image Recognition ability of machine learning in financial markets questions by Ready-Acanthaceae970
Hundreds of firms have already tried this years ago- there is no alpha left in this type of information. Source: work at a hedge fund and have friends at two sigma, citadel, and D. E. Shaw. If you want to do it as an academic exercise go ahead, but this is not how you make money.
big_ol_tender t1_jecny45 wrote
Reply to [P] Introducing Vicuna: An open-source language model based on LLaMA 13B by Business-Lead2679
Stop claiming fine tuned llama models as open source. It’s not open source and we can’t use them for anything real.