Submitted by manuelfraile t3_12327d1 in MachineLearning
I've recently discovered models such as ChatLLaMA that allows you to create a "ChatGPT" but you need Meta's LLaMA weights (yes, you can find them in torrents but that's not the point of the question). Similar limitations found in other cases.
Therefore I wanted to try to find an open source: dataset (in addition to hugging face), "base model", "chat model" AND that it is feasible to train with a commercial computer with a very good GPU (NVIDIA, etc.). With this get at least decent results.
Also would be interesting to distinguish between solutions with commercial limitations and those who don't.
Thanks!
• EDIT • A first solution I already found is this: https://github.com/databrickslabs/dolly based on this https://huggingface.co/EleutherAI/gpt-j-6B, but looking for some discussion and perhaps other/better solutions.
Hands0L0 t1_jdsvv9h wrote
You and everyone else here