Submitted by to4life4 t3_11zzgzc in MachineLearning
wendten t1_jdfh6ya wrote
best is a very vague term. Do you have access to a gpu cluster, or do you plan to run it on an office laptop. However id say the Alpaca model would be a good candidate. you can follow their guidance and make your own custom model from one of metas Llama models
sweatierorc t1_jdfwh5f wrote
LLama is not open-source (though it is gratis).
Edit: typo
ironmagnesiumzinc t1_jdfjdbj wrote
When fine-tuning alpaca, would it make sense to use an unsupervised approach with raw text data (say only 100mb) or would a supervised QA approach be way better?
to4life4 OP t1_jdfhhup wrote
"Best" I suppose meaning closest to the latest ChatGPT on the usual benchmarks.
First on my own gpu to test (3080ti), then on a cluster if I can prove out the concept.
Thanks I'll definitely look into Alpaca. It can be customized to work with human ratings of generated output?
Viewing a single comment thread. View all comments