yehiaserag
yehiaserag t1_jed4dee wrote
Reply to [P] Introducing Vicuna: An open-source language model based on LLaMA 13B by Business-Lead2679
I'm lost, it says open-source... and I can't see any mentioning of the weights, a download link or a huggingface repo.
On the website it says "We plan to release the model weights by providing a version of delta weights that build on the original LLaMA"
Please no lora for that, lora is always associated with degraded inference quality.
yehiaserag t1_jcrz2qt wrote
Reply to comment by yehiaserag in [P] The next generation of Stanford Alpaca by [deleted]
seems very similar
yehiaserag t1_jcru1ty wrote
Reply to [P] The next generation of Stanford Alpaca by [deleted]
Do you have a repo, a website, anywhere we can follow the progress of this?
yehiaserag t1_jcm31zk wrote
Reply to comment by [deleted] in [R] RWKV 14B ctx8192 is a zero-shot instruction-follower without finetuning, 23 token/s on 3090 after latest optimization (16G VRAM is enough, and you can stream layers to save more VRAM) by bo_peng
We say RWKV for short, the rest of the stuff is for a specific version
yehiaserag t1_jcj305q wrote
Reply to [R] RWKV 14B ctx8192 is a zero-shot instruction-follower without finetuning, 23 token/s on 3090 after latest optimization (16G VRAM is enough, and you can stream layers to save more VRAM) by bo_peng
How does that version compare to "RWKV-4-Pile-14B-20230228-ctx4096-test663"?
yehiaserag t1_iuendqq wrote
Reply to comment by Kaarssteun in Talked to people minimizing/negating potential AI impact in their field? eg: artists, coders... by kmtrp
Let's jerk together bro
yehiaserag t1_jegqni6 wrote
Reply to comment by light24bulbs in [P] Introducing Vicuna: An open-source language model based on LLaMA 13B by Business-Lead2679
There are lots of comparisons that show this, this is why ppl created alpaca native, to reach the quality described in the original paper