Submitted by floppy_llama t3_1266d02 in MachineLearning
A_Light_Spark t1_jeaim48 wrote
Reply to comment by saintshing in [R] LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention by floppy_llama
The real vip is in the comments again. TIL about rwkv!
Now I just need to read up on it and see if it can do sequence classification...
saintshing t1_jeaowjz wrote
I almost missed it too. There are too many new results.
The most crazy thing is it is all done by one person when the big techs all work on transformer models.
Viewing a single comment thread. View all comments