[R] LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention arxiv.org Submitted by floppy_llama t3_1266d02 on March 30, 2023 at 12:46 AM in MachineLearning 47 comments 233
_Arsenie_Boca_ t1_je8km8c wrote on March 30, 2023 at 4:56 AM Very interesting work! Though I find the explanation of the concrete approach (how the additional parameters are used by the LM) to be a bit vague. Does anyone have a deeper understanding? Is it using regular adapters? Permalink 4
Viewing a single comment thread. View all comments