Submitted by super_deap t3_11tmpc5 in MachineLearning
kittenkrazy t1_jcjxr0b wrote
I wonder if this can effectively be used in LLaMA. 32K context would be a game changer
Nhabls t1_jck9a4c wrote
Yeah just need enough training time and data to be able to train those 32k context layers effectively........................
fastinguy11 t1_jcle8cn wrote
Gpt4 32 k api when available ?
mrpogiface t1_jckmi7d wrote
Definitely, but you'd need to further fine-tune the model to "teach" it to make use of the additional context
super_deap OP t1_jcktps3 wrote
This
kreuzguy t1_jcliuwx wrote
Someone should definitely look into this!
Viewing a single comment thread. View all comments