Submitted by Not-Banksy t3_126a1dm in singularity
ActuatorMaterial2846 t1_je8luak wrote
Reply to comment by jetro30087 in When people refer to “training” an AI, what does that actually mean? by Not-Banksy
Interesting, curious what size this particular Llama model is, or is that not even relevant?
jetro30087 t1_je8mtjp wrote
This is a updated dataset for the 7b model, but you could train the others with the data. From anecdotal reports, the dataset seems to have a great impact on the model's performance than the parameter size up to a point. Less parameters means a faster model. More parameters mean the model can make longer responses.
Viewing a single comment thread. View all comments