Submitted by Cool_Abbreviations_9 t3_123b66w in MachineLearning
ntaylor- t1_je5qtl2 wrote
Reply to comment by was_der_Fall_ist in [D]GPT-4 might be able to tell you if it hallucinated by Cool_Abbreviations_9
Nope. It's the same as all neural networks using transformer architecture. Just a big old series of matrix multiplications with some non linear transformations at end of the day
was_der_Fall_ist t1_je6lfl9 wrote
Why are matrix multiplications mutually exclusive with complicated operations?
A computer just goes through a big series of 0s and 1s, yet through layers of abstraction they accomplish amazing things far more complicated than a naive person would think 0s and 1s could represent and do. Why not the same for a massive neural network trained via gradient descent to maximize a goal by means of matrix multiplication?
Viewing a single comment thread. View all comments