[D] Inner workings of the chatgpt memory Submitted by terserterseness t3_10fxryj on January 19, 2023 at 9:48 AM in MachineLearning 21 comments 35
andreichiffa t1_j50x4ky wrote on January 19, 2023 at 5:07 PM Reply to comment by DaLameLama in [D] Inner workings of the chatgpt memory by terserterseness Reported token size is 2048, but they likely do a hard attention mask. In about 1/4th of words Permalink Parent 9
Viewing a single comment thread. View all comments