Submitted by super_deap t3_11tmpc5 in MachineLearning
lmericle t1_jcln487 wrote
Reply to comment by felheartx in [D] PyTorch 2.0 Native Flash Attention 32k Context Window by super_deap
You will find that in hype circles such as NLP there's a lot of thought-terminating cliches passed around by people who are not so deep in the weeds. Someone says something with confidence, another person doesn't know how to vet it and so just blindly passes it on, and all of a sudden a hack becomes a rumor becomes dogma. It seems to me to be this way with context vs memory.
Put another way: it's the kind of attitude that says "No, Mr. Ford, what we wanted was faster horses".
Viewing a single comment thread. View all comments