Submitted by granddaddy t3_zjf45w in MachineLearning
Is there a way to get around GPT-3's 4k token limit?
Companies like Spellbook appear to have found a solution, with some people speculating what they have done on Twitter - e.g., summarizing the original document, looping in 4k chunks until the right answer is produced, etc.
I suspect multiple solutions have been applied.
I'd be curious if you have any ideas!
Relevant Tweet: https://twitter.com/AlphaMinus2/status/1600319547348639744
rafgro t1_izyke4t wrote
I've been sliding content window, summarizing chunks, chaining summaries, summarizing chained summaries, all while guiding attention (focus on X, ignore Y). I've had also limited success with storing all summaries separately, choosing most relevant summary based on task/question, and then answering with opened relevant context window in addition to the summaries, but it was too much pain (also financial) for very small gain in my case (but I imagine in legal environment it may be much more important to get every detail right).