Submitted by Destiny_Knight t3_11tab5h in singularity
Akimbo333 t1_jck73ph wrote
Reply to comment by Hands0L0 in Those who know... by Destiny_Knight
Well you could always ask it to continue the sentence
Hands0L0 t1_jck7ifi wrote
Not if there is a token limit.
I'm sorry, I don't think I was being clear. The token limit is tied to VRAM. You can load the 30b on a 3090 but it shallows up 20/24 gb of VRAM for the model and prompt alone. That gives you 4gb for returns
Akimbo333 t1_jcka9ef wrote
Oh ok. So you can't make it keep talking?
Hands0L0 t1_jckbm7h wrote
No, because the predictive text needs the entire conversation history context to predict what to say next, and the only way to store the conversation history is in RAM. If you run out of RAM you run out of room for returns.
Viewing a single comment thread. View all comments