-
Notifications
You must be signed in to change notification settings - Fork 53
Open
Description
Every next message added to chat is taking longer for llm to respond. I wonder if this is because I prepare the prompt from the whole history every time a new message is sent? Is there some better way to send next message and keep context without this outcome? I've noticed that if I generate prompt from just 2 last messages (user and empty assistant) it still works correctly, remembers previous context and works faster. Should I use this approach?
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels