Every next message added to chat is taking longer for llm to respond. I wonder if this is because I prepare the prompt from the whole history every time a new message is sent? Is there some better way to send next message and keep context without this outcome? I've noticed that if I generate prompt from just 2 last messages (user and empty assistant) it still works correctly, remembers previous context and works faster. Should I use this approach?