Memory settings
Last updated
Last updated
Memory refers to the content the AI model retains during a conversation. We recommend using the default memory length for optimal performance and cost-efficiency.
In Advanced Memory Mode, memory length determines how many dialogue rounds the AI will remember (one round = your message + AI's reply).
Higher turn counts mean:
More context is fed to the AI.
Higher token usage (text volume), which increases Ruby costs.
Click the + icon on the left side of the chat window.
We designed Advanced Memory Mode to let users have full controll over their chat experience. You can balance three key factors:
Memory retention (how much context the AI retains),
Chat costs (higher memory = more tokens/Ruby usage),
Model performance (different AI models have varying memory capacities, all of which are inherently limited).
This flexibility lets you prioritize what matters most: richer context, cost efficiency, or leveraging a model’s specific strengths.
Method 1: Reset Chats (Keep Context, Lower Costs)
Make a Summary: Paste this exact command into your chat:
(Execute command: Summarize all key events and memories <plot> to date)
→ The AI will generate a compact recap inside <plot>
tags.
Migrate to a New Chat:
Click Edit on the summary, copy the text.
Start a new chat, paste it under the greeting like this:
then you can continute to chat with the AI in a new chat.
→ Fresh chat = lower token count, same narrative flow!
Method 2: Pin Critical Plot Points
You can add pivotal details to your Persona (e.g., "The protagonist is allergic to roses"). This acts as a "cheat sheet" for the AI, hence reducing reliance on memory.
A: No. Costs are calculated based on the actual word count processed by the AI, not your memory length setting.
A: Our context search model ensures relevance, even with limited memory settings. For example:
If you chat 200+ rounds but only set memory to 30, the AI will prioritize retrieving critical early context (e.g., family details from the first few rounds) over less relevant later exchanges.
The AI still recalls key details without needing ultra-high memory limits.
A: Increasing memory length when in a long conversation will dramatically increase the input tokens and the price. Example:
Starting at 8 rounds → raising to 20 rounds forces the AI to process 20 rounds’ worth of text instead of 8.
More tokens = higher Ruby costs. Adjust the slider cautiously during long conversations!