Generation settings allow you to control the creativity of responses, the AI model used, and the number of tokens. You can adjust these parameters during a conversation to tailor the character’s behavior.
To change settings in a chat, click the gear icon next to the selected AI model and prompt in the site’s right-hand menu.

Note
Settings may vary depending on the selected AI model.
Generation Settings allow you to manage additional variables that affect the style and content of the character’s responses.
The Temperature parameter determines how predictable or creative a character’s responses will be. The default value is 0.7 and it can be set from 0 to 2.
A low temperature makes responses cautious, consistent, and clear—ideal for focused conversations. A high temperature makes characters more adventurous, generating diverse and creative responses that can be entertaining but less precise.
The Total Context Tokens parameter shows how many tokens are used for the entire chat context. It includes all messages and character responses, helping you monitor token usage and maintain the conversation.
This setting allows you to adjust the number of tokens for a character’s responses. Limits depend on the selected model. A higher limit does not guarantee longer responses—the character may reply briefly depending on the context.
Top-P controls the character’s word selection. Low values make responses stable and predictable, while high values make them more creative and unexpected. The default is 0.7, and it can be set from 0 to 1.
Presence Penalty and Frequency Penalty are settings that influence the diversity of a character’s responses.
These parameters help balance dialogue consistency with the character’s creativity. The default is 0.2, and they can be set from 0 to 2.