It would be good to be able to set the context length per chat/model/prompt/assistant.
At present the only way seems to be to go into the main settings and set it on the whole API.
Perhaps a slider in the chat parameters ranging from something like 0-128K with a self-updating numeric field (e.g. double click the current showing number to enter a precise number).
Here's a quick and dirty mock