chore: drop model_max_output_tokens (#7100)

This commit is contained in:
jif-oai
2025-11-21 18:42:54 +01:00
committed by GitHub
parent 2ae1f81d84
commit af65666561
6 changed files with 14 additions and 50 deletions

View File

@@ -30,7 +30,6 @@ model_provider = "openai"
# Optional manual model metadata. When unset, Codex auto-detects from model.
# Uncomment to force values.
# model_context_window = 128000 # tokens; default: auto for model
# model_max_output_tokens = 8192 # tokens; default: auto for model
# model_auto_compact_token_limit = 0 # disable/override auto; default: model family specific
# tool_output_token_limit = 10000 # tokens stored per tool output; default: 10000 for gpt-5.1-codex-max