mirror of
https://github.com/openai/codex.git
synced 2026-05-02 04:11:39 +03:00
Revert "fix: read max_output_tokens param from config" (#7088)
Reverts openai/codex#4139
This commit is contained in:
@@ -280,8 +280,6 @@ pub(crate) struct ResponsesApiRequest<'a> {
|
||||
pub(crate) prompt_cache_key: Option<String>,
|
||||
#[serde(skip_serializing_if = "Option::is_none")]
|
||||
pub(crate) text: Option<TextControls>,
|
||||
#[serde(skip_serializing_if = "Option::is_none")]
|
||||
pub(crate) max_output_tokens: Option<i64>,
|
||||
}
|
||||
|
||||
pub(crate) mod tools {
|
||||
@@ -465,7 +463,6 @@ mod tests {
|
||||
verbosity: Some(OpenAiVerbosity::Low),
|
||||
format: None,
|
||||
}),
|
||||
max_output_tokens: Some(10_000),
|
||||
};
|
||||
|
||||
let v = serde_json::to_value(&req).expect("json");
|
||||
@@ -504,7 +501,6 @@ mod tests {
|
||||
include: vec![],
|
||||
prompt_cache_key: None,
|
||||
text: Some(text_controls),
|
||||
max_output_tokens: Some(10_000),
|
||||
};
|
||||
|
||||
let v = serde_json::to_value(&req).expect("json");
|
||||
@@ -541,7 +537,6 @@ mod tests {
|
||||
include: vec![],
|
||||
prompt_cache_key: None,
|
||||
text: None,
|
||||
max_output_tokens: Some(10_000),
|
||||
};
|
||||
|
||||
let v = serde_json::to_value(&req).expect("json");
|
||||
|
||||
Reference in New Issue
Block a user