Files
codex/codex-rs/api-client/src/responses.rs
2025-11-12 16:18:24 +00:00

138 lines
4.6 KiB
Rust

use std::sync::Arc;
use codex_app_server_protocol::AuthMode;
use codex_otel::otel_event_manager::OtelEventManager;
use codex_protocol::ConversationId;
use futures::TryStreamExt;
use serde_json::Value;
use tracing::debug;
use tracing::trace;
use crate::auth::AuthProvider;
use crate::error::Error;
use crate::error::Result;
use crate::stream::WireResponseStream;
use codex_provider_config::ModelProviderInfo;
#[derive(Clone)]
/// Configuration for the OpenAI Responses API client (`/v1/responses`).
///
/// - `http_client`: Reqwest client used for HTTP requests.
/// - `provider`: Provider configuration (base URL, headers, retries, etc.).
/// - `conversation_id`: Used to set conversation/session headers and cache keys.
/// - `auth_provider`: Optional provider of auth context (e.g., ChatGPT login token).
/// - `otel_event_manager`: Telemetry event manager for request/stream instrumentation.
pub struct ResponsesApiClientConfig {
pub http_client: reqwest::Client,
pub provider: ModelProviderInfo,
pub conversation_id: ConversationId,
pub auth_provider: Option<Arc<dyn AuthProvider>>,
pub otel_event_manager: OtelEventManager,
pub extra_headers: Vec<(String, String)>,
}
#[derive(Clone)]
pub struct ResponsesApiClient {
config: ResponsesApiClientConfig,
}
impl ResponsesApiClient {
pub fn new(config: ResponsesApiClientConfig) -> Result<Self> {
Ok(Self { config })
}
}
impl ResponsesApiClient {
pub async fn stream_payload_wire(&self, payload_json: &Value) -> Result<WireResponseStream> {
if self.config.provider.wire_api != codex_provider_config::WireApi::Responses {
return Err(Error::UnsupportedOperation(
"ResponsesApiClient requires a Responses provider".to_string(),
));
}
let auth = crate::client::http::resolve_auth(&self.config.auth_provider).await;
trace!(
"POST to {}: {:?}",
self.config.provider.get_full_url(
auth.as_ref()
.map(|a| codex_provider_config::AuthContext {
mode: a.mode,
bearer_token: a.bearer_token.clone(),
account_id: a.account_id.clone(),
})
.as_ref()
),
serde_json::to_string(payload_json)
.unwrap_or_else(|_| "<unable to serialize payload>".to_string())
);
let mut owned_headers: Vec<(String, String)> = vec![
(
"conversation_id".to_string(),
self.config.conversation_id.to_string(),
),
(
"session_id".to_string(),
self.config.conversation_id.to_string(),
),
];
owned_headers.extend(self.config.extra_headers.iter().cloned());
let extra_headers = crate::client::http::header_pairs(&owned_headers);
let mut req_builder = crate::client::http::build_request(
&self.config.http_client,
&self.config.provider,
&auth,
&extra_headers,
)
.await?;
req_builder = req_builder
.header(reqwest::header::ACCEPT, "text/event-stream")
.json(payload_json);
if let Some(auth_ctx) = auth.as_ref()
&& auth_ctx.mode == AuthMode::ChatGPT
&& let Some(account_id) = auth_ctx.account_id.clone()
{
req_builder = req_builder.header("chatgpt-account-id", account_id);
}
let res = self
.config
.otel_event_manager
.log_request(0, || req_builder.send())
.await
.map_err(|source| Error::ResponseStreamFailed {
source,
request_id: None,
})?;
let snapshot = crate::client::rate_limits::parse_rate_limit_snapshot(res.headers());
let stream = res
.bytes_stream()
.map_err(|err| Error::ResponseStreamFailed {
source: err,
request_id: None,
});
let (tx_event, rx_event) = crate::client::sse::spawn_wire_stream(
stream,
&self.config.provider,
self.config.otel_event_manager.clone(),
crate::decode_wire::responses::WireResponsesSseDecoder,
);
if let Some(snapshot) = snapshot
&& tx_event
.send(Ok(crate::stream::WireEvent::RateLimits(snapshot)))
.await
.is_err()
{
debug!("receiver dropped rate limit snapshot event");
}
Ok(rx_event)
}
}