mirror of https://github.com/sigoden/aichat
You cannot select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
127 lines
3.5 KiB
Rust
127 lines
3.5 KiB
Rust
use super::{
|
|
catch_error, sse_stream, CompletionDetails, ExtraConfig, Model, ModelConfig, OpenAIClient,
|
|
PromptType, SendData, SsMmessage, SseHandler,
|
|
};
|
|
|
|
use crate::utils::PromptKind;
|
|
|
|
use anyhow::{anyhow, Result};
|
|
use reqwest::{Client as ReqwestClient, RequestBuilder};
|
|
use serde::Deserialize;
|
|
use serde_json::{json, Value};
|
|
|
|
const API_BASE: &str = "https://api.openai.com/v1";
|
|
|
|
#[derive(Debug, Clone, Deserialize, Default)]
|
|
pub struct OpenAIConfig {
|
|
pub name: Option<String>,
|
|
pub api_key: Option<String>,
|
|
pub api_base: Option<String>,
|
|
pub organization_id: Option<String>,
|
|
#[serde(default)]
|
|
pub models: Vec<ModelConfig>,
|
|
pub extra: Option<ExtraConfig>,
|
|
}
|
|
|
|
impl OpenAIClient {
|
|
config_get_fn!(api_key, get_api_key);
|
|
config_get_fn!(api_base, get_api_base);
|
|
|
|
pub const PROMPTS: [PromptType<'static>; 1] =
|
|
[("api_key", "API Key:", true, PromptKind::String)];
|
|
|
|
fn request_builder(&self, client: &ReqwestClient, data: SendData) -> Result<RequestBuilder> {
|
|
let api_key = self.get_api_key()?;
|
|
let api_base = self.get_api_base().unwrap_or_else(|_| API_BASE.to_string());
|
|
|
|
let body = openai_build_body(data, &self.model);
|
|
|
|
let url = format!("{api_base}/chat/completions");
|
|
|
|
debug!("OpenAI Request: {url} {body}");
|
|
|
|
let mut builder = client.post(url).bearer_auth(api_key).json(&body);
|
|
|
|
if let Some(organization_id) = &self.config.organization_id {
|
|
builder = builder.header("OpenAI-Organization", organization_id);
|
|
}
|
|
|
|
Ok(builder)
|
|
}
|
|
}
|
|
|
|
pub async fn openai_send_message(builder: RequestBuilder) -> Result<(String, CompletionDetails)> {
|
|
let res = builder.send().await?;
|
|
let status = res.status();
|
|
let data: Value = res.json().await?;
|
|
if status != 200 {
|
|
catch_error(&data, status.as_u16())?;
|
|
}
|
|
|
|
openai_extract_completion(&data)
|
|
}
|
|
|
|
pub async fn openai_send_message_streaming(
|
|
builder: RequestBuilder,
|
|
handler: &mut SseHandler,
|
|
) -> Result<()> {
|
|
let handle = |message: SsMmessage| -> Result<bool> {
|
|
if message.data == "[DONE]" {
|
|
return Ok(true);
|
|
}
|
|
let data: Value = serde_json::from_str(&message.data)?;
|
|
if let Some(text) = data["choices"][0]["delta"]["content"].as_str() {
|
|
handler.text(text)?;
|
|
}
|
|
Ok(false)
|
|
};
|
|
|
|
sse_stream(builder, handle).await
|
|
}
|
|
|
|
pub fn openai_build_body(data: SendData, model: &Model) -> Value {
|
|
let SendData {
|
|
messages,
|
|
temperature,
|
|
top_p,
|
|
stream,
|
|
} = data;
|
|
|
|
let mut body = json!({
|
|
"model": &model.name,
|
|
"messages": messages,
|
|
});
|
|
|
|
if let Some(v) = model.max_output_tokens {
|
|
body["max_tokens"] = v.into();
|
|
}
|
|
if let Some(v) = temperature {
|
|
body["temperature"] = v.into();
|
|
}
|
|
if let Some(v) = top_p {
|
|
body["top_p"] = v.into();
|
|
}
|
|
if stream {
|
|
body["stream"] = true.into();
|
|
}
|
|
body
|
|
}
|
|
|
|
pub fn openai_extract_completion(data: &Value) -> Result<(String, CompletionDetails)> {
|
|
let text = data["choices"][0]["message"]["content"]
|
|
.as_str()
|
|
.ok_or_else(|| anyhow!("Invalid response data: {data}"))?;
|
|
let details = CompletionDetails {
|
|
id: data["id"].as_str().map(|v| v.to_string()),
|
|
input_tokens: data["usage"]["prompt_tokens"].as_u64(),
|
|
output_tokens: data["usage"]["completion_tokens"].as_u64(),
|
|
};
|
|
Ok((text.to_string(), details))
|
|
}
|
|
|
|
impl_client_trait!(
|
|
OpenAIClient,
|
|
openai_send_message,
|
|
openai_send_message_streaming
|
|
);
|