diff --git a/Argcfile.sh b/Argcfile.sh index 4a2bb89..82e79e5 100755 --- a/Argcfile.sh +++ b/Argcfile.sh @@ -54,6 +54,7 @@ OPENAI_COMPATIBLE_PLATFORMS=( \ openai,gpt-3.5-turbo,https://api.openai.com/v1 \ anyscale,meta-llama/Meta-Llama-3-8B-Instruct,https://api.endpoints.anyscale.com/v1 \ deepinfra,meta-llama/Meta-Llama-3-8B-Instruct,https://api.deepinfra.com/v1/openai \ + deepseek,deepseek-chat,https://api.deepseek.com \ fireworks,accounts/fireworks/models/llama-v3-8b-instruct,https://api.fireworks.ai/inference/v1 \ groq,llama3-8b-8192,https://api.groq.com/openai/v1 \ mistral,mistral-small-latest,https://api.mistral.ai/v1 \ diff --git a/config.example.yaml b/config.example.yaml index ce16df8..b2ea51a 100644 --- a/config.example.yaml +++ b/config.example.yaml @@ -170,6 +170,11 @@ clients: api_base: https://api.moonshot.cn/v1 api_key: sk-xxx # ENV: {client}_API_KEY + # See https://platform.deepseek.com/api-docs/ + - type: openai-compatible + name: deepseek + api_key: sk-xxx # ENV: {client}_API_KEY + # See https://docs.endpoints.anyscale.com/ - type: openai-compatible name: anyscale diff --git a/models.yaml b/models.yaml index aa7fd45..2ecc937 100644 --- a/models.yaml +++ b/models.yaml @@ -157,18 +157,16 @@ # notes # - get max_output_tokens info from api error models: - - name: sonar-small-chat - max_input_tokens: 16384 - max_output_tokens?: 16384 - - name: sonar-small-online - max_input_tokens: 12000 - max_output_tokens?: 12288 - - name: sonar-medium-chat - max_input_tokens: 16384 - max_output_tokens?: 16384 - - name: sonar-medium-online - max_input_tokens: 12000 - max_output_tokens?: 12288 + - name: llama-3-sonar-small-32k-chat + max_input_tokens: 32768 + max_output_tokens?: 32768 + input_price: 0.2 + output_price: 0.2 + - name: llama-3-sonar-large-32k-chat + max_input_tokens: 32768 + max_output_tokens?: 32768 + input_price: 0.6 + output_price: 0.6 - name: llama-3-8b-instruct max_input_tokens: 8192 @@ -180,26 +178,11 @@ max_output_tokens?: 8192 input_price: 1 output_price: 1 - - name: codellama-70b-instruct - max_input_tokens: 16384 - max_output_tokens?: 16384 - input_price: 1 - output_price: 1 - - name: mistral-7b-instruct - max_input_tokens: 16384 - max_output_tokens?: 16384 - input_price: 0.2 - output_price: 0.2 - name: mixtral-8x7b-instruct max_input_tokens: 16384 max_output_tokens?: 16384 input_price: 0.6 output_price: 0.6 - - name: mixtral-8x22b-instruct - max_input_tokens: 16384 - max_output_tokens?: 16384 - input_price: 1 - output_price: 1 - platform: groq # docs: @@ -354,6 +337,9 @@ - name: '@cf/qwen/qwen1.5-14b-chat-awq' max_input_tokens: 4096 max_output_tokens: 4096 + - name: '@hf/thebloke/deepseek-coder-6.7b-instruct-awq' + max_input_tokens: 4096 + max_output_tokens: 4096 - name: '@hf/nexusflow/starling-lm-7b-beta' max_input_tokens: 4096 max_output_tokens: 4096 @@ -474,6 +460,20 @@ input_price: 8.4 output_price: 8.4 +- platform: deepseek + docs: + - https://platform.deepseek.com/api-docs/ + - https://platform.deepseek.com/api-docs/pricing + models: + - name: deepseek-chat + max_input_tokens: 32768 + input_price: 0.14 + output_price: 0.28 + - name: deepseek-coder + max_input_tokens: 16384 + input_price: 0.14 + output_price: 0.28 + - platform: anyscale # docs: # - https://docs.endpoints.anyscale.com/text-generation/query-a-model/#select-a-model diff --git a/src/client/mod.rs b/src/client/mod.rs index 07f5f57..ab28032 100644 --- a/src/client/mod.rs +++ b/src/client/mod.rs @@ -45,9 +45,10 @@ register_client!( (qianwen, "qianwen", QianwenConfig, QianwenClient), ); -pub const OPENAI_COMPATIBLE_PLATFORMS: [(&str, &str); 10] = [ +pub const OPENAI_COMPATIBLE_PLATFORMS: [(&str, &str); 11] = [ ("anyscale", "https://api.endpoints.anyscale.com/v1"), ("deepinfra", "https://api.deepinfra.com/v1/openai"), + ("deepseek", "https://api.deepseek.com"), ("fireworks", "https://api.fireworks.ai/inference/v1"), ("groq", "https://api.groq.com/openai/v1"), ("mistral", "https://api.mistral.ai/v1"),