feat: support deepseek client (#490)

pull/491/head
sigoden 2 weeks ago committed by GitHub
parent 5b33c7356f
commit 0071d84aa5
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

@ -54,6 +54,7 @@ OPENAI_COMPATIBLE_PLATFORMS=( \
openai,gpt-3.5-turbo,https://api.openai.com/v1 \
anyscale,meta-llama/Meta-Llama-3-8B-Instruct,https://api.endpoints.anyscale.com/v1 \
deepinfra,meta-llama/Meta-Llama-3-8B-Instruct,https://api.deepinfra.com/v1/openai \
deepseek,deepseek-chat,https://api.deepseek.com \
fireworks,accounts/fireworks/models/llama-v3-8b-instruct,https://api.fireworks.ai/inference/v1 \
groq,llama3-8b-8192,https://api.groq.com/openai/v1 \
mistral,mistral-small-latest,https://api.mistral.ai/v1 \

@ -170,6 +170,11 @@ clients:
api_base: https://api.moonshot.cn/v1
api_key: sk-xxx # ENV: {client}_API_KEY
# See https://platform.deepseek.com/api-docs/
- type: openai-compatible
name: deepseek
api_key: sk-xxx # ENV: {client}_API_KEY
# See https://docs.endpoints.anyscale.com/
- type: openai-compatible
name: anyscale

@ -157,18 +157,16 @@
# notes
# - get max_output_tokens info from api error
models:
- name: sonar-small-chat
max_input_tokens: 16384
max_output_tokens?: 16384
- name: sonar-small-online
max_input_tokens: 12000
max_output_tokens?: 12288
- name: sonar-medium-chat
max_input_tokens: 16384
max_output_tokens?: 16384
- name: sonar-medium-online
max_input_tokens: 12000
max_output_tokens?: 12288
- name: llama-3-sonar-small-32k-chat
max_input_tokens: 32768
max_output_tokens?: 32768
input_price: 0.2
output_price: 0.2
- name: llama-3-sonar-large-32k-chat
max_input_tokens: 32768
max_output_tokens?: 32768
input_price: 0.6
output_price: 0.6
- name: llama-3-8b-instruct
max_input_tokens: 8192
@ -180,26 +178,11 @@
max_output_tokens?: 8192
input_price: 1
output_price: 1
- name: codellama-70b-instruct
max_input_tokens: 16384
max_output_tokens?: 16384
input_price: 1
output_price: 1
- name: mistral-7b-instruct
max_input_tokens: 16384
max_output_tokens?: 16384
input_price: 0.2
output_price: 0.2
- name: mixtral-8x7b-instruct
max_input_tokens: 16384
max_output_tokens?: 16384
input_price: 0.6
output_price: 0.6
- name: mixtral-8x22b-instruct
max_input_tokens: 16384
max_output_tokens?: 16384
input_price: 1
output_price: 1
- platform: groq
# docs:
@ -354,6 +337,9 @@
- name: '@cf/qwen/qwen1.5-14b-chat-awq'
max_input_tokens: 4096
max_output_tokens: 4096
- name: '@hf/thebloke/deepseek-coder-6.7b-instruct-awq'
max_input_tokens: 4096
max_output_tokens: 4096
- name: '@hf/nexusflow/starling-lm-7b-beta'
max_input_tokens: 4096
max_output_tokens: 4096
@ -474,6 +460,20 @@
input_price: 8.4
output_price: 8.4
- platform: deepseek
docs:
- https://platform.deepseek.com/api-docs/
- https://platform.deepseek.com/api-docs/pricing
models:
- name: deepseek-chat
max_input_tokens: 32768
input_price: 0.14
output_price: 0.28
- name: deepseek-coder
max_input_tokens: 16384
input_price: 0.14
output_price: 0.28
- platform: anyscale
# docs:
# - https://docs.endpoints.anyscale.com/text-generation/query-a-model/#select-a-model

@ -45,9 +45,10 @@ register_client!(
(qianwen, "qianwen", QianwenConfig, QianwenClient),
);
pub const OPENAI_COMPATIBLE_PLATFORMS: [(&str, &str); 10] = [
pub const OPENAI_COMPATIBLE_PLATFORMS: [(&str, &str); 11] = [
("anyscale", "https://api.endpoints.anyscale.com/v1"),
("deepinfra", "https://api.deepinfra.com/v1/openai"),
("deepseek", "https://api.deepseek.com"),
("fireworks", "https://api.fireworks.ai/inference/v1"),
("groq", "https://api.groq.com/openai/v1"),
("mistral", "https://api.mistral.ai/v1"),

Loading…
Cancel
Save