From 646cee89920a09f92589b715d99c86ee2c29ea32 Mon Sep 17 00:00:00 2001 From: Yige Date: Mon, 23 Feb 2026 11:49:39 +0800 Subject: [PATCH] =?UTF-8?q?feat:=20=E6=96=B0=E5=A2=9E=209=20=E4=B8=AA=20Pr?= =?UTF-8?q?ovider=EF=BC=88baidu,=20tencent,=20xunfei,=20siliconflow,=20mis?= =?UTF-8?q?tral,=20xai,=20cohere,=20openrouter,=20perplexity=EF=BC=89?= =?UTF-8?q?=EF=BC=8CpresetDataVersion=205=E2=86=926?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- compute/providers/_index.json | 12 +++++- compute/providers/baidu.json | 50 ++++++++++++++++++++++ compute/providers/cohere.json | 65 +++++++++++++++++++++++++++++ compute/providers/mistral.json | 67 ++++++++++++++++++++++++++++++ compute/providers/openrouter.json | 64 ++++++++++++++++++++++++++++ compute/providers/perplexity.json | 67 ++++++++++++++++++++++++++++++ compute/providers/siliconflow.json | 65 +++++++++++++++++++++++++++++ compute/providers/tencent.json | 50 ++++++++++++++++++++++ compute/providers/xai.json | 50 ++++++++++++++++++++++ compute/providers/xunfei.json | 51 +++++++++++++++++++++++ manifest.json | 2 +- 11 files changed, 541 insertions(+), 2 deletions(-) create mode 100644 compute/providers/baidu.json create mode 100644 compute/providers/cohere.json create mode 100644 compute/providers/mistral.json create mode 100644 compute/providers/openrouter.json create mode 100644 compute/providers/perplexity.json create mode 100644 compute/providers/siliconflow.json create mode 100644 compute/providers/tencent.json create mode 100644 compute/providers/xai.json create mode 100644 compute/providers/xunfei.json diff --git a/compute/providers/_index.json b/compute/providers/_index.json index ab3ae66..548abbf 100644 --- a/compute/providers/_index.json +++ b/compute/providers/_index.json @@ -10,10 +10,20 @@ "zhipu", "baichuan", "lingyiwanwu", + "minimax", + "baidu", + "tencent", + "xunfei", + "siliconflow", "local-whisper", "ollama", "google", + "mistral", + "xai", + "cohere", "kling", - "stability" + "stability", + "openrouter", + "perplexity" ] } diff --git a/compute/providers/baidu.json b/compute/providers/baidu.json new file mode 100644 index 0000000..52846d7 --- /dev/null +++ b/compute/providers/baidu.json @@ -0,0 +1,50 @@ +{ + "id": "provider-baidu-001", + "provider": "baidu", + "label": "百度千帆", + "baseUrl": "https://qianfan.baidubce.com/v2", + "apiFormat": "openai-completions", + "apiKeyRef": "baidu", + "apiKeyVerified": false, + "enabled": false, + "status": "unconfigured", + "services": ["chat", "reasoning"], + "defaultTemperature": 1, + "defaultTopP": 1, + "models": [ + { + "modelName": "ernie-5.0-thinking-latest", + "displayName": "ERNIE 5.0", + "serviceType": ["chat", "reasoning"], + "description": "百度最新旗舰模型,支持深度推理、多模态和工具调用", + "contextWindow": 131072, + "maxOutputTokens": 16384, + "capabilities": [ + "chat", + "reasoning", + "vision", + "multimodal", + "tool_use" + ], + "inputPrice": 0.83, + "outputPrice": 3.33, + "extra": {} + }, + { + "modelName": "ernie-4.5-turbo-128k-latest", + "displayName": "ERNIE 4.5 Turbo 128K", + "serviceType": ["chat"], + "description": "百度高性价比长上下文模型,128K 窗口,适合日常对话和文档处理", + "contextWindow": 131072, + "maxOutputTokens": 8192, + "capabilities": [ + "chat", + "long_context", + "fast" + ], + "inputPrice": 0.11, + "outputPrice": 0.44, + "extra": {} + } + ] +} diff --git a/compute/providers/cohere.json b/compute/providers/cohere.json new file mode 100644 index 0000000..9322255 --- /dev/null +++ b/compute/providers/cohere.json @@ -0,0 +1,65 @@ +{ + "id": "provider-cohere-001", + "provider": "cohere", + "label": "Cohere", + "baseUrl": "https://api.cohere.com/compatibility/v1", + "apiFormat": "openai-completions", + "apiKeyRef": "cohere", + "apiKeyVerified": false, + "enabled": false, + "status": "unconfigured", + "services": ["chat", "embedding", "rerank"], + "defaultTemperature": 1, + "defaultTopP": 1, + "models": [ + { + "modelName": "command-a-03-2025", + "displayName": "Command A", + "serviceType": ["chat"], + "description": "Cohere 旗舰对话模型,262K 上下文,擅长 RAG 和工具调用", + "contextWindow": 262144, + "maxOutputTokens": 8192, + "capabilities": [ + "chat", + "reasoning", + "code", + "tool_use", + "rag", + "long_context" + ], + "inputPrice": 2.50, + "outputPrice": 10.00, + "extra": {} + }, + { + "modelName": "embed-v4.0", + "displayName": "Embed V4", + "serviceType": ["embedding"], + "description": "Cohere 最新 Embedding 模型,131K 上下文,多语言支持", + "contextWindow": 131072, + "maxOutputTokens": 0, + "capabilities": [ + "text_embedding", + "multilingual" + ], + "inputPrice": 0.12, + "outputPrice": 0, + "extra": {} + }, + { + "modelName": "rerank-v3.5", + "displayName": "Rerank V3.5", + "serviceType": ["rerank"], + "description": "Cohere 语义重排序模型,用于检索结果精排", + "contextWindow": 0, + "maxOutputTokens": 0, + "capabilities": [ + "rerank", + "semantic_reranking" + ], + "inputPrice": 2.00, + "outputPrice": 0, + "extra": {} + } + ] +} diff --git a/compute/providers/mistral.json b/compute/providers/mistral.json new file mode 100644 index 0000000..bab87e8 --- /dev/null +++ b/compute/providers/mistral.json @@ -0,0 +1,67 @@ +{ + "id": "provider-mistral-001", + "provider": "mistral", + "label": "Mistral AI", + "baseUrl": "https://api.mistral.ai/v1", + "apiFormat": "openai-completions", + "apiKeyRef": "mistral", + "apiKeyVerified": false, + "enabled": false, + "status": "unconfigured", + "services": ["chat", "fast"], + "defaultTemperature": 1, + "defaultTopP": 1, + "models": [ + { + "modelName": "mistral-large-latest", + "displayName": "Mistral Large 3", + "serviceType": ["chat"], + "description": "Mistral 旗舰模型,256K 上下文,支持视觉和工具调用", + "contextWindow": 256000, + "maxOutputTokens": 262144, + "capabilities": [ + "chat", + "reasoning", + "code", + "vision", + "tool_use", + "long_context" + ], + "inputPrice": 0.50, + "outputPrice": 1.50, + "extra": {} + }, + { + "modelName": "mistral-small-latest", + "displayName": "Mistral Small 3.2", + "serviceType": ["fast"], + "description": "Mistral 高效小模型,低延迟,适合快速推理场景", + "contextWindow": 130000, + "maxOutputTokens": 8192, + "capabilities": [ + "chat", + "code", + "fast", + "tool_use" + ], + "inputPrice": 0.10, + "outputPrice": 0.30, + "extra": {} + }, + { + "modelName": "codestral-latest", + "displayName": "Codestral", + "serviceType": ["chat"], + "description": "Mistral 专业代码模型,支持 Fill-in-the-Middle 补全", + "contextWindow": 256000, + "maxOutputTokens": 32768, + "capabilities": [ + "code", + "fill_in_the_middle" + ], + "inputPrice": 0.30, + "outputPrice": 0.90, + "extra": {} + } + ] +} diff --git a/compute/providers/openrouter.json b/compute/providers/openrouter.json new file mode 100644 index 0000000..b43d7af --- /dev/null +++ b/compute/providers/openrouter.json @@ -0,0 +1,64 @@ +{ + "id": "provider-openrouter-001", + "provider": "openrouter", + "label": "OpenRouter", + "baseUrl": "https://openrouter.ai/api/v1", + "apiFormat": "openai-completions", + "apiKeyRef": "openrouter", + "apiKeyVerified": false, + "enabled": false, + "status": "unconfigured", + "services": ["chat"], + "defaultTemperature": 1, + "defaultTopP": 1, + "models": [ + { + "modelName": "openrouter/auto", + "displayName": "OpenRouter Auto", + "serviceType": ["chat"], + "description": "OpenRouter 自动路由,智能选择最优模型", + "contextWindow": 200000, + "maxOutputTokens": 16384, + "capabilities": [ + "chat", + "auto_routing" + ], + "inputPrice": 0.00, + "outputPrice": 0.00, + "extra": {} + }, + { + "modelName": "google/gemini-2.5-flash-exp:free", + "displayName": "Gemini 2.5 Flash (免费)", + "serviceType": ["chat"], + "description": "通过 OpenRouter 免费使用的 Gemini 2.5 Flash,1M 上下文", + "contextWindow": 1048576, + "maxOutputTokens": 65535, + "capabilities": [ + "chat", + "reasoning", + "vision", + "long_context" + ], + "inputPrice": 0.00, + "outputPrice": 0.00, + "extra": {} + }, + { + "modelName": "qwen/qwen3-coder-480b:free", + "displayName": "Qwen3 Coder 480B (免费)", + "serviceType": ["chat"], + "description": "通过 OpenRouter 免费使用的 Qwen3 Coder 480B", + "contextWindow": 262144, + "maxOutputTokens": 32768, + "capabilities": [ + "code", + "reasoning", + "agent" + ], + "inputPrice": 0.00, + "outputPrice": 0.00, + "extra": {} + } + ] +} diff --git a/compute/providers/perplexity.json b/compute/providers/perplexity.json new file mode 100644 index 0000000..bcd5e17 --- /dev/null +++ b/compute/providers/perplexity.json @@ -0,0 +1,67 @@ +{ + "id": "provider-perplexity-001", + "provider": "perplexity", + "label": "Perplexity", + "baseUrl": "https://api.perplexity.ai", + "apiFormat": "openai-completions", + "apiKeyRef": "perplexity", + "apiKeyVerified": false, + "enabled": false, + "status": "unconfigured", + "services": ["chat"], + "defaultTemperature": 1, + "defaultTopP": 1, + "models": [ + { + "modelName": "sonar-pro", + "displayName": "Sonar Pro", + "serviceType": ["chat"], + "description": "Perplexity 旗舰搜索增强模型,内置实时联网搜索和引用", + "contextWindow": 200000, + "maxOutputTokens": 8192, + "capabilities": [ + "chat", + "web_search", + "reasoning", + "citation" + ], + "inputPrice": 3.00, + "outputPrice": 15.00, + "extra": {} + }, + { + "modelName": "sonar-reasoning-pro", + "displayName": "Sonar Reasoning Pro", + "serviceType": ["chat"], + "description": "Perplexity 深度推理模型,内置联网搜索和深度思考", + "contextWindow": 128000, + "maxOutputTokens": 8192, + "capabilities": [ + "chat", + "web_search", + "reasoning", + "deep_thinking" + ], + "inputPrice": 2.00, + "outputPrice": 8.00, + "extra": {} + }, + { + "modelName": "sonar", + "displayName": "Sonar", + "serviceType": ["chat"], + "description": "Perplexity 轻量搜索模型,低成本联网搜索", + "contextWindow": 128000, + "maxOutputTokens": 4096, + "capabilities": [ + "chat", + "web_search", + "citation", + "fast" + ], + "inputPrice": 1.00, + "outputPrice": 1.00, + "extra": {} + } + ] +} diff --git a/compute/providers/siliconflow.json b/compute/providers/siliconflow.json new file mode 100644 index 0000000..74a662e --- /dev/null +++ b/compute/providers/siliconflow.json @@ -0,0 +1,65 @@ +{ + "id": "provider-siliconflow-001", + "provider": "siliconflow", + "label": "硅基流动", + "baseUrl": "https://api.siliconflow.cn/v1", + "apiFormat": "openai-completions", + "apiKeyRef": "siliconflow", + "apiKeyVerified": false, + "enabled": false, + "status": "unconfigured", + "services": ["chat", "embedding"], + "defaultTemperature": 1, + "defaultTopP": 1, + "models": [ + { + "modelName": "Qwen/Qwen3-Coder-480B-A35B-Instruct", + "displayName": "Qwen3 Coder 480B (硅基)", + "serviceType": ["chat"], + "description": "开源最强代码模型,480B MoE 架构,擅长代码生成和 Agent 场景", + "contextWindow": 262144, + "maxOutputTokens": 32768, + "capabilities": [ + "code", + "reasoning", + "agent", + "tool_use" + ], + "inputPrice": 0.22, + "outputPrice": 1.00, + "extra": {} + }, + { + "modelName": "Qwen/Qwen3-235B-A22B-Instruct", + "displayName": "Qwen3 235B (硅基)", + "serviceType": ["chat"], + "description": "Qwen3 旗舰通用模型,235B MoE 架构,多语言能力突出", + "contextWindow": 262144, + "maxOutputTokens": 38912, + "capabilities": [ + "chat", + "reasoning", + "code", + "multilingual" + ], + "inputPrice": 0.35, + "outputPrice": 1.39, + "extra": {} + }, + { + "modelName": "BAAI/bge-m3", + "displayName": "BGE-M3 (硅基)", + "serviceType": ["embedding"], + "description": "多语言 Embedding 模型,支持 8K 上下文", + "contextWindow": 8192, + "maxOutputTokens": 0, + "capabilities": [ + "text_embedding", + "multilingual" + ], + "inputPrice": 0.005, + "outputPrice": 0, + "extra": {} + } + ] +} diff --git a/compute/providers/tencent.json b/compute/providers/tencent.json new file mode 100644 index 0000000..962170f --- /dev/null +++ b/compute/providers/tencent.json @@ -0,0 +1,50 @@ +{ + "id": "provider-tencent-001", + "provider": "tencent", + "label": "腾讯混元", + "baseUrl": "https://api.hunyuan.cloud.tencent.com/v1", + "apiFormat": "openai-completions", + "apiKeyRef": "tencent", + "apiKeyVerified": false, + "enabled": false, + "status": "unconfigured", + "services": ["chat", "reasoning"], + "defaultTemperature": 1, + "defaultTopP": 1, + "models": [ + { + "modelName": "hunyuan-2.0-thinking-20251109", + "displayName": "混元 2.0 Think", + "serviceType": ["reasoning"], + "description": "腾讯混元最新推理模型,262K 上下文,支持深度思考", + "contextWindow": 262144, + "maxOutputTokens": 16384, + "capabilities": [ + "chat", + "reasoning", + "deep_thinking", + "long_context" + ], + "inputPrice": 0.55, + "outputPrice": 2.21, + "extra": {} + }, + { + "modelName": "hunyuan-turbo-s", + "displayName": "混元 Turbo S", + "serviceType": ["chat"], + "description": "腾讯混元高速模型,262K 上下文,支持推理和工具调用", + "contextWindow": 262144, + "maxOutputTokens": 16384, + "capabilities": [ + "chat", + "reasoning", + "fast", + "tool_use" + ], + "inputPrice": 0.11, + "outputPrice": 0.28, + "extra": {} + } + ] +} diff --git a/compute/providers/xai.json b/compute/providers/xai.json new file mode 100644 index 0000000..0ba91a0 --- /dev/null +++ b/compute/providers/xai.json @@ -0,0 +1,50 @@ +{ + "id": "provider-xai-001", + "provider": "xai", + "label": "xAI Grok", + "baseUrl": "https://api.x.ai/v1", + "apiFormat": "openai-completions", + "apiKeyRef": "xai", + "apiKeyVerified": false, + "enabled": false, + "status": "unconfigured", + "services": ["chat", "reasoning"], + "defaultTemperature": 1, + "defaultTopP": 1, + "models": [ + { + "modelName": "grok-4-0709", + "displayName": "Grok 4", + "serviceType": ["chat", "reasoning"], + "description": "xAI 最新旗舰推理模型,262K 上下文,支持工具调用", + "contextWindow": 262144, + "maxOutputTokens": 8192, + "capabilities": [ + "chat", + "reasoning", + "code", + "tool_use" + ], + "inputPrice": 3.00, + "outputPrice": 15.00, + "extra": {} + }, + { + "modelName": "grok-4-1-fast", + "displayName": "Grok 4.1 Fast", + "serviceType": ["chat"], + "description": "xAI 超长上下文高速模型,2M token 窗口", + "contextWindow": 2000000, + "maxOutputTokens": 16384, + "capabilities": [ + "chat", + "reasoning", + "fast", + "long_context" + ], + "inputPrice": 0.20, + "outputPrice": 0.50, + "extra": {} + } + ] +} diff --git a/compute/providers/xunfei.json b/compute/providers/xunfei.json new file mode 100644 index 0000000..210d8ea --- /dev/null +++ b/compute/providers/xunfei.json @@ -0,0 +1,51 @@ +{ + "id": "provider-xunfei-001", + "provider": "xunfei", + "label": "科大讯飞", + "baseUrl": "https://spark-api-open.xf-yun.com/v1", + "apiFormat": "openai-completions", + "apiKeyRef": "xunfei", + "apiKeyVerified": false, + "enabled": false, + "status": "unconfigured", + "services": ["chat", "reasoning"], + "defaultTemperature": 1, + "defaultTopP": 1, + "models": [ + { + "modelName": "x1", + "displayName": "讯飞星火 X1", + "serviceType": ["reasoning"], + "description": "讯飞最新推理模型,擅长数学和代码推理", + "contextWindow": 32768, + "maxOutputTokens": 16384, + "capabilities": [ + "chat", + "reasoning", + "deep_thinking", + "math", + "code" + ], + "inputPrice": 2.92, + "outputPrice": 11.67, + "extra": {} + }, + { + "modelName": "4.0Ultra", + "displayName": "讯飞星火 4.0 Ultra", + "serviceType": ["chat"], + "description": "讯飞星火最强通用模型,支持推理、代码和工具调用", + "contextWindow": 32768, + "maxOutputTokens": 32768, + "capabilities": [ + "chat", + "reasoning", + "code", + "tool_use" + ], + "inputPrice": 2.92, + "outputPrice": 11.67, + "extra": {} + } + ] +} diff --git a/manifest.json b/manifest.json index 23d13cc..b064dce 100644 --- a/manifest.json +++ b/manifest.json @@ -1,6 +1,6 @@ { "version": "1.0.0", - "presetDataVersion": 5, + "presetDataVersion": 6, "updatedAt": "2026-02-23", "description": "DesireCore 官方配置中心" }