From 16b06ab168a2924db748100c91c54921e68f248a Mon Sep 17 00:00:00 2001 From: yi-ge Date: Mon, 23 Feb 2026 15:16:12 +0800 Subject: [PATCH] =?UTF-8?q?refactor:=20=E7=BB=9F=E4=B8=80=E6=89=80?= =?UTF-8?q?=E6=9C=89=20Provider=20label=20=E4=B8=BA=E3=80=8C=E5=85=AC?= =?UTF-8?q?=E5=8F=B8=20+=20=E6=A8=A1=E5=9E=8B=E5=93=81=E7=89=8C=E3=80=8D?= =?UTF-8?q?=E5=91=BD=E5=90=8D=E9=A3=8E=E6=A0=BC?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- compute/providers/anthropic.json | 12 +++---- compute/providers/baichuan.json | 10 +++--- compute/providers/baidu.json | 2 +- compute/providers/dashscope.json | 38 +++++++++++------------ compute/providers/google.json | 10 +++--- compute/providers/kling.json | 10 +++--- compute/providers/lingyiwanwu.json | 6 ++-- compute/providers/moonshot.json | 14 ++++----- compute/providers/openai.json | 50 +++++++++++++++--------------- compute/providers/perplexity.json | 2 +- compute/providers/stability.json | 4 +-- compute/providers/volcengine.json | 36 ++++++++++----------- compute/providers/xai.json | 2 +- compute/providers/zhipu.json | 18 +++++------ 14 files changed, 107 insertions(+), 107 deletions(-) diff --git a/compute/providers/anthropic.json b/compute/providers/anthropic.json index 0d722a0..fda1b21 100644 --- a/compute/providers/anthropic.json +++ b/compute/providers/anthropic.json @@ -1,7 +1,7 @@ { "id": "provider-anthropic-001", "provider": "anthropic", - "label": "Anthropic", + "label": "Anthropic Claude", "baseUrl": "https://api.anthropic.com/v1", "apiFormat": "anthropic-messages", "apiKeyRef": "anthropic", @@ -18,7 +18,7 @@ { "modelName": "claude-opus-4.6", "displayName": "Claude Opus 4.6", - "serviceType": "chat", + "serviceType": ["chat"], "description": "Anthropic 最新旗舰 Opus 模型,最强能力", "contextWindow": 200000, "maxOutputTokens": 32000, @@ -36,7 +36,7 @@ { "modelName": "claude-opus-4.5", "displayName": "Claude Opus 4.5", - "serviceType": "chat", + "serviceType": ["chat"], "description": "Anthropic Opus 4.5,顶级推理能力", "contextWindow": 200000, "maxOutputTokens": 32000, @@ -54,7 +54,7 @@ { "modelName": "claude-sonnet-4.5", "displayName": "Claude Sonnet 4.5", - "serviceType": "chat", + "serviceType": ["chat"], "description": "Anthropic Sonnet 4.5,高性能性价比", "contextWindow": 200000, "maxOutputTokens": 64000, @@ -72,7 +72,7 @@ { "modelName": "claude-sonnet-4", "displayName": "Claude Sonnet 4", - "serviceType": "chat", + "serviceType": ["chat"], "description": "Anthropic Sonnet 4 旗舰模型", "contextWindow": 200000, "maxOutputTokens": 64000, @@ -90,7 +90,7 @@ { "modelName": "claude-sonnet-4.5-computer", "displayName": "Claude Sonnet 4.5 Computer Use", - "serviceType": "computer_use", + "serviceType": ["computer_use"], "description": "Anthropic 最佳 Computer Use 模型", "contextWindow": 200000, "maxOutputTokens": 64000, diff --git a/compute/providers/baichuan.json b/compute/providers/baichuan.json index fefa236..40bbe40 100644 --- a/compute/providers/baichuan.json +++ b/compute/providers/baichuan.json @@ -1,7 +1,7 @@ { "id": "provider-baichuan-001", "provider": "baichuan", - "label": "百川智能", + "label": "百川 Baichuan", "baseUrl": "https://api.baichuan-ai.com/v1", "apiFormat": "openai-completions", "apiKeyRef": "baichuan", @@ -17,7 +17,7 @@ { "modelName": "Baichuan-M3-Plus", "displayName": "百川 M3-Plus", - "serviceType": "chat", + "serviceType": ["chat"], "description": "百川智能M3-Plus医疗增强模型,235B参数,幻觉率仅2.6%,API价格较前代下降70%", "contextWindow": 192000, "maxOutputTokens": 8192, @@ -36,7 +36,7 @@ { "modelName": "Baichuan-M3", "displayName": "百川 M3", - "serviceType": "chat", + "serviceType": ["chat"], "description": "百川智能M3新一代开源医疗增强模型,235B参数", "contextWindow": 192000, "maxOutputTokens": 8192, @@ -54,7 +54,7 @@ { "modelName": "Baichuan-M2-Plus", "displayName": "百川 M2-Plus", - "serviceType": "chat", + "serviceType": ["chat"], "description": "百川智能M2-Plus旗舰模型,最强推理和创作能力", "contextWindow": 128000, "maxOutputTokens": 8192, @@ -72,7 +72,7 @@ { "modelName": "Baichuan-M2", "displayName": "百川 M2", - "serviceType": "chat", + "serviceType": ["chat"], "description": "百川智能M2模型,高性能通用模型", "contextWindow": 128000, "maxOutputTokens": 8192, diff --git a/compute/providers/baidu.json b/compute/providers/baidu.json index 12634bb..81603b2 100644 --- a/compute/providers/baidu.json +++ b/compute/providers/baidu.json @@ -1,7 +1,7 @@ { "id": "provider-baidu-001", "provider": "baidu", - "label": "百度千帆", + "label": "百度 文心", "baseUrl": "https://qianfan.baidubce.com/v2", "apiFormat": "openai-completions", "apiKeyRef": "baidu", diff --git a/compute/providers/dashscope.json b/compute/providers/dashscope.json index d15e400..cf2c690 100644 --- a/compute/providers/dashscope.json +++ b/compute/providers/dashscope.json @@ -1,7 +1,7 @@ { "id": "provider-dashscope-001", "provider": "dashscope", - "label": "阿里云 DashScope", + "label": "阿里 通义千问", "baseUrl": "https://dashscope.aliyuncs.com/compatible-mode/v1", "apiFormat": "openai-completions", "apiKeyRef": "dashscope", @@ -27,7 +27,7 @@ { "modelName": "qwen3.5-max", "displayName": "阿里云 Qwen3.5-Max", - "serviceType": "chat", + "serviceType": ["chat"], "description": "通义千问3.5旗舰模型,397B参数,2026除夕发布,极低价格", "contextWindow": 252000, "maxOutputTokens": 16384, @@ -47,7 +47,7 @@ { "modelName": "qwen-max", "displayName": "阿里云 qwen-max", - "serviceType": "chat", + "serviceType": ["chat"], "description": "通义千问Max,千亿参数旗舰模型,最强性能", "contextWindow": 131072, "maxOutputTokens": 8192, @@ -65,7 +65,7 @@ { "modelName": "qwen-plus", "displayName": "阿里云 qwen-plus", - "serviceType": "chat", + "serviceType": ["chat"], "description": "通义千问Plus,支持思考模式,128K 上下文", "contextWindow": 131072, "maxOutputTokens": 8192, @@ -82,7 +82,7 @@ { "modelName": "qwen-turbo", "displayName": "阿里云 qwen-turbo", - "serviceType": "chat", + "serviceType": ["chat"], "description": "通义千问Turbo,支持100万上下文,高性价比", "contextWindow": 1000000, "maxOutputTokens": 8192, @@ -98,7 +98,7 @@ { "modelName": "qwen-long", "displayName": "阿里云 qwen-long", - "serviceType": "chat", + "serviceType": ["chat"], "description": "通义千问Long,支持1000万上下文,适合超长文档处理,降价97%", "contextWindow": 10000000, "maxOutputTokens": 8192, @@ -114,7 +114,7 @@ { "modelName": "qwen3-max", "displayName": "阿里云 Qwen3-Max", - "serviceType": "chat", + "serviceType": ["chat"], "description": "通义千问3代旗舰模型,252K 上下文", "contextWindow": 252000, "maxOutputTokens": 16384, @@ -134,7 +134,7 @@ { "modelName": "qwen3-vl-plus", "displayName": "阿里云 Qwen3-VL-Plus", - "serviceType": "vision", + "serviceType": ["vision"], "description": "通义千问3代视觉模型Plus版,支持图像理解", "contextWindow": 131072, "maxOutputTokens": 8192, @@ -152,7 +152,7 @@ { "modelName": "qwen3-vl-flash", "displayName": "阿里云 Qwen3-VL-Flash", - "serviceType": "vision", + "serviceType": ["vision"], "description": "通义千问3代视觉模型Flash版,高性价比", "contextWindow": 131072, "maxOutputTokens": 8192, @@ -169,7 +169,7 @@ { "modelName": "text-embedding-v3", "displayName": "text-embedding-v3", - "serviceType": "embedding", + "serviceType": ["embedding"], "description": "阿里云文本向量化", "capabilities": [ "text_embedding", @@ -182,7 +182,7 @@ { "modelName": "text-embedding-v4", "displayName": "阿里云 text-embedding-v4", - "serviceType": "embedding", + "serviceType": ["embedding"], "description": "阿里云百炼文本向量模型v4,支持多维度与多语种", "capabilities": [ "text_embedding", @@ -195,7 +195,7 @@ { "modelName": "qwen3-rerank", "displayName": "Qwen3-Rerank", - "serviceType": "rerank", + "serviceType": ["rerank"], "description": "检索重排序,支持 100+ 语种", "capabilities": [ "rerank", @@ -208,7 +208,7 @@ { "modelName": "cosyvoice-v2", "displayName": "CosyVoice V2", - "serviceType": "tts", + "serviceType": ["tts"], "description": "通义语音合成,支持中英文", "capabilities": [ "tts", @@ -232,7 +232,7 @@ { "modelName": "paraformer-v2", "displayName": "Paraformer V2", - "serviceType": "asr", + "serviceType": ["asr"], "description": "通义语音识别", "capabilities": [ "asr", @@ -252,7 +252,7 @@ { "modelName": "wanx-v2", "displayName": "通义万相 V2", - "serviceType": "image_gen", + "serviceType": ["image_gen"], "description": "通义万相文生图", "capabilities": [ "image_generation", @@ -275,7 +275,7 @@ { "modelName": "wanx-video", "displayName": "通义万相视频", - "serviceType": "video_gen", + "serviceType": ["video_gen"], "description": "通义万相文生视频", "capabilities": [ "video_generation", @@ -292,7 +292,7 @@ { "modelName": "cosyvoice-clone", "displayName": "CosyVoice 声音复刻", - "serviceType": "voice_clone", + "serviceType": ["voice_clone"], "description": "通义声音复刻,少量样本即可克隆", "capabilities": [ "voice_clone", @@ -307,7 +307,7 @@ { "modelName": "qwen-omni-turbo", "displayName": "Qwen-Omni-Turbo", - "serviceType": "omni", + "serviceType": ["omni"], "description": "通义全能模型,支持音视频文本多模态", "contextWindow": 32768, "capabilities": [ @@ -322,7 +322,7 @@ { "modelName": "qwen3-max-trans", "displayName": "Qwen3-Max(翻译)", - "serviceType": "translation", + "serviceType": ["translation"], "description": "通义千问翻译,中英日韩等多语种互译", "contextWindow": 131072, "capabilities": [ diff --git a/compute/providers/google.json b/compute/providers/google.json index aad7e8f..ccc9387 100644 --- a/compute/providers/google.json +++ b/compute/providers/google.json @@ -1,7 +1,7 @@ { "id": "provider-google-001", "provider": "google", - "label": "Google AI", + "label": "Google Gemini", "baseUrl": "https://generativelanguage.googleapis.com/v1beta", "apiFormat": "google-generative-ai", "apiKeyRef": "google", @@ -19,7 +19,7 @@ { "modelName": "gemini-2.5-pro", "displayName": "Gemini 2.5 Pro", - "serviceType": "chat", + "serviceType": ["chat"], "description": "Google 最新旗舰模型,100万上下文,多模态", "contextWindow": 1000000, "maxOutputTokens": 65536, @@ -38,7 +38,7 @@ { "modelName": "gemini-2.5-flash", "displayName": "Gemini 2.5 Flash", - "serviceType": "chat", + "serviceType": ["chat"], "description": "Google 高性价比模型,100万上下文,速度快", "contextWindow": 1000000, "maxOutputTokens": 65536, @@ -58,7 +58,7 @@ { "modelName": "gemini-2.5-pro-vision", "displayName": "Gemini 2.5 Pro Vision", - "serviceType": "vision", + "serviceType": ["vision"], "description": "Google 多模态旗舰,支持图像视频理解", "contextWindow": 1000000, "maxOutputTokens": 65536, @@ -76,7 +76,7 @@ { "modelName": "text-embedding-005", "displayName": "Text Embedding 005", - "serviceType": "embedding", + "serviceType": ["embedding"], "description": "Google 文本嵌入模型,768维度", "capabilities": [ "text_embedding", diff --git a/compute/providers/kling.json b/compute/providers/kling.json index 3b1af7c..738a645 100644 --- a/compute/providers/kling.json +++ b/compute/providers/kling.json @@ -1,7 +1,7 @@ { "id": "provider-kling-001", "provider": "kling", - "label": "可灵 AI", + "label": "快手 可灵", "baseUrl": "https://api.klingai.com/v1", "apiFormat": "openai-completions", "apiKeyRef": "kling", @@ -15,7 +15,7 @@ { "modelName": "kling-v2-5-turbo", "displayName": "可灵 V2.5 Turbo", - "serviceType": "video_gen", + "serviceType": ["video_gen"], "description": "可灵最新视频生成模型,高性价比", "capabilities": [ "video_generation", @@ -35,7 +35,7 @@ { "modelName": "kling-v2-5-turbo-pro", "displayName": "可灵 V2.5 Turbo Pro", - "serviceType": "video_gen", + "serviceType": ["video_gen"], "description": "可灵高品质视频生成", "capabilities": [ "video_generation", @@ -54,7 +54,7 @@ { "modelName": "kling-v2", "displayName": "可灵 V2", - "serviceType": "video_gen", + "serviceType": ["video_gen"], "description": "可灵文生视频旗舰模型", "capabilities": [ "video_generation", @@ -72,7 +72,7 @@ { "modelName": "kling-v2-master", "displayName": "可灵 V2 Master", - "serviceType": "video_gen", + "serviceType": ["video_gen"], "description": "可灵最高品质视频生成", "capabilities": [ "video_generation", diff --git a/compute/providers/lingyiwanwu.json b/compute/providers/lingyiwanwu.json index a104d56..6fb31ad 100644 --- a/compute/providers/lingyiwanwu.json +++ b/compute/providers/lingyiwanwu.json @@ -1,7 +1,7 @@ { "id": "provider-lingyiwanwu-001", "provider": "lingyiwanwu", - "label": "零一万物", + "label": "零一万物 Yi", "baseUrl": "https://api.lingyiwanwu.com/v1", "apiFormat": "openai-completions", "apiKeyRef": "lingyiwanwu", @@ -18,7 +18,7 @@ { "modelName": "yi-lightning", "displayName": "Yi Lightning", - "serviceType": "chat", + "serviceType": ["chat"], "description": "零一万物高性能模型,16K上下文,智能路由,极高性价比", "contextWindow": 16000, "maxOutputTokens": 4096, @@ -36,7 +36,7 @@ { "modelName": "yi-vision-v2", "displayName": "Yi Vision V2", - "serviceType": "vision", + "serviceType": ["vision"], "description": "零一万物复杂视觉任务模型,16K上下文,支持多图分析", "contextWindow": 16000, "maxOutputTokens": 4096, diff --git a/compute/providers/moonshot.json b/compute/providers/moonshot.json index 7c6f936..072d836 100644 --- a/compute/providers/moonshot.json +++ b/compute/providers/moonshot.json @@ -1,7 +1,7 @@ { "id": "provider-moonshot-001", "provider": "moonshot", - "label": "月之暗面", + "label": "月之暗面 Kimi", "baseUrl": "https://api.moonshot.cn/v1", "apiFormat": "openai-completions", "apiKeyRef": "moonshot", @@ -19,7 +19,7 @@ "modelName": "kimi-k2.5", "apiModelId": "kimi-2.5", "displayName": "Kimi K2.5", - "serviceType": "chat", + "serviceType": ["chat"], "description": "月之暗面Kimi K2.5,2026年1月发布的开源新一代模型", "contextWindow": 256000, "maxOutputTokens": 16384, @@ -38,7 +38,7 @@ { "modelName": "kimi-k2", "displayName": "Kimi K2", - "serviceType": "chat", + "serviceType": ["chat"], "description": "月之暗面Kimi K2,万亿参数MoE模型(320B激活),128K上下文,专为智能体设计", "contextWindow": 128000, "maxOutputTokens": 8192, @@ -57,7 +57,7 @@ { "modelName": "kimi-k2-thinking", "displayName": "Kimi K2 思考版", - "serviceType": "reasoning", + "serviceType": ["reasoning"], "description": "月之暗面Kimi K2思考版,256K上下文,深度推理能力", "contextWindow": 256000, "maxOutputTokens": 16384, @@ -75,7 +75,7 @@ { "modelName": "moonshot-v1-8k", "displayName": "Moonshot V1 8K", - "serviceType": "chat", + "serviceType": ["chat"], "description": "月之暗面标准模型,8K上下文", "contextWindow": 8192, "maxOutputTokens": 4096, @@ -90,7 +90,7 @@ { "modelName": "moonshot-v1-32k", "displayName": "Moonshot V1 32K", - "serviceType": "chat", + "serviceType": ["chat"], "description": "月之暗面标准模型,32K上下文", "contextWindow": 32768, "maxOutputTokens": 4096, @@ -106,7 +106,7 @@ { "modelName": "moonshot-v1-128k", "displayName": "Moonshot V1 128K", - "serviceType": "chat", + "serviceType": ["chat"], "description": "月之暗面标准模型,128K上下文", "contextWindow": 131072, "maxOutputTokens": 4096, diff --git a/compute/providers/openai.json b/compute/providers/openai.json index a217de1..4303b05 100644 --- a/compute/providers/openai.json +++ b/compute/providers/openai.json @@ -1,7 +1,7 @@ { "id": "provider-openai-001", "provider": "openai", - "label": "OpenAI", + "label": "OpenAI GPT", "baseUrl": "https://api.openai.com/v1", "apiFormat": "openai-completions", "apiKeyRef": "openai", @@ -26,7 +26,7 @@ { "modelName": "gpt-5.2", "displayName": "GPT-5.2", - "serviceType": "chat", + "serviceType": ["chat"], "description": "OpenAI 最新旗舰模型,400K 上下文,编码和智能体任务优化", "contextWindow": 400000, "maxOutputTokens": 128000, @@ -45,7 +45,7 @@ { "modelName": "gpt-5.2-pro", "displayName": "GPT-5.2 Pro", - "serviceType": "chat", + "serviceType": ["chat"], "description": "OpenAI GPT-5.2 Pro 专业版,最高性能", "contextWindow": 400000, "maxOutputTokens": 128000, @@ -64,7 +64,7 @@ { "modelName": "gpt-5.1", "displayName": "GPT-5.1", - "serviceType": "chat", + "serviceType": ["chat"], "description": "OpenAI GPT-5.1,400K 上下文,高性能旗舰", "contextWindow": 400000, "maxOutputTokens": 128000, @@ -83,7 +83,7 @@ { "modelName": "gpt-5", "displayName": "GPT-5", - "serviceType": "chat", + "serviceType": ["chat"], "description": "OpenAI GPT-5 新一代旗舰模型,统一所有模型能力", "contextWindow": 400000, "maxOutputTokens": 128000, @@ -102,7 +102,7 @@ { "modelName": "gpt-5-pro", "displayName": "GPT-5 Pro", - "serviceType": "chat", + "serviceType": ["chat"], "description": "OpenAI GPT-5 Pro 专业版高性能模型", "contextWindow": 400000, "maxOutputTokens": 128000, @@ -121,7 +121,7 @@ { "modelName": "gpt-5-mini", "displayName": "GPT-5-mini", - "serviceType": "chat", + "serviceType": ["chat"], "description": "高性价比模型,适合日常对话", "contextWindow": 128000, "maxOutputTokens": 32768, @@ -139,7 +139,7 @@ { "modelName": "gpt-5-nano", "displayName": "GPT-5-nano", - "serviceType": "fast", + "serviceType": ["fast"], "description": "极速响应模型,适合简单任务", "contextWindow": 128000, "maxOutputTokens": 16384, @@ -155,7 +155,7 @@ { "modelName": "gpt-4.1", "displayName": "GPT-4.1", - "serviceType": "chat", + "serviceType": ["chat"], "description": "OpenAI GPT-4.1,支持 1M 上下文,编码和指令遵循能力大幅提升", "contextWindow": 1000000, "maxOutputTokens": 32768, @@ -174,7 +174,7 @@ { "modelName": "gpt-4.1-mini", "displayName": "GPT-4.1 mini", - "serviceType": "chat", + "serviceType": ["chat"], "description": "OpenAI GPT-4.1 mini,1M 上下文,高性价比", "contextWindow": 1000000, "maxOutputTokens": 32768, @@ -192,7 +192,7 @@ { "modelName": "gpt-4.1-nano", "displayName": "GPT-4.1 nano", - "serviceType": "fast", + "serviceType": ["fast"], "description": "OpenAI GPT-4.1 nano,1M 上下文,极致性价比", "contextWindow": 1000000, "maxOutputTokens": 32768, @@ -209,7 +209,7 @@ { "modelName": "gpt-4o", "displayName": "GPT-4o", - "serviceType": "vision", + "serviceType": ["vision"], "description": "多模态模型,支持图像理解", "contextWindow": 128000, "maxOutputTokens": 16384, @@ -226,7 +226,7 @@ { "modelName": "gpt-4o-mini", "displayName": "GPT-4o mini", - "serviceType": "vision", + "serviceType": ["vision"], "description": "OpenAI GPT-4o mini,高性价比多模态模型", "contextWindow": 128000, "maxOutputTokens": 16384, @@ -243,7 +243,7 @@ { "modelName": "text-embedding-3-small", "displayName": "text-embedding-3-small", - "serviceType": "embedding", + "serviceType": ["embedding"], "description": "轻量文本向量化模型,1536 维", "capabilities": [ "text_embedding", @@ -255,7 +255,7 @@ { "modelName": "text-embedding-3-large", "displayName": "text-embedding-3-large", - "serviceType": "embedding", + "serviceType": ["embedding"], "description": "高精度文本向量化模型,3072 维", "capabilities": [ "text_embedding", @@ -267,7 +267,7 @@ { "modelName": "tts-1", "displayName": "TTS-1", - "serviceType": "tts", + "serviceType": ["tts"], "description": "标准语音合成", "capabilities": [ "tts" @@ -292,7 +292,7 @@ { "modelName": "tts-1-hd", "displayName": "TTS-1-HD", - "serviceType": "tts", + "serviceType": ["tts"], "description": "高清语音合成", "capabilities": [ "tts" @@ -317,7 +317,7 @@ { "modelName": "whisper-1", "displayName": "Whisper", - "serviceType": "asr", + "serviceType": ["asr"], "description": "通用语音识别", "capabilities": [ "asr", @@ -340,7 +340,7 @@ { "modelName": "o3", "displayName": "o3", - "serviceType": "responses", + "serviceType": ["responses"], "description": "OpenAI o3 推理模型,200K 上下文,支持工具调用", "contextWindow": 200000, "maxOutputTokens": 100000, @@ -359,7 +359,7 @@ { "modelName": "o3-pro", "displayName": "o3-pro", - "serviceType": "responses", + "serviceType": ["responses"], "description": "OpenAI o3-pro 高级推理模型,200K 上下文", "contextWindow": 200000, "maxOutputTokens": 100000, @@ -377,7 +377,7 @@ { "modelName": "o3-mini", "displayName": "o3-mini", - "serviceType": "responses", + "serviceType": ["responses"], "description": "OpenAI o3-mini 推理模型,128K 上下文", "contextWindow": 128000, "maxOutputTokens": 100000, @@ -394,7 +394,7 @@ { "modelName": "o4-mini", "displayName": "o4-mini", - "serviceType": "responses", + "serviceType": ["responses"], "description": "OpenAI o4-mini 推理模型,200K 上下文,支持工具调用", "contextWindow": 200000, "maxOutputTokens": 100000, @@ -412,7 +412,7 @@ { "modelName": "dall-e-3", "displayName": "DALL-E 3", - "serviceType": "image_gen", + "serviceType": ["image_gen"], "description": "高质量文生图模型", "capabilities": [ "image_generation", @@ -433,7 +433,7 @@ { "modelName": "gpt-4o-realtime", "displayName": "GPT-4o Realtime", - "serviceType": "omni", + "serviceType": ["omni"], "description": "全能实时模型,支持语音 + 文本 + 视觉", "contextWindow": 128000, "capabilities": [ @@ -454,7 +454,7 @@ { "modelName": "gpt-4o-realtime-preview", "displayName": "GPT-4o Realtime Preview", - "serviceType": "realtime_voice", + "serviceType": ["realtime_voice"], "description": "实时语音交互模型", "capabilities": [ "realtime", diff --git a/compute/providers/perplexity.json b/compute/providers/perplexity.json index 13d88bd..bea9710 100644 --- a/compute/providers/perplexity.json +++ b/compute/providers/perplexity.json @@ -1,7 +1,7 @@ { "id": "provider-perplexity-001", "provider": "perplexity", - "label": "Perplexity AI", + "label": "Perplexity Sonar", "baseUrl": "https://api.perplexity.ai", "apiFormat": "openai-completions", "apiKeyRef": "perplexity", diff --git a/compute/providers/stability.json b/compute/providers/stability.json index d037aed..80eb59b 100644 --- a/compute/providers/stability.json +++ b/compute/providers/stability.json @@ -1,7 +1,7 @@ { "id": "provider-stability-001", "provider": "stability", - "label": "Stability AI", + "label": "Stability SDXL", "baseUrl": "https://api.stability.ai/v2beta", "apiFormat": "openai-completions", "apiKeyRef": "stability", @@ -15,7 +15,7 @@ { "modelName": "stable-diffusion-3.5-large", "displayName": "Stable Diffusion 3.5 Large", - "serviceType": "image_gen", + "serviceType": ["image_gen"], "description": "Stability AI 旗舰文生图模型", "capabilities": [ "image_generation", diff --git a/compute/providers/volcengine.json b/compute/providers/volcengine.json index 7d738c2..a50a2f8 100644 --- a/compute/providers/volcengine.json +++ b/compute/providers/volcengine.json @@ -1,7 +1,7 @@ { "id": "provider-volcengine-001", "provider": "volcengine", - "label": "火山引擎", + "label": "火山 豆包", "baseUrl": "https://ark.cn-beijing.volces.com/api/v3", "apiFormat": "openai-completions", "apiKeyRef": "volcengine", @@ -24,7 +24,7 @@ { "modelName": "doubao-2.0-pro", "displayName": "豆包 2.0 Pro", - "serviceType": "chat", + "serviceType": ["chat"], "description": "火山引擎豆包2.0旗舰模型,256K上下文,相比Gemini 3 Pro和GPT 5.2有较大成本优势", "contextWindow": 256000, "maxOutputTokens": 16384, @@ -43,7 +43,7 @@ { "modelName": "doubao-seed-1.8", "displayName": "豆包 Seed-1.8", - "serviceType": "chat", + "serviceType": ["chat"], "description": "火山引擎豆包1.8旗舰模型,256K上下文,支持深度思考和多模态", "contextWindow": 256000, "maxOutputTokens": 16384, @@ -62,7 +62,7 @@ { "modelName": "doubao-seed-1.6", "displayName": "豆包 Seed-1.6", - "serviceType": "chat", + "serviceType": ["chat"], "description": "火山引擎豆包1.6旗舰模型,256K上下文", "contextWindow": 256000, "maxOutputTokens": 16384, @@ -81,7 +81,7 @@ { "modelName": "doubao-seed-1.6-thinking", "displayName": "豆包 Seed-1.6 思考版", - "serviceType": "chat", + "serviceType": ["chat"], "description": "火山引擎豆包1.6深度思考模型,256K上下文", "contextWindow": 256000, "maxOutputTokens": 32768, @@ -99,7 +99,7 @@ { "modelName": "doubao-seed-1.6-flash", "displayName": "豆包 Seed-1.6 Flash", - "serviceType": "chat", + "serviceType": ["chat"], "description": "火山引擎豆包1.6快速版,成本较前代下降63%", "contextWindow": 256000, "maxOutputTokens": 8192, @@ -116,7 +116,7 @@ { "modelName": "doubao-seed-1.6-lite", "displayName": "豆包 Seed-1.6 Lite", - "serviceType": "chat", + "serviceType": ["chat"], "description": "火山引擎豆包1.6轻量版,极致性价比", "contextWindow": 128000, "maxOutputTokens": 4096, @@ -132,7 +132,7 @@ { "modelName": "doubao-seed-1.6-vision", "displayName": "豆包 Seed-1.6 视觉版", - "serviceType": "vision", + "serviceType": ["vision"], "description": "火山引擎豆包1.6视觉模型,256K上下文,支持视频理解", "contextWindow": 256000, "maxOutputTokens": 16384, @@ -150,7 +150,7 @@ { "modelName": "doubao-seed-code", "displayName": "豆包 Seed-Code", - "serviceType": "chat", + "serviceType": ["chat"], "description": "火山引擎豆包编程专用模型,256K上下文", "contextWindow": 256000, "maxOutputTokens": 32768, @@ -169,7 +169,7 @@ { "modelName": "deepseek-v3.2", "displayName": "DeepSeek V3.2 (火山引擎)", - "serviceType": "chat", + "serviceType": ["chat"], "description": "火山引擎托管的DeepSeek V3.2,128K上下文", "contextWindow": 128000, "maxOutputTokens": 8000, @@ -189,7 +189,7 @@ { "modelName": "deepseek-r1", "displayName": "DeepSeek R1 (火山引擎)", - "serviceType": "reasoning", + "serviceType": ["reasoning"], "description": "火山引擎托管的DeepSeek R1推理模型,64K上下文", "contextWindow": 64000, "maxOutputTokens": 32768, @@ -210,7 +210,7 @@ { "modelName": "kimi-k2-volcengine", "displayName": "Kimi K2 (火山引擎)", - "serviceType": "chat", + "serviceType": ["chat"], "description": "火山引擎托管的Kimi K2,万亿参数MoE模型,128K上下文", "contextWindow": 128000, "maxOutputTokens": 8192, @@ -229,7 +229,7 @@ { "modelName": "glm-4-7", "displayName": "GLM-4.7 (火山引擎)", - "serviceType": "chat", + "serviceType": ["chat"], "description": "火山引擎托管的智谱GLM-4.7,200K上下文", "contextWindow": 200000, "maxOutputTokens": 200000, @@ -248,7 +248,7 @@ { "modelName": "doubao-embedding", "displayName": "豆包 Embedding", - "serviceType": "embedding", + "serviceType": ["embedding"], "description": "火山引擎豆包标准向量模型,2560维度", "contextWindow": 4096, "capabilities": [ @@ -263,7 +263,7 @@ { "modelName": "volc-mega-tts-clone", "displayName": "火山声音复刻", - "serviceType": "voice_clone", + "serviceType": ["voice_clone"], "description": "火山引擎声音复刻,支持极少样本", "capabilities": [ "voice_clone", @@ -278,7 +278,7 @@ { "modelName": "volc-realtime-voice", "displayName": "火山实时语音", - "serviceType": "realtime_voice", + "serviceType": ["realtime_voice"], "description": "火山引擎实时语音交互", "capabilities": [ "realtime", @@ -296,7 +296,7 @@ { "modelName": "volc-simultaneous", "displayName": "火山同声传译", - "serviceType": "simultaneous_interpret", + "serviceType": ["simultaneous_interpret"], "description": "火山引擎同声传译,支持中英双向", "capabilities": [ "simultaneous_interpretation", @@ -310,7 +310,7 @@ { "modelName": "volc-translation", "displayName": "火山翻译", - "serviceType": "translation", + "serviceType": ["translation"], "description": "火山引擎机器翻译,支持多语种互译", "capabilities": [ "translation", diff --git a/compute/providers/xai.json b/compute/providers/xai.json index 5d01a77..0ba91a0 100644 --- a/compute/providers/xai.json +++ b/compute/providers/xai.json @@ -1,7 +1,7 @@ { "id": "provider-xai-001", "provider": "xai", - "label": "xAI", + "label": "xAI Grok", "baseUrl": "https://api.x.ai/v1", "apiFormat": "openai-completions", "apiKeyRef": "xai", diff --git a/compute/providers/zhipu.json b/compute/providers/zhipu.json index 6e66bdd..cd6213f 100644 --- a/compute/providers/zhipu.json +++ b/compute/providers/zhipu.json @@ -1,7 +1,7 @@ { "id": "provider-zhipu-001", "provider": "zhipu", - "label": "智谱 AI", + "label": "智谱 GLM", "baseUrl": "https://open.bigmodel.cn/api/paas/v4", "apiFormat": "openai-completions", "apiKeyRef": "zhipu", @@ -20,7 +20,7 @@ { "modelName": "glm-5", "displayName": "GLM-5", - "serviceType": "chat", + "serviceType": ["chat"], "description": "智谱GLM-5,745B参数MoE架构,2026年2月发布,编程能力登顶", "contextWindow": 192000, "maxOutputTokens": 128000, @@ -40,7 +40,7 @@ { "modelName": "glm-5-plus", "displayName": "GLM-5-Plus", - "serviceType": "chat", + "serviceType": ["chat"], "description": "智谱GLM-5增强版,幻觉率仅2.6%,面向严肃医疗场景", "contextWindow": 192000, "maxOutputTokens": 128000, @@ -60,7 +60,7 @@ { "modelName": "glm-4.7", "displayName": "GLM-4.7", - "serviceType": "chat", + "serviceType": ["chat"], "description": "智谱GLM-4.7,面向超长上下文理解与生成,200K上下文", "contextWindow": 200000, "maxOutputTokens": 128000, @@ -79,7 +79,7 @@ { "modelName": "glm-4.7-thinking", "displayName": "GLM-4.7 Thinking", - "serviceType": "reasoning", + "serviceType": ["reasoning"], "description": "智谱GLM-4.7深度思考版,200K上下文", "contextWindow": 200000, "maxOutputTokens": 64000, @@ -97,7 +97,7 @@ { "modelName": "glm-4.7v", "displayName": "GLM-4.7V", - "serviceType": "vision", + "serviceType": ["vision"], "description": "智谱GLM-4.7多模态版,支持图像视频理解,200K上下文", "contextWindow": 200000, "maxOutputTokens": 8192, @@ -115,7 +115,7 @@ { "modelName": "glm-4.6", "displayName": "GLM-4.6", - "serviceType": "chat", + "serviceType": ["chat"], "description": "智谱GLM-4.6,最新一代模型,增强推理能力,128K上下文", "contextWindow": 128000, "maxOutputTokens": 8192, @@ -133,7 +133,7 @@ { "modelName": "glm-4.6v", "displayName": "GLM-4.6V", - "serviceType": "vision", + "serviceType": ["vision"], "description": "智谱GLM-4.6多模态版,支持图像理解,128K上下文", "contextWindow": 128000, "maxOutputTokens": 8192, @@ -150,7 +150,7 @@ { "modelName": "embedding-3", "displayName": "智谱 embedding-3", - "serviceType": "embedding", + "serviceType": ["embedding"], "description": "智谱嵌入模型v3,支持自定义维度,单条最大3072 tokens", "contextWindow": 3072, "capabilities": [