Skip to content

Commit 92ebdf2

Browse files
Jun-Howieqinxuye
andauthored
FEAT: support Deepseek-R1-0528 (#3539)
Co-authored-by: qinxuye <[email protected]>
1 parent a519d84 commit 92ebdf2

25 files changed

+527
-475
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -47,14 +47,14 @@ potential of cutting-edge AI models.
4747
- Support SGLang backend: [#1161](https://github.com/xorbitsai/inference/pull/1161)
4848
- Support LoRA for LLM and image models: [#1080](https://github.com/xorbitsai/inference/pull/1080)
4949
### New Models
50+
- Built-in support for [Deepseek-R1-0528](https://huggingface.co/deepseek-ai/DeepSeek-R1-0528): [#3539](https://github.com/xorbitsai/inference/pull/3539)
5051
- Built-in support for [Qwen3](https://qwenlm.github.io/blog/qwen3/): [#3347](https://github.com/xorbitsai/inference/pull/3347)
5152
- Built-in support for [Qwen2.5-Omni](https://github.com/QwenLM/Qwen2.5-Omni): [#3279](https://github.com/xorbitsai/inference/pull/3279)
5253
- Built-in support for [Skywork-OR1](https://github.com/SkyworkAI/Skywork-OR1): [#3274](https://github.com/xorbitsai/inference/pull/3274)
5354
- Built-in support for [GLM-4-0414](https://github.com/THUDM/GLM-4): [#3251](https://github.com/xorbitsai/inference/pull/3251)
5455
- Built-in support for [SeaLLMs-v3](https://github.com/DAMO-NLP-SG/DAMO-SeaLLMs): [#3248](https://github.com/xorbitsai/inference/pull/3248)
5556
- Built-in support for [paraformer-zh](https://huggingface.co/funasr/paraformer-zh): [#3236](https://github.com/xorbitsai/inference/pull/3236)
5657
- Built-in support for [InternVL3](https://internvl.github.io/blog/2025-04-11-InternVL-3.0/): [#3235](https://github.com/xorbitsai/inference/pull/3235)
57-
- Built-in support for [MegaTTS3](https://github.com/bytedance/MegaTTS3): [#3224](https://github.com/xorbitsai/inference/pull/3224)
5858
### Integrations
5959
- [Dify](https://docs.dify.ai/advanced/model-configuration/xinference): an LLMOps platform that enables developers (and even non-developers) to quickly build useful applications based on large language models, ensuring they are visual, operable, and improvable.
6060
- [FastGPT](https://github.com/labring/FastGPT): a knowledge-based platform built on the LLM, offers out-of-the-box data processing and model invocation capabilities, allows for workflow orchestration through Flow visualization.

README_zh_CN.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -43,14 +43,14 @@ Xorbits Inference(Xinference)是一个性能强大且功能全面的分布
4343
- 支持 SGLang 后端: [#1161](https://github.com/xorbitsai/inference/pull/1161)
4444
- 支持LLM和图像模型的LoRA: [#1080](https://github.com/xorbitsai/inference/pull/1080)
4545
### 新模型
46+
- 内置 [Deepseek-R1-0528](https://huggingface.co/deepseek-ai/DeepSeek-R1-0528): [#3539](https://github.com/xorbitsai/inference/pull/3539)
4647
- 内置 [Qwen3](https://qwenlm.github.io/blog/qwen3/): [#3347](https://github.com/xorbitsai/inference/pull/3347)
4748
- 内置 [Qwen2.5-Omni](https://github.com/QwenLM/Qwen2.5-Omni): [#3279](https://github.com/xorbitsai/inference/pull/3279)
4849
- 内置 [Skywork-OR1](https://github.com/SkyworkAI/Skywork-OR1): [#3274](https://github.com/xorbitsai/inference/pull/3274)
4950
- 内置 [GLM-4-0414](https://github.com/THUDM/GLM-4): [#3251](https://github.com/xorbitsai/inference/pull/3251)
5051
- 内置 [SeaLLMs-v3](https://github.com/DAMO-NLP-SG/DAMO-SeaLLMs): [#3248](https://github.com/xorbitsai/inference/pull/3248)
5152
- 内置 [paraformer-zh](https://huggingface.co/funasr/paraformer-zh): [#3236](https://github.com/xorbitsai/inference/pull/3236)
5253
- 内置 [InternVL3](https://internvl.github.io/blog/2025-04-11-InternVL-3.0/): [#3235](https://github.com/xorbitsai/inference/pull/3235)
53-
- 内置 [MegaTTS3](https://github.com/bytedance/MegaTTS3): [#3224](https://github.com/xorbitsai/inference/pull/3224)
5454
### 集成
5555
- [FastGPT](https://doc.fastai.site/docs/development/custom-models/xinference/):一个基于 LLM 大模型的开源 AI 知识库构建平台。提供了开箱即用的数据处理、模型调用、RAG 检索、可视化 AI 工作流编排等能力,帮助您轻松实现复杂的问答场景。
5656
- [Dify](https://docs.dify.ai/advanced/model-configuration/xinference): 一个涵盖了大型语言模型开发、部署、维护和优化的 LLMOps 平台。

doc/source/getting_started/installation.rst

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -60,7 +60,7 @@ Currently, supported models include:
6060
- ``codestral-v0.1``
6161
- ``Yi``, ``Yi-1.5``, ``Yi-chat``, ``Yi-1.5-chat``, ``Yi-1.5-chat-16k``
6262
- ``code-llama``, ``code-llama-python``, ``code-llama-instruct``
63-
- ``deepseek``, ``deepseek-coder``, ``deepseek-chat``, ``deepseek-coder-instruct``, ``deepseek-r1-distill-qwen``, ``deepseek-v2-chat``, ``deepseek-v2-chat-0628``, ``deepseek-v2.5``, ``deepseek-v3``, ``deepseek-r1``, ``deepseek-r1-distill-llama``
63+
- ``deepseek``, ``deepseek-coder``, ``deepseek-chat``, ``deepseek-coder-instruct``, ``deepseek-r1-distill-qwen``, ``deepseek-v2-chat``, ``deepseek-v2-chat-0628``, ``deepseek-v2.5``, ``deepseek-v3``, ``deepseek-v3-0324``, ``deepseek-r1``, ``deepseek-r1-0528``, ``deepseek-prover-v2``, ``deepseek-r1-distill-llama``
6464
- ``yi-coder``, ``yi-coder-chat``
6565
- ``codeqwen1.5``, ``codeqwen1.5-chat``
6666
- ``qwen2.5``, ``qwen2.5-coder``, ``qwen2.5-instruct``, ``qwen2.5-coder-instruct``, ``qwen2.5-instruct-1m``
@@ -74,11 +74,14 @@ Currently, supported models include:
7474
- ``codegeex4``
7575
- ``qwen1.5-chat``, ``qwen1.5-moe-chat``
7676
- ``qwen2-instruct``, ``qwen2-moe-instruct``
77+
- ``XiYanSQL-QwenCoder-2504``
7778
- ``QwQ-32B-Preview``, ``QwQ-32B``
7879
- ``marco-o1``
7980
- ``fin-r1``
8081
- ``seallms-v3``
81-
- ``skywork-or1-preview``
82+
- ``skywork-or1-preview``, ``skywork-or1``
83+
- ``HuatuoGPT-o1-Qwen2.5``, ``HuatuoGPT-o1-LLaMA-3.1``
84+
- ``DianJin-R1``
8285
- ``gemma-it``, ``gemma-2-it``, ``gemma-3-1b-it``
8386
- ``orion-chat``, ``orion-chat-rag``
8487
- ``c4ai-command-r-v01``

doc/source/locale/zh_CN/LC_MESSAGES/models/model_abilities/audio.po

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@ msgstr ""
2121

2222
#: ../../source/models/model_abilities/audio.rst:5
2323
msgid "Audio"
24-
msgstr ""
24+
msgstr "音频"
2525

2626
#: ../../source/models/model_abilities/audio.rst:7
2727
msgid "Learn how to turn audio into text or text into audio with Xinference."
@@ -358,7 +358,7 @@ msgstr "基本使用,加载模型 ``CosyVoice-300M-SFT``。"
358358
msgid ""
359359
"Please note that the latest CosyVoice 2.0 requires `use_flow_cache=True` "
360360
"for stream generation."
361-
msgstr ""
361+
msgstr "请注意,最新版本的 CosyVoice 2.0 在进行流式生成时需要设置 `use_flow_cache=True`。"
362362

363363
#: ../../source/models/model_abilities/audio.rst:422
364364
msgid ""

doc/source/models/builtin/audio/index.rst

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -55,6 +55,12 @@ The following is a list of built-in audio models in Xinference:
5555

5656
paraformer-zh
5757

58+
paraformer-zh-hotword
59+
60+
paraformer-zh-long
61+
62+
paraformer-zh-spk
63+
5864
sensevoicesmall
5965

6066
whisper-base
Lines changed: 19 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,19 @@
1+
.. _models_builtin_paraformer-zh-hotword:
2+
3+
=====================
4+
paraformer-zh-hotword
5+
=====================
6+
7+
- **Model Name:** paraformer-zh-hotword
8+
- **Model Family:** funasr
9+
- **Abilities:** ['audio2text']
10+
- **Multilingual:** False
11+
12+
Specifications
13+
^^^^^^^^^^^^^^
14+
15+
- **Model ID:** JunHowie/speech_paraformer-large-contextual_asr_nat-zh-cn-16k-common-vocab8404
16+
17+
Execute the following command to launch the model::
18+
19+
xinference launch --model-name paraformer-zh-hotword --model-type audio
Lines changed: 19 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,19 @@
1+
.. _models_builtin_paraformer-zh-long:
2+
3+
==================
4+
paraformer-zh-long
5+
==================
6+
7+
- **Model Name:** paraformer-zh-long
8+
- **Model Family:** funasr
9+
- **Abilities:** ['audio2text']
10+
- **Multilingual:** False
11+
12+
Specifications
13+
^^^^^^^^^^^^^^
14+
15+
- **Model ID:** JunHowie/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
16+
17+
Execute the following command to launch the model::
18+
19+
xinference launch --model-name paraformer-zh-long --model-type audio
Lines changed: 19 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,19 @@
1+
.. _models_builtin_paraformer-zh-spk:
2+
3+
=================
4+
paraformer-zh-spk
5+
=================
6+
7+
- **Model Name:** paraformer-zh-spk
8+
- **Model Family:** funasr
9+
- **Abilities:** ['audio2text']
10+
- **Multilingual:** False
11+
12+
Specifications
13+
^^^^^^^^^^^^^^
14+
15+
- **Model ID:** JunHowie/speech_paraformer-large-vad-punc-spk_asr_nat-zh-cn
16+
17+
Execute the following command to launch the model::
18+
19+
xinference launch --model-name paraformer-zh-spk --model-type audio

doc/source/models/builtin/llm/cogvlm2-video-llama3-chat.rst

Lines changed: 0 additions & 31 deletions
This file was deleted.

doc/source/models/builtin/llm/cogvlm2.rst

Lines changed: 0 additions & 47 deletions
This file was deleted.

0 commit comments

Comments
 (0)