-
Notifications
You must be signed in to change notification settings - Fork 502
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
DOC: update models for doc and readme (#2330)
- Loading branch information
Showing
19 changed files
with
874 additions
and
32 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file was deleted.
Oops, something went wrong.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,19 @@ | ||
.. _models_builtin_fishspeech-1.4: | ||
|
||
============== | ||
FishSpeech-1.4 | ||
============== | ||
|
||
- **Model Name:** FishSpeech-1.4 | ||
- **Model Family:** FishAudio | ||
- **Abilities:** text-to-audio | ||
- **Multilingual:** True | ||
|
||
Specifications | ||
^^^^^^^^^^^^^^ | ||
|
||
- **Model ID:** fishaudio/fish-speech-1.4 | ||
|
||
Execute the following command to launch the model:: | ||
|
||
xinference launch --model-name FishSpeech-1.4 --model-type audio |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,31 @@ | ||
.. _models_llm_deepseek-v2-chat-0628: | ||
|
||
======================================== | ||
deepseek-v2-chat-0628 | ||
======================================== | ||
|
||
- **Context Length:** 128000 | ||
- **Model Name:** deepseek-v2-chat-0628 | ||
- **Languages:** en, zh | ||
- **Abilities:** chat | ||
- **Description:** DeepSeek-V2-Chat-0628 is an improved version of DeepSeek-V2-Chat. | ||
|
||
Specifications | ||
^^^^^^^^^^^^^^ | ||
|
||
|
||
Model Spec 1 (pytorch, 236 Billion) | ||
++++++++++++++++++++++++++++++++++++++++ | ||
|
||
- **Model Format:** pytorch | ||
- **Model Size (in billions):** 236 | ||
- **Quantizations:** 4-bit, 8-bit, none | ||
- **Engines**: vLLM, Transformers, SGLang (vLLM and SGLang only available for quantization none) | ||
- **Model ID:** deepseek-ai/DeepSeek-V2-Chat-0628 | ||
- **Model Hubs**: `Hugging Face <https://huggingface.co/deepseek-ai/DeepSeek-V2-Chat-0628>`__, `ModelScope <https://modelscope.cn/models/deepseek-ai/DeepSeek-V2-Chat-0628>`__ | ||
|
||
Execute the following command to launch the model, remember to replace ``${quantization}`` with your | ||
chosen quantization method from the options listed above:: | ||
|
||
xinference launch --model-engine ${engine} --model-name deepseek-v2-chat-0628 --size-in-billions 236 --model-format pytorch --quantization ${quantization} | ||
|
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,47 @@ | ||
.. _models_llm_deepseek-v2-chat: | ||
|
||
======================================== | ||
deepseek-v2-chat | ||
======================================== | ||
|
||
- **Context Length:** 128000 | ||
- **Model Name:** deepseek-v2-chat | ||
- **Languages:** en, zh | ||
- **Abilities:** chat | ||
- **Description:** DeepSeek-V2, a strong Mixture-of-Experts (MoE) language model characterized by economical training and efficient inference. | ||
|
||
Specifications | ||
^^^^^^^^^^^^^^ | ||
|
||
|
||
Model Spec 1 (pytorch, 16 Billion) | ||
++++++++++++++++++++++++++++++++++++++++ | ||
|
||
- **Model Format:** pytorch | ||
- **Model Size (in billions):** 16 | ||
- **Quantizations:** 4-bit, 8-bit, none | ||
- **Engines**: vLLM, Transformers, SGLang (vLLM and SGLang only available for quantization none) | ||
- **Model ID:** deepseek-ai/DeepSeek-V2-Lite-Chat | ||
- **Model Hubs**: `Hugging Face <https://huggingface.co/deepseek-ai/DeepSeek-V2-Lite-Chat>`__, `ModelScope <https://modelscope.cn/models/deepseek-ai/DeepSeek-V2-Lite-Chat>`__ | ||
|
||
Execute the following command to launch the model, remember to replace ``${quantization}`` with your | ||
chosen quantization method from the options listed above:: | ||
|
||
xinference launch --model-engine ${engine} --model-name deepseek-v2-chat --size-in-billions 16 --model-format pytorch --quantization ${quantization} | ||
|
||
|
||
Model Spec 2 (pytorch, 236 Billion) | ||
++++++++++++++++++++++++++++++++++++++++ | ||
|
||
- **Model Format:** pytorch | ||
- **Model Size (in billions):** 236 | ||
- **Quantizations:** 4-bit, 8-bit, none | ||
- **Engines**: vLLM, Transformers, SGLang (vLLM and SGLang only available for quantization none) | ||
- **Model ID:** deepseek-ai/DeepSeek-V2-Chat | ||
- **Model Hubs**: `Hugging Face <https://huggingface.co/deepseek-ai/DeepSeek-V2-Chat>`__, `ModelScope <https://modelscope.cn/models/deepseek-ai/DeepSeek-V2-Chat>`__ | ||
|
||
Execute the following command to launch the model, remember to replace ``${quantization}`` with your | ||
chosen quantization method from the options listed above:: | ||
|
||
xinference launch --model-engine ${engine} --model-name deepseek-v2-chat --size-in-billions 236 --model-format pytorch --quantization ${quantization} | ||
|
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,31 @@ | ||
.. _models_llm_deepseek-v2.5: | ||
|
||
======================================== | ||
deepseek-v2.5 | ||
======================================== | ||
|
||
- **Context Length:** 128000 | ||
- **Model Name:** deepseek-v2.5 | ||
- **Languages:** en, zh | ||
- **Abilities:** chat | ||
- **Description:** DeepSeek-V2.5 is an upgraded version that combines DeepSeek-V2-Chat and DeepSeek-Coder-V2-Instruct. The new model integrates the general and coding abilities of the two previous versions. | ||
|
||
Specifications | ||
^^^^^^^^^^^^^^ | ||
|
||
|
||
Model Spec 1 (pytorch, 236 Billion) | ||
++++++++++++++++++++++++++++++++++++++++ | ||
|
||
- **Model Format:** pytorch | ||
- **Model Size (in billions):** 236 | ||
- **Quantizations:** 4-bit, 8-bit, none | ||
- **Engines**: vLLM, Transformers, SGLang (vLLM and SGLang only available for quantization none) | ||
- **Model ID:** deepseek-ai/DeepSeek-V2.5 | ||
- **Model Hubs**: `Hugging Face <https://huggingface.co/deepseek-ai/DeepSeek-V2.5>`__, `ModelScope <https://modelscope.cn/models/deepseek-ai/DeepSeek-V2.5>`__ | ||
|
||
Execute the following command to launch the model, remember to replace ``${quantization}`` with your | ||
chosen quantization method from the options listed above:: | ||
|
||
xinference launch --model-engine ${engine} --model-name deepseek-v2.5 --size-in-billions 236 --model-format pytorch --quantization ${quantization} | ||
|
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,47 @@ | ||
.. _models_llm_deepseek-v2: | ||
|
||
======================================== | ||
deepseek-v2 | ||
======================================== | ||
|
||
- **Context Length:** 128000 | ||
- **Model Name:** deepseek-v2 | ||
- **Languages:** en, zh | ||
- **Abilities:** generate | ||
- **Description:** DeepSeek-V2, a strong Mixture-of-Experts (MoE) language model characterized by economical training and efficient inference. | ||
|
||
Specifications | ||
^^^^^^^^^^^^^^ | ||
|
||
|
||
Model Spec 1 (pytorch, 16 Billion) | ||
++++++++++++++++++++++++++++++++++++++++ | ||
|
||
- **Model Format:** pytorch | ||
- **Model Size (in billions):** 16 | ||
- **Quantizations:** 4-bit, 8-bit, none | ||
- **Engines**: Transformers | ||
- **Model ID:** deepseek-ai/DeepSeek-V2-Lite | ||
- **Model Hubs**: `Hugging Face <https://huggingface.co/deepseek-ai/DeepSeek-V2-Lite>`__, `ModelScope <https://modelscope.cn/models/deepseek-ai/DeepSeek-V2-Lite>`__ | ||
|
||
Execute the following command to launch the model, remember to replace ``${quantization}`` with your | ||
chosen quantization method from the options listed above:: | ||
|
||
xinference launch --model-engine ${engine} --model-name deepseek-v2 --size-in-billions 16 --model-format pytorch --quantization ${quantization} | ||
|
||
|
||
Model Spec 2 (pytorch, 236 Billion) | ||
++++++++++++++++++++++++++++++++++++++++ | ||
|
||
- **Model Format:** pytorch | ||
- **Model Size (in billions):** 236 | ||
- **Quantizations:** 4-bit, 8-bit, none | ||
- **Engines**: Transformers | ||
- **Model ID:** deepseek-ai/DeepSeek-V2 | ||
- **Model Hubs**: `Hugging Face <https://huggingface.co/deepseek-ai/DeepSeek-V2>`__, `ModelScope <https://modelscope.cn/models/deepseek-ai/DeepSeek-V2>`__ | ||
|
||
Execute the following command to launch the model, remember to replace ``${quantization}`` with your | ||
chosen quantization method from the options listed above:: | ||
|
||
xinference launch --model-engine ${engine} --model-name deepseek-v2 --size-in-billions 236 --model-format pytorch --quantization ${quantization} | ||
|
Oops, something went wrong.