File tree Expand file tree Collapse file tree 4 files changed +68
-5
lines changed Expand file tree Collapse file tree 4 files changed +68
-5
lines changed Original file line number Diff line number Diff line change 1
1
<div align =" center " >
2
2
<img src =" docs/zh/_img/icon.png " width =" 450 " alt =" FlashTTS Logo " />
3
3
4
- [ 📘 Documentation] ( docs/zh/README.MD )
4
+ [ 📘 Documentation] ( docs/zh/README.MD ) | [ 📚 Deepwiki ] ( https://deepwiki.com/HuiResearch/FlashTTS )
5
5
6
6
[ 中文] ( README.MD ) | [ English] ( README_EN.MD )
7
7
@@ -212,6 +212,7 @@ flashtts infer \
212
212
--host 0.0.0.0 \
213
213
--port 8000
214
214
```
215
+ 详细部署说明,请参考:[ server.md] ( docs/zh/server/server.md )
215
216
216
217
## ⚡ 推理速度
217
218
Original file line number Diff line number Diff line change 1
1
<div align =" center " >
2
2
<img src =" docs/zh/_img/icon.png " width =" 450 " alt =" FlashTTS Logo " />
3
3
4
- [ 📘 Documentation] ( docs/en /README.MD )
4
+ [ 📘 Documentation] ( docs/zh /README.MD ) | [ 📚 Deepwiki ] ( https://deepwiki.com/HuiResearch/FlashTTS )
5
5
6
6
[ 中文] ( README.MD ) | [ English] ( README_EN.MD )
7
7
@@ -164,6 +164,8 @@ Server deployment:
164
164
--port 8000
165
165
```
166
166
167
+ For detailed deployment,please refer to: [ server.md] ( docs/en/server/server.md )
168
+
167
169
## ⚡ Inference Speed
168
170
169
171
Test environment: ` A800 GPU ` · Model: ` Spark-TTS-0.5B ` · Test script: [ speed_test.py] ( examples/speed_test.py )
Original file line number Diff line number Diff line change 5
5
1 . Refer to the installation guide: [ installation.md] ( ../get_started/installation.md )
6
6
2 . Start the server:
7
7
8
+ - spark tts
8
9
``` bash
9
10
flashtts serve \
10
11
--model_path Spark-TTS-0.5B \ # Change to your model path if needed
20
21
--host 0.0.0.0 \
21
22
--port 8000
22
23
```
24
+ - mega tts
25
+ ``` bash
26
+ flashtts serve \
27
+ --model_path MegaTTS3 \ # Change to your model path if needed
28
+ --backend vllm \ # vllm、sglang、torch、llama-cpp、mlx-lm任选一个
29
+ --llm_device cuda \
30
+ --tokenizer_device cuda \
31
+ --llm_attn_implementation sdpa \ # Recommended for torch backend
32
+ --torch_dtype " float16" \
33
+ --max_length 8192 \
34
+ --llm_gpu_memory_utilization 0.6 \
35
+ --host 0.0.0.0 \
36
+ --port 8000
37
+ ```
38
+ - orphpeus tts
39
+ ` ` ` bash
40
+ flashtts serve \
41
+ --model_path orpheus-3b-0.1-ft-bf16 \ # Change to your model path if needed
42
+ --lang english \
43
+ --backend vllm \ # vllm、sglang、torch、llama-cpp、mlx-lm任选一个
44
+ --llm_device cuda \
45
+ --detokenizer_device cuda \
46
+ --llm_attn_implementation sdpa \ # Recommended for torch backend
47
+ --torch_dtype "float16" \
48
+ --max_length 8192 \
49
+ --llm_gpu_memory_utilization 0.6 \
50
+ --host 0.0.0.0 \
51
+ --port 8000
52
+ ` ` `
23
53
24
54
3. Access the web interface:
25
55
```
Original file line number Diff line number Diff line change 4
4
5
5
1 . 参考安装文档: [ installation.md] ( ../get_started/installation.md )
6
6
2 . 启动服务:
7
+
8
+ - spark tts
7
9
``` bash
8
-
9
10
flashtts serve \
10
11
--model_path Spark-TTS-0.5B \ # 可修改为自己的模型地址
11
12
--backend vllm \ # vllm、sglang、torch、llama-cpp、mlx-lm任选一个
12
13
--llm_device cuda \
13
14
--tokenizer_device cuda \
14
15
--detokenizer_device cuda \
15
16
--wav2vec_attn_implementation sdpa \
16
- --llm_attn_implementation sdpa \ # 如果使用torch engine ,最好开启加速
17
+ --llm_attn_implementation sdpa \ # 如果backend为torch ,最好开启加速
17
18
--torch_dtype " bfloat16" \ # 对于spark-tts模型,不支持bfloat16的设备,只能设置为float32.
18
19
--max_length 32768 \
19
20
--llm_gpu_memory_utilization 0.6 \
20
21
--host 0.0.0.0 \
21
22
--port 8000
22
-
23
23
```
24
+ - mega tts
25
+ ` ` ` bash
26
+ flashtts serve \
27
+ --model_path MegaTTS3 \ # 可修改为自己的模型地址
28
+ --backend vllm \ # vllm、sglang、torch、llama-cpp、mlx-lm任选一个
29
+ --llm_device cuda \
30
+ --tokenizer_device cuda \
31
+ --llm_attn_implementation sdpa \ # 如果backend为torch,最好开启加速
32
+ --torch_dtype "float16" \
33
+ --max_length 8192 \
34
+ --llm_gpu_memory_utilization 0.6 \
35
+ --host 0.0.0.0 \
36
+ --port 8000
37
+ ` ` `
38
+ - orphpeus tts
39
+ ` ` ` bash
40
+ flashtts serve \
41
+ --model_path orpheus-3b-0.1-ft-bf16 \ # 可修改为自己的模型地址
42
+ --lang english \
43
+ --backend vllm \ # vllm、sglang、torch、llama-cpp、mlx-lm任选一个
44
+ --llm_device cuda \
45
+ --detokenizer_device cuda \
46
+ --llm_attn_implementation sdpa \ # 如果backend为torch,最好开启加速
47
+ --torch_dtype "float16" \
48
+ --max_length 8192 \
49
+ --llm_gpu_memory_utilization 0.6 \
50
+ --host 0.0.0.0 \
51
+ --port 8000
52
+ ` ` `
53
+
24
54
3. 在浏览器中访问页面
25
55
26
56
```
You can’t perform that action at this time.
0 commit comments