SWivid commited on
Commit
d3e15e3
·
1 Parent(s): e938b40

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -64,9 +64,10 @@ An initial guidance on Finetuning #57.
64
  To run inference with pretrained models, download the checkpoints from [🤗 Hugging Face](https://huggingface.co/SWivid/F5-TTS).
65
 
66
  Currently support up to 30s generation, which is the **TOTAL** length of prompt audio and the generated. Batch inference with chunks is supported by Gradio APP now.
67
- - To avoid inference failure, make sure you have seen through following instructions.
68
- - Uppercased letters will be uttered letter by letter, so use lowercased letter for normal words.
69
- - Add some spaces (blank: " ") or punctuations (e.g. "," ".") to explicitly introduce some pauses.
 
70
 
71
  ### Single Inference
72
 
 
64
  To run inference with pretrained models, download the checkpoints from [🤗 Hugging Face](https://huggingface.co/SWivid/F5-TTS).
65
 
66
  Currently support up to 30s generation, which is the **TOTAL** length of prompt audio and the generated. Batch inference with chunks is supported by Gradio APP now.
67
+ - To avoid possible inference failures, make sure you have seen through the following instructions.
68
+ - A longer prompt audio allows shorter generated output. The part longer than 30s cannot be generated properly. Consider split your text and do several separate inferences or leverage the local Gradio APP which enables a batch inference with chunks.
69
+ - Uppercased letters will be uttered letter by letter, so use lowercased letters for normal words.
70
+ - Add some spaces (blank: " ") or punctuations (e.g. "," ".") to explicitly introduce some pauses. If first few words skipped in code-switched generation (cuz different speed with different languages), this might help.
71
 
72
  ### Single Inference
73