Merge branch 'main' of https://github.com/QuentinFuxa/whisper_streaming_web
Browse files- README.md +3 -1
- src/__init__.py +0 -0
README.md
CHANGED
@@ -25,7 +25,7 @@ This project is based on [Whisper Streaming](https://github.com/ufal/whisper_str
|
|
25 |
|
26 |
#### π **Coming Soon**
|
27 |
|
28 |
-
- **Faster Word Validation** β Accelerate real-time transcription by validating high-confidence words immediately upon first appearance for whisper backends that
|
29 |
- **Enhanced Diarization Performance** β Optimize speaker identification by implementing longer steps for Diart processing and leveraging language-specific segmentation patterns to improve speaker boundary detection
|
30 |
|
31 |
|
@@ -78,6 +78,8 @@ This project is based on [Whisper Streaming](https://github.com/ufal/whisper_str
|
|
78 |
diart
|
79 |
```
|
80 |
|
|
|
|
|
81 |
|
82 |
3. **Run the FastAPI Server**:
|
83 |
|
|
|
25 |
|
26 |
#### π **Coming Soon**
|
27 |
|
28 |
+
- **Faster Word Validation** β Accelerate real-time transcription by validating high-confidence words immediately upon first appearance for whisper backends that return word & segment probabilities
|
29 |
- **Enhanced Diarization Performance** β Optimize speaker identification by implementing longer steps for Diart processing and leveraging language-specific segmentation patterns to improve speaker boundary detection
|
30 |
|
31 |
|
|
|
78 |
diart
|
79 |
```
|
80 |
|
81 |
+
Diart uses by default [pyannote.audio](https://github.com/pyannote/pyannote-audio) models from the _huggingface hub_. To use them, please follow the steps described [here](https://github.com/juanmc2005/diart?tab=readme-ov-file#get-access-to--pyannote-models).
|
82 |
+
|
83 |
|
84 |
3. **Run the FastAPI Server**:
|
85 |
|
src/__init__.py
DELETED
File without changes
|