Spaces:
Runtime error
Runtime error
tonic
commited on
Commit
Β·
33d9042
1
Parent(s):
89d01e6
Laion WhisperSpeech Demo
Browse files- README.md +3 -3
- app.py +61 -0
- requirements.txt +3 -1
- whisperspeech/__init__.py +1 -0
- whisperspeech/_modidx.py +615 -0
- whisperspeech/a2wav.py +45 -0
- whisperspeech/extract_acoustic.py +56 -0
- whisperspeech/fetch_models.py +17 -0
- whisperspeech/languages.py +131 -0
- whisperspeech/modules.py +331 -0
- whisperspeech/pipeline.py +93 -0
- whisperspeech/prepare_s2a_dataset.py +112 -0
- whisperspeech/prepare_t2s_dataset.py +111 -0
- whisperspeech/s2a_delar_mup_wds.py +688 -0
- whisperspeech/s2a_delar_mup_wds_mlang.py +564 -0
- whisperspeech/t2s_up_wds.py +442 -0
- whisperspeech/t2s_up_wds_mlang_enclm.py +519 -0
- whisperspeech/train.py +271 -0
- whisperspeech/train_multi.py +263 -0
- whisperspeech/utils.py +159 -0
- whisperspeech/vad.py +71 -0
- whisperspeech/vq_stoks.py +493 -0
- whisperspeech/wer_metrics.py +77 -0
- whisperspeech/wh_transcribe.py +146 -0
README.md
CHANGED
|
@@ -1,12 +1,12 @@
|
|
| 1 |
---
|
| 2 |
-
title:
|
| 3 |
-
emoji:
|
| 4 |
colorFrom: pink
|
| 5 |
colorTo: gray
|
| 6 |
sdk: gradio
|
| 7 |
sdk_version: 4.15.0
|
| 8 |
app_file: app.py
|
| 9 |
-
pinned:
|
| 10 |
license: mit
|
| 11 |
---
|
| 12 |
|
|
|
|
| 1 |
---
|
| 2 |
+
title: WhisperSpeech
|
| 3 |
+
emoji: π¬οΈπ¬π
|
| 4 |
colorFrom: pink
|
| 5 |
colorTo: gray
|
| 6 |
sdk: gradio
|
| 7 |
sdk_version: 4.15.0
|
| 8 |
app_file: app.py
|
| 9 |
+
pinned: True
|
| 10 |
license: mit
|
| 11 |
---
|
| 12 |
|
app.py
ADDED
|
@@ -0,0 +1,61 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import spaces
|
| 2 |
+
import gradio as gr
|
| 3 |
+
import os
|
| 4 |
+
from whisperspeech.pipeline import Pipeline
|
| 5 |
+
import torch
|
| 6 |
+
import torch.nn.functional as F
|
| 7 |
+
from whisperspeech.languages import LANGUAGES
|
| 8 |
+
from whisperspeech.pipeline import Pipeline
|
| 9 |
+
import tempfil
|
| 10 |
+
|
| 11 |
+
title = """#ππ»ββοΈ Welcome toπTonic'sπ¬οΈπ¬πWhisperSpeech
|
| 12 |
+
You can use this ZeroGPU Space to test out the current model [π¬οΈπ¬πcollabora/whisperspeech](https://huggingface.co/collabora/whisperspeech). π¬οΈπ¬πcollabora/whisperspeech is An Open Source text-to-speech system built by inverting Whisper. Previously known as spear-tts-pytorch. It's like Stable Diffusion but for speech β both powerful and easily customizable.
|
| 13 |
+
You can also use π¬οΈπ¬πWhisperSpeech by cloning this space. π§¬π¬π Simply click here: <a style="display:inline-block" href="https://huggingface.co/spaces/Tonic/laion-whisper?duplicate=true"><img src="https://img.shields.io/badge/-Duplicate%20Space-blue?labelColor=white&style=flat&logo=data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAAABAAAAAQCAYAAAAf8/9hAAAAAXNSR0IArs4c6QAAAP5JREFUOE+lk7FqAkEURY+ltunEgFXS2sZGIbXfEPdLlnxJyDdYB62sbbUKpLbVNhyYFzbrrA74YJlh9r079973psed0cvUD4A+4HoCjsA85X0Dfn/RBLBgBDxnQPfAEJgBY+A9gALA4tcbamSzS4xq4FOQAJgCDwV2CPKV8tZAJcAjMMkUe1vX+U+SMhfAJEHasQIWmXNN3abzDwHUrgcRGmYcgKe0bxrblHEB4E/pndMazNpSZGcsZdBlYJcEL9Afo75molJyM2FxmPgmgPqlWNLGfwZGG6UiyEvLzHYDmoPkDDiNm9JR9uboiONcBXrpY1qmgs21x1QwyZcpvxt9NS09PlsPAAAAAElFTkSuQmCC&logoWidth=14" alt="Duplicate Space"></a></h3>
|
| 14 |
+
Join us : πTeamTonicπ is always making cool demos! Join our active builder'sπ οΈcommunity π» [](https://discord.gg/GWpVpekp) On π€Huggingface: [TeamTonic](https://huggingface.co/TeamTonic) & [MultiTransformer](https://huggingface.co/MultiTransformer) On πGithub: [Polytonic](https://github.com/tonic-ai) & contribute to π [Poly](https://github.com/tonic-ai/poly) π€Big thanks to Yuvi Sharma and all the folks at huggingface for the community grant π€
|
| 15 |
+
"""
|
| 16 |
+
|
| 17 |
+
@spaces.GPU
|
| 18 |
+
|
| 19 |
+
def whisper_speech_demo(text, lang, speaker_audio=None, mix_lang=None, mix_text=None):
|
| 20 |
+
pipe = Pipeline(s2a_ref='collabora/whisperspeech:s2a-q4-tiny-en+pl.model')
|
| 21 |
+
|
| 22 |
+
# Use uploaded speaker audio if provided
|
| 23 |
+
speaker_url = None
|
| 24 |
+
if speaker_audio is not None:
|
| 25 |
+
speaker_url = speaker_audio.name
|
| 26 |
+
|
| 27 |
+
if mix_lang and mix_text:
|
| 28 |
+
mixed_langs = lang.split(',') + mix_lang.split(',')
|
| 29 |
+
mixed_texts = [text] + mix_text.split(',')
|
| 30 |
+
stoks = pipe.t2s.generate(mixed_texts, lang=mixed_langs)
|
| 31 |
+
audio_data = pipe.generate(stoks, speaker_url, lang=mixed_langs[0])
|
| 32 |
+
else:
|
| 33 |
+
audio_data = pipe.generate(text, speaker_url, lang)
|
| 34 |
+
|
| 35 |
+
with tempfile.NamedTemporaryFile(delete=False, suffix='.wav') as tmp_file:
|
| 36 |
+
tmp_file_name = tmp_file.name
|
| 37 |
+
with open(tmp_file_name, 'wb') as file:
|
| 38 |
+
file.write(audio_data)
|
| 39 |
+
|
| 40 |
+
return tmp_file_name
|
| 41 |
+
|
| 42 |
+
with gr.Blocks() as demo:
|
| 43 |
+
gr.Markdown(title)
|
| 44 |
+
with gr.Row():
|
| 45 |
+
text_input = gr.Textbox(label="Enter text")
|
| 46 |
+
lang_input = gr.Dropdown(choices=list(LANGUAGES.keys()), label="Language")
|
| 47 |
+
speaker_input = gr.File(label="Upload Speaker Audio (optional)", type="file", accepts=["audio/*"])
|
| 48 |
+
with gr.Row():
|
| 49 |
+
mix_lang_input = gr.Textbox(label="Mixed Languages (optional, comma-separated)", placeholder="e.g., en,pl")
|
| 50 |
+
mix_text_input = gr.Textbox(label="Mixed Texts (optional, for mixed languages)", placeholder="e.g., Hello, CzeΕΔ")
|
| 51 |
+
with gr.Row():
|
| 52 |
+
submit_button = gr.Button("Generate Speech")
|
| 53 |
+
output_audio = gr.Audio(label="Generated Speech")
|
| 54 |
+
|
| 55 |
+
submit_button.click(
|
| 56 |
+
whisper_speech_demo,
|
| 57 |
+
inputs=[text_input, lang_input, speaker_input, mix_lang_input, mix_text_input],
|
| 58 |
+
outputs=output_audio
|
| 59 |
+
)
|
| 60 |
+
|
| 61 |
+
demo.launch()
|
requirements.txt
CHANGED
|
@@ -1 +1,3 @@
|
|
| 1 |
-
|
|
|
|
|
|
|
|
|
| 1 |
+
torch
|
| 2 |
+
transformers
|
| 3 |
+
accelerate
|
whisperspeech/__init__.py
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
__version__ = "0.5.6"
|
whisperspeech/_modidx.py
ADDED
|
@@ -0,0 +1,615 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Autogenerated by nbdev
|
| 2 |
+
|
| 3 |
+
d = { 'settings': { 'branch': 'master',
|
| 4 |
+
'doc_baseurl': '/WhisperSpeech',
|
| 5 |
+
'doc_host': 'https://collabora.github.io',
|
| 6 |
+
'git_url': 'https://github.com/collabora/WhisperSpeech',
|
| 7 |
+
'lib_path': 'whisperspeech'},
|
| 8 |
+
'syms': { 'whisperspeech.a2wav': { 'whisperspeech.a2wav.Vocoder': ('6. quality-boosting vocoder.html#vocoder', 'whisperspeech/a2wav.py'),
|
| 9 |
+
'whisperspeech.a2wav.Vocoder.__init__': ( '6. quality-boosting vocoder.html#vocoder.__init__',
|
| 10 |
+
'whisperspeech/a2wav.py'),
|
| 11 |
+
'whisperspeech.a2wav.Vocoder.decode': ( '6. quality-boosting vocoder.html#vocoder.decode',
|
| 12 |
+
'whisperspeech/a2wav.py'),
|
| 13 |
+
'whisperspeech.a2wav.Vocoder.decode_to_file': ( '6. quality-boosting '
|
| 14 |
+
'vocoder.html#vocoder.decode_to_file',
|
| 15 |
+
'whisperspeech/a2wav.py'),
|
| 16 |
+
'whisperspeech.a2wav.Vocoder.decode_to_notebook': ( '6. quality-boosting '
|
| 17 |
+
'vocoder.html#vocoder.decode_to_notebook',
|
| 18 |
+
'whisperspeech/a2wav.py')},
|
| 19 |
+
'whisperspeech.extract_acoustic': { 'whisperspeech.extract_acoustic.extract_Atoks': ( '1. acoustic token '
|
| 20 |
+
'extraction.html#extract_atoks',
|
| 21 |
+
'whisperspeech/extract_acoustic.py'),
|
| 22 |
+
'whisperspeech.extract_acoustic.extract_acoustic': ( '1. acoustic token '
|
| 23 |
+
'extraction.html#extract_acoustic',
|
| 24 |
+
'whisperspeech/extract_acoustic.py'),
|
| 25 |
+
'whisperspeech.extract_acoustic.load': ( '1. acoustic token extraction.html#load',
|
| 26 |
+
'whisperspeech/extract_acoustic.py'),
|
| 27 |
+
'whisperspeech.extract_acoustic.load_model': ( '1. acoustic token '
|
| 28 |
+
'extraction.html#load_model',
|
| 29 |
+
'whisperspeech/extract_acoustic.py')},
|
| 30 |
+
'whisperspeech.extract_semb': { 'whisperspeech.extract_semb.encode_semantic': ( '2c. whisper semantic embedding '
|
| 31 |
+
'extraction.html#encode_semantic',
|
| 32 |
+
'whisperspeech/extract_semb.py'),
|
| 33 |
+
'whisperspeech.extract_semb.extract_semantic': ( '2c. whisper semantic embedding '
|
| 34 |
+
'extraction.html#extract_semantic',
|
| 35 |
+
'whisperspeech/extract_semb.py'),
|
| 36 |
+
'whisperspeech.extract_semb.load_model': ( '2c. whisper semantic embedding '
|
| 37 |
+
'extraction.html#load_model',
|
| 38 |
+
'whisperspeech/extract_semb.py')},
|
| 39 |
+
'whisperspeech.fetch_models': { 'whisperspeech.fetch_models.main': ( '0. download models.html#main',
|
| 40 |
+
'whisperspeech/fetch_models.py')},
|
| 41 |
+
'whisperspeech.modules': { 'whisperspeech.modules.Decoder': ('a. neural modules.html#decoder', 'whisperspeech/modules.py'),
|
| 42 |
+
'whisperspeech.modules.Decoder.__init__': ( 'a. neural modules.html#decoder.__init__',
|
| 43 |
+
'whisperspeech/modules.py'),
|
| 44 |
+
'whisperspeech.modules.Decoder.forward': ( 'a. neural modules.html#decoder.forward',
|
| 45 |
+
'whisperspeech/modules.py'),
|
| 46 |
+
'whisperspeech.modules.Encoder': ('a. neural modules.html#encoder', 'whisperspeech/modules.py'),
|
| 47 |
+
'whisperspeech.modules.Encoder.__init__': ( 'a. neural modules.html#encoder.__init__',
|
| 48 |
+
'whisperspeech/modules.py'),
|
| 49 |
+
'whisperspeech.modules.Encoder.forward': ( 'a. neural modules.html#encoder.forward',
|
| 50 |
+
'whisperspeech/modules.py'),
|
| 51 |
+
'whisperspeech.modules.LayerNorm': ('a. neural modules.html#layernorm', 'whisperspeech/modules.py'),
|
| 52 |
+
'whisperspeech.modules.LayerNorm.forward': ( 'a. neural modules.html#layernorm.forward',
|
| 53 |
+
'whisperspeech/modules.py'),
|
| 54 |
+
'whisperspeech.modules.LinearHead': ( 'a. neural modules.html#linearhead',
|
| 55 |
+
'whisperspeech/modules.py'),
|
| 56 |
+
'whisperspeech.modules.MultiHeadAttention': ( 'a. neural modules.html#multiheadattention',
|
| 57 |
+
'whisperspeech/modules.py'),
|
| 58 |
+
'whisperspeech.modules.MultiHeadAttention.__init__': ( 'a. neural '
|
| 59 |
+
'modules.html#multiheadattention.__init__',
|
| 60 |
+
'whisperspeech/modules.py'),
|
| 61 |
+
'whisperspeech.modules.MultiHeadAttention.forward': ( 'a. neural '
|
| 62 |
+
'modules.html#multiheadattention.forward',
|
| 63 |
+
'whisperspeech/modules.py'),
|
| 64 |
+
'whisperspeech.modules.MultiHeadAttention.qkv_attention_pth20': ( 'a. neural '
|
| 65 |
+
'modules.html#multiheadattention.qkv_attention_pth20',
|
| 66 |
+
'whisperspeech/modules.py'),
|
| 67 |
+
'whisperspeech.modules.MultiHeadAttention.qkv_attention_vanilla': ( 'a. neural '
|
| 68 |
+
'modules.html#multiheadattention.qkv_attention_vanilla',
|
| 69 |
+
'whisperspeech/modules.py'),
|
| 70 |
+
'whisperspeech.modules.MultiHeadAttention.qkv_attention_xformers': ( 'a. neural '
|
| 71 |
+
'modules.html#multiheadattention.qkv_attention_xformers',
|
| 72 |
+
'whisperspeech/modules.py'),
|
| 73 |
+
'whisperspeech.modules.QueryHead': ('a. neural modules.html#queryhead', 'whisperspeech/modules.py'),
|
| 74 |
+
'whisperspeech.modules.ResidualAttentionBlock': ( 'a. neural modules.html#residualattentionblock',
|
| 75 |
+
'whisperspeech/modules.py'),
|
| 76 |
+
'whisperspeech.modules.ResidualAttentionBlock.__init__': ( 'a. neural '
|
| 77 |
+
'modules.html#residualattentionblock.__init__',
|
| 78 |
+
'whisperspeech/modules.py'),
|
| 79 |
+
'whisperspeech.modules.ResidualAttentionBlock.forward': ( 'a. neural '
|
| 80 |
+
'modules.html#residualattentionblock.forward',
|
| 81 |
+
'whisperspeech/modules.py'),
|
| 82 |
+
'whisperspeech.modules.Rotary': ('a. neural modules.html#rotary', 'whisperspeech/modules.py'),
|
| 83 |
+
'whisperspeech.modules.Rotary.__init__': ( 'a. neural modules.html#rotary.__init__',
|
| 84 |
+
'whisperspeech/modules.py'),
|
| 85 |
+
'whisperspeech.modules.Rotary.forward': ( 'a. neural modules.html#rotary.forward',
|
| 86 |
+
'whisperspeech/modules.py'),
|
| 87 |
+
'whisperspeech.modules.SumDecoder': ( 'a. neural modules.html#sumdecoder',
|
| 88 |
+
'whisperspeech/modules.py'),
|
| 89 |
+
'whisperspeech.modules.SumDecoder.__init__': ( 'a. neural modules.html#sumdecoder.__init__',
|
| 90 |
+
'whisperspeech/modules.py'),
|
| 91 |
+
'whisperspeech.modules.SumDecoder.forward': ( 'a. neural modules.html#sumdecoder.forward',
|
| 92 |
+
'whisperspeech/modules.py'),
|
| 93 |
+
'whisperspeech.modules.apply_rotary_pos_emb': ( 'a. neural modules.html#apply_rotary_pos_emb',
|
| 94 |
+
'whisperspeech/modules.py'),
|
| 95 |
+
'whisperspeech.modules.init_transformer': ( 'a. neural modules.html#init_transformer',
|
| 96 |
+
'whisperspeech/modules.py'),
|
| 97 |
+
'whisperspeech.modules.rotate_half': ( 'a. neural modules.html#rotate_half',
|
| 98 |
+
'whisperspeech/modules.py'),
|
| 99 |
+
'whisperspeech.modules.sinusoids': ('a. neural modules.html#sinusoids', 'whisperspeech/modules.py')},
|
| 100 |
+
'whisperspeech.pipeline': { 'whisperspeech.pipeline.Pipeline': ('7. pipeline.html#pipeline', 'whisperspeech/pipeline.py'),
|
| 101 |
+
'whisperspeech.pipeline.Pipeline.__init__': ( '7. pipeline.html#pipeline.__init__',
|
| 102 |
+
'whisperspeech/pipeline.py'),
|
| 103 |
+
'whisperspeech.pipeline.Pipeline.generate': ( '7. pipeline.html#pipeline.generate',
|
| 104 |
+
'whisperspeech/pipeline.py'),
|
| 105 |
+
'whisperspeech.pipeline.Pipeline.generate_atoks': ( '7. pipeline.html#pipeline.generate_atoks',
|
| 106 |
+
'whisperspeech/pipeline.py'),
|
| 107 |
+
'whisperspeech.pipeline.Pipeline.generate_to_file': ( '7. pipeline.html#pipeline.generate_to_file',
|
| 108 |
+
'whisperspeech/pipeline.py'),
|
| 109 |
+
'whisperspeech.pipeline.Pipeline.generate_to_notebook': ( '7. '
|
| 110 |
+
'pipeline.html#pipeline.generate_to_notebook',
|
| 111 |
+
'whisperspeech/pipeline.py')},
|
| 112 |
+
'whisperspeech.prepare_s2a_dataset': { 'whisperspeech.prepare_s2a_dataset.flac_to_s2a_name': ( '4a. s2a dataset '
|
| 113 |
+
'preparation.html#flac_to_s2a_name',
|
| 114 |
+
'whisperspeech/prepare_s2a_dataset.py'),
|
| 115 |
+
'whisperspeech.prepare_s2a_dataset.prepare_s2a': ( '4a. s2a dataset '
|
| 116 |
+
'preparation.html#prepare_s2a',
|
| 117 |
+
'whisperspeech/prepare_s2a_dataset.py'),
|
| 118 |
+
'whisperspeech.prepare_s2a_dataset.resampler': ( '4a. s2a dataset '
|
| 119 |
+
'preparation.html#resampler',
|
| 120 |
+
'whisperspeech/prepare_s2a_dataset.py')},
|
| 121 |
+
'whisperspeech.prepare_t2s_dataset': { 'whisperspeech.prepare_t2s_dataset.Transcriber': ( '5a. t2s dataset '
|
| 122 |
+
'preparation.html#transcriber',
|
| 123 |
+
'whisperspeech/prepare_t2s_dataset.py'),
|
| 124 |
+
'whisperspeech.prepare_t2s_dataset.Transcriber.__init__': ( '5a. t2s dataset '
|
| 125 |
+
'preparation.html#transcriber.__init__',
|
| 126 |
+
'whisperspeech/prepare_t2s_dataset.py'),
|
| 127 |
+
'whisperspeech.prepare_t2s_dataset.Transcriber.transcribe': ( '5a. t2s dataset '
|
| 128 |
+
'preparation.html#transcriber.transcribe',
|
| 129 |
+
'whisperspeech/prepare_t2s_dataset.py'),
|
| 130 |
+
'whisperspeech.prepare_t2s_dataset.flac_to_t2s_name': ( '5a. t2s dataset '
|
| 131 |
+
'preparation.html#flac_to_t2s_name',
|
| 132 |
+
'whisperspeech/prepare_t2s_dataset.py'),
|
| 133 |
+
'whisperspeech.prepare_t2s_dataset.prepare_t2s': ( '5a. t2s dataset '
|
| 134 |
+
'preparation.html#prepare_t2s',
|
| 135 |
+
'whisperspeech/prepare_t2s_dataset.py')},
|
| 136 |
+
'whisperspeech.s2a_delar_mup_wds': { 'whisperspeech.s2a_delar_mup_wds.CMLMVisual': ( '4b. semantic to acoustic token '
|
| 137 |
+
'modeling.html#cmlmvisual',
|
| 138 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 139 |
+
'whisperspeech.s2a_delar_mup_wds.CMLMVisual.__init__': ( '4b. semantic to acoustic token '
|
| 140 |
+
'modeling.html#cmlmvisual.__init__',
|
| 141 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 142 |
+
'whisperspeech.s2a_delar_mup_wds.CMLMVisual.add_data': ( '4b. semantic to acoustic token '
|
| 143 |
+
'modeling.html#cmlmvisual.add_data',
|
| 144 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 145 |
+
'whisperspeech.s2a_delar_mup_wds.CMLMVisual.add_table_row': ( '4b. semantic to acoustic '
|
| 146 |
+
'token '
|
| 147 |
+
'modeling.html#cmlmvisual.add_table_row',
|
| 148 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 149 |
+
'whisperspeech.s2a_delar_mup_wds.CMLMVisual.hide': ( '4b. semantic to acoustic token '
|
| 150 |
+
'modeling.html#cmlmvisual.hide',
|
| 151 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 152 |
+
'whisperspeech.s2a_delar_mup_wds.CMLMVisual.on_iter': ( '4b. semantic to acoustic token '
|
| 153 |
+
'modeling.html#cmlmvisual.on_iter',
|
| 154 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 155 |
+
'whisperspeech.s2a_delar_mup_wds.CMLMVisual.plot': ( '4b. semantic to acoustic token '
|
| 156 |
+
'modeling.html#cmlmvisual.plot',
|
| 157 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 158 |
+
'whisperspeech.s2a_delar_mup_wds.CMLMVisual.show': ( '4b. semantic to acoustic token '
|
| 159 |
+
'modeling.html#cmlmvisual.show',
|
| 160 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 161 |
+
'whisperspeech.s2a_delar_mup_wds.DelSumDecoder': ( '4b. semantic to acoustic token '
|
| 162 |
+
'modeling.html#delsumdecoder',
|
| 163 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 164 |
+
'whisperspeech.s2a_delar_mup_wds.DelSumDecoder.__init__': ( '4b. semantic to acoustic '
|
| 165 |
+
'token '
|
| 166 |
+
'modeling.html#delsumdecoder.__init__',
|
| 167 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 168 |
+
'whisperspeech.s2a_delar_mup_wds.DelSumDecoder.forward': ( '4b. semantic to acoustic '
|
| 169 |
+
'token '
|
| 170 |
+
'modeling.html#delsumdecoder.forward',
|
| 171 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 172 |
+
'whisperspeech.s2a_delar_mup_wds.EmbeddingProjector': ( '4b. semantic to acoustic token '
|
| 173 |
+
'modeling.html#embeddingprojector',
|
| 174 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 175 |
+
'whisperspeech.s2a_delar_mup_wds.MultiHeadAttention': ( '4b. semantic to acoustic token '
|
| 176 |
+
'modeling.html#multiheadattention',
|
| 177 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 178 |
+
'whisperspeech.s2a_delar_mup_wds.MultiHeadAttention.__init__': ( '4b. semantic to '
|
| 179 |
+
'acoustic token '
|
| 180 |
+
'modeling.html#multiheadattention.__init__',
|
| 181 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 182 |
+
'whisperspeech.s2a_delar_mup_wds.MultiHeadAttention.forward': ( '4b. semantic to acoustic '
|
| 183 |
+
'token '
|
| 184 |
+
'modeling.html#multiheadattention.forward',
|
| 185 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 186 |
+
'whisperspeech.s2a_delar_mup_wds.MultiHeadAttention.qkv_attention_pth20': ( '4b. semantic '
|
| 187 |
+
'to acoustic '
|
| 188 |
+
'token '
|
| 189 |
+
'modeling.html#multiheadattention.qkv_attention_pth20',
|
| 190 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 191 |
+
'whisperspeech.s2a_delar_mup_wds.MultiHeadAttention.qkv_attention_xformers': ( '4b. '
|
| 192 |
+
'semantic '
|
| 193 |
+
'to '
|
| 194 |
+
'acoustic '
|
| 195 |
+
'token '
|
| 196 |
+
'modeling.html#multiheadattention.qkv_attention_xformers',
|
| 197 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 198 |
+
'whisperspeech.s2a_delar_mup_wds.ResidualAttentionBlock': ( '4b. semantic to acoustic '
|
| 199 |
+
'token '
|
| 200 |
+
'modeling.html#residualattentionblock',
|
| 201 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 202 |
+
'whisperspeech.s2a_delar_mup_wds.ResidualAttentionBlock.__init__': ( '4b. semantic to '
|
| 203 |
+
'acoustic token '
|
| 204 |
+
'modeling.html#residualattentionblock.__init__',
|
| 205 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 206 |
+
'whisperspeech.s2a_delar_mup_wds.ResidualAttentionBlock.forward': ( '4b. semantic to '
|
| 207 |
+
'acoustic token '
|
| 208 |
+
'modeling.html#residualattentionblock.forward',
|
| 209 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 210 |
+
'whisperspeech.s2a_delar_mup_wds.Rotary': ( '4b. semantic to acoustic token '
|
| 211 |
+
'modeling.html#rotary',
|
| 212 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 213 |
+
'whisperspeech.s2a_delar_mup_wds.Rotary.__init__': ( '4b. semantic to acoustic token '
|
| 214 |
+
'modeling.html#rotary.__init__',
|
| 215 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 216 |
+
'whisperspeech.s2a_delar_mup_wds.Rotary.forward': ( '4b. semantic to acoustic token '
|
| 217 |
+
'modeling.html#rotary.forward',
|
| 218 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 219 |
+
'whisperspeech.s2a_delar_mup_wds.SADelARTransformer': ( '4b. semantic to acoustic token '
|
| 220 |
+
'modeling.html#sadelartransformer',
|
| 221 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 222 |
+
'whisperspeech.s2a_delar_mup_wds.SADelARTransformer.__init__': ( '4b. semantic to '
|
| 223 |
+
'acoustic token '
|
| 224 |
+
'modeling.html#sadelartransformer.__init__',
|
| 225 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 226 |
+
'whisperspeech.s2a_delar_mup_wds.SADelARTransformer.device': ( '4b. semantic to acoustic '
|
| 227 |
+
'token '
|
| 228 |
+
'modeling.html#sadelartransformer.device',
|
| 229 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 230 |
+
'whisperspeech.s2a_delar_mup_wds.SADelARTransformer.embed_stoks': ( '4b. semantic to '
|
| 231 |
+
'acoustic token '
|
| 232 |
+
'modeling.html#sadelartransformer.embed_stoks',
|
| 233 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 234 |
+
'whisperspeech.s2a_delar_mup_wds.SADelARTransformer.forward': ( '4b. semantic to acoustic '
|
| 235 |
+
'token '
|
| 236 |
+
'modeling.html#sadelartransformer.forward',
|
| 237 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 238 |
+
'whisperspeech.s2a_delar_mup_wds.SADelARTransformer.generate': ( '4b. semantic to '
|
| 239 |
+
'acoustic token '
|
| 240 |
+
'modeling.html#sadelartransformer.generate',
|
| 241 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 242 |
+
'whisperspeech.s2a_delar_mup_wds.SADelARTransformer.get_extra_state': ( '4b. semantic to '
|
| 243 |
+
'acoustic token '
|
| 244 |
+
'modeling.html#sadelartransformer.get_extra_state',
|
| 245 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 246 |
+
'whisperspeech.s2a_delar_mup_wds.SADelARTransformer.get_metrics': ( '4b. semantic to '
|
| 247 |
+
'acoustic token '
|
| 248 |
+
'modeling.html#sadelartransformer.get_metrics',
|
| 249 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 250 |
+
'whisperspeech.s2a_delar_mup_wds.SADelARTransformer.init_transformer': ( '4b. semantic to '
|
| 251 |
+
'acoustic token '
|
| 252 |
+
'modeling.html#sadelartransformer.init_transformer',
|
| 253 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 254 |
+
'whisperspeech.s2a_delar_mup_wds.SADelARTransformer.load_checkpoint': ( '4b. semantic to '
|
| 255 |
+
'acoustic token '
|
| 256 |
+
'modeling.html#sadelartransformer.load_checkpoint',
|
| 257 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 258 |
+
'whisperspeech.s2a_delar_mup_wds.SADelARTransformer.load_frozen_semantic_embeddings': ( '4b. '
|
| 259 |
+
'semantic '
|
| 260 |
+
'to '
|
| 261 |
+
'acoustic '
|
| 262 |
+
'token '
|
| 263 |
+
'modeling.html#sadelartransformer.load_frozen_semantic_embeddings',
|
| 264 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 265 |
+
'whisperspeech.s2a_delar_mup_wds.SADelARTransformer.load_model': ( '4b. semantic to '
|
| 266 |
+
'acoustic token '
|
| 267 |
+
'modeling.html#sadelartransformer.load_model',
|
| 268 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 269 |
+
'whisperspeech.s2a_delar_mup_wds.SADelARTransformer.save_model': ( '4b. semantic to '
|
| 270 |
+
'acoustic token '
|
| 271 |
+
'modeling.html#sadelartransformer.save_model',
|
| 272 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 273 |
+
'whisperspeech.s2a_delar_mup_wds.SADelARTransformer.set_extra_state': ( '4b. semantic to '
|
| 274 |
+
'acoustic token '
|
| 275 |
+
'modeling.html#sadelartransformer.set_extra_state',
|
| 276 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 277 |
+
'whisperspeech.s2a_delar_mup_wds.SADelARTransformer.setup': ( '4b. semantic to acoustic '
|
| 278 |
+
'token '
|
| 279 |
+
'modeling.html#sadelartransformer.setup',
|
| 280 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 281 |
+
'whisperspeech.s2a_delar_mup_wds.Tunables': ( '4b. semantic to acoustic token '
|
| 282 |
+
'modeling.html#tunables',
|
| 283 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 284 |
+
'whisperspeech.s2a_delar_mup_wds.Tunables.__post_init__': ( '4b. semantic to acoustic '
|
| 285 |
+
'token '
|
| 286 |
+
'modeling.html#tunables.__post_init__',
|
| 287 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 288 |
+
'whisperspeech.s2a_delar_mup_wds.Tunables.upgrade': ( '4b. semantic to acoustic token '
|
| 289 |
+
'modeling.html#tunables.upgrade',
|
| 290 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 291 |
+
'whisperspeech.s2a_delar_mup_wds._make_model': ( '4b. semantic to acoustic token '
|
| 292 |
+
'modeling.html#_make_model',
|
| 293 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 294 |
+
'whisperspeech.s2a_delar_mup_wds.apply_rotary_pos_emb': ( '4b. semantic to acoustic token '
|
| 295 |
+
'modeling.html#apply_rotary_pos_emb',
|
| 296 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 297 |
+
'whisperspeech.s2a_delar_mup_wds.load_datasets': ( '4b. semantic to acoustic token '
|
| 298 |
+
'modeling.html#load_datasets',
|
| 299 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 300 |
+
'whisperspeech.s2a_delar_mup_wds.make_model': ( '4b. semantic to acoustic token '
|
| 301 |
+
'modeling.html#make_model',
|
| 302 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 303 |
+
'whisperspeech.s2a_delar_mup_wds.pad_samples': ( '4b. semantic to acoustic token '
|
| 304 |
+
'modeling.html#pad_samples',
|
| 305 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 306 |
+
'whisperspeech.s2a_delar_mup_wds.rand': ( '4b. semantic to acoustic token '
|
| 307 |
+
'modeling.html#rand',
|
| 308 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 309 |
+
'whisperspeech.s2a_delar_mup_wds.random_trunc': ( '4b. semantic to acoustic token '
|
| 310 |
+
'modeling.html#random_trunc',
|
| 311 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 312 |
+
'whisperspeech.s2a_delar_mup_wds.rotate_half': ( '4b. semantic to acoustic token '
|
| 313 |
+
'modeling.html#rotate_half',
|
| 314 |
+
'whisperspeech/s2a_delar_mup_wds.py'),
|
| 315 |
+
'whisperspeech.s2a_delar_mup_wds.speaker_id_extractor': ( '4b. semantic to acoustic token '
|
| 316 |
+
'modeling.html#speaker_id_extractor',
|
| 317 |
+
'whisperspeech/s2a_delar_mup_wds.py')},
|
| 318 |
+
'whisperspeech.t2s_up_wds': { 'whisperspeech.t2s_up_wds.CharTokenizer': ( '5b. text to semantic token '
|
| 319 |
+
'modeling.html#chartokenizer',
|
| 320 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 321 |
+
'whisperspeech.t2s_up_wds.CharTokenizer.decode': ( '5b. text to semantic token '
|
| 322 |
+
'modeling.html#chartokenizer.decode',
|
| 323 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 324 |
+
'whisperspeech.t2s_up_wds.CharTokenizer.encode': ( '5b. text to semantic token '
|
| 325 |
+
'modeling.html#chartokenizer.encode',
|
| 326 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 327 |
+
'whisperspeech.t2s_up_wds.Decoder': ( '5b. text to semantic token modeling.html#decoder',
|
| 328 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 329 |
+
'whisperspeech.t2s_up_wds.Decoder.__init__': ( '5b. text to semantic token '
|
| 330 |
+
'modeling.html#decoder.__init__',
|
| 331 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 332 |
+
'whisperspeech.t2s_up_wds.Decoder.forward': ( '5b. text to semantic token '
|
| 333 |
+
'modeling.html#decoder.forward',
|
| 334 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 335 |
+
'whisperspeech.t2s_up_wds.EmbeddingProjector': ( '5b. text to semantic token '
|
| 336 |
+
'modeling.html#embeddingprojector',
|
| 337 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 338 |
+
'whisperspeech.t2s_up_wds.Encoder': ( '5b. text to semantic token modeling.html#encoder',
|
| 339 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 340 |
+
'whisperspeech.t2s_up_wds.Encoder.__init__': ( '5b. text to semantic token '
|
| 341 |
+
'modeling.html#encoder.__init__',
|
| 342 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 343 |
+
'whisperspeech.t2s_up_wds.Encoder.forward': ( '5b. text to semantic token '
|
| 344 |
+
'modeling.html#encoder.forward',
|
| 345 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 346 |
+
'whisperspeech.t2s_up_wds.TSARTransformer': ( '5b. text to semantic token '
|
| 347 |
+
'modeling.html#tsartransformer',
|
| 348 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 349 |
+
'whisperspeech.t2s_up_wds.TSARTransformer.__init__': ( '5b. text to semantic token '
|
| 350 |
+
'modeling.html#tsartransformer.__init__',
|
| 351 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 352 |
+
'whisperspeech.t2s_up_wds.TSARTransformer.device': ( '5b. text to semantic token '
|
| 353 |
+
'modeling.html#tsartransformer.device',
|
| 354 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 355 |
+
'whisperspeech.t2s_up_wds.TSARTransformer.ensure_tokenizer': ( '5b. text to semantic token '
|
| 356 |
+
'modeling.html#tsartransformer.ensure_tokenizer',
|
| 357 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 358 |
+
'whisperspeech.t2s_up_wds.TSARTransformer.forward': ( '5b. text to semantic token '
|
| 359 |
+
'modeling.html#tsartransformer.forward',
|
| 360 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 361 |
+
'whisperspeech.t2s_up_wds.TSARTransformer.generate': ( '5b. text to semantic token '
|
| 362 |
+
'modeling.html#tsartransformer.generate',
|
| 363 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 364 |
+
'whisperspeech.t2s_up_wds.TSARTransformer.generate_batch': ( '5b. text to semantic token '
|
| 365 |
+
'modeling.html#tsartransformer.generate_batch',
|
| 366 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 367 |
+
'whisperspeech.t2s_up_wds.TSARTransformer.init_transformer': ( '5b. text to semantic token '
|
| 368 |
+
'modeling.html#tsartransformer.init_transformer',
|
| 369 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 370 |
+
'whisperspeech.t2s_up_wds.TSARTransformer.load_checkpoint': ( '5b. text to semantic token '
|
| 371 |
+
'modeling.html#tsartransformer.load_checkpoint',
|
| 372 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 373 |
+
'whisperspeech.t2s_up_wds.TSARTransformer.load_frozen_semantic_embeddings': ( '5b. text to '
|
| 374 |
+
'semantic token '
|
| 375 |
+
'modeling.html#tsartransformer.load_frozen_semantic_embeddings',
|
| 376 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 377 |
+
'whisperspeech.t2s_up_wds.TSARTransformer.load_model': ( '5b. text to semantic token '
|
| 378 |
+
'modeling.html#tsartransformer.load_model',
|
| 379 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 380 |
+
'whisperspeech.t2s_up_wds.TSARTransformer.save_model': ( '5b. text to semantic token '
|
| 381 |
+
'modeling.html#tsartransformer.save_model',
|
| 382 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 383 |
+
'whisperspeech.t2s_up_wds.TSARTransformer.setup': ( '5b. text to semantic token '
|
| 384 |
+
'modeling.html#tsartransformer.setup',
|
| 385 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 386 |
+
'whisperspeech.t2s_up_wds.Tunables': ( '5b. text to semantic token modeling.html#tunables',
|
| 387 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 388 |
+
'whisperspeech.t2s_up_wds.Tunables.__post_init__': ( '5b. text to semantic token '
|
| 389 |
+
'modeling.html#tunables.__post_init__',
|
| 390 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 391 |
+
'whisperspeech.t2s_up_wds._make_model': ( '5b. text to semantic token modeling.html#_make_model',
|
| 392 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 393 |
+
'whisperspeech.t2s_up_wds.ar_padder': ( '5b. text to semantic token modeling.html#ar_padder',
|
| 394 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 395 |
+
'whisperspeech.t2s_up_wds.build_speaker_map': ( '5b. text to semantic token '
|
| 396 |
+
'modeling.html#build_speaker_map',
|
| 397 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 398 |
+
'whisperspeech.t2s_up_wds.char_per_seconder': ( '5b. text to semantic token '
|
| 399 |
+
'modeling.html#char_per_seconder',
|
| 400 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 401 |
+
'whisperspeech.t2s_up_wds.load_datasets': ( '5b. text to semantic token '
|
| 402 |
+
'modeling.html#load_datasets',
|
| 403 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 404 |
+
'whisperspeech.t2s_up_wds.make_model': ( '5b. text to semantic token modeling.html#make_model',
|
| 405 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 406 |
+
'whisperspeech.t2s_up_wds.rand': ( '5b. text to semantic token modeling.html#rand',
|
| 407 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 408 |
+
'whisperspeech.t2s_up_wds.speaker_id_extractor': ( '5b. text to semantic token '
|
| 409 |
+
'modeling.html#speaker_id_extractor',
|
| 410 |
+
'whisperspeech/t2s_up_wds.py'),
|
| 411 |
+
'whisperspeech.t2s_up_wds.tokenizer': ( '5b. text to semantic token modeling.html#tokenizer',
|
| 412 |
+
'whisperspeech/t2s_up_wds.py')},
|
| 413 |
+
'whisperspeech.train': { 'whisperspeech.train.SimpleVisual': ('b1. training.html#simplevisual', 'whisperspeech/train.py'),
|
| 414 |
+
'whisperspeech.train.SimpleVisual.__init__': ( 'b1. training.html#simplevisual.__init__',
|
| 415 |
+
'whisperspeech/train.py'),
|
| 416 |
+
'whisperspeech.train.SimpleVisual.add_data': ( 'b1. training.html#simplevisual.add_data',
|
| 417 |
+
'whisperspeech/train.py'),
|
| 418 |
+
'whisperspeech.train.SimpleVisual.add_table_row': ( 'b1. training.html#simplevisual.add_table_row',
|
| 419 |
+
'whisperspeech/train.py'),
|
| 420 |
+
'whisperspeech.train.SimpleVisual.hide': ( 'b1. training.html#simplevisual.hide',
|
| 421 |
+
'whisperspeech/train.py'),
|
| 422 |
+
'whisperspeech.train.SimpleVisual.on_iter': ( 'b1. training.html#simplevisual.on_iter',
|
| 423 |
+
'whisperspeech/train.py'),
|
| 424 |
+
'whisperspeech.train.SimpleVisual.plot': ( 'b1. training.html#simplevisual.plot',
|
| 425 |
+
'whisperspeech/train.py'),
|
| 426 |
+
'whisperspeech.train.SimpleVisual.show': ( 'b1. training.html#simplevisual.show',
|
| 427 |
+
'whisperspeech/train.py'),
|
| 428 |
+
'whisperspeech.train.train': ('b1. training.html#train', 'whisperspeech/train.py'),
|
| 429 |
+
'whisperspeech.train.validate': ('b1. training.html#validate', 'whisperspeech/train.py')},
|
| 430 |
+
'whisperspeech.train_multi': { 'whisperspeech.train_multi.TrainingTask': ( 'b2. training (lightning).html#trainingtask',
|
| 431 |
+
'whisperspeech/train_multi.py'),
|
| 432 |
+
'whisperspeech.train_multi.TrainingTask.__init__': ( 'b2. training '
|
| 433 |
+
'(lightning).html#trainingtask.__init__',
|
| 434 |
+
'whisperspeech/train_multi.py'),
|
| 435 |
+
'whisperspeech.train_multi.TrainingTask.configure_optimizers': ( 'b2. training '
|
| 436 |
+
'(lightning).html#trainingtask.configure_optimizers',
|
| 437 |
+
'whisperspeech/train_multi.py'),
|
| 438 |
+
'whisperspeech.train_multi.TrainingTask.on_fit_start': ( 'b2. training '
|
| 439 |
+
'(lightning).html#trainingtask.on_fit_start',
|
| 440 |
+
'whisperspeech/train_multi.py'),
|
| 441 |
+
'whisperspeech.train_multi.TrainingTask.on_validation_epoch_end': ( 'b2. training '
|
| 442 |
+
'(lightning).html#trainingtask.on_validation_epoch_end',
|
| 443 |
+
'whisperspeech/train_multi.py'),
|
| 444 |
+
'whisperspeech.train_multi.TrainingTask.test_step': ( 'b2. training '
|
| 445 |
+
'(lightning).html#trainingtask.test_step',
|
| 446 |
+
'whisperspeech/train_multi.py'),
|
| 447 |
+
'whisperspeech.train_multi.TrainingTask.training_step': ( 'b2. training '
|
| 448 |
+
'(lightning).html#trainingtask.training_step',
|
| 449 |
+
'whisperspeech/train_multi.py'),
|
| 450 |
+
'whisperspeech.train_multi.TrainingTask.validation_step': ( 'b2. training '
|
| 451 |
+
'(lightning).html#trainingtask.validation_step',
|
| 452 |
+
'whisperspeech/train_multi.py'),
|
| 453 |
+
'whisperspeech.train_multi.parse_and_call': ( 'b2. training (lightning).html#parse_and_call',
|
| 454 |
+
'whisperspeech/train_multi.py')},
|
| 455 |
+
'whisperspeech.vad': { 'whisperspeech.vad.extract_segments': ( '1b. voice activity detection.html#extract_segments',
|
| 456 |
+
'whisperspeech/vad.py'),
|
| 457 |
+
'whisperspeech.vad.fix_dots_in_names': ( '1b. voice activity detection.html#fix_dots_in_names',
|
| 458 |
+
'whisperspeech/vad.py'),
|
| 459 |
+
'whisperspeech.vad.flac_to_vad_name': ( '1b. voice activity detection.html#flac_to_vad_name',
|
| 460 |
+
'whisperspeech/vad.py'),
|
| 461 |
+
'whisperspeech.vad.load_dataset': ( '1b. voice activity detection.html#load_dataset',
|
| 462 |
+
'whisperspeech/vad.py'),
|
| 463 |
+
'whisperspeech.vad.process_shard': ( '1b. voice activity detection.html#process_shard',
|
| 464 |
+
'whisperspeech/vad.py'),
|
| 465 |
+
'whisperspeech.vad.segment_audio': ( '1b. voice activity detection.html#segment_audio',
|
| 466 |
+
'whisperspeech/vad.py')},
|
| 467 |
+
'whisperspeech.verify_wds': { 'whisperspeech.verify_wds.process_shard': ( '0. verify webdataset archives.html#process_shard',
|
| 468 |
+
'whisperspeech/verify_wds.py')},
|
| 469 |
+
'whisperspeech.vq_stoks': { 'whisperspeech.vq_stoks.RQBottleneckTransformer': ( '2b. whisper quantization (semantic token) '
|
| 470 |
+
'model.html#rqbottlenecktransformer',
|
| 471 |
+
'whisperspeech/vq_stoks.py'),
|
| 472 |
+
'whisperspeech.vq_stoks.RQBottleneckTransformer.__init__': ( '2b. whisper quantization (semantic '
|
| 473 |
+
'token) '
|
| 474 |
+
'model.html#rqbottlenecktransformer.__init__',
|
| 475 |
+
'whisperspeech/vq_stoks.py'),
|
| 476 |
+
'whisperspeech.vq_stoks.RQBottleneckTransformer.decode_text': ( '2b. whisper quantization '
|
| 477 |
+
'(semantic token) '
|
| 478 |
+
'model.html#rqbottlenecktransformer.decode_text',
|
| 479 |
+
'whisperspeech/vq_stoks.py'),
|
| 480 |
+
'whisperspeech.vq_stoks.RQBottleneckTransformer.dequantize': ( '2b. whisper quantization (semantic '
|
| 481 |
+
'token) '
|
| 482 |
+
'model.html#rqbottlenecktransformer.dequantize',
|
| 483 |
+
'whisperspeech/vq_stoks.py'),
|
| 484 |
+
'whisperspeech.vq_stoks.RQBottleneckTransformer.device': ( '2b. whisper quantization (semantic '
|
| 485 |
+
'token) '
|
| 486 |
+
'model.html#rqbottlenecktransformer.device',
|
| 487 |
+
'whisperspeech/vq_stoks.py'),
|
| 488 |
+
'whisperspeech.vq_stoks.RQBottleneckTransformer.downsample_embeddings': ( '2b. whisper '
|
| 489 |
+
'quantization (semantic '
|
| 490 |
+
'token) '
|
| 491 |
+
'model.html#rqbottlenecktransformer.downsample_embeddings',
|
| 492 |
+
'whisperspeech/vq_stoks.py'),
|
| 493 |
+
'whisperspeech.vq_stoks.RQBottleneckTransformer.encode_audio': ( '2b. whisper quantization '
|
| 494 |
+
'(semantic token) '
|
| 495 |
+
'model.html#rqbottlenecktransformer.encode_audio',
|
| 496 |
+
'whisperspeech/vq_stoks.py'),
|
| 497 |
+
'whisperspeech.vq_stoks.RQBottleneckTransformer.encode_mel': ( '2b. whisper quantization (semantic '
|
| 498 |
+
'token) '
|
| 499 |
+
'model.html#rqbottlenecktransformer.encode_mel',
|
| 500 |
+
'whisperspeech/vq_stoks.py'),
|
| 501 |
+
'whisperspeech.vq_stoks.RQBottleneckTransformer.ensure_whisper': ( '2b. whisper quantization '
|
| 502 |
+
'(semantic token) '
|
| 503 |
+
'model.html#rqbottlenecktransformer.ensure_whisper',
|
| 504 |
+
'whisperspeech/vq_stoks.py'),
|
| 505 |
+
'whisperspeech.vq_stoks.RQBottleneckTransformer.extract_teacher': ( '2b. whisper quantization '
|
| 506 |
+
'(semantic token) '
|
| 507 |
+
'model.html#rqbottlenecktransformer.extract_teacher',
|
| 508 |
+
'whisperspeech/vq_stoks.py'),
|
| 509 |
+
'whisperspeech.vq_stoks.RQBottleneckTransformer.forward': ( '2b. whisper quantization (semantic '
|
| 510 |
+
'token) '
|
| 511 |
+
'model.html#rqbottlenecktransformer.forward',
|
| 512 |
+
'whisperspeech/vq_stoks.py'),
|
| 513 |
+
'whisperspeech.vq_stoks.RQBottleneckTransformer.get_metrics': ( '2b. whisper quantization '
|
| 514 |
+
'(semantic token) '
|
| 515 |
+
'model.html#rqbottlenecktransformer.get_metrics',
|
| 516 |
+
'whisperspeech/vq_stoks.py'),
|
| 517 |
+
'whisperspeech.vq_stoks.RQBottleneckTransformer.init_transformer': ( '2b. whisper quantization '
|
| 518 |
+
'(semantic token) '
|
| 519 |
+
'model.html#rqbottlenecktransformer.init_transformer',
|
| 520 |
+
'whisperspeech/vq_stoks.py'),
|
| 521 |
+
'whisperspeech.vq_stoks.RQBottleneckTransformer.load_checkpoint': ( '2b. whisper quantization '
|
| 522 |
+
'(semantic token) '
|
| 523 |
+
'model.html#rqbottlenecktransformer.load_checkpoint',
|
| 524 |
+
'whisperspeech/vq_stoks.py'),
|
| 525 |
+
'whisperspeech.vq_stoks.RQBottleneckTransformer.load_model': ( '2b. whisper quantization (semantic '
|
| 526 |
+
'token) '
|
| 527 |
+
'model.html#rqbottlenecktransformer.load_model',
|
| 528 |
+
'whisperspeech/vq_stoks.py'),
|
| 529 |
+
'whisperspeech.vq_stoks.RQBottleneckTransformer.quantize': ( '2b. whisper quantization (semantic '
|
| 530 |
+
'token) '
|
| 531 |
+
'model.html#rqbottlenecktransformer.quantize',
|
| 532 |
+
'whisperspeech/vq_stoks.py'),
|
| 533 |
+
'whisperspeech.vq_stoks.RQBottleneckTransformer.save_model': ( '2b. whisper quantization (semantic '
|
| 534 |
+
'token) '
|
| 535 |
+
'model.html#rqbottlenecktransformer.save_model',
|
| 536 |
+
'whisperspeech/vq_stoks.py'),
|
| 537 |
+
'whisperspeech.vq_stoks.RQBottleneckTransformer.setup': ( '2b. whisper quantization (semantic '
|
| 538 |
+
'token) '
|
| 539 |
+
'model.html#rqbottlenecktransformer.setup',
|
| 540 |
+
'whisperspeech/vq_stoks.py'),
|
| 541 |
+
'whisperspeech.vq_stoks.Tunables': ( '2b. whisper quantization (semantic token) '
|
| 542 |
+
'model.html#tunables',
|
| 543 |
+
'whisperspeech/vq_stoks.py'),
|
| 544 |
+
'whisperspeech.vq_stoks.Tunables.__post_init__': ( '2b. whisper quantization (semantic token) '
|
| 545 |
+
'model.html#tunables.__post_init__',
|
| 546 |
+
'whisperspeech/vq_stoks.py'),
|
| 547 |
+
'whisperspeech.vq_stoks.Tunables.upgrade': ( '2b. whisper quantization (semantic token) '
|
| 548 |
+
'model.html#tunables.upgrade',
|
| 549 |
+
'whisperspeech/vq_stoks.py'),
|
| 550 |
+
'whisperspeech.vq_stoks.add_masks': ( '2b. whisper quantization (semantic token) '
|
| 551 |
+
'model.html#add_masks',
|
| 552 |
+
'whisperspeech/vq_stoks.py'),
|
| 553 |
+
'whisperspeech.vq_stoks.derived_dataset': ( '2b. whisper quantization (semantic token) '
|
| 554 |
+
'model.html#derived_dataset',
|
| 555 |
+
'whisperspeech/vq_stoks.py'),
|
| 556 |
+
'whisperspeech.vq_stoks.load_datasets': ( '2b. whisper quantization (semantic token) '
|
| 557 |
+
'model.html#load_datasets',
|
| 558 |
+
'whisperspeech/vq_stoks.py'),
|
| 559 |
+
'whisperspeech.vq_stoks.logrand': ( '2b. whisper quantization (semantic token) model.html#logrand',
|
| 560 |
+
'whisperspeech/vq_stoks.py'),
|
| 561 |
+
'whisperspeech.vq_stoks.make_model': ( '2b. whisper quantization (semantic token) '
|
| 562 |
+
'model.html#make_model',
|
| 563 |
+
'whisperspeech/vq_stoks.py'),
|
| 564 |
+
'whisperspeech.vq_stoks.merge_in': ( '2b. whisper quantization (semantic token) '
|
| 565 |
+
'model.html#merge_in',
|
| 566 |
+
'whisperspeech/vq_stoks.py'),
|
| 567 |
+
'whisperspeech.vq_stoks.rand': ( '2b. whisper quantization (semantic token) model.html#rand',
|
| 568 |
+
'whisperspeech/vq_stoks.py'),
|
| 569 |
+
'whisperspeech.vq_stoks.tokenize_text': ( '2b. whisper quantization (semantic token) '
|
| 570 |
+
'model.html#tokenize_text',
|
| 571 |
+
'whisperspeech/vq_stoks.py')},
|
| 572 |
+
'whisperspeech.wer_metrics': { 'whisperspeech.wer_metrics.DfBuilder': ( 'c. word error rate metrics.html#dfbuilder',
|
| 573 |
+
'whisperspeech/wer_metrics.py'),
|
| 574 |
+
'whisperspeech.wer_metrics.DfBuilder.__init__': ( 'c. word error rate '
|
| 575 |
+
'metrics.html#dfbuilder.__init__',
|
| 576 |
+
'whisperspeech/wer_metrics.py'),
|
| 577 |
+
'whisperspeech.wer_metrics.DfBuilder.df': ( 'c. word error rate metrics.html#dfbuilder.df',
|
| 578 |
+
'whisperspeech/wer_metrics.py'),
|
| 579 |
+
'whisperspeech.wer_metrics.DfBuilder.push': ( 'c. word error rate metrics.html#dfbuilder.push',
|
| 580 |
+
'whisperspeech/wer_metrics.py'),
|
| 581 |
+
'whisperspeech.wer_metrics.WERStats': ( 'c. word error rate metrics.html#werstats',
|
| 582 |
+
'whisperspeech/wer_metrics.py'),
|
| 583 |
+
'whisperspeech.wer_metrics.WERStats.__init__': ( 'c. word error rate '
|
| 584 |
+
'metrics.html#werstats.__init__',
|
| 585 |
+
'whisperspeech/wer_metrics.py'),
|
| 586 |
+
'whisperspeech.wer_metrics.WERStats.push_sample': ( 'c. word error rate '
|
| 587 |
+
'metrics.html#werstats.push_sample',
|
| 588 |
+
'whisperspeech/wer_metrics.py'),
|
| 589 |
+
'whisperspeech.wer_metrics.librispeech_data': ( 'c. word error rate '
|
| 590 |
+
'metrics.html#librispeech_data',
|
| 591 |
+
'whisperspeech/wer_metrics.py'),
|
| 592 |
+
'whisperspeech.wer_metrics.whisper_normalize': ( 'c. word error rate '
|
| 593 |
+
'metrics.html#whisper_normalize',
|
| 594 |
+
'whisperspeech/wer_metrics.py')},
|
| 595 |
+
'whisperspeech.wh_transcribe': { 'whisperspeech.wh_transcribe.chunk_merger': ( '2a. whisper quantization dataset '
|
| 596 |
+
'preparation.html#chunk_merger',
|
| 597 |
+
'whisperspeech/wh_transcribe.py'),
|
| 598 |
+
'whisperspeech.wh_transcribe.flac_to_txt_name': ( '2a. whisper quantization dataset '
|
| 599 |
+
'preparation.html#flac_to_txt_name',
|
| 600 |
+
'whisperspeech/wh_transcribe.py'),
|
| 601 |
+
'whisperspeech.wh_transcribe.merge_in': ( '2a. whisper quantization dataset '
|
| 602 |
+
'preparation.html#merge_in',
|
| 603 |
+
'whisperspeech/wh_transcribe.py'),
|
| 604 |
+
'whisperspeech.wh_transcribe.process_shard': ( '2a. whisper quantization dataset '
|
| 605 |
+
'preparation.html#process_shard',
|
| 606 |
+
'whisperspeech/wh_transcribe.py'),
|
| 607 |
+
'whisperspeech.wh_transcribe.random_cutter': ( '2a. whisper quantization dataset '
|
| 608 |
+
'preparation.html#random_cutter',
|
| 609 |
+
'whisperspeech/wh_transcribe.py'),
|
| 610 |
+
'whisperspeech.wh_transcribe.split_to_chunks': ( '2a. whisper quantization dataset '
|
| 611 |
+
'preparation.html#split_to_chunks',
|
| 612 |
+
'whisperspeech/wh_transcribe.py'),
|
| 613 |
+
'whisperspeech.wh_transcribe.wds_compose': ( '2a. whisper quantization dataset '
|
| 614 |
+
'preparation.html#wds_compose',
|
| 615 |
+
'whisperspeech/wh_transcribe.py')}}}
|
whisperspeech/a2wav.py
ADDED
|
@@ -0,0 +1,45 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/6. Quality-boosting vocoder.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = ['Vocoder']
|
| 5 |
+
|
| 6 |
+
# %% ../nbs/6. Quality-boosting vocoder.ipynb 1
|
| 7 |
+
from vocos import Vocos
|
| 8 |
+
import torch
|
| 9 |
+
import torchaudio
|
| 10 |
+
|
| 11 |
+
# %% ../nbs/6. Quality-boosting vocoder.ipynb 2
|
| 12 |
+
class Vocoder:
|
| 13 |
+
def __init__(self, repo_id="charactr/vocos-encodec-24khz"):
|
| 14 |
+
self.vocos = Vocos.from_pretrained(repo_id).cuda()
|
| 15 |
+
|
| 16 |
+
def is_notebook(self):
|
| 17 |
+
try:
|
| 18 |
+
return get_ipython().__class__.__name__ == "ZMQInteractiveShell"
|
| 19 |
+
except:
|
| 20 |
+
return False
|
| 21 |
+
|
| 22 |
+
@torch.no_grad()
|
| 23 |
+
def decode(self, atoks):
|
| 24 |
+
if len(atoks.shape) == 3:
|
| 25 |
+
b,q,t = atoks.shape
|
| 26 |
+
atoks = atoks.permute(1,0,2)
|
| 27 |
+
else:
|
| 28 |
+
q,t = atoks.shape
|
| 29 |
+
|
| 30 |
+
features = self.vocos.codes_to_features(atoks)
|
| 31 |
+
bandwidth_id = torch.tensor({2:0,4:1,8:2}[q]).cuda()
|
| 32 |
+
return self.vocos.decode(features, bandwidth_id=bandwidth_id)
|
| 33 |
+
|
| 34 |
+
def decode_to_file(self, fname, atoks):
|
| 35 |
+
audio = self.decode(atoks)
|
| 36 |
+
torchaudio.save(fname, audio.cpu(), 24000)
|
| 37 |
+
if self.is_notebook():
|
| 38 |
+
from IPython.display import display, HTML, Audio
|
| 39 |
+
display(HTML(f'<a href="{fname}" target="_blank">Listen to {fname}</a>'))
|
| 40 |
+
|
| 41 |
+
def decode_to_notebook(self, atoks):
|
| 42 |
+
from IPython.display import display, HTML, Audio
|
| 43 |
+
|
| 44 |
+
audio = self.decode(atoks)
|
| 45 |
+
display(Audio(audio.cpu().numpy(), rate=24000))
|
whisperspeech/extract_acoustic.py
ADDED
|
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/1. Acoustic token extraction.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = ['load', 'load_model', 'extract_Atoks', 'extract_acoustic']
|
| 5 |
+
|
| 6 |
+
# %% ../nbs/1. Acoustic token extraction.ipynb 2
|
| 7 |
+
import torch
|
| 8 |
+
import torchaudio
|
| 9 |
+
import gc
|
| 10 |
+
|
| 11 |
+
from pathlib import Path
|
| 12 |
+
from fastcore.script import *
|
| 13 |
+
from fastprogress import progress_bar, master_bar
|
| 14 |
+
|
| 15 |
+
# %% ../nbs/1. Acoustic token extraction.ipynb 5
|
| 16 |
+
def load(fname, newsr=24000):
|
| 17 |
+
"""Load an audio file to the GPU and resample to `newsr`."""
|
| 18 |
+
x, sr = torchaudio.load(fname)
|
| 19 |
+
_tform = torchaudio.transforms.Resample(sr, newsr)
|
| 20 |
+
return _tform(x).cuda().unsqueeze(0)
|
| 21 |
+
|
| 22 |
+
# %% ../nbs/1. Acoustic token extraction.ipynb 6
|
| 23 |
+
def load_model():
|
| 24 |
+
"Load the pretrained EnCodec model"
|
| 25 |
+
from encodec.model import EncodecModel
|
| 26 |
+
model = EncodecModel.encodec_model_24khz()
|
| 27 |
+
model.set_target_bandwidth(1.5)
|
| 28 |
+
model.cuda().eval();
|
| 29 |
+
return model
|
| 30 |
+
|
| 31 |
+
# %% ../nbs/1. Acoustic token extraction.ipynb 7
|
| 32 |
+
def extract_Atoks(model, audio):
|
| 33 |
+
"""Extract EnCodec tokens for the given `audio` tensor (or file path)
|
| 34 |
+
using the given `model` (see `load_model`)."""
|
| 35 |
+
if isinstance(audio, (Path, str)):
|
| 36 |
+
audio = load(audio)
|
| 37 |
+
with torch.no_grad():
|
| 38 |
+
frames = torch.cat([model.encode(segment)[0][0]
|
| 39 |
+
for segment in torch.split(audio, 320*20000, dim=-1)], dim=-1)
|
| 40 |
+
return frames
|
| 41 |
+
|
| 42 |
+
# %% ../nbs/1. Acoustic token extraction.ipynb 8
|
| 43 |
+
@call_parse
|
| 44 |
+
def extract_acoustic(
|
| 45 |
+
srcdir:Path, # source dir, should contain *.flac files
|
| 46 |
+
outdir:Path, # output dir, will get the *.encodec files
|
| 47 |
+
):
|
| 48 |
+
"Convert audio files to .encodec files with tensors of tokens"
|
| 49 |
+
model = load_model()
|
| 50 |
+
outdir.mkdir(exist_ok=True, parents=True)
|
| 51 |
+
for name in progress_bar(list(srcdir.rglob('*.flac'))):
|
| 52 |
+
outname = outdir/name.with_suffix('.encodec').name
|
| 53 |
+
tokens = extract_Atoks(model, name)
|
| 54 |
+
torch.save(tokens, outname)
|
| 55 |
+
del tokens
|
| 56 |
+
gc.collect()
|
whisperspeech/fetch_models.py
ADDED
|
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/0. Download models.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = []
|
| 5 |
+
|
| 6 |
+
# %% ../nbs/0. Download models.ipynb 1
|
| 7 |
+
from fastcore.script import call_parse
|
| 8 |
+
import whisperx
|
| 9 |
+
import whisper
|
| 10 |
+
|
| 11 |
+
# %% ../nbs/0. Download models.ipynb 3
|
| 12 |
+
@call_parse
|
| 13 |
+
def main():
|
| 14 |
+
whisper.load_model('base.en')
|
| 15 |
+
whisper.load_model('small.en')
|
| 16 |
+
whisperx.vad.load_vad_model('cpu')
|
| 17 |
+
whisperx.asr.load_model('medium.en', "cpu", compute_type="float16", language='en')
|
whisperspeech/languages.py
ADDED
|
@@ -0,0 +1,131 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/B. Languages.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = ['to_id']
|
| 5 |
+
|
| 6 |
+
# %% ../nbs/B. Languages.ipynb 3
|
| 7 |
+
LANGUAGES = {
|
| 8 |
+
"en": "english",
|
| 9 |
+
"zh": "chinese",
|
| 10 |
+
"de": "german",
|
| 11 |
+
"es": "spanish",
|
| 12 |
+
"ru": "russian",
|
| 13 |
+
"ko": "korean",
|
| 14 |
+
"fr": "french",
|
| 15 |
+
"ja": "japanese",
|
| 16 |
+
"pt": "portuguese",
|
| 17 |
+
"tr": "turkish",
|
| 18 |
+
"pl": "polish",
|
| 19 |
+
"ca": "catalan",
|
| 20 |
+
"nl": "dutch",
|
| 21 |
+
"ar": "arabic",
|
| 22 |
+
"sv": "swedish",
|
| 23 |
+
"it": "italian",
|
| 24 |
+
"id": "indonesian",
|
| 25 |
+
"hi": "hindi",
|
| 26 |
+
"fi": "finnish",
|
| 27 |
+
"vi": "vietnamese",
|
| 28 |
+
"he": "hebrew",
|
| 29 |
+
"uk": "ukrainian",
|
| 30 |
+
"el": "greek",
|
| 31 |
+
"ms": "malay",
|
| 32 |
+
"cs": "czech",
|
| 33 |
+
"ro": "romanian",
|
| 34 |
+
"da": "danish",
|
| 35 |
+
"hu": "hungarian",
|
| 36 |
+
"ta": "tamil",
|
| 37 |
+
"no": "norwegian",
|
| 38 |
+
"th": "thai",
|
| 39 |
+
"ur": "urdu",
|
| 40 |
+
"hr": "croatian",
|
| 41 |
+
"bg": "bulgarian",
|
| 42 |
+
"lt": "lithuanian",
|
| 43 |
+
"la": "latin",
|
| 44 |
+
"mi": "maori",
|
| 45 |
+
"ml": "malayalam",
|
| 46 |
+
"cy": "welsh",
|
| 47 |
+
"sk": "slovak",
|
| 48 |
+
"te": "telugu",
|
| 49 |
+
"fa": "persian",
|
| 50 |
+
"lv": "latvian",
|
| 51 |
+
"bn": "bengali",
|
| 52 |
+
"sr": "serbian",
|
| 53 |
+
"az": "azerbaijani",
|
| 54 |
+
"sl": "slovenian",
|
| 55 |
+
"kn": "kannada",
|
| 56 |
+
"et": "estonian",
|
| 57 |
+
"mk": "macedonian",
|
| 58 |
+
"br": "breton",
|
| 59 |
+
"eu": "basque",
|
| 60 |
+
"is": "icelandic",
|
| 61 |
+
"hy": "armenian",
|
| 62 |
+
"ne": "nepali",
|
| 63 |
+
"mn": "mongolian",
|
| 64 |
+
"bs": "bosnian",
|
| 65 |
+
"kk": "kazakh",
|
| 66 |
+
"sq": "albanian",
|
| 67 |
+
"sw": "swahili",
|
| 68 |
+
"gl": "galician",
|
| 69 |
+
"mr": "marathi",
|
| 70 |
+
"pa": "punjabi",
|
| 71 |
+
"si": "sinhala",
|
| 72 |
+
"km": "khmer",
|
| 73 |
+
"sn": "shona",
|
| 74 |
+
"yo": "yoruba",
|
| 75 |
+
"so": "somali",
|
| 76 |
+
"af": "afrikaans",
|
| 77 |
+
"oc": "occitan",
|
| 78 |
+
"ka": "georgian",
|
| 79 |
+
"be": "belarusian",
|
| 80 |
+
"tg": "tajik",
|
| 81 |
+
"sd": "sindhi",
|
| 82 |
+
"gu": "gujarati",
|
| 83 |
+
"am": "amharic",
|
| 84 |
+
"yi": "yiddish",
|
| 85 |
+
"lo": "lao",
|
| 86 |
+
"uz": "uzbek",
|
| 87 |
+
"fo": "faroese",
|
| 88 |
+
"ht": "haitian creole",
|
| 89 |
+
"ps": "pashto",
|
| 90 |
+
"tk": "turkmen",
|
| 91 |
+
"nn": "nynorsk",
|
| 92 |
+
"mt": "maltese",
|
| 93 |
+
"sa": "sanskrit",
|
| 94 |
+
"lb": "luxembourgish",
|
| 95 |
+
"my": "myanmar",
|
| 96 |
+
"bo": "tibetan",
|
| 97 |
+
"tl": "tagalog",
|
| 98 |
+
"mg": "malagasy",
|
| 99 |
+
"as": "assamese",
|
| 100 |
+
"tt": "tatar",
|
| 101 |
+
"haw": "hawaiian",
|
| 102 |
+
"ln": "lingala",
|
| 103 |
+
"ha": "hausa",
|
| 104 |
+
"ba": "bashkir",
|
| 105 |
+
"jw": "javanese",
|
| 106 |
+
"su": "sundanese",
|
| 107 |
+
}
|
| 108 |
+
|
| 109 |
+
# %% ../nbs/B. Languages.ipynb 4
|
| 110 |
+
# language code lookup by name, with a few language aliases
|
| 111 |
+
TO_LANGUAGE_CODE = {
|
| 112 |
+
**{language: code for code, language in LANGUAGES.items()},
|
| 113 |
+
"burmese": "my",
|
| 114 |
+
"valencian": "ca",
|
| 115 |
+
"flemish": "nl",
|
| 116 |
+
"haitian": "ht",
|
| 117 |
+
"letzeburgesch": "lb",
|
| 118 |
+
"pushto": "ps",
|
| 119 |
+
"panjabi": "pa",
|
| 120 |
+
"moldavian": "ro",
|
| 121 |
+
"moldovan": "ro",
|
| 122 |
+
"sinhalese": "si",
|
| 123 |
+
"castilian": "es",
|
| 124 |
+
}
|
| 125 |
+
|
| 126 |
+
# %% ../nbs/B. Languages.ipynb 5
|
| 127 |
+
languages = tuple(LANGUAGES.keys())
|
| 128 |
+
|
| 129 |
+
# %% ../nbs/B. Languages.ipynb 6
|
| 130 |
+
def to_id(lang):
|
| 131 |
+
return languages.index(TO_LANGUAGE_CODE.get(lang, lang))
|
whisperspeech/modules.py
ADDED
|
@@ -0,0 +1,331 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/A. Neural modules.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = ['LayerNorm', 'LinearHead', 'QueryHead', 'init_transformer', 'sinusoids', 'MultiHeadAttention',
|
| 5 |
+
'ResidualAttentionBlock', 'BaseDecoder', 'EmbeddingProjector', 'FlexEmbeddings']
|
| 6 |
+
|
| 7 |
+
# %% ../nbs/A. Neural modules.ipynb 2
|
| 8 |
+
import torch
|
| 9 |
+
import numpy as np
|
| 10 |
+
import math
|
| 11 |
+
|
| 12 |
+
from torch import Tensor, nn
|
| 13 |
+
import torch.nn.functional as F
|
| 14 |
+
from typing import Dict, Iterable, Optional
|
| 15 |
+
|
| 16 |
+
# import xformers.ops as xops
|
| 17 |
+
|
| 18 |
+
# %% ../nbs/A. Neural modules.ipynb 3
|
| 19 |
+
# Code in this file is mostly borrowed from
|
| 20 |
+
# https://github.com/openai/whisper/blob/main/whisper/model.py
|
| 21 |
+
# and is under the MIT License
|
| 22 |
+
|
| 23 |
+
class LayerNorm(nn.LayerNorm):
|
| 24 |
+
def forward(self, x):
|
| 25 |
+
return super().forward(x.float()).type(x.dtype)
|
| 26 |
+
|
| 27 |
+
# Used in ΞΌP to initialize the weights and configure the optimizer
|
| 28 |
+
# These two layers map the transformer width into a fixed dimension
|
| 29 |
+
class LinearHead(nn.Linear):
|
| 30 |
+
pass
|
| 31 |
+
|
| 32 |
+
class QueryHead(nn.Linear):
|
| 33 |
+
pass
|
| 34 |
+
|
| 35 |
+
# based on https://github.com/karpathy/minGPT/blob/master/mingpt/model.py#L163
|
| 36 |
+
def init_transformer(m):
|
| 37 |
+
if isinstance(m, (nn.Linear, nn.Embedding)):
|
| 38 |
+
torch.nn.init.trunc_normal_(m.weight, std=.02)
|
| 39 |
+
if isinstance(m, nn.Linear) and m.bias is not None:
|
| 40 |
+
torch.nn.init.constant_(m.bias, 0)
|
| 41 |
+
elif isinstance(m, nn.LayerNorm):
|
| 42 |
+
torch.nn.init.constant_(m.bias, 0)
|
| 43 |
+
torch.nn.init.constant_(m.weight, 1.0)
|
| 44 |
+
|
| 45 |
+
# %% ../nbs/A. Neural modules.ipynb 4
|
| 46 |
+
def sinusoids(length, channels, max_timescale=10000):
|
| 47 |
+
"""Returns sinusoids for positional embedding"""
|
| 48 |
+
assert channels % 2 == 0
|
| 49 |
+
log_timescale_increment = np.log(max_timescale) / (channels // 2 - 1)
|
| 50 |
+
inv_timescales = torch.exp(-log_timescale_increment * torch.arange(channels // 2))
|
| 51 |
+
scaled_time = torch.arange(length)[:, np.newaxis] * inv_timescales[np.newaxis, :]
|
| 52 |
+
return torch.cat([torch.sin(scaled_time), torch.cos(scaled_time)], dim=1)
|
| 53 |
+
|
| 54 |
+
# %% ../nbs/A. Neural modules.ipynb 5
|
| 55 |
+
class MultiHeadAttention(nn.Module):
|
| 56 |
+
def __init__(self, n_state: int, n_head: int, qk_scale: float = 1, rope: bool = False, cross=False):
|
| 57 |
+
super().__init__()
|
| 58 |
+
self.n_state = n_state
|
| 59 |
+
self.n_head = n_head
|
| 60 |
+
self.sqrt_qk_scale = math.sqrt(qk_scale)
|
| 61 |
+
self.query = QueryHead(n_state, n_state)
|
| 62 |
+
self.key = nn.Linear(n_state, n_state, bias=False)
|
| 63 |
+
self.value = nn.Linear(n_state, n_state)
|
| 64 |
+
self.out = nn.Linear(n_state, n_state)
|
| 65 |
+
self.cross = cross
|
| 66 |
+
self.query_subsampling = 1
|
| 67 |
+
self.key_subsampling = 1
|
| 68 |
+
|
| 69 |
+
self.cached_kvx = None
|
| 70 |
+
self.register_buffer('k_cache', None)
|
| 71 |
+
self.register_buffer('v_cache', None)
|
| 72 |
+
|
| 73 |
+
self.rotary = None
|
| 74 |
+
if rope:
|
| 75 |
+
self.rotary = Rotary(n_state // n_head)
|
| 76 |
+
self.qkv = None
|
| 77 |
+
self.kv = None
|
| 78 |
+
|
| 79 |
+
def setup_kv_cache(self, max_batch_size, max_seq_len, dtype=torch.float32):
|
| 80 |
+
cache_shape = (max_batch_size, self.n_head, max_seq_len, self.n_state//self.n_head)
|
| 81 |
+
self.k_cache = torch.zeros(cache_shape, dtype=dtype, device=self.key.weight.device)
|
| 82 |
+
self.v_cache = torch.zeros(cache_shape, dtype=dtype, device=self.value.weight.device)
|
| 83 |
+
|
| 84 |
+
def merge_linears(self, layers, mults):
|
| 85 |
+
bias = [x.bias for x in layers if x.bias is not None][0]
|
| 86 |
+
din, dout = layers[0].weight.shape
|
| 87 |
+
new = nn.Linear(din, len(layers) * dout).to(layers[0].weight.device)
|
| 88 |
+
with torch.no_grad():
|
| 89 |
+
new.weight[:] = torch.cat([x.weight * m for x,m in zip(layers, mults)])
|
| 90 |
+
new.bias[:] = torch.cat([torch.zeros_like(bias) if x.bias is None else x.bias * m for x, m in zip(layers, mults)])
|
| 91 |
+
return new
|
| 92 |
+
|
| 93 |
+
def convert_for_eval(self):
|
| 94 |
+
if self.qkv or self.kv: raise AttributeError("already converted")
|
| 95 |
+
|
| 96 |
+
self.odim = self.key.weight.shape[1]
|
| 97 |
+
if self.cross:
|
| 98 |
+
self.q = self.merge_linears([self.query], [self.sqrt_qk_scale])
|
| 99 |
+
self.kv = self.merge_linears([self.key, self.value],
|
| 100 |
+
[self.sqrt_qk_scale, 1])
|
| 101 |
+
else:
|
| 102 |
+
self.qkv = self.merge_linears([self.query, self.key, self.value],
|
| 103 |
+
[self.sqrt_qk_scale, self.sqrt_qk_scale, 1])
|
| 104 |
+
|
| 105 |
+
def split_heads(self, x, x_positions, rope=False, subsampling=1):
|
| 106 |
+
x = x.view(*x.shape[:2], self.n_head, -1)
|
| 107 |
+
if rope:
|
| 108 |
+
x = rope_rotate(x, x_positions * subsampling, *self.rotary(x))
|
| 109 |
+
return x.permute(0, 2, 1, 3)
|
| 110 |
+
|
| 111 |
+
def forward(
|
| 112 |
+
self,
|
| 113 |
+
qx,
|
| 114 |
+
q_positions,
|
| 115 |
+
kvx,
|
| 116 |
+
kv_positions,
|
| 117 |
+
causal = False,
|
| 118 |
+
mask=None,
|
| 119 |
+
):
|
| 120 |
+
if self.qkv:
|
| 121 |
+
q,k,v = self.qkv(qx).split(self.odim, dim=-1)
|
| 122 |
+
elif self.kv:
|
| 123 |
+
q = self.q(qx)
|
| 124 |
+
k,v = self.kv(kvx).split(self.odim, dim=-1)
|
| 125 |
+
else:
|
| 126 |
+
q,k,v = None,None,None
|
| 127 |
+
|
| 128 |
+
if q is None: q = self.query(qx) * self.sqrt_qk_scale
|
| 129 |
+
q = self.split_heads(q, q_positions, rope = self.rotary, subsampling = self.query_subsampling)
|
| 130 |
+
|
| 131 |
+
if kvx is not self.cached_kvx:
|
| 132 |
+
if k is None: k = self.key(kvx) * self.sqrt_qk_scale
|
| 133 |
+
k = self.split_heads(k, kv_positions, rope = self.rotary, subsampling = self.key_subsampling)
|
| 134 |
+
if v is None: v = self.value(kvx)
|
| 135 |
+
v = self.split_heads(v, kv_positions)
|
| 136 |
+
if self.k_cache is not None:
|
| 137 |
+
self.k_cache[:,:,kv_positions] = k
|
| 138 |
+
self.v_cache[:,:,kv_positions] = v
|
| 139 |
+
|
| 140 |
+
if self.k_cache is not None:
|
| 141 |
+
k, v = self.k_cache, self.v_cache
|
| 142 |
+
|
| 143 |
+
if mask is not None:
|
| 144 |
+
mask = mask[q_positions]
|
| 145 |
+
|
| 146 |
+
wv = F.scaled_dot_product_attention(q, k, v, attn_mask=mask, dropout_p=0, is_causal=causal)
|
| 147 |
+
|
| 148 |
+
return self.out(wv.permute(0, 2, 1, 3).flatten(start_dim=2))
|
| 149 |
+
|
| 150 |
+
# %% ../nbs/A. Neural modules.ipynb 6
|
| 151 |
+
# modified from https://blog.eleuther.ai/rotary-embeddings/
|
| 152 |
+
|
| 153 |
+
import torch
|
| 154 |
+
|
| 155 |
+
class Rotary(torch.nn.Module):
|
| 156 |
+
def __init__(self, dim, base=10000):
|
| 157 |
+
super().__init__()
|
| 158 |
+
inv_freq = 1.0 / (base ** (torch.arange(0, dim, 2).float() / dim))
|
| 159 |
+
self.register_buffer("inv_freq", inv_freq)
|
| 160 |
+
self.seq_len_cached = None
|
| 161 |
+
self.cos_cached = None
|
| 162 |
+
self.sin_cached = None
|
| 163 |
+
|
| 164 |
+
def forward(self, x, seq_dim=1):
|
| 165 |
+
seq_len = x.shape[seq_dim]
|
| 166 |
+
if not self.seq_len_cached or seq_len > self.seq_len_cached:
|
| 167 |
+
self.seq_len_cached = 2500
|
| 168 |
+
# self.seq_len_cached = seq_len
|
| 169 |
+
|
| 170 |
+
t = torch.arange(self.seq_len_cached, device=x.device).type_as(self.inv_freq)
|
| 171 |
+
freqs = torch.einsum("i,j->ij", t, self.inv_freq)
|
| 172 |
+
emb = torch.cat((freqs, freqs), dim=-1).to(x.device)
|
| 173 |
+
self.cos_cached = emb.cos()[None, :, None, :]
|
| 174 |
+
self.sin_cached = emb.sin()[None, :, None, :]
|
| 175 |
+
return self.cos_cached, self.sin_cached
|
| 176 |
+
|
| 177 |
+
|
| 178 |
+
# rotary pos emb helpers:
|
| 179 |
+
def rotate_half(x):
|
| 180 |
+
x1, x2 = x[..., : x.shape[-1] // 2], x[..., x.shape[-1] // 2 :]
|
| 181 |
+
return torch.cat(
|
| 182 |
+
(-x2, x1), dim=len(x.shape)-1
|
| 183 |
+
)
|
| 184 |
+
|
| 185 |
+
def rope_rotate(x, positions, cos, sin):
|
| 186 |
+
return x * cos[:,positions] + rotate_half(x) * sin[:,positions]
|
| 187 |
+
|
| 188 |
+
# %% ../nbs/A. Neural modules.ipynb 7
|
| 189 |
+
class ResidualAttentionBlock(nn.Module):
|
| 190 |
+
def __init__(self, n_state: int, n_head: int, cross_attention: bool = False, rope: bool = False,
|
| 191 |
+
qk_scale: float = 1, ffn_mult: int = 4):
|
| 192 |
+
super().__init__()
|
| 193 |
+
self.attn = MultiHeadAttention(n_state, n_head, qk_scale=qk_scale, rope=rope)
|
| 194 |
+
self.attn_ln = LayerNorm(n_state)
|
| 195 |
+
|
| 196 |
+
self.cross_attn = (
|
| 197 |
+
MultiHeadAttention(n_state, n_head, qk_scale=qk_scale, rope=rope, cross=True) if cross_attention else None
|
| 198 |
+
)
|
| 199 |
+
self.cross_attn_ln = LayerNorm(n_state) if cross_attention else None
|
| 200 |
+
|
| 201 |
+
n_mlp = n_state * ffn_mult
|
| 202 |
+
self.mlp = nn.Sequential(
|
| 203 |
+
nn.Linear(n_state, n_mlp), nn.GELU(), nn.Linear(n_mlp, n_state)
|
| 204 |
+
)
|
| 205 |
+
self.mlp_ln = LayerNorm(n_state)
|
| 206 |
+
|
| 207 |
+
def setup_kv_cache(self, max_batch_size, max_seq_len, max_cross_seq_len=None):
|
| 208 |
+
self.attn.setup_kv_cache(max_batch_size, max_seq_len)
|
| 209 |
+
if self.cross_attn:
|
| 210 |
+
self.cross_attn.setup_kv_cache(max_batch_size, max_cross_seq_len)
|
| 211 |
+
|
| 212 |
+
def forward(
|
| 213 |
+
self,
|
| 214 |
+
x: Tensor,
|
| 215 |
+
x_positions: Tensor = None,
|
| 216 |
+
xa: Optional[Tensor] = None,
|
| 217 |
+
xa_positions: Optional[Tensor] = None,
|
| 218 |
+
causal = False,
|
| 219 |
+
mask=None,
|
| 220 |
+
):
|
| 221 |
+
lnx = self.attn_ln(x)
|
| 222 |
+
x = x + self.attn(lnx, x_positions, lnx, x_positions, causal=causal, mask=mask)
|
| 223 |
+
if self.cross_attn:
|
| 224 |
+
lnx = self.cross_attn_ln(x)
|
| 225 |
+
x = x + self.cross_attn(lnx, x_positions, xa, xa_positions)
|
| 226 |
+
x = x + self.mlp(self.mlp_ln(x))
|
| 227 |
+
return x
|
| 228 |
+
|
| 229 |
+
# %% ../nbs/A. Neural modules.ipynb 8
|
| 230 |
+
class BaseDecoder(nn.Module):
|
| 231 |
+
def __init__(self, depth=6, n_head=6, width=384, qk_scale=1, ffn_mult=4, length=2250, rope=False):
|
| 232 |
+
super().__init__()
|
| 233 |
+
self.length = length
|
| 234 |
+
self.width = width
|
| 235 |
+
self.layers = nn.ModuleList([
|
| 236 |
+
ResidualAttentionBlock(
|
| 237 |
+
self.width, n_head, qk_scale=qk_scale, ffn_mult=ffn_mult, cross_attention=True, rope=rope
|
| 238 |
+
) for _ in range(math.floor(depth))
|
| 239 |
+
])
|
| 240 |
+
|
| 241 |
+
self.ln_post = LayerNorm(width)
|
| 242 |
+
|
| 243 |
+
mask = torch.empty(length, length).fill_(-torch.inf).triu_(1)
|
| 244 |
+
self.register_buffer("mask", mask, persistent=False)
|
| 245 |
+
|
| 246 |
+
def forward(self, x, x_positions, xenc, xenc_positions):
|
| 247 |
+
for i,l in enumerate(self.layers):
|
| 248 |
+
x = l(x, x_positions, xenc, xenc_positions, causal=False, mask=self.mask)
|
| 249 |
+
|
| 250 |
+
x = self.ln_post(x)
|
| 251 |
+
|
| 252 |
+
return x
|
| 253 |
+
|
| 254 |
+
# %% ../nbs/A. Neural modules.ipynb 9
|
| 255 |
+
class EmbeddingProjector(nn.Linear):
|
| 256 |
+
pass
|
| 257 |
+
|
| 258 |
+
class FlexEmbeddings(nn.Module):
|
| 259 |
+
def __init__(self, codes, width, special_codes=None, frozen_width=None, special_embedding=None, unembed=True):
|
| 260 |
+
super().__init__()
|
| 261 |
+
self.codes = codes
|
| 262 |
+
self.special_codes = special_codes
|
| 263 |
+
if frozen_width is None: frozen_width = width
|
| 264 |
+
|
| 265 |
+
self.main = nn.Embedding(codes, frozen_width or width)
|
| 266 |
+
self.emb_to_hidden = EmbeddingProjector(frozen_width, width) if frozen_width != width else None
|
| 267 |
+
self.hidden_to_emb = EmbeddingProjector(width, frozen_width) if unembed and frozen_width != width else None
|
| 268 |
+
if special_codes:
|
| 269 |
+
self.special = special_embedding or nn.Embedding(special_codes, width)
|
| 270 |
+
|
| 271 |
+
self.register_buffer('merged_in', None)
|
| 272 |
+
self.register_buffer('merged_out', None)
|
| 273 |
+
self.register_buffer('bias_out', None)
|
| 274 |
+
|
| 275 |
+
def set_frozen_embeddings(self, values):
|
| 276 |
+
with torch.no_grad():
|
| 277 |
+
self.main.weight[:] = values
|
| 278 |
+
self.main.lr_scale = 0
|
| 279 |
+
|
| 280 |
+
@torch.no_grad()
|
| 281 |
+
def convert_for_eval(self):
|
| 282 |
+
if not self.special_codes: return
|
| 283 |
+
# in
|
| 284 |
+
main_w = self.main.weight
|
| 285 |
+
if self.emb_to_hidden is not None: main_w = self.emb_to_hidden(main_w)
|
| 286 |
+
weight = torch.cat([main_w, self.special.weight], dim=0)
|
| 287 |
+
self.merged_in = nn.Embedding(*weight.shape, _weight=weight)
|
| 288 |
+
|
| 289 |
+
# out
|
| 290 |
+
weight = self.main.weight
|
| 291 |
+
if self.hidden_to_emb: weight = weight @ self.hidden_to_emb.weight
|
| 292 |
+
self.merged_out = torch.cat([weight.T, self.special.weight.T], dim=1).T.contiguous() # T is for F.linear
|
| 293 |
+
if self.hidden_to_emb:
|
| 294 |
+
self.bias_out = torch.cat([
|
| 295 |
+
self.hidden_to_emb.bias @ self.main.weight.T,
|
| 296 |
+
torch.zeros(self.special.weight.shape[0], device=weight.device, dtype=weight.dtype)
|
| 297 |
+
], dim=0)
|
| 298 |
+
else:
|
| 299 |
+
self.bias_out = None
|
| 300 |
+
|
| 301 |
+
def forward(self, toks):
|
| 302 |
+
if not self.training and self.merged_in is not None:
|
| 303 |
+
return self.merged_in(toks)
|
| 304 |
+
|
| 305 |
+
if self.special_codes:
|
| 306 |
+
special_mask = toks >= self.codes
|
| 307 |
+
embs = self.main(torch.where(special_mask, 0, toks))
|
| 308 |
+
else:
|
| 309 |
+
embs = self.main(toks)
|
| 310 |
+
|
| 311 |
+
if self.emb_to_hidden: embs = self.emb_to_hidden(embs)
|
| 312 |
+
|
| 313 |
+
if self.special_codes:
|
| 314 |
+
embs[special_mask] = self.special(toks[special_mask] - self.codes).to(embs.dtype)
|
| 315 |
+
|
| 316 |
+
return embs
|
| 317 |
+
|
| 318 |
+
def unembed(self, embs):
|
| 319 |
+
if not self.training and self.merged_out is not None:
|
| 320 |
+
return F.linear(embs, self.merged_out, self.bias_out) # embs @ self.merged_out + self.bias_out
|
| 321 |
+
|
| 322 |
+
orig_embs = embs
|
| 323 |
+
if self.hidden_to_emb: embs = self.hidden_to_emb(embs)
|
| 324 |
+
|
| 325 |
+
main_logits = (embs @ self.main.weight.to(embs.dtype).T).float()
|
| 326 |
+
|
| 327 |
+
if not self.special_codes:
|
| 328 |
+
return main_logits
|
| 329 |
+
|
| 330 |
+
special_logits = (orig_embs @ self.special.weight.to(orig_embs.dtype).T).float()
|
| 331 |
+
return torch.cat([main_logits, special_logits], dim=-1)
|
whisperspeech/pipeline.py
ADDED
|
@@ -0,0 +1,93 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/7. Pipeline.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = ['Pipeline']
|
| 5 |
+
|
| 6 |
+
# %% ../nbs/7. Pipeline.ipynb 1
|
| 7 |
+
import torch
|
| 8 |
+
from whisperspeech.t2s_up_wds_mlang_enclm import TSARTransformer
|
| 9 |
+
from whisperspeech.s2a_delar_mup_wds_mlang import SADelARTransformer
|
| 10 |
+
from whisperspeech.a2wav import Vocoder
|
| 11 |
+
import traceback
|
| 12 |
+
from pathlib import Path
|
| 13 |
+
|
| 14 |
+
# %% ../nbs/7. Pipeline.ipynb 2
|
| 15 |
+
class Pipeline:
|
| 16 |
+
default_speaker = torch.tensor(
|
| 17 |
+
[-0.2929, -0.4503, 0.4155, -0.1417, 0.0473, -0.1624, -0.2322, 0.7071,
|
| 18 |
+
0.4800, 0.5496, 0.0410, 0.6236, 0.4729, 0.0587, 0.2194, -0.0466,
|
| 19 |
+
-0.3036, 0.0497, 0.5028, -0.1703, 0.5039, -0.6464, 0.3857, -0.7350,
|
| 20 |
+
-0.1605, 0.4808, 0.5397, -0.4851, 0.1774, -0.8712, 0.5789, 0.1785,
|
| 21 |
+
-0.1417, 0.3039, 0.4232, -0.0186, 0.2685, 0.6153, -0.3103, -0.5706,
|
| 22 |
+
-0.4494, 0.3394, -0.6184, -0.3617, 1.1041, -0.1178, -0.1885, 0.1997,
|
| 23 |
+
0.5571, -0.2906, -0.0477, -0.4048, -0.1062, 1.4779, 0.1639, -0.3712,
|
| 24 |
+
-0.1776, -0.0568, -0.6162, 0.0110, -0.0207, -0.1319, -0.3854, 0.7248,
|
| 25 |
+
0.0343, 0.5724, 0.0670, 0.0486, -0.3813, 0.1738, 0.3017, 1.0502,
|
| 26 |
+
0.1550, 0.5708, 0.0366, 0.5093, 0.0294, -0.7091, -0.8220, -0.1583,
|
| 27 |
+
-0.2343, 0.1366, 0.7372, -0.0631, 0.1505, 0.4600, -0.1252, -0.5245,
|
| 28 |
+
0.7523, -0.0386, -0.2587, 1.0066, -0.2037, 0.1617, -0.3800, 0.2790,
|
| 29 |
+
0.0184, -0.5111, -0.7291, 0.1627, 0.2367, -0.0192, 0.4822, -0.4458,
|
| 30 |
+
0.1457, -0.5884, 0.1909, 0.2563, -0.2035, -0.0377, 0.7771, 0.2139,
|
| 31 |
+
0.3801, 0.6047, -0.6043, -0.2563, -0.0726, 0.3856, 0.3217, 0.0823,
|
| 32 |
+
-0.1302, 0.3287, 0.5693, 0.2453, 0.8231, 0.0072, 1.0327, 0.6065,
|
| 33 |
+
-0.0620, -0.5572, 0.5220, 0.2485, 0.1520, 0.0222, -0.2179, -0.7392,
|
| 34 |
+
-0.3855, 0.1822, 0.1042, 0.7133, 0.3583, 0.0606, -0.0424, -0.9189,
|
| 35 |
+
-0.4882, -0.5480, -0.5719, -0.1660, -0.3439, -0.5814, -0.2542, 0.0197,
|
| 36 |
+
0.4942, 0.0915, -0.0420, -0.0035, 0.5578, 0.1051, -0.0891, 0.2348,
|
| 37 |
+
0.6876, -0.6685, 0.8215, -0.3692, -0.3150, -0.0462, -0.6806, -0.2661,
|
| 38 |
+
-0.0308, -0.0050, 0.6756, -0.1647, 1.0734, 0.0049, 0.4969, 0.0259,
|
| 39 |
+
-0.8949, 0.0731, 0.0886, 0.3442, -0.1433, -0.6804, 0.2204, 0.1859,
|
| 40 |
+
0.2702, 0.1699, -0.1443, -0.9614, 0.3261, 0.1718, 0.3545, -0.0686]
|
| 41 |
+
)
|
| 42 |
+
|
| 43 |
+
def __init__(self, t2s_ref=None, s2a_ref=None, optimize=True, torch_compile=False):
|
| 44 |
+
args = dict()
|
| 45 |
+
try:
|
| 46 |
+
if t2s_ref:
|
| 47 |
+
args["ref"] = t2s_ref
|
| 48 |
+
self.t2s = TSARTransformer.load_model(**args).cuda()
|
| 49 |
+
if optimize: self.t2s.optimize(torch_compile=torch_compile)
|
| 50 |
+
except:
|
| 51 |
+
print("Failed to load the T2S model:")
|
| 52 |
+
print(traceback.format_exc())
|
| 53 |
+
try:
|
| 54 |
+
if s2a_ref:
|
| 55 |
+
args["ref"] = s2a_ref
|
| 56 |
+
self.s2a = SADelARTransformer.load_model(**args).cuda()
|
| 57 |
+
if optimize: self.s2a.optimize(torch_compile=torch_compile)
|
| 58 |
+
except:
|
| 59 |
+
print("Failed to load the S2A model:")
|
| 60 |
+
print(traceback.format_exc())
|
| 61 |
+
self.vocoder = Vocoder()
|
| 62 |
+
self.encoder = None
|
| 63 |
+
|
| 64 |
+
def extract_spk_emb(self, fname):
|
| 65 |
+
"""Extracts a speaker embedding from the first 30 seconds of the give audio file.
|
| 66 |
+
"""
|
| 67 |
+
import torchaudio
|
| 68 |
+
if self.encoder is None:
|
| 69 |
+
from speechbrain.pretrained import EncoderClassifier
|
| 70 |
+
self.encoder = EncoderClassifier.from_hparams("speechbrain/spkrec-ecapa-voxceleb",
|
| 71 |
+
savedir="~/.cache/speechbrain/",
|
| 72 |
+
run_opts={"device": "cuda"})
|
| 73 |
+
samples, sr = torchaudio.load(fname)
|
| 74 |
+
samples = self.encoder.audio_normalizer(samples[0,:30*sr], sr)
|
| 75 |
+
spk_emb = self.encoder.encode_batch(samples)
|
| 76 |
+
return spk_emb[0,0]
|
| 77 |
+
|
| 78 |
+
def generate_atoks(self, text, speaker=None, lang='en', cps=15, step_callback=None):
|
| 79 |
+
if speaker is None: speaker = self.default_speaker
|
| 80 |
+
elif isinstance(speaker, (str, Path)): speaker = self.extract_spk_emb(speaker)
|
| 81 |
+
text = text.replace("\n", " ")
|
| 82 |
+
stoks = self.t2s.generate(text, cps=cps, lang=lang, step=step_callback)
|
| 83 |
+
atoks = self.s2a.generate(stoks, speaker.unsqueeze(0), step=step_callback)
|
| 84 |
+
return atoks
|
| 85 |
+
|
| 86 |
+
def generate(self, text, speaker=None, lang='en', cps=15, step_callback=None):
|
| 87 |
+
return self.vocoder.decode(self.generate_atoks(text, speaker, lang=lang, cps=cps, step_callback=step_callback))
|
| 88 |
+
|
| 89 |
+
def generate_to_file(self, fname, text, speaker=None, lang='en', cps=15, step_callback=None):
|
| 90 |
+
self.vocoder.decode_to_file(fname, self.generate_atoks(text, speaker, lang=lang, cps=cps, step_callback=None))
|
| 91 |
+
|
| 92 |
+
def generate_to_notebook(self, text, speaker=None, lang='en', cps=15, step_callback=None):
|
| 93 |
+
self.vocoder.decode_to_notebook(self.generate_atoks(text, speaker, lang=lang, cps=cps, step_callback=None))
|
whisperspeech/prepare_s2a_dataset.py
ADDED
|
@@ -0,0 +1,112 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/4A. S2A dataset preparation.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = ['flac_to_s2a_name']
|
| 5 |
+
|
| 6 |
+
# %% ../nbs/4A. S2A dataset preparation.ipynb 2
|
| 7 |
+
import sys
|
| 8 |
+
import os
|
| 9 |
+
import itertools
|
| 10 |
+
from pathlib import Path
|
| 11 |
+
|
| 12 |
+
import numpy as np
|
| 13 |
+
import torch
|
| 14 |
+
import torchaudio
|
| 15 |
+
import torch.nn.functional as F
|
| 16 |
+
from torch.profiler import profile, record_function, ProfilerActivity
|
| 17 |
+
|
| 18 |
+
from fastprogress import progress_bar
|
| 19 |
+
from fastcore.script import *
|
| 20 |
+
|
| 21 |
+
import whisper
|
| 22 |
+
from . import vad, wh_transcribe, vq_stoks, extract_acoustic
|
| 23 |
+
import webdataset as wds
|
| 24 |
+
|
| 25 |
+
# %% ../nbs/4A. S2A dataset preparation.ipynb 4
|
| 26 |
+
def flac_to_s2a_name(input):
|
| 27 |
+
if '-flac-' in input:
|
| 28 |
+
return input.rsplit("/", 1)[1].replace('flac', 's2a') + ".gz"
|
| 29 |
+
else:
|
| 30 |
+
return input.rsplit("/", 1)[1].replace('raw', 's2a') + ".gz"
|
| 31 |
+
|
| 32 |
+
# %% ../nbs/4A. S2A dataset preparation.ipynb 6
|
| 33 |
+
def resampler(newsr = 24000, key = 'samples_24k'):
|
| 34 |
+
_last_sr = None
|
| 35 |
+
tform = None
|
| 36 |
+
|
| 37 |
+
def _resample(samples):
|
| 38 |
+
for s in samples:
|
| 39 |
+
sr = s['sample_rate']
|
| 40 |
+
if sr != newsr:
|
| 41 |
+
if sr != _last_sr: tform = torchaudio.transforms.Resample(sr, newsr)
|
| 42 |
+
s[key] = tform(s['samples'])
|
| 43 |
+
else:
|
| 44 |
+
s[key] = s['samples']
|
| 45 |
+
yield s
|
| 46 |
+
|
| 47 |
+
return _resample
|
| 48 |
+
|
| 49 |
+
# %% ../nbs/4A. S2A dataset preparation.ipynb 9
|
| 50 |
+
@call_parse
|
| 51 |
+
def prepare_s2a(
|
| 52 |
+
input:str, # FLAC webdataset file path (or - to read the names from stdin)
|
| 53 |
+
proc_dataset_path:Path, # processed VAD files path
|
| 54 |
+
output:str=None, # output file name
|
| 55 |
+
vq_model:str="collabora/spear-tts-pytorch:whisper-vq-stoks.model", # the model path (use repo_id:filename to download it from hugginface)
|
| 56 |
+
n_samples:int=None, # process a limited amount of samples
|
| 57 |
+
batch_size:int=1, # process several segments at once
|
| 58 |
+
fix_dots:bool=False, # fix dots in file names
|
| 59 |
+
):
|
| 60 |
+
if ":" in vq_model:
|
| 61 |
+
repo, fname = vq_model.split(":", 1)
|
| 62 |
+
vq_model = vq_stoks.RQBottleneckTransformer.load_model(repo, fname).cuda()
|
| 63 |
+
else:
|
| 64 |
+
vq_model = vq_stoks.RQBottleneckTransformer.load_model(local_filename=vq_model).cuda()
|
| 65 |
+
amodel = extract_acoustic.load_model()
|
| 66 |
+
amodel.set_target_bandwidth(3)
|
| 67 |
+
|
| 68 |
+
if input == "-":
|
| 69 |
+
input = [f.strip() for f in sys.stdin.readlines()]
|
| 70 |
+
assert output, "please provide the output shard name"
|
| 71 |
+
else:
|
| 72 |
+
if output is None: output = flac_to_s2a_name(input)
|
| 73 |
+
input = [input]
|
| 74 |
+
|
| 75 |
+
total = n_samples//batch_size if n_samples else 'noinfer'
|
| 76 |
+
|
| 77 |
+
ds = wds.WebDataset(input, shardshuffle=True, rename_files=vad.fix_dots_in_names if fix_dots else None).compose(
|
| 78 |
+
wds.decode(wds.torch_audio),
|
| 79 |
+
wds.select(lambda x: 'wav' in x or 'flac' in x),
|
| 80 |
+
vq_stoks.merge_in(vq_stoks.derived_dataset(proc_dataset_path, 'vad')),
|
| 81 |
+
wds.map_dict(**{"vad.npy":wh_transcribe.chunk_merger}),
|
| 82 |
+
lambda x: wh_transcribe.split_to_chunks(x),
|
| 83 |
+
resampler(),
|
| 84 |
+
resampler(16000, 'samples_16k'),
|
| 85 |
+
wds.to_tuple('__key__', 'rpad_s', 'samples_16k', 'samples_24k'),
|
| 86 |
+
wds.batched(64),
|
| 87 |
+
)
|
| 88 |
+
|
| 89 |
+
dl = wds.WebLoader(ds, num_workers=4, batch_size=None).unbatched().shuffle(2000).batched(batch_size)
|
| 90 |
+
|
| 91 |
+
speakers = set()
|
| 92 |
+
tmp = output+".tmp"
|
| 93 |
+
with wds.TarWriter(tmp) as sink:
|
| 94 |
+
for keys, rpad_ss, samples, samples24k in progress_bar(dl, total=total):
|
| 95 |
+
with record_function('to_cuda'):
|
| 96 |
+
samples, samples24k = samples.cuda(), samples24k.unsqueeze(1).cuda()
|
| 97 |
+
with record_function('encodec'):
|
| 98 |
+
atoks = amodel.encode(samples24k)[0][0]
|
| 99 |
+
with record_function('vq_stoks'):
|
| 100 |
+
stoks = vq_model.encode_audio(samples)
|
| 101 |
+
with record_function('from_cuda'):
|
| 102 |
+
atoks, stoks = atoks.cpu().numpy().astype(np.int16), stoks.cpu().numpy().astype(np.int16)
|
| 103 |
+
for key, rpad_s, _atoks, _stoks in zip(keys, rpad_ss, atoks, stoks):
|
| 104 |
+
speakers.add(key.split('/')[1])
|
| 105 |
+
sink.write({
|
| 106 |
+
"__key__": key,
|
| 107 |
+
"atoks.npy": _atoks[:,:int(-rpad_s * 75)],
|
| 108 |
+
"stoks.npy": _stoks[:int(-rpad_s * 25)],
|
| 109 |
+
})
|
| 110 |
+
with open(output+".speakers.txt", "w") as f: f.write("\n".join(speakers))
|
| 111 |
+
if not n_samples:
|
| 112 |
+
os.rename(tmp, output)
|
whisperspeech/prepare_t2s_dataset.py
ADDED
|
@@ -0,0 +1,111 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/5A. T2S dataset preparation.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = []
|
| 5 |
+
|
| 6 |
+
# %% ../nbs/5A. T2S dataset preparation.ipynb 2
|
| 7 |
+
import sys
|
| 8 |
+
import os
|
| 9 |
+
import itertools
|
| 10 |
+
from pathlib import Path
|
| 11 |
+
|
| 12 |
+
import numpy as np
|
| 13 |
+
import torch
|
| 14 |
+
import torchaudio
|
| 15 |
+
import torch.nn.functional as F
|
| 16 |
+
from torch.profiler import profile, record_function, ProfilerActivity
|
| 17 |
+
|
| 18 |
+
from fastprogress import progress_bar
|
| 19 |
+
from fastcore.script import *
|
| 20 |
+
|
| 21 |
+
import whisper, whisperx
|
| 22 |
+
from . import vad, wh_transcribe, vq_stoks, extract_acoustic
|
| 23 |
+
import webdataset as wds
|
| 24 |
+
|
| 25 |
+
# %% ../nbs/5A. T2S dataset preparation.ipynb 4
|
| 26 |
+
def flac_to_t2s_name(input):
|
| 27 |
+
return input.rsplit("/", 1)[1].replace('flac', 't2s') + ".gz"
|
| 28 |
+
|
| 29 |
+
# %% ../nbs/5A. T2S dataset preparation.ipynb 6
|
| 30 |
+
class Transcriber:
|
| 31 |
+
"""
|
| 32 |
+
A helper class to transcribe a batch of 30 second audio chunks.
|
| 33 |
+
"""
|
| 34 |
+
def __init__(self, model_size, lang=False):
|
| 35 |
+
self.model = whisperx.asr.load_model(model_size, "cuda", compute_type="float16", language=lang)
|
| 36 |
+
# without calling vad_model at least once the rest segfaults for some reason...
|
| 37 |
+
self.model.vad_model({"waveform": torch.zeros(1, 16000), "sample_rate": 16000})
|
| 38 |
+
|
| 39 |
+
def transcribe(self, batch):
|
| 40 |
+
batch = whisper.log_mel_spectrogram(batch)
|
| 41 |
+
embs = self.model.model.encode(batch.cpu().numpy())
|
| 42 |
+
return self.model.tokenizer.tokenizer.decode_batch([x.sequences_ids[0] for x in
|
| 43 |
+
self.model.model.model.generate(
|
| 44 |
+
embs,
|
| 45 |
+
[self.model.model.get_prompt(self.model.tokenizer, [], without_timestamps=True)]*len(batch),
|
| 46 |
+
)])
|
| 47 |
+
|
| 48 |
+
# %% ../nbs/5A. T2S dataset preparation.ipynb 7
|
| 49 |
+
@call_parse
|
| 50 |
+
def prepare_t2s(
|
| 51 |
+
input:str, # FLAC webdataset file path (or - to read the names from stdin)
|
| 52 |
+
proc_dataset_path:Path, # processed VAD files path
|
| 53 |
+
output:str=None, # output file name
|
| 54 |
+
vq_model:str="collabora/spear-tts-pytorch:whisper-vq-stoks.model", # the model path (use repo_id:filename to download it from hugginface)
|
| 55 |
+
n_samples:int=None, # process a limited amount of samples
|
| 56 |
+
batch_size:int=1, # process several segments at once
|
| 57 |
+
transcription_model:str="small.en",
|
| 58 |
+
):
|
| 59 |
+
if ":" in vq_model:
|
| 60 |
+
repo, fname = vq_model.split(":", 1)
|
| 61 |
+
vq_model = vq_stoks.RQBottleneckTransformer.load_model(repo, fname).cuda()
|
| 62 |
+
else:
|
| 63 |
+
vq_model = vq_stoks.RQBottleneckTransformer.load_model(local_filename=vq_model).cuda()
|
| 64 |
+
transcriber = Transcriber(transcription_model)
|
| 65 |
+
|
| 66 |
+
if input == "-":
|
| 67 |
+
input = [f.strip() for f in sys.stdin.readlines()]
|
| 68 |
+
assert output, "please provide the output shard name"
|
| 69 |
+
else:
|
| 70 |
+
if output is None: output = flac_to_t2s_name(input)
|
| 71 |
+
input = [input]
|
| 72 |
+
|
| 73 |
+
total = n_samples//batch_size if n_samples else 'noinfer'
|
| 74 |
+
if n_samples: print(f"Benchmarking run of {n_samples} samples ({total} batches)")
|
| 75 |
+
|
| 76 |
+
ds = wds.WebDataset(input, shardshuffle=True, rename_files=vad.fix_dots_in_names).compose(
|
| 77 |
+
wds.decode(wds.torch_audio),
|
| 78 |
+
vq_stoks.merge_in(vq_stoks.derived_dataset(proc_dataset_path, 'vad')),
|
| 79 |
+
wds.map_dict(**{"vad.npy": lambda s: wh_transcribe.chunk_merger(s, wh_transcribe.random_cutter)}),
|
| 80 |
+
lambda x: wh_transcribe.split_to_chunks(x),
|
| 81 |
+
# drop the first and last segment because they tend to be inaccurate
|
| 82 |
+
# (the transcriptions don't have the "LibriVox" header and "end of chapter" suffix)
|
| 83 |
+
wds.select(lambda x: x['i'] != 0 and x['i'] != x['imax']),
|
| 84 |
+
wds.to_tuple('__key__', 'rpad', 'samples'),
|
| 85 |
+
wds.batched(64),
|
| 86 |
+
)
|
| 87 |
+
|
| 88 |
+
dl = wds.WebLoader(ds, num_workers=4, batch_size=None).unbatched().shuffle(2000).batched(batch_size)
|
| 89 |
+
|
| 90 |
+
speakers = set()
|
| 91 |
+
tmp = output+".tmp"
|
| 92 |
+
with wds.TarWriter(tmp) as sink:
|
| 93 |
+
for keys, rpads, samples in progress_bar(dl, total=total):
|
| 94 |
+
with record_function('to_cuda'):
|
| 95 |
+
csamples = samples.cuda()
|
| 96 |
+
with record_function('transcribe'):
|
| 97 |
+
txts = transcriber.transcribe(csamples)
|
| 98 |
+
with record_function('vq_stoks'):
|
| 99 |
+
stoks = vq_model.encode_audio(csamples)
|
| 100 |
+
with record_function('from_cuda'):
|
| 101 |
+
stoks = stoks.cpu().numpy().astype(np.int16)
|
| 102 |
+
for key, rpad, txt, _stoks in zip(keys, rpads, txts, stoks):
|
| 103 |
+
speakers.add(key.split('/')[1])
|
| 104 |
+
sink.write({
|
| 105 |
+
"__key__": key,
|
| 106 |
+
"txt": txt,
|
| 107 |
+
"stoks.npy": _stoks[:int(-rpad/16000 * 25)],
|
| 108 |
+
})
|
| 109 |
+
with open(output+".speakers.txt", "w") as f: f.write("\n".join(speakers))
|
| 110 |
+
if not n_samples:
|
| 111 |
+
os.rename(tmp, output)
|
whisperspeech/s2a_delar_mup_wds.py
ADDED
|
@@ -0,0 +1,688 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/4B. Semantic to acoustic token modeling.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = ['load_datasets', 'CMLMVisual', 'Rotary', 'rotate_half', 'apply_rotary_pos_emb', 'ResidualAttentionBlock',
|
| 5 |
+
'MultiHeadAttention', 'DelSumDecoder', 'EmbeddingProjector', 'rand', 'Tunables', 'SADelARTransformer']
|
| 6 |
+
|
| 7 |
+
# %% ../nbs/4B. Semantic to acoustic token modeling.ipynb 1
|
| 8 |
+
import io
|
| 9 |
+
import time
|
| 10 |
+
import math
|
| 11 |
+
import random
|
| 12 |
+
import dataclasses
|
| 13 |
+
|
| 14 |
+
# %% ../nbs/4B. Semantic to acoustic token modeling.ipynb 2
|
| 15 |
+
import torch
|
| 16 |
+
import torch.nn as nn
|
| 17 |
+
import torch.nn.functional as F
|
| 18 |
+
from torch.profiler import profile, record_function, ProfilerActivity, schedule
|
| 19 |
+
from fastcore.basics import store_attr
|
| 20 |
+
from huggingface_hub import hf_hub_download
|
| 21 |
+
|
| 22 |
+
# %% ../nbs/4B. Semantic to acoustic token modeling.ipynb 3
|
| 23 |
+
from pathlib import Path
|
| 24 |
+
import json
|
| 25 |
+
from fastprogress import progress_bar, master_bar
|
| 26 |
+
import webdataset as wds
|
| 27 |
+
|
| 28 |
+
# %% ../nbs/4B. Semantic to acoustic token modeling.ipynb 4
|
| 29 |
+
from .train import *
|
| 30 |
+
from .modules import *
|
| 31 |
+
from . import vq_stoks
|
| 32 |
+
|
| 33 |
+
# %% ../nbs/4B. Semantic to acoustic token modeling.ipynb 8
|
| 34 |
+
def rand(start, end):
|
| 35 |
+
return random.random() * (end - start) + start
|
| 36 |
+
|
| 37 |
+
# %% ../nbs/4B. Semantic to acoustic token modeling.ipynb 9
|
| 38 |
+
def random_trunc(random_trunc_p, atoks_len = 2250, stoks_len = 750):
|
| 39 |
+
atoks_per_second = atoks_len / 30
|
| 40 |
+
def _trunc(samples):
|
| 41 |
+
for s in samples:
|
| 42 |
+
if random.random() < random_trunc_p:
|
| 43 |
+
seconds = rand(0.3, 30)
|
| 44 |
+
s['atoks.npy'] = s['atoks.npy'][:,:math.ceil(seconds * atoks_per_second)]
|
| 45 |
+
s['stoks.npy'] = s['stoks.npy'][:math.ceil(s['atoks.npy'].shape[-1]/atoks_len*stoks_len)]
|
| 46 |
+
yield s
|
| 47 |
+
return _trunc
|
| 48 |
+
|
| 49 |
+
def pad_samples(atoks_len = 2250, stoks_len = 750, stoks_pad_token = 4096):
|
| 50 |
+
def _pad(samples):
|
| 51 |
+
for s in samples:
|
| 52 |
+
s['stoks.npy'] = F.pad(torch.tensor(s['stoks.npy']), (0, stoks_len - s['stoks.npy'].shape[-1]), value=stoks_pad_token)
|
| 53 |
+
s['atoks.npy'] = F.pad(torch.tensor(s['atoks.npy']), (0, atoks_len - s['atoks.npy'].shape[-1]), value=-100)
|
| 54 |
+
yield s
|
| 55 |
+
return _pad
|
| 56 |
+
|
| 57 |
+
# %% ../nbs/4B. Semantic to acoustic token modeling.ipynb 10
|
| 58 |
+
def speaker_id_extractor(speaker_map):
|
| 59 |
+
def _extractor(samples):
|
| 60 |
+
for s in samples:
|
| 61 |
+
s['speaker'] = torch.tensor(speaker_map[s['__key__'].split("/")[1]])
|
| 62 |
+
yield s
|
| 63 |
+
return _extractor
|
| 64 |
+
|
| 65 |
+
# %% ../nbs/4B. Semantic to acoustic token modeling.ipynb 14
|
| 66 |
+
def load_datasets(
|
| 67 |
+
input:str, # webdataset folder
|
| 68 |
+
samples:int, # samples per epoch
|
| 69 |
+
subsample:float=1, # use a fraction of the files
|
| 70 |
+
val_samples:int=512,
|
| 71 |
+
random_trunc_p:float=0,# probability of truncating the input to less than 30 seconds
|
| 72 |
+
stoks_pad_token=4096,
|
| 73 |
+
):
|
| 74 |
+
|
| 75 |
+
if isinstance(input, (Path, str)):
|
| 76 |
+
path = Path(input)
|
| 77 |
+
if path.is_dir():
|
| 78 |
+
glob = '*-s2a-*.tar.gz'
|
| 79 |
+
else:
|
| 80 |
+
glob = path.name
|
| 81 |
+
path = path.parent
|
| 82 |
+
input = Path(path).glob(glob)
|
| 83 |
+
elif isinstance(input, list):
|
| 84 |
+
pass
|
| 85 |
+
else:
|
| 86 |
+
raise ArgumentError("input should be either a list or a path with an optional glob specifier")
|
| 87 |
+
shards = [str(x) for x in input]
|
| 88 |
+
|
| 89 |
+
speakers = set()
|
| 90 |
+
for shard in shards:
|
| 91 |
+
with open(shard+'.speakers.txt') as f: speakers = speakers.union(set(x.strip() for x in f.readlines()))
|
| 92 |
+
speakers = {id:i for i,id in enumerate(sorted(speakers))}
|
| 93 |
+
|
| 94 |
+
def ds(shards, length):
|
| 95 |
+
ds = wds.WebDataset(wds.ResampledShards(shards)).compose(
|
| 96 |
+
wds.decode(),
|
| 97 |
+
speaker_id_extractor(speakers),
|
| 98 |
+
random_trunc(random_trunc_p) if random_trunc_p > 0 else lambda x: x,
|
| 99 |
+
pad_samples(stoks_pad_token=stoks_pad_token),
|
| 100 |
+
wds.to_tuple('stoks.npy', 'atoks.npy', 'speaker'),
|
| 101 |
+
wds.batched(64),
|
| 102 |
+
)
|
| 103 |
+
ds.speakers = speakers
|
| 104 |
+
ds.total_samples = length
|
| 105 |
+
return ds.compose(wds.slice(length // 64)).with_epoch(length // 64).with_length(length // 64)
|
| 106 |
+
|
| 107 |
+
return (
|
| 108 |
+
ds(shards[1:], samples),
|
| 109 |
+
ds(shards[:1], val_samples),
|
| 110 |
+
)
|
| 111 |
+
|
| 112 |
+
# %% ../nbs/4B. Semantic to acoustic token modeling.ipynb 33
|
| 113 |
+
import pylab as plt
|
| 114 |
+
import fastprogress
|
| 115 |
+
import IPython
|
| 116 |
+
import numpy as np
|
| 117 |
+
|
| 118 |
+
class CMLMVisual:
|
| 119 |
+
"""Visualize training progress"""
|
| 120 |
+
def __init__ (self, model, masterbar, total_steps):
|
| 121 |
+
self.model = model
|
| 122 |
+
self.masterbar = masterbar
|
| 123 |
+
self.total_steps = total_steps
|
| 124 |
+
self.epochs = total_steps // masterbar.main_bar.total
|
| 125 |
+
|
| 126 |
+
gs = plt.GridSpec(3, 1, height_ratios=[2,2,1])
|
| 127 |
+
graph_fig = plt.figure(figsize=(10,6))
|
| 128 |
+
self.graph_fig = graph_fig
|
| 129 |
+
self.loss_p = graph_fig.add_subplot(gs[0])
|
| 130 |
+
self.acc_p = graph_fig.add_subplot(gs[1], sharex=self.loss_p)
|
| 131 |
+
self.acc_p.tick_params('x', labelbottom=False)
|
| 132 |
+
self.lr_p = graph_fig.add_subplot(gs[2], sharex=self.loss_p)
|
| 133 |
+
self.lr_p.tick_params('x', labelbottom=False)
|
| 134 |
+
self.graph_out = None
|
| 135 |
+
|
| 136 |
+
self.its = []
|
| 137 |
+
self.train_losses = []
|
| 138 |
+
self.val_losses = []
|
| 139 |
+
self.lr_history = []
|
| 140 |
+
self.acc = np.nan
|
| 141 |
+
self.acc_history = []
|
| 142 |
+
self.pacc_history = []
|
| 143 |
+
|
| 144 |
+
def show(self):
|
| 145 |
+
self.start_t = time.time()
|
| 146 |
+
self.masterbar.write(["samples", "train", "val", "time"], table=True)
|
| 147 |
+
self.graph_out = display(self.graph_fig, display_id=True)
|
| 148 |
+
self.acc_out = display(IPython.display.HTML(''), display_id=True)
|
| 149 |
+
|
| 150 |
+
def hide(self):
|
| 151 |
+
if self.graph_out is not None:
|
| 152 |
+
self.graph_out.update(IPython.display.HTML(''))
|
| 153 |
+
|
| 154 |
+
def plot(self):
|
| 155 |
+
loss_p, acc_p, lr_p = self.loss_p, self.acc_p, self.lr_p
|
| 156 |
+
loss_p.clear()
|
| 157 |
+
loss_p.plot(self.its, self.train_losses)
|
| 158 |
+
loss_p.plot(self.its, self.val_losses)
|
| 159 |
+
loss_p.set_xlim(0, self.total_steps)
|
| 160 |
+
loss_p.set_yscale('log')
|
| 161 |
+
acc_p.clear()
|
| 162 |
+
for k in self.acc_history[-1].keys():
|
| 163 |
+
acc_p.plot(self.its, [x[k] for x in self.acc_history], ':')
|
| 164 |
+
# acc_p.plot(self.its, np.stack(self.pacc_history), label=range(len(self.pacc_history[0])))
|
| 165 |
+
lr_p.clear()
|
| 166 |
+
lrs = np.array(self.lr_history)
|
| 167 |
+
lr_p.plot(self.its, lrs)
|
| 168 |
+
self.graph_out.update(self.graph_fig)
|
| 169 |
+
|
| 170 |
+
def add_data(self, it, lr, train_loss, val_los):
|
| 171 |
+
self.its.append(it)
|
| 172 |
+
self.train_losses.append(train_loss)
|
| 173 |
+
self.val_losses.append(val_los)
|
| 174 |
+
self.lr_history.append(lr)
|
| 175 |
+
metrics = self.model.get_metrics()
|
| 176 |
+
self.acc_history.append(metrics)
|
| 177 |
+
# self.acc_out.update(f"Accuracy: {self.entropy_history[-1]:.2f}")
|
| 178 |
+
# self.pacc_history.append((self.model.pval_true / self.model.pval_total).cpu().numpy())
|
| 179 |
+
# if self.acc_history:
|
| 180 |
+
html = "<h5>Accuracies:</h5><table>"
|
| 181 |
+
html += "<thead>"+(''.join([f"<td>{k}<td>" for k,x in metrics.items()]))+"</thead>"
|
| 182 |
+
html += "<tr>"+(''.join([f"<td>{x*100:.1f}%<td>" for k,x in metrics.items()]))+"</tr>"
|
| 183 |
+
html += "</table>"
|
| 184 |
+
self.acc_out.update(IPython.display.HTML(html))
|
| 185 |
+
self.plot()
|
| 186 |
+
|
| 187 |
+
def add_table_row(self, it, avg_train_loss, val_loss):
|
| 188 |
+
elapsed_t = time.time() - self.start_t
|
| 189 |
+
self.masterbar.write([it, f"{avg_train_loss:.5f}", f"{val_loss:.5f}", fastprogress.core.format_time(elapsed_t)], table=True)
|
| 190 |
+
|
| 191 |
+
def on_iter(self, bar, it, avg_train_loss, val_loss):
|
| 192 |
+
epoch = math.ceil(it / self.total_steps * self.epochs)
|
| 193 |
+
bar.comment = f"#{epoch}/{self.epochs} loss: {avg_train_loss:.3f} / {val_loss:.3f}"
|
| 194 |
+
|
| 195 |
+
# %% ../nbs/4B. Semantic to acoustic token modeling.ipynb 34
|
| 196 |
+
# modified from https://blog.eleuther.ai/rotary-embeddings/
|
| 197 |
+
import torch
|
| 198 |
+
|
| 199 |
+
class Rotary(torch.nn.Module):
|
| 200 |
+
def __init__(self, dim, base=10000):
|
| 201 |
+
super().__init__()
|
| 202 |
+
inv_freq = 1.0 / (base ** (torch.arange(0, dim, 2).float() / dim))
|
| 203 |
+
self.register_buffer("inv_freq", inv_freq)
|
| 204 |
+
self.seq_len_cached = None
|
| 205 |
+
self.cos_cached = None
|
| 206 |
+
self.sin_cached = None
|
| 207 |
+
|
| 208 |
+
def forward(self, x, seq_dim=1):
|
| 209 |
+
seq_len = x.shape[seq_dim]
|
| 210 |
+
if seq_len != self.seq_len_cached:
|
| 211 |
+
self.seq_len_cached = seq_len
|
| 212 |
+
t = torch.arange(x.shape[seq_dim], device=x.device).type_as(self.inv_freq)
|
| 213 |
+
freqs = torch.einsum("i,j->ij", t, self.inv_freq)
|
| 214 |
+
emb = torch.cat((freqs, freqs), dim=-1).to(x.device)
|
| 215 |
+
self.cos_cached = emb.cos()[None, :, None, :]
|
| 216 |
+
self.sin_cached = emb.sin()[None, :, None, :]
|
| 217 |
+
return self.cos_cached, self.sin_cached
|
| 218 |
+
|
| 219 |
+
|
| 220 |
+
# rotary pos emb helpers:
|
| 221 |
+
def rotate_half(x):
|
| 222 |
+
x1, x2 = x[..., : x.shape[-1] // 2], x[..., x.shape[-1] // 2 :]
|
| 223 |
+
return torch.cat(
|
| 224 |
+
(-x2, x1), dim=-1
|
| 225 |
+
)
|
| 226 |
+
|
| 227 |
+
#@torch.jit.script
|
| 228 |
+
def apply_rotary_pos_emb(q, k, cos, sin):
|
| 229 |
+
return (q * cos[:,:q.shape[1]]) + (rotate_half(q) * sin[:,:q.shape[1]]), (k * cos) + (rotate_half(k) * sin)
|
| 230 |
+
|
| 231 |
+
# %% ../nbs/4B. Semantic to acoustic token modeling.ipynb 35
|
| 232 |
+
from torch import Tensor, nn
|
| 233 |
+
import torch.nn.functional as F
|
| 234 |
+
from typing import Dict, Iterable, Optional
|
| 235 |
+
|
| 236 |
+
class ResidualAttentionBlock(nn.Module):
|
| 237 |
+
def __init__(self, n_state: int, n_head: int, cross_attention: bool = False, rope: bool = False,
|
| 238 |
+
qk_scale: float = 1, ffn_mult: int = 4):
|
| 239 |
+
super().__init__()
|
| 240 |
+
|
| 241 |
+
self.attn = MultiHeadAttention(n_state, n_head, qk_scale=qk_scale, rope=rope)
|
| 242 |
+
self.attn_ln = LayerNorm(n_state)
|
| 243 |
+
|
| 244 |
+
self.cross_attn = (
|
| 245 |
+
MultiHeadAttention(n_state, n_head, qk_scale=qk_scale, rope=rope) if cross_attention else None
|
| 246 |
+
)
|
| 247 |
+
self.cross_attn_ln = LayerNorm(n_state) if cross_attention else None
|
| 248 |
+
|
| 249 |
+
n_mlp = n_state * ffn_mult
|
| 250 |
+
self.mlp = nn.Sequential(
|
| 251 |
+
nn.Linear(n_state, n_mlp), nn.GELU(), nn.Linear(n_mlp, n_state)
|
| 252 |
+
)
|
| 253 |
+
self.mlp_ln = LayerNorm(n_state)
|
| 254 |
+
|
| 255 |
+
def forward(
|
| 256 |
+
self,
|
| 257 |
+
x: Tensor,
|
| 258 |
+
xa: Optional[Tensor] = None,
|
| 259 |
+
causal = False,
|
| 260 |
+
kv_cache: Optional[dict] = None,
|
| 261 |
+
):
|
| 262 |
+
x = x + self.attn(self.attn_ln(x), causal=causal, kv_cache=kv_cache)[0]
|
| 263 |
+
if self.cross_attn:
|
| 264 |
+
x = x + self.cross_attn(self.cross_attn_ln(x), xa, kv_cache=kv_cache)[0]
|
| 265 |
+
x = x + self.mlp(self.mlp_ln(x))
|
| 266 |
+
return x
|
| 267 |
+
|
| 268 |
+
class MultiHeadAttention(nn.Module):
|
| 269 |
+
def __init__(self, n_state: int, n_head: int, qk_scale: float = 1, rope: bool = False):
|
| 270 |
+
super().__init__()
|
| 271 |
+
self.n_head = n_head
|
| 272 |
+
self.sqrt_qk_scale = math.sqrt(qk_scale)
|
| 273 |
+
self.query = QueryHead(n_state, n_state)
|
| 274 |
+
self.key = nn.Linear(n_state, n_state, bias=False)
|
| 275 |
+
self.value = nn.Linear(n_state, n_state)
|
| 276 |
+
self.out = nn.Linear(n_state, n_state)
|
| 277 |
+
|
| 278 |
+
self.rotary = None
|
| 279 |
+
if rope:
|
| 280 |
+
self.rotary = Rotary(n_state // n_head)
|
| 281 |
+
|
| 282 |
+
def forward(
|
| 283 |
+
self,
|
| 284 |
+
x: Tensor,
|
| 285 |
+
xa: Optional[Tensor] = None,
|
| 286 |
+
causal = False,
|
| 287 |
+
kv_cache: Optional[dict] = None,
|
| 288 |
+
):
|
| 289 |
+
q = self.query(x)
|
| 290 |
+
|
| 291 |
+
if kv_cache is None or xa is None or self.key not in kv_cache:
|
| 292 |
+
# hooks, if installed (i.e. kv_cache is not None), will prepend the cached kv tensors;
|
| 293 |
+
# otherwise, perform key/value projections for self- or cross-attention as usual.
|
| 294 |
+
k = self.key(x if xa is None else xa)
|
| 295 |
+
v = self.value(x if xa is None else xa)
|
| 296 |
+
else:
|
| 297 |
+
# for cross-attention, calculate keys and values once and reuse in subsequent calls.
|
| 298 |
+
k = kv_cache[self.key]
|
| 299 |
+
v = kv_cache[self.value]
|
| 300 |
+
|
| 301 |
+
if self.sqrt_qk_scale != 1:
|
| 302 |
+
q *= self.sqrt_qk_scale
|
| 303 |
+
k *= self.sqrt_qk_scale
|
| 304 |
+
|
| 305 |
+
wv, qk = self.qkv_attention_pth20(q, k, v, causal)
|
| 306 |
+
# wv, qk = self.qkv_attention_xformers(q, k, v, causal)
|
| 307 |
+
|
| 308 |
+
return self.out(wv), qk
|
| 309 |
+
|
| 310 |
+
def qkv_attention_pth20(
|
| 311 |
+
self, q: Tensor, k: Tensor, v: Tensor, causal = False
|
| 312 |
+
):
|
| 313 |
+
n_batch, n_ctx, n_state = q.shape
|
| 314 |
+
q = q.view(*q.shape[:2], self.n_head, -1)
|
| 315 |
+
k = k.view(*k.shape[:2], self.n_head, -1)
|
| 316 |
+
v = v.view(*v.shape[:2], self.n_head, -1).permute(0, 2, 1, 3)
|
| 317 |
+
|
| 318 |
+
#print('before rot:', q.shape, k.shape)
|
| 319 |
+
if self.rotary:
|
| 320 |
+
q, k = apply_rotary_pos_emb(q, k, *self.rotary(k))
|
| 321 |
+
#print(' after rot:', q.shape, k.shape)
|
| 322 |
+
|
| 323 |
+
k = k.permute(0, 2, 1, 3)
|
| 324 |
+
q = q.permute(0, 2, 1, 3)
|
| 325 |
+
# modified for better performance under PyTorch 2.0
|
| 326 |
+
wv = F.scaled_dot_product_attention(q, k, v, attn_mask=None, dropout_p=0, is_causal=causal)
|
| 327 |
+
|
| 328 |
+
# previously we've returned q@k which we don't have now
|
| 329 |
+
# since it's not actually used anywhere else, let's just keep two return values for compatibility
|
| 330 |
+
return wv.permute(0, 2, 1, 3).flatten(start_dim=2), None
|
| 331 |
+
|
| 332 |
+
def qkv_attention_xformers(
|
| 333 |
+
self, q: Tensor, k: Tensor, v: Tensor, causal = False
|
| 334 |
+
):
|
| 335 |
+
n_batch, n_ctx, n_state = q.shape
|
| 336 |
+
q = q.view(*q.shape[:2], self.n_head, -1)
|
| 337 |
+
k = k.view(*k.shape[:2], self.n_head, -1)
|
| 338 |
+
v = v.view(*v.shape[:2], self.n_head, -1)
|
| 339 |
+
|
| 340 |
+
if self.rotary:
|
| 341 |
+
q, k = apply_rotary_pos_emb(q, k, *self.rotary(k))
|
| 342 |
+
|
| 343 |
+
bias = xops.LowerTriangularMask() if causal else None
|
| 344 |
+
wv = xops.memory_efficient_attention(q,k,v, attn_bias=bias)
|
| 345 |
+
|
| 346 |
+
# previously we've returned q@k which we don't have now
|
| 347 |
+
# since it's not actually used anywhere else, let's just keep two return values for compatibility
|
| 348 |
+
return wv.flatten(start_dim=2), None
|
| 349 |
+
|
| 350 |
+
# %% ../nbs/4B. Semantic to acoustic token modeling.ipynb 36
|
| 351 |
+
class DelSumDecoder(nn.Module):
|
| 352 |
+
def __init__(self, depth=6, n_head=6, head_width=64, qk_scale=1, ffn_mult=4, length=2250, codes=1024, quantizers=8, linear_heads=True, rope=False, pos_embs=None):
|
| 353 |
+
super().__init__()
|
| 354 |
+
self.length = length
|
| 355 |
+
width = n_head * head_width
|
| 356 |
+
self.width = width
|
| 357 |
+
self.codes = codes
|
| 358 |
+
self.quantizers = quantizers
|
| 359 |
+
self.linear_heads = linear_heads
|
| 360 |
+
|
| 361 |
+
self.embeddings = nn.ModuleList([nn.Embedding(codes+1, width) for _ in range(quantizers)])
|
| 362 |
+
if pos_embs is not None:
|
| 363 |
+
self.register_buffer("positional_embedding", pos_embs)
|
| 364 |
+
|
| 365 |
+
self.layers = nn.ModuleList([
|
| 366 |
+
ResidualAttentionBlock(width, n_head, qk_scale=qk_scale, ffn_mult=ffn_mult, cross_attention=True, rope=rope) for _ in range(math.floor(depth))
|
| 367 |
+
])
|
| 368 |
+
|
| 369 |
+
self.ln_post = LayerNorm(width)
|
| 370 |
+
|
| 371 |
+
if self.linear_heads:
|
| 372 |
+
self.heads = LinearHead(width, (codes+1) * quantizers, bias=False)
|
| 373 |
+
else:
|
| 374 |
+
self.splitter = nn.Sequential(
|
| 375 |
+
nn.Linear(width, width * quantizers),
|
| 376 |
+
nn.GELU(),
|
| 377 |
+
)
|
| 378 |
+
self.heads = nn.ModuleList([
|
| 379 |
+
LinearHead(width, codes+1, bias=True) for _ in range(quantizers)
|
| 380 |
+
])
|
| 381 |
+
|
| 382 |
+
def forward(self, toks, xenc):
|
| 383 |
+
b,_,n = toks.shape
|
| 384 |
+
newn = min(n+1, self.length)
|
| 385 |
+
embs = torch.zeros((b,newn,self.width), dtype=xenc.dtype, device=xenc.device)
|
| 386 |
+
for i in range(self.quantizers):
|
| 387 |
+
embs[:,:i+1] += self.embeddings[i](torch.tensor([self.codes], device=xenc.device))
|
| 388 |
+
if i < n:
|
| 389 |
+
embs[:,i+1:] += self.embeddings[i](toks[:,i,:newn-i-1])
|
| 390 |
+
|
| 391 |
+
x = embs.to(xenc.dtype)
|
| 392 |
+
|
| 393 |
+
for l in self.layers:
|
| 394 |
+
x = l(x, xenc, causal=True)
|
| 395 |
+
x = self.ln_post(x)
|
| 396 |
+
|
| 397 |
+
if self.linear_heads:
|
| 398 |
+
logits = self.heads(x).view(b,newn,self.quantizers,self.codes+1).permute(0,2,1,3)
|
| 399 |
+
else:
|
| 400 |
+
split = self.splitter(x).view(b,newn,self.quantizers,self.width)
|
| 401 |
+
logits = torch.stack([self.heads[q](split[:,:,q]) for q in range(self.quantizers)], dim=1)
|
| 402 |
+
|
| 403 |
+
return logits
|
| 404 |
+
|
| 405 |
+
class EmbeddingProjector(nn.Linear):
|
| 406 |
+
pass
|
| 407 |
+
|
| 408 |
+
def rand(start, end):
|
| 409 |
+
return random.random() * (end - start) + start
|
| 410 |
+
|
| 411 |
+
@dataclasses.dataclass
|
| 412 |
+
class Tunables:
|
| 413 |
+
init_std :float = 9
|
| 414 |
+
embeddings_std :float = 0.2
|
| 415 |
+
embeddings_lr_scale: float = 10
|
| 416 |
+
output_mult :float = 5.6
|
| 417 |
+
# FIXME: try separate mults for self and cross attention
|
| 418 |
+
query_mult :float = .3
|
| 419 |
+
encoder_depth_ratio :float = 0.25
|
| 420 |
+
linear_heads :bool = False
|
| 421 |
+
rope :bool = True
|
| 422 |
+
|
| 423 |
+
lr0 :float = 3e-3
|
| 424 |
+
clip_gradient_norm :float = 2
|
| 425 |
+
weight_decay :float = 1e-3
|
| 426 |
+
warmup_steps :float = 2000
|
| 427 |
+
|
| 428 |
+
random :bool = False
|
| 429 |
+
|
| 430 |
+
def __post_init__(self):
|
| 431 |
+
# randomize the hyperparams if requested
|
| 432 |
+
if self.random:
|
| 433 |
+
self.init_std = 2*10**rand(0,1)
|
| 434 |
+
self.embeddings_std = 10**rand(-1.7,-0.22)
|
| 435 |
+
self.embeddings_lr_scale = 2**rand(2,4)
|
| 436 |
+
self.output_mult = 2**rand(1.5,3)
|
| 437 |
+
self.query_mult = 2**rand(-3,-1.3)
|
| 438 |
+
self.encoder_depth_ratio = random.choice([0.25,0.5])
|
| 439 |
+
self.linear_heads = False
|
| 440 |
+
self.rope = True
|
| 441 |
+
|
| 442 |
+
self.lr0 = 3e-3
|
| 443 |
+
self.clip_gradient_norm = 10**rand(-1,1)
|
| 444 |
+
self.warmup_steps = 100*(10**rand(1.18,1.3))
|
| 445 |
+
|
| 446 |
+
@staticmethod
|
| 447 |
+
def upgrade(args):
|
| 448 |
+
args = {k:v for k,v in args.items()}
|
| 449 |
+
def old_default(name, value):
|
| 450 |
+
if name not in args: args[name] = value
|
| 451 |
+
old_default('rope', False)
|
| 452 |
+
old_default('linear_heads', True)
|
| 453 |
+
return args
|
| 454 |
+
|
| 455 |
+
class SADelARTransformer(nn.Module):
|
| 456 |
+
def __init__(self, depth=3, ctx_n=2250, stoks_len=750, stoks_codes=4097, stoks_width=None, spk_width=None, n_head=3, head_width=64, ffn_mult=4,
|
| 457 |
+
quantizers=8, speaker_map={"1":0}, tunables=Tunables()):
|
| 458 |
+
super().__init__()
|
| 459 |
+
self.quantizers = quantizers
|
| 460 |
+
width = n_head * head_width
|
| 461 |
+
store_attr("depth,ctx_n,stoks_len,stoks_codes,stoks_width,spk_width,n_head,head_width,ffn_mult,quantizers,speaker_map")
|
| 462 |
+
self.width = width
|
| 463 |
+
self.base_width = 3 * head_width
|
| 464 |
+
self.tunables = tunables
|
| 465 |
+
|
| 466 |
+
if stoks_width is None: stoks_width = width
|
| 467 |
+
if spk_width is None: spk_width = width
|
| 468 |
+
self.emb_factor = width != stoks_width
|
| 469 |
+
self.spk_factor = width != spk_width
|
| 470 |
+
|
| 471 |
+
if tunables.rope:
|
| 472 |
+
self.positional_embeddings = None
|
| 473 |
+
else:
|
| 474 |
+
self.register_buffer('positional_embeddings', sinusoids(ctx_n, width))
|
| 475 |
+
|
| 476 |
+
self.speaker_embedding = nn.Embedding(len(speaker_map), width)
|
| 477 |
+
self.semantic_embedding = nn.Embedding(stoks_codes, stoks_width)
|
| 478 |
+
if self.emb_factor:
|
| 479 |
+
self.emb_to_hidden = nn.Linear(stoks_width, width)
|
| 480 |
+
|
| 481 |
+
if self.spk_factor:
|
| 482 |
+
self.spk_to_hidden = EmbeddingProjector(spk_width, width)
|
| 483 |
+
|
| 484 |
+
qk_scale = self.tunables.query_mult * 8 / math.sqrt(head_width)
|
| 485 |
+
|
| 486 |
+
encoder_depth = int(depth * 2 * tunables.encoder_depth_ratio)
|
| 487 |
+
decoder_depth = depth * 2 - encoder_depth
|
| 488 |
+
self.encoder = nn.Sequential(*[
|
| 489 |
+
ResidualAttentionBlock(width, n_head, qk_scale=qk_scale, ffn_mult=ffn_mult, rope=tunables.rope) for _ in range(encoder_depth)
|
| 490 |
+
])
|
| 491 |
+
self.ln_post = LayerNorm(width)
|
| 492 |
+
|
| 493 |
+
self.decoder = DelSumDecoder(pos_embs=self.positional_embeddings, qk_scale=qk_scale,
|
| 494 |
+
length=ctx_n, n_head=n_head, head_width=head_width, ffn_mult=ffn_mult,
|
| 495 |
+
depth=decoder_depth, quantizers=quantizers,
|
| 496 |
+
linear_heads=tunables.linear_heads, rope=tunables.rope)
|
| 497 |
+
|
| 498 |
+
self.register_buffer('val_true', torch.zeros(self.quantizers).cuda())
|
| 499 |
+
self.register_buffer('val_total', torch.zeros(self.quantizers).cuda())
|
| 500 |
+
self.apply(self.init_transformer)
|
| 501 |
+
|
| 502 |
+
def setup(self, device):
|
| 503 |
+
pass
|
| 504 |
+
|
| 505 |
+
def load_frozen_semantic_embeddings(self, vqmodel):
|
| 506 |
+
with torch.no_grad():
|
| 507 |
+
self.semantic_embedding.weight[:] = vqmodel.rq.layers[0]._codebook.embed[0]
|
| 508 |
+
self.semantic_embedding.lr_scale = 0
|
| 509 |
+
|
| 510 |
+
def init_transformer(self, m):
|
| 511 |
+
if isinstance(m, LinearHead):
|
| 512 |
+
m.no_weight_decay = True
|
| 513 |
+
torch.nn.init.constant_(m.weight, 0)
|
| 514 |
+
elif isinstance(m, QueryHead):
|
| 515 |
+
m.lr_scale = 1/(m.weight.shape[1] / self.base_width)
|
| 516 |
+
torch.nn.init.constant_(m.weight, 0)
|
| 517 |
+
elif isinstance(m, nn.Embedding):
|
| 518 |
+
m.no_weight_decay = True
|
| 519 |
+
m.lr_scale = self.tunables.embeddings_lr_scale
|
| 520 |
+
std = self.tunables.embeddings_std
|
| 521 |
+
torch.nn.init.trunc_normal_(m.weight, std=std, a=-3*std, b=3*std)
|
| 522 |
+
elif isinstance(m, EmbeddingProjector):
|
| 523 |
+
m.lr_scale = self.tunables.embeddings_lr_scale/2
|
| 524 |
+
std = self.tunables.init_std
|
| 525 |
+
torch.nn.init.trunc_normal_(m.weight, std=std, a=-3*std, b=3*std)
|
| 526 |
+
elif isinstance(m, nn.Linear):
|
| 527 |
+
m.lr_scale = 1/(m.weight.shape[1] / self.base_width)
|
| 528 |
+
std = self.tunables.init_std / m.weight.shape[1]
|
| 529 |
+
torch.nn.init.trunc_normal_(m.weight, std=std, a=-3*std, b=3*std)
|
| 530 |
+
if m.bias is not None:
|
| 531 |
+
torch.nn.init.trunc_normal_(m.bias, std=std, a=-3*std, b=3*std)
|
| 532 |
+
elif isinstance(m, nn.LayerNorm):
|
| 533 |
+
m.no_weight_decay = True
|
| 534 |
+
torch.nn.init.constant_(m.bias, 0)
|
| 535 |
+
torch.nn.init.constant_(m.weight, 1)
|
| 536 |
+
|
| 537 |
+
def embed_stoks(self, Stoks):
|
| 538 |
+
b,n = Stoks.shape
|
| 539 |
+
if self.stoks_len == 1500:
|
| 540 |
+
# converts 50 toks/s to 75 toks/s by adding padding between every two tokens
|
| 541 |
+
x = Stoks.reshape(b,n//2,2)
|
| 542 |
+
x = x.repeat_interleave(2, -1)[:,:,:3]
|
| 543 |
+
x[:,:,1] = 1024
|
| 544 |
+
x = x.reshape(b,n//2*3)
|
| 545 |
+
else:
|
| 546 |
+
# it's a lot easier with 25 toks/s
|
| 547 |
+
x = Stoks.repeat_interleave(3, -1)
|
| 548 |
+
# embed semantic tokens
|
| 549 |
+
Sembs = self.semantic_embedding(x.to(torch.long))
|
| 550 |
+
if self.emb_factor:
|
| 551 |
+
Sembs = self.emb_to_hidden(Sembs)
|
| 552 |
+
return Sembs
|
| 553 |
+
|
| 554 |
+
def forward(self, Stoks, Atoks, speakers, noloss=False):
|
| 555 |
+
Atoks = Atoks.to(torch.long)
|
| 556 |
+
semb = self.embed_stoks(Stoks)
|
| 557 |
+
with record_function("encoder"):
|
| 558 |
+
if self.positional_embeddings is not None: semb = semb + self.positional_embeddings
|
| 559 |
+
xenc = self.ln_post(self.encoder(semb))
|
| 560 |
+
# xenc = torch.zeros_like(xenc)
|
| 561 |
+
with record_function("decoder"):
|
| 562 |
+
Atoks_gt = Atoks.clone()
|
| 563 |
+
Atoks_gt[Atoks == -100] = 1024
|
| 564 |
+
# we can randomize speaker ids during validation to measure
|
| 565 |
+
# the importance of the speaker embedding vs. just the acoustic prompt/prefix
|
| 566 |
+
# if not self.training: speakers = speakers[torch.randperm(speakers.nelement())]
|
| 567 |
+
spk_embs = self.speaker_embedding(speakers)
|
| 568 |
+
if self.spk_factor: spk_embs = self.spk_to_hidden(spk_embs)
|
| 569 |
+
logits = self.decoder(Atoks_gt, xenc + spk_embs.unsqueeze(1))
|
| 570 |
+
logits *= self.tunables.output_mult / (self.width / self.base_width)
|
| 571 |
+
|
| 572 |
+
if noloss:
|
| 573 |
+
return logits
|
| 574 |
+
|
| 575 |
+
with record_function("loss"):
|
| 576 |
+
N = Atoks.shape[-1]
|
| 577 |
+
loss = 0
|
| 578 |
+
for i in range(self.quantizers):
|
| 579 |
+
loss += F.cross_entropy(logits[:,i,i:].reshape(-1,logits.shape[-1]), Atoks[:,i,:N-i].reshape(-1))
|
| 580 |
+
loss /= self.quantizers
|
| 581 |
+
|
| 582 |
+
if not self.training:
|
| 583 |
+
for i in range(self.quantizers):
|
| 584 |
+
Atoks_i = Atoks[:,i,:N-i]
|
| 585 |
+
valid_Atoks = Atoks_i != -100
|
| 586 |
+
self.val_true[i] += (logits[:,i,i:].argmax(-1)[valid_Atoks] == Atoks_i[valid_Atoks]).float().sum()
|
| 587 |
+
self.val_total[i] += valid_Atoks.float().sum()
|
| 588 |
+
|
| 589 |
+
return logits, loss
|
| 590 |
+
|
| 591 |
+
def get_metrics(self):
|
| 592 |
+
metrics = {
|
| 593 |
+
f'acc_{i}':x.item() for i,x in enumerate(self.val_true / self.val_total)
|
| 594 |
+
}
|
| 595 |
+
self.val_true[:] = 0
|
| 596 |
+
self.val_total[:] = 0
|
| 597 |
+
return metrics
|
| 598 |
+
|
| 599 |
+
#
|
| 600 |
+
# inference
|
| 601 |
+
#
|
| 602 |
+
@classmethod
|
| 603 |
+
def load_model(cls, repo_id="collabora/whisperspeech", filename="s2a_up_wds.model", local_filename=None):
|
| 604 |
+
if not local_filename:
|
| 605 |
+
local_filename = hf_hub_download(repo_id=repo_id, filename=filename)
|
| 606 |
+
spec = torch.load(local_filename)
|
| 607 |
+
if '_extra_state' not in spec['state_dict']: spec['state_dict']['_extra_state'] = { 'speaker_map': spec['config']['speaker_map'] }
|
| 608 |
+
model = cls(**spec['config'], tunables=Tunables(**Tunables.upgrade(spec['tunables'])))
|
| 609 |
+
model.load_state_dict(spec['state_dict'])
|
| 610 |
+
model.eval()
|
| 611 |
+
return model
|
| 612 |
+
|
| 613 |
+
def get_extra_state(self):
|
| 614 |
+
return { 'speaker_map': self.speaker_map }
|
| 615 |
+
|
| 616 |
+
def set_extra_state(self, st):
|
| 617 |
+
self.speaker_map = st['speaker_map']
|
| 618 |
+
|
| 619 |
+
def load_checkpoint(self, local_filename):
|
| 620 |
+
spec = torch.load(local_filename, map_location='cpu')
|
| 621 |
+
assert 'pytorch-lightning_version' in spec, 'not a valid PyTorch Lightning checkpoint'
|
| 622 |
+
state_dict = {k.replace('model.', ''):v
|
| 623 |
+
for k,v in spec['state_dict'].items()}
|
| 624 |
+
self.load_state_dict(state_dict)
|
| 625 |
+
return self
|
| 626 |
+
|
| 627 |
+
def save_model(self, fname):
|
| 628 |
+
torch.save(dict(config = self.__stored_args__,
|
| 629 |
+
tunables = dataclasses.asdict(self.tunables),
|
| 630 |
+
state_dict = self.state_dict()), fname)
|
| 631 |
+
|
| 632 |
+
@property
|
| 633 |
+
def device(self):
|
| 634 |
+
return next(self.parameters()).device
|
| 635 |
+
|
| 636 |
+
@torch.no_grad()
|
| 637 |
+
def generate(self, stoks, speakers, N=None, T=0.7, top_k=None, show_progress_bar=True):
|
| 638 |
+
dev = self.device
|
| 639 |
+
if self.stoks_len == 1500:
|
| 640 |
+
N = N or len(stoks) * 3 // 2
|
| 641 |
+
else:
|
| 642 |
+
N = N or len(stoks) * 3
|
| 643 |
+
stoks = F.pad(stoks.to(dev), (0, self.stoks_len - len(stoks)), value=self.stoks_codes-1).unsqueeze(0)
|
| 644 |
+
speakers = torch.tensor([self.speaker_map[spk] for spk in speakers], device=dev)
|
| 645 |
+
toks = torch.zeros((1,self.quantizers,N), dtype=torch.long, device=dev)
|
| 646 |
+
it = range(0,N)
|
| 647 |
+
if show_progress_bar: it = progress_bar(it)
|
| 648 |
+
for i in it:
|
| 649 |
+
p = self(stoks, toks[:,:,:i], speakers, noloss=True)
|
| 650 |
+
last_p = p[0,:,-1]
|
| 651 |
+
if top_k:
|
| 652 |
+
last_p[last_p < torch.topk(last_p, top_k).values[:,-1,None]] = -torch.inf
|
| 653 |
+
for j,tok in enumerate(torch.multinomial((last_p / float(T)).softmax(-1), 1)):
|
| 654 |
+
toks[0,j,max(0,i-j)] = tok
|
| 655 |
+
if toks[0,0,i] == 1024: return toks[0,:,:i]
|
| 656 |
+
return toks[0]
|
| 657 |
+
|
| 658 |
+
# %% ../nbs/4B. Semantic to acoustic token modeling.ipynb 37
|
| 659 |
+
def _make_model(size:str, quantizers:int=4, tunables:Tunables=Tunables(), dataset:torch.utils.data.Dataset=None, **kwargs):
|
| 660 |
+
assert(dataset is not None)
|
| 661 |
+
kwargs = dict(speaker_map=dataset.speakers, quantizers=quantizers, tunables=tunables, **kwargs)
|
| 662 |
+
if size == 'micro':
|
| 663 |
+
return SADelARTransformer(depth=4, n_head=3, ffn_mult=2, **kwargs)
|
| 664 |
+
if size == 'tiny-narrow':
|
| 665 |
+
return SADelARTransformer(depth=4, n_head=6, ffn_mult=1, **kwargs)
|
| 666 |
+
if size == 'tiny':
|
| 667 |
+
return SADelARTransformer(depth=4, n_head=6, **kwargs)
|
| 668 |
+
if size == 'base':
|
| 669 |
+
return SADelARTransformer(depth=6, n_head=8, **kwargs)
|
| 670 |
+
if size == 'base-deep':
|
| 671 |
+
return SADelARTransformer(depth=9, n_head=8, **kwargs)
|
| 672 |
+
if size == 'base-wide':
|
| 673 |
+
return SADelARTransformer(depth=6, n_head=12, **kwargs)
|
| 674 |
+
if size == 'small/2':
|
| 675 |
+
return SADelARTransformer(depth=9, n_head=12, **kwargs)
|
| 676 |
+
if size == 'small':
|
| 677 |
+
return SADelARTransformer(depth=12, n_head=12, **kwargs)
|
| 678 |
+
if size == 'medium':
|
| 679 |
+
return SADelARTransformer(depth=24, n_head=16, **kwargs)
|
| 680 |
+
|
| 681 |
+
def make_model(size:str, quantizers:int=4, frozen_embeddings_model:str=None, tunables:Tunables=Tunables(), dataset:torch.utils.data.Dataset=None):
|
| 682 |
+
if frozen_embeddings_model:
|
| 683 |
+
vqmodel = vq_stoks.RQBottleneckTransformer.load_model(frozen_embeddings_model)
|
| 684 |
+
model = _make_model(size, quantizers, tunables, dataset, stoks_codes=vqmodel.vq_codes+1, stoks_width=vqmodel.rq.layers[0]._codebook.embed[0].shape[-1])
|
| 685 |
+
model.load_frozen_semantic_embeddings(vqmodel)
|
| 686 |
+
else:
|
| 687 |
+
model = _make_model(size, quantizers, tunables, dataset)
|
| 688 |
+
return model
|
whisperspeech/s2a_delar_mup_wds_mlang.py
ADDED
|
@@ -0,0 +1,564 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/4B. Multi-language semantic to acoustic token modeling.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = ['load_dataset', 'DelSumEmbedding', 'DelSumHead', 'rand', 'Tunables', 'SADelARTransformer']
|
| 5 |
+
|
| 6 |
+
# %% ../nbs/4B. Multi-language semantic to acoustic token modeling.ipynb 1
|
| 7 |
+
import io
|
| 8 |
+
import time
|
| 9 |
+
import math
|
| 10 |
+
import random
|
| 11 |
+
import dataclasses
|
| 12 |
+
|
| 13 |
+
# %% ../nbs/4B. Multi-language semantic to acoustic token modeling.ipynb 2
|
| 14 |
+
import torch
|
| 15 |
+
import torch.nn as nn
|
| 16 |
+
import torch.nn.functional as F
|
| 17 |
+
import numpy as np
|
| 18 |
+
from torch.profiler import profile, record_function, ProfilerActivity, schedule
|
| 19 |
+
from fastcore.basics import store_attr
|
| 20 |
+
from huggingface_hub import hf_hub_download
|
| 21 |
+
|
| 22 |
+
# %% ../nbs/4B. Multi-language semantic to acoustic token modeling.ipynb 3
|
| 23 |
+
from pathlib import Path
|
| 24 |
+
import json
|
| 25 |
+
from fastprogress import progress_bar, master_bar
|
| 26 |
+
|
| 27 |
+
# %% ../nbs/4B. Multi-language semantic to acoustic token modeling.ipynb 4
|
| 28 |
+
from .modules import *
|
| 29 |
+
|
| 30 |
+
# %% ../nbs/4B. Multi-language semantic to acoustic token modeling.ipynb 8
|
| 31 |
+
def rand(start, end):
|
| 32 |
+
return random.random() * (end - start) + start
|
| 33 |
+
|
| 34 |
+
# %% ../nbs/4B. Multi-language semantic to acoustic token modeling.ipynb 9
|
| 35 |
+
def random_trunc(random_trunc_p, atoks_len = 2250, stoks_len = 750):
|
| 36 |
+
atoks_per_second = atoks_len / 30
|
| 37 |
+
def _trunc(samples):
|
| 38 |
+
for s in samples:
|
| 39 |
+
if random.random() < random_trunc_p:
|
| 40 |
+
seconds = rand(0.3, 30)
|
| 41 |
+
s['atoks.npy'] = s['atoks.npy'][:,:math.ceil(seconds * atoks_per_second)]
|
| 42 |
+
s['stoks.npy'] = s['stoks.npy'][:math.ceil(s['atoks.npy'].shape[-1]/atoks_len*stoks_len)]
|
| 43 |
+
yield s
|
| 44 |
+
return _trunc
|
| 45 |
+
|
| 46 |
+
def pad_samples(atoks_len = 2250, stoks_len = 750, stoks_pad_token = 4096):
|
| 47 |
+
def _pad(samples):
|
| 48 |
+
for s in samples:
|
| 49 |
+
s['stoks.npy'] = F.pad(torch.tensor(s['stoks.npy']), (1, stoks_len - s['stoks.npy'].shape[-1]-1), value=stoks_pad_token)
|
| 50 |
+
s['out_stoks'] = F.pad(torch.tensor(s['stoks.npy']), (0, stoks_len - s['stoks.npy'].shape[-1]), value=stoks_pad_token)
|
| 51 |
+
s['atoks.npy'] = F.pad(torch.tensor(s['atoks.npy']), (0, atoks_len - s['atoks.npy'].shape[-1]), value=-100)
|
| 52 |
+
yield s
|
| 53 |
+
return _pad
|
| 54 |
+
|
| 55 |
+
# %% ../nbs/4B. Multi-language semantic to acoustic token modeling.ipynb 10
|
| 56 |
+
def make_speaker_map(shards):
|
| 57 |
+
speakers = set()
|
| 58 |
+
for shard in shards:
|
| 59 |
+
with open(shard+'.speakers.txt') as f: speakers = speakers.union(set(x.strip() for x in f.readlines()))
|
| 60 |
+
return {id:i for i,id in enumerate(sorted(speakers))}
|
| 61 |
+
|
| 62 |
+
def speaker_id_extractor(speaker_map):
|
| 63 |
+
def _extractor(samples):
|
| 64 |
+
for s in samples:
|
| 65 |
+
s['speaker'] = torch.tensor(speaker_map[s['__key__'].split("/")[1]])
|
| 66 |
+
yield s
|
| 67 |
+
return _extractor
|
| 68 |
+
|
| 69 |
+
# %% ../nbs/4B. Multi-language semantic to acoustic token modeling.ipynb 27
|
| 70 |
+
def load_dataset(
|
| 71 |
+
atoks_shard_spec:str, # webdataset folder
|
| 72 |
+
stoks_shard_dir:str, # stoks webdataset base dir
|
| 73 |
+
samples:int, # samples per epoch
|
| 74 |
+
random_trunc_p:float=0,# probability of truncating the input to less than 30 seconds
|
| 75 |
+
vq_codes:int=4096,
|
| 76 |
+
language:str='en',
|
| 77 |
+
weight:float=1,
|
| 78 |
+
validation:bool=False,
|
| 79 |
+
exclude_files:str=None,
|
| 80 |
+
randomize_speakers:bool=False,
|
| 81 |
+
):
|
| 82 |
+
import webdataset as wds
|
| 83 |
+
from whisperspeech import utils
|
| 84 |
+
|
| 85 |
+
shards = utils.shard_glob(atoks_shard_spec)
|
| 86 |
+
excludes = {x for file in exclude_files.split() for x in utils.readlines(file)} if exclude_files else set()
|
| 87 |
+
|
| 88 |
+
def check_for_nan(s):
|
| 89 |
+
if torch.tensor(s['spk_emb.npy']).isnan().any(): print("found NaN:", s['__key__'])
|
| 90 |
+
return s
|
| 91 |
+
|
| 92 |
+
def set_language(x):
|
| 93 |
+
x['language'] = language
|
| 94 |
+
return x
|
| 95 |
+
|
| 96 |
+
same_on_all_nodes = lambda urls: urls # will only be used for validation
|
| 97 |
+
ds = wds.WebDataset(shards, resampled=not validation, nodesplitter=same_on_all_nodes).compose(
|
| 98 |
+
wds.decode(),
|
| 99 |
+
utils.merge_in(utils.derived_dataset('maxvad-stoks', base='atoks-3kbps', suffix='', dir=stoks_shard_dir)),
|
| 100 |
+
wds.map(check_for_nan),
|
| 101 |
+
wds.select(lambda s: s['__key__'] not in excludes),
|
| 102 |
+
wds.map_dict(**{'spk_emb.npy':np.nan_to_num}), # remove nans from the speaker embedding model
|
| 103 |
+
random_trunc(random_trunc_p) if random_trunc_p > 0 else lambda x: x,
|
| 104 |
+
pad_samples(stoks_pad_token=vq_codes-1),
|
| 105 |
+
wds.map(set_language),
|
| 106 |
+
wds.to_tuple('stoks.npy', 'atoks.npy', 'spk_emb.npy', 'language', 'out_stoks'),
|
| 107 |
+
wds.shuffle(20000, initial=20000),
|
| 108 |
+
wds.batched(64),
|
| 109 |
+
)
|
| 110 |
+
if randomize_speakers:
|
| 111 |
+
rng = np.random.default_rng()
|
| 112 |
+
ds = ds.compose(
|
| 113 |
+
wds.map_tuple(None, None, lambda x: rng.permutation(x), None),
|
| 114 |
+
)
|
| 115 |
+
if validation:
|
| 116 |
+
ds = ds.slice(samples // 64)
|
| 117 |
+
ds.total_samples = samples
|
| 118 |
+
ds.weight = weight
|
| 119 |
+
|
| 120 |
+
return ds
|
| 121 |
+
|
| 122 |
+
# %% ../nbs/4B. Multi-language semantic to acoustic token modeling.ipynb 37
|
| 123 |
+
class DelSumEmbedding(nn.Module):
|
| 124 |
+
def __init__(self, n_head=6, head_width=64, atoks_width=None, length=2250, codes=1024, quantizers=8, pos_embs=None):
|
| 125 |
+
super().__init__()
|
| 126 |
+
self.length = length
|
| 127 |
+
width = n_head * head_width
|
| 128 |
+
if atoks_width is None: atoks_width = width
|
| 129 |
+
self.width = width
|
| 130 |
+
self.quantizers = quantizers
|
| 131 |
+
|
| 132 |
+
emb = None
|
| 133 |
+
embs = []
|
| 134 |
+
for _ in range(quantizers):
|
| 135 |
+
emb = FlexEmbeddings(codes, width, special_codes=2, frozen_width=atoks_width,
|
| 136 |
+
special_embedding=emb and emb.special)
|
| 137 |
+
embs.append(emb)
|
| 138 |
+
self.embeddings = nn.ModuleList(embs)
|
| 139 |
+
if pos_embs is not None:
|
| 140 |
+
self.register_buffer("positional_embedding", pos_embs)
|
| 141 |
+
|
| 142 |
+
def forward(self, toks, xenc):
|
| 143 |
+
with record_function("embeddings"):
|
| 144 |
+
b,_,n = toks.shape
|
| 145 |
+
newn = min(n, self.length)
|
| 146 |
+
|
| 147 |
+
embs = torch.zeros((b,newn,self.width), dtype=xenc.dtype, device=xenc.device)
|
| 148 |
+
for i in range(self.quantizers):
|
| 149 |
+
embs[:, :] += self.embeddings[i](toks[:,i,:])
|
| 150 |
+
|
| 151 |
+
x = embs.to(xenc.dtype)
|
| 152 |
+
return x
|
| 153 |
+
|
| 154 |
+
# %% ../nbs/4B. Multi-language semantic to acoustic token modeling.ipynb 38
|
| 155 |
+
class DelSumHead(nn.Module):
|
| 156 |
+
def __init__(self, quantizers=8, n_head=6, head_width=64):
|
| 157 |
+
super().__init__()
|
| 158 |
+
self.width = n_head * head_width
|
| 159 |
+
self.quantizers = quantizers
|
| 160 |
+
self.splitter = nn.Sequential(
|
| 161 |
+
nn.Linear(self.width, self.width * quantizers),
|
| 162 |
+
nn.GELU(),
|
| 163 |
+
)
|
| 164 |
+
|
| 165 |
+
def forward(self, x, embeddings=None):
|
| 166 |
+
b, newn, _ = x.shape
|
| 167 |
+
with record_function("splitter"):
|
| 168 |
+
split = self.splitter(x).view(b,newn,self.quantizers,self.width)
|
| 169 |
+
with record_function("unembed"):
|
| 170 |
+
logits = torch.stack([embeddings[q].unembed(split[:,:,q]) for q in range(self.quantizers)], dim=1)
|
| 171 |
+
return logits
|
| 172 |
+
|
| 173 |
+
def rand(start, end):
|
| 174 |
+
return random.random() * (end - start) + start
|
| 175 |
+
|
| 176 |
+
@dataclasses.dataclass
|
| 177 |
+
class Tunables:
|
| 178 |
+
init_std :float = 9
|
| 179 |
+
embeddings_std :float = 0.2
|
| 180 |
+
embeddings_lr_scale: float = 10
|
| 181 |
+
output_mult :float = 5.6
|
| 182 |
+
# FIXME: try separate mults for self and cross attention
|
| 183 |
+
query_mult :float = .3
|
| 184 |
+
encoder_depth_ratio :float = 0.25
|
| 185 |
+
linear_heads :bool = False
|
| 186 |
+
rope :bool = True
|
| 187 |
+
|
| 188 |
+
lr0 :float = 3e-3
|
| 189 |
+
clip_gradient_norm :float = 2
|
| 190 |
+
weight_decay :float = 1e-3
|
| 191 |
+
warmup_steps :float = 2000
|
| 192 |
+
|
| 193 |
+
random :bool = False
|
| 194 |
+
|
| 195 |
+
def __post_init__(self):
|
| 196 |
+
# randomize the hyperparams if requested
|
| 197 |
+
if self.random:
|
| 198 |
+
self.init_std = 2*10**rand(0,1)
|
| 199 |
+
self.embeddings_std = 10**rand(-1.7,-0.22)
|
| 200 |
+
self.embeddings_lr_scale = 2**rand(2,4)
|
| 201 |
+
self.output_mult = 2**rand(1.5,3)
|
| 202 |
+
self.query_mult = 2**rand(-3,-1.3)
|
| 203 |
+
self.encoder_depth_ratio = random.choice([0.25,0.5])
|
| 204 |
+
self.linear_heads = False
|
| 205 |
+
self.rope = True
|
| 206 |
+
|
| 207 |
+
self.lr0 = 3e-3
|
| 208 |
+
self.clip_gradient_norm = 10**rand(-1,1)
|
| 209 |
+
self.warmup_steps = 100*(10**rand(1.18,1.3))
|
| 210 |
+
|
| 211 |
+
@staticmethod
|
| 212 |
+
def upgrade(args):
|
| 213 |
+
args = {k:v for k,v in args.items()}
|
| 214 |
+
def old_default(name, value):
|
| 215 |
+
if name not in args: args[name] = value
|
| 216 |
+
old_default('rope', False)
|
| 217 |
+
old_default('linear_heads', True)
|
| 218 |
+
return args
|
| 219 |
+
|
| 220 |
+
class SADelARTransformer(nn.Module):
|
| 221 |
+
def __init__(self, depth=3, ctx_n=2250,
|
| 222 |
+
stoks_len=750, stoks_codes=4097, stoks_width=None,
|
| 223 |
+
spk_width=None,
|
| 224 |
+
atoks_width=None,
|
| 225 |
+
n_head=3, head_width=64, ffn_mult=4,
|
| 226 |
+
quantizers=8, speaker_map={"1":0}, tunables=Tunables()):
|
| 227 |
+
super().__init__()
|
| 228 |
+
self.quantizers = quantizers
|
| 229 |
+
self.codes = 1024
|
| 230 |
+
width = n_head * head_width
|
| 231 |
+
store_attr("depth,ctx_n,stoks_len,stoks_codes,stoks_width,spk_width,atoks_width,n_head,head_width,ffn_mult,quantizers,speaker_map")
|
| 232 |
+
self.width = width
|
| 233 |
+
self.base_width = 3 * head_width
|
| 234 |
+
self.tunables = tunables
|
| 235 |
+
|
| 236 |
+
if stoks_width is None: stoks_width = width
|
| 237 |
+
if spk_width is None: spk_width = width
|
| 238 |
+
self.emb_factor = width != stoks_width
|
| 239 |
+
self.spk_factor = width != spk_width
|
| 240 |
+
|
| 241 |
+
if tunables.rope:
|
| 242 |
+
self.positional_embeddings = None
|
| 243 |
+
else:
|
| 244 |
+
self.register_buffer('positional_embeddings', sinusoids(ctx_n, width))
|
| 245 |
+
|
| 246 |
+
# self.speaker_embedding = nn.Embedding(len(speaker_map), spk_width)
|
| 247 |
+
self.semantic_embedding = nn.Embedding(stoks_codes, stoks_width)
|
| 248 |
+
if self.emb_factor:
|
| 249 |
+
self.emb_to_hidden = nn.Linear(stoks_width, width)
|
| 250 |
+
self.hidden_to_emb = nn.Linear(width, stoks_width)
|
| 251 |
+
|
| 252 |
+
if self.spk_factor:
|
| 253 |
+
self.spk_to_hidden = nn.Linear(spk_width, width)
|
| 254 |
+
|
| 255 |
+
qk_scale = self.tunables.query_mult * 8 / math.sqrt(head_width)
|
| 256 |
+
|
| 257 |
+
encoder_depth = int(depth * 2 * tunables.encoder_depth_ratio)
|
| 258 |
+
decoder_depth = depth * 2 - encoder_depth
|
| 259 |
+
self.encoder = nn.Sequential(*[
|
| 260 |
+
ResidualAttentionBlock(width, n_head, qk_scale=qk_scale, ffn_mult=ffn_mult, rope=tunables.rope) for _ in range(encoder_depth)
|
| 261 |
+
]) # FIXME: enclm requires causal attention here
|
| 262 |
+
self.ln_post = LayerNorm(width)
|
| 263 |
+
|
| 264 |
+
self.embds = DelSumEmbedding(
|
| 265 |
+
pos_embs=self.positional_embeddings, length=ctx_n,
|
| 266 |
+
n_head=n_head, head_width=head_width, atoks_width=atoks_width,
|
| 267 |
+
quantizers=quantizers,
|
| 268 |
+
)
|
| 269 |
+
self.decoder = BaseDecoder(qk_scale=qk_scale, length=ctx_n,
|
| 270 |
+
n_head=n_head, width=n_head * head_width,
|
| 271 |
+
ffn_mult=ffn_mult, depth=decoder_depth,
|
| 272 |
+
rope=tunables.rope)
|
| 273 |
+
self.head = DelSumHead(n_head=n_head, head_width=head_width, quantizers=quantizers)
|
| 274 |
+
for l in self.decoder.layers:
|
| 275 |
+
l.cross_attn.key_subsampling = 3
|
| 276 |
+
# for l in self.encoder:
|
| 277 |
+
# l.attn.key_subsampling = 3
|
| 278 |
+
# l.attn.query_subsampling = 3
|
| 279 |
+
|
| 280 |
+
self.register_buffer('val_true', torch.zeros(self.quantizers).cuda())
|
| 281 |
+
self.register_buffer('val_total', torch.zeros(self.quantizers).cuda())
|
| 282 |
+
self.apply(self.init_transformer)
|
| 283 |
+
|
| 284 |
+
def setup(self, device):
|
| 285 |
+
pass
|
| 286 |
+
|
| 287 |
+
def load_frozen_semantic_embeddings(self, vqmodel):
|
| 288 |
+
with torch.no_grad():
|
| 289 |
+
self.semantic_embedding.weight[:] = vqmodel.rq.layers[0]._codebook.embed[0]
|
| 290 |
+
self.semantic_embedding.lr_scale = 0
|
| 291 |
+
|
| 292 |
+
def load_frozen_acoustic_embeddings(self, amodel):
|
| 293 |
+
for i in range(self.quantizers):
|
| 294 |
+
self.decoder.embeddings[i].set_frozen_embeddings(amodel.quantizer.vq.layers[i].codebook)
|
| 295 |
+
|
| 296 |
+
def init_transformer(self, m):
|
| 297 |
+
if isinstance(m, LinearHead):
|
| 298 |
+
m.no_weight_decay = True
|
| 299 |
+
torch.nn.init.constant_(m.weight, 0)
|
| 300 |
+
elif isinstance(m, QueryHead):
|
| 301 |
+
m.lr_scale = 1/(m.weight.shape[1] / self.base_width)
|
| 302 |
+
torch.nn.init.constant_(m.weight, 0)
|
| 303 |
+
elif isinstance(m, nn.Embedding):
|
| 304 |
+
m.no_weight_decay = True
|
| 305 |
+
m.lr_scale = self.tunables.embeddings_lr_scale
|
| 306 |
+
std = self.tunables.embeddings_std
|
| 307 |
+
torch.nn.init.trunc_normal_(m.weight, std=std, a=-3*std, b=3*std)
|
| 308 |
+
# elif isinstance(m, EmbeddingProjector):
|
| 309 |
+
# m.lr_scale = self.tunables.embeddings_lr_scale #1/(m.weight.shape[1] / self.base_width)
|
| 310 |
+
# m.lr_scale = 2/(m.weight.shape[1] / self.base_width)
|
| 311 |
+
# std = self.tunables.init_std / m.weight.shape[1]
|
| 312 |
+
# torch.nn.init.trunc_normal_(m.weight, std=std, a=-3*std, b=3*std)
|
| 313 |
+
elif isinstance(m, nn.Linear):
|
| 314 |
+
m.lr_scale = 1/(m.weight.shape[1] / self.base_width)
|
| 315 |
+
std = self.tunables.init_std / m.weight.shape[1]
|
| 316 |
+
torch.nn.init.trunc_normal_(m.weight, std=std, a=-3*std, b=3*std)
|
| 317 |
+
if m.bias is not None:
|
| 318 |
+
torch.nn.init.trunc_normal_(m.bias, std=std, a=-3*std, b=3*std)
|
| 319 |
+
elif isinstance(m, nn.LayerNorm):
|
| 320 |
+
m.no_weight_decay = True
|
| 321 |
+
torch.nn.init.constant_(m.bias, 0)
|
| 322 |
+
torch.nn.init.constant_(m.weight, 1)
|
| 323 |
+
|
| 324 |
+
def embed_stoks(self, Stoks):
|
| 325 |
+
b,n = Stoks.shape
|
| 326 |
+
if self.stoks_len == 1500:
|
| 327 |
+
# converts 50 toks/s to 75 toks/s by adding padding between every two tokens
|
| 328 |
+
x = Stoks.reshape(b,n//2,2)
|
| 329 |
+
x = x.repeat_interleave(2, -1)[:,:,:3]
|
| 330 |
+
x[:,:,1] = 1024
|
| 331 |
+
x = x.reshape(b,n//2*3)
|
| 332 |
+
else:
|
| 333 |
+
# it's a lot easier with 25 toks/s
|
| 334 |
+
# x = Stoks.repeat_interleave(3, -1)
|
| 335 |
+
x = Stoks
|
| 336 |
+
# embed semantic tokens
|
| 337 |
+
Sembs = self.semantic_embedding(x.to(torch.long))
|
| 338 |
+
if self.emb_factor:
|
| 339 |
+
Sembs = self.emb_to_hidden(Sembs)
|
| 340 |
+
return Sembs
|
| 341 |
+
|
| 342 |
+
def _encoder(self, semb, positions):
|
| 343 |
+
x = semb
|
| 344 |
+
for l in self.encoder: x = l(x, positions)
|
| 345 |
+
return self.ln_post(x)
|
| 346 |
+
|
| 347 |
+
def run_encoder(self, Stoks, speakers):
|
| 348 |
+
semb = self.embed_stoks(Stoks)
|
| 349 |
+
with record_function("encoder"):
|
| 350 |
+
if self.positional_embeddings is not None: semb = semb + self.positional_embeddings
|
| 351 |
+
positions = torch.arange(0, semb.shape[1], device=semb.device)
|
| 352 |
+
xenc = self._encoder(semb, positions)
|
| 353 |
+
if self.training:
|
| 354 |
+
enc_logits = (self.hidden_to_emb(xenc) @ self.semantic_embedding.weight.to(xenc.dtype).T).float()
|
| 355 |
+
enc_logits = enc_logits * self.tunables.output_mult / (self.width / self.base_width)
|
| 356 |
+
else:
|
| 357 |
+
enc_logits = None
|
| 358 |
+
# print(xenc.shape, speakers.shape)
|
| 359 |
+
spk_embs = F.normalize(speakers, dim=-1) # use extracted embeddings
|
| 360 |
+
if self.spk_factor: spk_embs = self.spk_to_hidden(spk_embs)
|
| 361 |
+
return xenc + spk_embs.unsqueeze(1), positions, enc_logits
|
| 362 |
+
|
| 363 |
+
def forward(self, Stoks, Atoks, speakers, langs=None, out_stoks=None, noloss=False, xenc=None, xenc_positions=None, atoks_positions=None):
|
| 364 |
+
if xenc is None:
|
| 365 |
+
Atoks = Atoks.to(torch.long)
|
| 366 |
+
out_stoks = out_stoks.to(torch.long)
|
| 367 |
+
Atoks_gt = Atoks.clone()
|
| 368 |
+
Atoks_gt[Atoks == -100] = 1024
|
| 369 |
+
xenc, enc_logits = self.run_encoder(Stoks, speakers)
|
| 370 |
+
else:
|
| 371 |
+
Atoks_gt = Atoks
|
| 372 |
+
with record_function("decoder"):
|
| 373 |
+
embs = self.embds(Atoks, xenc)
|
| 374 |
+
if atoks_positions is None: atoks_positions = torch.arange(0, embs.shape[1], device=embs.device)
|
| 375 |
+
x = self.decoder(embs, atoks_positions, xenc, xenc_positions)
|
| 376 |
+
logits = self.head(x, embeddings=self.embds.embeddings)
|
| 377 |
+
logits *= self.tunables.output_mult / (self.width / self.base_width)
|
| 378 |
+
|
| 379 |
+
if noloss:
|
| 380 |
+
return logits
|
| 381 |
+
|
| 382 |
+
with record_function("loss"):
|
| 383 |
+
N = Atoks.shape[-1]
|
| 384 |
+
loss = 0
|
| 385 |
+
for i in range(self.quantizers):
|
| 386 |
+
loss += F.cross_entropy(logits[:,i,i:].reshape(-1,logits.shape[-1]), Atoks[:,i,:N-i].reshape(-1))
|
| 387 |
+
if self.training and i == 0:
|
| 388 |
+
loss *= 5
|
| 389 |
+
loss /= self.quantizers
|
| 390 |
+
if self.training:
|
| 391 |
+
loss += 0.1 * F.cross_entropy(enc_logits.transpose(-1,-2), out_stoks)
|
| 392 |
+
|
| 393 |
+
if not self.training:
|
| 394 |
+
for i in range(self.quantizers):
|
| 395 |
+
Atoks_i = Atoks[:,i,:N-i]
|
| 396 |
+
valid_Atoks = Atoks_i != -100
|
| 397 |
+
self.val_true[i] += (logits[:,i,i:].argmax(-1)[valid_Atoks] == Atoks_i[valid_Atoks]).float().sum()
|
| 398 |
+
self.val_total[i] += valid_Atoks.float().sum()
|
| 399 |
+
|
| 400 |
+
return logits, loss
|
| 401 |
+
|
| 402 |
+
def get_metrics(self):
|
| 403 |
+
metrics = {
|
| 404 |
+
f'acc_{i}':x.item() for i,x in enumerate(self.val_true / self.val_total)
|
| 405 |
+
}
|
| 406 |
+
self.val_true[:] = 0
|
| 407 |
+
self.val_total[:] = 0
|
| 408 |
+
return metrics
|
| 409 |
+
|
| 410 |
+
#
|
| 411 |
+
# inference
|
| 412 |
+
#
|
| 413 |
+
@classmethod
|
| 414 |
+
def load_model(cls, ref="collabora/whisperspeech:s2a-q4-small-en+pl.model",
|
| 415 |
+
repo_id=None, filename=None, local_filename=None):
|
| 416 |
+
if repo_id is None and filename is None and local_filename is None:
|
| 417 |
+
if ":" in ref:
|
| 418 |
+
repo_id, filename = ref.split(":", 1)
|
| 419 |
+
else:
|
| 420 |
+
local_filename = ref
|
| 421 |
+
if not local_filename:
|
| 422 |
+
local_filename = hf_hub_download(repo_id=repo_id, filename=filename)
|
| 423 |
+
spec = torch.load(local_filename)
|
| 424 |
+
if '_extra_state' not in spec['state_dict']: spec['state_dict']['_extra_state'] = { 'speaker_map': spec['config']['speaker_map'] }
|
| 425 |
+
model = cls(**spec['config'], tunables=Tunables(**Tunables.upgrade(spec['tunables'])))
|
| 426 |
+
model.load_state_dict(spec['state_dict'])
|
| 427 |
+
model.eval()
|
| 428 |
+
return model
|
| 429 |
+
|
| 430 |
+
def get_extra_state(self):
|
| 431 |
+
return { 'speaker_map': self.speaker_map }
|
| 432 |
+
|
| 433 |
+
def set_extra_state(self, st):
|
| 434 |
+
self.speaker_map = st['speaker_map']
|
| 435 |
+
|
| 436 |
+
def load_checkpoint(self, local_filename):
|
| 437 |
+
spec = torch.load(local_filename, map_location='cpu')
|
| 438 |
+
assert 'pytorch-lightning_version' in spec, 'not a valid PyTorch Lightning checkpoint'
|
| 439 |
+
state_dict = {k.replace('model.', ''):v
|
| 440 |
+
for k,v in spec['state_dict'].items()}
|
| 441 |
+
self.load_state_dict(state_dict)
|
| 442 |
+
return self
|
| 443 |
+
|
| 444 |
+
def save_model(self, fname):
|
| 445 |
+
torch.save(dict(config = self.__stored_args__,
|
| 446 |
+
tunables = dataclasses.asdict(self.tunables),
|
| 447 |
+
state_dict = self.state_dict()), fname)
|
| 448 |
+
|
| 449 |
+
def switch_dtypes(self, dtype=torch.float16):
|
| 450 |
+
self.dtype = dtype
|
| 451 |
+
for n,m in self.named_modules():
|
| 452 |
+
# convert every leaf layer apart from the LayerNorms
|
| 453 |
+
if isinstance(m, (nn.Linear, nn.Embedding)):
|
| 454 |
+
m.to(dtype)
|
| 455 |
+
# take care of buffers ([kv]_cache, masks) that are not in the leaf layers
|
| 456 |
+
for bn,b in m.named_buffers(recurse=False):
|
| 457 |
+
setattr(m,bn,b.to(dtype))
|
| 458 |
+
|
| 459 |
+
def optimize(self, max_batch_size=1, dtype=torch.float16, torch_compile=True):
|
| 460 |
+
for emb in self.embds.embeddings:
|
| 461 |
+
emb.convert_for_eval()
|
| 462 |
+
for l in self.encoder:
|
| 463 |
+
l.attn.convert_for_eval()
|
| 464 |
+
for l in self.decoder.layers:
|
| 465 |
+
l.attn.convert_for_eval()
|
| 466 |
+
l.cross_attn.convert_for_eval()
|
| 467 |
+
l.setup_kv_cache(max_batch_size, self.ctx_n, self.stoks_len)
|
| 468 |
+
self.switch_dtypes(dtype)
|
| 469 |
+
if torch_compile:
|
| 470 |
+
self.generate_next = torch.compile(self.generate_next, mode="reduce-overhead", fullgraph=True)
|
| 471 |
+
|
| 472 |
+
@property
|
| 473 |
+
def device(self):
|
| 474 |
+
return next(self.parameters()).device
|
| 475 |
+
|
| 476 |
+
# from https://github.com/pytorch-labs/gpt-fast/blob/main/generate.py
|
| 477 |
+
def multinomial_sample_one_no_sync(self, probs_sort): # Does multinomial sampling without a cuda synchronization
|
| 478 |
+
q = torch.empty_like(probs_sort).exponential_(1)
|
| 479 |
+
return torch.argmax(probs_sort / q, dim=-1, keepdim=True).to(dtype=torch.int)
|
| 480 |
+
|
| 481 |
+
def logits_to_probs(self, logits, T=1.0, top_k=None):
|
| 482 |
+
logits = logits / max(T, 1e-5)
|
| 483 |
+
|
| 484 |
+
if top_k is not None:
|
| 485 |
+
v, _ = torch.topk(logits, min(top_k, logits.size(-1)))
|
| 486 |
+
pivot = v.select(-1, -1).unsqueeze(-1)
|
| 487 |
+
logits = torch.where(logits < pivot, -float("Inf"), logits)
|
| 488 |
+
probs = torch.nn.functional.softmax(logits, dim=-1)
|
| 489 |
+
return probs
|
| 490 |
+
|
| 491 |
+
def sample(self, logits, T=1.0, top_k=None):
|
| 492 |
+
probs = self.logits_to_probs(logits[0,:,-1], T, top_k)
|
| 493 |
+
idx_next = self.multinomial_sample_one_no_sync(probs)
|
| 494 |
+
return idx_next
|
| 495 |
+
|
| 496 |
+
def generate_one(self, toks, positions, langs, xenc, xenc_positions, T, top_k):
|
| 497 |
+
probs = self(None, toks, None, langs, noloss=True, xenc=xenc, xenc_positions=xenc_positions, atoks_positions=positions)
|
| 498 |
+
return self.sample(probs, T, top_k)
|
| 499 |
+
|
| 500 |
+
def generate_next(self, *args, **kwargs):
|
| 501 |
+
return self.generate_one(*args, **kwargs)
|
| 502 |
+
|
| 503 |
+
@torch.no_grad()
|
| 504 |
+
def generate(self, stoks, speakers, langs=None, N=None, T=0.7, top_k=None, show_progress_bar=True, step=None, subsample_enc=False):
|
| 505 |
+
dev = self.device
|
| 506 |
+
N = N or len(stoks) * 3
|
| 507 |
+
stoks = F.pad(stoks.to(dev), (1, self.stoks_len - len(stoks)-1), value=self.stoks_codes-1).unsqueeze(0)
|
| 508 |
+
speakers = speakers.to(device=dev, dtype=self.dtype)
|
| 509 |
+
toks = torch.full((1,self.quantizers,2250), self.codes+1, dtype=torch.long, device=dev)
|
| 510 |
+
it = range(1,min(N,2250-1))
|
| 511 |
+
if show_progress_bar: it = progress_bar(it)
|
| 512 |
+
with record_function("encode"):
|
| 513 |
+
xenc, xenc_positions, _ = self.run_encoder(stoks, speakers)
|
| 514 |
+
toks_positions = torch.arange(N, device=dev)
|
| 515 |
+
with record_function("prefill"):
|
| 516 |
+
toks[0,0,1] = self.generate_one(toks[:,:,:1], toks_positions[:1], langs, xenc, xenc_positions, T, top_k)[0,0]
|
| 517 |
+
with torch.backends.cuda.sdp_kernel(enable_flash=False, enable_mem_efficient=False, enable_math=True):
|
| 518 |
+
for i in it:
|
| 519 |
+
with record_function("generate_one"):
|
| 520 |
+
toks[0,:i+1,i+1] = self.generate_next(toks[:,:,i:i+1], toks_positions[i:i+1], langs, xenc, xenc_positions, T, top_k)[:i+1,0]
|
| 521 |
+
|
| 522 |
+
# for profiling, debugging or early exit
|
| 523 |
+
if step is not None: step()
|
| 524 |
+
# shift tokens
|
| 525 |
+
toks = toks[:,:,1:N]
|
| 526 |
+
for j in range(self.quantizers):
|
| 527 |
+
toks[0, j] = torch.roll(toks[0, j], -j)
|
| 528 |
+
return toks[0]
|
| 529 |
+
|
| 530 |
+
# %% ../nbs/4B. Multi-language semantic to acoustic token modeling.ipynb 39
|
| 531 |
+
def _make_model(size:str, quantizers:int=4, tunables:Tunables=Tunables(), **kwargs):
|
| 532 |
+
kwargs = dict(quantizers=quantizers, tunables=tunables, **kwargs)
|
| 533 |
+
if size == 'micro':
|
| 534 |
+
return SADelARTransformer(depth=4, n_head=3, ffn_mult=2, **kwargs)
|
| 535 |
+
if size == 'tiny-narrow':
|
| 536 |
+
return SADelARTransformer(depth=4, n_head=6, ffn_mult=1, **kwargs)
|
| 537 |
+
if size == 'tiny':
|
| 538 |
+
return SADelARTransformer(depth=4, n_head=6, **kwargs)
|
| 539 |
+
if size == 'base':
|
| 540 |
+
return SADelARTransformer(depth=6, n_head=8, **kwargs)
|
| 541 |
+
if size == 'base-deep':
|
| 542 |
+
return SADelARTransformer(depth=9, n_head=8, **kwargs)
|
| 543 |
+
if size == 'base-wide':
|
| 544 |
+
return SADelARTransformer(depth=6, n_head=12, **kwargs)
|
| 545 |
+
if size == 'small/2':
|
| 546 |
+
return SADelARTransformer(depth=9, n_head=12, **kwargs)
|
| 547 |
+
if size == 'small':
|
| 548 |
+
return SADelARTransformer(depth=12, n_head=12, **kwargs)
|
| 549 |
+
if size == 'medium':
|
| 550 |
+
return SADelARTransformer(depth=24, n_head=16, **kwargs)
|
| 551 |
+
|
| 552 |
+
def make_model(size:str, quantizers:int=4, frozen_embeddings_model:str=None, frozen_acoustic_embeddings:bool=False, spk_width:int=None, tunables:Tunables=Tunables(), dataset=None):
|
| 553 |
+
from encodec.model import EncodecModel
|
| 554 |
+
from whisperspeech import vq_stoks
|
| 555 |
+
|
| 556 |
+
amodel = EncodecModel.encodec_model_24khz() if frozen_acoustic_embeddings else None
|
| 557 |
+
vqmodel = vq_stoks.RQBottleneckTransformer.load_model(frozen_embeddings_model) if frozen_embeddings_model else None
|
| 558 |
+
model = _make_model(size, quantizers, tunables,
|
| 559 |
+
spk_width=spk_width,
|
| 560 |
+
atoks_width=amodel and amodel.quantizer.vq.layers[0]._codebook.embed.shape[-1],
|
| 561 |
+
stoks_codes=vqmodel.vq_codes+1, stoks_width=vqmodel.rq.layers[0]._codebook.embed[0].shape[-1])
|
| 562 |
+
if vqmodel: model.load_frozen_semantic_embeddings(vqmodel)
|
| 563 |
+
if amodel: model.load_frozen_acoustic_embeddings(amodel)
|
| 564 |
+
return model
|
whisperspeech/t2s_up_wds.py
ADDED
|
@@ -0,0 +1,442 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/5B. Text to semantic token modeling.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = ['load_datasets', 'rand', 'Tunables', 'Encoder', 'Decoder', 'TSARTransformer', 'make_model']
|
| 5 |
+
|
| 6 |
+
# %% ../nbs/5B. Text to semantic token modeling.ipynb 1
|
| 7 |
+
import dataclasses
|
| 8 |
+
import random
|
| 9 |
+
import math
|
| 10 |
+
import torch
|
| 11 |
+
import torch.nn as nn
|
| 12 |
+
import torch.nn.functional as F
|
| 13 |
+
from torch.profiler import record_function
|
| 14 |
+
|
| 15 |
+
from huggingface_hub import hf_hub_download
|
| 16 |
+
from fastcore.basics import store_attr
|
| 17 |
+
from fastprogress import progress_bar
|
| 18 |
+
|
| 19 |
+
import webdataset as wds
|
| 20 |
+
|
| 21 |
+
# %% ../nbs/5B. Text to semantic token modeling.ipynb 2
|
| 22 |
+
from pathlib import Path
|
| 23 |
+
import pylab as plt
|
| 24 |
+
import pandas as pd
|
| 25 |
+
import numpy as np
|
| 26 |
+
|
| 27 |
+
# %% ../nbs/5B. Text to semantic token modeling.ipynb 3
|
| 28 |
+
import whisper
|
| 29 |
+
from whisperspeech.train import *
|
| 30 |
+
from whisperspeech.modules import *
|
| 31 |
+
from whisperspeech import vq_stoks
|
| 32 |
+
|
| 33 |
+
# %% ../nbs/5B. Text to semantic token modeling.ipynb 8
|
| 34 |
+
import re
|
| 35 |
+
|
| 36 |
+
class CharTokenizer:
|
| 37 |
+
"""Trivial tokenizer β just use UTF-8 bytes"""
|
| 38 |
+
eot = 0
|
| 39 |
+
|
| 40 |
+
def encode(self, txt):
|
| 41 |
+
return list(bytes(txt.strip(), 'utf-8'))
|
| 42 |
+
|
| 43 |
+
def decode(self, tokens):
|
| 44 |
+
return bytes(tokens).decode('utf-8')
|
| 45 |
+
|
| 46 |
+
def tokenizer(ikey, okey, length):
|
| 47 |
+
"""Tokenizes a transcript"""
|
| 48 |
+
tok = CharTokenizer()
|
| 49 |
+
def _tokenizer(samples):
|
| 50 |
+
for s in samples:
|
| 51 |
+
toks = torch.tensor(tok.encode(s[ikey]))
|
| 52 |
+
s[okey] = F.pad(toks, (0, length - toks.shape[-1]), value=tok.eot)
|
| 53 |
+
yield s
|
| 54 |
+
return _tokenizer
|
| 55 |
+
|
| 56 |
+
def ar_padder(ikey, okey, length, pad_token):
|
| 57 |
+
"""Pads the tokens for autoregresive training"""
|
| 58 |
+
def _ar_padder(samples):
|
| 59 |
+
for s in samples:
|
| 60 |
+
toks = s[ikey]
|
| 61 |
+
if isinstance(toks, (list, np.ndarray)): toks = torch.tensor(toks)
|
| 62 |
+
toks = toks.to(torch.long)
|
| 63 |
+
s['in_' +okey] = F.pad(toks, (1, length - toks.shape[-1] - 1), value=pad_token)
|
| 64 |
+
s['out_'+okey] = F.pad(toks, (0, length - toks.shape[-1]), value=pad_token)
|
| 65 |
+
yield s
|
| 66 |
+
return _ar_padder
|
| 67 |
+
|
| 68 |
+
def char_per_seconder(txt_key, stoks_key, cps_key, stoks_per_second=25):
|
| 69 |
+
"""Adds the characters per second metric to the input data"""
|
| 70 |
+
def _char_per_seconder(samples):
|
| 71 |
+
for s in samples:
|
| 72 |
+
secs = s[stoks_key].shape[-1] / stoks_per_second
|
| 73 |
+
s[cps_key] = len(s[txt_key]) / secs
|
| 74 |
+
yield s
|
| 75 |
+
return _char_per_seconder
|
| 76 |
+
|
| 77 |
+
# %% ../nbs/5B. Text to semantic token modeling.ipynb 9
|
| 78 |
+
def build_speaker_map(shards):
|
| 79 |
+
speakers = set()
|
| 80 |
+
for shard in shards:
|
| 81 |
+
with open(shard+'.speakers.txt') as f: speakers = speakers.union(set(x.strip() for x in f.readlines()))
|
| 82 |
+
return {id:i for i,id in enumerate(speakers)}
|
| 83 |
+
|
| 84 |
+
def speaker_id_extractor(speaker_map):
|
| 85 |
+
def _extractor(samples):
|
| 86 |
+
for s in samples:
|
| 87 |
+
s['speaker'] = torch.tensor(speaker_map[s['__key__'].split("/")[1]])
|
| 88 |
+
yield s
|
| 89 |
+
return _extractor
|
| 90 |
+
|
| 91 |
+
# %% ../nbs/5B. Text to semantic token modeling.ipynb 10
|
| 92 |
+
def load_datasets(
|
| 93 |
+
input:str, # webdataset folder or shard list
|
| 94 |
+
samples:int, # samples per epoch
|
| 95 |
+
subsample:float=1, # use a fraction of the files
|
| 96 |
+
val_samples:int=512,
|
| 97 |
+
vq_codes:int=4096,
|
| 98 |
+
):
|
| 99 |
+
if isinstance(input, (Path, str)):
|
| 100 |
+
path = Path(input)
|
| 101 |
+
if path.is_dir():
|
| 102 |
+
glob = '*-t2s-*.tar.gz'
|
| 103 |
+
else:
|
| 104 |
+
glob = path.name
|
| 105 |
+
path = path.parent
|
| 106 |
+
input = Path(path).glob(glob)
|
| 107 |
+
elif isinstance(input, list):
|
| 108 |
+
pass
|
| 109 |
+
else:
|
| 110 |
+
raise ArgumentError("input should be either a list of a path with an optional glob specifier")
|
| 111 |
+
shards = [str(x) for x in input]
|
| 112 |
+
|
| 113 |
+
speaker_map = build_speaker_map(shards)
|
| 114 |
+
|
| 115 |
+
def ds(shards, length):
|
| 116 |
+
ds = wds.WebDataset(wds.ResampledShards(shards)).compose(
|
| 117 |
+
wds.decode(),
|
| 118 |
+
speaker_id_extractor(speaker_map),
|
| 119 |
+
wds.select(lambda s: s['stoks.npy'].shape[-1] > 12), # select samples > .5s
|
| 120 |
+
tokenizer('txt', 'ttoks', length=550),
|
| 121 |
+
ar_padder('stoks.npy', 'stoks', length=750, pad_token=vq_codes-1),
|
| 122 |
+
char_per_seconder('txt', 'stoks.npy', 'cps', stoks_per_second=25),
|
| 123 |
+
wds.to_tuple('ttoks', 'speaker', 'cps', 'in_stoks', 'out_stoks'),
|
| 124 |
+
wds.batched(64)
|
| 125 |
+
)
|
| 126 |
+
ds.speakers = speaker_map
|
| 127 |
+
ds.total_samples = length
|
| 128 |
+
ds.stoks_len = 750
|
| 129 |
+
ds.stoks_codes = vq_codes
|
| 130 |
+
ds.ttoks_len = 550
|
| 131 |
+
return ds.compose(wds.slice(length // 64)).with_epoch(length // 64).with_length(length // 64)
|
| 132 |
+
|
| 133 |
+
return (
|
| 134 |
+
ds(shards[1:], samples),
|
| 135 |
+
ds(shards[:1], val_samples),
|
| 136 |
+
)
|
| 137 |
+
|
| 138 |
+
# %% ../nbs/5B. Text to semantic token modeling.ipynb 14
|
| 139 |
+
def rand(start, end):
|
| 140 |
+
return random.random() * (end - start) + start
|
| 141 |
+
|
| 142 |
+
@dataclasses.dataclass
|
| 143 |
+
class Tunables:
|
| 144 |
+
init_std :float = 1
|
| 145 |
+
embeddings_std :float = .01
|
| 146 |
+
embeddings_lr_scale: float = 5
|
| 147 |
+
embedding_projector_lr_scale: float = 2.5
|
| 148 |
+
output_mult :float = .35
|
| 149 |
+
query_mult :float = 1
|
| 150 |
+
encoder_depth_ratio :float = 0.25
|
| 151 |
+
eot_dropout_p :float = .5
|
| 152 |
+
cps_input: bool = True
|
| 153 |
+
cps_bins: int = 32
|
| 154 |
+
|
| 155 |
+
lr0 :float = 1.5e-3
|
| 156 |
+
clip_gradient_norm :float = .2
|
| 157 |
+
weight_decay :float = 1e-1
|
| 158 |
+
warmup_steps :float = 4000
|
| 159 |
+
|
| 160 |
+
random :bool = False
|
| 161 |
+
|
| 162 |
+
def __post_init__(self):
|
| 163 |
+
# randomize the hyperparams if requested
|
| 164 |
+
if self.random:
|
| 165 |
+
self.init_std = 10**rand(-1,1)
|
| 166 |
+
self.embeddings_std = 10**rand(-3,-.7)
|
| 167 |
+
self.embeddings_lr_scale = rand(2,6)
|
| 168 |
+
self.output_mult = rand(0.25,0.65)
|
| 169 |
+
self.query_mult = 2**rand(-2,3)
|
| 170 |
+
self.encoder_depth_ratio = 0.25
|
| 171 |
+
|
| 172 |
+
self.lr0 = rand(1,5)*1e-3
|
| 173 |
+
self.clip_gradient_norm = 10**rand(-3,0)
|
| 174 |
+
self.warmup_steps = 100*(10**rand(1,1.85))
|
| 175 |
+
|
| 176 |
+
# %% ../nbs/5B. Text to semantic token modeling.ipynb 15
|
| 177 |
+
class EmbeddingProjector(nn.Linear):
|
| 178 |
+
pass
|
| 179 |
+
|
| 180 |
+
# %% ../nbs/5B. Text to semantic token modeling.ipynb 16
|
| 181 |
+
class Encoder(nn.Module):
|
| 182 |
+
def __init__(self, depth=6, width=384, n_head=6, length=1500, codes=1024, emb_width=384, ffn_mult=4, pos_embs=None, tunables=Tunables()):
|
| 183 |
+
super().__init__()
|
| 184 |
+
self.emb_width = emb_width
|
| 185 |
+
|
| 186 |
+
self.emb_factor = width != emb_width
|
| 187 |
+
|
| 188 |
+
self.embedding = nn.Embedding(codes, emb_width)
|
| 189 |
+
if self.emb_factor:
|
| 190 |
+
self.emb_to_hidden = EmbeddingProjector(emb_width, width)
|
| 191 |
+
|
| 192 |
+
if pos_embs is None: pos_embs = sinusoids(length, width)
|
| 193 |
+
self.register_buffer("positional_embedding", pos_embs)
|
| 194 |
+
|
| 195 |
+
self.layers = nn.Sequential(*[
|
| 196 |
+
ResidualAttentionBlock(width, n_head,
|
| 197 |
+
qk_scale=tunables.query_mult*8/math.sqrt(width/n_head), ffn_mult=ffn_mult) for _ in range(depth)
|
| 198 |
+
])
|
| 199 |
+
|
| 200 |
+
self.ln_post = LayerNorm(width)
|
| 201 |
+
|
| 202 |
+
def forward(self, Stoks):
|
| 203 |
+
xin = self.embedding(Stoks)
|
| 204 |
+
if self.emb_factor:
|
| 205 |
+
xin = self.emb_to_hidden(xin)
|
| 206 |
+
|
| 207 |
+
assert xin.shape[1:] == self.positional_embedding.shape, "incorrect semantic token shape"
|
| 208 |
+
xin = (xin + self.positional_embedding).to(xin.dtype)
|
| 209 |
+
|
| 210 |
+
return self.ln_post(self.layers(xin))
|
| 211 |
+
|
| 212 |
+
# %% ../nbs/5B. Text to semantic token modeling.ipynb 17
|
| 213 |
+
class Decoder(nn.Module):
|
| 214 |
+
def __init__(self, depth=6, stoks_width=384, width=384, n_head=6, length=1500, codes=1024, ffn_mult=4, pos_embs=None, tunables=Tunables()):
|
| 215 |
+
super().__init__()
|
| 216 |
+
self.length = length
|
| 217 |
+
self.codes = codes
|
| 218 |
+
self.width = width
|
| 219 |
+
self.stoks_width = stoks_width
|
| 220 |
+
|
| 221 |
+
self.emb_factor = width != stoks_width
|
| 222 |
+
|
| 223 |
+
# embed semantic tokens
|
| 224 |
+
self.embedding = nn.Embedding(codes, stoks_width)
|
| 225 |
+
if self.emb_factor:
|
| 226 |
+
self.emb_to_hidden = EmbeddingProjector(stoks_width, width)
|
| 227 |
+
self.hidden_to_emb = EmbeddingProjector(width, stoks_width)
|
| 228 |
+
|
| 229 |
+
if pos_embs is None: pos_embs = sinusoids(length, width)
|
| 230 |
+
self.register_buffer("positional_embedding", pos_embs)
|
| 231 |
+
|
| 232 |
+
self.layers = nn.ModuleList([
|
| 233 |
+
ResidualAttentionBlock(width, n_head, cross_attention=True,
|
| 234 |
+
qk_scale=tunables.query_mult*8/math.sqrt(width/n_head), ffn_mult=ffn_mult) for _ in range(depth)
|
| 235 |
+
])
|
| 236 |
+
self.ln_post = LayerNorm(width)
|
| 237 |
+
|
| 238 |
+
def forward(self, Stoks, xenc, cps=None):
|
| 239 |
+
Sembs = self.embedding(Stoks)
|
| 240 |
+
|
| 241 |
+
if self.emb_factor:
|
| 242 |
+
Sembs = self.emb_to_hidden(Sembs)
|
| 243 |
+
|
| 244 |
+
xin = (Sembs + self.positional_embedding[:Sembs.shape[1]]).to(xenc.dtype)
|
| 245 |
+
if cps is not None: xin = xin + cps
|
| 246 |
+
|
| 247 |
+
x = xin
|
| 248 |
+
for l in self.layers: x = l(x, xenc, causal=True)
|
| 249 |
+
|
| 250 |
+
x = self.ln_post(x)
|
| 251 |
+
|
| 252 |
+
if self.emb_factor:
|
| 253 |
+
x = self.hidden_to_emb(x)
|
| 254 |
+
|
| 255 |
+
logits = (x @ self.embedding.weight.to(x.dtype).T).float()
|
| 256 |
+
return logits
|
| 257 |
+
|
| 258 |
+
# %% ../nbs/5B. Text to semantic token modeling.ipynb 18
|
| 259 |
+
class TSARTransformer(nn.Module):
|
| 260 |
+
def __init__(self, depth=6, n_head=6, head_width=64, ffn_mult=4, language='en',
|
| 261 |
+
ttoks_len=200, ttoks_codes=50364, ttoks_width=None,
|
| 262 |
+
stoks_len=1500, stoks_codes=1024, stoks_width=None,
|
| 263 |
+
tunables=Tunables()):
|
| 264 |
+
assert language == 'en', "only english is supported right now"
|
| 265 |
+
super().__init__()
|
| 266 |
+
store_attr("depth,n_head,head_width,ffn_mult,stoks_width,ttoks_width,ttoks_len,stoks_len,ttoks_codes,stoks_codes,language")
|
| 267 |
+
|
| 268 |
+
width = n_head * head_width
|
| 269 |
+
self.width = width
|
| 270 |
+
self.base_width = 3 * head_width
|
| 271 |
+
self.tunables = tunables
|
| 272 |
+
if self.stoks_width is None: self.stoks_width = self.width
|
| 273 |
+
if self.ttoks_width is None: self.ttoks_width = self.width
|
| 274 |
+
|
| 275 |
+
if tunables.cps_input:
|
| 276 |
+
self.cps_embeddings = nn.Embedding(tunables.cps_bins, self.width)
|
| 277 |
+
else:
|
| 278 |
+
self.cps_embeddings = None
|
| 279 |
+
|
| 280 |
+
encoder_depth = int(depth * 2 * tunables.encoder_depth_ratio)
|
| 281 |
+
decoder_depth = depth * 2 - encoder_depth
|
| 282 |
+
tformer_args = dict(width=width, n_head=n_head, ffn_mult=ffn_mult, tunables=tunables)
|
| 283 |
+
self.encoder = Encoder(length=ttoks_len, codes=ttoks_codes, emb_width=self.ttoks_width, depth=encoder_depth, **tformer_args)
|
| 284 |
+
self.decoder = Decoder(length=stoks_len, codes=stoks_codes, stoks_width=self.stoks_width, depth=decoder_depth, **tformer_args)
|
| 285 |
+
|
| 286 |
+
self.tokenizer = None
|
| 287 |
+
|
| 288 |
+
self.apply(self.init_transformer)
|
| 289 |
+
|
| 290 |
+
def load_frozen_semantic_embeddings(self, vqmodel):
|
| 291 |
+
with torch.no_grad():
|
| 292 |
+
self.decoder.embedding.weight[:] = vqmodel.rq.layers[0]._codebook.embed[0]
|
| 293 |
+
self.decoder.embedding.lr_scale = 0
|
| 294 |
+
|
| 295 |
+
def setup(self, device):
|
| 296 |
+
pass
|
| 297 |
+
|
| 298 |
+
def init_transformer(self, m):
|
| 299 |
+
if isinstance(m, LinearHead):
|
| 300 |
+
m.no_weight_decay = True
|
| 301 |
+
torch.nn.init.constant_(m.weight, 0)
|
| 302 |
+
elif isinstance(m, QueryHead):
|
| 303 |
+
m.lr_scale = 1/(m.weight.shape[1] / self.base_width)
|
| 304 |
+
torch.nn.init.constant_(m.weight, 0)
|
| 305 |
+
elif isinstance(m, nn.Embedding):
|
| 306 |
+
m.no_weight_decay = True
|
| 307 |
+
m.lr_scale = self.tunables.embeddings_lr_scale
|
| 308 |
+
std = self.tunables.embeddings_std
|
| 309 |
+
torch.nn.init.trunc_normal_(m.weight, std=std, a=-3*std, b=3*std)
|
| 310 |
+
elif isinstance(m, EmbeddingProjector):
|
| 311 |
+
m.lr_scale = self.tunables.embedding_projector_lr_scale
|
| 312 |
+
std = self.tunables.init_std
|
| 313 |
+
torch.nn.init.trunc_normal_(m.weight, std=std, a=-3*std, b=3*std)
|
| 314 |
+
elif isinstance(m, nn.Linear):
|
| 315 |
+
m.lr_scale = 1/(m.weight.shape[1] / self.base_width)
|
| 316 |
+
std = self.tunables.init_std / m.weight.shape[1]
|
| 317 |
+
torch.nn.init.trunc_normal_(m.weight, std=std, a=-3*std, b=3*std)
|
| 318 |
+
if m.bias is not None:
|
| 319 |
+
torch.nn.init.trunc_normal_(m.bias, std=std, a=-3*std, b=3*std)
|
| 320 |
+
elif isinstance(m, nn.LayerNorm):
|
| 321 |
+
m.no_weight_decay = True
|
| 322 |
+
torch.nn.init.constant_(m.bias, 0)
|
| 323 |
+
torch.nn.init.constant_(m.weight, 1)
|
| 324 |
+
|
| 325 |
+
def forward(self, Ttoks, speakers, cpss, in_stoks, out_stoks=None, loss=True):
|
| 326 |
+
with record_function("encoder"):
|
| 327 |
+
xenc = self.encoder(Ttoks.to(torch.long))
|
| 328 |
+
with record_function("decoder"):
|
| 329 |
+
if self.cps_embeddings:
|
| 330 |
+
cps_bin = (cpss / 20 * self.tunables.cps_bins).to(torch.long)
|
| 331 |
+
cps_bin[cps_bin >= self.tunables.cps_bins] = self.tunables.cps_bins-1
|
| 332 |
+
cps_embs = self.cps_embeddings(cps_bin).unsqueeze(1)
|
| 333 |
+
else:
|
| 334 |
+
cps_embs = None
|
| 335 |
+
logits = self.decoder(in_stoks, xenc, cps=cps_embs) * self.tunables.output_mult / (self.width / self.base_width)
|
| 336 |
+
if loss is not None:
|
| 337 |
+
with record_function("loss"):
|
| 338 |
+
loss = F.cross_entropy(logits.transpose(-1,-2), out_stoks)#, reduction='none')
|
| 339 |
+
return logits, loss
|
| 340 |
+
|
| 341 |
+
#
|
| 342 |
+
# inference
|
| 343 |
+
#
|
| 344 |
+
@classmethod
|
| 345 |
+
def load_model(cls, repo_id="collabora/whisperspeech", filename="t2s_up_wds.model", local_filename=None):
|
| 346 |
+
if not local_filename:
|
| 347 |
+
local_filename = hf_hub_download(repo_id=repo_id, filename=filename)
|
| 348 |
+
spec = torch.load(local_filename)
|
| 349 |
+
model = cls(**spec['config'], tunables=Tunables(**spec['tunables']))
|
| 350 |
+
model.load_state_dict(spec['state_dict'])
|
| 351 |
+
model.eval()
|
| 352 |
+
return model
|
| 353 |
+
|
| 354 |
+
def load_checkpoint(self, local_filename):
|
| 355 |
+
spec = torch.load(local_filename, map_location='cpu')
|
| 356 |
+
assert 'pytorch-lightning_version' in spec, 'not a valid PyTorch Lightning checkpoint'
|
| 357 |
+
state_dict = {k.replace('model.', ''):v
|
| 358 |
+
for k,v in spec['state_dict'].items()}
|
| 359 |
+
self.load_state_dict(state_dict)
|
| 360 |
+
return self
|
| 361 |
+
|
| 362 |
+
def save_model(self, fname):
|
| 363 |
+
torch.save(dict(config = self.__stored_args__,
|
| 364 |
+
tunables = dataclasses.asdict(self.tunables),
|
| 365 |
+
state_dict = self.state_dict()), fname)
|
| 366 |
+
|
| 367 |
+
def ensure_tokenizer(self):
|
| 368 |
+
assert not self.training
|
| 369 |
+
if self.tokenizer is None: self.tokenizer = CharTokenizer()
|
| 370 |
+
#whisper.tokenizer.get_tokenizer(multilingual=True)
|
| 371 |
+
|
| 372 |
+
@property
|
| 373 |
+
def device(self):
|
| 374 |
+
return next(self.parameters()).device
|
| 375 |
+
|
| 376 |
+
@torch.no_grad()
|
| 377 |
+
def generate(self, txt, cps=15, N=None, T=0.7, top_k=None, show_progress_bar=True):
|
| 378 |
+
self.ensure_tokenizer()
|
| 379 |
+
N = N or self.stoks_len
|
| 380 |
+
dev = self.device
|
| 381 |
+
ttoks = torch.tensor(self.tokenizer.encode(txt), device=dev)
|
| 382 |
+
ttoks = F.pad(ttoks, (0, self.ttoks_len - len(ttoks)), value=self.tokenizer.eot).unsqueeze(0)
|
| 383 |
+
cpss = torch.tensor([cps], device=dev)
|
| 384 |
+
toks = torch.zeros((1,N), dtype=torch.long, device=dev)
|
| 385 |
+
toks[0,0] = self.stoks_codes-1
|
| 386 |
+
it = range(1,N)
|
| 387 |
+
if show_progress_bar: it = progress_bar(it)
|
| 388 |
+
for i in it:
|
| 389 |
+
p, _ = self(ttoks, None, cpss, toks[:,:i], loss=None)
|
| 390 |
+
last_p = p[0,-1]
|
| 391 |
+
if top_k:
|
| 392 |
+
last_p[last_p < torch.topk(last_p, top_k).values[-1,None]] = -torch.inf
|
| 393 |
+
tok = torch.multinomial((last_p / float(T)).softmax(-1), 1)
|
| 394 |
+
toks[0,i] = tok
|
| 395 |
+
if toks[0,i] == self.stoks_codes-1: return toks[0,1:i]
|
| 396 |
+
return toks[0,1:]
|
| 397 |
+
|
| 398 |
+
@torch.no_grad()
|
| 399 |
+
def generate_batch(self, txts, N=None, T=1.1, top_k=7, show_progress_bar=True):
|
| 400 |
+
self.ensure_tokenizer()
|
| 401 |
+
N = self.stoks_len
|
| 402 |
+
dev = self.device
|
| 403 |
+
ttoks = []
|
| 404 |
+
for txt in txts:
|
| 405 |
+
ttoks_ = torch.tensor(self.tokenizer.encode(txt), device=dev)
|
| 406 |
+
ttoks_ = F.pad(ttoks_, (0, self.ttoks_len - len(ttoks_)), value=self.tokenizer.eot).unsqueeze(0)
|
| 407 |
+
ttoks.append(ttoks_)
|
| 408 |
+
ttoks = torch.cat(ttoks, dim=0)
|
| 409 |
+
toks = torch.zeros((len(ttoks),N), dtype=torch.long, device=dev)
|
| 410 |
+
it = range(N)
|
| 411 |
+
if show_progress_bar: it = progress_bar(it)
|
| 412 |
+
for i in it:
|
| 413 |
+
p, _ = self(ttoks, toks[:,:i], loss=None)
|
| 414 |
+
last_p = p[:,-1]
|
| 415 |
+
if top_k:
|
| 416 |
+
last_p[last_p < torch.topk(last_p, top_k).values[:,-1,None]] = -torch.inf
|
| 417 |
+
tok = torch.multinomial((last_p / float(T)).softmax(-1), 1)
|
| 418 |
+
toks[:,i] = tok[:,0]
|
| 419 |
+
if (toks[:,i] == self.stoks_codes-1).all(): return toks[:,:i]
|
| 420 |
+
return toks
|
| 421 |
+
|
| 422 |
+
# %% ../nbs/5B. Text to semantic token modeling.ipynb 19
|
| 423 |
+
def _make_model(size:str, tunables:Tunables=Tunables(), dataset=None, **kwargs):
|
| 424 |
+
kwargs = dict(stoks_len = dataset.stoks_len, ttoks_len = dataset.ttoks_len, tunables=tunables, **kwargs)
|
| 425 |
+
if 'stoks_codes' not in kwargs: kwargs['stoks_codes'] = dataset.stoks_codes
|
| 426 |
+
if size == 'micro':
|
| 427 |
+
return TSARTransformer(depth=2, n_head=3, ffn_mult=1, **kwargs)
|
| 428 |
+
if size == 'tiny':
|
| 429 |
+
return TSARTransformer(depth=4, n_head=6, **kwargs)
|
| 430 |
+
if size == 'base':
|
| 431 |
+
return TSARTransformer(depth=6, n_head=8, **kwargs)
|
| 432 |
+
if size == 'small':
|
| 433 |
+
return TSARTransformer(depth=12, n_head=16, **kwargs)
|
| 434 |
+
|
| 435 |
+
def make_model(size:str, frozen_embeddings_model:str=None, tunables:Tunables=Tunables(), dataset:torch.utils.data.Dataset=None):
|
| 436 |
+
if frozen_embeddings_model:
|
| 437 |
+
vqmodel = vq_stoks.RQBottleneckTransformer.load_model(frozen_embeddings_model)
|
| 438 |
+
model = _make_model(size, tunables, dataset, stoks_codes=vqmodel.vq_codes+1, stoks_width=vqmodel.rq.layers[0]._codebook.embed[0].shape[-1])
|
| 439 |
+
model.load_frozen_semantic_embeddings(vqmodel)
|
| 440 |
+
else:
|
| 441 |
+
model = _make_model(size, quantizers, tunables, dataset)
|
| 442 |
+
return model
|
whisperspeech/t2s_up_wds_mlang_enclm.py
ADDED
|
@@ -0,0 +1,519 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/5B. Multi-lang text to semantic token modeling.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = ['load_dataset', 'rand', 'Tunables', 'T2SEmbedding', 'Encoder', 'TSARTransformer', 'make_model']
|
| 5 |
+
|
| 6 |
+
# %% ../nbs/5B. Multi-lang text to semantic token modeling.ipynb 1
|
| 7 |
+
import dataclasses
|
| 8 |
+
import random
|
| 9 |
+
import math
|
| 10 |
+
import itertools
|
| 11 |
+
import torch
|
| 12 |
+
import torch.nn as nn
|
| 13 |
+
import torch.nn.functional as F
|
| 14 |
+
from torch.profiler import record_function
|
| 15 |
+
|
| 16 |
+
from huggingface_hub import hf_hub_download
|
| 17 |
+
from fastcore.basics import store_attr
|
| 18 |
+
from fastprogress import progress_bar
|
| 19 |
+
|
| 20 |
+
from pathlib import Path
|
| 21 |
+
|
| 22 |
+
# %% ../nbs/5B. Multi-lang text to semantic token modeling.ipynb 2
|
| 23 |
+
from whisperspeech.modules import *
|
| 24 |
+
from whisperspeech import languages
|
| 25 |
+
|
| 26 |
+
# %% ../nbs/5B. Multi-lang text to semantic token modeling.ipynb 6
|
| 27 |
+
import re
|
| 28 |
+
|
| 29 |
+
class CharTokenizer:
|
| 30 |
+
"""Trivial tokenizer β just use UTF-8 bytes"""
|
| 31 |
+
eot = 0
|
| 32 |
+
|
| 33 |
+
def encode(self, txt):
|
| 34 |
+
return list(bytes(txt.strip(), 'utf-8'))
|
| 35 |
+
|
| 36 |
+
def decode(self, tokens):
|
| 37 |
+
return bytes(tokens).decode('utf-8')
|
| 38 |
+
|
| 39 |
+
def tokenizer(ikey, okey, length):
|
| 40 |
+
"""Tokenizes a transcript"""
|
| 41 |
+
tok = CharTokenizer()
|
| 42 |
+
def _tokenizer(samples):
|
| 43 |
+
for s in samples:
|
| 44 |
+
toks = torch.tensor(tok.encode(s[ikey]))
|
| 45 |
+
s[okey] = F.pad(toks, (0, length - toks.shape[-1]), value=tok.eot)
|
| 46 |
+
yield s
|
| 47 |
+
return _tokenizer
|
| 48 |
+
|
| 49 |
+
def ar_padder(ikey, okey, length, pad_token):
|
| 50 |
+
"""Pads the tokens for autoregresive training"""
|
| 51 |
+
import numpy as np
|
| 52 |
+
|
| 53 |
+
def _ar_padder(samples):
|
| 54 |
+
for s in samples:
|
| 55 |
+
toks = s[ikey]
|
| 56 |
+
if isinstance(toks, (list, np.ndarray)): toks = torch.tensor(toks)
|
| 57 |
+
toks = toks.to(torch.long)
|
| 58 |
+
s['in_' +okey] = F.pad(toks, (1, length - toks.shape[-1] - 1), value=pad_token)
|
| 59 |
+
s['out_'+okey] = F.pad(toks, (0, length - toks.shape[-1]), value=pad_token)
|
| 60 |
+
yield s
|
| 61 |
+
return _ar_padder
|
| 62 |
+
|
| 63 |
+
def char_per_seconder(txt_key, stoks_key, cps_key, stoks_per_second=25):
|
| 64 |
+
"""Adds the characters per second metric to the input data"""
|
| 65 |
+
def _char_per_seconder(samples):
|
| 66 |
+
for s in samples:
|
| 67 |
+
secs = s[stoks_key].shape[-1] / stoks_per_second
|
| 68 |
+
s[cps_key] = len(s[txt_key]) / secs
|
| 69 |
+
yield s
|
| 70 |
+
return _char_per_seconder
|
| 71 |
+
|
| 72 |
+
# %% ../nbs/5B. Multi-lang text to semantic token modeling.ipynb 7
|
| 73 |
+
def load_dataset(
|
| 74 |
+
txt_shard_spec:str, # transcription webdataset shards
|
| 75 |
+
stoks_shard_dir:str, # stoks webdataset base dir
|
| 76 |
+
samples:int, # samples per epoch
|
| 77 |
+
txt_kind:str='small.en-txt',
|
| 78 |
+
vq_codes:int=4096,
|
| 79 |
+
language:str='en',
|
| 80 |
+
weight:float=1,
|
| 81 |
+
validation:bool=False,
|
| 82 |
+
exclude_files:str=None,
|
| 83 |
+
):
|
| 84 |
+
import webdataset as wds
|
| 85 |
+
from whisperspeech import utils
|
| 86 |
+
|
| 87 |
+
shards = utils.shard_glob(txt_shard_spec)
|
| 88 |
+
excludes = {x for file in exclude_files.split() for x in utils.readlines(file)} if exclude_files else set()
|
| 89 |
+
|
| 90 |
+
language = languages.to_id(language)
|
| 91 |
+
|
| 92 |
+
def set_language(x):
|
| 93 |
+
x['language'] = language
|
| 94 |
+
return x
|
| 95 |
+
|
| 96 |
+
same_on_all_nodes = lambda urls: urls # will only be used for validation
|
| 97 |
+
ds = wds.WebDataset(shards, resampled=not validation, nodesplitter=same_on_all_nodes).compose(
|
| 98 |
+
wds.decode(),
|
| 99 |
+
utils.merge_in(utils.derived_dataset('eqvad-stoks', base=txt_kind, suffix='', dir=stoks_shard_dir)),
|
| 100 |
+
# discard validation samples, select samples > .5s
|
| 101 |
+
wds.select(lambda s: s['__key__'] not in excludes and s['stoks.npy'].shape[-1] > 12),
|
| 102 |
+
tokenizer('txt', 'ttoks', length=550),
|
| 103 |
+
ar_padder('stoks.npy', 'stoks', length=750, pad_token=vq_codes-1),
|
| 104 |
+
ar_padder('ttoks', 'ttoks', length=550, pad_token=CharTokenizer.eot),
|
| 105 |
+
char_per_seconder('txt', 'stoks.npy', 'cps', stoks_per_second=25),
|
| 106 |
+
wds.map(set_language),
|
| 107 |
+
wds.to_tuple('in_ttoks', 'out_ttoks', 'language', 'cps', 'in_stoks', 'out_stoks'),
|
| 108 |
+
wds.shuffle(20000, initial=20000),
|
| 109 |
+
wds.batched(64)
|
| 110 |
+
)
|
| 111 |
+
if validation:
|
| 112 |
+
ds = ds.slice(samples // 64)
|
| 113 |
+
ds.total_samples = samples
|
| 114 |
+
ds.stoks_len = 750
|
| 115 |
+
ds.stoks_codes = vq_codes
|
| 116 |
+
ds.ttoks_len = 550
|
| 117 |
+
ds.weight = weight
|
| 118 |
+
|
| 119 |
+
return ds
|
| 120 |
+
|
| 121 |
+
# %% ../nbs/5B. Multi-lang text to semantic token modeling.ipynb 14
|
| 122 |
+
def rand(start, end):
|
| 123 |
+
return random.random() * (end - start) + start
|
| 124 |
+
|
| 125 |
+
@dataclasses.dataclass
|
| 126 |
+
class Tunables:
|
| 127 |
+
init_std :float = 1
|
| 128 |
+
embeddings_std :float = .01
|
| 129 |
+
embeddings_lr_scale: float = 5
|
| 130 |
+
embedding_projector_lr_scale: float = 2.5
|
| 131 |
+
output_mult :float = .35
|
| 132 |
+
query_mult :float = 1
|
| 133 |
+
encoder_depth_ratio :float = 0.25
|
| 134 |
+
eot_dropout_p :float = .5
|
| 135 |
+
cps_input: bool = True
|
| 136 |
+
cps_bins: int = 32
|
| 137 |
+
|
| 138 |
+
lr0 :float = 1.5e-3
|
| 139 |
+
clip_gradient_norm :float = .2
|
| 140 |
+
weight_decay :float = 1e-1
|
| 141 |
+
warmup_steps :float = 4000
|
| 142 |
+
|
| 143 |
+
random :bool = False
|
| 144 |
+
|
| 145 |
+
def __post_init__(self):
|
| 146 |
+
# randomize the hyperparams if requested
|
| 147 |
+
if self.random:
|
| 148 |
+
self.init_std = 10**rand(-1,1)
|
| 149 |
+
self.embeddings_std = 10**rand(-3,-.7)
|
| 150 |
+
self.embeddings_lr_scale = rand(2,6)
|
| 151 |
+
self.output_mult = rand(0.25,0.65)
|
| 152 |
+
self.query_mult = 2**rand(-2,3)
|
| 153 |
+
self.encoder_depth_ratio = 0.25
|
| 154 |
+
|
| 155 |
+
self.lr0 = rand(1,5)*1e-3
|
| 156 |
+
self.clip_gradient_norm = 10**rand(-3,0)
|
| 157 |
+
self.warmup_steps = 100*(10**rand(1,1.85))
|
| 158 |
+
|
| 159 |
+
# %% ../nbs/5B. Multi-lang text to semantic token modeling.ipynb 15
|
| 160 |
+
class T2SEmbedding(nn.Module):
|
| 161 |
+
def __init__(self, length=1500, codes=1024, width=384, pos_embs=None, stoks_width=384):
|
| 162 |
+
super().__init__()
|
| 163 |
+
self.embedding = FlexEmbeddings(codes, width, special_codes=1, frozen_width=stoks_width)
|
| 164 |
+
if pos_embs is None: pos_embs = sinusoids(length, width)
|
| 165 |
+
self.register_buffer("positional_embedding", pos_embs)
|
| 166 |
+
|
| 167 |
+
def forward(self, Stoks, xenc, cps=None, offset=0):
|
| 168 |
+
Sembs = self.embedding(Stoks)
|
| 169 |
+
xin = (Sembs + self.positional_embedding[offset : offset + Sembs.shape[1]]).to(xenc.dtype)
|
| 170 |
+
if cps is not None: xin = xin + cps
|
| 171 |
+
return xin, offset
|
| 172 |
+
|
| 173 |
+
# %% ../nbs/5B. Multi-lang text to semantic token modeling.ipynb 16
|
| 174 |
+
class Encoder(nn.Module):
|
| 175 |
+
def __init__(self, depth=6, width=384, n_head=6, length=1500, codes=1024, emb_width=384, ffn_mult=4, pos_embs=None, tunables=Tunables()):
|
| 176 |
+
super().__init__()
|
| 177 |
+
self.emb_width = emb_width
|
| 178 |
+
|
| 179 |
+
self.embedding = FlexEmbeddings(codes, width, frozen_width=emb_width)
|
| 180 |
+
|
| 181 |
+
if pos_embs is None: pos_embs = sinusoids(length, width)
|
| 182 |
+
self.register_buffer("positional_embedding", pos_embs)
|
| 183 |
+
|
| 184 |
+
self.layers = nn.ModuleList([
|
| 185 |
+
ResidualAttentionBlock(width, n_head,
|
| 186 |
+
qk_scale=tunables.query_mult*8/math.sqrt(width/n_head), ffn_mult=ffn_mult) for _ in range(depth)
|
| 187 |
+
])
|
| 188 |
+
|
| 189 |
+
self.ln_post = LayerNorm(width)
|
| 190 |
+
|
| 191 |
+
mask = torch.empty(length, length).fill_(-torch.inf).triu_(1)
|
| 192 |
+
self.register_buffer("mask", mask, persistent=False)
|
| 193 |
+
|
| 194 |
+
def forward(self, Stoks, positions, lang_emb=None):
|
| 195 |
+
xin = self.embedding(Stoks)
|
| 196 |
+
|
| 197 |
+
if lang_emb is not None: xin += lang_emb
|
| 198 |
+
|
| 199 |
+
# assert xin.shape[1:] == self.positional_embedding.shape, "incorrect semantic token shape"
|
| 200 |
+
x = (xin +
|
| 201 |
+
self.positional_embedding[positions]).to(xin.dtype)
|
| 202 |
+
|
| 203 |
+
for l in self.layers: x = l(x, positions, causal=False, mask=self.mask)
|
| 204 |
+
|
| 205 |
+
return self.ln_post(x)
|
| 206 |
+
|
| 207 |
+
# %% ../nbs/5B. Multi-lang text to semantic token modeling.ipynb 17
|
| 208 |
+
class TSARTransformer(nn.Module):
|
| 209 |
+
def __init__(self, depth=6, n_head=6, head_width=64, ffn_mult=4,
|
| 210 |
+
ttoks_len=200, ttoks_codes=256, ttoks_width=None,
|
| 211 |
+
stoks_len=1500, stoks_codes=1024, stoks_width=None,
|
| 212 |
+
tunables=Tunables()):
|
| 213 |
+
super().__init__()
|
| 214 |
+
store_attr("depth,n_head,head_width,ffn_mult,stoks_width,ttoks_width,ttoks_len,stoks_len,ttoks_codes,stoks_codes")
|
| 215 |
+
|
| 216 |
+
width = n_head * head_width
|
| 217 |
+
self.width = width
|
| 218 |
+
self.base_width = 3 * head_width
|
| 219 |
+
self.tunables = tunables
|
| 220 |
+
if self.stoks_width is None: self.stoks_width = self.width
|
| 221 |
+
if self.ttoks_width is None: self.ttoks_width = self.width
|
| 222 |
+
|
| 223 |
+
self.lang_embeddings = nn.Embedding(len(languages.languages), width)
|
| 224 |
+
if tunables.cps_input:
|
| 225 |
+
self.cps_embeddings = nn.Embedding(tunables.cps_bins, self.width)
|
| 226 |
+
else:
|
| 227 |
+
self.cps_embeddings = None
|
| 228 |
+
|
| 229 |
+
encoder_depth = int(depth * 2 * tunables.encoder_depth_ratio)
|
| 230 |
+
decoder_depth = depth * 2 - encoder_depth
|
| 231 |
+
tformer_args = dict(width=width, n_head=n_head, ffn_mult=ffn_mult, tunables=tunables)
|
| 232 |
+
self.encoder = Encoder(length=ttoks_len, codes=ttoks_codes, emb_width=self.ttoks_width, depth=encoder_depth, **tformer_args)
|
| 233 |
+
self.embeddings = T2SEmbedding(length=stoks_len, codes=stoks_codes, width=width, stoks_width=self.stoks_width)
|
| 234 |
+
|
| 235 |
+
self.decoder = BaseDecoder(
|
| 236 |
+
length=stoks_len,
|
| 237 |
+
depth=decoder_depth,
|
| 238 |
+
qk_scale=tunables.query_mult*8/math.sqrt(width/n_head),
|
| 239 |
+
width=width, n_head=n_head, ffn_mult=ffn_mult,
|
| 240 |
+
)
|
| 241 |
+
self.tokenizer = None
|
| 242 |
+
|
| 243 |
+
self.apply(self.init_transformer)
|
| 244 |
+
|
| 245 |
+
def load_frozen_semantic_embeddings(self, vqmodel):
|
| 246 |
+
self.embeddings.embedding.set_frozen_embeddings(vqmodel.rq.layers[0]._codebook.embed[0])
|
| 247 |
+
|
| 248 |
+
def setup(self, device):
|
| 249 |
+
pass
|
| 250 |
+
|
| 251 |
+
def init_transformer(self, m):
|
| 252 |
+
if isinstance(m, LinearHead):
|
| 253 |
+
m.no_weight_decay = True
|
| 254 |
+
torch.nn.init.constant_(m.weight, 0)
|
| 255 |
+
elif isinstance(m, QueryHead):
|
| 256 |
+
m.lr_scale = 1/(m.weight.shape[1] / self.base_width)
|
| 257 |
+
torch.nn.init.constant_(m.weight, 0)
|
| 258 |
+
elif isinstance(m, nn.Embedding):
|
| 259 |
+
m.no_weight_decay = True
|
| 260 |
+
m.lr_scale = self.tunables.embeddings_lr_scale
|
| 261 |
+
std = self.tunables.embeddings_std
|
| 262 |
+
torch.nn.init.trunc_normal_(m.weight, std=std, a=-3*std, b=3*std)
|
| 263 |
+
elif isinstance(m, EmbeddingProjector):
|
| 264 |
+
m.lr_scale = self.tunables.embedding_projector_lr_scale
|
| 265 |
+
std = self.tunables.init_std
|
| 266 |
+
torch.nn.init.trunc_normal_(m.weight, std=std, a=-3*std, b=3*std)
|
| 267 |
+
elif isinstance(m, nn.Linear):
|
| 268 |
+
m.lr_scale = 1/(m.weight.shape[1] / self.base_width)
|
| 269 |
+
std = self.tunables.init_std / m.weight.shape[1]
|
| 270 |
+
torch.nn.init.trunc_normal_(m.weight, std=std, a=-3*std, b=3*std)
|
| 271 |
+
if m.bias is not None:
|
| 272 |
+
torch.nn.init.trunc_normal_(m.bias, std=std, a=-3*std, b=3*std)
|
| 273 |
+
elif isinstance(m, nn.LayerNorm):
|
| 274 |
+
m.no_weight_decay = True
|
| 275 |
+
torch.nn.init.constant_(m.bias, 0)
|
| 276 |
+
torch.nn.init.constant_(m.weight, 1)
|
| 277 |
+
|
| 278 |
+
def _embed_cps(self, cpss):
|
| 279 |
+
if self.cps_embeddings is None: return None
|
| 280 |
+
|
| 281 |
+
cps_bin = (cpss / 20 * self.tunables.cps_bins).to(torch.long)
|
| 282 |
+
cps_bin[cps_bin >= self.tunables.cps_bins] = self.tunables.cps_bins-1
|
| 283 |
+
return self.cps_embeddings(cps_bin).unsqueeze(1)
|
| 284 |
+
|
| 285 |
+
def run_encoder(self, in_ttoks, languages, cpss):
|
| 286 |
+
if len(languages.shape) != 3: lang_embs = self.lang_embeddings(languages)
|
| 287 |
+
else: lang_embs = languages
|
| 288 |
+
if len(lang_embs.shape) == 2: lang_embs = lang_embs.unsqueeze(1)
|
| 289 |
+
|
| 290 |
+
cps_emb = self._embed_cps(cpss)
|
| 291 |
+
|
| 292 |
+
with record_function("encoder"):
|
| 293 |
+
positions = torch.arange(0, in_ttoks.shape[1], device=in_ttoks.device)
|
| 294 |
+
xenc = self.encoder(in_ttoks.to(torch.long), positions, lang_emb=lang_embs)
|
| 295 |
+
|
| 296 |
+
return xenc, positions, cps_emb
|
| 297 |
+
|
| 298 |
+
def forward(self, in_ttoks, out_ttoks, languages, cpss, in_stoks, in_stoks_positions, out_stoks=None, loss=True, offset=None, xenc=None, xenc_positions=None, cps_emb=None):
|
| 299 |
+
if xenc is None:
|
| 300 |
+
xenc, cps_emb = self.run_encoder(in_ttoks, languages, cpss)
|
| 301 |
+
|
| 302 |
+
with record_function("decoder"):
|
| 303 |
+
x = (self.embeddings.embedding(in_stoks) +
|
| 304 |
+
self.embeddings.positional_embedding[in_stoks_positions] +
|
| 305 |
+
cps_emb).to(xenc[0].dtype)
|
| 306 |
+
x = self.decoder(x, in_stoks_positions, xenc, xenc_positions)
|
| 307 |
+
logits = self.embeddings.embedding.unembed(x)
|
| 308 |
+
logits = logits * self.tunables.output_mult / (self.width / self.base_width)
|
| 309 |
+
|
| 310 |
+
if loss is not None:
|
| 311 |
+
enc_logits = self.encoder.embedding.unembed(xenc[0])
|
| 312 |
+
enc_logits = enc_logits * self.tunables.output_mult / (self.width / self.base_width)
|
| 313 |
+
with record_function("loss"):
|
| 314 |
+
loss = F.cross_entropy(logits.transpose(-1,-2), out_stoks)
|
| 315 |
+
if self.training:
|
| 316 |
+
loss += 0.1 * F.cross_entropy(enc_logits.transpose(-1,-2), out_ttoks)
|
| 317 |
+
|
| 318 |
+
return logits, loss
|
| 319 |
+
|
| 320 |
+
#
|
| 321 |
+
# inference
|
| 322 |
+
#
|
| 323 |
+
@classmethod
|
| 324 |
+
def load_model(cls, ref="collabora/whisperspeech:t2s-small-en+pl.model",
|
| 325 |
+
repo_id=None, filename=None, local_filename=None):
|
| 326 |
+
if repo_id is None and filename is None and local_filename is None:
|
| 327 |
+
if ":" in ref:
|
| 328 |
+
repo_id, filename = ref.split(":", 1)
|
| 329 |
+
else:
|
| 330 |
+
local_filename = ref
|
| 331 |
+
if not local_filename:
|
| 332 |
+
local_filename = hf_hub_download(repo_id=repo_id, filename=filename)
|
| 333 |
+
spec = torch.load(local_filename)
|
| 334 |
+
model = cls(**spec['config'], tunables=Tunables(**spec['tunables']))
|
| 335 |
+
model.load_state_dict(spec['state_dict'])
|
| 336 |
+
model.eval()
|
| 337 |
+
return model
|
| 338 |
+
|
| 339 |
+
def load_checkpoint(self, local_filename):
|
| 340 |
+
spec = torch.load(local_filename, map_location='cpu')
|
| 341 |
+
assert 'pytorch-lightning_version' in spec, 'not a valid PyTorch Lightning checkpoint'
|
| 342 |
+
state_dict = {k.replace('model.', ''):v
|
| 343 |
+
for k,v in spec['state_dict'].items()}
|
| 344 |
+
self.load_state_dict(state_dict)
|
| 345 |
+
return self
|
| 346 |
+
|
| 347 |
+
def save_model(self, fname):
|
| 348 |
+
torch.save(dict(config = self.__stored_args__,
|
| 349 |
+
tunables = dataclasses.asdict(self.tunables),
|
| 350 |
+
state_dict = self.state_dict()), fname)
|
| 351 |
+
|
| 352 |
+
def ensure_tokenizer(self):
|
| 353 |
+
assert not self.training
|
| 354 |
+
if self.tokenizer is None: self.tokenizer = CharTokenizer()
|
| 355 |
+
|
| 356 |
+
def switch_dtypes(self, dtype=torch.float16):
|
| 357 |
+
self.dtype = dtype
|
| 358 |
+
for n,m in self.named_modules():
|
| 359 |
+
# convert every leaf layer apart from the LayerNorms
|
| 360 |
+
if isinstance(m, (nn.Linear, nn.Embedding)):
|
| 361 |
+
m.to(dtype)
|
| 362 |
+
# take care of buffers ([kv]_cache, masks) that are not in the leaf layers
|
| 363 |
+
for bn,b in m.named_buffers(recurse=False):
|
| 364 |
+
setattr(m,bn,b.to(dtype))
|
| 365 |
+
|
| 366 |
+
def optimize(self, max_batch_size=1, dtype=torch.float16, torch_compile=True):
|
| 367 |
+
for emb in [self.embeddings.embedding, self.embeddings.embedding]:
|
| 368 |
+
emb.convert_for_eval()
|
| 369 |
+
for l in self.encoder.layers:
|
| 370 |
+
l.attn.convert_for_eval()
|
| 371 |
+
for l in self.decoder.layers:
|
| 372 |
+
l.attn.convert_for_eval()
|
| 373 |
+
l.cross_attn.convert_for_eval()
|
| 374 |
+
l.setup_kv_cache(max_batch_size, self.stoks_len, self.ttoks_len)
|
| 375 |
+
self.switch_dtypes(dtype)
|
| 376 |
+
if torch_compile:
|
| 377 |
+
self.generate_next = torch.compile(self.generate_next, mode="reduce-overhead", fullgraph=True)
|
| 378 |
+
|
| 379 |
+
@property
|
| 380 |
+
def device(self):
|
| 381 |
+
return next(self.parameters()).device
|
| 382 |
+
|
| 383 |
+
# from https://github.com/pytorch-labs/gpt-fast/blob/main/generate.py
|
| 384 |
+
def multinomial_sample_one_no_sync(self, probs_sort): # Does multinomial sampling without a cuda synchronization
|
| 385 |
+
q = torch.empty_like(probs_sort).exponential_(1)
|
| 386 |
+
return torch.argmax(probs_sort / q, dim=-1, keepdim=True).to(dtype=torch.int)
|
| 387 |
+
|
| 388 |
+
def logits_to_probs(self, logits, T=1.0, top_k=None):
|
| 389 |
+
logits = logits / max(T, 1e-5)
|
| 390 |
+
|
| 391 |
+
logits[self.embeddings.embedding.codes:] = -torch.inf
|
| 392 |
+
if top_k is not None:
|
| 393 |
+
v, _ = torch.topk(logits, min(top_k, logits.size(-1)))
|
| 394 |
+
pivot = v.select(-1, -1).unsqueeze(-1)
|
| 395 |
+
logits = torch.where(logits < pivot, -float("Inf"), logits)
|
| 396 |
+
|
| 397 |
+
probs = torch.nn.functional.softmax(logits, dim=-1)
|
| 398 |
+
return probs
|
| 399 |
+
|
| 400 |
+
def sample(self, logits, T=1.0, top_k=None):
|
| 401 |
+
probs = self.logits_to_probs(logits[0,-1], T, top_k)
|
| 402 |
+
idx_next = self.multinomial_sample_one_no_sync(probs)
|
| 403 |
+
return idx_next
|
| 404 |
+
|
| 405 |
+
def generate_one(self, toks, toks_positions, cps_emb, xenc, xenc_positions, T, top_k):
|
| 406 |
+
probs, _ = self(None, None, None, None, toks, toks_positions, loss=None, xenc=xenc, xenc_positions=xenc_positions, cps_emb=cps_emb)
|
| 407 |
+
return self.sample(probs, T, top_k)
|
| 408 |
+
|
| 409 |
+
def generate_next(self, *args, **kwargs):
|
| 410 |
+
return self.generate_one(*args, **kwargs)
|
| 411 |
+
|
| 412 |
+
@torch.no_grad()
|
| 413 |
+
def prep(self, txt, cps=15, lang="en"):
|
| 414 |
+
dev = self.device
|
| 415 |
+
ttoks = torch.tensor(self.tokenizer.encode(txt), device=dev)
|
| 416 |
+
ttoks = F.pad(ttoks, (0, self.ttoks_len - len(ttoks)), value=self.tokenizer.eot).unsqueeze(0)
|
| 417 |
+
cpss = torch.tensor([cps], device=dev)
|
| 418 |
+
langs = torch.tensor([languages.to_id(lang)], device=dev)
|
| 419 |
+
return ttoks, cpss, langs
|
| 420 |
+
|
| 421 |
+
@torch.no_grad()
|
| 422 |
+
def generate(self, txt, cps=15, lang="en", N=None, T=0.7, top_k=None, step=None, show_progress_bar=True):
|
| 423 |
+
self.ensure_tokenizer()
|
| 424 |
+
N = N or self.stoks_len
|
| 425 |
+
dev = self.device
|
| 426 |
+
ttoks = []
|
| 427 |
+
langs = []
|
| 428 |
+
if isinstance(lang, list):
|
| 429 |
+
lang0 = lang[0]
|
| 430 |
+
assert isinstance(txt, list), "lang and txt have to be both lists or strings"
|
| 431 |
+
for txt, lang in zip(txt, lang):
|
| 432 |
+
tt = self.tokenizer.encode(txt)
|
| 433 |
+
ttoks += tt
|
| 434 |
+
langs += [languages.to_id(lang)] * len(tt)
|
| 435 |
+
elif isinstance(lang, torch.Tensor):
|
| 436 |
+
langs = lang
|
| 437 |
+
ttoks = self.tokenizer.encode(txt)
|
| 438 |
+
else:
|
| 439 |
+
lang0 = lang
|
| 440 |
+
ttoks = self.tokenizer.encode(txt)
|
| 441 |
+
langs = torch.tensor([languages.to_id(lang)], device=dev).unsqueeze(0)
|
| 442 |
+
ttoks = torch.tensor(ttoks, device=dev)
|
| 443 |
+
ttoks = F.pad(ttoks, (1, self.ttoks_len - len(ttoks) - 1), value=self.tokenizer.eot).unsqueeze(0)
|
| 444 |
+
cpss = torch.tensor([cps], device=dev)
|
| 445 |
+
if not isinstance(langs, torch.Tensor):
|
| 446 |
+
langs = torch.tensor(langs, device=dev)
|
| 447 |
+
langs = F.pad(langs, (1, self.ttoks_len - len(langs) - 1), value=languages.to_id(lang0)).unsqueeze(0)
|
| 448 |
+
it = range(0,N-1)
|
| 449 |
+
if show_progress_bar: it = progress_bar(it)
|
| 450 |
+
|
| 451 |
+
toks = torch.zeros((1,N), dtype=torch.long, device=dev)
|
| 452 |
+
toks[:,0] = self.stoks_codes-1
|
| 453 |
+
toks_positions = torch.arange(N, device=dev)
|
| 454 |
+
with record_function("encode"):
|
| 455 |
+
xenc, xenc_positions, cps_emb = self.run_encoder(ttoks, langs, cpss)
|
| 456 |
+
toks_positions = torch.arange(N+1, device=dev)
|
| 457 |
+
# contrary to S2A this model works without prefill and is actually a tiny bit faster
|
| 458 |
+
# with record_function("prefill"):
|
| 459 |
+
# toks[0,1] = self.generate_one(toks[:,:1], toks_positions[:1], cps_emb, xenc, xenc_positions, T, top_k)
|
| 460 |
+
with torch.backends.cuda.sdp_kernel(enable_flash=False, enable_mem_efficient=False, enable_math=True):
|
| 461 |
+
for i in it:
|
| 462 |
+
toks[0,i+1] = self.generate_next(toks[:,i:i+1], toks_positions[i:i+1], cps_emb, xenc, xenc_positions, T, top_k)
|
| 463 |
+
if i % 25 == 0 and toks[0,i+1] == self.stoks_codes-1: return toks[0,:i+1]
|
| 464 |
+
|
| 465 |
+
# for profiling, debugging or early exit
|
| 466 |
+
if step is not None: step()
|
| 467 |
+
return toks[0,:]
|
| 468 |
+
|
| 469 |
+
@torch.no_grad()
|
| 470 |
+
def generate_batch(self, txts, N=None, T=1.1, top_k=7, show_progress_bar=True):
|
| 471 |
+
self.ensure_tokenizer()
|
| 472 |
+
N = self.stoks_len
|
| 473 |
+
dev = self.device
|
| 474 |
+
ttoks = []
|
| 475 |
+
for txt in txts:
|
| 476 |
+
ttoks_ = torch.tensor(self.tokenizer.encode(txt), device=dev)
|
| 477 |
+
ttoks_ = F.pad(ttoks_, (0, self.ttoks_len - len(ttoks_)), value=self.tokenizer.eot).unsqueeze(0)
|
| 478 |
+
ttoks.append(ttoks_)
|
| 479 |
+
ttoks = torch.cat(ttoks, dim=0)
|
| 480 |
+
toks = torch.zeros((len(ttoks),N), dtype=torch.long, device=dev)
|
| 481 |
+
it = range(N)
|
| 482 |
+
if show_progress_bar: it = progress_bar(it)
|
| 483 |
+
for i in it:
|
| 484 |
+
p, _ = self(ttoks, toks[:,:i], loss=None)
|
| 485 |
+
last_p = p[:,-1]
|
| 486 |
+
if top_k:
|
| 487 |
+
last_p[last_p < torch.topk(last_p, top_k).values[:,-1,None]] = -torch.inf
|
| 488 |
+
tok = torch.multinomial((last_p / float(T)).softmax(-1), 1)
|
| 489 |
+
toks[:,i] = tok[:,0]
|
| 490 |
+
if (toks[:,i] == self.stoks_codes-1).all(): return toks[:,:i]
|
| 491 |
+
return toks
|
| 492 |
+
|
| 493 |
+
# %% ../nbs/5B. Multi-lang text to semantic token modeling.ipynb 18
|
| 494 |
+
def _make_model(size:str, tunables:Tunables=Tunables(), dataset=None, **kwargs):
|
| 495 |
+
kwargs = dict(stoks_len = dataset.stoks_len, ttoks_len = dataset.ttoks_len, tunables=tunables, **kwargs)
|
| 496 |
+
if 'stoks_codes' not in kwargs: kwargs['stoks_codes'] = dataset.stoks_codes
|
| 497 |
+
if size == 'micro':
|
| 498 |
+
return TSARTransformer(depth=2, n_head=3, ffn_mult=1, **kwargs)
|
| 499 |
+
if size == 'tiny':
|
| 500 |
+
return TSARTransformer(depth=4, n_head=6, **kwargs)
|
| 501 |
+
if size == 'base':
|
| 502 |
+
return TSARTransformer(depth=6, n_head=8, **kwargs)
|
| 503 |
+
if size == 'small':
|
| 504 |
+
return TSARTransformer(depth=12, n_head=12, **kwargs)
|
| 505 |
+
if size == 'small+':
|
| 506 |
+
return TSARTransformer(depth=12, n_head=16, **kwargs)
|
| 507 |
+
if size == 'medium':
|
| 508 |
+
return TSARTransformer(depth=24, n_head=16, **kwargs)
|
| 509 |
+
|
| 510 |
+
def make_model(size:str, frozen_embeddings_model:str=None, tunables:Tunables=Tunables(), dataset:torch.utils.data.Dataset=None):
|
| 511 |
+
from whisperspeech import vq_stoks
|
| 512 |
+
|
| 513 |
+
if frozen_embeddings_model:
|
| 514 |
+
vqmodel = vq_stoks.RQBottleneckTransformer.load_model(frozen_embeddings_model)
|
| 515 |
+
model = _make_model(size, tunables, dataset, stoks_codes=vqmodel.vq_codes+1, stoks_width=vqmodel.rq.layers[0]._codebook.embed[0].shape[-1])
|
| 516 |
+
model.load_frozen_semantic_embeddings(vqmodel)
|
| 517 |
+
else:
|
| 518 |
+
model = _make_model(size, tunables, dataset, mode=mode)
|
| 519 |
+
return model
|
whisperspeech/train.py
ADDED
|
@@ -0,0 +1,271 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/B1. Training.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = ['SimpleVisual', 'validate', 'train']
|
| 5 |
+
|
| 6 |
+
# %% ../nbs/B1. Training.ipynb 2
|
| 7 |
+
import io
|
| 8 |
+
import time
|
| 9 |
+
import random
|
| 10 |
+
from pathlib import Path
|
| 11 |
+
|
| 12 |
+
from fastprogress import progress_bar, master_bar
|
| 13 |
+
import fastprogress
|
| 14 |
+
|
| 15 |
+
import numpy as np
|
| 16 |
+
import pylab as plt
|
| 17 |
+
import math
|
| 18 |
+
|
| 19 |
+
import IPython
|
| 20 |
+
|
| 21 |
+
import torch
|
| 22 |
+
import torch.nn as nn
|
| 23 |
+
from torch.utils.data.dataloader import DataLoader
|
| 24 |
+
from torch.profiler import record_function
|
| 25 |
+
|
| 26 |
+
import webdataset as wds
|
| 27 |
+
|
| 28 |
+
torch.backends.cudnn.benchmark = True
|
| 29 |
+
torch.backends.cudnn.enabled = True
|
| 30 |
+
torch.backends.cuda.matmul.allow_tf32 = True
|
| 31 |
+
torch.set_float32_matmul_precision('medium')
|
| 32 |
+
|
| 33 |
+
# %% ../nbs/B1. Training.ipynb 3
|
| 34 |
+
class SimpleVisual:
|
| 35 |
+
def __init__ (self, model, masterbar, total_steps):
|
| 36 |
+
self.model = model
|
| 37 |
+
self.masterbar = masterbar
|
| 38 |
+
self.total_steps = total_steps
|
| 39 |
+
self.epochs = total_steps // masterbar.main_bar.total
|
| 40 |
+
|
| 41 |
+
gs = plt.GridSpec(2, 1, height_ratios=[3,1])
|
| 42 |
+
graph_fig = plt.figure(figsize=(10,6))
|
| 43 |
+
self.graph_fig = graph_fig
|
| 44 |
+
self.loss_p = graph_fig.add_subplot(gs[0])
|
| 45 |
+
self.lr_p = graph_fig.add_subplot(gs[1], sharex=self.loss_p)
|
| 46 |
+
self.lr_p.tick_params('x', labelbottom=False)
|
| 47 |
+
self.graph_out = None
|
| 48 |
+
|
| 49 |
+
self.its = []
|
| 50 |
+
self.train_losses = []
|
| 51 |
+
self.val_losses = []
|
| 52 |
+
self.lr_history = []
|
| 53 |
+
|
| 54 |
+
def show(self):
|
| 55 |
+
self.start_t = time.time()
|
| 56 |
+
self.masterbar.write(["samples", "train", "val", "time"], table=True)
|
| 57 |
+
self.graph_out = display(self.graph_fig, display_id=True, clear=True)
|
| 58 |
+
|
| 59 |
+
def hide(self):
|
| 60 |
+
if self.graph_out is not None:
|
| 61 |
+
self.graph_out.update(IPython.display.HTML(''))
|
| 62 |
+
|
| 63 |
+
def plot(self):
|
| 64 |
+
loss_p, lr_p = self.loss_p, self.lr_p
|
| 65 |
+
loss_p.clear()
|
| 66 |
+
loss_p.plot(self.its, self.train_losses)
|
| 67 |
+
loss_p.plot(self.its, self.val_losses)
|
| 68 |
+
loss_p.set_xlim(0, self.total_steps)
|
| 69 |
+
loss_p.set_yscale('log')
|
| 70 |
+
lr_p.clear()
|
| 71 |
+
lrs = np.array(self.lr_history)
|
| 72 |
+
lr_p.plot(self.its, lrs)
|
| 73 |
+
self.graph_out.update(self.graph_fig)
|
| 74 |
+
|
| 75 |
+
def add_data(self, it, lr, train_loss, val_los):
|
| 76 |
+
self.its.append(it)
|
| 77 |
+
self.train_losses.append(train_loss)
|
| 78 |
+
self.val_losses.append(val_los)
|
| 79 |
+
self.lr_history.append(lr)
|
| 80 |
+
self.plot()
|
| 81 |
+
|
| 82 |
+
def add_table_row(self, it, avg_train_loss, val_loss):
|
| 83 |
+
elapsed_t = time.time() - self.start_t
|
| 84 |
+
self.masterbar.write([it, f"{avg_train_loss:.5f}", f"{val_loss:.5f}", fastprogress.core.format_time(elapsed_t)], table=True)
|
| 85 |
+
|
| 86 |
+
def on_iter(self, bar, it, avg_train_loss, val_loss):
|
| 87 |
+
epoch = math.ceil(it / self.total_steps * self.epochs)
|
| 88 |
+
bar.comment = f"#{epoch}/{self.epochs} loss: {avg_train_loss:.3f} / {val_loss:.3f}"
|
| 89 |
+
|
| 90 |
+
# %% ../nbs/B1. Training.ipynb 4
|
| 91 |
+
# FIXME: we need to keep this synchronised with the validation code below...
|
| 92 |
+
def validate(model, val, half=True, bs=16, drop_last=False, dl_workers=8, device="cuda"):
|
| 93 |
+
if isinstance(val, torch.utils.data.IterableDataset):
|
| 94 |
+
val_loader = wds.WebLoader(val, batch_size=None, num_workers=dl_workers, drop_last=drop_last) \
|
| 95 |
+
.unbatched().shuffle(1024).batched(bs)
|
| 96 |
+
else:
|
| 97 |
+
val_loader = DataLoader(val, batch_size=bs, num_workers=dl_workers, pin_memory=True, drop_last=drop_last)
|
| 98 |
+
|
| 99 |
+
with torch.no_grad():
|
| 100 |
+
val_loss = 0
|
| 101 |
+
val_samples = 0
|
| 102 |
+
for args in val_loader:
|
| 103 |
+
args = [x.to(device, non_blocking=True) for x in args]
|
| 104 |
+
with torch.autocast(device_type=device, dtype=torch.float16 if half else torch.float32, enabled=device!='cpu'):
|
| 105 |
+
ps, loss = model(*args)
|
| 106 |
+
N = args[0].shape[0]
|
| 107 |
+
val_loss += loss.mean().item() * N
|
| 108 |
+
val_samples += N
|
| 109 |
+
val_loss = val_loss / val_samples
|
| 110 |
+
|
| 111 |
+
return val_loss
|
| 112 |
+
|
| 113 |
+
# %% ../nbs/B1. Training.ipynb 5
|
| 114 |
+
def train(checkpoint_path, model, train, val, half=True, bs=16, lr=1e-4, drop_last=False,
|
| 115 |
+
weight_decay=0.1, warmup_steps=10000, epochs=10, clip_gradient_norm=None,
|
| 116 |
+
dl_workers=8, visual_class = SimpleVisual, profiler=None,
|
| 117 |
+
run_valid_every_iters=8000, table_row_every_iters=80000, chkpt_every_iters=None,
|
| 118 |
+
device="cuda", trainable_params=None):
|
| 119 |
+
if chkpt_every_iters is None:
|
| 120 |
+
chkpt_every_iters = table_row_every_iters
|
| 121 |
+
|
| 122 |
+
mb = master_bar(range(epochs))
|
| 123 |
+
if isinstance(train, torch.utils.data.IterableDataset):
|
| 124 |
+
pct_start = min(0.3, warmup_steps / (epochs * (train.total_samples//bs)))
|
| 125 |
+
visual = visual_class(model, mb, epochs * train.total_samples)
|
| 126 |
+
# pct_start = min(0.3, warmup_steps / (epochs * len(train)))
|
| 127 |
+
# visual = visual_class(model, mb, epochs*len(train)*bs)
|
| 128 |
+
else:
|
| 129 |
+
pct_start = min(0.3, warmup_steps / (epochs * len(train) / bs))
|
| 130 |
+
visual = visual_class(model, mb, epochs*len(train))
|
| 131 |
+
model.visual = visual
|
| 132 |
+
|
| 133 |
+
Path(checkpoint_path).mkdir(exist_ok=True)
|
| 134 |
+
|
| 135 |
+
if isinstance(train, torch.utils.data.IterableDataset):
|
| 136 |
+
# train_loader = DataLoader(train, batch_size=None, num_workers=dl_workers, pin_memory=True, drop_last=False, shuffle=False)
|
| 137 |
+
# val_loader = DataLoader(val, batch_size=None, num_workers=dl_workers, pin_memory=True, drop_last=False)
|
| 138 |
+
train_loader = wds.WebLoader(train, batch_size=None, num_workers=dl_workers, drop_last=drop_last) \
|
| 139 |
+
.unbatched().shuffle(1024).batched(bs, partial=False)
|
| 140 |
+
val_loader = wds.WebLoader(val, batch_size=None, num_workers=dl_workers, drop_last=drop_last) \
|
| 141 |
+
.unbatched().shuffle(1024).batched(bs)
|
| 142 |
+
else:
|
| 143 |
+
train_loader = DataLoader(train, batch_size=bs, num_workers=dl_workers, pin_memory=True, drop_last=drop_last, shuffle=True)
|
| 144 |
+
val_loader = DataLoader(val, batch_size=bs, num_workers=dl_workers, pin_memory=True, drop_last=drop_last)
|
| 145 |
+
|
| 146 |
+
val_loss = torch.nan
|
| 147 |
+
avg_train_loss = torch.nan
|
| 148 |
+
|
| 149 |
+
if hasattr(model, 'setup'):
|
| 150 |
+
model.setup(device)
|
| 151 |
+
|
| 152 |
+
try:
|
| 153 |
+
scheduler = None
|
| 154 |
+
|
| 155 |
+
if trainable_params is None: trainable_params = model.parameters()
|
| 156 |
+
all_params = set(trainable_params)
|
| 157 |
+
customized_params = set()
|
| 158 |
+
groups = []
|
| 159 |
+
group_map = {}
|
| 160 |
+
for name,m in model.named_modules():
|
| 161 |
+
if hasattr(m, 'no_weight_decay') or hasattr(m, 'lr_scale'):
|
| 162 |
+
m_trainable = [x for x in m.parameters() if x in all_params]
|
| 163 |
+
if not m_trainable: continue
|
| 164 |
+
customized_params |= set(m_trainable)
|
| 165 |
+
m_wd = 0 if hasattr(m, 'no_weight_decay') else weight_decay
|
| 166 |
+
m_lr = lr * getattr(m, 'lr_scale', 1)
|
| 167 |
+
group = group_map.get((m_wd, m_lr), None)
|
| 168 |
+
if not group:
|
| 169 |
+
group = {"params": [], "names": [], "weight_decay": m_wd, "lr": m_lr}
|
| 170 |
+
groups.append(group)
|
| 171 |
+
group_map[(m_wd, m_lr)] = group
|
| 172 |
+
group['params'] += m_trainable
|
| 173 |
+
group['names'].append(name)
|
| 174 |
+
|
| 175 |
+
other_params = all_params - customized_params
|
| 176 |
+
|
| 177 |
+
if other_params:
|
| 178 |
+
groups = groups + [
|
| 179 |
+
{"names": ["other"], "params": list(other_params), "weight_decay": weight_decay },
|
| 180 |
+
]
|
| 181 |
+
|
| 182 |
+
optimizer = torch.optim.AdamW(lr=lr, betas=(0.9, 0.95), fused=device!='cpu', params=groups)
|
| 183 |
+
model._optimizer = optimizer
|
| 184 |
+
scaler = torch.cuda.amp.GradScaler(enabled=half)
|
| 185 |
+
scheduler = torch.optim.lr_scheduler.OneCycleLR(
|
| 186 |
+
optimizer, pct_start=pct_start, steps_per_epoch=math.ceil(train.total_samples/bs), epochs=epochs,
|
| 187 |
+
max_lr=[pg.get('lr', lr) for pg in groups],
|
| 188 |
+
final_div_factor=25)
|
| 189 |
+
|
| 190 |
+
it = 0
|
| 191 |
+
next_val_it = it + 50
|
| 192 |
+
next_chkpt_it = chkpt_every_iters
|
| 193 |
+
next_table_it = table_row_every_iters
|
| 194 |
+
|
| 195 |
+
visual.show()
|
| 196 |
+
|
| 197 |
+
running_loss = [0]
|
| 198 |
+
|
| 199 |
+
for epoch in mb:
|
| 200 |
+
bar = progress_bar(train_loader, total=train.total_samples//bs, parent=mb)
|
| 201 |
+
for args in bar:
|
| 202 |
+
with record_function("forward"):
|
| 203 |
+
args = [x.to(device, non_blocking=True) for x in args]
|
| 204 |
+
|
| 205 |
+
# zero the parameter gradients
|
| 206 |
+
optimizer.zero_grad(set_to_none=True)
|
| 207 |
+
|
| 208 |
+
with torch.autocast(device_type=device, dtype=torch.float16 if half else torch.float32, enabled=device!='cpu'):
|
| 209 |
+
ps, loss = model(*args)
|
| 210 |
+
loss = loss.mean()
|
| 211 |
+
|
| 212 |
+
with record_function("backward"):
|
| 213 |
+
scaler.scale(loss).backward()
|
| 214 |
+
|
| 215 |
+
if clip_gradient_norm:
|
| 216 |
+
scaler.unscale_(optimizer)
|
| 217 |
+
# Since the gradients of optimizer's assigned params are unscaled, clips as usual:
|
| 218 |
+
torch.nn.utils.clip_grad_norm_(model.parameters(), clip_gradient_norm)
|
| 219 |
+
|
| 220 |
+
scaler.step(optimizer)
|
| 221 |
+
scaler.update()
|
| 222 |
+
|
| 223 |
+
scheduler.step()
|
| 224 |
+
|
| 225 |
+
if profiler is not None: profiler.step()
|
| 226 |
+
|
| 227 |
+
with record_function("running_loss"):
|
| 228 |
+
running_loss.append(loss.item())
|
| 229 |
+
running_loss = running_loss[-5:]
|
| 230 |
+
avg_train_loss = sum(running_loss)/len(running_loss)
|
| 231 |
+
|
| 232 |
+
if it >= next_chkpt_it:
|
| 233 |
+
with record_function("checkpoint"):
|
| 234 |
+
next_chkpt_it += chkpt_every_iters
|
| 235 |
+
torch.save(model.state_dict(), f'{checkpoint_path}/{it:08d}.pt')
|
| 236 |
+
|
| 237 |
+
if it >= next_val_it:
|
| 238 |
+
next_val_it += run_valid_every_iters
|
| 239 |
+
with record_function("validation"):
|
| 240 |
+
with record_function("model.eval"):
|
| 241 |
+
model.eval()
|
| 242 |
+
with torch.no_grad():
|
| 243 |
+
val_loss = 0
|
| 244 |
+
val_samples = 0
|
| 245 |
+
for args in val_loader:
|
| 246 |
+
args = [x.to(device, non_blocking=True) for x in args]
|
| 247 |
+
with torch.autocast(device_type=device, dtype=torch.float16 if half else torch.float32, enabled=device!='cpu'):
|
| 248 |
+
ps, loss = model(*args)
|
| 249 |
+
N = args[0].shape[0]
|
| 250 |
+
val_loss += loss.mean().item() * N
|
| 251 |
+
val_samples += N
|
| 252 |
+
val_loss = val_loss / val_samples
|
| 253 |
+
with record_function("model.train"):
|
| 254 |
+
model.train()
|
| 255 |
+
with record_function("plotting"):
|
| 256 |
+
visual.add_data(it, scheduler.get_last_lr(), avg_train_loss, val_loss)
|
| 257 |
+
|
| 258 |
+
if it >= next_table_it:
|
| 259 |
+
visual.add_table_row(it, avg_train_loss, val_loss)
|
| 260 |
+
next_table_it += table_row_every_iters
|
| 261 |
+
|
| 262 |
+
it += bs
|
| 263 |
+
visual.on_iter(bar, it, avg_train_loss, val_loss)
|
| 264 |
+
except KeyboardInterrupt:
|
| 265 |
+
mb.write(f"interrupted")
|
| 266 |
+
mb.show()
|
| 267 |
+
pass
|
| 268 |
+
finally:
|
| 269 |
+
visual.add_table_row(it, avg_train_loss, val_loss)
|
| 270 |
+
mb.show()
|
| 271 |
+
visual.hide()
|
whisperspeech/train_multi.py
ADDED
|
@@ -0,0 +1,263 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/B2. Training (Lightning).ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = []
|
| 5 |
+
|
| 6 |
+
# %% ../nbs/B2. Training (Lightning).ipynb 2
|
| 7 |
+
import io
|
| 8 |
+
import time
|
| 9 |
+
import random
|
| 10 |
+
from pathlib import Path
|
| 11 |
+
|
| 12 |
+
from fastprogress import progress_bar, master_bar
|
| 13 |
+
import fastprogress
|
| 14 |
+
import wandb
|
| 15 |
+
|
| 16 |
+
import numpy as np
|
| 17 |
+
import pylab as plt
|
| 18 |
+
|
| 19 |
+
import torch
|
| 20 |
+
import torch.nn as nn
|
| 21 |
+
from torch.utils.data.dataloader import DataLoader
|
| 22 |
+
from torch.profiler import record_function
|
| 23 |
+
|
| 24 |
+
# %% ../nbs/B2. Training (Lightning).ipynb 3
|
| 25 |
+
import lightning.pytorch as pl
|
| 26 |
+
import math
|
| 27 |
+
|
| 28 |
+
class TrainingTask(pl.LightningModule):
|
| 29 |
+
def __init__(self, model, model_hparams=None):
|
| 30 |
+
super().__init__()
|
| 31 |
+
self.model = model
|
| 32 |
+
self.model_hparams = model_hparams
|
| 33 |
+
|
| 34 |
+
def on_fit_start(self):
|
| 35 |
+
if getattr(self.model, 'setup'):
|
| 36 |
+
self.model.setup(self.device)
|
| 37 |
+
|
| 38 |
+
def configure_optimizers(self):
|
| 39 |
+
""" Initialize AdamW optimizer"""
|
| 40 |
+
lr = self.model_hparams['lr0']
|
| 41 |
+
weight_decay = self.model_hparams['weight_decay']
|
| 42 |
+
|
| 43 |
+
all_params = set(model.parameters())
|
| 44 |
+
customized_params = set()
|
| 45 |
+
groups = []
|
| 46 |
+
group_map = {}
|
| 47 |
+
for name,m in model.named_modules():
|
| 48 |
+
if hasattr(m, 'no_weight_decay') or hasattr(m, 'lr_scale'):
|
| 49 |
+
customized_params |= set(m.parameters())
|
| 50 |
+
m_wd = 0 if hasattr(m, 'no_weight_decay') else weight_decay
|
| 51 |
+
m_lr = lr * getattr(m, 'lr_scale', 1)
|
| 52 |
+
group = group_map.get((m_wd, m_lr), None)
|
| 53 |
+
if not group:
|
| 54 |
+
group = {"params": [], "names": [], "weight_decay": m_wd, "lr": m_lr}
|
| 55 |
+
groups.append(group)
|
| 56 |
+
group_map[(m_wd, m_lr)] = group
|
| 57 |
+
group['params'] += m.parameters()
|
| 58 |
+
group['names'].append(name)
|
| 59 |
+
|
| 60 |
+
other_params = all_params - customized_params
|
| 61 |
+
|
| 62 |
+
param_groups = groups + [
|
| 63 |
+
{"names": ["other"], "params": list(other_params), "weight_decay": weight_decay },
|
| 64 |
+
]
|
| 65 |
+
|
| 66 |
+
optimizer = torch.optim.AdamW(lr=lr, betas=(0.9, 0.95), params=param_groups)
|
| 67 |
+
|
| 68 |
+
# modified from https://github.com/Lightning-AI/lightning/issues/5449#issuecomment-1501597319
|
| 69 |
+
def num_steps_per_epoch() -> int:
|
| 70 |
+
"""Get number of steps"""
|
| 71 |
+
# Accessing _data_source is flaky and might break
|
| 72 |
+
dataset = self.trainer.fit_loop._data_source.dataloader()
|
| 73 |
+
dataset_size = len(dataset)
|
| 74 |
+
# math.ceil so always overestimate (underestimating throws exceptions)
|
| 75 |
+
num_steps = math.ceil(dataset_size / self.trainer.accumulate_grad_batches)
|
| 76 |
+
return num_steps
|
| 77 |
+
|
| 78 |
+
total_steps = self.model_hparams['epochs'] * num_steps_per_epoch()
|
| 79 |
+
self.model_hparams['pct_start'] = min(0.3, self.model_hparams['warmup_steps'] / total_steps)
|
| 80 |
+
|
| 81 |
+
print(f"{self.model_hparams['epochs']=} epochs x {num_steps_per_epoch()=} steps")
|
| 82 |
+
|
| 83 |
+
lr_scheduler = torch.optim.lr_scheduler.OneCycleLR(
|
| 84 |
+
optimizer,
|
| 85 |
+
pct_start=self.model_hparams['pct_start'],
|
| 86 |
+
max_lr=[pg.get('lr', lr) for pg in param_groups],
|
| 87 |
+
steps_per_epoch=num_steps_per_epoch(),
|
| 88 |
+
epochs=int(self.model_hparams['epochs']),
|
| 89 |
+
final_div_factor=25
|
| 90 |
+
)
|
| 91 |
+
|
| 92 |
+
return [optimizer], [{'scheduler': lr_scheduler, 'interval': 'step'}]
|
| 93 |
+
|
| 94 |
+
def training_step(self, train_batch, batch_idx):
|
| 95 |
+
train_logits, train_loss = self.model.forward(*train_batch)
|
| 96 |
+
|
| 97 |
+
self.log("train_loss", train_loss, sync_dist=True)
|
| 98 |
+
return train_loss
|
| 99 |
+
|
| 100 |
+
def validation_step(self, val_batch, batch_idx):
|
| 101 |
+
val_logits, val_loss = self.model.forward(*val_batch)
|
| 102 |
+
|
| 103 |
+
self.log("val_loss", val_loss, sync_dist=True)
|
| 104 |
+
return val_loss
|
| 105 |
+
|
| 106 |
+
def on_validation_epoch_end(self):
|
| 107 |
+
if hasattr(self.model, 'get_metrics'):
|
| 108 |
+
self.log_dict({'metrics/'+k:v for k,v in self.model.get_metrics().items()}, sync_dist=True)
|
| 109 |
+
|
| 110 |
+
def test_step(self, val_batch, batch_idx):
|
| 111 |
+
test_logits, test_loss = self.model.forward(*val_batch)
|
| 112 |
+
|
| 113 |
+
self.log("test_loss", test_loss, sync_dist=True)
|
| 114 |
+
return test_loss
|
| 115 |
+
|
| 116 |
+
# %% ../nbs/B2. Training (Lightning).ipynb 4
|
| 117 |
+
from fastcore.script import anno_parser
|
| 118 |
+
import shlex
|
| 119 |
+
|
| 120 |
+
# watch out: we can only pass Python values as keyword arguments (not positional)
|
| 121 |
+
# everything else has to be a string
|
| 122 |
+
def parse_and_call(name, fun, args, kwargs={}, log_to_wandb=True):
|
| 123 |
+
p = anno_parser(fun)
|
| 124 |
+
args = p.parse_args(args).__dict__
|
| 125 |
+
args.pop('xtra'); args.pop('pdb')
|
| 126 |
+
args.update({k:v for k, v in kwargs.items()})
|
| 127 |
+
if log_to_wandb and type(wandb_logger.experiment.config) == wandb.sdk.wandb_config.Config:
|
| 128 |
+
wandb_logger.experiment.config[name] = {k:v for k,v in args.items() if k not in ['dataset', 'tunables']}
|
| 129 |
+
return fun(**args)
|
| 130 |
+
|
| 131 |
+
# %% ../nbs/B2. Training (Lightning).ipynb 8
|
| 132 |
+
import argparse
|
| 133 |
+
|
| 134 |
+
parser = argparse.ArgumentParser()
|
| 135 |
+
parser.add_argument('--task', type=str, help='Task to train')
|
| 136 |
+
parser.add_argument('--seed', type=int, default=0, help='Global training seed')
|
| 137 |
+
parser.add_argument('--batch-size', type=int, default=16, help='total batch size for all GPUs')
|
| 138 |
+
parser.add_argument('--workers', type=int, default=8, help='max dataloader workers (per RANK in DDP mode)')
|
| 139 |
+
parser.add_argument('--input-dir', type=str, default='', help='input data path') # fixed in the model for now
|
| 140 |
+
parser.add_argument("--checkpoint-dir", type=str, default="./checkpoints/", help="directory to save the checkpoints")
|
| 141 |
+
parser.add_argument('--epochs', type=int, default=10, help='total training epochs')
|
| 142 |
+
parser.add_argument('--validate-every-n-steps', type=int, default=500, help='how training steps to run between validations')
|
| 143 |
+
parser.add_argument('--weight-decay', type=float, default=1e-2, help='optimizer weight decay')
|
| 144 |
+
parser.add_argument('--lr0', type=float, default=1e-4, help='optimizer initial learning rate')
|
| 145 |
+
parser.add_argument('--clip-gradient-norm', type=float, default=None, help='enable gradient norm clipping')
|
| 146 |
+
parser.add_argument('--accumulate-grad-batches', type=int, default=1, help='perform the optimizer step only after going through several batches of samples')
|
| 147 |
+
parser.add_argument('--precision', type=str, default="16-mixed", help="floating point precision")
|
| 148 |
+
parser.add_argument('--warmup-steps', type=int, default=10000, help='total number steps during which the learning rate rises (defaults to 10k updates)')
|
| 149 |
+
parser.add_argument('--tunables', type=str, default="", help='tunable hyperparameters')
|
| 150 |
+
parser.add_argument('--resume-from', type=Path, default=None, help='resume training from the given checkpoint')
|
| 151 |
+
parser.add_argument('--strategy', type=str, default='ddp', help='distributed training strategy')
|
| 152 |
+
parser.add_argument('--wandb-suffix', type=str, default=None, help='W&B project name suffix')
|
| 153 |
+
parser.add_argument('--wandb-task-name', type=str, default=None, help='Task name for the W&B project name')
|
| 154 |
+
|
| 155 |
+
args = parser.parse_args().__dict__
|
| 156 |
+
|
| 157 |
+
task_args: list = shlex.split(args.pop("task"))
|
| 158 |
+
task_name, task_args = task_args[0], task_args[1:]
|
| 159 |
+
input_args: list = shlex.split(args.pop("input_dir"))
|
| 160 |
+
checkpoint_dir: str = args.pop("checkpoint_dir")
|
| 161 |
+
num_workers: int = args.pop("workers")
|
| 162 |
+
batch_size: int = args.pop("batch_size")
|
| 163 |
+
epochs: int = args.pop("epochs")
|
| 164 |
+
tunables_args: list = shlex.split(args.pop("tunables"))
|
| 165 |
+
|
| 166 |
+
hyp_params = {}
|
| 167 |
+
hyp_params['batch_size'] = batch_size
|
| 168 |
+
hyp_params['warmup_steps'] = args['warmup_steps']
|
| 169 |
+
hyp_params['weight_decay'] = args['weight_decay']
|
| 170 |
+
hyp_params['clip_gradient_norm'] = args['clip_gradient_norm']
|
| 171 |
+
hyp_params['accumulate_grad_batches'] = args['accumulate_grad_batches']
|
| 172 |
+
hyp_params['precision'] = args['precision']
|
| 173 |
+
hyp_params['lr0'] = args['lr0']
|
| 174 |
+
hyp_params['epochs'] = epochs
|
| 175 |
+
hyp_params['strategy'] = args['strategy']
|
| 176 |
+
|
| 177 |
+
# %% ../nbs/B2. Training (Lightning).ipynb 9
|
| 178 |
+
from lightning.pytorch.loggers import WandbLogger
|
| 179 |
+
from lightning.pytorch.callbacks import LearningRateMonitor
|
| 180 |
+
import datetime
|
| 181 |
+
import webdataset as wds
|
| 182 |
+
import importlib
|
| 183 |
+
|
| 184 |
+
torch.set_float32_matmul_precision('medium')
|
| 185 |
+
|
| 186 |
+
project = f"WhisperSpeech-{args['wandb_task_name'] or task_name}"
|
| 187 |
+
if args['wandb_suffix']:
|
| 188 |
+
project += "-"+args['wandb_suffix']
|
| 189 |
+
|
| 190 |
+
wandb_logger = WandbLogger(project=project)
|
| 191 |
+
|
| 192 |
+
ckpt_callback = pl.callbacks.ModelCheckpoint(
|
| 193 |
+
dirpath=f'{task_name}-{epochs}e',
|
| 194 |
+
filename=task_name+"-{epoch}-{step}-{val_loss:.2f}",
|
| 195 |
+
monitor="val_loss",
|
| 196 |
+
save_top_k=4,
|
| 197 |
+
train_time_interval=datetime.timedelta(minutes=5),
|
| 198 |
+
)
|
| 199 |
+
|
| 200 |
+
lr_monitor_callback = LearningRateMonitor(logging_interval='step')
|
| 201 |
+
|
| 202 |
+
from torch.utils.data import DataLoader
|
| 203 |
+
|
| 204 |
+
task = importlib.import_module("whisperspeech."+task_name)
|
| 205 |
+
|
| 206 |
+
train_ds, val_ds = parse_and_call('dataset', task.load_datasets, input_args)
|
| 207 |
+
|
| 208 |
+
tunables = None
|
| 209 |
+
if hasattr(task, "Tunables"):
|
| 210 |
+
import dataclasses
|
| 211 |
+
tunables = parse_and_call('tunables', task.Tunables, tunables_args, log_to_wandb=False)
|
| 212 |
+
if type(wandb_logger.experiment.config) == wandb.sdk.wandb_config.Config:
|
| 213 |
+
wandb_logger.experiment.config['tunables'] = dataclasses.asdict(tunables)
|
| 214 |
+
|
| 215 |
+
for name in ["lr0", "clip_gradient_norm", "weight_decay", "warmup_steps"]:
|
| 216 |
+
val = getattr(tunables, name, None)
|
| 217 |
+
if val is not None: hyp_params[name] = val
|
| 218 |
+
|
| 219 |
+
if isinstance(train_ds, torch.utils.data.IterableDataset):
|
| 220 |
+
dl_batch_size, dl_shuffle = None, False
|
| 221 |
+
pin_memory = False
|
| 222 |
+
else:
|
| 223 |
+
dl_batch_size, dl_shuffle = batch_size, True
|
| 224 |
+
pin_memory = True
|
| 225 |
+
|
| 226 |
+
val_loader = wds.WebLoader(val_ds,
|
| 227 |
+
batch_size=dl_batch_size,
|
| 228 |
+
num_workers=num_workers,
|
| 229 |
+
drop_last=False,
|
| 230 |
+
pin_memory=pin_memory).unbatched().shuffle(1024).batched(batch_size).with_length(val_ds.total_samples // batch_size)
|
| 231 |
+
|
| 232 |
+
train_loader = wds.WebLoader(train_ds,
|
| 233 |
+
batch_size=dl_batch_size,
|
| 234 |
+
num_workers=num_workers,
|
| 235 |
+
drop_last=False,
|
| 236 |
+
shuffle=dl_shuffle,
|
| 237 |
+
pin_memory=pin_memory).unbatched().shuffle(1024).batched(batch_size).with_length(train_ds.total_samples // batch_size)
|
| 238 |
+
|
| 239 |
+
model_kwargs = dict(dataset=train_ds)
|
| 240 |
+
if tunables is not None: model_kwargs['tunables'] = tunables
|
| 241 |
+
model = parse_and_call('model', task.make_model, task_args, model_kwargs)
|
| 242 |
+
|
| 243 |
+
task = TrainingTask(model, model_hparams=hyp_params)
|
| 244 |
+
|
| 245 |
+
trainer = pl.Trainer(strategy=hyp_params['strategy'],
|
| 246 |
+
max_epochs=hyp_params['epochs'],
|
| 247 |
+
accelerator="gpu",
|
| 248 |
+
profiler="simple",
|
| 249 |
+
precision=hyp_params['precision'],
|
| 250 |
+
gradient_clip_val=hyp_params['clip_gradient_norm'],
|
| 251 |
+
accumulate_grad_batches=hyp_params['accumulate_grad_batches'],
|
| 252 |
+
val_check_interval=args.pop("validate_every_n_steps"),
|
| 253 |
+
enable_checkpointing=True,
|
| 254 |
+
logger=wandb_logger,
|
| 255 |
+
callbacks=[ckpt_callback, lr_monitor_callback])
|
| 256 |
+
|
| 257 |
+
if type(wandb_logger.experiment.config) == wandb.sdk.wandb_config.Config:
|
| 258 |
+
wandb_logger.experiment.config.update(hyp_params)
|
| 259 |
+
|
| 260 |
+
kwargs = {}
|
| 261 |
+
if 'resume_from' in args:
|
| 262 |
+
kwargs['ckpt_path'] = args['resume_from']
|
| 263 |
+
trainer.fit(model=task, train_dataloaders=train_loader, val_dataloaders=val_loader, **kwargs)
|
whisperspeech/utils.py
ADDED
|
@@ -0,0 +1,159 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/D. Common dataset utilities.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = ['shard_glob', 'join_datasets', 'resampler', 'derived_name', 'derived_dataset', 'merge_in', 'AtomicTarWriter',
|
| 5 |
+
'readlines']
|
| 6 |
+
|
| 7 |
+
# %% ../nbs/D. Common dataset utilities.ipynb 1
|
| 8 |
+
import os
|
| 9 |
+
import torch
|
| 10 |
+
import torchaudio
|
| 11 |
+
from pathlib import Path
|
| 12 |
+
import webdataset as wds
|
| 13 |
+
from contextlib import contextmanager
|
| 14 |
+
|
| 15 |
+
import torch.nn.functional as F
|
| 16 |
+
|
| 17 |
+
# %% ../nbs/D. Common dataset utilities.ipynb 2
|
| 18 |
+
def shard_glob(input):
|
| 19 |
+
if '{' in input:
|
| 20 |
+
return wds.shardlists.expand_urls(input)
|
| 21 |
+
if isinstance(input, (Path, str)):
|
| 22 |
+
path = Path(input)
|
| 23 |
+
if path.is_dir():
|
| 24 |
+
glob = '*.tar.gz'
|
| 25 |
+
else:
|
| 26 |
+
glob = path.name
|
| 27 |
+
path = path.parent
|
| 28 |
+
input = Path(path).glob(glob)
|
| 29 |
+
else:
|
| 30 |
+
raise ArgumentError("input should be either a list or a path with an optional glob specifier")
|
| 31 |
+
return [str(x) for x in input]
|
| 32 |
+
|
| 33 |
+
# %% ../nbs/D. Common dataset utilities.ipynb 3
|
| 34 |
+
class join_datasets(torch.utils.data.IterableDataset):
|
| 35 |
+
def __init__(self, datasets):
|
| 36 |
+
self.datasets = datasets
|
| 37 |
+
|
| 38 |
+
def __iter__(self):
|
| 39 |
+
probs = torch.tensor([getattr(ds, 'weight', 1) for ds in self.datasets], dtype=torch.float)
|
| 40 |
+
its = [iter(ds) for ds in self.datasets]
|
| 41 |
+
while True:
|
| 42 |
+
try:
|
| 43 |
+
yield next(its[torch.multinomial(probs, 1)])
|
| 44 |
+
except StopIteration:
|
| 45 |
+
return
|
| 46 |
+
|
| 47 |
+
def __len__(self):
|
| 48 |
+
return sum([ds.total_samples for ds in self.datasets])
|
| 49 |
+
|
| 50 |
+
# %% ../nbs/D. Common dataset utilities.ipynb 5
|
| 51 |
+
def resampler(newsr = 24000, key = 'samples_24k'):
|
| 52 |
+
_last_sr = None
|
| 53 |
+
tform = None
|
| 54 |
+
|
| 55 |
+
def _resample(samples):
|
| 56 |
+
for s in samples:
|
| 57 |
+
sr = s['sample_rate']
|
| 58 |
+
if sr != newsr:
|
| 59 |
+
if sr != _last_sr: tform = torchaudio.transforms.Resample(sr, newsr)
|
| 60 |
+
s[key] = tform(s['samples'])
|
| 61 |
+
else:
|
| 62 |
+
s[key] = s['samples']
|
| 63 |
+
yield s
|
| 64 |
+
|
| 65 |
+
return _resample
|
| 66 |
+
|
| 67 |
+
# %% ../nbs/D. Common dataset utilities.ipynb 6
|
| 68 |
+
def derived_name(input, kind, base="audio", suffix=".gz", dir=None):
|
| 69 |
+
dir = Path(dir) if dir else Path(input).parent
|
| 70 |
+
return str(dir/(Path(input).name.replace(f"-{base}-", f"-{kind}-") + suffix))
|
| 71 |
+
|
| 72 |
+
# %% ../nbs/D. Common dataset utilities.ipynb 7
|
| 73 |
+
def derived_dataset(kind, base='audio', suffix=".gz", decoders=[], dir=None):
|
| 74 |
+
def deriver(url):
|
| 75 |
+
url = str(derived_name(url, kind, base=base, suffix=suffix, dir=dir))
|
| 76 |
+
return wds.WebDataset(
|
| 77 |
+
wds.SimpleShardList([url])
|
| 78 |
+
).decode(*decoders)
|
| 79 |
+
return deriver
|
| 80 |
+
|
| 81 |
+
# %% ../nbs/D. Common dataset utilities.ipynb 8
|
| 82 |
+
def merge_in(dataset_fun):
|
| 83 |
+
"""Merge a dataset into the current one returning samples with the union of keys. Pass in a function
|
| 84 |
+
that takes a URL of a sample and returns a dataset for it (called everytime the URL changes).
|
| 85 |
+
|
| 86 |
+
It requires (and validates) that both datasets have the same ordering of keys so you have
|
| 87 |
+
to use it before any sample shuffling. Shard shuffling is ok.
|
| 88 |
+
"""
|
| 89 |
+
def merge_loop(main_samples):
|
| 90 |
+
#print("new merge loop:", dataset_fun)
|
| 91 |
+
merged_samples = None
|
| 92 |
+
cur_url = None
|
| 93 |
+
i = None
|
| 94 |
+
for s in main_samples:
|
| 95 |
+
url = s['__url__']
|
| 96 |
+
if url != cur_url:
|
| 97 |
+
# this will open a new file when we get the first sample with a new __url__
|
| 98 |
+
merged_samples = iter(dataset_fun(url))
|
| 99 |
+
cur_url = url
|
| 100 |
+
try:
|
| 101 |
+
merge_s = next(merged_samples)
|
| 102 |
+
except StopIteration:
|
| 103 |
+
# if the original shard got repeated we won't observe a __url__ change
|
| 104 |
+
# in this case restart the dataset from the beginning
|
| 105 |
+
merged_samples = iter(dataset_fun(url))
|
| 106 |
+
merge_s = next(merged_samples)
|
| 107 |
+
assert merge_s['__key__'] == s['__key__'], f"sample keys don't match: {merge_s['__key__']}, {s['__key__']} in file {s['__url__']}"
|
| 108 |
+
news = {}
|
| 109 |
+
news.update(merge_s)
|
| 110 |
+
news.update(s)
|
| 111 |
+
yield news
|
| 112 |
+
return merge_loop
|
| 113 |
+
|
| 114 |
+
# %% ../nbs/D. Common dataset utilities.ipynb 9
|
| 115 |
+
def split_to_chunks(stream, ikey='vad.npy', metakeys=[], pad_to_seconds=30, random_shift=False):
|
| 116 |
+
for s in stream:
|
| 117 |
+
audio, sr = s['audio']
|
| 118 |
+
imax = len(s[ikey]) - 1
|
| 119 |
+
for i,(ts,te) in enumerate(s[ikey]):
|
| 120 |
+
samples = audio[0,int(ts*sr):int(te*sr)]
|
| 121 |
+
if pad_to_seconds is not None:
|
| 122 |
+
padding = pad_to_seconds*sr-samples.shape[-1]
|
| 123 |
+
lpad = random.randint(0, padding) if random_shift else 0
|
| 124 |
+
samples = F.pad(samples, (lpad, padding-lpad))
|
| 125 |
+
subs = {"__key__": s['__key__'] + f"_{i:03d}",
|
| 126 |
+
"src_key": s['__key__'],
|
| 127 |
+
"__url__": s['__url__'],
|
| 128 |
+
"i": i, "imax": imax,
|
| 129 |
+
"tstart": ts, "tend": te, "total_seconds": audio.shape[-1]/sr,
|
| 130 |
+
"lpad": lpad, "rpad": padding-lpad,
|
| 131 |
+
"lpad_s": lpad/sr, "rpad_s": (padding-lpad)/sr,
|
| 132 |
+
"samples": samples, "sample_rate": sr}
|
| 133 |
+
for k in metakeys:
|
| 134 |
+
subs[k] = s[k][i]
|
| 135 |
+
yield subs
|
| 136 |
+
|
| 137 |
+
# %% ../nbs/D. Common dataset utilities.ipynb 10
|
| 138 |
+
def vad_dataset(shards, ikey='vad.npy', kind='vad'):
|
| 139 |
+
return wds.WebDataset(shards).compose(
|
| 140 |
+
wds.decode(wds.torch_audio),
|
| 141 |
+
merge_in(derived_dataset(kind)),
|
| 142 |
+
wds.select(lambda x: 'wav' in x or 'flac' in x or 'mp3' in x or 'ogg' in x), # skip samples without audio
|
| 143 |
+
wds.rename(audio="flac;mp3;wav;ogg"),
|
| 144 |
+
lambda x: split_to_chunks(x, ikey=ikey),
|
| 145 |
+
)
|
| 146 |
+
|
| 147 |
+
# %% ../nbs/D. Common dataset utilities.ipynb 11
|
| 148 |
+
@contextmanager
|
| 149 |
+
def AtomicTarWriter(name, throwaway=False):
|
| 150 |
+
tmp = name+".tmp"
|
| 151 |
+
with wds.TarWriter(tmp, compress=name.endswith('gz')) as sink:
|
| 152 |
+
yield sink
|
| 153 |
+
if not throwaway:
|
| 154 |
+
os.rename(tmp, name)
|
| 155 |
+
|
| 156 |
+
# %% ../nbs/D. Common dataset utilities.ipynb 12
|
| 157 |
+
def readlines(fname):
|
| 158 |
+
with open(fname) as file:
|
| 159 |
+
return [line.rstrip() for line in file]
|
whisperspeech/vad.py
ADDED
|
@@ -0,0 +1,71 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/1B. Voice activity detection.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = []
|
| 5 |
+
|
| 6 |
+
# %% ../nbs/1B. Voice activity detection.ipynb 3
|
| 7 |
+
import os
|
| 8 |
+
import torch
|
| 9 |
+
import torchaudio
|
| 10 |
+
|
| 11 |
+
from pathlib import Path
|
| 12 |
+
from fastprogress import progress_bar
|
| 13 |
+
from fastcore.script import call_parse
|
| 14 |
+
|
| 15 |
+
import whisperx
|
| 16 |
+
import random
|
| 17 |
+
import numpy as np
|
| 18 |
+
import webdataset as wds
|
| 19 |
+
|
| 20 |
+
# %% ../nbs/1B. Voice activity detection.ipynb 5
|
| 21 |
+
# some of the original file names have a dot in their name
|
| 22 |
+
# webdataset does not like it so let's patch it
|
| 23 |
+
def fix_dots_in_names(name):
|
| 24 |
+
name, ext = name.rsplit('.', 1)
|
| 25 |
+
return ".".join((name.replace('.', '_'), ext))
|
| 26 |
+
|
| 27 |
+
def load_dataset(url, decode=True, rename_files=None):
|
| 28 |
+
ds = wds.WebDataset(url, rename_files=rename_files)
|
| 29 |
+
if not decode: return ds
|
| 30 |
+
return ds.decode(wds.torch_audio)
|
| 31 |
+
|
| 32 |
+
# %% ../nbs/1B. Voice activity detection.ipynb 7
|
| 33 |
+
def extract_segments(vad_result, max_duration):
|
| 34 |
+
binarize = whisperx.vad.Binarize(max_duration=max_duration)
|
| 35 |
+
segments = binarize(vad_result)
|
| 36 |
+
return [(x.start, x.end) for x in segments.get_timeline()]
|
| 37 |
+
|
| 38 |
+
def segment_audio(vad_model, audio, sr=16000):
|
| 39 |
+
vad_result = vad_model({"waveform": audio, "sample_rate": sr})
|
| 40 |
+
return extract_segments(vad_result, 30)
|
| 41 |
+
|
| 42 |
+
# %% ../nbs/1B. Voice activity detection.ipynb 13
|
| 43 |
+
def flac_to_vad_name(input):
|
| 44 |
+
if '-flac-' in input:
|
| 45 |
+
return input.rsplit("/", 1)[1].replace('flac', 'vad') + ".gz"
|
| 46 |
+
else:
|
| 47 |
+
return input.rsplit("/", 1)[1].replace('raw', 'vad') + ".gz"
|
| 48 |
+
|
| 49 |
+
@call_parse
|
| 50 |
+
def process_shard(
|
| 51 |
+
input:str, # input shard URL/path
|
| 52 |
+
output:str=None, # output shard URL/path
|
| 53 |
+
fix_dots:bool=False, # fix dots in LibriLight filenames
|
| 54 |
+
):
|
| 55 |
+
if output is None: output = flac_to_vad_name(input)
|
| 56 |
+
|
| 57 |
+
ds = torch.utils.data.DataLoader(load_dataset(input, rename_files=fix_dots_in_names if fix_dots else None), num_workers=2, batch_size=None)
|
| 58 |
+
vad_model = whisperx.vad.load_vad_model('cuda')
|
| 59 |
+
|
| 60 |
+
tmp = output+".tmp"
|
| 61 |
+
with wds.TarWriter(tmp) as sink:
|
| 62 |
+
for s in progress_bar(ds, total='noinfer'):
|
| 63 |
+
audio, sr = s.get('flac', s.get('wav', (None, None)))
|
| 64 |
+
if audio is None:
|
| 65 |
+
print(f"warning: '{s['__key__']}' does not contain an audio file")
|
| 66 |
+
continue
|
| 67 |
+
sink.write({
|
| 68 |
+
"__key__": s['__key__'],
|
| 69 |
+
"vad.npy": np.array(segment_audio(vad_model, audio, sr=sr), dtype=np.float16)
|
| 70 |
+
})
|
| 71 |
+
os.rename(tmp, output)
|
whisperspeech/vq_stoks.py
ADDED
|
@@ -0,0 +1,493 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/2B. Whisper quantization (semantic token) model.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = ['RQBottleneckTransformer', 'make_model']
|
| 5 |
+
|
| 6 |
+
# %% ../nbs/2B. Whisper quantization (semantic token) model.ipynb 2
|
| 7 |
+
import io
|
| 8 |
+
import sys
|
| 9 |
+
import time
|
| 10 |
+
import torch
|
| 11 |
+
import torchaudio
|
| 12 |
+
|
| 13 |
+
# %% ../nbs/2B. Whisper quantization (semantic token) model.ipynb 3
|
| 14 |
+
from pathlib import Path
|
| 15 |
+
import json
|
| 16 |
+
from fastprogress import progress_bar, master_bar
|
| 17 |
+
import fastprogress
|
| 18 |
+
import numpy as np
|
| 19 |
+
import pylab as plt
|
| 20 |
+
import pandas as pd
|
| 21 |
+
import random
|
| 22 |
+
|
| 23 |
+
import whisper
|
| 24 |
+
from huggingface_hub import hf_hub_download
|
| 25 |
+
from fastcore.basics import store_attr
|
| 26 |
+
|
| 27 |
+
from torch import nn
|
| 28 |
+
import torch.optim as optim
|
| 29 |
+
import torch.nn.functional as F
|
| 30 |
+
from torch.utils.data.dataloader import DataLoader
|
| 31 |
+
import webdataset as wds
|
| 32 |
+
from . import utils
|
| 33 |
+
|
| 34 |
+
from vector_quantize_pytorch import ResidualVQ
|
| 35 |
+
|
| 36 |
+
from fastcore.script import *
|
| 37 |
+
|
| 38 |
+
# %% ../nbs/2B. Whisper quantization (semantic token) model.ipynb 9
|
| 39 |
+
def merge_in(dataset_fun):
|
| 40 |
+
"""Merge a dataset into the current one returning samples with the union of keys. Pass in a function
|
| 41 |
+
that takes a URL of a sample and returns a dataset for it (called everytime the URL changes).
|
| 42 |
+
|
| 43 |
+
It requires (and validates) that both datasets have the same ordering of keys so you have
|
| 44 |
+
to use it before any sample shuffling. Shard shuffling is ok.
|
| 45 |
+
"""
|
| 46 |
+
def merge_loop(main_samples):
|
| 47 |
+
#print("new merge loop:", dataset_fun)
|
| 48 |
+
merged_samples = None
|
| 49 |
+
cur_url = None
|
| 50 |
+
i = None
|
| 51 |
+
for s in main_samples:
|
| 52 |
+
url = s['__url__']
|
| 53 |
+
if url != cur_url:
|
| 54 |
+
# this will open a new file when we get the first sample with a new __url__
|
| 55 |
+
merged_samples = iter(dataset_fun(url))
|
| 56 |
+
cur_url = url
|
| 57 |
+
try:
|
| 58 |
+
merge_s = next(merged_samples)
|
| 59 |
+
except StopIteration:
|
| 60 |
+
# if the original shard got repeated we won't observe a __url__ change
|
| 61 |
+
# in this case restart the dataset from the beginning
|
| 62 |
+
merged_samples = iter(dataset_fun(url))
|
| 63 |
+
merge_s = next(merged_samples)
|
| 64 |
+
assert merge_s['__key__'] == s['__key__'], f"sample keys don't match: {merge_s['__key__']}, {s['__key__']} in file {s['__url__']}"
|
| 65 |
+
news = {}
|
| 66 |
+
news.update(merge_s)
|
| 67 |
+
news.update(s)
|
| 68 |
+
yield news
|
| 69 |
+
return merge_loop
|
| 70 |
+
|
| 71 |
+
# %% ../nbs/2B. Whisper quantization (semantic token) model.ipynb 10
|
| 72 |
+
def derived_dataset(kind, key='audio'):
|
| 73 |
+
def deriver(url):
|
| 74 |
+
url = str(Path(url).parent/(Path(url).name.replace(key, kind) + ".gz"))
|
| 75 |
+
return wds.WebDataset(
|
| 76 |
+
wds.SimpleShardList([url])
|
| 77 |
+
).decode()
|
| 78 |
+
return deriver
|
| 79 |
+
|
| 80 |
+
# %% ../nbs/2B. Whisper quantization (semantic token) model.ipynb 17
|
| 81 |
+
def add_masks(samples):
|
| 82 |
+
for s in samples:
|
| 83 |
+
seconds = s['tend'] - s['tstart']
|
| 84 |
+
# a mask (downsampled to the Whisper encoder token rate of 50/s) is used
|
| 85 |
+
# to teach the model the concept of padding
|
| 86 |
+
# this let's us decode shorter sequences later
|
| 87 |
+
mask = torch.zeros(30*16000//320, dtype=torch.bool)
|
| 88 |
+
mask[:int(seconds * 16000) // 320] = 1
|
| 89 |
+
s['mask'] = mask
|
| 90 |
+
yield s
|
| 91 |
+
|
| 92 |
+
def tokenize_text(samples, ttoks_size=200, model="base.en", language="en"):
|
| 93 |
+
multilingual = not model.endswith(".en")
|
| 94 |
+
tokenizer = whisper.tokenizer.get_tokenizer(multilingual, language=language, task="transcribe")
|
| 95 |
+
for s in samples:
|
| 96 |
+
ttoks = tokenizer.encode(s['txt'])
|
| 97 |
+
tokens = list(tokenizer.sot_sequence) + ttoks
|
| 98 |
+
rpad = ttoks_size - len(tokens)
|
| 99 |
+
s['in_ttoks'] = F.pad(torch.tensor(tokens), (0, rpad), value=tokenizer.eot)
|
| 100 |
+
s['out_ttoks'] = F.pad(torch.tensor(tokens[1:] + [tokenizer.eot]), (0, rpad), value=-100)
|
| 101 |
+
yield s
|
| 102 |
+
|
| 103 |
+
# %% ../nbs/2B. Whisper quantization (semantic token) model.ipynb 22
|
| 104 |
+
def load_dataset(
|
| 105 |
+
shard_spec:str,
|
| 106 |
+
proc_dataset_path:Path, # processed VAD and txt files
|
| 107 |
+
samples:int, # set the per-GPU sample count
|
| 108 |
+
txt_label:str="base.en-txt", # the label of the files containing transcriptions
|
| 109 |
+
model:str="base.en",
|
| 110 |
+
key:str="flac",
|
| 111 |
+
language:str=None,
|
| 112 |
+
validation:bool=False,
|
| 113 |
+
):
|
| 114 |
+
from . import wh_transcribe
|
| 115 |
+
shards = utils.shard_glob(shard_spec)
|
| 116 |
+
|
| 117 |
+
if not language and model.endswith('en'): language = 'en'
|
| 118 |
+
assert language, "please provide the dataset language for multilang models"
|
| 119 |
+
|
| 120 |
+
same_on_all_nodes = lambda urls: urls # will only be used for validation
|
| 121 |
+
ds = wds.WebDataset(shards, resampled=not validation, nodesplitter=same_on_all_nodes).compose(
|
| 122 |
+
wds.decode(wds.torch_audio),
|
| 123 |
+
wds.select(lambda x: 'wav' in x or 'flac' in x or 'mp3' in x or 'ogg' in x), # skip samples without audio
|
| 124 |
+
wds.rename(audio="flac;mp3;wav;ogg"),
|
| 125 |
+
merge_in(derived_dataset(proc_dataset_path, 'vad', key=key)),
|
| 126 |
+
wds.map_dict(**{"vad.npy":wh_transcribe.chunk_merger}),
|
| 127 |
+
wh_transcribe.split_to_chunks,
|
| 128 |
+
utils.resampler(16000, 'samples_16k'),
|
| 129 |
+
merge_in(derived_dataset(proc_dataset_path, txt_label, key=key)),
|
| 130 |
+
)
|
| 131 |
+
if 'librilight' in shards[0]:
|
| 132 |
+
ds = ds.compose(
|
| 133 |
+
# drop the first and last segment because they tend to be inaccurate
|
| 134 |
+
# (the transcriptions don't have the "LibriVox" headers and "end of chapter" suffixes)
|
| 135 |
+
wds.select(lambda x: x['i'] != 0 and x['i'] != x['imax']),
|
| 136 |
+
)
|
| 137 |
+
ds = ds.compose(
|
| 138 |
+
add_masks,
|
| 139 |
+
lambda x: tokenize_text(x, model=model, language=language),
|
| 140 |
+
wds.to_tuple('samples_16k', 'mask', 'in_ttoks', 'out_ttoks'),
|
| 141 |
+
wds.batched(32),
|
| 142 |
+
)
|
| 143 |
+
ds.total_samples = samples
|
| 144 |
+
|
| 145 |
+
return ds
|
| 146 |
+
|
| 147 |
+
# %% ../nbs/2B. Whisper quantization (semantic token) model.ipynb 28
|
| 148 |
+
from whisperspeech.train import *
|
| 149 |
+
from whisperspeech.modules import *
|
| 150 |
+
|
| 151 |
+
# %% ../nbs/2B. Whisper quantization (semantic token) model.ipynb 29
|
| 152 |
+
import dataclasses
|
| 153 |
+
|
| 154 |
+
def rand(start, end):
|
| 155 |
+
return random.random() * (end - start) + start
|
| 156 |
+
|
| 157 |
+
def logrand(start, end):
|
| 158 |
+
return 10**rand(math.log10(start), math.log10(end))
|
| 159 |
+
|
| 160 |
+
@dataclasses.dataclass
|
| 161 |
+
class Tunables:
|
| 162 |
+
init_std :float = 1.5
|
| 163 |
+
embeddings_std :float = 4.5e-2
|
| 164 |
+
embeddings_lr_scale: float = 1
|
| 165 |
+
output_mult :float = 1
|
| 166 |
+
query_mult :float = 2
|
| 167 |
+
rope :bool = True
|
| 168 |
+
mask_embs :bool = True # force embeddings corresponding to the input audio padding to a constant value
|
| 169 |
+
downsample_conv: bool = False
|
| 170 |
+
downsample_mean: bool = True
|
| 171 |
+
|
| 172 |
+
codebook_dim: int = 32
|
| 173 |
+
codebook_decay: float = 0.9
|
| 174 |
+
|
| 175 |
+
lr0 :float = .9e-3
|
| 176 |
+
clip_gradient_norm :float = 2
|
| 177 |
+
weight_decay :float = 1e-3
|
| 178 |
+
warmup_steps :float = 850
|
| 179 |
+
|
| 180 |
+
random :bool = False
|
| 181 |
+
|
| 182 |
+
def __post_init__(self):
|
| 183 |
+
# randomize the hyperparams if requested
|
| 184 |
+
if self.random:
|
| 185 |
+
self.init_std = logrand(1, 2)
|
| 186 |
+
self.embeddings_std = logrand(3e-2,6e-2)
|
| 187 |
+
self.embeddings_lr_scale = 2**rand(0,3)
|
| 188 |
+
self.output_mult = 2**rand(-3,3)
|
| 189 |
+
self.query_mult = logrand(1,8)
|
| 190 |
+
self.codebook_dim = int(logrand(30,50))
|
| 191 |
+
self.codebook_decay = logrand(0.86,0.95)
|
| 192 |
+
self.rope = True
|
| 193 |
+
self.mask_embs = True
|
| 194 |
+
self.downsample_mean = True
|
| 195 |
+
|
| 196 |
+
self.lr0 = logrand(.8e-3,1e-3)
|
| 197 |
+
self.clip_gradient_norm = 10**rand(-1,1)
|
| 198 |
+
self.warmup_steps = logrand(700,1000)
|
| 199 |
+
|
| 200 |
+
@staticmethod
|
| 201 |
+
def upgrade(args):
|
| 202 |
+
args = {k:v for k,v in args.items()}
|
| 203 |
+
def old_default(name, value):
|
| 204 |
+
if name not in args: args[name] = value
|
| 205 |
+
old_default('output_mult', 1)
|
| 206 |
+
old_default('query_mult', 1)
|
| 207 |
+
old_default('rope', False)
|
| 208 |
+
old_default('mask_embs', False)
|
| 209 |
+
old_default('downsample_conv', False)
|
| 210 |
+
old_default('downsample_mean', False)
|
| 211 |
+
if 'encoder_depth_ratio' in args: del args['encoder_depth_ratio']
|
| 212 |
+
if 'vq_codes' in args: del args['vq_codes']
|
| 213 |
+
return args
|
| 214 |
+
|
| 215 |
+
# %% ../nbs/2B. Whisper quantization (semantic token) model.ipynb 30
|
| 216 |
+
import math
|
| 217 |
+
|
| 218 |
+
# %% ../nbs/2B. Whisper quantization (semantic token) model.ipynb 31
|
| 219 |
+
class RQBottleneckTransformer(nn.Module):
|
| 220 |
+
def __init__(self, vq_codes=512, q_depth=12, depth=1, n_head=2, head_width=64, ffn_mult=4,
|
| 221 |
+
codebook_dim=2, threshold_ema_dead_code=2, use_cosine_sim = False, kl_loss_mul=1,
|
| 222 |
+
downsample=1,
|
| 223 |
+
whisper_model_name='tiny.en', tunables=Tunables()):
|
| 224 |
+
super().__init__()
|
| 225 |
+
width = n_head * head_width
|
| 226 |
+
store_attr("codebook_dim,vq_codes,q_depth,n_head,head_width,ffn_mult,depth,use_cosine_sim,downsample,whisper_model_name")
|
| 227 |
+
self.width = width
|
| 228 |
+
self.base_width = 3 * head_width
|
| 229 |
+
self.vq_codes = vq_codes
|
| 230 |
+
self.tunables = tunables
|
| 231 |
+
self.stoks_len = 1500//downsample
|
| 232 |
+
self.stoks_per_sec = self.stoks_len//30
|
| 233 |
+
|
| 234 |
+
qk_scale = self.tunables.query_mult * 8 / math.sqrt(head_width)
|
| 235 |
+
|
| 236 |
+
self.kl_loss_mul = kl_loss_mul
|
| 237 |
+
|
| 238 |
+
n_mlp = width * ffn_mult
|
| 239 |
+
self.mlp = nn.Sequential(
|
| 240 |
+
nn.Linear(width, n_mlp), nn.GELU(), nn.Linear(n_mlp, width)
|
| 241 |
+
)
|
| 242 |
+
self.mlp_ln = LayerNorm(width)
|
| 243 |
+
|
| 244 |
+
if tunables.downsample_conv:
|
| 245 |
+
self.downsample_conv = nn.Conv1d(width, width, kernel_size=3, stride=downsample, padding=1)
|
| 246 |
+
else:
|
| 247 |
+
self.downsample_conv = None
|
| 248 |
+
|
| 249 |
+
if tunables.mask_embs: vq_codes = vq_codes + 1
|
| 250 |
+
self.rq = ResidualVQ(
|
| 251 |
+
dim = width,
|
| 252 |
+
codebook_size = vq_codes, # codebook size
|
| 253 |
+
decay = tunables.codebook_decay, # the exponential moving average decay, lower means the dictionary will change faster
|
| 254 |
+
commitment_weight = 1., # the weight on the commitment loss
|
| 255 |
+
threshold_ema_dead_code = threshold_ema_dead_code,
|
| 256 |
+
use_cosine_sim = use_cosine_sim,
|
| 257 |
+
codebook_dim = codebook_dim,
|
| 258 |
+
num_quantizers= 1,
|
| 259 |
+
)
|
| 260 |
+
|
| 261 |
+
self.ce_lossf = nn.CrossEntropyLoss(ignore_index=-100)
|
| 262 |
+
self.kl_lossf = nn.KLDivLoss(reduction='batchmean')
|
| 263 |
+
|
| 264 |
+
self.positional_embedding = nn.Embedding(1500, width) # FIXME: should be self.stoks_len
|
| 265 |
+
|
| 266 |
+
self.out_blocks = nn.Sequential(*[
|
| 267 |
+
ResidualAttentionBlock(width, n_head, qk_scale=qk_scale, ffn_mult=ffn_mult, rope=tunables.rope) for _ in range(depth)
|
| 268 |
+
])
|
| 269 |
+
self.ln_post = LayerNorm(width)
|
| 270 |
+
|
| 271 |
+
self.whmodel = None
|
| 272 |
+
|
| 273 |
+
self.apply(self.init_transformer)
|
| 274 |
+
self.register_buffer('val_true', torch.zeros(1).cuda())
|
| 275 |
+
self.register_buffer('val_total', torch.zeros(1).cuda())
|
| 276 |
+
|
| 277 |
+
def setup(self, device):
|
| 278 |
+
self.ensure_whisper(device)
|
| 279 |
+
|
| 280 |
+
def init_transformer(self, m):
|
| 281 |
+
if isinstance(m, LinearHead):
|
| 282 |
+
m.no_weight_decay = True
|
| 283 |
+
torch.nn.init.constant_(m.weight, 0)
|
| 284 |
+
elif isinstance(m, QueryHead):
|
| 285 |
+
m.lr_scale = 1/(m.weight.shape[1] / self.base_width)
|
| 286 |
+
torch.nn.init.constant_(m.weight, 0)
|
| 287 |
+
elif isinstance(m, nn.Embedding):
|
| 288 |
+
m.no_weight_decay = True
|
| 289 |
+
m.lr_scale = self.tunables.embeddings_lr_scale
|
| 290 |
+
std = self.tunables.embeddings_std
|
| 291 |
+
torch.nn.init.trunc_normal_(m.weight, std=std, a=-3*std, b=3*std)
|
| 292 |
+
elif isinstance(m, nn.Linear):
|
| 293 |
+
m.lr_scale = 1/(m.weight.shape[1] / self.base_width)
|
| 294 |
+
std = self.tunables.init_std / m.weight.shape[1]
|
| 295 |
+
torch.nn.init.trunc_normal_(m.weight, std=std, a=-3*std, b=3*std)
|
| 296 |
+
if m.bias is not None:
|
| 297 |
+
torch.nn.init.trunc_normal_(m.bias, std=std, a=-3*std, b=3*std)
|
| 298 |
+
elif isinstance(m, nn.LayerNorm):
|
| 299 |
+
m.no_weight_decay = True
|
| 300 |
+
torch.nn.init.constant_(m.bias, 0)
|
| 301 |
+
torch.nn.init.constant_(m.weight, 1)
|
| 302 |
+
|
| 303 |
+
@property
|
| 304 |
+
def device(self):
|
| 305 |
+
return next(self.parameters()).device
|
| 306 |
+
|
| 307 |
+
#
|
| 308 |
+
# training
|
| 309 |
+
#
|
| 310 |
+
@torch.no_grad()
|
| 311 |
+
def extract_teacher(self, samples, input_toks, output_toks):
|
| 312 |
+
embs = self.whmodel[0].encoder(whisper.log_mel_spectrogram(samples))
|
| 313 |
+
teacher_logits = self.whmodel[0].decoder(input_toks, embs)
|
| 314 |
+
# set teacher logits to 0 for padding positions so KLDivLoss ignores them
|
| 315 |
+
teacher_logits[output_toks == -100] = 0
|
| 316 |
+
return embs, teacher_logits
|
| 317 |
+
|
| 318 |
+
def downsample_embeddings(self, x):
|
| 319 |
+
if self.downsample_conv is not None:
|
| 320 |
+
return x[:,::self.downsample] + self.downsample_conv(x.transpose(-1,-2)).transpose(-2,-1)
|
| 321 |
+
elif self.tunables.downsample_mean:
|
| 322 |
+
bs,slen,depth = x.shape
|
| 323 |
+
return x.reshape(bs,slen//self.downsample,self.downsample,depth).mean(-2)
|
| 324 |
+
else:
|
| 325 |
+
return x[:,::self.downsample]
|
| 326 |
+
|
| 327 |
+
def forward(self, samples, mask, input_toks, output_toks):
|
| 328 |
+
embs, teacher_logits = self.extract_teacher(samples, input_toks, output_toks)
|
| 329 |
+
|
| 330 |
+
x = self.downsample_embeddings(embs)
|
| 331 |
+
x = x + self.mlp(self.mlp_ln(x))
|
| 332 |
+
# VQ bottleneck
|
| 333 |
+
quantized, self.indices, self.commit_loss = self.rq(x)
|
| 334 |
+
self.commit_loss = self.commit_loss.mean()
|
| 335 |
+
|
| 336 |
+
x = quantized.repeat_interleave(self.downsample, -2)
|
| 337 |
+
project_out = getattr(self.rq, 'project_out', None) or self.rq.layers[0].project_out
|
| 338 |
+
if self.tunables.mask_embs: x[~mask] = project_out(self.rq.layers[0]._codebook.embed[0,self.vq_codes])
|
| 339 |
+
positions = torch.arange(0, x.shape[-2], dtype=torch.long, device=x.device)
|
| 340 |
+
x = x + self.positional_embedding(positions)
|
| 341 |
+
x = self.ln_post(self.out_blocks(x))
|
| 342 |
+
|
| 343 |
+
logits = self.whmodel[0].decoder(input_toks, x)
|
| 344 |
+
self.ce_loss = self.ce_lossf(logits.view(-1,logits.shape[-1]), output_toks.view(-1))
|
| 345 |
+
self.kl_loss = self.kl_lossf(F.log_softmax(logits, dim=-1), F.softmax(teacher_logits, dim=-1))
|
| 346 |
+
loss = self.ce_loss + self.kl_loss_mul * self.kl_loss + self.commit_loss
|
| 347 |
+
|
| 348 |
+
if not self.training:
|
| 349 |
+
valid_toks = output_toks != -100
|
| 350 |
+
self.val_true += (logits.argmax(-1)[valid_toks] == output_toks[valid_toks]).float().sum()
|
| 351 |
+
self.val_total += valid_toks.float().sum()
|
| 352 |
+
|
| 353 |
+
return x, loss
|
| 354 |
+
|
| 355 |
+
def get_metrics(self):
|
| 356 |
+
metrics = {
|
| 357 |
+
'acc_0': (self.val_true / self.val_total).item(),
|
| 358 |
+
}
|
| 359 |
+
self.val_true[:] = 0
|
| 360 |
+
self.val_total[:] = 0
|
| 361 |
+
return metrics
|
| 362 |
+
|
| 363 |
+
#
|
| 364 |
+
# inference
|
| 365 |
+
#
|
| 366 |
+
@classmethod
|
| 367 |
+
def load_model(cls, ref="collabora/spear-tts-pytorch:whisper-vq-stoks-medium-en+pl.model",
|
| 368 |
+
repo_id=None, filename=None, local_filename=None):
|
| 369 |
+
if repo_id is None and filename is None and local_filename is None:
|
| 370 |
+
if ":" in ref:
|
| 371 |
+
repo_id, filename = ref.split(":", 1)
|
| 372 |
+
else:
|
| 373 |
+
local_filename = ref
|
| 374 |
+
if not local_filename:
|
| 375 |
+
local_filename = hf_hub_download(repo_id=repo_id, filename=filename)
|
| 376 |
+
spec = torch.load(local_filename)
|
| 377 |
+
vqmodel = cls(**spec['config'], tunables=Tunables(**Tunables.upgrade(spec.get('tunables', {}))))
|
| 378 |
+
vqmodel.load_state_dict(spec['state_dict'])
|
| 379 |
+
vqmodel.eval()
|
| 380 |
+
return vqmodel
|
| 381 |
+
|
| 382 |
+
def load_checkpoint(self, local_filename):
|
| 383 |
+
spec = torch.load(local_filename, map_location='cpu')
|
| 384 |
+
assert 'pytorch-lightning_version' in spec, 'not a valid PyTorch Lightning checkpoint'
|
| 385 |
+
state_dict = {k.replace('model.', ''):v
|
| 386 |
+
for k,v in spec['state_dict'].items()}
|
| 387 |
+
self.load_state_dict(state_dict)
|
| 388 |
+
return self
|
| 389 |
+
|
| 390 |
+
def save_model(self, fname, store_parameters=True):
|
| 391 |
+
torch.save(dict(config = self.__stored_args__,
|
| 392 |
+
tunables = dataclasses.asdict(self.tunables),
|
| 393 |
+
state_dict = self.state_dict() if store_parameters else None), fname)
|
| 394 |
+
|
| 395 |
+
def ensure_whisper(self, device):
|
| 396 |
+
# the list wrapper is a hack to make sure the whole of Whisper is not sucked into self.parameters()
|
| 397 |
+
if self.whmodel is None: self.whmodel = [whisper.load_model(self.whisper_model_name, device=device)]
|
| 398 |
+
self.decoding_options = whisper.DecodingOptions()
|
| 399 |
+
multilingual = not self.whisper_model_name.endswith('.en')
|
| 400 |
+
self.tokenizer = whisper.tokenizer.get_tokenizer(multilingual)
|
| 401 |
+
|
| 402 |
+
def quantize(self, embs):
|
| 403 |
+
x = self.downsample_embeddings(embs)
|
| 404 |
+
x = x + self.mlp(self.mlp_ln(x))
|
| 405 |
+
_, stoks, _ = self.rq(x)
|
| 406 |
+
if self.q_depth == 1:
|
| 407 |
+
stoks = stoks.squeeze(-1)
|
| 408 |
+
return stoks
|
| 409 |
+
|
| 410 |
+
def dequantize(self, stoks):
|
| 411 |
+
assert self.q_depth == 1
|
| 412 |
+
assert len(stoks.shape) == 1, "batch processing is not supported"
|
| 413 |
+
if isinstance(stoks, np.ndarray): stoks = torch.tensor(stoks)
|
| 414 |
+
# remove padding
|
| 415 |
+
padding = torch.nonzero(stoks == self.vq_codes)
|
| 416 |
+
if padding.any(): stoks = stoks[:padding[0,0]]
|
| 417 |
+
stoks = F.pad(stoks, (0,self.stoks_len - stoks.shape[-1]), value=self.vq_codes if self.tunables.mask_embs else 0)
|
| 418 |
+
x = self.rq.layers[0]._codebook.embed[0,stoks.to(torch.long).view(-1)]
|
| 419 |
+
x = x.repeat_interleave(self.downsample, -2)
|
| 420 |
+
project_out = getattr(self.rq, 'project_out', None) or self.rq.layers[0].project_out
|
| 421 |
+
x = project_out(x).unsqueeze(0)
|
| 422 |
+
positions = torch.arange(0, x.shape[-2], dtype=torch.long, device=x.device)
|
| 423 |
+
x = x + self.positional_embedding(positions)
|
| 424 |
+
return self.ln_post(self.out_blocks(x))
|
| 425 |
+
|
| 426 |
+
def encode_audio(self, audio):
|
| 427 |
+
if isinstance(audio, str):
|
| 428 |
+
x, sr = torchaudio.load(audio)
|
| 429 |
+
x = torchaudio.transforms.Resample(sr, 16000)(x)[0]
|
| 430 |
+
audio = x.unsqueeze(0)
|
| 431 |
+
return self.encode_mel(whisper.log_mel_spectrogram(audio).to(self.device))
|
| 432 |
+
|
| 433 |
+
def encode_mel(self, mel):
|
| 434 |
+
assert len(mel.shape) == 3, "invalid mel spectrogram shape, expect (batch,chn,time)"
|
| 435 |
+
self.ensure_whisper(self.device)
|
| 436 |
+
n = mel.shape[-1]
|
| 437 |
+
if n > whisper.audio.N_FRAMES:
|
| 438 |
+
padding = 0
|
| 439 |
+
padded = mel[:,:,:whisper.audio.N_FRAMES]
|
| 440 |
+
else:
|
| 441 |
+
padding = -n % whisper.audio.N_FRAMES
|
| 442 |
+
padded = F.pad(mel, (0, padding), value=-1.5)
|
| 443 |
+
embs = self.whmodel[0].encoder(padded)#.to(self.whmodel[0].device))#[:,:n//2]
|
| 444 |
+
stoks = self.quantize(embs)
|
| 445 |
+
if self.tunables.mask_embs:
|
| 446 |
+
return stoks[:,:n//2//self.downsample]
|
| 447 |
+
else:
|
| 448 |
+
return stoks
|
| 449 |
+
|
| 450 |
+
def decode_text(self, stoks, decoding_options=None):
|
| 451 |
+
self.ensure_whisper(self.device)
|
| 452 |
+
if decoding_options is None: decoding_options = self.decoding_options
|
| 453 |
+
embs = self.dequantize(stoks).to(self.whmodel[0].device)
|
| 454 |
+
return self.whmodel[0].decode(embs, decoding_options)
|
| 455 |
+
|
| 456 |
+
# %% ../nbs/2B. Whisper quantization (semantic token) model.ipynb 33
|
| 457 |
+
def make_model(size:str, tunables:Tunables=Tunables(), dataset:torch.utils.data.Dataset=None):
|
| 458 |
+
if size == 'base.en-2d-4096c':
|
| 459 |
+
model = RQBottleneckTransformer(codebook_dim=32, vq_codes=4096, q_depth=1, n_head=8, depth=1,
|
| 460 |
+
downsample=2, threshold_ema_dead_code=0, use_cosine_sim=True,
|
| 461 |
+
whisper_model_name=size.split("-")[0], tunables=tunables)
|
| 462 |
+
return model
|
| 463 |
+
if size == 'base.en-2d-512c':
|
| 464 |
+
model = RQBottleneckTransformer(codebook_dim=32, vq_codes=512, q_depth=1, n_head=8, depth=1,
|
| 465 |
+
downsample=2, threshold_ema_dead_code=0, use_cosine_sim=True,
|
| 466 |
+
whisper_model_name=size.split("-")[0], tunables=tunables)
|
| 467 |
+
return model
|
| 468 |
+
if size == 'base.en-2d-512c-dim64':
|
| 469 |
+
model = RQBottleneckTransformer(codebook_dim=64, vq_codes=512, q_depth=1, n_head=8, depth=1,
|
| 470 |
+
downsample=2, threshold_ema_dead_code=0, use_cosine_sim=True,
|
| 471 |
+
whisper_model_name=size.split("-")[0], tunables=tunables)
|
| 472 |
+
return model
|
| 473 |
+
if size == 'base-2d-512c-dim64':
|
| 474 |
+
model = RQBottleneckTransformer(codebook_dim=64, vq_codes=512, q_depth=1, n_head=8, depth=1,
|
| 475 |
+
downsample=2, threshold_ema_dead_code=0, use_cosine_sim=True,
|
| 476 |
+
whisper_model_name=size.split("-")[0], tunables=tunables)
|
| 477 |
+
return model
|
| 478 |
+
if size == 'base-2d-1024c-dim64':
|
| 479 |
+
model = RQBottleneckTransformer(codebook_dim=64, vq_codes=1024, q_depth=1, n_head=8, depth=1,
|
| 480 |
+
downsample=2, threshold_ema_dead_code=0, use_cosine_sim=True,
|
| 481 |
+
whisper_model_name=size.split("-")[0], tunables=tunables)
|
| 482 |
+
return model
|
| 483 |
+
if size == 'medium-2d-512c-dim64':
|
| 484 |
+
model = RQBottleneckTransformer(codebook_dim=64, vq_codes=512, q_depth=1, n_head=16, depth=1,
|
| 485 |
+
downsample=2, threshold_ema_dead_code=0, use_cosine_sim=True,
|
| 486 |
+
whisper_model_name=size.split("-")[0], tunables=tunables)
|
| 487 |
+
return model
|
| 488 |
+
if size == 'medium-2d-1024c-dim64':
|
| 489 |
+
model = RQBottleneckTransformer(codebook_dim=64, vq_codes=1024, q_depth=1, n_head=16, depth=1,
|
| 490 |
+
downsample=2, threshold_ema_dead_code=0, use_cosine_sim=True,
|
| 491 |
+
whisper_model_name=size.split("-")[0], tunables=tunables)
|
| 492 |
+
return model
|
| 493 |
+
raise ArgumentError(f"invalid model size: {size}")
|
whisperspeech/wer_metrics.py
ADDED
|
@@ -0,0 +1,77 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/C. Word error rate metrics.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = ['librispeech_data', 'DfBuilder', 'WERStats']
|
| 5 |
+
|
| 6 |
+
# %% ../nbs/C. Word error rate metrics.ipynb 2
|
| 7 |
+
import jiwer
|
| 8 |
+
from whisper_normalizer.english import EnglishTextNormalizer
|
| 9 |
+
|
| 10 |
+
import torchaudio
|
| 11 |
+
from pathlib import Path
|
| 12 |
+
import pandas as pd
|
| 13 |
+
|
| 14 |
+
# %% ../nbs/C. Word error rate metrics.ipynb 3
|
| 15 |
+
engnorm = EnglishTextNormalizer()
|
| 16 |
+
def whisper_normalize(x):
|
| 17 |
+
if type(x) == list:
|
| 18 |
+
return [engnorm(y) for y in x]
|
| 19 |
+
else:
|
| 20 |
+
return engnorm(x)
|
| 21 |
+
|
| 22 |
+
default_transform = jiwer.transforms.Compose([
|
| 23 |
+
jiwer.transforms.ToLowerCase(),
|
| 24 |
+
jiwer.transforms.ExpandCommonEnglishContractions(),
|
| 25 |
+
whisper_normalize,
|
| 26 |
+
jiwer.transforms.RemoveMultipleSpaces(),
|
| 27 |
+
jiwer.transforms.Strip(),
|
| 28 |
+
jiwer.transforms.RemovePunctuation(),
|
| 29 |
+
jiwer.transforms.ReduceToListOfListOfWords(),
|
| 30 |
+
])
|
| 31 |
+
|
| 32 |
+
# %% ../nbs/C. Word error rate metrics.ipynb 5
|
| 33 |
+
def librispeech_data(datadir, sample_rate=16000):
|
| 34 |
+
for file in Path(datadir).rglob('*.txt'):
|
| 35 |
+
for line in file.read_text().split('\n'):
|
| 36 |
+
if not line: continue
|
| 37 |
+
idx, text = line.split(" ", 1)
|
| 38 |
+
x, sr = torchaudio.load((file.parent/idx).with_suffix('.flac'))
|
| 39 |
+
if sr != sample_rate:
|
| 40 |
+
x = torchaudio.transforms.Resample(sr, self.sample_rate)(x)
|
| 41 |
+
yield x, text
|
| 42 |
+
|
| 43 |
+
# %% ../nbs/C. Word error rate metrics.ipynb 6
|
| 44 |
+
class DfBuilder:
|
| 45 |
+
def __init__(self):
|
| 46 |
+
self.data = {}
|
| 47 |
+
|
| 48 |
+
def push(self, **kwargs):
|
| 49 |
+
for k,v in kwargs.items():
|
| 50 |
+
if k not in self.data:
|
| 51 |
+
self.data[k] = [v]
|
| 52 |
+
else:
|
| 53 |
+
self.data[k].append(v)
|
| 54 |
+
|
| 55 |
+
def df(self):
|
| 56 |
+
return pd.DataFrame(self.data)
|
| 57 |
+
|
| 58 |
+
# %% ../nbs/C. Word error rate metrics.ipynb 7
|
| 59 |
+
class WERStats(DfBuilder):
|
| 60 |
+
def __init__(self, transform=default_transform):
|
| 61 |
+
super().__init__()
|
| 62 |
+
self.reference_transform = transform
|
| 63 |
+
self.hypothesis_transform = transform
|
| 64 |
+
|
| 65 |
+
def push_sample(self, snd, gt_text, text, idx=None):
|
| 66 |
+
if snd is not None: self.push(secs = snd.shape[-1]/16000)
|
| 67 |
+
diff = jiwer.process_words(gt_text, text, reference_transform=self.reference_transform, hypothesis_transform=self.hypothesis_transform)
|
| 68 |
+
self.push(
|
| 69 |
+
idx = idx,
|
| 70 |
+
gt_text = gt_text,
|
| 71 |
+
text = text,
|
| 72 |
+
wer = diff.wer,
|
| 73 |
+
mer = diff.mer,
|
| 74 |
+
wil = diff.wil,
|
| 75 |
+
wip = diff.wip,
|
| 76 |
+
)
|
| 77 |
+
return diff
|
whisperspeech/wh_transcribe.py
ADDED
|
@@ -0,0 +1,146 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# AUTOGENERATED! DO NOT EDIT! File to edit: ../nbs/2A. Whisper quantization dataset preparation.ipynb.
|
| 2 |
+
|
| 3 |
+
# %% auto 0
|
| 4 |
+
__all__ = []
|
| 5 |
+
|
| 6 |
+
# %% ../nbs/2A. Whisper quantization dataset preparation.ipynb 3
|
| 7 |
+
import os
|
| 8 |
+
import io
|
| 9 |
+
import time
|
| 10 |
+
import torch
|
| 11 |
+
import torchaudio
|
| 12 |
+
|
| 13 |
+
# %% ../nbs/2A. Whisper quantization dataset preparation.ipynb 4
|
| 14 |
+
from pathlib import Path
|
| 15 |
+
import json
|
| 16 |
+
from fastprogress import progress_bar, master_bar
|
| 17 |
+
import numpy as np
|
| 18 |
+
import random
|
| 19 |
+
|
| 20 |
+
import whisper
|
| 21 |
+
|
| 22 |
+
from torch import nn
|
| 23 |
+
import torch.nn.functional as F
|
| 24 |
+
from torch.utils.data.dataloader import DataLoader
|
| 25 |
+
|
| 26 |
+
from fastcore.script import *
|
| 27 |
+
|
| 28 |
+
from . import vad
|
| 29 |
+
import webdataset as wds
|
| 30 |
+
|
| 31 |
+
# %% ../nbs/2A. Whisper quantization dataset preparation.ipynb 9
|
| 32 |
+
# let's make it a bit more conservative
|
| 33 |
+
# with full 30 second chunks it sometimes misses a small part of the transcript
|
| 34 |
+
def random_cutter(dur):
|
| 35 |
+
if random.random() < 0.5:
|
| 36 |
+
return dur > 28 * (random.random()*0.95+0.05)
|
| 37 |
+
else:
|
| 38 |
+
return dur > 28
|
| 39 |
+
|
| 40 |
+
def chunk_merger(segments, should_cut=lambda x: x > 28):
|
| 41 |
+
if len(segments) == 0: return segments
|
| 42 |
+
curr_start = segments[0][0]
|
| 43 |
+
curr_end = 0
|
| 44 |
+
merged = []
|
| 45 |
+
|
| 46 |
+
for ts,te in segments:
|
| 47 |
+
if should_cut(te - curr_start) and curr_end - curr_start > 0:
|
| 48 |
+
merged.append((curr_start, curr_end))
|
| 49 |
+
curr_start = ts
|
| 50 |
+
curr_end = te
|
| 51 |
+
merged.append((curr_start, curr_end))
|
| 52 |
+
return merged
|
| 53 |
+
|
| 54 |
+
# %% ../nbs/2A. Whisper quantization dataset preparation.ipynb 18
|
| 55 |
+
def merge_in(*datasets):
|
| 56 |
+
"""Merge multiple datasets into the current one returning samples with the union of keys.
|
| 57 |
+
|
| 58 |
+
It requires (and validates) all datasets to have the same ordering of keys so you have
|
| 59 |
+
to use it before any sample shuffling. Shard shuffling is ok.
|
| 60 |
+
"""
|
| 61 |
+
def merge_loop(main_samples):
|
| 62 |
+
for samples in zip(*[main_samples]+[iter(x) for x in datasets]):
|
| 63 |
+
key = samples[0]['__key__']
|
| 64 |
+
news = {}
|
| 65 |
+
for s in samples:
|
| 66 |
+
assert s['__key__'] == key
|
| 67 |
+
news.update(s)
|
| 68 |
+
yield news
|
| 69 |
+
return merge_loop
|
| 70 |
+
|
| 71 |
+
# %% ../nbs/2A. Whisper quantization dataset preparation.ipynb 19
|
| 72 |
+
import copy
|
| 73 |
+
|
| 74 |
+
# %% ../nbs/2A. Whisper quantization dataset preparation.ipynb 20
|
| 75 |
+
# a workaround for https://github.com/webdataset/webdataset/issues/297
|
| 76 |
+
# should be possible to use ds.compose here
|
| 77 |
+
def wds_compose(ds, *args):
|
| 78 |
+
ds = copy.copy(ds)
|
| 79 |
+
ds.pipeline = copy.copy(ds.pipeline)
|
| 80 |
+
for f in args:
|
| 81 |
+
ds.append(f)
|
| 82 |
+
return ds
|
| 83 |
+
|
| 84 |
+
# %% ../nbs/2A. Whisper quantization dataset preparation.ipynb 24
|
| 85 |
+
def split_to_chunks(stream, pad_to_seconds=30, random_shift=False):
|
| 86 |
+
for s in stream:
|
| 87 |
+
audio, sr = s.get('flac', s.get('wav', (None, None)))
|
| 88 |
+
if audio is None:
|
| 89 |
+
print(f"warning: '{s['__key__']}' does not contain an audio file")
|
| 90 |
+
continue
|
| 91 |
+
imax = len(s['vad.npy']) - 1
|
| 92 |
+
for i,(ts,te) in enumerate(s['vad.npy']):
|
| 93 |
+
samples = audio[0,int(ts*sr):int(te*sr)]
|
| 94 |
+
if pad_to_seconds is not None:
|
| 95 |
+
padding = pad_to_seconds*sr-samples.shape[-1]
|
| 96 |
+
lpad = random.randint(0, padding) if random_shift else 0
|
| 97 |
+
samples = F.pad(samples, (lpad, padding-lpad))
|
| 98 |
+
yield {"__key__": s['__key__'] + f"_{i:03d}",
|
| 99 |
+
"__url__": s['__url__'],
|
| 100 |
+
"i": i, "imax": imax,
|
| 101 |
+
"tstart": ts, "tend": te, "total_seconds": audio.shape[-1]/sr,
|
| 102 |
+
"lpad": lpad, "rpad": padding-lpad,
|
| 103 |
+
"lpad_s": lpad/sr, "rpad_s": (padding-lpad)/sr,
|
| 104 |
+
"samples": samples, "sample_rate": sr}
|
| 105 |
+
|
| 106 |
+
# %% ../nbs/2A. Whisper quantization dataset preparation.ipynb 38
|
| 107 |
+
def flac_to_txt_name(input, model_size):
|
| 108 |
+
return input.rsplit("/", 1)[1].replace('flac', f'{model_size}-txt') + ".gz"
|
| 109 |
+
|
| 110 |
+
@call_parse
|
| 111 |
+
def process_shard(
|
| 112 |
+
input:str, # input shard URL/path
|
| 113 |
+
output:str=None, # output shard URL/path
|
| 114 |
+
bs:int=None, # batch size (16 uses around 11GB of VRAM)
|
| 115 |
+
n_samples:int=None, # limit the number of samples (useful for quick benchmarking)
|
| 116 |
+
whisper_model:str="base.en" # Whisper model size
|
| 117 |
+
):
|
| 118 |
+
if output is None: output = flac_to_txt_name(input, whisper_model)
|
| 119 |
+
if bs is None: bs = 16
|
| 120 |
+
if n_samples is None: n_samples = 'noinfer'
|
| 121 |
+
else: n_samples = n_samples // bs
|
| 122 |
+
|
| 123 |
+
ds = wds_compose(vad.load_dataset(input),
|
| 124 |
+
merge_in(wds.WebDataset(vad.flac_to_vad_name(input)).decode()),
|
| 125 |
+
wds.map_dict(**{"vad.npy":chunk_merger}),
|
| 126 |
+
split_to_chunks,
|
| 127 |
+
wds.to_tuple('__key__', 'samples'),
|
| 128 |
+
wds.batched(bs),
|
| 129 |
+
)
|
| 130 |
+
dl = DataLoader(ds, num_workers=2, batch_size=None)
|
| 131 |
+
|
| 132 |
+
whmodel = whisper.load_model(whisper_model)
|
| 133 |
+
decoding_options = whisper.DecodingOptions(language='en')
|
| 134 |
+
|
| 135 |
+
tmp = output+".tmp"
|
| 136 |
+
with wds.TarWriter(tmp) as sink:
|
| 137 |
+
for keys, samples in progress_bar(dl, total=n_samples):
|
| 138 |
+
with torch.no_grad():
|
| 139 |
+
embs = whmodel.encoder(whisper.log_mel_spectrogram(samples).cuda())
|
| 140 |
+
decs = whmodel.decode(embs, decoding_options)
|
| 141 |
+
for key, dec in zip(keys, decs):
|
| 142 |
+
sink.write({
|
| 143 |
+
"__key__": key,
|
| 144 |
+
"txt": dec.text,
|
| 145 |
+
})
|
| 146 |
+
os.rename(tmp, output)
|