File size: 483 Bytes
05b45a5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
"""Text processing pipeline."""

from .normalizer import normalize_text
from .phonemizer import phonemize
from .text_processor import process_text_chunk, smart_split
from .vocabulary import tokenize


def process_text(text: str) -> list[int]:
    """Process text into token IDs (for backward compatibility)."""
    return process_text_chunk(text)


__all__ = [
    "normalize_text",
    "phonemize",
    "tokenize",
    "process_text",
    "process_text_chunk",
    "smart_split",
]