Update orpheus-tts/engine_class.py
Browse files
orpheus-tts/engine_class.py
CHANGED
@@ -78,7 +78,7 @@ class OrpheusModel:
|
|
78 |
gpu_memory_utilization=0.85,
|
79 |
max_model_len=8192,
|
80 |
trust_remote_code=True,
|
81 |
-
enforce_eager=
|
82 |
**self.engine_kwargs
|
83 |
)
|
84 |
|
@@ -107,8 +107,9 @@ class OrpheusModel:
|
|
107 |
# end_tokens = [128009, 128260]
|
108 |
# Aber für Audio-Generierung brauchen wir auch das Audio-Start-Token!
|
109 |
|
|
|
110 |
start_token = torch.tensor([[128259]], dtype=torch.int64)
|
111 |
-
end_tokens = torch.tensor([[128009, 128260
|
112 |
input_ids = self.tokenizer(full_prompt, return_tensors="pt").input_ids
|
113 |
|
114 |
print(f"DEBUG KARTOFFEL: Original prompt: '{full_prompt}'")
|
|
|
78 |
gpu_memory_utilization=0.85,
|
79 |
max_model_len=8192,
|
80 |
trust_remote_code=True,
|
81 |
+
enforce_eager=True, # Disable CUDA graphs for better compatibility
|
82 |
**self.engine_kwargs
|
83 |
)
|
84 |
|
|
|
107 |
# end_tokens = [128009, 128260]
|
108 |
# Aber für Audio-Generierung brauchen wir auch das Audio-Start-Token!
|
109 |
|
110 |
+
# Exakt wie in der Referenz-Implementierung
|
111 |
start_token = torch.tensor([[128259]], dtype=torch.int64)
|
112 |
+
end_tokens = torch.tensor([[128009, 128260]], dtype=torch.int64)
|
113 |
input_ids = self.tokenizer(full_prompt, return_tensors="pt").input_ids
|
114 |
|
115 |
print(f"DEBUG KARTOFFEL: Original prompt: '{full_prompt}'")
|