aka7774 commited on
Commit
153a23f
·
1 Parent(s): c873483

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +9 -9
app.py CHANGED
@@ -1,11 +1,11 @@
1
  import gradio as gr
2
  import whisper
3
- from faster_whisper import WhisperModel
4
 
5
  model_size = 'large-v3'
6
- #model = whisper.load_model(model_size)
7
  #model = WhisperModel(model_size, device="cuda", compute_type="float16")
8
- model = WhisperModel(model_size, compute_type="float16")
9
 
10
  # or run on GPU with INT8
11
  # model = WhisperModel(model_size, device="cuda", compute_type="int8_float16")
@@ -16,14 +16,14 @@ def speech_to_text(audio_file, _model_size):
16
  global model_size, model
17
  if model_size != _model_size:
18
  model_size = _model_size
19
- #model = whisper.load_model(model_size)
20
- model = WhisperModel(model_size, compute_type="float16")
21
 
22
- #result = model.transcribe(audio_file)
23
- segments, info = model.transcribe(audio_file, beam_size=5)
24
 
25
- # return result["text"]
26
- return "".join([segment.text for segment in segments])
27
 
28
  gr.Interface(
29
  fn=speech_to_text,
 
1
  import gradio as gr
2
  import whisper
3
+ #from faster_whisper import WhisperModel
4
 
5
  model_size = 'large-v3'
6
+ model = whisper.load_model(model_size)
7
  #model = WhisperModel(model_size, device="cuda", compute_type="float16")
8
+ #model = WhisperModel(model_size, compute_type="float16")
9
 
10
  # or run on GPU with INT8
11
  # model = WhisperModel(model_size, device="cuda", compute_type="int8_float16")
 
16
  global model_size, model
17
  if model_size != _model_size:
18
  model_size = _model_size
19
+ model = whisper.load_model(model_size)
20
+ #model = WhisperModel(model_size, compute_type="float16")
21
 
22
+ result = model.transcribe(audio_file)
23
+ #segments, info = model.transcribe(audio_file, beam_size=5)
24
 
25
+ return result["text"]
26
+ #return "".join([segment.text for segment in segments])
27
 
28
  gr.Interface(
29
  fn=speech_to_text,