yoru_tomosu commited on
Commit
010a816
1 Parent(s): 3b2268f

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +16 -4
app.py CHANGED
@@ -20,25 +20,37 @@ def transcribe(audio):
20
  mel = whisper.log_mel_spectrogram(audio).to(model.device)
21
 
22
 
23
- # detect the spoken language
24
- # mel = mel.to("float32") # もしくは mel = mel.to("float32")
25
  _, probs = model.detect_language(mel)
26
 
27
  print(f"Detected language: {max(probs, key=probs.get)}")
28
  detect_lang = max(probs, key=probs.get)
29
 
 
 
30
  # decode the audio
31
  # options = whisper.DecodingOptions()
32
  options = whisper.DecodingOptions(fp16 = False)
33
  result = whisper.decode(model, mel, options)
34
 
35
-
 
 
 
 
 
 
 
 
 
36
  translated_text = translate(result.text, "JA")
37
  return translated_text
38
 
 
 
39
  import gradio as gr
40
 
41
- title = 'Video to Translator'
42
 
43
  inputs = gr.Video()
44
  outputs = gr.Text()
 
20
  mel = whisper.log_mel_spectrogram(audio).to(model.device)
21
 
22
 
23
+ # detect the spoken language
 
24
  _, probs = model.detect_language(mel)
25
 
26
  print(f"Detected language: {max(probs, key=probs.get)}")
27
  detect_lang = max(probs, key=probs.get)
28
 
29
+
30
+
31
  # decode the audio
32
  # options = whisper.DecodingOptions()
33
  options = whisper.DecodingOptions(fp16 = False)
34
  result = whisper.decode(model, mel, options)
35
 
36
+ if detect_lang == "en":
37
+ print("Text: ", result.text)
38
+ translated_text = translate(result.text, "JA")
39
+ print("translated_text: ", translated_text)
40
+ # generated_video = text_to_speech(translated_text)
41
+ # print("generated_video 01: ", generated_video)
42
+ elif detect_lang == "ja":
43
+ print("Text: ", result.text)
44
+ translated_text = translate(result.text, "EN-US")
45
+
46
  translated_text = translate(result.text, "JA")
47
  return translated_text
48
 
49
+
50
+
51
  import gradio as gr
52
 
53
+ title = 'Translator_Video'
54
 
55
  inputs = gr.Video()
56
  outputs = gr.Text()