THP2903 commited on
Commit
5037272
·
verified ·
1 Parent(s): 62b8b49

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -2
app.py CHANGED
@@ -24,6 +24,7 @@ def process_video_audio(video_path):
24
  train_visual = pt.zeros([1, 120, 120, 3, 10])
25
  train_audio_wave = pt.zeros([1, 261540])
26
  train_audio_cnn = pt.zeros([1, 150, 512, 1])
 
27
 
28
  mfcc = torchaudio.transforms.MFCC(n_mfcc=150, melkwargs={"n_fft": 1022, "n_mels": 150})
29
 
@@ -70,7 +71,7 @@ def process_video_audio(video_path):
70
  def predict_emotion(video_path):
71
  last_frame, audio_path, train_visual, train_audio_wave, train_audio_cnn = process_video_audio(video_path)
72
 
73
- model = load_model("model_vui_ve1024.keras")
74
 
75
  predictions = model.predict({
76
  "input_visual": train_visual,
@@ -82,7 +83,7 @@ def predict_emotion(video_path):
82
  return last_frame, audio_path, predicted_label
83
 
84
  def predict_emotion_gradio(video_path):
85
- emotion_dict = {0: 'neutral', 1: 'calm', 2: 'happy', 3: 'sad', 4: 'angry', 5: 'fearful'}
86
  last_frame, audio_path, predicted_label = predict_emotion(video_path)
87
  predicted_emotion = emotion_dict[predicted_label]
88
  return last_frame, audio_path, predicted_emotion
 
24
  train_visual = pt.zeros([1, 120, 120, 3, 10])
25
  train_audio_wave = pt.zeros([1, 261540])
26
  train_audio_cnn = pt.zeros([1, 150, 512, 1])
27
+
28
 
29
  mfcc = torchaudio.transforms.MFCC(n_mfcc=150, melkwargs={"n_fft": 1022, "n_mels": 150})
30
 
 
71
  def predict_emotion(video_path):
72
  last_frame, audio_path, train_visual, train_audio_wave, train_audio_cnn = process_video_audio(video_path)
73
 
74
+ model = load_model("model_vui_ve2392.keras")
75
 
76
  predictions = model.predict({
77
  "input_visual": train_visual,
 
83
  return last_frame, audio_path, predicted_label
84
 
85
  def predict_emotion_gradio(video_path):
86
+ emotion_dict = {0: 'neutral', 1: 'calm', 2: 'happy', 3: 'sad', 4: 'angry', 5: 'fearful', 6: 'disgust', 7: 'surprised'}
87
  last_frame, audio_path, predicted_label = predict_emotion(video_path)
88
  predicted_emotion = emotion_dict[predicted_label]
89
  return last_frame, audio_path, predicted_emotion