Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -24,6 +24,7 @@ def process_video_audio(video_path):
|
|
| 24 |
train_visual = pt.zeros([1, 120, 120, 3, 10])
|
| 25 |
train_audio_wave = pt.zeros([1, 261540])
|
| 26 |
train_audio_cnn = pt.zeros([1, 150, 512, 1])
|
|
|
|
| 27 |
|
| 28 |
mfcc = torchaudio.transforms.MFCC(n_mfcc=150, melkwargs={"n_fft": 1022, "n_mels": 150})
|
| 29 |
|
|
@@ -70,7 +71,7 @@ def process_video_audio(video_path):
|
|
| 70 |
def predict_emotion(video_path):
|
| 71 |
last_frame, audio_path, train_visual, train_audio_wave, train_audio_cnn = process_video_audio(video_path)
|
| 72 |
|
| 73 |
-
model = load_model("
|
| 74 |
|
| 75 |
predictions = model.predict({
|
| 76 |
"input_visual": train_visual,
|
|
@@ -82,7 +83,7 @@ def predict_emotion(video_path):
|
|
| 82 |
return last_frame, audio_path, predicted_label
|
| 83 |
|
| 84 |
def predict_emotion_gradio(video_path):
|
| 85 |
-
emotion_dict = {0: 'neutral', 1: 'calm', 2: 'happy', 3: 'sad', 4: 'angry', 5: 'fearful'}
|
| 86 |
last_frame, audio_path, predicted_label = predict_emotion(video_path)
|
| 87 |
predicted_emotion = emotion_dict[predicted_label]
|
| 88 |
return last_frame, audio_path, predicted_emotion
|
|
|
|
| 24 |
train_visual = pt.zeros([1, 120, 120, 3, 10])
|
| 25 |
train_audio_wave = pt.zeros([1, 261540])
|
| 26 |
train_audio_cnn = pt.zeros([1, 150, 512, 1])
|
| 27 |
+
|
| 28 |
|
| 29 |
mfcc = torchaudio.transforms.MFCC(n_mfcc=150, melkwargs={"n_fft": 1022, "n_mels": 150})
|
| 30 |
|
|
|
|
| 71 |
def predict_emotion(video_path):
|
| 72 |
last_frame, audio_path, train_visual, train_audio_wave, train_audio_cnn = process_video_audio(video_path)
|
| 73 |
|
| 74 |
+
model = load_model("model_vui_ve2392.keras")
|
| 75 |
|
| 76 |
predictions = model.predict({
|
| 77 |
"input_visual": train_visual,
|
|
|
|
| 83 |
return last_frame, audio_path, predicted_label
|
| 84 |
|
| 85 |
def predict_emotion_gradio(video_path):
|
| 86 |
+
emotion_dict = {0: 'neutral', 1: 'calm', 2: 'happy', 3: 'sad', 4: 'angry', 5: 'fearful', 6: 'disgust', 7: 'surprised'}
|
| 87 |
last_frame, audio_path, predicted_label = predict_emotion(video_path)
|
| 88 |
predicted_emotion = emotion_dict[predicted_label]
|
| 89 |
return last_frame, audio_path, predicted_emotion
|