Spaces:
Running
Running
Commit ·
5b08ad0
1
Parent(s): 32ff648
fix?
Browse files
app.py
CHANGED
|
@@ -4,10 +4,10 @@ import json
|
|
| 4 |
import torch
|
| 5 |
import gc
|
| 6 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 7 |
-
from functools import wraps
|
| 8 |
import threading
|
| 9 |
import sys
|
| 10 |
import argparse
|
|
|
|
| 11 |
|
| 12 |
# 1. SETUP
|
| 13 |
EVAL_MODEL = "HuggingFaceTB/SmolLM2-1.7B-Instruct"
|
|
@@ -165,7 +165,6 @@ def evaluate_submission(file_obj, debug=False):
|
|
| 165 |
print(f"{'='*60}\n")
|
| 166 |
|
| 167 |
# Import with a unique module name each time
|
| 168 |
-
import time
|
| 169 |
module_name = f"student_module_{int(time.time() * 1000000)}"
|
| 170 |
spec = importlib.util.spec_from_file_location(module_name, file_path)
|
| 171 |
student_module = importlib.util.module_from_spec(spec)
|
|
@@ -174,6 +173,8 @@ def evaluate_submission(file_obj, debug=False):
|
|
| 174 |
|
| 175 |
report = [f"## Results:\n"]
|
| 176 |
|
|
|
|
|
|
|
| 177 |
# --- EXERCISE 1 ---
|
| 178 |
ex1_passed = 0
|
| 179 |
ex1_timeout = False
|
|
|
|
| 4 |
import torch
|
| 5 |
import gc
|
| 6 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
|
|
| 7 |
import threading
|
| 8 |
import sys
|
| 9 |
import argparse
|
| 10 |
+
import time
|
| 11 |
|
| 12 |
# 1. SETUP
|
| 13 |
EVAL_MODEL = "HuggingFaceTB/SmolLM2-1.7B-Instruct"
|
|
|
|
| 165 |
print(f"{'='*60}\n")
|
| 166 |
|
| 167 |
# Import with a unique module name each time
|
|
|
|
| 168 |
module_name = f"student_module_{int(time.time() * 1000000)}"
|
| 169 |
spec = importlib.util.spec_from_file_location(module_name, file_path)
|
| 170 |
student_module = importlib.util.module_from_spec(spec)
|
|
|
|
| 173 |
|
| 174 |
report = [f"## Results:\n"]
|
| 175 |
|
| 176 |
+
print("### Loaded student module successfully.")
|
| 177 |
+
|
| 178 |
# --- EXERCISE 1 ---
|
| 179 |
ex1_passed = 0
|
| 180 |
ex1_timeout = False
|