File size: 3,293 Bytes
81fef9c
 
 
 
 
69dc570
81fef9c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
69dc570
 
81fef9c
69dc570
81fef9c
 
 
 
 
 
 
 
 
 
 
 
 
69dc570
 
81fef9c
69dc570
 
 
 
81fef9c
 
 
 
69dc570
 
 
 
 
 
 
 
 
 
81fef9c
 
 
 
 
 
 
69dc570
 
 
81fef9c
 
 
 
 
 
 
 
 
69dc570
81fef9c
69dc570
81fef9c
 
 
 
 
69dc570
 
 
 
 
 
 
 
81fef9c
 
69dc570
 
81fef9c
 
69dc570
81fef9c
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
import json
import os
from datetime import datetime, timezone

from src.display.formatting import styled_error, styled_message, styled_warning
from src.envs import API, EVAL_REQUESTS_PATH, LOCAL_EVAL_REQUESTS_PATH, QUEUE_REPO, TOKEN, has_remote_backend
from src.submission.check_validity import (
    already_submitted_models,
    check_model_card,
    get_model_size,
    is_model_on_hub,
)

REQUESTED_MODELS = None
USERS_TO_SUBMISSION_DATES = None

def add_new_eval(
    model: str,
    revision: str,
    precision: str,
):
    global REQUESTED_MODELS
    global USERS_TO_SUBMISSION_DATES
    requests_path = EVAL_REQUESTS_PATH if has_remote_backend() else LOCAL_EVAL_REQUESTS_PATH

    if not REQUESTED_MODELS:
        REQUESTED_MODELS, USERS_TO_SUBMISSION_DATES = already_submitted_models(requests_path)

    user_name = ""
    model_path = model
    if "/" in model:
        user_name = model.split("/")[0]
        model_path = model.split("/")[1]

    precision = precision.split(" ")[0]
    current_time = datetime.now(timezone.utc).strftime("%Y-%m-%dT%H:%M:%SZ")

    if revision == "":
        revision = "main"

    # Seems good, creating the eval
    print("Adding new eval")

    license = "?"
    model_size = 0
    likes = 0
    if has_remote_backend():
        model_on_hub, error, _ = is_model_on_hub(model_name=model, revision=revision, token=TOKEN, test_tokenizer=True)
        if not model_on_hub:
            return styled_error(f'Model "{model}" {error}')

        try:
            model_info = API.model_info(repo_id=model, revision=revision)
            model_size = get_model_size(model_info=model_info, precision=precision)
            likes = model_info.likes
            license = model_info.cardData.get("license", "?")
            modelcard_OK, error_msg = check_model_card(model)
            if not modelcard_OK:
                return styled_error(error_msg)
        except Exception:
            return styled_error("Could not get your model information from the Hub.")

    eval_entry = {
        "model": model,
        "revision": revision,
        "precision": precision,
        "status": "PENDING",
        "submitted_time": current_time,
        "model_type": "pretrained",
        "weight_type": "Original",
        "likes": likes,
        "params": model_size,
        "license": license,
    }

    # Check for duplicate submission
    if f"{model}_{revision}_{precision}" in REQUESTED_MODELS:
        return styled_warning("This model has been already submitted.")

    print("Creating eval file")
    OUT_DIR = f"{requests_path}/{user_name}"
    os.makedirs(OUT_DIR, exist_ok=True)
    out_path = f"{OUT_DIR}/{model_path}_eval_request_False_{precision}_Original.json"

    with open(out_path, "w") as f:
        f.write(json.dumps(eval_entry))

    print("Uploading eval file")
    if has_remote_backend():
        API.upload_file(
            path_or_fileobj=out_path,
            path_in_repo=out_path.split("eval-queue/")[1],
            repo_id=QUEUE_REPO,
            repo_type="dataset",
            commit_message=f"Add {model} to eval queue",
        )

    # Remove the local file
    if has_remote_backend():
        os.remove(out_path)

    return styled_message(
        "Your request has been submitted to the evaluation queue."
    )