Update metric.py
Browse files
metric.py
CHANGED
|
@@ -58,7 +58,7 @@ def heading(text: str) -> str:
|
|
| 58 |
return "\n" + f" {text} ".center(80, "=") + "\n"
|
| 59 |
|
| 60 |
|
| 61 |
-
def chat_llm_batch(model_id, prompts, limit=
|
| 62 |
parameters = TextGenerationParameters(
|
| 63 |
# decoding_method=DecodingMethod.SAMPLE, max_new_tokens=128, min_new_tokens=30, temperature=0, top_k=50, top_p=1, random_seed=42
|
| 64 |
decoding_method=DecodingMethod.SAMPLE, max_new_tokens=128, min_new_tokens=30, temperature=0, top_k=1, top_p=1, random_seed=42,
|
|
@@ -518,7 +518,7 @@ def compute(params):
|
|
| 518 |
'phase2_overall_rating' # phase 2 overall rating
|
| 519 |
]
|
| 520 |
mixtral_eval_batch_size = 16
|
| 521 |
-
limit =
|
| 522 |
public_category_score_dict = {}
|
| 523 |
private_category_score_dict = {}
|
| 524 |
for key_ in all_keys:
|
|
|
|
| 58 |
return "\n" + f" {text} ".center(80, "=") + "\n"
|
| 59 |
|
| 60 |
|
| 61 |
+
def chat_llm_batch(model_id, prompts, limit= 20):
|
| 62 |
parameters = TextGenerationParameters(
|
| 63 |
# decoding_method=DecodingMethod.SAMPLE, max_new_tokens=128, min_new_tokens=30, temperature=0, top_k=50, top_p=1, random_seed=42
|
| 64 |
decoding_method=DecodingMethod.SAMPLE, max_new_tokens=128, min_new_tokens=30, temperature=0, top_k=1, top_p=1, random_seed=42,
|
|
|
|
| 518 |
'phase2_overall_rating' # phase 2 overall rating
|
| 519 |
]
|
| 520 |
mixtral_eval_batch_size = 16
|
| 521 |
+
limit = 100 # concurrency limit
|
| 522 |
public_category_score_dict = {}
|
| 523 |
private_category_score_dict = {}
|
| 524 |
for key_ in all_keys:
|