Mr-FineTuner commited on
Commit
0f1cbd5
·
verified ·
1 Parent(s): 255f2c1

Add model card with exact and within-1 confusion matrices and per-class metrics

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -24,7 +24,7 @@ This is a fine-tuned version of `unsloth/gemma-7b-bnb-4bit` for CEFR-level sente
24
  - **Other Metrics**:
25
  - Perplexity: 5.344
26
  - Diversity (Unique Sentences): 0.100
27
- - Inference Time (ms): 5869.541
28
  - Model Size (GB): 4.8
29
  - Robustness (F1): 0.045
30
  - **Confusion Matrix (Exact Matches)**:
@@ -51,8 +51,8 @@ This is a fine-tuned version of `unsloth/gemma-7b-bnb-4bit` for CEFR-level sente
51
  ```python
52
  from transformers import AutoModelForCausalLM, AutoTokenizer
53
 
54
- model = AutoModelForCausalLM.from_pretrained("Mr-FineTuner/Test_02_llama_trainPercen_myValidator")
55
- tokenizer = AutoTokenizer.from_pretrained("Mr-FineTuner/Test_02_llama_trainPercen_myValidator")
56
 
57
  # Example inference
58
  prompt = "<|user|>Generate a CEFR B1 level sentence.<|end|>"
 
24
  - **Other Metrics**:
25
  - Perplexity: 5.344
26
  - Diversity (Unique Sentences): 0.100
27
+ - Inference Time (ms): 5802.883
28
  - Model Size (GB): 4.8
29
  - Robustness (F1): 0.045
30
  - **Confusion Matrix (Exact Matches)**:
 
51
  ```python
52
  from transformers import AutoModelForCausalLM, AutoTokenizer
53
 
54
+ model = AutoModelForCausalLM.from_pretrained("Mr-FineTuner/Test_02_llama_trainPercen_myValidator_2ndTry")
55
+ tokenizer = AutoTokenizer.from_pretrained("Mr-FineTuner/Test_02_llama_trainPercen_myValidator_2ndTry")
56
 
57
  # Example inference
58
  prompt = "<|user|>Generate a CEFR B1 level sentence.<|end|>"