MackenzieTest / README.md
mackenzietechdocs's picture
Add Artificial Analysis evaluations for ministral-14b
203131a verified
|
raw
history blame
1.35 kB
---
license: mit
library_name: transformers
pipeline_tag: text-generation
base_model:
- deepseek-ai/DeepSeek-V3.2-Exp-Base
base_model_relation: finetune
model-index:
- name: MyPrivateModel
results:
- task:
type: text-generation
name: Text Generation
dataset:
name: DummyEval Benchmark
type: dummy-eval-benchmark
metrics:
- type: accuracy
value: 0.0
name: accuracy
---
# MyPrivateModel
This is a **placeholder model card** for a private repository used for experimentation and evaluation tooling.
## Model description
- **Owner:** my-username
- **Base model:** `deepseek-ai/DeepSeek-V3.2-Exp-Base`
- **Intended use:** Internal testing of evaluation workflows, HF Skills tasks, and model card metadata.
## Evaluation
The `model-index` block above contains **dummy evaluation data** (0.0 accuracy on a fictitious `DummyEval Benchmark`).
This exists only to ensure the model card passes Hugging Face metadata validation and can be used to test:
- `hf_model_evaluation` tools
- `import-aa`, `extract-readme`, and related commands
- Leaderboard / evaluation display on the Hub
## Limitations
- The evaluation numbers are **not real** and must not be used as performance claims.
- This repository is intended only for private experimentation.