bfcl-trending / README.md
dvilasuero's picture
Upload README.md with huggingface_hub
1fbd39e verified
---
title: Inspect Evals/bfcl
emoji: 📊
colorFrom: blue
colorTo: purple
sdk: docker
sdk_version: "latest"
pinned: false
---
# bfcl
This eval was run using [evaljobs](https://github.com/dvsrepo/evaljobs).
## Command
```bash
evaljobs inspect_evals/bfcl \
--model hf-inference-providers/moonshotai/Kimi-K2-Thinking,hf-inference-providers/meta-llama/Llama-3.1-8B-Instruct,hf-inference-providers/allenai/Olmo-3-32B-Think,hf-inference-providers/openai/gpt-oss-20b,hf-inference-providers/zai-org/GLM-4.6,hf-inference-providers/allenai/Olmo-3-7B-Instruct,hf-inference-providers/allenai/Olmo-3-7B-Think,hf-inference-providers/openai/gpt-oss-120b,hf-inference-providers/deepseek-ai/DeepSeek-V3.2-Exp,hf-inference-providers/meta-llama/Llama-3.2-3B-Instruct \
--name bfcl-trending
```
## Run with other models
To run this eval with a different model, use:
```bash
evaljobs inspect_evals/bfcl \
--model <your-model> \
--name <your-name> \
--flavor cpu-basic
```
## Inspect eval command
The eval was executed with:
```bash
inspect eval-set inspect_evals/bfcl \
--model hf-inference-providers/moonshotai/Kimi-K2-Thinking,hf-inference-providers/meta-llama/Llama-3.1-8B-Instruct,hf-inference-providers/allenai/Olmo-3-32B-Think,hf-inference-providers/openai/gpt-oss-20b,hf-inference-providers/zai-org/GLM-4.6,hf-inference-providers/allenai/Olmo-3-7B-Instruct,hf-inference-providers/allenai/Olmo-3-7B-Think,hf-inference-providers/openai/gpt-oss-120b,hf-inference-providers/deepseek-ai/DeepSeek-V3.2-Exp,hf-inference-providers/meta-llama/Llama-3.2-3B-Instruct \
--log-shared \
--log-buffer 100
```