Text Generation
Transformers
Safetensors
qwen3
llama-factory
full
Generated from Trainer
conversational
text-generation-inference
Instructions to use DCAgent/a1-magicoder with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use DCAgent/a1-magicoder with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="DCAgent/a1-magicoder") messages = [ {"role": "user", "content": "Who are you?"}, ] pipe(messages)# Load model directly from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("DCAgent/a1-magicoder") model = AutoModelForCausalLM.from_pretrained("DCAgent/a1-magicoder") messages = [ {"role": "user", "content": "Who are you?"}, ] inputs = tokenizer.apply_chat_template( messages, add_generation_prompt=True, tokenize=True, return_dict=True, return_tensors="pt", ).to(model.device) outputs = model.generate(**inputs, max_new_tokens=40) print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:])) - Notebooks
- Google Colab
- Kaggle
- Local Apps
- vLLM
How to use DCAgent/a1-magicoder with vLLM:
Install from pip and serve model
# Install vLLM from pip: pip install vllm # Start the vLLM server: vllm serve "DCAgent/a1-magicoder" # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "DCAgent/a1-magicoder", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'Use Docker
docker model run hf.co/DCAgent/a1-magicoder
- SGLang
How to use DCAgent/a1-magicoder with SGLang:
Install from pip and serve model
# Install SGLang from pip: pip install sglang # Start the SGLang server: python3 -m sglang.launch_server \ --model-path "DCAgent/a1-magicoder" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "DCAgent/a1-magicoder", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }'Use Docker images
docker run --gpus all \ --shm-size 32g \ -p 30000:30000 \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HF_TOKEN=<secret>" \ --ipc=host \ lmsysorg/sglang:latest \ python3 -m sglang.launch_server \ --model-path "DCAgent/a1-magicoder" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "DCAgent/a1-magicoder", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }' - Docker Model Runner
How to use DCAgent/a1-magicoder with Docker Model Runner:
docker model run hf.co/DCAgent/a1-magicoder
Upload folder using huggingface_hub
Browse files- README.md +1 -1
- all_results.json +12 -12
- model-00001-of-00004.safetensors +1 -1
- model-00002-of-00004.safetensors +1 -1
- model-00003-of-00004.safetensors +1 -1
- model-00004-of-00004.safetensors +1 -1
- run_summary.json +2 -2
- train_results.json +12 -12
- trainer_log.jsonl +0 -0
- training_loss.png +0 -0
README.md
CHANGED
|
@@ -16,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 16 |
|
| 17 |
# sft_a1_magicoder__Qwen3-8B
|
| 18 |
|
| 19 |
-
This model is a fine-tuned version of [Qwen/Qwen3-8B](https://huggingface.co/Qwen/Qwen3-8B) on the /e/scratch/jureap59/raoof1/sft_data/hf_hub/datasets--DCAgent--
|
| 20 |
|
| 21 |
## Model description
|
| 22 |
|
|
|
|
| 16 |
|
| 17 |
# sft_a1_magicoder__Qwen3-8B
|
| 18 |
|
| 19 |
+
This model is a fine-tuned version of [Qwen/Qwen3-8B](https://huggingface.co/Qwen/Qwen3-8B) on the /e/scratch/jureap59/raoof1/sft_data/hf_hub/datasets--DCAgent--Magicoder-Evol-Instruct-110K-sandboxes-1_10k_glm_4.7_traces_jupiter/snapshots/5aeec71d5cdc4cd588e71eed903e3ab2f7b35051_thinking_preprocessed dataset.
|
| 20 |
|
| 21 |
## Model description
|
| 22 |
|
all_results.json
CHANGED
|
@@ -1,16 +1,16 @@
|
|
| 1 |
{
|
| 2 |
-
"achieved_tflops_per_gpu": 0.
|
| 3 |
-
"achieved_tflops_per_gpu_theoretical":
|
| 4 |
"epoch": 7.0,
|
| 5 |
"loss_nan_ranks": 0,
|
| 6 |
-
"loss_rank_avg": 0.
|
| 7 |
-
"mfu_percent": 0.
|
| 8 |
-
"mfu_percent_theoretical":
|
| 9 |
-
"total_flos":
|
| 10 |
-
"train_loss": 0.
|
| 11 |
-
"train_runtime":
|
| 12 |
-
"train_samples_per_second":
|
| 13 |
-
"train_steps_per_second": 0.
|
| 14 |
-
"valid_targets_mean":
|
| 15 |
-
"valid_targets_min":
|
| 16 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"achieved_tflops_per_gpu": 0.002110187398451811,
|
| 3 |
+
"achieved_tflops_per_gpu_theoretical": 650.4776760620482,
|
| 4 |
"epoch": 7.0,
|
| 5 |
"loss_nan_ranks": 0,
|
| 6 |
+
"loss_rank_avg": 0.2384590357542038,
|
| 7 |
+
"mfu_percent": 0.00014912985148069337,
|
| 8 |
+
"mfu_percent_theoretical": 45.97015378530376,
|
| 9 |
+
"total_flos": 688651439570944.0,
|
| 10 |
+
"train_loss": 0.26049303336732077,
|
| 11 |
+
"train_runtime": 20396.6316,
|
| 12 |
+
"train_samples_per_second": 3.064,
|
| 13 |
+
"train_steps_per_second": 0.192,
|
| 14 |
+
"valid_targets_mean": 4082.4,
|
| 15 |
+
"valid_targets_min": 866
|
| 16 |
}
|
model-00001-of-00004.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4902257696
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:daffa32e7444e736a1c517d62c2a55d9f96260371213ac37e97ac9827b55b478
|
| 3 |
size 4902257696
|
model-00002-of-00004.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4915960368
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dfe4bb96e03e07782586ca63884f59543a475c5932ac4ccf440848b3736ff540
|
| 3 |
size 4915960368
|
model-00003-of-00004.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4983068496
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cc931313941b618ca7de44a63927fa3a1e43ed4b723a103c01493d683889d17b
|
| 3 |
size 4983068496
|
model-00004-of-00004.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1580230264
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7ab55fd4c4fd1cdb8d41f5aee283940231d28eba52cc99ebf70cf3cbf11b598d
|
| 3 |
size 1580230264
|
run_summary.json
CHANGED
|
@@ -1,10 +1,10 @@
|
|
| 1 |
{
|
| 2 |
-
"agent_name": "
|
| 3 |
"training_start": null,
|
| 4 |
"training_end": null,
|
| 5 |
"created_by": "raoof1",
|
| 6 |
"base_model_name": "Qwen/Qwen3-8B",
|
| 7 |
-
"dataset_name": "/e/scratch/jureap59/raoof1/sft_data/hf_hub/datasets--DCAgent--
|
| 8 |
"training_type": "SFT",
|
| 9 |
"training_parameters": "https://huggingface.co/DCAgent/a1-magicoder/blob/main/config.json",
|
| 10 |
"wandb_link": null,
|
|
|
|
| 1 |
{
|
| 2 |
+
"agent_name": "5aeec71d5cdc4cd588e71eed903e3ab2f7b35051_thinking_preprocessed",
|
| 3 |
"training_start": null,
|
| 4 |
"training_end": null,
|
| 5 |
"created_by": "raoof1",
|
| 6 |
"base_model_name": "Qwen/Qwen3-8B",
|
| 7 |
+
"dataset_name": "/e/scratch/jureap59/raoof1/sft_data/hf_hub/datasets--DCAgent--Magicoder-Evol-Instruct-110K-sandboxes-1_10k_glm_4.7_traces_jupiter/snapshots/5aeec71d5cdc4cd588e71eed903e3ab2f7b35051_thinking_preprocessed",
|
| 8 |
"training_type": "SFT",
|
| 9 |
"training_parameters": "https://huggingface.co/DCAgent/a1-magicoder/blob/main/config.json",
|
| 10 |
"wandb_link": null,
|
train_results.json
CHANGED
|
@@ -1,16 +1,16 @@
|
|
| 1 |
{
|
| 2 |
-
"achieved_tflops_per_gpu": 0.
|
| 3 |
-
"achieved_tflops_per_gpu_theoretical":
|
| 4 |
"epoch": 7.0,
|
| 5 |
"loss_nan_ranks": 0,
|
| 6 |
-
"loss_rank_avg": 0.
|
| 7 |
-
"mfu_percent": 0.
|
| 8 |
-
"mfu_percent_theoretical":
|
| 9 |
-
"total_flos":
|
| 10 |
-
"train_loss": 0.
|
| 11 |
-
"train_runtime":
|
| 12 |
-
"train_samples_per_second":
|
| 13 |
-
"train_steps_per_second": 0.
|
| 14 |
-
"valid_targets_mean":
|
| 15 |
-
"valid_targets_min":
|
| 16 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"achieved_tflops_per_gpu": 0.002110187398451811,
|
| 3 |
+
"achieved_tflops_per_gpu_theoretical": 650.4776760620482,
|
| 4 |
"epoch": 7.0,
|
| 5 |
"loss_nan_ranks": 0,
|
| 6 |
+
"loss_rank_avg": 0.2384590357542038,
|
| 7 |
+
"mfu_percent": 0.00014912985148069337,
|
| 8 |
+
"mfu_percent_theoretical": 45.97015378530376,
|
| 9 |
+
"total_flos": 688651439570944.0,
|
| 10 |
+
"train_loss": 0.26049303336732077,
|
| 11 |
+
"train_runtime": 20396.6316,
|
| 12 |
+
"train_samples_per_second": 3.064,
|
| 13 |
+
"train_steps_per_second": 0.192,
|
| 14 |
+
"valid_targets_mean": 4082.4,
|
| 15 |
+
"valid_targets_min": 866
|
| 16 |
}
|
trainer_log.jsonl
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
training_loss.png
CHANGED
|
|