YAML Metadata Warning: empty or missing yaml metadata in repo card

Check out the documentation for more information.

DeepSeek-R1-Distill-Qwen-7B β€” Python Code Fine-tune

A LoRA fine-tuned version of DeepSeek-R1-Distill-Qwen-7B specialized for Python code generation.

Model Details

Model Description

Model Sources

Uses

Direct Use

Generate Python code from natural language instructions. Examples:

  • Writing functions, classes, algorithms
  • Async/await patterns
  • Data structures and error handling

Out-of-Scope Use

  • Not intended for other programming languages
  • Not suitable for production security-critical code without review

Bias, Risks, and Limitations

Generated code should always be reviewed before use in production. The model may occasionally produce syntactically incorrect code, particularly for complex async patterns.

Training Details

Training Data

iamtarun/python_code_instructions_18k_alpaca β€” 18,612 Python code instruction/response pairs.

  • Train split: 17,681 examples
  • Validation split: 931 examples

Training Hyperparameters

Parameter Value
Method LoRA
LoRA Rank 8
LoRA Layers 8
Learning Rate 5e-6
Batch Size 2
Iterations 2000
Quantization 4-bit

Technical Specifications

Compute Infrastructure

Hardware

  • Apple MacBook Pro M4 β€” 16 GB unified memory

Software

  • MLX (Apple Silicon optimized)
  • M-Courtyard fine-tuning app

Model Card Authors

Armand β€” @ArmandS11

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support