Qwen2-0.5B Fine-tuned with Ray Tune Hyperparameter Optimization
This model is a fine-tuned version of Qwen/Qwen2-0.5B-Instruct, optimized using Ray Tune with ASHA scheduler for automated hyperparameter search across 6 trials to find the optimal configuration.
This model was fine-tuned using the IMDB dataset on 2 × T4 16GB GPUs with automated hyperparameter optimization across learning rate, batch size, weight decay, epochs, and warmup steps.
For detailed implementation, Ray Tune configurations, and hyperparameter search setup, please check out the project repository.
- Downloads last month
- 1
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support