AI & ML interests

None defined yet.

sweatSmile 
posted an update 6 days ago
view post
Post
368
Just published a hands-on guide on building a Kubernetes cluster from scratch on AWS EC2 using kubeadm, no managed services, no shortcuts.

If you want to truly understand how the control plane and workers communicate, how pod networking works with Flannel, and how to lock down access with security groups ,then this is the kind of exercise that makes it click.

The guide covers a full 3-node setup (1 control plane + 2 workers) on Amazon Linux 2023, from instance provisioning all the way to deploying your first workload.



Read it here 👉 https://www.amitchoubey.dev/posts/kubernetes-cluster-aws-ec2-kubeadm/
sweatSmile 
posted an update 7 months ago
sweatSmile 
posted an update 7 months ago
view post
Post
2801
Teaching a 7B Model to Be Just the Right Amount of Snark

Ever wondered if a language model could get sarcasm? I fine-tuned Mistral-7B using LoRA and 4-bit quantisation—on just ~720 hand-picked sarcastic prompt–response pairs from Reddit, Twitter, and real-life conversations.

The challenge? Keeping it sarcastic but still helpful.

LoRA rank 16 to avoid overfitting

4-bit NF4 quantization to fit on limited GPU memory

10 carefully monitored epochs so it didn’t turn into a full-time comedian

Result: a model that understands “Oh great, another meeting” exactly as you mean it.

Read the full journey, tech details, and lessons learned on my blog:
Fine-Tuning Mistral-7B for Sarcasm with LoRA and 4-Bit Quantisation

Try the model here on Hugging Face: sweatSmile/Mistral-7B-Instruct-v0.1-Sarcasm.

sweatSmile 
posted an update 7 months ago
view post
Post
2840
Qwen3 is the latest version of the Qwen language models. It's smarter, faster, and now understands 119 languages instead of just 29.
It can do both deep reasoning and quick answers using a single model, depending on what you need.
The models range in size from small (0.6B) to huge (235B), with smart ways to save compute.
It's trained on 36 trillion tokens and fine-tuned in four steps to boost performance.
Qwen3 performs as well as or better than many top models, including some from big companies.
It’s fully open-source under licence. Amazing!!!


https://github.com/QwenLM/Qwen3/blob/main/Qwen3_Technical_Report.pdf