Nymbo's picture
Building on HF

Nymbo PRO

Nymbo

AI & ML interests

I like unrestricted, free, utilitarian stuff. I tend to archive good spaces because I don't trust y'all to keep them good :) Most spaces with runtime errors just need a restart. Paused spaces should work but require a GPU, duplicate to use.

Recent Activity

updated a collection about 14 hours ago
Hub Models
liked a model about 14 hours ago
moonshotai/Kimi-K2.5
reacted to IlyasMoutawwakil's post with 🔥 about 14 hours ago
Transformers v5 just landed! 🚀 It significantly unifies and reduces modeling code across architectures, while opening the door to a whole new class of performance optimizations. My favorite new feature? 🤔 The new dynamic weight loader + converter. Here’s why 👇 Over the last few months, the core Transformers maintainers built an incredibly fast weight loader, capable of converting tensors on the fly while loading them in parallel threads. This means we’re no longer constrained by how parameters are laid out inside the safetensors weight files. In practice, this unlocks two big things: - Much more modular modeling code. You can now clearly see how architectures build on top of each other (DeepSeek v2 → v3, Qwen v2 → v3 → MoE, etc.). This makes shared bottlenecks obvious and lets us optimize the right building blocks once, for all model families. - Performance optimizations beyond what torch.compile can do alone. torch.compile operates on the computation graph, but it can’t change parameter layouts. With the new loader, we can restructure weights at load time: fusing MoE expert projections, merging attention QKV projections, and enabling more compute-dense kernels that simply weren’t possible before. Personally, I'm honored to have contributed in this direction, including the work on optimizing MoE implementations and making modeling code more torch-exportable, so these optimizations can be ported cleanly across runtimes. Overall, Transformers v5 is a strong signal of where the community and industry are converging: Modularity and Performance, without sacrificing Flexibility. Transformers v5 makes its signature from_pretrained an entrypoint where you can mix and match: - Parallelism - Quantization - Custom kernels - Flash/Paged attention - Continuous batching - ... Kudos to everyone involved! I highly recommend the: Release notes: https://github.com/huggingface/transformers/releases/tag/v5.0.0 Blog post: https://huggingface.co/blog/transformers-v5
View all activity

Organizations

Stanford AI's profile picture DataScienceEngineering's profile picture AI그림 채널 저장소 (비공식)'s profile picture AI FILMS's profile picture GEM benchmark's profile picture Zhejiang University's profile picture CognitiveAIForHealth's profile picture OpenGVLab's profile picture MusicAI's profile picture CVPR Demo Track's profile picture BigScience Biomedical Datasets's profile picture LLMs's profile picture ONNXConfig for all's profile picture Gradio-Themes-Party's profile picture video-p2p-library's profile picture Gradio-Blocks-Party's profile picture scikit-learn's profile picture lora concepts library's profile picture Open-Source AI Meetup's profile picture IRANGPT's profile picture Arabic Machine Learning 's profile picture DataScienceGuild's profile picture Pseudo Lab's profile picture East China Normal University's profile picture Kornia AI's profile picture Platzi Community's profile picture Tune a video concepts library's profile picture SIGGRAPH 2022's profile picture Keras Dreambooth Event's profile picture Stable Diffusion Dreambooth Concepts Library's profile picture The Waifu Research Department's profile picture Musika's profile picture M.O.F.U.'s profile picture AI Indonesia Community's profile picture HackMIT's profile picture Blog-explorers's profile picture OpenSky's profile picture CyberHarem's profile picture Tensor Diffusion's profile picture ICCV2023's profile picture ICML2023's profile picture OpenOrca's profile picture Media Party 2023's profile picture OpenLLM France's profile picture huggingPartyParis's profile picture Multi🤖Transformers's profile picture Team Tonic's profile picture The Collectionists's profile picture Niansuh AI's profile picture That Time I got Reincarnated as a Hugging Face Organization's profile picture ZeroGPU Explorers's profile picture LibreWeb's profile picture Project Fluently's profile picture LocalLLaMA's profile picture Journalists on Hugging Face's profile picture Major TOM's profile picture MLX Community's profile picture INNOVA AI's profile picture Pine's profile picture Social Post Explorers's profile picture Cohere Labs Community's profile picture M4-ai's profile picture Dev Mode Explorers's profile picture Chinese LLMs on Hugging Face's profile picture Stable Diffusion Community (Unofficial, Non-profit)'s profile picture Hugging Face for Legal's profile picture ONNX Community's profile picture Hugging Face Discord Community's profile picture Nerdy Face's profile picture open/ acc's profile picture AI Starter Pack's profile picture Smol Community's profile picture LiteRT Community (FKA TFLite)'s profile picture Hugging Face MCP Course's profile picture Agents-MCP-Hackathon's profile picture MCP-1st-Birthday's profile picture Hugging Face Skills's profile picture 25daysofagents's profile picture