If @clem comments on this post within the week with a task for me to classify with text and a paramater size for the model, within 48 hours I will create a new dataset, train the model, and post it all. Paramater size must be bellow 50 million params, and task can be text only, and genuinely possible.
if I complete it, Clem must tell everyone on Twitter/x to follow me on huggingface and link it
why did 36 people unfollow me đ we are back in the hundreds. if you become my 500th follower and have proof I'll give you 5 dollars worth of openrouter credits as an API key
⢠**Qwen3 update** â Alibabaâs Qwen team released an update to its Qwen âŻ3 model. The latest Qwen3â235BâA22BâThinkingâ2507 has 235 billion parameters with 22 billion active (MoE), supports a 256âŻk context and introduces a reasoning modeă887277669479733â L94-L103ă, making it âagentâready.â
⢠**Qwen3âCoder** â The 480 billionâparameter Qwen3âCoder activates 35 billion parameters and can handle 256kâ1M token contexts. It tops SWEâBench and CodeForces benchmarks and can generate, refactor and debug code across languagesă813759831752062â L31-L69ăă813759831752062â L96-L113ă.
⢠**HRM** â MITâs Hierarchical Reasoning Model uses separate highâlevel and lowâlevel modules; with only ~27 million parameters and 1 k training examples it outperforms chainâofâthought LLMs on reasoning tasksă961622371027367â L74-L93ă.
⢠**ASIâArch** â An arXiv paper presents ASIâArch, an autonomous AI research system that ran 1 1773 experiments (~20k GPU hours) and discovered 106 new linearâattention architecturesă159627632585686â L49-L73ă.
Other headlines: OpenAI & SoftBank are building a compact data centeră594568962713797â L92-L99ă; Lloyds Bank launched the Athena AI assistantă594568962713797â L218-L226ă; Yahoo Japan plans daily use of generative AIă594568962713797â L73-L79ă; and AIâagent tokens like FET, Virtuals and OriginTrail are powering Web3 automationă178973618089356â L24-L63ă.
### ChatGPT Agent Corner OpenAIâs ChatGPT Agent unifies browsing, coding, document analysis, slide creation and scheduling into one interfaceă619815554729605â L14-L24ă. It scores 41.6Â % on Humanityâs Last Exam and 45.5Â % on SpreadsheetBenchă619815554729605â L42-L52ă, and is available to Pro, Plus and Team usersă619815554729605â L65-L69ă.
We zoomed pass 70, 80, and 90 followers in a day. That is insane.
As a thank you, ive published one of my largest and most expensive dataset (rip my 2 dollars, most of my datasets are free to make) to the hub. To be used for text summarization in english, check it out! ProCreations/simple-summaries
New dataset I made! ProCreations/quality-fiction This is a synthetically made dataset with about 200 examples. It contains high quality fiction/fantasy story to use for AI training.
What is AI? Some weird wizards at big companies summoning brains? Humans behind the scenes just responding to requests? Nope. Itâs literally JUST math. Some people know this already, but not what actually goes on behind the scenes.