AetherMind_SRL: How I beat 7B models on MMLU with 184M params and a $300 GPU I’m Sameer, a solo researcher from Iraq working on a single RTX 3050 8GB laptop.Today I’m releasing AetherMind_SRL – a 184M-parameter NLI model that was trained only on tasks (SNLI, MNLI, ANLI, and a small clinical Alzheimer’s dataset). It was never fine-tuned or even shown a single MMLU question during training.Yet here are the zero-shot MMLU (57 subjects) results:Model MMLU Zero-Shot Training Data AetherMind_SRL (me) 184M 36.05 % Only NLI (SNLI/MNLI/ANLI + ADNI) DeBERTa-v3-base 278M ~30.8 % General pre-training BERT-large 340M 27–30 % General pre-training LLaMA-1 7B 7B 34–35 % Massive text corpus LLaMA-2 7B 7B ~45 % Bigger + better data
Yes – my 184M model beats every classic 300–400M model and the original 7-billion-parameter LLaMA-1, all while running at 300+ samples/sec on a $300 laptop GPU.How did this happen?I built a standardized self-improvement loop called AetherMind Self-Reflective Learning (SRL) v1.0:Train normally on NLI Let the model predict on hard adversarial data (ANLI) Log every mistake + low-confidence case Build a balanced “SMART” buffer (60% errors + 40% correct anchors) Fine-tune with tiny LR and error-weighted loss Repeat until stable That’s it. No external knowledge, no MMLU data, no cluster. Just pure reasoning transfer from entailment/contradiction patterns → real-world knowledge.Try it yourself python from transformers import pipeline import torch
Need Help Getting arXiv Endorsement for My AI Research Paper
Hi everyone, I hope you're doing well. I’m trying to publish my new AI research paper on arXiv under the cs.AI category, but I currently need an endorser who is already authorized for cs.AI submissions.
If anyone here is registered as a cs.AI endorser and is willing to help, I would truly appreciate it.
Here is the official arXiv endorsement request link:
My research: It’s part of the AetherMind project — a self-reflective NLI reasoning system inspired by human cognitive consistency and used also in Alzheimer’s research. If needed, I can share the abstract or full PDF.
Need Help Getting arXiv Endorsement for My AI Research Paper
Hi everyone, I hope you're doing well. I’m trying to publish my new AI research paper on arXiv under the cs.AI category, but I currently need an endorser who is already authorized for cs.AI submissions.
If anyone here is registered as a cs.AI endorser and is willing to help, I would truly appreciate it.
Here is the official arXiv endorsement request link:
My research: It’s part of the AetherMind project — a self-reflective NLI reasoning system inspired by human cognitive consistency and used also in Alzheimer’s research. If needed, I can share the abstract or full PDF.
Thank you so much to anyone who can support.
— Sameer S.Najm
reacted to aufklarer's
post with 🔥about 1 month ago
New Research Release: AetherMind_SRL — Self-Reflective Learning for Robust Natural Language Inference (NLI) By: Sameer S. Najm
After more than a year of continuous work, experimentation, and refinement, I am proud to officially publish AetherMind_SRL, a self-improving Transformer model trained using Self-Reflective Learning (SRL) — a technique that enables the model to learn from its own mistakes.
This research integrates:
🔹 Knowledge Distillation from DeBERTa-v3-base 🔹 Self-Reflective Learning loops 🔹 Adversarial ANLI training 🔹 Clinical Alzheimer’s reasoning (ADNI) 🔹 SMART Error Buffers for hard-example mining
What makes it unique? AetherMind_SRL continuously improves through structured error logs, balanced correction buffers, and clinical-domain adaptation. It achieves strong performance across general NLI benchmarks, adversarial datasets, and Alzheimer’s-focused reasoning tasks.
AetherMind-KD-Student is a 184M-parameter Natural Language Inference (NLI) model distilled from a DeBERTa-v3 teacher using a multi-stage, adversarial-aware knowledge distillation pipeline. The model is designed to provide:
High accuracy on standard NLI benchmarks Strong robustness on adversarial datasets Excellent zero-shot generalization to unseen datasets High inference efficiency on consumer GPUs
This makes it suitable for research and practical applications that require fast and reliable sentence-level reasoning. samerzaher80/AetherMind-KD-Student
reacted to evalstate's
post with 🤗about 1 month ago