Artha-1: A Compact Liquid-Autoencoder Language Model
Repository: vyomie/artha-1
Model Type: Hybrid LLM with Liquid Neural Network Core
Architecture: Custom Autoencoder + Liquid Neural Network
Model Size: ~400M parameters
Format: PyTorch .pth + Python plug-and-play pipeline
Usage: Plug-and-play via from model import Pipeline
Affordable Custom LLM
Artha-1 is a Liquid Neural Network-powered language model pipeline** trained and deployed end-to-end by an independent teen researcher. Unlike traditional LLMs that require millions of dollars, proprietary infrastructure, and industrial compute clusters, Artha-1 was trained in under 3 days using accessible datasets and mid-range hardware.
What Makes It the most efficient?
LNN-Based Reasoning Core:
This is the first open-source LLM to integrate a Liquid Neural Network (LNN) core for deep, dynamic reasoning inside compressed latent space.Built With ~Zero Budget:
Trained using just local GPUs and open datasets, no enterprise backing or funding was involved. This makes Artha-1 arguably the cheapest working LLM architecture available to the public.Created by a Teen Researcher:
From architecture design and training to deployment and packaging, every step was executed by an independent teen developer, proving that you donβt need a PhD or billion-dollar lab to innovate in AI.
Summary
Artha-1 is a compact and efficient language model designed with an unconventional architecture combining a pretrained autoencoder with a Liquid Neural Network (LNN) core. The model emphasizes interpretability, small footprint, and ease of use for experimentation and lightweight reasoning tasks. Built with simplicity and modularity in mind, Artha-1 is ideal for research, tinkering, or educational use, and runs efficiently on consumer-grade hardware.
Architecture
- Encoder: Bottleneck-T5 autoencoder (
thesephist/contra-bottleneck-t5-base-wikipedia) - Core Processor: Liquid Neural Network (LNN) with dynamic temporal memory
- Decoder: Same T5 decoder via latent perturbation and reconstruction
- Interface: Python
Pipelineclass (plug-and-play)
Intended Use
- Lightweight reasoning tasks
- Prompt-based experimentation
- Research on alternative LLM architectures
- Educational demos for architecture breakdowns
- Fine-tuning or distillation experiments for compact models
π« Not Intended Use
- Do not deploy in high-stakes environments (medical, legal, safety-critical tasks)
- Not optimized for factual correctness or robustness
- Not meant to replace larger foundational models (GPT, LLaMA, Claude, etc.)
Model Architecture
This model is a two-part system:
- A Bottleneck T5 autoencoder for text-to-latent and latent-to-text conversion, adapted from
thesephist/contra-bottleneck-t5-base-wikipedia - A custom Liquid Neural Network (LNN) core trained to perform latent-level reasoning on compressed embeddings.
The LNN consists of multiple gated recurrent layers designed for temporal and structural memory propagation, allowing for highly expressive representations at low parameter count.
Training Details
- Training Data: Synthetic question-answer dataset generated using open-source LLMs
- Latent Size: 768
- LNN Hidden Units: 4000
- Training Duration: ~2β3 days on mid-range GPUs
- Optimizer: AdamW with SWA (Stochastic Weight Averaging)
- Loss Function: Cosine similarity between predicted and true bottleneck embeddings
How to Use
Install dependancies
pip install arthaLM
Initializing Pipeline
from arthaLM import Pipeline
pipe = Pipeline(model_name="vyomie/artha-1")
Prompting
print(pipe("Hello, how are you!"))
(OR)
Make sure you have the following dependencies installed:
pip install torch transformers==4.36.1 huggingface_hub
Import required packages
import os
import sys
import importlib.util
from huggingface_hub import snapshot_download
Now, download custom Pipeline dynamically
snapshot_download("vyomie/artha-1", local_dir="/tmp/vyomie_artha-1", local_dir_use_symlinks=False)
spec = importlib.util.spec_from_file_location("model", "/tmp/vyomie_artha-1/model.py")
model = importlib.util.module_from_spec(spec)
sys.modules["model"] = model
spec.loader.exec_module(model)
Initialize the Pipeline
Pipe = Pipeline("vyomie/artha-1")
Output = Pipe("Hi, how are you!")
print(Output)
- Downloads last month
- 2
