--- license: mit ---
Orion-BiX Logo
Homepage Hugging Face Discord GitHub TabTune GitHub
# Orion-BiX: Bi-Axial Meta-Learning Model for Tabular In-Context Learning **Orion-BiX** is an advanced tabular foundation model that combines **Bi-Axial Attention** with **Meta-Learning** capabilities for few-shot tabular classification. The model extends the TabICL architecture with alternating attention patterns and episode-based training. The model is part of **Orion**, a family of tabular foundation models with various enhancements. ### Key Innovations 1. **Bi-Axial Attention**: Alternating attention patterns (Standard → Grouped → Hierarchical → Relational) that capture multi-scale feature interactions within tabular data 2. **Meta-Learning with k-NN Support Selection**: Episode-based training with intelligent support set selection using similarity metrics 3. **Three-Component Architecture**: Column embedding (Set Transformer), Bi-Axial row interaction, and In-Context Learning prediction ### Architecture Overview ``` Input → tf_col (Set Transformer) → Bi-Axial Attention → tf_icl (ICL) → Output ``` **Component Details:** - **tf_col (Column Embedder)**: Set Transformer for statistical distribution learning across features - **Bi-Axial Attention**: Replaces standard RowInteraction with alternating attention patterns: - Standard Cross-Feature Attention - Grouped Feature Attention - Hierarchical Feature Attention - Relational Feature Attention - CLS Token Aggregation - **tf_icl (ICL Predictor)**: In-context learning module for few-shot prediction ## Usage ### Direct (OrionBiX Python package) ```python from orion_bix.sklearn import OrionBiXClassifier # Initialize and use clf = OrionBiXClassifier() clf.fit(X_train, y_train) predictions = clf.predict(X_test) ``` This code will automatically download the pre-trained model from Hugging Face and use a GPU if available. ### Via TabTune (unified TFM library) OrionBix can be used either directly via its own Python package or through [TabTune](https://github.com/Lexsi-Labs/TabTune), which provides a unified interface over several tabular foundation models. ```python from tabtune import TabularPipeline pipeline = TabularPipeline( model_name="OrionBix", # use OrionBix through TabTune tuning_strategy="inference", # zero-shot / in-context mode tuning_params={"device": "cuda"} # or "cpu" ) pipeline.fit(X_train, y_train) predictions = pipeline.predict(X_test) ``` When used through TabTune, the OrionBiX weights are automatically downloaded from this Hugging Face repository on first use, and TabTune handles model-aware preprocessing for you. ## Installation ### Via TabTune (recommended if you want multiple tabular FMs) ```bash pip install tabtune ``` This installs TabTune and its built-in OrionBiX integration; no separate orion-bix install is required. ### From the OrionBiX source #### Option 1: From the local clone ```bash cd orion-bix pip install -e . ``` #### Option 2: From the Git Remote ```bash pip install git+https://github.com/Lexsi-Labs/Orion-BiX.git ``` ## Citation If you use Orion-BiX in your research, please cite our [paper](https://arxiv.org/abs/2512.00181):: ```bibtex @misc{bouadi2025orionbix, title={Orion-Bix: Bi-Axial Attention for Tabular In-Context Learning}, author={Mohamed Bouadi and Pratinav Seth and Aditya Tanna and Vinay Kumar Sankarapu}, year={2025}, eprint={2512.00181}, archivePrefix={arXiv}, primaryClass={cs.LG}, url={https://arxiv.org/abs/2512.00181}, } ```