--- license: apache-2.0 tags: - llm - hyperbolic - geometry - adapter - peft - research base_model: - Qwen/Qwen2.5-7B pipeline_tag: text-generation library_name: peft --- # ManifoldGL – Information‑Geometric Adapter for LLMs ManifoldGL is a parameter‑efficient adapter that enforces **hyperbolic geometry** on the latent space of large language models. It treats the meaning of a token as a **fiber** over a hyperbolic base manifold (a Poincaré ball), rather than a single vector in flat Euclidean space. Latent states are projected onto the ball, and attentions are computed using geodesic distance. A sheaf‑theoretic consistency loss and natural gradient optimization maintain semantic structure during training. ## Motivation and theoretical background Modern LLMs embed tokens in a Euclidean vector space. While convenient, Euclidean geometry has limited capacity to represent hierarchical structures: flat space grows polynomially, whereas hierarchical trees expand exponentially. By contrast, **hyperbolic space** grows exponentially and preserves both local and global relationships in a hierarchy【247949143190903†L115-L124】. Hyperbolic embeddings outperform Euclidean ones for lexical entailment, similarity and analogy tasks【247949143190903†L154-L169】. ManifoldGL leverages these properties by modelling the latent space as a fiber bundle over a hyperbolic base: each point in the Poincaré ball encodes a context, and its fiber contains a distribution of semantic components. ## Results on ARC‑AGI benchmark ManifoldGL fine‑tuned on Qwen2.5‑7B improves task accuracy on the ARC‑AGI benchmark from **12.4 %** to **28.7 %**, a **131.5 % relative improvement**. The model also achieves a **Manifold Faithfulness Rate (MFR) of 94.2 %**, indicating high adherence to the hyperbolic constraints, and maintains a curvature close to the target κ = ‑1 (mean ‑0.98 ± 0.04). Ablation studies show that removing curvature regularization, natural gradients, sheaf consistency or the hyperbolic target significantly reduces accuracy; the Euclidean target ablation causes the largest drop (–10.9 %), highlighting the importance of hyperbolic geometry. ## Files in this repository This model card accompanies adapter weights trained with ManifoldGL. The files follow the structure of the original repository: - `adapter_config.json` – configuration for PEFT/LoRA loading - `pytorch_adapter.bin` – adapter weights - `README.md` – this model card ## Quick start ```python from transformers import AutoModelForCausalLM from peft import PeftModel # Load the base model (Qwen2.5-7B) base_model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen2.5-7B") # Load the ManifoldGL adapter model = PeftModel.from_pretrained(base_model, "jesusvilela/manifoldgl") # Now use model.generate(...) to generate text with hyperbolic adapters ``` ## Usage This adapter can be loaded with [PEFT](https://github.com/huggingface/peft) on top of any compatible Qwen2.5‑7B model. During generation, latent states are projected into hyperbolic space and meaning is represented as fibers. We recommend using FP32 precision for maximum stability. ## Citation If you use ManifoldGL in your work, please cite the accompanying thesis and repository.