Lexa-Delta / README.md
unfavalen's picture
Update README.md
65d8dda verified
---
license: apache-2.0
tags:
- reasoning
- multilingual
- transformer
- robi-labs
- delta
- lexa
- lexa-family
- lexa-delta
pipeline_tag: text-generation
---
# Model Card for Lexa-Delta
Lexa-Delta is a multilingual reasoning large language model, developed by **Robi Labs**, designed for structured reasoning and natural conversation across multiple languages. It has been trained independently by Robi Labs.
---
## Model Details
### Model Description
* **Developed by:** Robi Labs
* **Model type:** Causal language model (decoder-only transformer)
* **Language(s):** Multilingual (English, Spanish, French, German, Chinese, Hindi, and more)
* **License:** Custom License (see LICENSE file)
### Model Sources
* **Repository:** [https://huggingface.co/RobiLabs/Lexa-Delta](https://huggingface.co/RobiLabs/Lexa-Delta)
* **Website:** [https://labs.robiai.com](https://labs.robiai.com)
* **Lexa Chat:** [https://lexa.chat](https://lexa.chat) (coming soon)
* **Socials:**
* [Twitter/X](https://twitter.com/justlexait)
* [LinkedIn](https://www.linkedin.com/company/robilabsai)
* [Instagram](https://www.instagram.com/robilabs)
---
## Uses
### Direct Use
Lexa-Delta can be used directly for:
* Multilingual question answering
* Chain-of-thought reasoning
* Conversational AI assistants
* Educational support (explaining concepts across languages)
### Downstream Use
* Fine-tuning for domain-specific tasks (e.g., legal, medical, educational)
* Integration into applications and chat platforms
### Out-of-Scope Use
* Disallowed or harmful content generation
* High-stakes decision making without expert human oversight
---
## Bias, Risks, and Limitations
* May inherit biases from multilingual training data
* Reasoning ability may vary by language
* Can generate incorrect or hallucinated outputs
### Recommendations
Users should:
* Verify important information independently
* Avoid high-stakes reliance without human review
* Use responsibly with awareness of multilingual limitations
---
## How to Get Started with the Model
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("RobiLabs/Lexa-Delta", device_map="auto")
tokenizer = AutoTokenizer.from_pretrained("RobiLabs/Lexa-Delta")
messages = [
{"role": "system", "content": "You are Lexa-Delta, a multilingual reasoning model from Robi Labs."},
{"role": "user", "content": "What is the capital of Armenia?"}
]
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to(model.device)
outputs = model.generate(inputs, max_new_tokens=200)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
---
## Training Details
### Training Data
* Multilingual reasoning data (sources undisclosed)
### Training Procedure
* **Method:** Full training
* **Precision:** Mixed precision
* **Compute:** B200 GPUs
#### Training Hyperparameters
* **Learning rate:** 2e-4
* **Sequence length:** 4096 tokens
* **Gradient accumulation:** enabled
---
## Environmental Impact
* **Hardware Type:** B200 GPUs
* **Region:** Not disclosed
* **Carbon Emitted:** Not disclosed
---
## Technical Specifications
### Model Architecture and Objective
* Decoder-only transformer
* Objective: Causal language modeling with reasoning-oriented training
### Compute Infrastructure
* **Hardware:** B200 GPUs
* **Software:** PyTorch, Hugging Face Transformers
---
## Citation
```bibtex
@misc{lexa-delta,
title={Lexa-Delta: A Multilingual Reasoning LLM},
author={Robi Labs},
year={2025},
howpublished={\url{https://huggingface.co/RobiLabs/Lexa-Delta}},
}
```
---
## Model Card Authors
[Robi Labs](https://labs.robiai.com)
## Model Card Contact
* Website: [labs.robiai.com](https://labs.robiai.com)
* Email: [labs@robiai.com](mailto:labs@robiai.com)