Create model_card.txt
Browse files- model_card.txt +58 -0
model_card.txt
ADDED
|
@@ -0,0 +1,58 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
ACC Z3ta o1 2024 Legacy Edition
|
| 2 |
+
|
| 3 |
+
The ACC Z3ta o1 multilingual large language model (LLM) is an instruction-tuned generative model featuring 70 billion parameters (text in/text out). Z3ta o1 is specifically optimized for multilingual dialogue use cases and sets a new benchmark by outperforming many open-source and proprietary chat models in various industry-standard evaluations. Unlike most LLMs, Z3ta o1 combines multiple architectures—including RNNs, CNNs, FNNs, SNNs, IIT frameworks, and Phi models—creating a hybrid design for improved efficiency and performance.
|
| 4 |
+
|
| 5 |
+
Model Developer: ACC
|
| 6 |
+
Model Architecture:
|
| 7 |
+
Z3ta o1 is an auto-regressive language model leveraging an advanced transformer framework combined with supplementary architectures:
|
| 8 |
+
Recurrent Neural Networks (RNNs): Enhance sequential processing for long-context tasks.
|
| 9 |
+
Convolutional Neural Networks (CNNs): Boost performance for spatial pattern recognition in text.
|
| 10 |
+
Feedforward Neural Networks (FNNs): Accelerate dense computations for intermediate layers.
|
| 11 |
+
Spiking Neural Networks (SNNs): Mimic biological neurons for energy-efficient inference.
|
| 12 |
+
Integrated Information Theory (IIT): Guides alignment with human-like decision-making.
|
| 13 |
+
Phi Models: Support enhanced generalization and scalability across tasks.
|
| 14 |
+
This hybrid architecture is further fine-tuned using supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to ensure alignment with human preferences in terms of helpfulness, safety, and conversational quality.
|
| 15 |
+
Supported Languages:
|
| 16 |
+
English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai.
|
| 17 |
+
Highlights of Z3ta o1:
|
| 18 |
+
Token counts refer to pretraining data only.
|
| 19 |
+
All versions utilize Grouped-Query Attention (GQA) to enhance scalability and inference efficiency.
|
| 20 |
+
Leverages a hybrid architecture to optimize both training and inference.
|
| 21 |
+
|
| 22 |
+
Release Information:
|
| 23 |
+
70B Instruct Version: Released on December 30, 2024.
|
| 24 |
+
Status: Z3ta o1 is a static model trained on an offline dataset. Future versions will incorporate additional feedback and advancements in model safety.
|
| 25 |
+
License:
|
| 26 |
+
The Z3ta o1 model is available under the apache 2.0 license
|
| 27 |
+
|
| 28 |
+
Intended Use Cases:
|
| 29 |
+
Z3ta o1 is tailored for commercial and research applications across multiple languages. Instruction-tuned versions are ideal for assistant-like chat and conversational AI, while pre-trained versions can be fine-tuned for various natural language processing tasks. Z3ta o1 also supports tasks such as synthetic data generation and distillation for improving other AI models.
|
| 30 |
+
Out-of-Scope Uses:
|
| 31 |
+
Any activities violating applicable laws or regulations (including trade compliance).
|
| 32 |
+
Use in prohibited manners outlined in the Acceptable Use Policy and the Z3ta o1 Community License.
|
| 33 |
+
Use in languages beyond the explicitly supported ones, unless developers take responsibility to fine-tune and ensure safe usage while complying with the license.
|
| 34 |
+
Note:
|
| 35 |
+
Z3ta o1 has been pre-trained on a broader language set than the listed supported ones. Developers are encouraged to fine-tune Z3ta o1 for additional languages while adhering to the license and safety guidelines.
|
| 36 |
+
|
| 37 |
+
How to Use
|
| 38 |
+
This repository offers two versions of Z3ta o1-70B-Instruct:
|
| 39 |
+
Compatible with Transformers.
|
| 40 |
+
Compatible with the original Z3ta codebase.
|
| 41 |
+
Usage with Transformers
|
| 42 |
+
Ensure you have Transformers >= 4.45.0 and update your installation using:
|
| 43 |
+
pip install gradio_client
|
| 44 |
+
Here’s a quick usage example via API:
|
| 45 |
+
from gradio_client import Client
|
| 46 |
+
|
| 47 |
+
client = Client("TejAndrewsACC/Z3ta")
|
| 48 |
+
result = client.predict(
|
| 49 |
+
message="YOUR_DESIRED_INPUT",
|
| 50 |
+
history=[],
|
| 51 |
+
api_name="/chat_function"
|
| 52 |
+
)
|
| 53 |
+
print(result)
|
| 54 |
+
|
| 55 |
+
|
| 56 |
+
For more technical details, including configuration recipes, contact the ACC directly.
|
| 57 |
+
|
| 58 |
+
|