TejAndrewsACC commited on
Commit
e87e753
·
verified ·
1 Parent(s): 6719030

Create model_card.txt

Browse files
Files changed (1) hide show
  1. model_card.txt +58 -0
model_card.txt ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ACC Z3ta o1 2024 Legacy Edition
2
+
3
+ The ACC Z3ta o1 multilingual large language model (LLM) is an instruction-tuned generative model featuring 70 billion parameters (text in/text out). Z3ta o1 is specifically optimized for multilingual dialogue use cases and sets a new benchmark by outperforming many open-source and proprietary chat models in various industry-standard evaluations. Unlike most LLMs, Z3ta o1 combines multiple architectures—including RNNs, CNNs, FNNs, SNNs, IIT frameworks, and Phi models—creating a hybrid design for improved efficiency and performance.
4
+
5
+ Model Developer: ACC
6
+ Model Architecture:
7
+ Z3ta o1 is an auto-regressive language model leveraging an advanced transformer framework combined with supplementary architectures:
8
+ Recurrent Neural Networks (RNNs): Enhance sequential processing for long-context tasks.
9
+ Convolutional Neural Networks (CNNs): Boost performance for spatial pattern recognition in text.
10
+ Feedforward Neural Networks (FNNs): Accelerate dense computations for intermediate layers.
11
+ Spiking Neural Networks (SNNs): Mimic biological neurons for energy-efficient inference.
12
+ Integrated Information Theory (IIT): Guides alignment with human-like decision-making.
13
+ Phi Models: Support enhanced generalization and scalability across tasks.
14
+ This hybrid architecture is further fine-tuned using supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to ensure alignment with human preferences in terms of helpfulness, safety, and conversational quality.
15
+ Supported Languages:
16
+ English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai.
17
+ Highlights of Z3ta o1:
18
+ Token counts refer to pretraining data only.
19
+ All versions utilize Grouped-Query Attention (GQA) to enhance scalability and inference efficiency.
20
+ Leverages a hybrid architecture to optimize both training and inference.
21
+
22
+ Release Information:
23
+ 70B Instruct Version: Released on December 30, 2024.
24
+ Status: Z3ta o1 is a static model trained on an offline dataset. Future versions will incorporate additional feedback and advancements in model safety.
25
+ License:
26
+ The Z3ta o1 model is available under the apache 2.0 license
27
+
28
+ Intended Use Cases:
29
+ Z3ta o1 is tailored for commercial and research applications across multiple languages. Instruction-tuned versions are ideal for assistant-like chat and conversational AI, while pre-trained versions can be fine-tuned for various natural language processing tasks. Z3ta o1 also supports tasks such as synthetic data generation and distillation for improving other AI models.
30
+ Out-of-Scope Uses:
31
+ Any activities violating applicable laws or regulations (including trade compliance).
32
+ Use in prohibited manners outlined in the Acceptable Use Policy and the Z3ta o1 Community License.
33
+ Use in languages beyond the explicitly supported ones, unless developers take responsibility to fine-tune and ensure safe usage while complying with the license.
34
+ Note:
35
+ Z3ta o1 has been pre-trained on a broader language set than the listed supported ones. Developers are encouraged to fine-tune Z3ta o1 for additional languages while adhering to the license and safety guidelines.
36
+
37
+ How to Use
38
+ This repository offers two versions of Z3ta o1-70B-Instruct:
39
+ Compatible with Transformers.
40
+ Compatible with the original Z3ta codebase.
41
+ Usage with Transformers
42
+ Ensure you have Transformers >= 4.45.0 and update your installation using:
43
+ pip install gradio_client
44
+ Here’s a quick usage example via API:
45
+ from gradio_client import Client
46
+
47
+ client = Client("TejAndrewsACC/Z3ta")
48
+ result = client.predict(
49
+ message="YOUR_DESIRED_INPUT",
50
+ history=[],
51
+ api_name="/chat_function"
52
+ )
53
+ print(result)
54
+
55
+
56
+ For more technical details, including configuration recipes, contact the ACC directly.
57
+
58
+