SoteriaInitiative commited on
Commit
8725ca8
·
verified ·
1 Parent(s): 98b0e04

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -11
README.md CHANGED
@@ -13,15 +13,15 @@ tags:
13
 
14
  A model fined tuned for sanctions and AML related OFAC FAQ questions with the Swiss AI
15
  Apertus 8B Instruct model which was then used as teacher and distilled to TinyLlama 1.1B. The model is 6-7 X smaller than the original. Quantization to INT8 should allow even low-memory CPU inference
16
- deployments if model latency is not a primary concern. FEFT LoRA adapter are included for use with base model.
17
 
18
 
19
  ## Model Details
20
 
21
  ### Model Description
22
 
23
- The model includes INT8 quantized weights for CPU inference and LoRA adopters for GPU inference with
24
- amatching base.
25
 
26
 
27
  - **Developed by:** Soteria Initiative
@@ -37,12 +37,12 @@ amatching base.
37
  <!-- Provide the basic links for the model. -->
38
 
39
  - **Repository:** https://huggingface.co/SoteriaInitiative/Apertus-8B-Instruct-OFAC-FAQ
40
- - **Demo:** https://huggingface.co/SoteriaInitiative/Apertus-8B-Instruct-OFAC-FAQnotebook.ipynb
41
 
42
  ## Uses
43
 
44
  Use for chat or assistant applications where compliance or financial crime analysis need to
45
- Get answers regarding FATF or OFAC FAQ matters.
46
 
47
  ### Direct Use
48
 
@@ -57,7 +57,7 @@ This model is not intended for production deployment.
57
  ## Bias, Risks, and Limitations
58
 
59
  The model is fine tuned for FATF and OFAC FAQ matters and hence should be restricted to such
60
- Use cases where this is of a concern.
61
 
62
 
63
  ### Recommendations
@@ -130,13 +130,9 @@ The following sources where used for fine tuning:
130
  ### Training Procedure
131
 
132
  Supervised fine tuning has been applied to the Apertus 8B Instruct model with a training dataset
133
- Containing of FAQ question/answer pairs as well as FATF titles and recommendation pairs.
134
 
135
 
136
- #### Training Hyperparameters
137
-
138
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
139
-
140
 
141
  ## Evaluation
142
 
 
13
 
14
  A model fined tuned for sanctions and AML related OFAC FAQ questions with the Swiss AI
15
  Apertus 8B Instruct model which was then used as teacher and distilled to TinyLlama 1.1B. The model is 6-7 X smaller than the original. Quantization to INT8 should allow even low-memory CPU inference
16
+ deployments if model latency is not a primary concern. PEFT LoRA adapter are included for use with base model.
17
 
18
 
19
  ## Model Details
20
 
21
  ### Model Description
22
 
23
+ The model includes INT8 quantized weights for CPU inference and a LoRA adapter for GPU inference with
24
+ a matching base.
25
 
26
 
27
  - **Developed by:** Soteria Initiative
 
37
  <!-- Provide the basic links for the model. -->
38
 
39
  - **Repository:** https://huggingface.co/SoteriaInitiative/Apertus-8B-Instruct-OFAC-FAQ
40
+ - **Demo:** _WIP_
41
 
42
  ## Uses
43
 
44
  Use for chat or assistant applications where compliance or financial crime analysis need to
45
+ get answers regarding FATF or OFAC FAQ matters.
46
 
47
  ### Direct Use
48
 
 
57
  ## Bias, Risks, and Limitations
58
 
59
  The model is fine tuned for FATF and OFAC FAQ matters and hence should be restricted to such
60
+ use cases where this is of a concern.
61
 
62
 
63
  ### Recommendations
 
130
  ### Training Procedure
131
 
132
  Supervised fine tuning has been applied to the Apertus 8B Instruct model with a training dataset
133
+ of FAQ question/answer pairs as well as FATF titles and recommendation pairs.
134
 
135
 
 
 
 
 
136
 
137
  ## Evaluation
138