RichardErkhov commited on
Commit
1576bff
·
verified ·
1 Parent(s): 59ddce8

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +63 -0
README.md ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ LLM-ADE_tiny-v0.001 - AWQ
11
+ - Model creator: https://huggingface.co/InvestmentResearchAI/
12
+ - Original model: https://huggingface.co/InvestmentResearchAI/LLM-ADE_tiny-v0.001/
13
+
14
+
15
+
16
+
17
+ Original model description:
18
+ ---
19
+ language:
20
+ - en
21
+ license: mit
22
+ tags:
23
+ - finance
24
+ pipeline_tag: text-generation
25
+ widget:
26
+ - example_title: Easy
27
+ text: '<|im_start|>user
28
+
29
+ How do call options benefit the buyer?<|im_end|>
30
+
31
+ <|im_start|>assistant
32
+
33
+ '
34
+ - example_title: Medium
35
+ text: '<|im_start|>user
36
+
37
+ Why might a trader choose to quickly exit a losing position, even if they still
38
+ believe in the original trade idea?<|im_end|>
39
+
40
+ <|im_start|>assistant
41
+
42
+ '
43
+ - example_title: Hard
44
+ text: '<|im_start|>user
45
+
46
+ In the context of Harry Markowitz''s Portfolio Selection theory, what does an
47
+ ''efficient'' portfolio refer to?<|im_end|>
48
+
49
+ <|im_start|>assistant
50
+
51
+ '
52
+ inference:
53
+ parameters:
54
+ temperature: 0.2
55
+ min_new_tokens: 20
56
+ max_new_tokens: 250
57
+ ---
58
+ # AlphaBlind Tiny v0.001
59
+
60
+ Our Proof-of-Concept (POC) for the LLM-ADE framework (https://arxiv.org/abs/2404.13028). A very early, initial version of TinyLlama processing and ingesting llm-ade-fin_data-subset-earnings-10k and other financial data with the LLM-ADE framework.
61
+
62
+ Note: This model has not been thoroughly tested, and is very small - it can run on a Macbook Pro. Please do not use this version of the model as is.
63
+