LeroyDyer commited on
Commit
71f6e01
·
verified ·
1 Parent(s): bd612e7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +229 -16
README.md CHANGED
@@ -1,23 +1,236 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
- base_model: LeroyDyer/_Spydaz_Web_AI_AGI_R1_Top_Student
 
 
 
 
 
3
  tags:
4
- - text-generation-inference
5
- - transformers
6
- - unsloth
7
- - mistral
8
- - trl
9
- - sft
10
- license: apache-2.0
11
- language:
12
- - en
13
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
- # Uploaded model
 
16
 
17
- - **Developed by:** LeroyDyer
18
- - **License:** apache-2.0
19
- - **Finetuned from model :** LeroyDyer/_Spydaz_Web_AI_AGI_R1_Top_Student
20
 
21
- This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
22
 
23
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
 
 
 
 
1
+ Hugging Face's logo
2
+ Hugging Face
3
+ Models
4
+ Datasets
5
+ Spaces
6
+ Posts
7
+ Docs
8
+ Enterprise
9
+ Pricing
10
+
11
+
12
+
13
+
14
+ LeroyDyer
15
+ /
16
+ _Spydaz_Web_AI_AGI_R1_Top_Student
17
+
18
+ like
19
+ 1
20
+ Text Generation
21
+ Transformers
22
+ Safetensors
23
+ mistral
24
+ mergekit
25
+ Merge
26
+ conversational
27
+ text-generation-inference
28
+ Inference Endpoints
29
+ Model card
30
+ Files and versions
31
+ Community
32
+ Settings
33
+ _Spydaz_Web_AI_AGI_R1_Top_Student/
34
+ README.md
35
+ Metadata UI
36
+ license
37
+
38
+ datasets
39
+
40
+ language
41
+
42
+ metrics
43
+
44
+ base_model
45
+
46
+
47
+
48
+
49
+
50
+ new_version
51
+
52
+ pipeline_tag
53
+
54
+
55
+ Auto-detected
56
+ library_name
57
+
58
+
59
+ tags
60
+
61
+
62
+
63
+ Eval Results
64
+
65
+ View doc
66
+
67
+ 1
68
+ 2
69
+ 3
70
+ 4
71
+ 5
72
+ 6
73
+ 7
74
+ 8
75
+ 9
76
+ 10
77
+ 11
78
+ 12
79
+ 13
80
+ 14
81
+ 15
82
+ 16
83
+ 17
84
+ 18
85
+ 19
86
+ 20
87
+ 21
88
+ 22
89
+ 23
90
+ 24
91
+ 25
92
+ 26
93
+ 27
94
+ 28
95
+ 29
96
+ 30
97
+ 31
98
+ 32
99
+ 33
100
+ 34
101
+ 35
102
+ 36
103
+ 37
104
+ 38
105
+ 39
106
+ 40
107
+ 41
108
+ 42
109
+ 43
110
+ 44
111
+ 45
112
+ 46
113
+ 47
114
+ 48
115
+ 49
116
+ 50
117
+ 51
118
+ 52
119
+ 53
120
+ 54
121
+ 55
122
+ 56
123
+ 57
124
+ 58
125
+ 59
126
+ 60
127
+ 61
128
+ 62
129
+ 63
130
+ 64
131
+ 65
132
+ 66
133
+ 67
134
+ 68
135
+ 69
136
+
137
+
138
+
139
+
140
+
141
+
142
+
143
+
144
+
145
+
146
+
147
+
148
+
149
+
150
  ---
151
+ base_model:
152
+ - LeroyDyer/_Spydaz_Web_AI_AGI_R1_OmG_Coder
153
+ - LeroyDyer/_Spydaz_Web_AI_AGI_R1_Math_AdvancedStudent
154
+ - LeroyDyer/LCARS_TOP_SCORE
155
+ - LeroyDyer/_Spydaz_Web_AI_AGI_R1_Math_001
156
+ library_name: transformers
157
  tags:
158
+ - mergekit
159
+ - merge
160
+
 
 
 
 
 
 
161
  ---
162
+ Rank 1588
163
+ # Winners create more winners, while losers do the opposite.
164
+ Success is a game of winners.
165
+
166
+ — # Leroy Dyer (1972-Present)
167
+ <img src="https://cdn-avatars.huggingface.co/v1/production/uploads/65d883893a52cd9bcd8ab7cf/tRsCJlHNZo1D02kBTmfy9.jpeg" width="300"/>
168
+
169
+ # The Human AI .
170
+
171
+ # PERSONAL NOTE :
172
+
173
+ Sad to hear the leaderboard benchmarks stopped !
174
+ ut this model focussed on the bbh Collection and the mmlu collection - as well as the henricks maths collection :
175
+ I would expect that the musr went down as the model had already begun to miss on those tests .. But was still very high for most model at (20+)
176
+
177
+
178
+ this would have been the model that showed that All Sections in the green Aligning with the motto that it is not the size of the model but the training the moel has had :
179
+
180
+ There is a justification amoung the sellers of large ai ! they believe the more complexity and more parameters the better the model will perform !
181
+ Ie: Throw money at it : when indeed there was a 1.5b model topping the maths board !
182
+ SO its unjustified that the parameter size equals the inteligence of the model !
183
+
184
+ There is a technique to creating larger sized models , as extending models actually damages the model . but stacking various experts on top DOES ! make a difference in the models performance : as some training for modles actually throws other skill off !
185
+ So for a general inteligence it would have to be a multi expert model : with some internal reasoning chain between each espert in the stack !
186
+ Perhaps even a langchain graph as an interneal structure of llms which communicate between each other finally coming to concensous and responding !@
187
+
188
+ IE : a deepseek type model !
189
+ then extracting all the layers of the traied model back into a single moel structure effectivly merging the stacked models into a single tensor stack ! and realisning the model to agentic training datas @
190
+
191
+ so an actual agentic network ! instead of a external grpah or chain an internal chain !
192
+ now we can get to the genral inteligence bits : as we will need to add modalitys to allow for the model to be a true general intelugence !
193
+ we are not quite there yet ! as technology cannot handle the processing required until the gpu and cpus catch up ! the calculations need to be perfromed ont he gpu and not the cpu as this process is being hijack by gpu manafacturers and devlopers !
194
+
195
+ then we will be able to breath again ! creating full functioning grpahically rich near reality and highly agentic systems !
196
+
197
+
198
+ # Deep thinking Model - Highly Trained om Multiple Datasets
199
+
200
+ The base model has been created as a new staarting point : It has been fully primed with various types of chains of thoughts and step by step solutions : enabling for reward training to take place . this model has been trained with various languges ( not intensivly ), enabling for cross languge understanding ;
201
+ Here we create a valid start point for agent based modelling , As we find that some training actually affects existing knowledge , hence agents become a thing ! or if you prefr, distillations ....
202
+ These agents can be medical , technical , roleplayers etc .
203
+
204
+ This model was trained on various datasets , such as the basic math ones . As well as some adaced reasoning tasks. here we experiment with various styles of data from finacial to medical to coding (althugh this seemms to have an issue with very long context ,, as the servers seems to crash out a lot whe pushing larger cotext and rewards - suggestion , only 1 sample perstep can solve it), very impressive with its diagnsis skill for medical.
205
+
206
+ # SpydazWeb AI (7b Mistral) (512k)
207
+
208
+ This model has been trained to perform with contexts of 512k , although in training it has been trained mainly with the 2048 for general usage : the long context aspect also allows fro advanced projects and sumarys as well as image and audio translationns and generations:
209
+
210
+ Highly trained as well as methodolgy oriented , this model has been trained on the reAct Prcess and other structured processes . hence structured outputs (json) are very highly trained as well as orchestration of other agents and tasks : the model has been trained for tools use as well as funtion use : as well as custom processes and tools : some tools do not need code either as thier implication means the model may even generate a tool or artifct to perfrom the task :
211
+
212
+
213
+
214
+ A New genrea of AI ! This is Trained to give highly detailed humanized responses : Performs tasks well, a Very good model for multipupose use : the model has been trained to become more human in its reposes as well as role playing and story telling : This latest model has been trained on Conversations with a desire to respond with expressive emotive content , As well as discussions on various topics: It has also been focused on conversations by human interactions. hence there maybe NFSW contet in the model : This has no way inhibited its other tasks which were also aligned using the new intensive and Expressive prompt :
215
+ ## Thinking Humanly:
216
+
217
+ AI aims to model human thought, a goal of cognitive science across fields like psychology and computer science.
218
+ ## Thinking Rationally:
219
+
220
+ AI also seeks to formalize “laws of thought” through logic, though human thinking is often inconsistent and uncertain.
221
+ ## Acting Humanly:
222
 
223
+ Turing's test evaluates AI by its ability to mimic human behavior convincingly, encompassing skills like reasoning and language.
224
+ ## Acting Rationally:
225
 
226
+ Russell and Norvig advocate for AI that acts rationally to achieve the best outcomes, integrating reasoning and adaptability to environments.
 
 
227
 
 
228
 
229
+ Commit directly to the
230
+ main
231
+ branch
232
+ Open as a pull request to the
233
+ main
234
+ branch
235
+ Commit changes
236
+ Upload images, audio, and videos by dragging in the text input, pasting, or clicking here.