Sarverott commited on
Commit
ba91bc8
·
1 Parent(s): cd20d11

preparing scaffoldings of concept to be aimed

Browse files
Taskfile.yml CHANGED
@@ -8,5 +8,4 @@ vars:
8
  tasks:
9
  default:
10
  cmds:
11
- - echo "{{.GREETING}}"
12
- silent: true
 
8
  tasks:
9
  default:
10
  cmds:
11
+ - uv run jupyter notebook
 
bin/knowledge.db ADDED
Binary file (8.19 kB). View file
 
pyproject.toml CHANGED
@@ -4,7 +4,36 @@ version = "0.1.0"
4
  description = "Add your description here"
5
  readme = "README.md"
6
  requires-python = ">=3.12"
7
- dependencies = []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
 
9
  [dependency-groups]
10
  dev = [
@@ -23,5 +52,7 @@ dev = [
23
  "notebook>=7.5.4",
24
  "ollama>=0.6.1",
25
  "pillow>=12.1.1",
 
 
26
  "transformers>=5.2.0",
27
  ]
 
4
  description = "Add your description here"
5
  readme = "README.md"
6
  requires-python = ">=3.12"
7
+ dependencies = [
8
+ "astropy>=7.2.0",
9
+ "beautifulsoup4>=4.14.3",
10
+ "chromadb>=1.5.2",
11
+ "crewai[litellm]>=0.95.0",
12
+ "datasets>=4.6.1",
13
+ "diffusers[torch]>=0.36.0",
14
+ "fastapi>=0.135.1",
15
+ "gguf>=0.18.0",
16
+ "gitpython>=3.1.46",
17
+ "libretranslatepy>=2.1.4",
18
+ "litellm[proxy]>=1.82.0",
19
+ "markdownify>=1.2.2",
20
+ "numpy>=2.4.2",
21
+ "ollama>=0.6.1",
22
+ "openai>=2.24.0",
23
+ "opencv-contrib-python-headless>=4.13.0.92",
24
+ "paramiko>=4.0.0",
25
+ "pillow>=12.1.1",
26
+ "pygithub>=2.8.1",
27
+ "requests>=2.32.5",
28
+ "safetensors>=0.7.0",
29
+ "scikit-learn>=1.8.0",
30
+ "setuptools>=82.0.0",
31
+ "tokenizers>=0.22.2",
32
+ "torch>=2.10.0",
33
+ "transformers>=5.2.0",
34
+ "wheel>=0.46.3",
35
+ "wikipedia>=1.4.0",
36
+ ]
37
 
38
  [dependency-groups]
39
  dev = [
 
52
  "notebook>=7.5.4",
53
  "ollama>=0.6.1",
54
  "pillow>=12.1.1",
55
+ "scapy>=2.7.0",
56
+ "sherlock-project>=0.16.0",
57
  "transformers>=5.2.0",
58
  ]
src/main.py ADDED
File without changes
src/mindset/expression.py ADDED
File without changes
src/mindset/interpretation.py ADDED
File without changes
src/mindset/logics.py ADDED
File without changes
src/mindset/main.py ADDED
File without changes
src/mindset/memories.py ADDED
File without changes
src/mindset/thinking.py ADDED
File without changes
src/tests.ipynb ADDED
@@ -0,0 +1,390 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cells": [
3
+ {
4
+ "cell_type": "code",
5
+ "execution_count": 2,
6
+ "id": "9a4eb4ec",
7
+ "metadata": {},
8
+ "outputs": [
9
+ {
10
+ "data": {
11
+ "text/plain": [
12
+ "<sqlite3.Cursor at 0x7f13e777f5c0>"
13
+ ]
14
+ },
15
+ "execution_count": 2,
16
+ "metadata": {},
17
+ "output_type": "execute_result"
18
+ }
19
+ ],
20
+ "source": [
21
+ "import sqlite3\n",
22
+ "con = sqlite3.connect(\"../bin/knowledge.db\")\n",
23
+ "\n",
24
+ "cur = con.cursor()\n",
25
+ "\n",
26
+ "cur.execute(\"CREATE TABLE glossary(id, title, description, hashlabel)\")"
27
+ ]
28
+ },
29
+ {
30
+ "cell_type": "code",
31
+ "execution_count": 1,
32
+ "id": "75eef0d4",
33
+ "metadata": {},
34
+ "outputs": [],
35
+ "source": [
36
+ "import wikipedia"
37
+ ]
38
+ },
39
+ {
40
+ "cell_type": "code",
41
+ "execution_count": 3,
42
+ "id": "718bcc93-4e4a-43ad-993e-73788e350947",
43
+ "metadata": {},
44
+ "outputs": [],
45
+ "source": [
46
+ "import chromadb"
47
+ ]
48
+ },
49
+ {
50
+ "cell_type": "code",
51
+ "execution_count": 5,
52
+ "id": "a54e8544-ec7d-4600-9d4e-87f2af110cfe",
53
+ "metadata": {},
54
+ "outputs": [],
55
+ "source": [
56
+ "import git"
57
+ ]
58
+ },
59
+ {
60
+ "cell_type": "code",
61
+ "execution_count": 1,
62
+ "id": "467d2b08-35e6-4d9c-9412-dec84b8ad155",
63
+ "metadata": {},
64
+ "outputs": [
65
+ {
66
+ "data": {
67
+ "text/plain": [
68
+ "<ollama._client.Client at 0x7f431f2aa8a0>"
69
+ ]
70
+ },
71
+ "execution_count": 1,
72
+ "metadata": {},
73
+ "output_type": "execute_result"
74
+ }
75
+ ],
76
+ "source": [
77
+ "import ollama\n",
78
+ "\n",
79
+ "client = ollama.Client(\n",
80
+ " host='http://azrael.snet:11434'\n",
81
+ ")\n",
82
+ "\n",
83
+ "client"
84
+ ]
85
+ },
86
+ {
87
+ "cell_type": "code",
88
+ "execution_count": 5,
89
+ "id": "7eebda8a-ca83-4dec-82e5-a2b3ab00148b",
90
+ "metadata": {
91
+ "scrolled": true
92
+ },
93
+ "outputs": [
94
+ {
95
+ "name": "stdout",
96
+ "output_type": "stream",
97
+ "text": [
98
+ "tinytest-gguf:latest\n",
99
+ "gpt-oss:120b\n",
100
+ "mathstral:latest\n",
101
+ "wizardlm:7b-q2_K\n",
102
+ "phi4-mini-reasoning:latest\n",
103
+ "smallthinker:latest\n",
104
+ "aya-expanse:latest\n",
105
+ "yarn-llama2:latest\n",
106
+ "meditron:latest\n",
107
+ "xwinlm:latest\n",
108
+ "reflection:latest\n",
109
+ "exaone3.5:7.8b\n",
110
+ "granite3.1-dense:8b\n",
111
+ "llava-phi3:latest\n",
112
+ "falcon3:1b\n",
113
+ "codegemma:2b\n",
114
+ "starcoder2:3b\n",
115
+ "mistral-small:22b\n",
116
+ "llama3.1:8b\n",
117
+ "qwen3:1.7b\n",
118
+ "qwen3-coder:latest\n",
119
+ "deepseek-r1:8b\n",
120
+ "deepseek-r1:14b\n",
121
+ "gpt-oss:latest\n",
122
+ "llama4:latest\n",
123
+ "embeddinggemma:latest\n",
124
+ "paraphrase-multilingual:latest\n",
125
+ "all-minilm:latest\n",
126
+ "bge-m3:latest\n",
127
+ "notus:latest\n",
128
+ "goliath:latest\n",
129
+ "minicpm-v:latest\n",
130
+ "sarverott/EON-alfa:latest\n",
131
+ "nomic-embed-text:latest\n",
132
+ "sarverott/EON-beta:latest\n",
133
+ "EON-beta___Chronus:latest\n",
134
+ "gemma3:latest\n",
135
+ "EON-alfa---Pozeralka-snap2:latest\n",
136
+ "EON-alfa---Pozeralka:latest\n",
137
+ "llama3.2:latest\n",
138
+ "gemma3:1b\n",
139
+ "cognitivetech/obook_title:latest\n",
140
+ "msn/artl:latest\n",
141
+ "wizardcoder:latest\n",
142
+ "wizard-math:70b\n",
143
+ "wizard-math:13b\n",
144
+ "wizard-math:latest\n",
145
+ "smollm:latest\n",
146
+ "yi:latest\n",
147
+ "orca-mini:latest\n",
148
+ "olmo2:latest\n",
149
+ "snowflake-arctic-embed:latest\n",
150
+ "duckdb-nsql:latest\n",
151
+ "mistrallite:latest\n",
152
+ "command-r7b:latest\n",
153
+ "tulu3:latest\n",
154
+ "r1-1776:latest\n",
155
+ "dbrx:latest\n",
156
+ "granite3-guardian:latest\n",
157
+ "alfred:latest\n",
158
+ "sailor2:latest\n",
159
+ "command-a:latest\n",
160
+ "starcoder2:latest\n",
161
+ "deepseek-v3:latest\n",
162
+ "mistral-nemo:latest\n",
163
+ "tinyllama:latest\n",
164
+ "mxbai-embed-large:latest\n",
165
+ "phi:latest\n",
166
+ "codellama:latest\n",
167
+ "gemma2:latest\n",
168
+ "llava:latest\n",
169
+ "qwen2.5:latest\n",
170
+ "mistral:latest\n",
171
+ "qwq:latest\n",
172
+ "qwen2:latest\n",
173
+ "phi4:14b-fp16\n",
174
+ "OpSor-license-master:latest\n",
175
+ "the-omnilibris-license-specialist:latest\n",
176
+ "the-hermetic-officer___BOS:latest\n",
177
+ "nous-hermes2:latest\n",
178
+ "the-library-master___omnilibris-shakespeare-KFT_docs:latest\n",
179
+ "deepseek-r1:7b\n",
180
+ "deepseek-apokryf-overviewed-and-topics-discused:latest\n",
181
+ "deepseek-for-anubis:latest\n",
182
+ "deepseek-r1:latest\n"
183
+ ]
184
+ },
185
+ {
186
+ "data": {
187
+ "text/plain": [
188
+ "[None,\n",
189
+ " None,\n",
190
+ " None,\n",
191
+ " None,\n",
192
+ " None,\n",
193
+ " None,\n",
194
+ " None,\n",
195
+ " None,\n",
196
+ " None,\n",
197
+ " None,\n",
198
+ " None,\n",
199
+ " None,\n",
200
+ " None,\n",
201
+ " None,\n",
202
+ " None,\n",
203
+ " None,\n",
204
+ " None,\n",
205
+ " None,\n",
206
+ " None,\n",
207
+ " None,\n",
208
+ " None,\n",
209
+ " None,\n",
210
+ " None,\n",
211
+ " None,\n",
212
+ " None,\n",
213
+ " None,\n",
214
+ " None,\n",
215
+ " None,\n",
216
+ " None,\n",
217
+ " None,\n",
218
+ " None,\n",
219
+ " None,\n",
220
+ " None,\n",
221
+ " None,\n",
222
+ " None,\n",
223
+ " None,\n",
224
+ " None,\n",
225
+ " None,\n",
226
+ " None,\n",
227
+ " None,\n",
228
+ " None,\n",
229
+ " None,\n",
230
+ " None,\n",
231
+ " None,\n",
232
+ " None,\n",
233
+ " None,\n",
234
+ " None,\n",
235
+ " None,\n",
236
+ " None,\n",
237
+ " None,\n",
238
+ " None,\n",
239
+ " None,\n",
240
+ " None,\n",
241
+ " None,\n",
242
+ " None,\n",
243
+ " None,\n",
244
+ " None,\n",
245
+ " None,\n",
246
+ " None,\n",
247
+ " None,\n",
248
+ " None,\n",
249
+ " None,\n",
250
+ " None,\n",
251
+ " None,\n",
252
+ " None,\n",
253
+ " None,\n",
254
+ " None,\n",
255
+ " None,\n",
256
+ " None,\n",
257
+ " None,\n",
258
+ " None,\n",
259
+ " None,\n",
260
+ " None,\n",
261
+ " None,\n",
262
+ " None,\n",
263
+ " None,\n",
264
+ " None,\n",
265
+ " None,\n",
266
+ " None,\n",
267
+ " None,\n",
268
+ " None,\n",
269
+ " None,\n",
270
+ " None,\n",
271
+ " None,\n",
272
+ " None]"
273
+ ]
274
+ },
275
+ "execution_count": 5,
276
+ "metadata": {},
277
+ "output_type": "execute_result"
278
+ }
279
+ ],
280
+ "source": [
281
+ "[print(llmEntity.model) for llmEntity in client.list().models]"
282
+ ]
283
+ },
284
+ {
285
+ "cell_type": "code",
286
+ "execution_count": 6,
287
+ "id": "8f6b4e83-53cb-4a81-91ca-01479c80ca62",
288
+ "metadata": {},
289
+ "outputs": [
290
+ {
291
+ "data": {
292
+ "text/plain": [
293
+ "ProgressResponse(status='success', completed=None, total=None, digest=None)"
294
+ ]
295
+ },
296
+ "execution_count": 6,
297
+ "metadata": {},
298
+ "output_type": "execute_result"
299
+ }
300
+ ],
301
+ "source": [
302
+ "client.pull(\"translategemma:4b\")\n",
303
+ "client.pull(\"qwen3.5:0.8b\")\n",
304
+ "client.pull(\"qwen3.5:latest\")\n",
305
+ "client.pull(\"lfm2.5-thinking:latest\")"
306
+ ]
307
+ },
308
+ {
309
+ "cell_type": "code",
310
+ "execution_count": 8,
311
+ "id": "f273ce0b-a7af-44d6-bb5c-14b7998b3947",
312
+ "metadata": {},
313
+ "outputs": [
314
+ {
315
+ "data": {
316
+ "text/plain": [
317
+ "ShowResponse(modified_at=datetime.datetime(2026, 2, 27, 8, 7, 46, 993669, tzinfo=TzInfo(0)), template='{{- range .Messages }}<|{{ .Role }}|>\\n{{ .Content }}</s>\\n{{ end }}<|assistant|>\\n', modelfile='# Modelfile generated by \"ollama show\"\\n# To build a new Modelfile based on this, replace FROM with:\\n# FROM tinytest-gguf:latest\\n\\nFROM /media/sarverott/setternet-A22/House_of_Anubis/dynamic-morgue-valut/OLLAMA/models/blobs/sha256-2af3b81862c6be03c769683af18efdadb2c33f60ff32ab6f83e42c043d6c7816\\nTEMPLATE \"{{- range .Messages }}<|{{ .Role }}|>\\n{{ .Content }}</s>\\n{{ end }}<|assistant|>\\n\"\\nPARAMETER stop <|system|>\\nPARAMETER stop </s>\\nPARAMETER stop <|user|>\\nPARAMETER stop <|assistant|>\\n', license=None, details=ModelDetails(parent_model='/media/sarverott/setternet-A22/House_of_Anubis/dynamic-morgue-valut/OLLAMA/models/blobs/sha256-2af3b81862c6be03c769683af18efdadb2c33f60ff32ab6f83e42c043d6c7816', format='gguf', family='llama', families=['llama'], parameter_size='1.1B', quantization_level='Q4_0'), modelinfo={'general.architecture': 'llama', 'general.file_type': 2, 'general.parameter_count': 1100048384, 'general.quantization_version': 2, 'llama.attention.head_count': 32, 'llama.attention.head_count_kv': 4, 'llama.attention.layer_norm_rms_epsilon': 1e-05, 'llama.block_count': 22, 'llama.context_length': 2048, 'llama.embedding_length': 2048, 'llama.feed_forward_length': 5632, 'llama.rope.dimension_count': 64, 'llama.rope.freq_base': 10000, 'tokenizer.ggml.bos_token_id': 1, 'tokenizer.ggml.eos_token_id': 2, 'tokenizer.ggml.merges': None, 'tokenizer.ggml.model': 'llama', 'tokenizer.ggml.padding_token_id': 2, 'tokenizer.ggml.scores': None, 'tokenizer.ggml.token_type': None, 'tokenizer.ggml.tokens': None, 'tokenizer.ggml.unknown_token_id': 0}, parameters='stop \"<|system|>\"\\nstop \"</s>\"\\nstop \"<|user|>\"\\nstop \"<|assistant|>\"', capabilities=['completion'])"
318
+ ]
319
+ },
320
+ "execution_count": 8,
321
+ "metadata": {},
322
+ "output_type": "execute_result"
323
+ }
324
+ ],
325
+ "source": [
326
+ "#client.show(\"tinytest-gguf:latest\")\n",
327
+ "#gemma3:latest\n",
328
+ "#gemma3n:latest\n",
329
+ "#gemma3:1b\n",
330
+ "#gemma2:latest\n",
331
+ "#codegemma:latest\n",
332
+ "#shieldgemma:latest\n",
333
+ "#gemma:latest\n",
334
+ "\n"
335
+ ]
336
+ },
337
+ {
338
+ "cell_type": "code",
339
+ "execution_count": 9,
340
+ "id": "83c1605f-e3fe-41e5-b6cb-1eebd450a253",
341
+ "metadata": {},
342
+ "outputs": [
343
+ {
344
+ "data": {
345
+ "text/plain": [
346
+ "ProgressResponse(status='success', completed=None, total=None, digest=None)"
347
+ ]
348
+ },
349
+ "execution_count": 9,
350
+ "metadata": {},
351
+ "output_type": "execute_result"
352
+ }
353
+ ],
354
+ "source": [
355
+ "client.pull(\"functiongemma:latest\")"
356
+ ]
357
+ },
358
+ {
359
+ "cell_type": "code",
360
+ "execution_count": null,
361
+ "id": "5feb7263-e18e-4c98-916d-d4f4fa3bba1d",
362
+ "metadata": {},
363
+ "outputs": [],
364
+ "source": [
365
+ "[\"embeddinggemma:latest\",\"functiongemma:latest\",\"translategemma:4b\","
366
+ ]
367
+ }
368
+ ],
369
+ "metadata": {
370
+ "kernelspec": {
371
+ "display_name": "Python 3 (ipykernel)",
372
+ "language": "python",
373
+ "name": "python3"
374
+ },
375
+ "language_info": {
376
+ "codemirror_mode": {
377
+ "name": "ipython",
378
+ "version": 3
379
+ },
380
+ "file_extension": ".py",
381
+ "mimetype": "text/x-python",
382
+ "name": "python",
383
+ "nbconvert_exporter": "python",
384
+ "pygments_lexer": "ipython3",
385
+ "version": "3.12.12"
386
+ }
387
+ },
388
+ "nbformat": 4,
389
+ "nbformat_minor": 5
390
+ }
uv.lock CHANGED
The diff for this file is too large to render. See raw diff