RthItalia commited on
Commit
b72aa93
·
verified ·
1 Parent(s): 18f9e9a

Release AICE-v1: README

Browse files
Files changed (1) hide show
  1. README.md +133 -0
README.md ADDED
@@ -0,0 +1,133 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - it
5
+ pipeline_tag: text-generation
6
+ tags:
7
+ - gguf
8
+ - code
9
+ - instruct
10
+ - llama
11
+ ---
12
+
13
+ # MODEL_CARD - PINDARO AI CODE
14
+
15
+ Date: 2026-03-02
16
+ Model path: `e:\Pindaro\PINDARO AI CODE`
17
+
18
+ ## 1. Model Identity
19
+ - Name: `PINDARO AI CODE`
20
+ - Family: LLaMA-style causal LM
21
+ - Intended role: coding assistant
22
+ - Format support:
23
+ - Hugging Face (`model.safetensors`)
24
+ - No GGUF artifact in this folder
25
+
26
+ ## 2. Technical Specs
27
+ - Architecture: `LlamaForCausalLM`
28
+ - `model_type`: `llama`
29
+ - Layers: `22`
30
+ - Hidden size: `2048`
31
+ - Attention heads: `32`
32
+ - KV heads: `4`
33
+ - Intermediate size: `5632`
34
+ - Max context: `2048`
35
+ - Vocab size: `32002`
36
+ - Tensor count in safetensors: `201`
37
+ - Parameter count (computed): `1,100,056,576`
38
+ - Dtype in config: `float16`
39
+
40
+ ## 3. Chat / Prompt Format
41
+ Template is aligned to registered special tokens:
42
+ - `<|noesis|>` (id `32000`)
43
+ - `<|end|>` (id `32001`)
44
+
45
+ Configured template:
46
+ ````jinja
47
+ {{ bos_token }}{% for message in messages %}<|noesis|>
48
+ {% if message['role'] == 'system' %}### System
49
+ {{ message['content'] }}
50
+ {% elif message['role'] == 'user' %}### Question
51
+ {{ message['content'] }}
52
+ {% elif message['role'] == 'assistant' %}### Answer
53
+ {{ message['content'] }}
54
+ {% endif %}<|end|>
55
+ {% endfor %}{% if add_generation_prompt %}<|noesis|>
56
+ ### Answer
57
+ ```
58
+ {% endif %}
59
+ ````
60
+
61
+ ## 4. Local Artifact Integrity (SHA256)
62
+ - `model.safetensors`: `F77C27B8BABF9FCAB83A7DC68BA58934E8C8C031C9F10B4B73E802D4FBFE0CEC`
63
+ - `config.json`: `B37C45060F3E2F5F9B91903C9CCB32F3C21076E809954FDA6C01D987CD8F25CC`
64
+ - `generation_config.json`: `6FF47E725C0EC6D0F1895670DE7EE68E61A4F99703F6C8E89AEA6AB14EA02DC3`
65
+ - `tokenizer_config.json`: `4AE48D4963835B0767DC6510D1DF8AE76E636ED35BE021F02F5B344353BDFA21`
66
+ - `tokenizer.model`: `9E556AFD44213B6BD1BE2B850EBBBD98F5481437A8021AFAF58EE7FB1818D347`
67
+
68
+ ## 5. Smoke Tests (2026-03-02)
69
+ Environment:
70
+ - Python `3.11.9`
71
+ - Transformers `4.57.3`
72
+ - Torch `2.10.0+cpu`
73
+
74
+ Results:
75
+ - AutoConfig load: PASS
76
+ - AutoTokenizer load: PASS
77
+ - AutoModel load: PASS
78
+ - Chat-template render: PASS
79
+ - Template special-token alignment: PASS
80
+ - Deterministic generation: PASS
81
+
82
+ Observed non-blocking warning:
83
+ - Folder name with spaces may trigger a Python module-name warning in some runtimes.
84
+
85
+ ## 6. Known Issues
86
+ 1. Folder-name warning risk
87
+ - `PINDARO AI CODE` has spaces; some tools warn on module naming.
88
+
89
+ 2. Attention-mask warning in some calls
90
+ - As `pad_token` equals `eos_token`, pass `attention_mask` explicitly for stable behavior.
91
+
92
+ ## 7. Recommended Next Steps
93
+ 1. Optional packaging cleanup
94
+ - Rename folder to a no-space slug (example: `PINDARO_AI_CODE`) when compatible with your deployment scripts.
95
+
96
+ 2. Add coding eval gate
97
+ - HumanEval pass@1
98
+ - MBPP subset
99
+ - Prompt-format adherence checks
100
+
101
+ ## 8. Usage Example
102
+ ```python
103
+ import torch
104
+ from transformers import AutoTokenizer, AutoModelForCausalLM
105
+
106
+ path = r"e:\Pindaro\PINDARO AI CODE"
107
+ tokenizer = AutoTokenizer.from_pretrained(path, local_files_only=True)
108
+ model = AutoModelForCausalLM.from_pretrained(path, local_files_only=True, dtype=torch.float16)
109
+
110
+ messages = [
111
+ {"role": "system", "content": "You are a coding assistant."},
112
+ {"role": "user", "content": "Write a Python function add(a, b)."},
113
+ ]
114
+
115
+ inputs = tokenizer.apply_chat_template(
116
+ messages,
117
+ tokenize=True,
118
+ add_generation_prompt=True,
119
+ return_tensors="pt",
120
+ )
121
+ outputs = model.generate(inputs, max_new_tokens=80, do_sample=False)
122
+ print(tokenizer.decode(outputs[0], skip_special_tokens=False))
123
+ ```
124
+
125
+ ## 9. Limitations and Safety
126
+ - No training-data statement is included in this folder.
127
+ - No official benchmark sheet is included.
128
+ - Code generation can be plausible but wrong; always run tests.
129
+
130
+ ## 10. Release Readiness
131
+ Current status: READY FOR LOCAL USE.
132
+ - Packaging/runtime blockers are resolved.
133
+ - Remaining items are evaluation and packaging polish.