luigicfilho commited on
Commit
274b51f
·
verified ·
1 Parent(s): 836901d

commit first version

Browse files

Add new model, statistically generated

README.md CHANGED
@@ -1,3 +1,60 @@
1
  ---
 
 
2
  license: mit
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
  license: mit
5
+ library_name: transformers
6
+ tags:
7
+ - generated
8
+ - numerical-generation
9
+ - weight-space
10
+ - text-generation
11
+ inference: true
12
+ model_creator: LCDev-LLMGen
13
+ widget:
14
+ - text: "The future of AI is"
15
+ example_title: "AI Future"
16
+ - text: "Once upon a time"
17
+ example_title: "Story"
18
  ---
19
+
20
+ # Numera: The Numerically Generated Model
21
+
22
+ This model was automatically generated using **LCDev-Numera**, a proprietary tool for numerical model generation.
23
+
24
+ ## Model Details
25
+
26
+ - **Model Name:** Numera (Gen-1)
27
+ - **Generated By:** LCDev-Numera
28
+ - **Base Architecture:** GPT-2
29
+ - **Type:** Statistical Weight Generation
30
+ - **Date Generated:** 2026-01-29
31
+
32
+ ## Intended Use
33
+
34
+ This model is intended for research into:
35
+ - Weight space analysis of Large Language Models.
36
+ - Statistical properties of model weights.
37
+ - Experimental initialization checkpoints.
38
+
39
+ **Note:** This model is a statistical approximation and not a trained model. It may exhibit repetitive behaviors or lack specific factual knowledge.
40
+
41
+ ## How to Use
42
+
43
+ ```python
44
+ from transformers import AutoTokenizer, AutoModelForCausalLM
45
+
46
+ model_name = "./Numera-v1" # Or the huggingface repo id if uploaded
47
+
48
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
49
+ model = AutoModelForCausalLM.from_pretrained(model_name)
50
+
51
+ prompt = "The future of AI is"
52
+ inputs = tokenizer(prompt, return_tensors="pt")
53
+
54
+ outputs = model.generate(**inputs, max_length=50)
55
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
56
+ ```
57
+
58
+ ## License
59
+
60
+ MIT
config.json ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_num_labels": 1,
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPT2LMHeadModel"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "bos_token_id": 50256,
9
+ "dtype": "float32",
10
+ "embd_pdrop": 0.1,
11
+ "eos_token_id": 50256,
12
+ "id2label": {
13
+ "0": "LABEL_0"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "label2id": {
17
+ "LABEL_0": 0
18
+ },
19
+ "layer_norm_epsilon": 1e-05,
20
+ "model_type": "gpt2",
21
+ "n_ctx": 1024,
22
+ "n_embd": 768,
23
+ "n_head": 12,
24
+ "n_inner": null,
25
+ "n_layer": 6,
26
+ "n_positions": 1024,
27
+ "reorder_and_upcast_attn": false,
28
+ "resid_pdrop": 0.1,
29
+ "scale_attn_by_inverse_layer_idx": false,
30
+ "scale_attn_weights": true,
31
+ "summary_activation": null,
32
+ "summary_first_dropout": 0.1,
33
+ "summary_proj_to_labels": true,
34
+ "summary_type": "cls_index",
35
+ "summary_use_proj": true,
36
+ "task_specific_params": {
37
+ "text-generation": {
38
+ "do_sample": true,
39
+ "max_length": 50
40
+ }
41
+ },
42
+ "transformers_version": "4.57.3",
43
+ "use_cache": true,
44
+ "vocab_size": 50257,
45
+ "weight_space_meta": {
46
+ "architecture": "gpt2",
47
+ "family": "Numera",
48
+ "generated_by": "LCDev-Numera",
49
+ "generation_method": "genf1",
50
+ "noise_scale": 0.0001,
51
+ "pca_components": 0.999,
52
+ "timestamp": "2026-01-29T12:15:11.135955"
53
+ }
54
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "eos_token_id": 50256,
5
+ "transformers_version": "4.57.3"
6
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a0007e0d519cd4899c2ccb809fe9f50255f78d66635580c715fde7305e975c3
3
+ size 327657928
special_tokens_map.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<|endoftext|>",
3
+ "eos_token": "<|endoftext|>",
4
+ "unk_token": "<|endoftext|>"
5
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "50256": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ }
12
+ },
13
+ "bos_token": "<|endoftext|>",
14
+ "clean_up_tokenization_spaces": false,
15
+ "eos_token": "<|endoftext|>",
16
+ "extra_special_tokens": {},
17
+ "model_max_length": 1024,
18
+ "tokenizer_class": "GPT2Tokenizer",
19
+ "unk_token": "<|endoftext|>"
20
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff