Phase-Technologies commited on
Commit
9c90934
·
verified ·
1 Parent(s): 0f06821

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -9
README.md CHANGED
@@ -1,14 +1,17 @@
1
  ---
2
- library_name: 'custom'
3
  tags:
4
- - custom-architecture
5
- - numpy
6
- - chatbot
7
- - text-generation
8
  license: mit
9
  metrics:
10
- - loss
11
- - perplexity
 
 
 
12
  ---
13
 
14
  # HRAN Chatbot Model Card
@@ -17,7 +20,7 @@ metrics:
17
 
18
  The architecture is strictly derived from concepts in Simon Haykin's Neural Networks and Learning Machines (3rd Ed.), actively challenging modern transformer defaults by replacing dot-product attention and standard activations with biologically and mathematically grounded alternatives.
19
 
20
- * **Developer**: Phase-Technologies
21
  * **Model Type**: Custom Sequence-to-Sequence Language Model
22
  * **Parameters**: ~1.01 Million
23
  * **Framework**: Pure NumPy
@@ -32,6 +35,11 @@ HRAN abandons several standard transformer conventions in favor of experimental
32
  * **Lateral Inhibition Gate (Ch.9)**: Introduces competitive learning where winning activations are amplified and weak ones suppressed, producing sparse, discriminative representations.
33
  * **Wiener-SNR Gradient Scaling (Ch.3)**: Scales parameter updates by local signal-to-noise ratio, allowing high-signal weights to learn quickly while suppressing noisy weight updates.
34
 
 
 
 
 
 
35
  ## Training Data
36
  The model was trained on a highly curated, 100% original dataset of 235 question-answer pairs (augmented to 1,040 samples). The dataset spans deep topics including neural network architecture, philosophy, physics, mathematics, and Haykin's specific theories.
37
 
@@ -80,4 +88,4 @@ model.load(weights_path)
80
  # 5. Generate Text
81
  response = hran.generate_response(model, tokenizer, "What is attention?")
82
  print(response)
83
- ```
 
1
  ---
2
+ library_name: custom
3
  tags:
4
+ - custom-architecture
5
+ - numpy
6
+ - chatbot
7
+ - text-generation
8
  license: mit
9
  metrics:
10
+ - loss
11
+ - perplexity
12
+ language:
13
+ - en
14
+ pipeline_tag: text-generation
15
  ---
16
 
17
  # HRAN Chatbot Model Card
 
20
 
21
  The architecture is strictly derived from concepts in Simon Haykin's Neural Networks and Learning Machines (3rd Ed.), actively challenging modern transformer defaults by replacing dot-product attention and standard activations with biologically and mathematically grounded alternatives.
22
 
23
+ * **Developer**: Soham Pal
24
  * **Model Type**: Custom Sequence-to-Sequence Language Model
25
  * **Parameters**: ~1.01 Million
26
  * **Framework**: Pure NumPy
 
35
  * **Lateral Inhibition Gate (Ch.9)**: Introduces competitive learning where winning activations are amplified and weak ones suppressed, producing sparse, discriminative representations.
36
  * **Wiener-SNR Gradient Scaling (Ch.3)**: Scales parameter updates by local signal-to-noise ratio, allowing high-signal weights to learn quickly while suppressing noisy weight updates.
37
 
38
+ ## Loss Graph
39
+
40
+ ![HRAN Training Loss](hran_training_loss.png)
41
+
42
+
43
  ## Training Data
44
  The model was trained on a highly curated, 100% original dataset of 235 question-answer pairs (augmented to 1,040 samples). The dataset spans deep topics including neural network architecture, philosophy, physics, mathematics, and Haykin's specific theories.
45
 
 
88
  # 5. Generate Text
89
  response = hran.generate_response(model, tokenizer, "What is attention?")
90
  print(response)
91
+ ```