Spaces:
Build error
Build error
Commit
·
7855334
1
Parent(s):
84ec87c
better about and readme docs
Browse files- README.md +7 -1
- about_package.md +10 -4
README.md
CHANGED
|
@@ -14,7 +14,13 @@ license: mit
|
|
| 14 |
# Numpy-Neuron
|
| 15 |
|
| 16 |
A small, simple neural network framework built using only [numpy](https://numpy.org) and python (duh).
|
| 17 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 18 |
|
| 19 |
```py
|
| 20 |
from sklearn import datasets
|
|
|
|
| 14 |
# Numpy-Neuron
|
| 15 |
|
| 16 |
A small, simple neural network framework built using only [numpy](https://numpy.org) and python (duh).
|
| 17 |
+
|
| 18 |
+
## Install
|
| 19 |
+
|
| 20 |
+
`pip install numpy_neuron`
|
| 21 |
+
|
| 22 |
+
|
| 23 |
+
## Example
|
| 24 |
|
| 25 |
```py
|
| 26 |
from sklearn import datasets
|
about_package.md
CHANGED
|
@@ -1,7 +1,13 @@
|
|
| 1 |
# Numpy-Neuron
|
| 2 |
|
| 3 |
A small, simple neural network framework built using only [numpy](https://numpy.org) and python (duh).
|
| 4 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5 |
|
| 6 |
```py
|
| 7 |
from sklearn import datasets
|
|
@@ -71,11 +77,11 @@ if __name__ == "__main__":
|
|
| 71 |
train_nn_classifier()
|
| 72 |
```
|
| 73 |
|
| 74 |
-
|
| 75 |
## Roadmap
|
| 76 |
|
| 77 |
**Optimizers**
|
| 78 |
-
I would love to add the ability to modify the learning rate over each epoch to ensure
|
| 79 |
-
that the gradient descent algorithm does not get stuck in local minima as easily.
|
| 80 |
|
|
|
|
|
|
|
|
|
|
| 81 |
|
|
|
|
| 1 |
# Numpy-Neuron
|
| 2 |
|
| 3 |
A small, simple neural network framework built using only [numpy](https://numpy.org) and python (duh).
|
| 4 |
+
|
| 5 |
+
## Install
|
| 6 |
+
|
| 7 |
+
`pip install numpy_neuron`
|
| 8 |
+
|
| 9 |
+
|
| 10 |
+
## Example
|
| 11 |
|
| 12 |
```py
|
| 13 |
from sklearn import datasets
|
|
|
|
| 77 |
train_nn_classifier()
|
| 78 |
```
|
| 79 |
|
|
|
|
| 80 |
## Roadmap
|
| 81 |
|
| 82 |
**Optimizers**
|
|
|
|
|
|
|
| 83 |
|
| 84 |
+
Currently the learning rate in a NN object is static during training. I would like
|
| 85 |
+
to work on developing at least the functionality for the Adam optimizer at some point.
|
| 86 |
+
This would help prevent getting stuck in local minima of the loss function.
|
| 87 |
|