YModel1.1 / README.md
SnifferCaptain's picture
Update README.md
2da1b07 verified
---
license: mit
language:
- zh
- en
pipeline_tag: text-generation
tags:
- agent
---
# YModel1.1
## structure
- using SnifferCaptain's LoE (lack of expert) layer as feed forward.
- using SnifferCaptain's PEGA (Position Embedding Gate Attention) as Transformer attention layer
- using additional identity link between ffn's intermediate part.