YModel1.1 / README.md
SnifferCaptain's picture
Update README.md
2da1b07 verified
metadata
license: mit
language:
  - zh
  - en
pipeline_tag: text-generation
tags:
  - agent

YModel1.1

structure

  • using SnifferCaptain's LoE (lack of expert) layer as feed forward.
  • using SnifferCaptain's PEGA (Position Embedding Gate Attention) as Transformer attention layer
  • using additional identity link between ffn's intermediate part.