File size: 1,119 Bytes
6630eef 74d16cc 6630eef 2693b63 bc52414 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 | ---
license: mit
---
# Model Card for mpyt5_e15
<!-- Provide a quick summary of what the model is/does. [Optional] -->
事前に自然言語だけでなくPythonを学習したモデル
# Training Details
## Training Data
<!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
Python Code (1.05GB)
## Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
- MLM
- python vocab (https://huggingface.co/kkuramitsu/mt5-pytoken)
### Preprocessing
mT5 + Python
### Speeds, Sizes, Times
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
- mT5-small(300M Paramators)
- max_length = 128
# Model Version
- *epoch5: https://huggingface.co/Roy029/mpyt5_e5
- *epoch10: This Model
- *epoch15: https://huggingface.co/Roy029/mpyt5_e15
- *epoch20: https://huggingface.co/Roy029/mpyt5_e20 |