TallyFormer-Finance-51M
TallyFormer-Finance-51M is a 51M parameter transformer model specialized for finance instruction-following tasks.
Model Details
- Architecture: decoder-only transformer
- Parameters: ~ 51M (~197 MB in fp16 / bf16)
- Domain: Finance
- Context length: 256 tokens
- Tokenizer: GPT-2 base + custom special tokens (
<|user|>,<|assistant|>)
References
- GitHub repository: TallyFormer-Finance-51M
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support