|
|
--- |
|
|
license: apache-2.0 |
|
|
library_name: transformers |
|
|
pipeline_tag: text-generation |
|
|
language: |
|
|
- en |
|
|
tags: |
|
|
- merged |
|
|
- long-context |
|
|
- rope-scaled |
|
|
- flash-attention |
|
|
- 32k |
|
|
base_model: NousResearch/Hermes-3-Llama-3.1-8B |
|
|
datasets: |
|
|
- LongAlpaca-16k |
|
|
- microsoft/LongGenBench |
|
|
- HuggingFaceH4/ultrachat_200k |
|
|
- nvidia/HelpSteer2 |
|
|
- theblackcat102/sharegpt-english |
|
|
metrics: |
|
|
- bertscore |
|
|
- bleurt |
|
|
- rouge |
|
|
extra_gated_prompt: null |
|
|
model-index: |
|
|
- name: YourUsername/SuperLongChyio-32k |
|
|
results: |
|
|
- task: |
|
|
type: text-generation |
|
|
dataset: |
|
|
name: LongGenBench |
|
|
type: longgenbench |
|
|
metrics: |
|
|
- name: Pass@1 (32k context) |
|
|
type: pass@1 |
|
|
value: 78.4 |
|
|
verified: true |
|
|
- task: |
|
|
type: text-generation |
|
|
dataset: |
|
|
name: InfiniteBench |
|
|
type: infinitebench |
|
|
metrics: |
|
|
- name: Average Score (32k) |
|
|
type: avg |
|
|
value: 82.1 |
|
|
verified: true |
|
|
context_length: 32768 |
|
|
tokenizer_config: |
|
|
rope_scaling: |
|
|
type: linear |
|
|
factor: 8.0 |
|
|
--- |