Update README.md
Browse files
README.md
CHANGED
|
@@ -15,9 +15,9 @@ tags:
|
|
| 15 |
|
| 16 |
YingLong model is introduced in this [paper](xxxxxxxx) (coming soon). This version is pre-trained on **78B** time points. More details can be found at our [github](https://github.com/wxie9/YingLong/).
|
| 17 |
|
|
|
|
| 18 |
|
| 19 |
## Quickstart
|
| 20 |
-
We currently only suppoert Ampere, Ada, or Hopper GPUs (e.g., A100, RTX 3090, RTX 4090, H100).
|
| 21 |
|
| 22 |
```bash
|
| 23 |
pip install xformers transformers
|
|
@@ -26,13 +26,15 @@ git clone https://github.com/Dao-AILab/flash-attention && cd flash-attention
|
|
| 26 |
cd csrc/rotary && pip install .
|
| 27 |
cd ../layer_norm && pip install .
|
| 28 |
```
|
|
|
|
|
|
|
| 29 |
|
| 30 |
```python
|
| 31 |
import torch
|
| 32 |
from transformers import AutoModelForCausalLM
|
| 33 |
|
| 34 |
# load pretrain model
|
| 35 |
-
model = AutoModelForCausalLM.from_pretrained('
|
| 36 |
|
| 37 |
# prepare input
|
| 38 |
batch_size, lookback_length = 1, 2880
|
|
@@ -48,22 +50,8 @@ print(output.shape)
|
|
| 48 |
A notebook example is also provided [here](https://github.com/wxie9/YingLong/blob/main/quickstart_zero_shot.ipynb). The sample codes for long-term forecasting tasks and gift-eval tasks are provided at [link](https://github.com/wxie9/YingLong/tree/main).
|
| 49 |
|
| 50 |
|
| 51 |
-
## Specification
|
| 52 |
-
|
| 53 |
-
<!-- * Architecture: Causal Transformer (Decoder-only)
|
| 54 |
-
* Pre-training Scale: 260B time points
|
| 55 |
-
* Context Length: up to 2880
|
| 56 |
-
* Parameter Count: 84M
|
| 57 |
-
* Patch Length: 96
|
| 58 |
-
* Number of Layers: 8
|
| 59 |
-
-->
|
| 60 |
-
|
| 61 |
-
<!-- ## Acknowledgments
|
| 62 |
-
|
| 63 |
-
This work was supported by the National Natural Science Foundation of China (62022050 and U2342217), the BNRist Innovation Fund (BNR2024RC01010), and the National Engineering Research Center for Big Data Software.
|
| 64 |
|
| 65 |
-
The model is mostly built from the Internet public time series dataset, which comes from different research teams and providers. We sincerely thank all individuals and organizations who have contributed the data. Without their generous sharing, this model would not have existed.
|
| 66 |
-
-->
|
| 67 |
|
| 68 |
## Citation
|
| 69 |
Coming soon...
|
|
|
|
| 15 |
|
| 16 |
YingLong model is introduced in this [paper](xxxxxxxx) (coming soon). This version is pre-trained on **78B** time points. More details can be found at our [github](https://github.com/wxie9/YingLong/).
|
| 17 |
|
| 18 |
+
|
| 19 |
|
| 20 |
## Quickstart
|
|
|
|
| 21 |
|
| 22 |
```bash
|
| 23 |
pip install xformers transformers
|
|
|
|
| 26 |
cd csrc/rotary && pip install .
|
| 27 |
cd ../layer_norm && pip install .
|
| 28 |
```
|
| 29 |
+
The flash attention is not required. If you use V100 or other GPU doesn't support flash attention, just change the FlashAttention2Available = RequirementCache("flash-attn>=2.0.0.post1") to
|
| 30 |
+
FlashAttention2Available = False in the model.py file. It should be able to run.
|
| 31 |
|
| 32 |
```python
|
| 33 |
import torch
|
| 34 |
from transformers import AutoModelForCausalLM
|
| 35 |
|
| 36 |
# load pretrain model
|
| 37 |
+
model = AutoModelForCausalLM.from_pretrained('qcw2333/YingLong_50m', trust_remote_code=True,torch_dtype=torch.bfloat16).cuda()
|
| 38 |
|
| 39 |
# prepare input
|
| 40 |
batch_size, lookback_length = 1, 2880
|
|
|
|
| 50 |
A notebook example is also provided [here](https://github.com/wxie9/YingLong/blob/main/quickstart_zero_shot.ipynb). The sample codes for long-term forecasting tasks and gift-eval tasks are provided at [link](https://github.com/wxie9/YingLong/tree/main).
|
| 51 |
|
| 52 |
|
| 53 |
+
<!-- ## Specification -->
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 54 |
|
|
|
|
|
|
|
| 55 |
|
| 56 |
## Citation
|
| 57 |
Coming soon...
|