qcw2333 commited on
Commit
5b93afa
·
verified ·
1 Parent(s): aa012a5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -17
README.md CHANGED
@@ -15,9 +15,9 @@ tags:
15
 
16
  YingLong model is introduced in this [paper](xxxxxxxx) (coming soon). This version is pre-trained on **78B** time points. More details can be found at our [github](https://github.com/wxie9/YingLong/).
17
 
 
18
 
19
  ## Quickstart
20
- We currently only suppoert Ampere, Ada, or Hopper GPUs (e.g., A100, RTX 3090, RTX 4090, H100).
21
 
22
  ```bash
23
  pip install xformers transformers
@@ -26,13 +26,15 @@ git clone https://github.com/Dao-AILab/flash-attention && cd flash-attention
26
  cd csrc/rotary && pip install .
27
  cd ../layer_norm && pip install .
28
  ```
 
 
29
 
30
  ```python
31
  import torch
32
  from transformers import AutoModelForCausalLM
33
 
34
  # load pretrain model
35
- model = AutoModelForCausalLM.from_pretrained('qcw1314/YingLong_50m', trust_remote_code=True,torch_dtype=torch.bfloat16).cuda()
36
 
37
  # prepare input
38
  batch_size, lookback_length = 1, 2880
@@ -48,22 +50,8 @@ print(output.shape)
48
  A notebook example is also provided [here](https://github.com/wxie9/YingLong/blob/main/quickstart_zero_shot.ipynb). The sample codes for long-term forecasting tasks and gift-eval tasks are provided at [link](https://github.com/wxie9/YingLong/tree/main).
49
 
50
 
51
- ## Specification
52
-
53
- <!-- * Architecture: Causal Transformer (Decoder-only)
54
- * Pre-training Scale: 260B time points
55
- * Context Length: up to 2880
56
- * Parameter Count: 84M
57
- * Patch Length: 96
58
- * Number of Layers: 8
59
- -->
60
-
61
- <!-- ## Acknowledgments
62
-
63
- This work was supported by the National Natural Science Foundation of China (62022050 and U2342217), the BNRist Innovation Fund (BNR2024RC01010), and the National Engineering Research Center for Big Data Software.
64
 
65
- The model is mostly built from the Internet public time series dataset, which comes from different research teams and providers. We sincerely thank all individuals and organizations who have contributed the data. Without their generous sharing, this model would not have existed.
66
- -->
67
 
68
  ## Citation
69
  Coming soon...
 
15
 
16
  YingLong model is introduced in this [paper](xxxxxxxx) (coming soon). This version is pre-trained on **78B** time points. More details can be found at our [github](https://github.com/wxie9/YingLong/).
17
 
18
+
19
 
20
  ## Quickstart
 
21
 
22
  ```bash
23
  pip install xformers transformers
 
26
  cd csrc/rotary && pip install .
27
  cd ../layer_norm && pip install .
28
  ```
29
+ The flash attention is not required. If you use V100 or other GPU doesn't support flash attention, just change the FlashAttention2Available = RequirementCache("flash-attn>=2.0.0.post1") to
30
+ FlashAttention2Available = False in the model.py file. It should be able to run.
31
 
32
  ```python
33
  import torch
34
  from transformers import AutoModelForCausalLM
35
 
36
  # load pretrain model
37
+ model = AutoModelForCausalLM.from_pretrained('qcw2333/YingLong_50m', trust_remote_code=True,torch_dtype=torch.bfloat16).cuda()
38
 
39
  # prepare input
40
  batch_size, lookback_length = 1, 2880
 
50
  A notebook example is also provided [here](https://github.com/wxie9/YingLong/blob/main/quickstart_zero_shot.ipynb). The sample codes for long-term forecasting tasks and gift-eval tasks are provided at [link](https://github.com/wxie9/YingLong/tree/main).
51
 
52
 
53
+ <!-- ## Specification -->
 
 
 
 
 
 
 
 
 
 
 
 
54
 
 
 
55
 
56
  ## Citation
57
  Coming soon...