ADAPT-Chase commited on
Commit
45266d3
·
verified ·
1 Parent(s): 7c43246

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +2 -0
  2. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/.optimizer.pt.87fdNN +3 -0
  3. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/.optimizer.pt.eqAZ1p +3 -0
  4. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/model-00001-of-00004.safetensors +3 -0
  5. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/model-00002-of-00004.safetensors +3 -0
  6. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/model-00003-of-00004.safetensors +3 -0
  7. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/model-00004-of-00004.safetensors +3 -0
  8. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/optimizer.pt.gz +3 -0
  9. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/tokenizer.json +0 -0
  10. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/tokenizer_config.json +207 -0
  11. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/trainer_state.json +721 -0
  12. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/vocab.json +0 -0
  13. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/added_tokens.json +24 -0
  14. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/config.json +29 -0
  15. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/generation_config.json +6 -0
  16. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/merges.txt +0 -0
  17. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/model-00001-of-00004.safetensors +3 -0
  18. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/model-00002-of-00004.safetensors +3 -0
  19. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/model-00003-of-00004.safetensors +3 -0
  20. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/model-00004-of-00004.safetensors +3 -0
  21. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/model.safetensors.index.json +346 -0
  22. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/optimizer.pt +3 -0
  23. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/model-00001-of-00004.safetensors +3 -0
  24. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/model-00002-of-00004.safetensors +3 -0
  25. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/model-00003-of-00004.safetensors +3 -0
  26. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/model-00004-of-00004.safetensors +3 -0
  27. platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/optimizer_backups.tar.gz +3 -0
  28. platform/aiml/models/vllm/torch_compile_cache/069ce3ea2b/rank_0_0/triton_cache/JGBJTSHFQO5ZMSSDR5JQ4N6IDDC3AKTV33H2WXFG26XWQP3IZ4ZA/triton_poi_fused_cat_1.ptx +528 -0
  29. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/__grp__triton_red_fused__to_copy_mean_pow_4.json +1 -0
  30. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.cubin +0 -0
  31. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.json +1 -0
  32. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.llir +169 -0
  33. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.ptx +469 -0
  34. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.ttgir +106 -0
  35. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.ttir +97 -0
  36. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/5TS244IYEIF4EECWF7IIP4LY7WII7Y4ZJDYZBCSZG4CRQWGLCERQ/__grp__triton_poi_fused_add_mul_sub_6.json +1 -0
  37. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/5TS244IYEIF4EECWF7IIP4LY7WII7Y4ZJDYZBCSZG4CRQWGLCERQ/triton_poi_fused_add_mul_sub_6.json +1 -0
  38. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/5TS244IYEIF4EECWF7IIP4LY7WII7Y4ZJDYZBCSZG4CRQWGLCERQ/triton_poi_fused_add_mul_sub_6.ttgir +196 -0
  39. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/5TS244IYEIF4EECWF7IIP4LY7WII7Y4ZJDYZBCSZG4CRQWGLCERQ/triton_poi_fused_add_mul_sub_6.ttir +195 -0
  40. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/7HDVHUPTEOX4FHYEQOTFHD5U573A7K7C62TYWKVZQQZ4QTR7356A/triton_red_fused__to_copy_mean_pow_4.json +1 -0
  41. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/7HDVHUPTEOX4FHYEQOTFHD5U573A7K7C62TYWKVZQQZ4QTR7356A/triton_red_fused__to_copy_mean_pow_4.llir +143 -0
  42. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/7HDVHUPTEOX4FHYEQOTFHD5U573A7K7C62TYWKVZQQZ4QTR7356A/triton_red_fused__to_copy_mean_pow_4.ptx +415 -0
  43. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/7HDVHUPTEOX4FHYEQOTFHD5U573A7K7C62TYWKVZQQZ4QTR7356A/triton_red_fused__to_copy_mean_pow_4.ttgir +96 -0
  44. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/CUASPVXQ7OI3JPFN55O2D57GL57SRLSB6LBTAARTQDPYHVXGYBSA/__grp__triton_poi_fused_add_mul_sub_6.json +1 -0
  45. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/CUASPVXQ7OI3JPFN55O2D57GL57SRLSB6LBTAARTQDPYHVXGYBSA/triton_poi_fused_add_mul_sub_6.llir +301 -0
  46. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/CUASPVXQ7OI3JPFN55O2D57GL57SRLSB6LBTAARTQDPYHVXGYBSA/triton_poi_fused_add_mul_sub_6.ptx +516 -0
  47. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/CUASPVXQ7OI3JPFN55O2D57GL57SRLSB6LBTAARTQDPYHVXGYBSA/triton_poi_fused_add_mul_sub_6.ttgir +196 -0
  48. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/CUASPVXQ7OI3JPFN55O2D57GL57SRLSB6LBTAARTQDPYHVXGYBSA/triton_poi_fused_add_mul_sub_6.ttir +195 -0
  49. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/D6B7FVPYPXXUJ4C3PXEACCPD4ZWL5EA6LGYT663ZACLUNJHJ5LAQ/__grp__triton_poi_fused_add_mul_sub_5.json +1 -0
  50. platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/D6B7FVPYPXXUJ4C3PXEACCPD4ZWL5EA6LGYT663ZACLUNJHJ5LAQ/triton_poi_fused_add_mul_sub_5.json +1 -0
.gitattributes CHANGED
@@ -196,3 +196,5 @@ platform/aiml/experiments/91b6033272a21bdbeef81b7999c45580a468795118fde6064492aa
196
  platform/aiml/experiments/89e6ca00b860ff181bc81f98651b5a6b422436a06d1f42e11e63def64d7ec59b filter=lfs diff=lfs merge=lfs -text
197
  platform/aiml/experiments/0cf14170a81e7da42e358eee102faa5f6900028f8cbf1c6f64d8f2014991cae3 filter=lfs diff=lfs merge=lfs -text
198
  platform/aiml/models/onnx/model.onnx_data filter=lfs diff=lfs merge=lfs -text
 
 
 
196
  platform/aiml/experiments/89e6ca00b860ff181bc81f98651b5a6b422436a06d1f42e11e63def64d7ec59b filter=lfs diff=lfs merge=lfs -text
197
  platform/aiml/experiments/0cf14170a81e7da42e358eee102faa5f6900028f8cbf1c6f64d8f2014991cae3 filter=lfs diff=lfs merge=lfs -text
198
  platform/aiml/models/onnx/model.onnx_data filter=lfs diff=lfs merge=lfs -text
199
+ platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/.optimizer.pt.87fdNN filter=lfs diff=lfs merge=lfs -text
200
+ platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/.optimizer.pt.eqAZ1p filter=lfs diff=lfs merge=lfs -text
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/.optimizer.pt.87fdNN ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba35e9808ea31db0c13b2677f2c1d5831dddbc954e25eccec3aaa796cf63a8a7
3
+ size 12309495808
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/.optimizer.pt.eqAZ1p ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f382e0071e505853f8d44b2d29a7b5234e07a48890a8608a14f03bb2822ab351
3
+ size 1320943616
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/model-00001-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5df3a13765fd51d1ae877873d54fe1c92bbde967203b87f81523d7abb8316391
3
+ size 4877660776
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/model-00002-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8bb462f37b7088cecad77e77ec61799ecc2960c9919a26f63d39a82d8db53b0e
3
+ size 4932751008
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/model-00003-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89f20ea1338c9b8e2c4f5fc87cd326394bac2691a8f6efca39dac0f9603d95c8
3
+ size 4330865200
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/model-00004-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86b07761727f98fe5be039799ddf14eb7a6621309797c4b4ebec1bef1cbf94f2
3
+ size 1089994880
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/optimizer.pt.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c7a653ee014e70773fed127fd6e35a702fb69c49f3add499255a7d478a2e510
3
+ size 1518862336
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/tokenizer_config.json ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
199
+ "clean_up_tokenization_spaces": false,
200
+ "eos_token": "<|endoftext|>",
201
+ "errors": "replace",
202
+ "model_max_length": 131072,
203
+ "pad_token": "<|endoftext|>",
204
+ "split_special_tokens": false,
205
+ "tokenizer_class": "Qwen2Tokenizer",
206
+ "unk_token": null
207
+ }
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/trainer_state.json ADDED
@@ -0,0 +1,721 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 21.333333333333332,
5
+ "eval_steps": 500,
6
+ "global_step": 1000,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.21333333333333335,
13
+ "grad_norm": 206.0,
14
+ "learning_rate": 1.1778563015312134e-07,
15
+ "loss": 3.4341,
16
+ "step": 10
17
+ },
18
+ {
19
+ "epoch": 0.4266666666666667,
20
+ "grad_norm": 992.0,
21
+ "learning_rate": 2.3557126030624267e-07,
22
+ "loss": 3.4156,
23
+ "step": 20
24
+ },
25
+ {
26
+ "epoch": 0.64,
27
+ "grad_norm": 2608.0,
28
+ "learning_rate": 3.53356890459364e-07,
29
+ "loss": 3.4391,
30
+ "step": 30
31
+ },
32
+ {
33
+ "epoch": 0.8533333333333334,
34
+ "grad_norm": 438.0,
35
+ "learning_rate": 4.7114252061248535e-07,
36
+ "loss": 3.3685,
37
+ "step": 40
38
+ },
39
+ {
40
+ "epoch": 1.0666666666666667,
41
+ "grad_norm": 238.0,
42
+ "learning_rate": 5.889281507656066e-07,
43
+ "loss": 3.3025,
44
+ "step": 50
45
+ },
46
+ {
47
+ "epoch": 1.28,
48
+ "grad_norm": 100.0,
49
+ "learning_rate": 7.06713780918728e-07,
50
+ "loss": 3.2549,
51
+ "step": 60
52
+ },
53
+ {
54
+ "epoch": 1.4933333333333334,
55
+ "grad_norm": 1784.0,
56
+ "learning_rate": 8.244994110718493e-07,
57
+ "loss": 3.1491,
58
+ "step": 70
59
+ },
60
+ {
61
+ "epoch": 1.7066666666666666,
62
+ "grad_norm": 89.5,
63
+ "learning_rate": 9.422850412249707e-07,
64
+ "loss": 2.9619,
65
+ "step": 80
66
+ },
67
+ {
68
+ "epoch": 1.92,
69
+ "grad_norm": 13312.0,
70
+ "learning_rate": 1.060070671378092e-06,
71
+ "loss": 2.8322,
72
+ "step": 90
73
+ },
74
+ {
75
+ "epoch": 2.1333333333333333,
76
+ "grad_norm": 808.0,
77
+ "learning_rate": 1.1778563015312133e-06,
78
+ "loss": 2.9565,
79
+ "step": 100
80
+ },
81
+ {
82
+ "epoch": 2.3466666666666667,
83
+ "grad_norm": 188.0,
84
+ "learning_rate": 1.2956419316843347e-06,
85
+ "loss": 2.7616,
86
+ "step": 110
87
+ },
88
+ {
89
+ "epoch": 2.56,
90
+ "grad_norm": 194.0,
91
+ "learning_rate": 1.413427561837456e-06,
92
+ "loss": 2.6079,
93
+ "step": 120
94
+ },
95
+ {
96
+ "epoch": 2.7733333333333334,
97
+ "grad_norm": 1256.0,
98
+ "learning_rate": 1.5312131919905772e-06,
99
+ "loss": 2.5776,
100
+ "step": 130
101
+ },
102
+ {
103
+ "epoch": 2.986666666666667,
104
+ "grad_norm": 49.75,
105
+ "learning_rate": 1.6489988221436987e-06,
106
+ "loss": 2.6403,
107
+ "step": 140
108
+ },
109
+ {
110
+ "epoch": 3.2,
111
+ "grad_norm": 876.0,
112
+ "learning_rate": 1.76678445229682e-06,
113
+ "loss": 2.4019,
114
+ "step": 150
115
+ },
116
+ {
117
+ "epoch": 3.413333333333333,
118
+ "grad_norm": 169.0,
119
+ "learning_rate": 1.8845700824499414e-06,
120
+ "loss": 2.2964,
121
+ "step": 160
122
+ },
123
+ {
124
+ "epoch": 3.626666666666667,
125
+ "grad_norm": 256.0,
126
+ "learning_rate": 2.002355712603063e-06,
127
+ "loss": 2.2499,
128
+ "step": 170
129
+ },
130
+ {
131
+ "epoch": 3.84,
132
+ "grad_norm": 1576.0,
133
+ "learning_rate": 2.120141342756184e-06,
134
+ "loss": 2.1755,
135
+ "step": 180
136
+ },
137
+ {
138
+ "epoch": 4.053333333333334,
139
+ "grad_norm": 592.0,
140
+ "learning_rate": 2.2379269729093053e-06,
141
+ "loss": 2.0546,
142
+ "step": 190
143
+ },
144
+ {
145
+ "epoch": 4.266666666666667,
146
+ "grad_norm": 12288.0,
147
+ "learning_rate": 2.3557126030624266e-06,
148
+ "loss": 1.9888,
149
+ "step": 200
150
+ },
151
+ {
152
+ "epoch": 4.48,
153
+ "grad_norm": 370.0,
154
+ "learning_rate": 2.473498233215548e-06,
155
+ "loss": 1.9038,
156
+ "step": 210
157
+ },
158
+ {
159
+ "epoch": 4.693333333333333,
160
+ "grad_norm": 254.0,
161
+ "learning_rate": 2.5912838633686695e-06,
162
+ "loss": 1.7628,
163
+ "step": 220
164
+ },
165
+ {
166
+ "epoch": 4.906666666666666,
167
+ "grad_norm": 241.0,
168
+ "learning_rate": 2.7090694935217903e-06,
169
+ "loss": 1.7098,
170
+ "step": 230
171
+ },
172
+ {
173
+ "epoch": 5.12,
174
+ "grad_norm": 29952.0,
175
+ "learning_rate": 2.826855123674912e-06,
176
+ "loss": 1.4924,
177
+ "step": 240
178
+ },
179
+ {
180
+ "epoch": 5.333333333333333,
181
+ "grad_norm": 13824.0,
182
+ "learning_rate": 2.9446407538280332e-06,
183
+ "loss": 1.2908,
184
+ "step": 250
185
+ },
186
+ {
187
+ "epoch": 5.546666666666667,
188
+ "grad_norm": 14336.0,
189
+ "learning_rate": 3.0624263839811545e-06,
190
+ "loss": 1.1921,
191
+ "step": 260
192
+ },
193
+ {
194
+ "epoch": 5.76,
195
+ "grad_norm": 418.0,
196
+ "learning_rate": 3.1802120141342757e-06,
197
+ "loss": 1.108,
198
+ "step": 270
199
+ },
200
+ {
201
+ "epoch": 5.973333333333334,
202
+ "grad_norm": 360.0,
203
+ "learning_rate": 3.2979976442873974e-06,
204
+ "loss": 1.0499,
205
+ "step": 280
206
+ },
207
+ {
208
+ "epoch": 6.1866666666666665,
209
+ "grad_norm": 936.0,
210
+ "learning_rate": 3.415783274440518e-06,
211
+ "loss": 0.9391,
212
+ "step": 290
213
+ },
214
+ {
215
+ "epoch": 6.4,
216
+ "grad_norm": 231.0,
217
+ "learning_rate": 3.53356890459364e-06,
218
+ "loss": 0.9643,
219
+ "step": 300
220
+ },
221
+ {
222
+ "epoch": 6.613333333333333,
223
+ "grad_norm": 17152.0,
224
+ "learning_rate": 3.651354534746761e-06,
225
+ "loss": 0.9206,
226
+ "step": 310
227
+ },
228
+ {
229
+ "epoch": 6.826666666666666,
230
+ "grad_norm": 29184.0,
231
+ "learning_rate": 3.7691401648998828e-06,
232
+ "loss": 0.8702,
233
+ "step": 320
234
+ },
235
+ {
236
+ "epoch": 7.04,
237
+ "grad_norm": 552.0,
238
+ "learning_rate": 3.886925795053004e-06,
239
+ "loss": 0.7931,
240
+ "step": 330
241
+ },
242
+ {
243
+ "epoch": 7.253333333333333,
244
+ "grad_norm": 696.0,
245
+ "learning_rate": 4.004711425206126e-06,
246
+ "loss": 0.8089,
247
+ "step": 340
248
+ },
249
+ {
250
+ "epoch": 7.466666666666667,
251
+ "grad_norm": 616.0,
252
+ "learning_rate": 4.122497055359246e-06,
253
+ "loss": 0.7657,
254
+ "step": 350
255
+ },
256
+ {
257
+ "epoch": 7.68,
258
+ "grad_norm": 43.75,
259
+ "learning_rate": 4.240282685512368e-06,
260
+ "loss": 0.7359,
261
+ "step": 360
262
+ },
263
+ {
264
+ "epoch": 7.8933333333333335,
265
+ "grad_norm": 496.0,
266
+ "learning_rate": 4.358068315665489e-06,
267
+ "loss": 0.6411,
268
+ "step": 370
269
+ },
270
+ {
271
+ "epoch": 8.106666666666667,
272
+ "grad_norm": 76.0,
273
+ "learning_rate": 4.475853945818611e-06,
274
+ "loss": 0.6151,
275
+ "step": 380
276
+ },
277
+ {
278
+ "epoch": 8.32,
279
+ "grad_norm": 6656.0,
280
+ "learning_rate": 4.593639575971732e-06,
281
+ "loss": 0.5215,
282
+ "step": 390
283
+ },
284
+ {
285
+ "epoch": 8.533333333333333,
286
+ "grad_norm": 5280.0,
287
+ "learning_rate": 4.711425206124853e-06,
288
+ "loss": 0.5172,
289
+ "step": 400
290
+ },
291
+ {
292
+ "epoch": 8.746666666666666,
293
+ "grad_norm": 134.0,
294
+ "learning_rate": 4.829210836277974e-06,
295
+ "loss": 0.4355,
296
+ "step": 410
297
+ },
298
+ {
299
+ "epoch": 8.96,
300
+ "grad_norm": 24.0,
301
+ "learning_rate": 4.946996466431096e-06,
302
+ "loss": 0.2923,
303
+ "step": 420
304
+ },
305
+ {
306
+ "epoch": 9.173333333333334,
307
+ "grad_norm": 7.84375,
308
+ "learning_rate": 5.064782096584218e-06,
309
+ "loss": 0.1739,
310
+ "step": 430
311
+ },
312
+ {
313
+ "epoch": 9.386666666666667,
314
+ "grad_norm": 2.8125,
315
+ "learning_rate": 5.182567726737339e-06,
316
+ "loss": 0.066,
317
+ "step": 440
318
+ },
319
+ {
320
+ "epoch": 9.6,
321
+ "grad_norm": 2064.0,
322
+ "learning_rate": 5.300353356890459e-06,
323
+ "loss": 0.0894,
324
+ "step": 450
325
+ },
326
+ {
327
+ "epoch": 9.813333333333333,
328
+ "grad_norm": 118.0,
329
+ "learning_rate": 5.418138987043581e-06,
330
+ "loss": 0.0747,
331
+ "step": 460
332
+ },
333
+ {
334
+ "epoch": 10.026666666666667,
335
+ "grad_norm": 9.9375,
336
+ "learning_rate": 5.535924617196703e-06,
337
+ "loss": 0.0359,
338
+ "step": 470
339
+ },
340
+ {
341
+ "epoch": 10.24,
342
+ "grad_norm": 3.0625,
343
+ "learning_rate": 5.653710247349824e-06,
344
+ "loss": 0.0307,
345
+ "step": 480
346
+ },
347
+ {
348
+ "epoch": 10.453333333333333,
349
+ "grad_norm": 4.03125,
350
+ "learning_rate": 5.771495877502945e-06,
351
+ "loss": 0.0245,
352
+ "step": 490
353
+ },
354
+ {
355
+ "epoch": 10.666666666666666,
356
+ "grad_norm": 2.53125,
357
+ "learning_rate": 5.8892815076560664e-06,
358
+ "loss": 0.0238,
359
+ "step": 500
360
+ },
361
+ {
362
+ "epoch": 10.88,
363
+ "grad_norm": 356.0,
364
+ "learning_rate": 6.0070671378091885e-06,
365
+ "loss": 0.0228,
366
+ "step": 510
367
+ },
368
+ {
369
+ "epoch": 11.093333333333334,
370
+ "grad_norm": 0.97265625,
371
+ "learning_rate": 6.124852767962309e-06,
372
+ "loss": 0.032,
373
+ "step": 520
374
+ },
375
+ {
376
+ "epoch": 11.306666666666667,
377
+ "grad_norm": 8512.0,
378
+ "learning_rate": 6.24263839811543e-06,
379
+ "loss": 0.0266,
380
+ "step": 530
381
+ },
382
+ {
383
+ "epoch": 11.52,
384
+ "grad_norm": 0.52734375,
385
+ "learning_rate": 6.360424028268551e-06,
386
+ "loss": 0.0234,
387
+ "step": 540
388
+ },
389
+ {
390
+ "epoch": 11.733333333333333,
391
+ "grad_norm": 1.0,
392
+ "learning_rate": 6.4782096584216735e-06,
393
+ "loss": 0.0216,
394
+ "step": 550
395
+ },
396
+ {
397
+ "epoch": 11.946666666666667,
398
+ "grad_norm": 0.60546875,
399
+ "learning_rate": 6.595995288574795e-06,
400
+ "loss": 0.021,
401
+ "step": 560
402
+ },
403
+ {
404
+ "epoch": 12.16,
405
+ "grad_norm": 0.55859375,
406
+ "learning_rate": 6.713780918727916e-06,
407
+ "loss": 0.0204,
408
+ "step": 570
409
+ },
410
+ {
411
+ "epoch": 12.373333333333333,
412
+ "grad_norm": 0.55078125,
413
+ "learning_rate": 6.831566548881036e-06,
414
+ "loss": 0.0205,
415
+ "step": 580
416
+ },
417
+ {
418
+ "epoch": 12.586666666666666,
419
+ "grad_norm": 0.51171875,
420
+ "learning_rate": 6.949352179034159e-06,
421
+ "loss": 0.0205,
422
+ "step": 590
423
+ },
424
+ {
425
+ "epoch": 12.8,
426
+ "grad_norm": 0.392578125,
427
+ "learning_rate": 7.06713780918728e-06,
428
+ "loss": 0.0205,
429
+ "step": 600
430
+ },
431
+ {
432
+ "epoch": 13.013333333333334,
433
+ "grad_norm": 0.50390625,
434
+ "learning_rate": 7.184923439340401e-06,
435
+ "loss": 0.0202,
436
+ "step": 610
437
+ },
438
+ {
439
+ "epoch": 13.226666666666667,
440
+ "grad_norm": 0.455078125,
441
+ "learning_rate": 7.302709069493522e-06,
442
+ "loss": 0.0202,
443
+ "step": 620
444
+ },
445
+ {
446
+ "epoch": 13.44,
447
+ "grad_norm": 0.81640625,
448
+ "learning_rate": 7.420494699646644e-06,
449
+ "loss": 0.0208,
450
+ "step": 630
451
+ },
452
+ {
453
+ "epoch": 13.653333333333332,
454
+ "grad_norm": 0.431640625,
455
+ "learning_rate": 7.5382803297997655e-06,
456
+ "loss": 0.0202,
457
+ "step": 640
458
+ },
459
+ {
460
+ "epoch": 13.866666666666667,
461
+ "grad_norm": 0.40625,
462
+ "learning_rate": 7.656065959952887e-06,
463
+ "loss": 0.0209,
464
+ "step": 650
465
+ },
466
+ {
467
+ "epoch": 14.08,
468
+ "grad_norm": 0.71484375,
469
+ "learning_rate": 7.773851590106007e-06,
470
+ "loss": 0.0207,
471
+ "step": 660
472
+ },
473
+ {
474
+ "epoch": 14.293333333333333,
475
+ "grad_norm": 0.61328125,
476
+ "learning_rate": 7.89163722025913e-06,
477
+ "loss": 0.0199,
478
+ "step": 670
479
+ },
480
+ {
481
+ "epoch": 14.506666666666666,
482
+ "grad_norm": 0.375,
483
+ "learning_rate": 8.009422850412251e-06,
484
+ "loss": 0.0197,
485
+ "step": 680
486
+ },
487
+ {
488
+ "epoch": 14.72,
489
+ "grad_norm": 0.71484375,
490
+ "learning_rate": 8.127208480565372e-06,
491
+ "loss": 0.021,
492
+ "step": 690
493
+ },
494
+ {
495
+ "epoch": 14.933333333333334,
496
+ "grad_norm": 0.373046875,
497
+ "learning_rate": 8.244994110718492e-06,
498
+ "loss": 0.0209,
499
+ "step": 700
500
+ },
501
+ {
502
+ "epoch": 15.146666666666667,
503
+ "grad_norm": 0.435546875,
504
+ "learning_rate": 8.362779740871614e-06,
505
+ "loss": 0.0201,
506
+ "step": 710
507
+ },
508
+ {
509
+ "epoch": 15.36,
510
+ "grad_norm": 0.5859375,
511
+ "learning_rate": 8.480565371024736e-06,
512
+ "loss": 0.0207,
513
+ "step": 720
514
+ },
515
+ {
516
+ "epoch": 15.573333333333334,
517
+ "grad_norm": 0.6875,
518
+ "learning_rate": 8.598351001177857e-06,
519
+ "loss": 0.0206,
520
+ "step": 730
521
+ },
522
+ {
523
+ "epoch": 15.786666666666667,
524
+ "grad_norm": 0.51953125,
525
+ "learning_rate": 8.716136631330979e-06,
526
+ "loss": 0.0201,
527
+ "step": 740
528
+ },
529
+ {
530
+ "epoch": 16.0,
531
+ "grad_norm": 0.57421875,
532
+ "learning_rate": 8.8339222614841e-06,
533
+ "loss": 0.02,
534
+ "step": 750
535
+ },
536
+ {
537
+ "epoch": 16.213333333333335,
538
+ "grad_norm": 0.375,
539
+ "learning_rate": 8.951707891637221e-06,
540
+ "loss": 0.02,
541
+ "step": 760
542
+ },
543
+ {
544
+ "epoch": 16.426666666666666,
545
+ "grad_norm": 0.380859375,
546
+ "learning_rate": 9.069493521790342e-06,
547
+ "loss": 0.0208,
548
+ "step": 770
549
+ },
550
+ {
551
+ "epoch": 16.64,
552
+ "grad_norm": 0.67578125,
553
+ "learning_rate": 9.187279151943464e-06,
554
+ "loss": 0.0198,
555
+ "step": 780
556
+ },
557
+ {
558
+ "epoch": 16.85333333333333,
559
+ "grad_norm": 0.484375,
560
+ "learning_rate": 9.305064782096584e-06,
561
+ "loss": 0.0201,
562
+ "step": 790
563
+ },
564
+ {
565
+ "epoch": 17.066666666666666,
566
+ "grad_norm": 0.62890625,
567
+ "learning_rate": 9.422850412249706e-06,
568
+ "loss": 0.0206,
569
+ "step": 800
570
+ },
571
+ {
572
+ "epoch": 17.28,
573
+ "grad_norm": 0.37109375,
574
+ "learning_rate": 9.540636042402828e-06,
575
+ "loss": 0.0211,
576
+ "step": 810
577
+ },
578
+ {
579
+ "epoch": 17.493333333333332,
580
+ "grad_norm": 0.45703125,
581
+ "learning_rate": 9.658421672555949e-06,
582
+ "loss": 0.0205,
583
+ "step": 820
584
+ },
585
+ {
586
+ "epoch": 17.706666666666667,
587
+ "grad_norm": 0.52734375,
588
+ "learning_rate": 9.77620730270907e-06,
589
+ "loss": 0.0198,
590
+ "step": 830
591
+ },
592
+ {
593
+ "epoch": 17.92,
594
+ "grad_norm": 0.451171875,
595
+ "learning_rate": 9.893992932862191e-06,
596
+ "loss": 0.0208,
597
+ "step": 840
598
+ },
599
+ {
600
+ "epoch": 18.133333333333333,
601
+ "grad_norm": 0.43359375,
602
+ "learning_rate": 1.0011778563015313e-05,
603
+ "loss": 0.0205,
604
+ "step": 850
605
+ },
606
+ {
607
+ "epoch": 18.346666666666668,
608
+ "grad_norm": 0.7109375,
609
+ "learning_rate": 1.0129564193168435e-05,
610
+ "loss": 0.0199,
611
+ "step": 860
612
+ },
613
+ {
614
+ "epoch": 18.56,
615
+ "grad_norm": 0.3203125,
616
+ "learning_rate": 1.0247349823321556e-05,
617
+ "loss": 0.0198,
618
+ "step": 870
619
+ },
620
+ {
621
+ "epoch": 18.773333333333333,
622
+ "grad_norm": 0.5390625,
623
+ "learning_rate": 1.0365135453474678e-05,
624
+ "loss": 0.0202,
625
+ "step": 880
626
+ },
627
+ {
628
+ "epoch": 18.986666666666668,
629
+ "grad_norm": 0.54296875,
630
+ "learning_rate": 1.0482921083627797e-05,
631
+ "loss": 0.0203,
632
+ "step": 890
633
+ },
634
+ {
635
+ "epoch": 19.2,
636
+ "grad_norm": 0.4453125,
637
+ "learning_rate": 1.0600706713780919e-05,
638
+ "loss": 0.02,
639
+ "step": 900
640
+ },
641
+ {
642
+ "epoch": 19.413333333333334,
643
+ "grad_norm": 0.94140625,
644
+ "learning_rate": 1.071849234393404e-05,
645
+ "loss": 0.0199,
646
+ "step": 910
647
+ },
648
+ {
649
+ "epoch": 19.626666666666665,
650
+ "grad_norm": 0.5,
651
+ "learning_rate": 1.0836277974087161e-05,
652
+ "loss": 0.0206,
653
+ "step": 920
654
+ },
655
+ {
656
+ "epoch": 19.84,
657
+ "grad_norm": 0.78125,
658
+ "learning_rate": 1.0954063604240283e-05,
659
+ "loss": 0.0201,
660
+ "step": 930
661
+ },
662
+ {
663
+ "epoch": 20.053333333333335,
664
+ "grad_norm": 0.56640625,
665
+ "learning_rate": 1.1071849234393405e-05,
666
+ "loss": 0.0206,
667
+ "step": 940
668
+ },
669
+ {
670
+ "epoch": 20.266666666666666,
671
+ "grad_norm": 0.59765625,
672
+ "learning_rate": 1.1189634864546526e-05,
673
+ "loss": 0.0196,
674
+ "step": 950
675
+ },
676
+ {
677
+ "epoch": 20.48,
678
+ "grad_norm": 0.53515625,
679
+ "learning_rate": 1.1307420494699648e-05,
680
+ "loss": 0.0207,
681
+ "step": 960
682
+ },
683
+ {
684
+ "epoch": 20.693333333333335,
685
+ "grad_norm": 0.71875,
686
+ "learning_rate": 1.1425206124852768e-05,
687
+ "loss": 0.0199,
688
+ "step": 970
689
+ },
690
+ {
691
+ "epoch": 20.906666666666666,
692
+ "grad_norm": 0.59765625,
693
+ "learning_rate": 1.154299175500589e-05,
694
+ "loss": 0.0201,
695
+ "step": 980
696
+ },
697
+ {
698
+ "epoch": 21.12,
699
+ "grad_norm": 0.4609375,
700
+ "learning_rate": 1.1660777385159012e-05,
701
+ "loss": 0.0203,
702
+ "step": 990
703
+ },
704
+ {
705
+ "epoch": 21.333333333333332,
706
+ "grad_norm": 0.69140625,
707
+ "learning_rate": 1.1778563015312133e-05,
708
+ "loss": 0.0204,
709
+ "step": 1000
710
+ }
711
+ ],
712
+ "logging_steps": 10,
713
+ "max_steps": 16974,
714
+ "num_input_tokens_seen": 0,
715
+ "num_train_epochs": 369,
716
+ "save_steps": 500,
717
+ "total_flos": 5.6257256712687206e+17,
718
+ "train_batch_size": 4,
719
+ "trial_name": null,
720
+ "trial_params": null
721
+ }
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1000/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/workspace/models/qwen3-8b",
3
+ "architectures": [
4
+ "Qwen2ForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "eos_token_id": 151643,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 3584,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 18944,
13
+ "max_position_embeddings": 131072,
14
+ "max_window_layers": 28,
15
+ "model_type": "qwen2",
16
+ "num_attention_heads": 28,
17
+ "num_hidden_layers": 28,
18
+ "num_key_value_heads": 4,
19
+ "rms_norm_eps": 1e-06,
20
+ "rope_theta": 1000000.0,
21
+ "sliding_window": 131072,
22
+ "tie_word_embeddings": false,
23
+ "torch_dtype": "bfloat16",
24
+ "transformers_version": "4.40.0",
25
+ "use_cache": false,
26
+ "use_mrope": false,
27
+ "use_sliding_window": false,
28
+ "vocab_size": 152064
29
+ }
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "eos_token_id": 151643,
4
+ "max_new_tokens": 2048,
5
+ "transformers_version": "4.40.0"
6
+ }
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/model-00001-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78595fcb02354ece82b04234a1a09b889499322107ff7b792b42e4eead8dab9e
3
+ size 4877660776
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/model-00002-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a2c7f4e1ad82f21b057802ed2fdbc63013d5e89f8dcea732d339045c3798e21
3
+ size 4932751008
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/model-00003-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0483bd4620ada6ddb194eb8df375ca2eabce8065f09817633a7d5bbe5d887c7
3
+ size 4330865200
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/model-00004-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:623b711a92d36cf9017a90a7a87675f88c895820799ae17f98886f72a13ba5d1
3
+ size 1089994880
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/model.safetensors.index.json ADDED
@@ -0,0 +1,346 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 15231233024
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00004-of-00004.safetensors",
7
+ "model.embed_tokens.weight": "model-00001-of-00004.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
10
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
11
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
12
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
13
+ "model.layers.0.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
14
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
15
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
16
+ "model.layers.0.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
17
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
18
+ "model.layers.0.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
19
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
20
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00004.safetensors",
21
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
22
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
23
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
24
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
25
+ "model.layers.1.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
26
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
27
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
28
+ "model.layers.1.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
29
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
30
+ "model.layers.1.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
31
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
32
+ "model.layers.10.input_layernorm.weight": "model-00002-of-00004.safetensors",
33
+ "model.layers.10.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
34
+ "model.layers.10.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
35
+ "model.layers.10.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
36
+ "model.layers.10.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
37
+ "model.layers.10.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
38
+ "model.layers.10.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
39
+ "model.layers.10.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
40
+ "model.layers.10.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
41
+ "model.layers.10.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
42
+ "model.layers.10.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
43
+ "model.layers.10.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
44
+ "model.layers.11.input_layernorm.weight": "model-00002-of-00004.safetensors",
45
+ "model.layers.11.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
46
+ "model.layers.11.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
47
+ "model.layers.11.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
48
+ "model.layers.11.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
49
+ "model.layers.11.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
50
+ "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
51
+ "model.layers.11.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
52
+ "model.layers.11.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
53
+ "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
54
+ "model.layers.11.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
55
+ "model.layers.11.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
56
+ "model.layers.12.input_layernorm.weight": "model-00002-of-00004.safetensors",
57
+ "model.layers.12.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
58
+ "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
59
+ "model.layers.12.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
60
+ "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
61
+ "model.layers.12.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
62
+ "model.layers.12.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
63
+ "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
64
+ "model.layers.12.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
65
+ "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
66
+ "model.layers.12.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
67
+ "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
68
+ "model.layers.13.input_layernorm.weight": "model-00002-of-00004.safetensors",
69
+ "model.layers.13.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
70
+ "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
71
+ "model.layers.13.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
72
+ "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
73
+ "model.layers.13.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
74
+ "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
75
+ "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
76
+ "model.layers.13.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
77
+ "model.layers.13.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
78
+ "model.layers.13.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
79
+ "model.layers.13.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
80
+ "model.layers.14.input_layernorm.weight": "model-00002-of-00004.safetensors",
81
+ "model.layers.14.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
82
+ "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
83
+ "model.layers.14.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
84
+ "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
85
+ "model.layers.14.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
86
+ "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
87
+ "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
88
+ "model.layers.14.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
89
+ "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
90
+ "model.layers.14.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
91
+ "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
92
+ "model.layers.15.input_layernorm.weight": "model-00002-of-00004.safetensors",
93
+ "model.layers.15.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
94
+ "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
95
+ "model.layers.15.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
96
+ "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
97
+ "model.layers.15.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
98
+ "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
99
+ "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
100
+ "model.layers.15.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
101
+ "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
102
+ "model.layers.15.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
103
+ "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
104
+ "model.layers.16.input_layernorm.weight": "model-00002-of-00004.safetensors",
105
+ "model.layers.16.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
106
+ "model.layers.16.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
107
+ "model.layers.16.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
108
+ "model.layers.16.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
109
+ "model.layers.16.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
110
+ "model.layers.16.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
111
+ "model.layers.16.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
112
+ "model.layers.16.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
113
+ "model.layers.16.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
114
+ "model.layers.16.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
115
+ "model.layers.16.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
116
+ "model.layers.17.input_layernorm.weight": "model-00002-of-00004.safetensors",
117
+ "model.layers.17.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
118
+ "model.layers.17.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
119
+ "model.layers.17.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
120
+ "model.layers.17.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
121
+ "model.layers.17.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
122
+ "model.layers.17.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
123
+ "model.layers.17.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
124
+ "model.layers.17.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
125
+ "model.layers.17.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
126
+ "model.layers.17.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
127
+ "model.layers.17.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
128
+ "model.layers.18.input_layernorm.weight": "model-00003-of-00004.safetensors",
129
+ "model.layers.18.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
130
+ "model.layers.18.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
131
+ "model.layers.18.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
132
+ "model.layers.18.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
133
+ "model.layers.18.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
134
+ "model.layers.18.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
135
+ "model.layers.18.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
136
+ "model.layers.18.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
137
+ "model.layers.18.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
138
+ "model.layers.18.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
139
+ "model.layers.18.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
140
+ "model.layers.19.input_layernorm.weight": "model-00003-of-00004.safetensors",
141
+ "model.layers.19.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
142
+ "model.layers.19.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
143
+ "model.layers.19.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
144
+ "model.layers.19.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
145
+ "model.layers.19.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
146
+ "model.layers.19.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
147
+ "model.layers.19.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
148
+ "model.layers.19.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
149
+ "model.layers.19.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
150
+ "model.layers.19.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
151
+ "model.layers.19.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
152
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00004.safetensors",
153
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
154
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
155
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
156
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
157
+ "model.layers.2.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
158
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
159
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
160
+ "model.layers.2.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
161
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
162
+ "model.layers.2.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
163
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
164
+ "model.layers.20.input_layernorm.weight": "model-00003-of-00004.safetensors",
165
+ "model.layers.20.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
166
+ "model.layers.20.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
167
+ "model.layers.20.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
168
+ "model.layers.20.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
169
+ "model.layers.20.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
170
+ "model.layers.20.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
171
+ "model.layers.20.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
172
+ "model.layers.20.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
173
+ "model.layers.20.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
174
+ "model.layers.20.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
175
+ "model.layers.20.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
176
+ "model.layers.21.input_layernorm.weight": "model-00003-of-00004.safetensors",
177
+ "model.layers.21.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
178
+ "model.layers.21.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
179
+ "model.layers.21.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
180
+ "model.layers.21.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
181
+ "model.layers.21.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
182
+ "model.layers.21.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
183
+ "model.layers.21.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
184
+ "model.layers.21.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
185
+ "model.layers.21.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
186
+ "model.layers.21.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
187
+ "model.layers.21.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
188
+ "model.layers.22.input_layernorm.weight": "model-00003-of-00004.safetensors",
189
+ "model.layers.22.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
190
+ "model.layers.22.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
191
+ "model.layers.22.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
192
+ "model.layers.22.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
193
+ "model.layers.22.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
194
+ "model.layers.22.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
195
+ "model.layers.22.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
196
+ "model.layers.22.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
197
+ "model.layers.22.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
198
+ "model.layers.22.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
199
+ "model.layers.22.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
200
+ "model.layers.23.input_layernorm.weight": "model-00003-of-00004.safetensors",
201
+ "model.layers.23.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
202
+ "model.layers.23.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
203
+ "model.layers.23.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
204
+ "model.layers.23.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
205
+ "model.layers.23.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
206
+ "model.layers.23.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
207
+ "model.layers.23.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
208
+ "model.layers.23.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
209
+ "model.layers.23.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
210
+ "model.layers.23.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
211
+ "model.layers.23.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
212
+ "model.layers.24.input_layernorm.weight": "model-00003-of-00004.safetensors",
213
+ "model.layers.24.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
214
+ "model.layers.24.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
215
+ "model.layers.24.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
216
+ "model.layers.24.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
217
+ "model.layers.24.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
218
+ "model.layers.24.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
219
+ "model.layers.24.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
220
+ "model.layers.24.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
221
+ "model.layers.24.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
222
+ "model.layers.24.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
223
+ "model.layers.24.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
224
+ "model.layers.25.input_layernorm.weight": "model-00003-of-00004.safetensors",
225
+ "model.layers.25.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
226
+ "model.layers.25.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
227
+ "model.layers.25.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
228
+ "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
229
+ "model.layers.25.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
230
+ "model.layers.25.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
231
+ "model.layers.25.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
232
+ "model.layers.25.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
233
+ "model.layers.25.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
234
+ "model.layers.25.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
235
+ "model.layers.25.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
236
+ "model.layers.26.input_layernorm.weight": "model-00003-of-00004.safetensors",
237
+ "model.layers.26.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
238
+ "model.layers.26.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
239
+ "model.layers.26.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
240
+ "model.layers.26.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
241
+ "model.layers.26.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
242
+ "model.layers.26.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
243
+ "model.layers.26.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
244
+ "model.layers.26.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
245
+ "model.layers.26.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
246
+ "model.layers.26.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
247
+ "model.layers.26.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
248
+ "model.layers.27.input_layernorm.weight": "model-00003-of-00004.safetensors",
249
+ "model.layers.27.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
250
+ "model.layers.27.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
251
+ "model.layers.27.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
252
+ "model.layers.27.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
253
+ "model.layers.27.self_attn.k_proj.bias": "model-00003-of-00004.safetensors",
254
+ "model.layers.27.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
255
+ "model.layers.27.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
256
+ "model.layers.27.self_attn.q_proj.bias": "model-00003-of-00004.safetensors",
257
+ "model.layers.27.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
258
+ "model.layers.27.self_attn.v_proj.bias": "model-00003-of-00004.safetensors",
259
+ "model.layers.27.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
260
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00004.safetensors",
261
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
262
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
263
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
264
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
265
+ "model.layers.3.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
266
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
267
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
268
+ "model.layers.3.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
269
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
270
+ "model.layers.3.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
271
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
272
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00004.safetensors",
273
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
274
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
275
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
276
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
277
+ "model.layers.4.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
278
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
279
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
280
+ "model.layers.4.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
281
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
282
+ "model.layers.4.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
283
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
284
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00004.safetensors",
285
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
286
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
287
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
288
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
289
+ "model.layers.5.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
290
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
291
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
292
+ "model.layers.5.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
293
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
294
+ "model.layers.5.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
295
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
296
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00004.safetensors",
297
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
298
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
299
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
300
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
301
+ "model.layers.6.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
302
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
303
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
304
+ "model.layers.6.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
305
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
306
+ "model.layers.6.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
307
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
308
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00004.safetensors",
309
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
310
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
311
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
312
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
313
+ "model.layers.7.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
314
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
315
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
316
+ "model.layers.7.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
317
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
318
+ "model.layers.7.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
319
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
320
+ "model.layers.8.input_layernorm.weight": "model-00002-of-00004.safetensors",
321
+ "model.layers.8.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
322
+ "model.layers.8.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
323
+ "model.layers.8.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
324
+ "model.layers.8.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
325
+ "model.layers.8.self_attn.k_proj.bias": "model-00001-of-00004.safetensors",
326
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
327
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
328
+ "model.layers.8.self_attn.q_proj.bias": "model-00001-of-00004.safetensors",
329
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
330
+ "model.layers.8.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
331
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
332
+ "model.layers.9.input_layernorm.weight": "model-00002-of-00004.safetensors",
333
+ "model.layers.9.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
334
+ "model.layers.9.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
335
+ "model.layers.9.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
336
+ "model.layers.9.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
337
+ "model.layers.9.self_attn.k_proj.bias": "model-00002-of-00004.safetensors",
338
+ "model.layers.9.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
339
+ "model.layers.9.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
340
+ "model.layers.9.self_attn.q_proj.bias": "model-00002-of-00004.safetensors",
341
+ "model.layers.9.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
342
+ "model.layers.9.self_attn.v_proj.bias": "model-00002-of-00004.safetensors",
343
+ "model.layers.9.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
344
+ "model.norm.weight": "model-00003-of-00004.safetensors"
345
+ }
346
+ }
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/checkpoint-1500/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a5775162a628571050b537846c48d16bdfaecbc97e6c43cd3fe53cbf8ddd27c
3
+ size 20272906496
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/model-00001-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e9e1f4ecd9cf369559f2cd494d8f5649f7e90d20f44ea71d8f0e6a325001528
3
+ size 4877660776
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/model-00002-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b34d179c53fb1e609ce1ccc2b667e5df47ce8b3fd32cdcfca8d78761a10210f2
3
+ size 4932751008
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/model-00003-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20c567bfad427e3b07b93fa3330f1e2bed093a03c67a7607297b7053a058d629
3
+ size 4330865200
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/model-00004-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e00171701c7bcf6f10f6a9c69f54380fd9c076ec43d9815c5673ca63766f8187
3
+ size 1089994880
platform/aiml/checkpoints/qwen3-8b-elizabeth-sft/optimizer_backups.tar.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:322092b747b3f363735641babcd9fd1d1f1a3f47f4073202be56cb2a4614063e
3
+ size 1518600192
platform/aiml/models/vllm/torch_compile_cache/069ce3ea2b/rank_0_0/triton_cache/JGBJTSHFQO5ZMSSDR5JQ4N6IDDC3AKTV33H2WXFG26XWQP3IZ4ZA/triton_poi_fused_cat_1.ptx ADDED
@@ -0,0 +1,528 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ //
2
+ // Generated by LLVM NVPTX Back-End
3
+ //
4
+
5
+ .version 8.4
6
+ .target sm_90a
7
+ .address_size 64
8
+
9
+ // .globl triton_poi_fused_cat_1 // -- Begin function triton_poi_fused_cat_1
10
+ .extern .func __assertfail
11
+ (
12
+ .param .b64 __assertfail_param_0,
13
+ .param .b64 __assertfail_param_1,
14
+ .param .b32 __assertfail_param_2,
15
+ .param .b64 __assertfail_param_3,
16
+ .param .b64 __assertfail_param_4
17
+ )
18
+ .noreturn;
19
+ .global .align 1 .b8 assertFunc_1[8] = {117, 110, 107, 110, 111, 119, 110};
20
+ .global .align 1 .b8 assertFile_1[134] = {47, 104, 111, 109, 101, 47, 120, 47, 46, 99, 97, 99, 104, 101, 47, 118, 108, 108, 109, 47, 116, 111, 114, 99, 104, 95, 99, 111, 109, 112, 105, 108, 101, 95, 99, 97, 99, 104, 101, 47, 48, 54, 57, 99, 101, 51, 101, 97, 50, 98, 47, 114, 97, 110, 107, 95, 48, 95, 48, 47, 105, 110, 100, 117, 99, 116, 111, 114, 95, 99, 97, 99, 104, 101, 47, 122, 109, 47, 99, 122, 109, 119, 113, 108, 110, 51, 112, 109, 100, 53, 105, 116, 102, 53, 110, 98, 118, 110, 106, 117, 114, 102, 122, 115, 120, 103, 50, 52, 120, 51, 114, 120, 117, 103, 97, 103, 107, 116, 114, 107, 54, 109, 101, 102, 119, 114, 101, 97, 97, 104, 46, 112, 121};
21
+ .global .align 1 .b8 assertMessage_1[68] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 108, 46, 98, 114, 111, 97, 100, 99, 97, 115, 116, 95, 116, 111, 40, 116, 109, 112, 50, 56, 44, 32, 91, 88, 66, 76, 79, 67, 75, 93, 41, 32, 60, 32, 49, 51, 49, 48, 55, 50};
22
+ .global .align 1 .b8 assertFunc_0[8] = {117, 110, 107, 110, 111, 119, 110};
23
+ .global .align 1 .b8 assertFile_0[134] = {47, 104, 111, 109, 101, 47, 120, 47, 46, 99, 97, 99, 104, 101, 47, 118, 108, 108, 109, 47, 116, 111, 114, 99, 104, 95, 99, 111, 109, 112, 105, 108, 101, 95, 99, 97, 99, 104, 101, 47, 48, 54, 57, 99, 101, 51, 101, 97, 50, 98, 47, 114, 97, 110, 107, 95, 48, 95, 48, 47, 105, 110, 100, 117, 99, 116, 111, 114, 95, 99, 97, 99, 104, 101, 47, 122, 109, 47, 99, 122, 109, 119, 113, 108, 110, 51, 112, 109, 100, 53, 105, 116, 102, 53, 110, 98, 118, 110, 106, 117, 114, 102, 122, 115, 120, 103, 50, 52, 120, 51, 114, 120, 117, 103, 97, 103, 107, 116, 114, 107, 54, 109, 101, 102, 119, 114, 101, 97, 97, 104, 46, 112, 121};
24
+ .global .align 1 .b8 assertMessage_0[68] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 108, 46, 98, 114, 111, 97, 100, 99, 97, 115, 116, 95, 116, 111, 40, 116, 109, 112, 49, 48, 44, 32, 91, 88, 66, 76, 79, 67, 75, 93, 41, 32, 60, 32, 49, 51, 49, 48, 55, 50};
25
+ // @triton_poi_fused_cat_1
26
+ .visible .entry triton_poi_fused_cat_1(
27
+ .param .u64 .ptr .global .align 1 triton_poi_fused_cat_1_param_0,
28
+ .param .u64 .ptr .global .align 1 triton_poi_fused_cat_1_param_1,
29
+ .param .u64 .ptr .global .align 1 triton_poi_fused_cat_1_param_2,
30
+ .param .u64 .ptr .global .align 1 triton_poi_fused_cat_1_param_3,
31
+ .param .u32 triton_poi_fused_cat_1_param_4,
32
+ .param .u64 .ptr .global .align 1 triton_poi_fused_cat_1_param_5
33
+ )
34
+ .reqntid 256, 1, 1
35
+ {
36
+ .reg .pred %p<27>;
37
+ .reg .b16 %rs<17>;
38
+ .reg .b32 %r<58>;
39
+ .reg .f32 %f<29>;
40
+ .reg .b64 %rd<54>;
41
+ .loc 1 18 0 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:18:0
42
+ $L__func_begin0:
43
+ .loc 1 18 0 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:18:0
44
+
45
+ // %bb.0:
46
+ ld.param.u32 %r12, [triton_poi_fused_cat_1_param_4];
47
+ ld.param.u64 %rd5, [triton_poi_fused_cat_1_param_0];
48
+ ld.param.u64 %rd13, [triton_poi_fused_cat_1_param_1];
49
+ $L__tmp0:
50
+ .loc 1 19 28 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:19:28
51
+ mov.u32 %r15, %ctaid.x;
52
+ .loc 1 19 33 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:19:33
53
+ shl.b32 %r16, %r15, 9;
54
+ .loc 1 20 36 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:20:36
55
+ mov.u32 %r17, %tid.x;
56
+ shl.b32 %r18, %r17, 1;
57
+ and.b32 %r19, %r18, 510;
58
+ .loc 1 20 23 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:20:23
59
+ or.b32 %r1, %r19, %r16;
60
+ .loc 1 21 21 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:21:21
61
+ setp.lt.s32 %p6, %r1, %r12;
62
+ .loc 1 23 21 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:23:21
63
+ bfe.s32 %r20, %r15, 22, 1;
64
+ shr.u32 %r21, %r20, 25;
65
+ add.s32 %r22, %r1, %r21;
66
+ shr.s32 %r23, %r22, 7;
67
+ .loc 1 22 19 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:22:19
68
+ and.b32 %r24, %r22, -128;
69
+ sub.s32 %r2, %r1, %r24;
70
+ .loc 1 23 28 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:23:28
71
+ mul.hi.s32 %r25, %r23, -1840700269;
72
+ add.s32 %r26, %r25, %r23;
73
+ shr.u32 %r27, %r26, 31;
74
+ shr.u32 %r28, %r26, 4;
75
+ add.s32 %r29, %r28, %r27;
76
+ mul.lo.s32 %r30, %r29, 28;
77
+ sub.s32 %r31, %r23, %r30;
78
+ .loc 1 24 19 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:24:19
79
+ mul.hi.s32 %r32, %r1, -1840700269;
80
+ add.s32 %r33, %r32, %r1;
81
+ shr.u32 %r34, %r33, 31;
82
+ shr.s32 %r35, %r33, 11;
83
+ add.s32 %r36, %r35, %r34;
84
+ .loc 1 30 18 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:30:18
85
+ setp.lt.s32 %p7, %r2, 64;
86
+ .loc 1 31 34 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:31:34
87
+ shl.b32 %r3, %r31, 7;
88
+ .loc 1 31 44 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:31:44
89
+ mul.lo.s32 %r4, %r36, 4608;
90
+ .loc 1 31 39 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:31:39
91
+ add.s32 %r5, %r3, %r4;
92
+ .loc 1 31 50 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:31:50
93
+ add.s32 %r37, %r5, %r2;
94
+ .loc 1 31 30 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:31:30
95
+ mul.wide.s32 %rd14, %r37, 2;
96
+ add.s64 %rd8, %rd5, %rd14;
97
+ .loc 1 31 64 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:31:64
98
+ and.pred %p5, %p7, %p6;
99
+ not.pred %p8, %p5;
100
+ mov.b32 %r14, 0;
101
+ .loc 1 31 56 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:31:56
102
+ // begin inline asm
103
+ mov.u32 %r13, %r14;
104
+ @%p5 ld.global.L1::evict_last.b32 { %r13 }, [ %rd8 + 0 ];
105
+ // end inline asm
106
+ .loc 1 32 30 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:32:30
107
+ mul.wide.s32 %rd15, %r36, 8;
108
+ add.s64 %rd12, %rd13, %rd15;
109
+ .loc 1 32 35 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:32:35
110
+ // begin inline asm
111
+ mov.u64 %rd9, 0x0;
112
+ @%p5 ld.global.L1::evict_last.b64 { %rd9 }, [ %rd12 + 0 ];
113
+ // end inline asm
114
+ // begin inline asm
115
+ mov.u64 %rd11, 0x0;
116
+ @%p5 ld.global.L1::evict_last.b64 { %rd11 }, [ %rd12 + 0 ];
117
+ // end inline asm
118
+ .loc 1 36 33 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:36:33
119
+ shr.u64 %rd16, %rd9, 46;
120
+ and.b64 %rd17, %rd16, 131072;
121
+ add.s64 %rd2, %rd17, %rd9;
122
+ .loc 1 37 65 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:37:65
123
+ setp.lt.u64 %p9, %rd2, 131072;
124
+ .loc 1 37 128 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:37:128
125
+ or.pred %p10, %p8, %p9;
126
+ @%p10 bra $L__BB0_2;
127
+ bra.uni $L__BB0_1;
128
+ $L__BB0_2:
129
+ .loc 1 0 128 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:0:128
130
+ ld.param.u64 %rd6, [triton_poi_fused_cat_1_param_2];
131
+ .loc 1 37 128 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:37:128
132
+ bar.sync 0;
133
+ .loc 1 38 31 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:38:31
134
+ shl.b64 %rd26, %rd2, 8;
135
+ add.s64 %rd27, %rd6, %rd26;
136
+ mul.wide.s32 %rd28, %r2, 2;
137
+ add.s64 %rd18, %rd27, %rd28;
138
+ .loc 1 38 50 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:38:50
139
+ // begin inline asm
140
+ mov.u32 %r38, %r14;
141
+ @%p5 ld.global.L1::evict_last.b32 { %r38 }, [ %rd18 + 0 ];
142
+ // end inline asm
143
+ .loc 1 40 45 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:40:45
144
+ add.s32 %r46, %r4, %r3;
145
+ add.s32 %r47, %r46, 64;
146
+ .loc 1 40 56 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:40:56
147
+ add.s32 %r48, %r47, %r2;
148
+ .loc 1 40 31 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:40:31
149
+ mul.wide.s32 %rd29, %r48, 2;
150
+ add.s64 %rd19, %rd5, %rd29;
151
+ .loc 1 40 62 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:40:62
152
+ // begin inline asm
153
+ mov.u32 %r40, %r14;
154
+ @%p5 ld.global.L1::evict_last.b32 { %r40 }, [ %rd19 + 0 ];
155
+ // end inline asm
156
+ .loc 1 41 31 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:41:31
157
+ add.s64 %rd20, %rd18, 128;
158
+ .loc 1 41 55 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:41:55
159
+ // begin inline asm
160
+ mov.u32 %r42, %r14;
161
+ @%p5 ld.global.L1::evict_last.b32 { %r42 }, [ %rd20 + 0 ];
162
+ // end inline asm
163
+ .loc 1 46 20 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:46:20
164
+ setp.gt.s32 %p18, %r2, 63;
165
+ .loc 1 49 64 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:49:64
166
+ add.s32 %r10, %r2, -64;
167
+ .loc 1 49 56 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:49:56
168
+ add.s32 %r49, %r47, %r10;
169
+ .loc 1 49 31 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:49:31
170
+ mul.wide.s32 %rd30, %r49, 2;
171
+ add.s64 %rd21, %rd5, %rd30;
172
+ .loc 1 49 78 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:49:78
173
+ and.pred %p16, %p18, %p6;
174
+ not.pred %p19, %p16;
175
+ .loc 1 49 70 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:49:70
176
+ // begin inline asm
177
+ mov.u32 %r44, %r14;
178
+ @%p16 ld.global.L1::evict_last.b32 { %r44 }, [ %rd21 + 0 ];
179
+ // end inline asm
180
+ .loc 1 50 36 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:50:36
181
+ // begin inline asm
182
+ mov.u64 %rd22, 0x0;
183
+ @%p16 ld.global.L1::evict_last.b64 { %rd22 }, [ %rd12 + 0 ];
184
+ // end inline asm
185
+ // begin inline asm
186
+ mov.u64 %rd24, 0x0;
187
+ @%p16 ld.global.L1::evict_last.b64 { %rd24 }, [ %rd12 + 0 ];
188
+ // end inline asm
189
+ .loc 1 54 35 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:54:35
190
+ shr.u64 %rd31, %rd22, 46;
191
+ and.b64 %rd32, %rd31, 131072;
192
+ add.s64 %rd4, %rd32, %rd22;
193
+ .loc 1 55 65 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:55:65
194
+ setp.lt.u64 %p20, %rd4, 131072;
195
+ .loc 1 55 129 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:55:129
196
+ or.pred %p21, %p19, %p20;
197
+ @%p21 bra $L__BB0_4;
198
+ bra.uni $L__BB0_3;
199
+ $L__BB0_4:
200
+ .loc 1 0 129 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:0:129
201
+ ld.param.u64 %rd7, [triton_poi_fused_cat_1_param_3];
202
+ cvt.s64.s32 %rd3, %r2;
203
+ .loc 1 55 129 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:55:129
204
+ bar.sync 0;
205
+ .loc 1 56 31 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:56:31
206
+ shl.b64 %rd37, %rd4, 8;
207
+ add.s64 %rd38, %rd6, %rd37;
208
+ shl.b64 %rd39, %rd3, 1;
209
+ add.s64 %rd35, %rd38, %rd39;
210
+ add.s64 %rd33, %rd35, -128;
211
+ mov.b32 %r51, 0;
212
+ .loc 1 56 58 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:56:58
213
+ // begin inline asm
214
+ mov.u32 %r50, %r51;
215
+ @%p16 ld.global.L1::evict_last.b32 { %r50 }, [ %rd33 + 0 ];
216
+ // end inline asm
217
+ .loc 1 58 51 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:58:51
218
+ add.s32 %r57, %r5, %r10;
219
+ .loc 1 58 31 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:58:31
220
+ mul.wide.s32 %rd40, %r57, 2;
221
+ add.s64 %rd34, %rd5, %rd40;
222
+ .loc 1 58 65 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:58:65
223
+ // begin inline asm
224
+ mov.u32 %r52, %r51;
225
+ @%p16 ld.global.L1::evict_last.b32 { %r52 }, [ %rd34 + 0 ];
226
+ // end inline asm
227
+ .loc 1 59 63 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:59:63
228
+ // begin inline asm
229
+ mov.u32 %r54, %r51;
230
+ @%p16 ld.global.L1::evict_last.b32 { %r54 }, [ %rd35 + 0 ];
231
+ // end inline asm
232
+ .loc 1 65 25 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:65:25
233
+ mul.wide.s32 %rd41, %r1, 2;
234
+ add.s64 %rd36, %rd7, %rd41;
235
+ .loc 1 49 129 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:49:129
236
+ mov.b32 {%rs1, %rs2}, %r44;
237
+ cvt.f32.bf16 %f1, %rs1;
238
+ cvt.f32.bf16 %f2, %rs2;
239
+ .loc 1 31 114 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:31:114
240
+ mov.b32 {%rs3, %rs4}, %r13;
241
+ cvt.f32.bf16 %f3, %rs3;
242
+ cvt.f32.bf16 %f4, %rs4;
243
+ .loc 1 38 108 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:38:108
244
+ mov.b32 {%rs5, %rs6}, %r38;
245
+ cvt.f32.bf16 %f5, %rs5;
246
+ cvt.f32.bf16 %f6, %rs6;
247
+ .loc 1 40 120 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:40:120
248
+ mov.b32 {%rs7, %rs8}, %r40;
249
+ cvt.f32.bf16 %f7, %rs8;
250
+ cvt.f32.bf16 %f8, %rs7;
251
+ .loc 1 41 113 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:41:113
252
+ mov.b32 {%rs9, %rs10}, %r42;
253
+ cvt.f32.bf16 %f9, %rs10;
254
+ cvt.f32.bf16 %f10, %rs9;
255
+ .loc 1 42 20 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:42:20
256
+ mul.f32 %f11, %f8, %f10;
257
+ mul.f32 %f12, %f7, %f9;
258
+ .loc 1 43 20 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:43:20
259
+ neg.f32 %f13, %f12;
260
+ fma.rn.f32 %f14, %f4, %f6, %f13;
261
+ neg.f32 %f15, %f11;
262
+ fma.rn.f32 %f16, %f3, %f5, %f15;
263
+ .loc 1 56 117 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:56:117
264
+ mov.b32 {%rs11, %rs12}, %r50;
265
+ cvt.f32.bf16 %f17, %rs11;
266
+ cvt.f32.bf16 %f18, %rs12;
267
+ .loc 1 58 124 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:58:124
268
+ mov.b32 {%rs13, %rs14}, %r52;
269
+ cvt.f32.bf16 %f19, %rs14;
270
+ cvt.f32.bf16 %f20, %rs13;
271
+ .loc 1 59 122 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:59:122
272
+ mov.b32 {%rs15, %rs16}, %r54;
273
+ cvt.f32.bf16 %f21, %rs16;
274
+ cvt.f32.bf16 %f22, %rs15;
275
+ .loc 1 60 20 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:60:20
276
+ mul.f32 %f23, %f20, %f22;
277
+ mul.f32 %f24, %f19, %f21;
278
+ .loc 1 61 20 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:61:20
279
+ fma.rn.f32 %f25, %f2, %f18, %f24;
280
+ fma.rn.f32 %f26, %f1, %f17, %f23;
281
+ .loc 1 0 0 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:0:0
282
+ selp.f32 %f27, %f16, %f26, %p7;
283
+ selp.f32 %f28, %f14, %f25, %p7;
284
+ .loc 1 65 37 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:65:37
285
+ cvt.rn.bf16x2.f32 %r56, %f28, %f27;
286
+ // begin inline asm
287
+ @%p6 st.global.b32 [ %rd36 + 0 ], { %r56 };
288
+ // end inline asm
289
+ .loc 1 65 4 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:65:4
290
+ ret;
291
+ $L__BB0_1:
292
+ .loc 1 37 128 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:37:128
293
+ mov.u64 %rd48, assertMessage_0;
294
+ cvta.global.u64 %rd49, %rd48;
295
+ mov.u64 %rd50, assertFile_0;
296
+ cvta.global.u64 %rd51, %rd50;
297
+ mov.u64 %rd52, assertFunc_0;
298
+ cvta.global.u64 %rd53, %rd52;
299
+ { // callseq 4, 0
300
+ .param .b64 param0;
301
+ st.param.b64 [param0], %rd49;
302
+ .param .b64 param1;
303
+ st.param.b64 [param1], %rd51;
304
+ .param .b32 param2;
305
+ st.param.b32 [param2], 37;
306
+ .param .b64 param3;
307
+ st.param.b64 [param3], %rd53;
308
+ .param .b64 param4;
309
+ st.param.b64 [param4], 1;
310
+ call.uni
311
+ __assertfail,
312
+ (
313
+ param0,
314
+ param1,
315
+ param2,
316
+ param3,
317
+ param4
318
+ );
319
+ } // callseq 4
320
+ trap;
321
+ $L__BB0_3:
322
+ .loc 1 55 129 // czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py:55:129
323
+ mov.u64 %rd42, assertMessage_1;
324
+ cvta.global.u64 %rd43, %rd42;
325
+ mov.u64 %rd44, assertFile_1;
326
+ cvta.global.u64 %rd45, %rd44;
327
+ mov.u64 %rd46, assertFunc_1;
328
+ cvta.global.u64 %rd47, %rd46;
329
+ { // callseq 3, 0
330
+ .param .b64 param0;
331
+ st.param.b64 [param0], %rd43;
332
+ .param .b64 param1;
333
+ st.param.b64 [param1], %rd45;
334
+ .param .b32 param2;
335
+ st.param.b32 [param2], 55;
336
+ .param .b64 param3;
337
+ st.param.b64 [param3], %rd47;
338
+ .param .b64 param4;
339
+ st.param.b64 [param4], 1;
340
+ call.uni
341
+ __assertfail,
342
+ (
343
+ param0,
344
+ param1,
345
+ param2,
346
+ param3,
347
+ param4
348
+ );
349
+ } // callseq 3
350
+ trap;
351
+ $L__tmp1:
352
+ $L__func_end0:
353
+ // -- End function
354
+ }
355
+ .file 1 "/home/x/.cache/vllm/torch_compile_cache/069ce3ea2b/rank_0_0/inductor_cache/zm/czmwqln3pmd5itf5nbvnjurfzsxg24x3rxugagktrk6mefwreaah.py"
356
+ .section .debug_abbrev
357
+ {
358
+ .b8 1 // Abbreviation Code
359
+ .b8 17 // DW_TAG_compile_unit
360
+ .b8 0 // DW_CHILDREN_no
361
+ .b8 37 // DW_AT_producer
362
+ .b8 8 // DW_FORM_string
363
+ .b8 19 // DW_AT_language
364
+ .b8 5 // DW_FORM_data2
365
+ .b8 3 // DW_AT_name
366
+ .b8 8 // DW_FORM_string
367
+ .b8 16 // DW_AT_stmt_list
368
+ .b8 6 // DW_FORM_data4
369
+ .b8 27 // DW_AT_comp_dir
370
+ .b8 8 // DW_FORM_string
371
+ .b8 0 // EOM(1)
372
+ .b8 0 // EOM(2)
373
+ .b8 0 // EOM(3)
374
+ }
375
+ .section .debug_info
376
+ {
377
+ .b32 155 // Length of Unit
378
+ .b8 2 // DWARF version number
379
+ .b8 0
380
+ .b32 .debug_abbrev // Offset Into Abbrev. Section
381
+ .b8 8 // Address Size (in bytes)
382
+ .b8 1 // Abbrev [1] 0xb:0x94 DW_TAG_compile_unit
383
+ .b8 116 // DW_AT_producer
384
+ .b8 114
385
+ .b8 105
386
+ .b8 116
387
+ .b8 111
388
+ .b8 110
389
+ .b8 0
390
+ .b8 2 // DW_AT_language
391
+ .b8 0
392
+ .b8 99 // DW_AT_name
393
+ .b8 122
394
+ .b8 109
395
+ .b8 119
396
+ .b8 113
397
+ .b8 108
398
+ .b8 110
399
+ .b8 51
400
+ .b8 112
401
+ .b8 109
402
+ .b8 100
403
+ .b8 53
404
+ .b8 105
405
+ .b8 116
406
+ .b8 102
407
+ .b8 53
408
+ .b8 110
409
+ .b8 98
410
+ .b8 118
411
+ .b8 110
412
+ .b8 106
413
+ .b8 117
414
+ .b8 114
415
+ .b8 102
416
+ .b8 122
417
+ .b8 115
418
+ .b8 120
419
+ .b8 103
420
+ .b8 50
421
+ .b8 52
422
+ .b8 120
423
+ .b8 51
424
+ .b8 114
425
+ .b8 120
426
+ .b8 117
427
+ .b8 103
428
+ .b8 97
429
+ .b8 103
430
+ .b8 107
431
+ .b8 116
432
+ .b8 114
433
+ .b8 107
434
+ .b8 54
435
+ .b8 109
436
+ .b8 101
437
+ .b8 102
438
+ .b8 119
439
+ .b8 114
440
+ .b8 101
441
+ .b8 97
442
+ .b8 97
443
+ .b8 104
444
+ .b8 46
445
+ .b8 112
446
+ .b8 121
447
+ .b8 0
448
+ .b32 .debug_line // DW_AT_stmt_list
449
+ .b8 47 // DW_AT_comp_dir
450
+ .b8 104
451
+ .b8 111
452
+ .b8 109
453
+ .b8 101
454
+ .b8 47
455
+ .b8 120
456
+ .b8 47
457
+ .b8 46
458
+ .b8 99
459
+ .b8 97
460
+ .b8 99
461
+ .b8 104
462
+ .b8 101
463
+ .b8 47
464
+ .b8 118
465
+ .b8 108
466
+ .b8 108
467
+ .b8 109
468
+ .b8 47
469
+ .b8 116
470
+ .b8 111
471
+ .b8 114
472
+ .b8 99
473
+ .b8 104
474
+ .b8 95
475
+ .b8 99
476
+ .b8 111
477
+ .b8 109
478
+ .b8 112
479
+ .b8 105
480
+ .b8 108
481
+ .b8 101
482
+ .b8 95
483
+ .b8 99
484
+ .b8 97
485
+ .b8 99
486
+ .b8 104
487
+ .b8 101
488
+ .b8 47
489
+ .b8 48
490
+ .b8 54
491
+ .b8 57
492
+ .b8 99
493
+ .b8 101
494
+ .b8 51
495
+ .b8 101
496
+ .b8 97
497
+ .b8 50
498
+ .b8 98
499
+ .b8 47
500
+ .b8 114
501
+ .b8 97
502
+ .b8 110
503
+ .b8 107
504
+ .b8 95
505
+ .b8 48
506
+ .b8 95
507
+ .b8 48
508
+ .b8 47
509
+ .b8 105
510
+ .b8 110
511
+ .b8 100
512
+ .b8 117
513
+ .b8 99
514
+ .b8 116
515
+ .b8 111
516
+ .b8 114
517
+ .b8 95
518
+ .b8 99
519
+ .b8 97
520
+ .b8 99
521
+ .b8 104
522
+ .b8 101
523
+ .b8 47
524
+ .b8 122
525
+ .b8 109
526
+ .b8 0
527
+ }
528
+ .section .debug_macinfo { }
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/__grp__triton_red_fused__to_copy_mean_pow_4.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"child_paths": {"triton_red_fused__to_copy_mean_pow_4.ttir": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.ttir", "triton_red_fused__to_copy_mean_pow_4.ttgir": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.ttgir", "triton_red_fused__to_copy_mean_pow_4.llir": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.llir", "triton_red_fused__to_copy_mean_pow_4.ptx": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.ptx", "triton_red_fused__to_copy_mean_pow_4.cubin": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.cubin", "triton_red_fused__to_copy_mean_pow_4.json": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.json"}}
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.cubin ADDED
Binary file (11.3 kB). View file
 
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"hash": "dde87391e8e537d55078ec911de8aa99b0ae8675c8e0e555a1cd93c960c9a280", "target": {"backend": "cuda", "arch": 90, "warp_size": 32}, "num_warps": 16, "num_ctas": 1, "num_stages": 1, "num_buffers_warp_spec": 0, "num_consumer_groups": 0, "reg_dec_producer": 0, "reg_inc_consumer": 0, "maxnreg": null, "cluster_dims": [1, 1, 1], "ptx_version": null, "enable_fp_fusion": true, "launch_cooperative_grid": false, "supported_fp8_dtypes": ["fp8e4b15", "fp8e4nv", "fp8e5"], "deprecated_fp8_dtypes": ["fp8e4b15"], "default_dot_input_precision": "tf32", "allowed_dot_input_precisions": ["tf32", "tf32x3", "ieee"], "max_num_imprecise_acc_default": 1073741824, "extern_libs": [["libdevice", "/home/x/hfenv/lib/python3.12/site-packages/triton/backends/nvidia/lib/libdevice.10.bc"]], "debug": true, "backend_name": "cuda", "sanitize_overflow": false, "arch": "sm90", "triton_version": "3.3.1", "shared": 256, "tmem_size": 0, "global_scratch_size": 0, "global_scratch_align": 1, "name": "triton_red_fused__to_copy_mean_pow_4"}
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.llir ADDED
@@ -0,0 +1,169 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ; ModuleID = 'LLVMDialectModule'
2
+ source_filename = "LLVMDialectModule"
3
+ target datalayout = "e-p3:32:32-p4:32:32-p5:32:32-p6:32:32-i64:64-i128:128-v16:16-v32:32-n16:32:64"
4
+
5
+ @global_smem = external addrspace(3) global [0 x i8], align 16
6
+
7
+ define ptx_kernel void @triton_red_fused__to_copy_mean_pow_4(ptr addrspace(1) %0, ptr addrspace(1) %1, i32 %2, i32 %3, ptr addrspace(1) readnone captures(none) %4) local_unnamed_addr !dbg !6 {
8
+ %6 = tail call i32 @llvm.nvvm.read.ptx.sreg.ctaid.x(), !dbg !9
9
+ %7 = shl i32 %6, 6, !dbg !10
10
+ %8 = tail call i32 @llvm.nvvm.read.ptx.sreg.tid.x(), !dbg !11
11
+ %9 = lshr i32 %8, 3, !dbg !11
12
+ %10 = and i32 %9, 63, !dbg !11
13
+ %11 = or disjoint i32 %10, %7, !dbg !12
14
+ %12 = icmp slt i32 %11, %2, !dbg !13
15
+ %13 = shl i32 %8, 3, !dbg !14
16
+ %14 = and i32 %13, 56, !dbg !14
17
+ %15 = sdiv i32 %11, 8, !dbg !15
18
+ %16 = mul i32 %15, 8, !dbg !16
19
+ %.decomposed = sub i32 %11, %16, !dbg !16
20
+ %17 = shl nsw i32 %.decomposed, 7, !dbg !17
21
+ %18 = mul i32 %15, 7168, !dbg !18
22
+ %19 = add nsw i32 %17, 5120
23
+ %20 = add i32 %19, %18
24
+ %21 = or disjoint i32 %20, %14, !dbg !19
25
+ %22 = sext i32 %21 to i64, !dbg !20
26
+ %23 = getelementptr bfloat, ptr addrspace(1) %0, i64 %22, !dbg !20
27
+ %24 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, $4;\0A\09mov.u32 $1, $5;\0A\09mov.u32 $2, $6;\0A\09mov.u32 $3, $7;\0A\09@$9 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $8 + 0 ];", "=r,=r,=r,=r,r,r,r,r,l,b"(i32 0, i32 0, i32 0, i32 0, ptr addrspace(1) %23, i1 %12) #3, !dbg !21
28
+ %25 = extractvalue { i32, i32, i32, i32 } %24, 0, !dbg !21
29
+ %26 = bitcast i32 %25 to <2 x bfloat>, !dbg !21
30
+ %27 = extractvalue { i32, i32, i32, i32 } %24, 1, !dbg !21
31
+ %28 = bitcast i32 %27 to <2 x bfloat>, !dbg !21
32
+ %29 = extractvalue { i32, i32, i32, i32 } %24, 2, !dbg !21
33
+ %30 = bitcast i32 %29 to <2 x bfloat>, !dbg !21
34
+ %31 = extractvalue { i32, i32, i32, i32 } %24, 3, !dbg !21
35
+ %32 = bitcast i32 %31 to <2 x bfloat>, !dbg !21
36
+ %33 = or disjoint i32 %14, 64, !dbg !22
37
+ %34 = or disjoint i32 %20, %33, !dbg !19
38
+ %35 = sext i32 %34 to i64, !dbg !20
39
+ %36 = getelementptr bfloat, ptr addrspace(1) %0, i64 %35, !dbg !20
40
+ %37 = tail call { i32, i32, i32, i32 } asm sideeffect "mov.u32 $0, $4;\0A\09mov.u32 $1, $5;\0A\09mov.u32 $2, $6;\0A\09mov.u32 $3, $7;\0A\09@$9 ld.global.L1::evict_first.v4.b32 { $0, $1, $2, $3 }, [ $8 + 0 ];", "=r,=r,=r,=r,r,r,r,r,l,b"(i32 0, i32 0, i32 0, i32 0, ptr addrspace(1) %36, i1 %12) #3, !dbg !21
41
+ %38 = extractvalue { i32, i32, i32, i32 } %37, 0, !dbg !21
42
+ %39 = bitcast i32 %38 to <2 x bfloat>, !dbg !21
43
+ %40 = extractvalue { i32, i32, i32, i32 } %37, 1, !dbg !21
44
+ %41 = bitcast i32 %40 to <2 x bfloat>, !dbg !21
45
+ %42 = extractvalue { i32, i32, i32, i32 } %37, 2, !dbg !21
46
+ %43 = bitcast i32 %42 to <2 x bfloat>, !dbg !21
47
+ %44 = extractvalue { i32, i32, i32, i32 } %37, 3, !dbg !21
48
+ %45 = bitcast i32 %44 to <2 x bfloat>, !dbg !21
49
+ %46 = fpext <2 x bfloat> %26 to <2 x float>, !dbg !23
50
+ %47 = fmul <2 x float> %46, %46, !dbg !24
51
+ %48 = fpext <2 x bfloat> %39 to <2 x float>, !dbg !23
52
+ %49 = fmul <2 x float> %48, %48, !dbg !24
53
+ %50 = fadd <2 x float> %47, %49, !dbg !25
54
+ %51 = fpext <2 x bfloat> %28 to <2 x float>, !dbg !23
55
+ %52 = fmul <2 x float> %51, %51, !dbg !24
56
+ %53 = fpext <2 x bfloat> %41 to <2 x float>, !dbg !23
57
+ %54 = fmul <2 x float> %53, %53, !dbg !24
58
+ %55 = fadd <2 x float> %52, %54, !dbg !25
59
+ %56 = fpext <2 x bfloat> %30 to <2 x float>, !dbg !23
60
+ %57 = fmul <2 x float> %56, %56, !dbg !24
61
+ %58 = fpext <2 x bfloat> %43 to <2 x float>, !dbg !23
62
+ %59 = fmul <2 x float> %58, %58, !dbg !24
63
+ %60 = fadd <2 x float> %57, %59, !dbg !25
64
+ %61 = fpext <2 x bfloat> %32 to <2 x float>, !dbg !23
65
+ %62 = fmul <2 x float> %61, %61, !dbg !24
66
+ %63 = fpext <2 x bfloat> %45 to <2 x float>, !dbg !23
67
+ %64 = fmul <2 x float> %63, %63, !dbg !24
68
+ %65 = fadd <2 x float> %62, %64, !dbg !25
69
+ %66 = and i32 %8, 63, !dbg !11
70
+ %67 = or disjoint i32 %7, %66, !dbg !12
71
+ %68 = icmp slt i32 %67, %2, !dbg !13
72
+ %shift = shufflevector <2 x float> %50, <2 x float> poison, <2 x i32> <i32 1, i32 poison>, !dbg !26
73
+ %69 = fadd <2 x float> %50, %shift, !dbg !26
74
+ %70 = fadd <2 x float> %55, %69, !dbg !26
75
+ %shift10 = shufflevector <2 x float> %55, <2 x float> poison, <2 x i32> <i32 1, i32 poison>, !dbg !26
76
+ %71 = fadd <2 x float> %shift10, %70, !dbg !26
77
+ %72 = fadd <2 x float> %60, %71, !dbg !26
78
+ %shift11 = shufflevector <2 x float> %60, <2 x float> poison, <2 x i32> <i32 1, i32 poison>, !dbg !26
79
+ %73 = fadd <2 x float> %shift11, %72, !dbg !26
80
+ %74 = fadd <2 x float> %65, %73, !dbg !26
81
+ %shift12 = shufflevector <2 x float> %65, <2 x float> poison, <2 x i32> <i32 1, i32 poison>, !dbg !26
82
+ %75 = fadd <2 x float> %shift12, %74, !dbg !26
83
+ %76 = extractelement <2 x float> %75, i64 0, !dbg !26
84
+ %77 = select i1 %12, float %76, float 0.000000e+00, !dbg !26
85
+ %78 = bitcast float %77 to i32, !dbg !31
86
+ %79 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %78, i32 4, i32 31), !dbg !31
87
+ %80 = bitcast i32 %79 to float, !dbg !31
88
+ %81 = fadd float %77, %80, !dbg !26
89
+ %82 = bitcast float %81 to i32, !dbg !31
90
+ %83 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %82, i32 2, i32 31), !dbg !31
91
+ %84 = bitcast i32 %83 to float, !dbg !31
92
+ %85 = fadd float %81, %84, !dbg !26
93
+ %86 = bitcast float %85 to i32, !dbg !31
94
+ %87 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %86, i32 1, i32 31), !dbg !31
95
+ %88 = bitcast i32 %87 to float, !dbg !31
96
+ %89 = fadd float %85, %88, !dbg !26
97
+ %90 = getelementptr inbounds nuw float, ptr addrspace(3) @global_smem, i32 %10, !dbg !32
98
+ %91 = bitcast float %89 to <1 x i32>, !dbg !32
99
+ tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %90, <1 x i32> %91, i1 true) #3, !dbg !32
100
+ tail call void @llvm.nvvm.barrier0(), !dbg !32
101
+ %92 = getelementptr inbounds nuw float, ptr addrspace(3) @global_smem, i32 %66, !dbg !32
102
+ %93 = load i32, ptr addrspace(3) %92, align 4, !dbg !32
103
+ %94 = sext i32 %67 to i64, !dbg !33
104
+ %95 = getelementptr float, ptr addrspace(1) %1, i64 %94, !dbg !33
105
+ %96 = and i32 %8, 448, !dbg !34
106
+ %97 = icmp eq i32 %96, 0, !dbg !34
107
+ %98 = and i1 %97, %68, !dbg !34
108
+ tail call void asm sideeffect "@$2 st.global.b32 [ $1 + 0 ], { $0 };", "r,l,b"(i32 %93, ptr addrspace(1) %95, i1 %98) #3, !dbg !34
109
+ ret void, !dbg !35
110
+ }
111
+
112
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none)
113
+ declare noundef i32 @llvm.nvvm.read.ptx.sreg.ctaid.x() #0
114
+
115
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none)
116
+ declare noundef i32 @llvm.nvvm.read.ptx.sreg.tid.x() #0
117
+
118
+ ; Function Attrs: convergent nocallback nounwind memory(inaccessiblemem: readwrite)
119
+ declare i32 @llvm.nvvm.shfl.sync.bfly.i32(i32, i32, i32, i32) #1
120
+
121
+ ; Function Attrs: convergent nocallback nounwind
122
+ declare void @llvm.nvvm.barrier0() #2
123
+
124
+ attributes #0 = { mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none) }
125
+ attributes #1 = { convergent nocallback nounwind memory(inaccessiblemem: readwrite) }
126
+ attributes #2 = { convergent nocallback nounwind }
127
+ attributes #3 = { nounwind }
128
+
129
+ !llvm.module.flags = !{!0, !1}
130
+ !llvm.dbg.cu = !{!2}
131
+ !nvvm.annotations = !{!4}
132
+ !llvm.ident = !{!5}
133
+
134
+ !0 = !{i32 2, !"Debug Info Version", i32 3}
135
+ !1 = !{i32 4, !"nvvm-reflect-ftz", i32 1}
136
+ !2 = distinct !DICompileUnit(language: DW_LANG_C, file: !3, producer: "triton", isOptimized: true, runtimeVersion: 0, emissionKind: LineTablesOnly)
137
+ !3 = !DIFile(filename: "ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py", directory: "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty")
138
+ !4 = !{ptr @triton_red_fused__to_copy_mean_pow_4, !"reqntidx", i32 512}
139
+ !5 = !{!"clang version 3.8.0 (tags/RELEASE_380/final)"}
140
+ !6 = distinct !DISubprogram(name: "triton_red_fused__to_copy_mean_pow_4", linkageName: "triton_red_fused__to_copy_mean_pow_4", scope: !3, file: !3, line: 18, type: !7, scopeLine: 18, spFlags: DISPFlagDefinition | DISPFlagOptimized, unit: !2)
141
+ !7 = !DISubroutineType(cc: DW_CC_normal, types: !8)
142
+ !8 = !{}
143
+ !9 = !DILocation(line: 22, column: 28, scope: !6)
144
+ !10 = !DILocation(line: 22, column: 33, scope: !6)
145
+ !11 = !DILocation(line: 23, column: 44, scope: !6)
146
+ !12 = !DILocation(line: 23, column: 23, scope: !6)
147
+ !13 = !DILocation(line: 24, column: 21, scope: !6)
148
+ !14 = !DILocation(line: 25, column: 37, scope: !6)
149
+ !15 = !DILocation(line: 28, column: 19, scope: !6)
150
+ !16 = !DILocation(line: 27, column: 19, scope: !6)
151
+ !17 = !DILocation(line: 37, column: 52, scope: !6)
152
+ !18 = !DILocation(line: 37, column: 62, scope: !6)
153
+ !19 = !DILocation(line: 37, column: 57, scope: !6)
154
+ !20 = !DILocation(line: 37, column: 34, scope: !6)
155
+ !21 = !DILocation(line: 37, column: 67, scope: !6)
156
+ !22 = !DILocation(line: 32, column: 31, scope: !6)
157
+ !23 = !DILocation(line: 37, column: 129, scope: !6)
158
+ !24 = !DILocation(line: 39, column: 22, scope: !6)
159
+ !25 = !DILocation(line: 41, column: 23, scope: !6)
160
+ !26 = !DILocation(line: 256, column: 15, scope: !27, inlinedAt: !30)
161
+ !27 = distinct !DILexicalBlockFile(scope: !29, file: !28, discriminator: 0)
162
+ !28 = !DIFile(filename: "standard.py", directory: "/home/x/hfenv/lib/python3.12/site-packages/triton/language")
163
+ !29 = distinct !DILexicalBlockFile(scope: !6, file: !28, discriminator: 0)
164
+ !30 = !DILocation(line: 43, column: 25, scope: !6)
165
+ !31 = !DILocation(line: 286, column: 36, scope: !29, inlinedAt: !30)
166
+ !32 = !DILocation(line: 43, column: 28, scope: !6)
167
+ !33 = !DILocation(line: 44, column: 25, scope: !6)
168
+ !34 = !DILocation(line: 44, column: 36, scope: !6)
169
+ !35 = !DILocation(line: 44, column: 4, scope: !6)
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.ptx ADDED
@@ -0,0 +1,469 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ //
2
+ // Generated by LLVM NVPTX Back-End
3
+ //
4
+
5
+ .version 8.4
6
+ .target sm_90a
7
+ .address_size 64
8
+
9
+ // .globl triton_red_fused__to_copy_mean_pow_4 // -- Begin function triton_red_fused__to_copy_mean_pow_4
10
+ .extern .shared .align 16 .b8 global_smem[];
11
+ // @triton_red_fused__to_copy_mean_pow_4
12
+ .visible .entry triton_red_fused__to_copy_mean_pow_4(
13
+ .param .u64 .ptr .global .align 1 triton_red_fused__to_copy_mean_pow_4_param_0,
14
+ .param .u64 .ptr .global .align 1 triton_red_fused__to_copy_mean_pow_4_param_1,
15
+ .param .u32 triton_red_fused__to_copy_mean_pow_4_param_2,
16
+ .param .u32 triton_red_fused__to_copy_mean_pow_4_param_3,
17
+ .param .u64 .ptr .global .align 1 triton_red_fused__to_copy_mean_pow_4_param_4
18
+ )
19
+ .reqntid 512, 1, 1
20
+ {
21
+ .reg .pred %p<7>;
22
+ .reg .b16 %rs<17>;
23
+ .reg .b32 %r<51>;
24
+ .reg .f32 %f<47>;
25
+ .reg .b64 %rd<13>;
26
+ .loc 1 18 0 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:18:0
27
+ $L__func_begin0:
28
+ .loc 1 18 0 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:18:0
29
+
30
+ // %bb.0:
31
+ ld.param.u64 %rd4, [triton_red_fused__to_copy_mean_pow_4_param_0];
32
+ ld.param.u64 %rd5, [triton_red_fused__to_copy_mean_pow_4_param_1];
33
+ $L__tmp0:
34
+ .loc 1 22 28 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:22:28
35
+ mov.u32 %r20, %ctaid.x;
36
+ .loc 1 22 33 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:22:33
37
+ shl.b32 %r21, %r20, 6;
38
+ ld.param.u32 %r22, [triton_red_fused__to_copy_mean_pow_4_param_2];
39
+ .loc 1 23 44 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:23:44
40
+ mov.u32 %r23, %tid.x;
41
+ bfe.u32 %r24, %r23, 3, 6;
42
+ .loc 1 23 23 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:23:23
43
+ or.b32 %r25, %r24, %r21;
44
+ .loc 1 24 21 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:24:21
45
+ setp.lt.s32 %p1, %r25, %r22;
46
+ .loc 1 25 37 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:25:37
47
+ shl.b32 %r26, %r23, 3;
48
+ and.b32 %r27, %r26, 56;
49
+ .loc 1 28 19 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:28:19
50
+ bfe.s32 %r28, %r20, 25, 1;
51
+ shr.u32 %r29, %r28, 29;
52
+ add.s32 %r30, %r25, %r29;
53
+ shr.u32 %r31, %r30, 3;
54
+ .loc 1 27 19 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:27:19
55
+ and.b32 %r32, %r30, 33554424;
56
+ sub.s32 %r33, %r25, %r32;
57
+ .loc 1 37 52 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:37:52
58
+ shl.b32 %r34, %r33, 7;
59
+ mad.lo.s32 %r35, %r31, 7168, %r34;
60
+ add.s32 %r36, %r35, 5120;
61
+ .loc 1 37 57 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:37:57
62
+ or.b32 %r37, %r36, %r27;
63
+ .loc 1 37 34 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:37:34
64
+ mul.wide.s32 %rd6, %r37, 2;
65
+ add.s64 %rd1, %rd4, %rd6;
66
+ mov.b32 %r5, 0;
67
+ .loc 1 37 67 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:37:67
68
+ // begin inline asm
69
+ mov.u32 %r1, %r5;
70
+ mov.u32 %r2, %r5;
71
+ mov.u32 %r3, %r5;
72
+ mov.u32 %r4, %r5;
73
+ @%p1 ld.global.L1::evict_first.v4.b32 { %r1, %r2, %r3, %r4 }, [ %rd1 + 0 ];
74
+ // end inline asm
75
+ .loc 1 37 34 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:37:34
76
+ cvt.s64.s32 %rd7, %r36;
77
+ cvt.u64.u32 %rd8, %r27;
78
+ or.b64 %rd9, %rd7, %rd8;
79
+ shl.b64 %rd10, %rd9, 1;
80
+ add.s64 %rd11, %rd4, %rd10;
81
+ add.s64 %rd2, %rd11, 128;
82
+ .loc 1 37 67 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:37:67
83
+ // begin inline asm
84
+ mov.u32 %r9, %r5;
85
+ mov.u32 %r10, %r5;
86
+ mov.u32 %r11, %r5;
87
+ mov.u32 %r12, %r5;
88
+ @%p1 ld.global.L1::evict_first.v4.b32 { %r9, %r10, %r11, %r12 }, [ %rd2 + 0 ];
89
+ // end inline asm
90
+ .loc 1 37 129 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:37:129
91
+ mov.b32 {%rs1, %rs2}, %r1;
92
+ cvt.f32.bf16 %f1, %rs1;
93
+ cvt.f32.bf16 %f2, %rs2;
94
+ mov.b32 {%rs3, %rs4}, %r9;
95
+ cvt.f32.bf16 %f3, %rs4;
96
+ cvt.f32.bf16 %f4, %rs3;
97
+ .loc 1 39 22 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:39:22
98
+ mul.f32 %f5, %f4, %f4;
99
+ mul.f32 %f6, %f3, %f3;
100
+ .loc 1 41 23 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:41:23
101
+ fma.rn.f32 %f7, %f2, %f2, %f6;
102
+ fma.rn.f32 %f8, %f1, %f1, %f5;
103
+ .loc 1 37 129 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:37:129
104
+ mov.b32 {%rs5, %rs6}, %r2;
105
+ cvt.f32.bf16 %f9, %rs5;
106
+ cvt.f32.bf16 %f10, %rs6;
107
+ mov.b32 {%rs7, %rs8}, %r10;
108
+ cvt.f32.bf16 %f11, %rs8;
109
+ cvt.f32.bf16 %f12, %rs7;
110
+ .loc 1 39 22 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:39:22
111
+ mul.f32 %f13, %f12, %f12;
112
+ mul.f32 %f14, %f11, %f11;
113
+ .loc 1 41 23 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:41:23
114
+ fma.rn.f32 %f15, %f10, %f10, %f14;
115
+ fma.rn.f32 %f16, %f9, %f9, %f13;
116
+ .loc 1 37 129 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:37:129
117
+ mov.b32 {%rs9, %rs10}, %r3;
118
+ cvt.f32.bf16 %f17, %rs9;
119
+ cvt.f32.bf16 %f18, %rs10;
120
+ mov.b32 {%rs11, %rs12}, %r11;
121
+ cvt.f32.bf16 %f19, %rs12;
122
+ cvt.f32.bf16 %f20, %rs11;
123
+ .loc 1 39 22 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:39:22
124
+ mul.f32 %f21, %f20, %f20;
125
+ mul.f32 %f22, %f19, %f19;
126
+ .loc 1 41 23 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:41:23
127
+ fma.rn.f32 %f23, %f18, %f18, %f22;
128
+ fma.rn.f32 %f24, %f17, %f17, %f21;
129
+ .loc 1 37 129 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:37:129
130
+ mov.b32 {%rs13, %rs14}, %r4;
131
+ cvt.f32.bf16 %f25, %rs13;
132
+ cvt.f32.bf16 %f26, %rs14;
133
+ mov.b32 {%rs15, %rs16}, %r12;
134
+ cvt.f32.bf16 %f27, %rs16;
135
+ cvt.f32.bf16 %f28, %rs15;
136
+ .loc 1 39 22 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:39:22
137
+ mul.f32 %f29, %f28, %f28;
138
+ mul.f32 %f30, %f27, %f27;
139
+ .loc 1 41 23 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:41:23
140
+ fma.rn.f32 %f31, %f26, %f26, %f30;
141
+ fma.rn.f32 %f32, %f25, %f25, %f29;
142
+ .loc 1 23 44 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:23:44
143
+ and.b32 %r38, %r23, 63;
144
+ .loc 1 23 23 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:23:23
145
+ or.b32 %r39, %r21, %r38;
146
+ .loc 1 24 21 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:24:21
147
+ setp.lt.s32 %p5, %r39, %r22;
148
+ $L__tmp1:
149
+ .loc 2 256 15 // standard.py:256:15
150
+ add.f32 %f33, %f8, %f7;
151
+ add.f32 %f34, %f16, %f33;
152
+ add.f32 %f35, %f15, %f34;
153
+ add.f32 %f36, %f24, %f35;
154
+ add.f32 %f37, %f23, %f36;
155
+ add.f32 %f38, %f32, %f37;
156
+ add.f32 %f39, %f31, %f38;
157
+ selp.f32 %f40, %f39, 0f00000000, %p1;
158
+ .loc 2 286 36 // standard.py:286:36
159
+ mov.b32 %r40, %f40;
160
+ shfl.sync.bfly.b32 %r41, %r40, 4, 31, -1;
161
+ mov.b32 %f41, %r41;
162
+ .loc 2 256 15 // standard.py:256:15
163
+ add.f32 %f42, %f40, %f41;
164
+ .loc 2 286 36 // standard.py:286:36
165
+ mov.b32 %r42, %f42;
166
+ shfl.sync.bfly.b32 %r43, %r42, 2, 31, -1;
167
+ mov.b32 %f43, %r43;
168
+ .loc 2 256 15 // standard.py:256:15
169
+ add.f32 %f44, %f42, %f43;
170
+ .loc 2 286 36 // standard.py:286:36
171
+ mov.b32 %r44, %f44;
172
+ shfl.sync.bfly.b32 %r45, %r44, 1, 31, -1;
173
+ mov.b32 %f45, %r45;
174
+ .loc 2 256 15 // standard.py:256:15
175
+ add.f32 %f46, %f44, %f45;
176
+ $L__tmp2:
177
+ .loc 1 43 28 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:43:28
178
+ shl.b32 %r46, %r24, 2;
179
+ mov.u32 %r47, global_smem;
180
+ add.s32 %r17, %r47, %r46;
181
+ mov.b32 %r18, %f46;
182
+ mov.pred %p3, -1;
183
+ // begin inline asm
184
+ @%p3 st.shared.b32 [ %r17 + 0 ], %r18;
185
+ // end inline asm
186
+ bar.sync 0;
187
+ shl.b32 %r48, %r38, 2;
188
+ add.s32 %r49, %r47, %r48;
189
+ ld.shared.u32 %r19, [%r49];
190
+ .loc 1 44 25 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:44:25
191
+ mul.wide.s32 %rd12, %r39, 4;
192
+ add.s64 %rd3, %rd5, %rd12;
193
+ .loc 1 44 36 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:44:36
194
+ and.b32 %r50, %r23, 448;
195
+ setp.eq.s32 %p6, %r50, 0;
196
+ and.pred %p4, %p6, %p5;
197
+ // begin inline asm
198
+ @%p4 st.global.b32 [ %rd3 + 0 ], { %r19 };
199
+ // end inline asm
200
+ .loc 1 44 4 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:44:4
201
+ ret;
202
+ $L__tmp3:
203
+ $L__func_end0:
204
+ // -- End function
205
+ }
206
+ .file 1 "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py"
207
+ .file 2 "/home/x/hfenv/lib/python3.12/site-packages/triton/language/standard.py"
208
+ .section .debug_abbrev
209
+ {
210
+ .b8 1 // Abbreviation Code
211
+ .b8 17 // DW_TAG_compile_unit
212
+ .b8 1 // DW_CHILDREN_yes
213
+ .b8 37 // DW_AT_producer
214
+ .b8 8 // DW_FORM_string
215
+ .b8 19 // DW_AT_language
216
+ .b8 5 // DW_FORM_data2
217
+ .b8 3 // DW_AT_name
218
+ .b8 8 // DW_FORM_string
219
+ .b8 16 // DW_AT_stmt_list
220
+ .b8 6 // DW_FORM_data4
221
+ .b8 27 // DW_AT_comp_dir
222
+ .b8 8 // DW_FORM_string
223
+ .b8 0 // EOM(1)
224
+ .b8 0 // EOM(2)
225
+ .b8 2 // Abbreviation Code
226
+ .b8 46 // DW_TAG_subprogram
227
+ .b8 0 // DW_CHILDREN_no
228
+ .b8 3 // DW_AT_name
229
+ .b8 8 // DW_FORM_string
230
+ .b8 32 // DW_AT_inline
231
+ .b8 11 // DW_FORM_data1
232
+ .b8 0 // EOM(1)
233
+ .b8 0 // EOM(2)
234
+ .b8 3 // Abbreviation Code
235
+ .b8 46 // DW_TAG_subprogram
236
+ .b8 1 // DW_CHILDREN_yes
237
+ .b8 17 // DW_AT_low_pc
238
+ .b8 1 // DW_FORM_addr
239
+ .b8 18 // DW_AT_high_pc
240
+ .b8 1 // DW_FORM_addr
241
+ .b8 49 // DW_AT_abstract_origin
242
+ .b8 19 // DW_FORM_ref4
243
+ .b8 0 // EOM(1)
244
+ .b8 0 // EOM(2)
245
+ .b8 4 // Abbreviation Code
246
+ .b8 29 // DW_TAG_inlined_subroutine
247
+ .b8 0 // DW_CHILDREN_no
248
+ .b8 49 // DW_AT_abstract_origin
249
+ .b8 19 // DW_FORM_ref4
250
+ .b8 17 // DW_AT_low_pc
251
+ .b8 1 // DW_FORM_addr
252
+ .b8 18 // DW_AT_high_pc
253
+ .b8 1 // DW_FORM_addr
254
+ .b8 88 // DW_AT_call_file
255
+ .b8 11 // DW_FORM_data1
256
+ .b8 89 // DW_AT_call_line
257
+ .b8 11 // DW_FORM_data1
258
+ .b8 87 // DW_AT_call_column
259
+ .b8 11 // DW_FORM_data1
260
+ .b8 0 // EOM(1)
261
+ .b8 0 // EOM(2)
262
+ .b8 0 // EOM(3)
263
+ }
264
+ .section .debug_info
265
+ {
266
+ .b32 241 // Length of Unit
267
+ .b8 2 // DWARF version number
268
+ .b8 0
269
+ .b32 .debug_abbrev // Offset Into Abbrev. Section
270
+ .b8 8 // Address Size (in bytes)
271
+ .b8 1 // Abbrev [1] 0xb:0xea DW_TAG_compile_unit
272
+ .b8 116 // DW_AT_producer
273
+ .b8 114
274
+ .b8 105
275
+ .b8 116
276
+ .b8 111
277
+ .b8 110
278
+ .b8 0
279
+ .b8 2 // DW_AT_language
280
+ .b8 0
281
+ .b8 99 // DW_AT_name
282
+ .b8 116
283
+ .b8 121
284
+ .b8 100
285
+ .b8 106
286
+ .b8 97
287
+ .b8 55
288
+ .b8 107
289
+ .b8 116
290
+ .b8 114
291
+ .b8 101
292
+ .b8 119
293
+ .b8 120
294
+ .b8 113
295
+ .b8 101
296
+ .b8 104
297
+ .b8 107
298
+ .b8 105
299
+ .b8 122
300
+ .b8 106
301
+ .b8 105
302
+ .b8 55
303
+ .b8 113
304
+ .b8 122
305
+ .b8 105
306
+ .b8 52
307
+ .b8 119
308
+ .b8 122
309
+ .b8 113
310
+ .b8 107
311
+ .b8 106
312
+ .b8 99
313
+ .b8 115
314
+ .b8 119
315
+ .b8 106
316
+ .b8 107
317
+ .b8 113
318
+ .b8 110
319
+ .b8 120
320
+ .b8 97
321
+ .b8 108
322
+ .b8 102
323
+ .b8 101
324
+ .b8 121
325
+ .b8 51
326
+ .b8 109
327
+ .b8 111
328
+ .b8 104
329
+ .b8 103
330
+ .b8 103
331
+ .b8 111
332
+ .b8 116
333
+ .b8 46
334
+ .b8 112
335
+ .b8 121
336
+ .b8 0
337
+ .b32 .debug_line // DW_AT_stmt_list
338
+ .b8 47 // DW_AT_comp_dir
339
+ .b8 104
340
+ .b8 111
341
+ .b8 109
342
+ .b8 101
343
+ .b8 47
344
+ .b8 120
345
+ .b8 47
346
+ .b8 46
347
+ .b8 99
348
+ .b8 97
349
+ .b8 99
350
+ .b8 104
351
+ .b8 101
352
+ .b8 47
353
+ .b8 118
354
+ .b8 108
355
+ .b8 108
356
+ .b8 109
357
+ .b8 47
358
+ .b8 116
359
+ .b8 111
360
+ .b8 114
361
+ .b8 99
362
+ .b8 104
363
+ .b8 95
364
+ .b8 99
365
+ .b8 111
366
+ .b8 109
367
+ .b8 112
368
+ .b8 105
369
+ .b8 108
370
+ .b8 101
371
+ .b8 95
372
+ .b8 99
373
+ .b8 97
374
+ .b8 99
375
+ .b8 104
376
+ .b8 101
377
+ .b8 47
378
+ .b8 57
379
+ .b8 48
380
+ .b8 98
381
+ .b8 52
382
+ .b8 53
383
+ .b8 98
384
+ .b8 99
385
+ .b8 101
386
+ .b8 48
387
+ .b8 50
388
+ .b8 47
389
+ .b8 114
390
+ .b8 97
391
+ .b8 110
392
+ .b8 107
393
+ .b8 95
394
+ .b8 48
395
+ .b8 95
396
+ .b8 48
397
+ .b8 47
398
+ .b8 105
399
+ .b8 110
400
+ .b8 100
401
+ .b8 117
402
+ .b8 99
403
+ .b8 116
404
+ .b8 111
405
+ .b8 114
406
+ .b8 95
407
+ .b8 99
408
+ .b8 97
409
+ .b8 99
410
+ .b8 104
411
+ .b8 101
412
+ .b8 47
413
+ .b8 116
414
+ .b8 121
415
+ .b8 0
416
+ .b8 2 // Abbrev [2] 0x9f:0x27 DW_TAG_subprogram
417
+ .b8 116 // DW_AT_name
418
+ .b8 114
419
+ .b8 105
420
+ .b8 116
421
+ .b8 111
422
+ .b8 110
423
+ .b8 95
424
+ .b8 114
425
+ .b8 101
426
+ .b8 100
427
+ .b8 95
428
+ .b8 102
429
+ .b8 117
430
+ .b8 115
431
+ .b8 101
432
+ .b8 100
433
+ .b8 95
434
+ .b8 95
435
+ .b8 116
436
+ .b8 111
437
+ .b8 95
438
+ .b8 99
439
+ .b8 111
440
+ .b8 112
441
+ .b8 121
442
+ .b8 95
443
+ .b8 109
444
+ .b8 101
445
+ .b8 97
446
+ .b8 110
447
+ .b8 95
448
+ .b8 112
449
+ .b8 111
450
+ .b8 119
451
+ .b8 95
452
+ .b8 52
453
+ .b8 0
454
+ .b8 1 // DW_AT_inline
455
+ .b8 3 // Abbrev [3] 0xc6:0x2e DW_TAG_subprogram
456
+ .b64 $L__func_begin0 // DW_AT_low_pc
457
+ .b64 $L__func_end0 // DW_AT_high_pc
458
+ .b32 159 // DW_AT_abstract_origin
459
+ .b8 4 // Abbrev [4] 0xdb:0x18 DW_TAG_inlined_subroutine
460
+ .b32 159 // DW_AT_abstract_origin
461
+ .b64 $L__tmp1 // DW_AT_low_pc
462
+ .b64 $L__tmp2 // DW_AT_high_pc
463
+ .b8 1 // DW_AT_call_file
464
+ .b8 43 // DW_AT_call_line
465
+ .b8 25 // DW_AT_call_column
466
+ .b8 0 // End Of Children Mark
467
+ .b8 0 // End Of Children Mark
468
+ }
469
+ .section .debug_macinfo { }
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.ttgir ADDED
@@ -0,0 +1,106 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #blocked = #ttg.blocked<{sizePerThread = [1, 8], threadsPerWarp = [4, 8], warpsPerCTA = [16, 1], order = [1, 0]}>
2
+ #blocked1 = #ttg.blocked<{sizePerThread = [1, 1], threadsPerWarp = [32, 1], warpsPerCTA = [2, 8], order = [0, 1]}>
3
+ #loc = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":18:0)
4
+ #loc1 = loc(unknown)
5
+ #loc27 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":43:25)
6
+ #loc34 = loc(callsite(#loc1 at #loc27))
7
+ module attributes {"ttg.num-ctas" = 1 : i32, "ttg.num-warps" = 16 : i32, ttg.target = "cuda:90", "ttg.threads-per-warp" = 32 : i32} {
8
+ tt.func public @triton_red_fused__to_copy_mean_pow_4(%arg0: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":18:0), %arg1: !tt.ptr<f32> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":18:0), %arg2: i32 loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":18:0), %arg3: i32 {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":18:0)) attributes {noinline = false} {
9
+ %cst = arith.constant dense<0.000000e+00> : tensor<64x64xf32, #blocked> loc(#loc1)
10
+ %c0_i32 = arith.constant 0 : i32 loc(#loc1)
11
+ %c128_i32 = arith.constant 128 : i32 loc(#loc1)
12
+ %cst_0 = arith.constant dense<0.000000e+00> : tensor<64x64xbf16, #blocked> loc(#loc1)
13
+ %c64_i32 = arith.constant 64 : i32 loc(#loc1)
14
+ %cst_1 = arith.constant dense<8> : tensor<64x1xi32, #blocked> loc(#loc1)
15
+ %cst_2 = arith.constant dense<7168> : tensor<64x1xi32, #blocked> loc(#loc1)
16
+ %cst_3 = arith.constant dense<128> : tensor<64x1xi32, #blocked> loc(#loc1)
17
+ %cst_4 = arith.constant dense<5120> : tensor<1x64xi32, #blocked> loc(#loc1)
18
+ %cst_5 = arith.constant dense<128> : tensor<1x64xi32, #blocked> loc(#loc1)
19
+ %0 = tt.get_program_id x : i32 loc(#loc2)
20
+ %1 = arith.muli %0, %c64_i32 : i32 loc(#loc3)
21
+ %2 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #ttg.slice<{dim = 1, parent = #blocked}>> loc(#loc4)
22
+ %3 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #ttg.slice<{dim = 1, parent = #blocked1}>> loc(#loc4)
23
+ %4 = tt.expand_dims %2 {axis = 1 : i32} : tensor<64xi32, #ttg.slice<{dim = 1, parent = #blocked}>> -> tensor<64x1xi32, #blocked> loc(#loc4)
24
+ %5 = tt.expand_dims %3 {axis = 1 : i32} : tensor<64xi32, #ttg.slice<{dim = 1, parent = #blocked1}>> -> tensor<64x1xi32, #blocked1> loc(#loc4)
25
+ %6 = tt.splat %1 : i32 -> tensor<64x1xi32, #blocked> loc(#loc5)
26
+ %7 = tt.splat %1 : i32 -> tensor<64x1xi32, #blocked1> loc(#loc5)
27
+ %8 = arith.addi %6, %4 : tensor<64x1xi32, #blocked> loc(#loc5)
28
+ %9 = arith.addi %7, %5 : tensor<64x1xi32, #blocked1> loc(#loc5)
29
+ %10 = tt.splat %arg2 : i32 -> tensor<64x1xi32, #blocked> loc(#loc6)
30
+ %11 = tt.splat %arg2 : i32 -> tensor<64x1xi32, #blocked1> loc(#loc6)
31
+ %12 = arith.cmpi slt, %8, %10 : tensor<64x1xi32, #blocked> loc(#loc6)
32
+ %13 = arith.cmpi slt, %9, %11 : tensor<64x1xi32, #blocked1> loc(#loc6)
33
+ %14 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32, #ttg.slice<{dim = 0, parent = #blocked}>> loc(#loc7)
34
+ %15 = tt.expand_dims %14 {axis = 0 : i32} : tensor<64xi32, #ttg.slice<{dim = 0, parent = #blocked}>> -> tensor<1x64xi32, #blocked> loc(#loc7)
35
+ %16 = arith.remsi %8, %cst_1 : tensor<64x1xi32, #blocked> loc(#loc8)
36
+ %17 = arith.divsi %8, %cst_1 : tensor<64x1xi32, #blocked> loc(#loc9)
37
+ %18 = arith.muli %16, %cst_3 : tensor<64x1xi32, #blocked> loc(#loc10)
38
+ %19 = tt.broadcast %18 : tensor<64x1xi32, #blocked> -> tensor<64x64xi32, #blocked> loc(#loc11)
39
+ %20 = arith.muli %17, %cst_2 : tensor<64x1xi32, #blocked> loc(#loc12)
40
+ %21 = tt.broadcast %20 : tensor<64x1xi32, #blocked> -> tensor<64x64xi32, #blocked> loc(#loc13)
41
+ %22 = tt.splat %arg0 : !tt.ptr<bf16> -> tensor<64x64x!tt.ptr<bf16>, #blocked> loc(#loc14)
42
+ %23 = tt.broadcast %12 : tensor<64x1xi1, #blocked> -> tensor<64x64xi1, #blocked> loc(#loc15)
43
+ %24 = scf.for %arg4 = %c0_i32 to %c128_i32 step %c64_i32 iter_args(%arg5 = %cst) -> (tensor<64x64xf32, #blocked>) : i32 {
44
+ %30 = tt.splat %arg4 : i32 -> tensor<1x64xi32, #blocked> loc(#loc17)
45
+ %31 = arith.addi %30, %15 : tensor<1x64xi32, #blocked> loc(#loc17)
46
+ %32 = arith.cmpi slt, %31, %cst_5 : tensor<1x64xi32, #blocked> loc(#loc18)
47
+ %33 = arith.addi %31, %cst_4 : tensor<1x64xi32, #blocked> loc(#loc19)
48
+ %34 = tt.broadcast %33 : tensor<1x64xi32, #blocked> -> tensor<64x64xi32, #blocked> loc(#loc11)
49
+ %35 = arith.addi %34, %19 : tensor<64x64xi32, #blocked> loc(#loc11)
50
+ %36 = arith.addi %35, %21 : tensor<64x64xi32, #blocked> loc(#loc13)
51
+ %37 = tt.addptr %22, %36 : tensor<64x64x!tt.ptr<bf16>, #blocked>, tensor<64x64xi32, #blocked> loc(#loc14)
52
+ %38 = tt.broadcast %32 : tensor<1x64xi1, #blocked> -> tensor<64x64xi1, #blocked> loc(#loc15)
53
+ %39 = arith.andi %38, %23 : tensor<64x64xi1, #blocked> loc(#loc15)
54
+ %40 = tt.load %37, %39, %cst_0 evictionPolicy = evict_first : tensor<64x64x!tt.ptr<bf16>, #blocked> loc(#loc20)
55
+ %41 = arith.extf %40 : tensor<64x64xbf16, #blocked> to tensor<64x64xf32, #blocked> loc(#loc21)
56
+ %42 = arith.mulf %41, %41 : tensor<64x64xf32, #blocked> loc(#loc22)
57
+ %43 = arith.addf %arg5, %42 : tensor<64x64xf32, #blocked> loc(#loc23)
58
+ %44 = arith.select %39, %43, %arg5 : tensor<64x64xi1, #blocked>, tensor<64x64xf32, #blocked> loc(#loc24)
59
+ scf.yield %44 : tensor<64x64xf32, #blocked> loc(#loc25)
60
+ } loc(#loc16)
61
+ %25 = "tt.reduce"(%24) <{axis = 1 : i32}> ({
62
+ ^bb0(%arg4: f32 loc(callsite(#loc1 at #loc27)), %arg5: f32 loc(callsite(#loc1 at #loc27))):
63
+ %30 = arith.addf %arg4, %arg5 : f32 loc(#loc36)
64
+ tt.reduce.return %30 : f32 loc(#loc33)
65
+ }) : (tensor<64x64xf32, #blocked>) -> tensor<64xf32, #ttg.slice<{dim = 1, parent = #blocked}>> loc(#loc33)
66
+ %26 = ttg.convert_layout %25 : tensor<64xf32, #ttg.slice<{dim = 1, parent = #blocked}>> -> tensor<64xf32, #ttg.slice<{dim = 1, parent = #blocked1}>> loc(#loc29)
67
+ %27 = tt.expand_dims %26 {axis = 1 : i32} : tensor<64xf32, #ttg.slice<{dim = 1, parent = #blocked1}>> -> tensor<64x1xf32, #blocked1> loc(#loc29)
68
+ %28 = tt.splat %arg1 : !tt.ptr<f32> -> tensor<64x1x!tt.ptr<f32>, #blocked1> loc(#loc30)
69
+ %29 = tt.addptr %28, %9 : tensor<64x1x!tt.ptr<f32>, #blocked1>, tensor<64x1xi32, #blocked1> loc(#loc30)
70
+ tt.store %29, %27, %13 : tensor<64x1x!tt.ptr<f32>, #blocked1> loc(#loc31)
71
+ tt.return loc(#loc32)
72
+ } loc(#loc)
73
+ } loc(#loc)
74
+ #loc2 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":22:28)
75
+ #loc3 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":22:33)
76
+ #loc4 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":23:44)
77
+ #loc5 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":23:23)
78
+ #loc6 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":24:21)
79
+ #loc7 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":25:37)
80
+ #loc8 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":27:19)
81
+ #loc9 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":28:19)
82
+ #loc10 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:52)
83
+ #loc11 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:48)
84
+ #loc12 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:62)
85
+ #loc13 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:57)
86
+ #loc14 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:34)
87
+ #loc15 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:77)
88
+ #loc16 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":31:40)
89
+ #loc17 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":32:31)
90
+ #loc18 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":33:29)
91
+ #loc19 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:41)
92
+ #loc20 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:67)
93
+ #loc21 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:129)
94
+ #loc22 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":39:22)
95
+ #loc23 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":41:23)
96
+ #loc24 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":42:48)
97
+ #loc25 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":42:8)
98
+ #loc26 = loc("/home/x/hfenv/lib/python3.12/site-packages/triton/language/standard.py":286:36)
99
+ #loc28 = loc("/home/x/hfenv/lib/python3.12/site-packages/triton/language/standard.py":256:15)
100
+ #loc29 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":43:28)
101
+ #loc30 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":44:25)
102
+ #loc31 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":44:36)
103
+ #loc32 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":44:4)
104
+ #loc33 = loc(callsite(#loc26 at #loc27))
105
+ #loc35 = loc(callsite(#loc28 at #loc26))
106
+ #loc36 = loc(callsite(#loc35 at #loc27))
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/3XUHHEPI4U35KUDY5SIR32FKTGYK5BTVZDQOKVNBZWJ4SYGJUKAA/triton_red_fused__to_copy_mean_pow_4.ttir ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #loc = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":18:0)
2
+ #loc1 = loc(unknown)
3
+ #loc28 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":43:25)
4
+ #loc35 = loc(callsite(#loc1 at #loc28))
5
+ module {
6
+ tt.func public @triton_red_fused__to_copy_mean_pow_4(%arg0: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":18:0), %arg1: !tt.ptr<f32> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":18:0), %arg2: i32 loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":18:0), %arg3: i32 {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":18:0)) attributes {noinline = false} {
7
+ %cst = arith.constant dense<0.000000e+00> : tensor<64x64xbf16> loc(#loc1)
8
+ %c128_i32 = arith.constant 128 : i32 loc(#loc1)
9
+ %c0_i32 = arith.constant 0 : i32 loc(#loc1)
10
+ %cst_0 = arith.constant dense<7168> : tensor<64x1xi32> loc(#loc1)
11
+ %cst_1 = arith.constant dense<128> : tensor<64x1xi32> loc(#loc1)
12
+ %cst_2 = arith.constant dense<5120> : tensor<1x64xi32> loc(#loc1)
13
+ %cst_3 = arith.constant dense<128> : tensor<1x64xi32> loc(#loc1)
14
+ %cst_4 = arith.constant dense<0.000000e+00> : tensor<64x64xf32> loc(#loc1)
15
+ %cst_5 = arith.constant dense<8> : tensor<64x1xi32> loc(#loc1)
16
+ %c64_i32 = arith.constant 64 : i32 loc(#loc1)
17
+ %0 = tt.get_program_id x : i32 loc(#loc2)
18
+ %1 = arith.muli %0, %c64_i32 : i32 loc(#loc3)
19
+ %2 = tt.make_range {end = 64 : i32, start = 0 : i32} : tensor<64xi32> loc(#loc4)
20
+ %3 = tt.expand_dims %2 {axis = 1 : i32} : tensor<64xi32> -> tensor<64x1xi32> loc(#loc5)
21
+ %4 = tt.splat %1 : i32 -> tensor<64x1xi32> loc(#loc6)
22
+ %5 = arith.addi %4, %3 : tensor<64x1xi32> loc(#loc6)
23
+ %6 = tt.splat %arg2 : i32 -> tensor<64x1xi32> loc(#loc7)
24
+ %7 = arith.cmpi slt, %5, %6 : tensor<64x1xi32> loc(#loc7)
25
+ %8 = tt.expand_dims %2 {axis = 0 : i32} : tensor<64xi32> -> tensor<1x64xi32> loc(#loc8)
26
+ %9 = arith.remsi %5, %cst_5 : tensor<64x1xi32> loc(#loc9)
27
+ %10 = arith.divsi %5, %cst_5 : tensor<64x1xi32> loc(#loc10)
28
+ %11 = scf.for %arg4 = %c0_i32 to %c128_i32 step %c64_i32 iter_args(%arg5 = %cst_4) -> (tensor<64x64xf32>) : i32 {
29
+ %16 = tt.splat %arg4 : i32 -> tensor<1x64xi32> loc(#loc12)
30
+ %17 = arith.addi %16, %8 : tensor<1x64xi32> loc(#loc12)
31
+ %18 = arith.cmpi slt, %17, %cst_3 : tensor<1x64xi32> loc(#loc13)
32
+ %19 = arith.addi %17, %cst_2 : tensor<1x64xi32> loc(#loc14)
33
+ %20 = arith.muli %9, %cst_1 : tensor<64x1xi32> loc(#loc15)
34
+ %21 = tt.broadcast %19 : tensor<1x64xi32> -> tensor<64x64xi32> loc(#loc16)
35
+ %22 = tt.broadcast %20 : tensor<64x1xi32> -> tensor<64x64xi32> loc(#loc16)
36
+ %23 = arith.addi %21, %22 : tensor<64x64xi32> loc(#loc16)
37
+ %24 = arith.muli %10, %cst_0 : tensor<64x1xi32> loc(#loc17)
38
+ %25 = tt.broadcast %24 : tensor<64x1xi32> -> tensor<64x64xi32> loc(#loc18)
39
+ %26 = arith.addi %23, %25 : tensor<64x64xi32> loc(#loc18)
40
+ %27 = tt.splat %arg0 : !tt.ptr<bf16> -> tensor<64x64x!tt.ptr<bf16>> loc(#loc19)
41
+ %28 = tt.addptr %27, %26 : tensor<64x64x!tt.ptr<bf16>>, tensor<64x64xi32> loc(#loc19)
42
+ %29 = tt.broadcast %18 : tensor<1x64xi1> -> tensor<64x64xi1> loc(#loc20)
43
+ %30 = tt.broadcast %7 : tensor<64x1xi1> -> tensor<64x64xi1> loc(#loc20)
44
+ %31 = arith.andi %29, %30 : tensor<64x64xi1> loc(#loc20)
45
+ %32 = tt.load %28, %31, %cst evictionPolicy = evict_first : tensor<64x64x!tt.ptr<bf16>> loc(#loc21)
46
+ %33 = arith.extf %32 : tensor<64x64xbf16> to tensor<64x64xf32> loc(#loc22)
47
+ %34 = arith.mulf %33, %33 : tensor<64x64xf32> loc(#loc23)
48
+ %35 = arith.addf %arg5, %34 : tensor<64x64xf32> loc(#loc24)
49
+ %36 = arith.select %31, %35, %arg5 : tensor<64x64xi1>, tensor<64x64xf32> loc(#loc25)
50
+ scf.yield %36 : tensor<64x64xf32> loc(#loc26)
51
+ } loc(#loc11)
52
+ %12 = "tt.reduce"(%11) <{axis = 1 : i32}> ({
53
+ ^bb0(%arg4: f32 loc(callsite(#loc1 at #loc28)), %arg5: f32 loc(callsite(#loc1 at #loc28))):
54
+ %16 = arith.addf %arg4, %arg5 : f32 loc(#loc37)
55
+ tt.reduce.return %16 : f32 loc(#loc34)
56
+ }) : (tensor<64x64xf32>) -> tensor<64xf32> loc(#loc34)
57
+ %13 = tt.expand_dims %12 {axis = 1 : i32} : tensor<64xf32> -> tensor<64x1xf32> loc(#loc30)
58
+ %14 = tt.splat %arg1 : !tt.ptr<f32> -> tensor<64x1x!tt.ptr<f32>> loc(#loc31)
59
+ %15 = tt.addptr %14, %5 : tensor<64x1x!tt.ptr<f32>>, tensor<64x1xi32> loc(#loc31)
60
+ tt.store %15, %13, %7 : tensor<64x1x!tt.ptr<f32>> loc(#loc32)
61
+ tt.return loc(#loc33)
62
+ } loc(#loc)
63
+ } loc(#loc)
64
+ #loc2 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":22:28)
65
+ #loc3 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":22:33)
66
+ #loc4 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":23:36)
67
+ #loc5 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":23:44)
68
+ #loc6 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":23:23)
69
+ #loc7 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":24:21)
70
+ #loc8 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":25:37)
71
+ #loc9 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":27:19)
72
+ #loc10 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":28:19)
73
+ #loc11 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":31:40)
74
+ #loc12 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":32:31)
75
+ #loc13 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":33:29)
76
+ #loc14 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:41)
77
+ #loc15 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:52)
78
+ #loc16 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:48)
79
+ #loc17 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:62)
80
+ #loc18 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:57)
81
+ #loc19 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:34)
82
+ #loc20 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:77)
83
+ #loc21 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:67)
84
+ #loc22 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:129)
85
+ #loc23 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":39:22)
86
+ #loc24 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":41:23)
87
+ #loc25 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":42:48)
88
+ #loc26 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":42:8)
89
+ #loc27 = loc("/home/x/hfenv/lib/python3.12/site-packages/triton/language/standard.py":286:36)
90
+ #loc29 = loc("/home/x/hfenv/lib/python3.12/site-packages/triton/language/standard.py":256:15)
91
+ #loc30 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":43:28)
92
+ #loc31 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":44:25)
93
+ #loc32 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":44:36)
94
+ #loc33 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":44:4)
95
+ #loc34 = loc(callsite(#loc27 at #loc28))
96
+ #loc36 = loc(callsite(#loc29 at #loc27))
97
+ #loc37 = loc(callsite(#loc36 at #loc28))
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/5TS244IYEIF4EECWF7IIP4LY7WII7Y4ZJDYZBCSZG4CRQWGLCERQ/__grp__triton_poi_fused_add_mul_sub_6.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"child_paths": {"triton_poi_fused_add_mul_sub_6.ttir": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/5TS244IYEIF4EECWF7IIP4LY7WII7Y4ZJDYZBCSZG4CRQWGLCERQ/triton_poi_fused_add_mul_sub_6.ttir", "triton_poi_fused_add_mul_sub_6.ttgir": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/5TS244IYEIF4EECWF7IIP4LY7WII7Y4ZJDYZBCSZG4CRQWGLCERQ/triton_poi_fused_add_mul_sub_6.ttgir", "triton_poi_fused_add_mul_sub_6.llir": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/5TS244IYEIF4EECWF7IIP4LY7WII7Y4ZJDYZBCSZG4CRQWGLCERQ/triton_poi_fused_add_mul_sub_6.llir", "triton_poi_fused_add_mul_sub_6.ptx": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/5TS244IYEIF4EECWF7IIP4LY7WII7Y4ZJDYZBCSZG4CRQWGLCERQ/triton_poi_fused_add_mul_sub_6.ptx", "triton_poi_fused_add_mul_sub_6.cubin": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/5TS244IYEIF4EECWF7IIP4LY7WII7Y4ZJDYZBCSZG4CRQWGLCERQ/triton_poi_fused_add_mul_sub_6.cubin", "triton_poi_fused_add_mul_sub_6.json": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/5TS244IYEIF4EECWF7IIP4LY7WII7Y4ZJDYZBCSZG4CRQWGLCERQ/triton_poi_fused_add_mul_sub_6.json"}}
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/5TS244IYEIF4EECWF7IIP4LY7WII7Y4ZJDYZBCSZG4CRQWGLCERQ/triton_poi_fused_add_mul_sub_6.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"hash": "ece5ae7118220bc210562fd087f178fd908fe39948f1908a5937051858cb1123", "target": {"backend": "cuda", "arch": 90, "warp_size": 32}, "num_warps": 4, "num_ctas": 1, "num_stages": 1, "num_buffers_warp_spec": 0, "num_consumer_groups": 0, "reg_dec_producer": 0, "reg_inc_consumer": 0, "maxnreg": null, "cluster_dims": [1, 1, 1], "ptx_version": null, "enable_fp_fusion": true, "launch_cooperative_grid": false, "supported_fp8_dtypes": ["fp8e4b15", "fp8e4nv", "fp8e5"], "deprecated_fp8_dtypes": ["fp8e4b15"], "default_dot_input_precision": "tf32", "allowed_dot_input_precisions": ["tf32", "tf32x3", "ieee"], "max_num_imprecise_acc_default": 1073741824, "extern_libs": [["libdevice", "/home/x/hfenv/lib/python3.12/site-packages/triton/backends/nvidia/lib/libdevice.10.bc"]], "debug": true, "backend_name": "cuda", "sanitize_overflow": false, "arch": "sm90", "triton_version": "3.3.1", "shared": 0, "tmem_size": 0, "global_scratch_size": 0, "global_scratch_align": 1, "name": "triton_poi_fused_add_mul_sub_6"}
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/5TS244IYEIF4EECWF7IIP4LY7WII7Y4ZJDYZBCSZG4CRQWGLCERQ/triton_poi_fused_add_mul_sub_6.ttgir ADDED
@@ -0,0 +1,196 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #blocked = #ttg.blocked<{sizePerThread = [8], threadsPerWarp = [32], warpsPerCTA = [4], order = [0]}>
2
+ #loc = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0)
3
+ module attributes {"ttg.num-ctas" = 1 : i32, "ttg.num-warps" = 4 : i32, ttg.target = "cuda:90", "ttg.threads-per-warp" = 32 : i32} {
4
+ tt.func public @triton_poi_fused_add_mul_sub_6(%arg0: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg1: !tt.ptr<f32> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg2: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg3: !tt.ptr<i64> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg4: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg5: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg6: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg7: i32 {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0)) attributes {noinline = false} {
5
+ %cst = arith.constant dense<1.280000e+02> : tensor<1024xf32, #blocked> loc(#loc1)
6
+ %cst_0 = arith.constant dense<9.99999997E-7> : tensor<1024xf32, #blocked> loc(#loc1)
7
+ %cst_1 = arith.constant dense<128> : tensor<1024xi64, #blocked> loc(#loc1)
8
+ %cst_2 = arith.constant dense<40960> : tensor<1024xi64, #blocked> loc(#loc1)
9
+ %cst_3 = arith.constant dense<0> : tensor<1024xi64, #blocked> loc(#loc1)
10
+ %c1024_i32 = arith.constant 1024 : i32 loc(#loc1)
11
+ %cst_4 = arith.constant dense<64> : tensor<1024xi32, #blocked> loc(#loc1)
12
+ %cst_5 = arith.constant dense<8> : tensor<1024xi32, #blocked> loc(#loc1)
13
+ %cst_6 = arith.constant dense<512> : tensor<1024xi32, #blocked> loc(#loc1)
14
+ %cst_7 = arith.constant dense<5120> : tensor<1024xi32, #blocked> loc(#loc1)
15
+ %cst_8 = arith.constant dense<128> : tensor<1024xi32, #blocked> loc(#loc1)
16
+ %cst_9 = arith.constant dense<7168> : tensor<1024xi32, #blocked> loc(#loc1)
17
+ %cst_10 = arith.constant dense<5184> : tensor<1024xi32, #blocked> loc(#loc1)
18
+ %cst_11 = arith.constant dense<true> : tensor<1024xi1, #blocked> loc(#loc1)
19
+ %0 = tt.get_program_id x : i32 loc(#loc2)
20
+ %1 = arith.muli %0, %c1024_i32 : i32 loc(#loc3)
21
+ %2 = tt.make_range {end = 1024 : i32, start = 0 : i32} : tensor<1024xi32, #blocked> loc(#loc4)
22
+ %3 = tt.splat %1 : i32 -> tensor<1024xi32, #blocked> loc(#loc5)
23
+ %4 = arith.addi %3, %2 : tensor<1024xi32, #blocked> loc(#loc5)
24
+ %5 = tt.splat %arg7 : i32 -> tensor<1024xi32, #blocked> loc(#loc6)
25
+ %6 = arith.cmpi slt, %4, %5 : tensor<1024xi32, #blocked> loc(#loc6)
26
+ %7 = arith.remsi %4, %cst_4 : tensor<1024xi32, #blocked> loc(#loc7)
27
+ %8 = arith.divsi %4, %cst_4 : tensor<1024xi32, #blocked> loc(#loc8)
28
+ %9 = arith.remsi %8, %cst_5 : tensor<1024xi32, #blocked> loc(#loc9)
29
+ %10 = arith.divsi %4, %cst_6 : tensor<1024xi32, #blocked> loc(#loc10)
30
+ %11 = arith.addi %7, %cst_7 : tensor<1024xi32, #blocked> loc(#loc11)
31
+ %12 = arith.muli %9, %cst_8 : tensor<1024xi32, #blocked> loc(#loc12)
32
+ %13 = arith.addi %11, %12 : tensor<1024xi32, #blocked> loc(#loc13)
33
+ %14 = arith.muli %10, %cst_9 : tensor<1024xi32, #blocked> loc(#loc14)
34
+ %15 = arith.addi %13, %14 : tensor<1024xi32, #blocked> loc(#loc15)
35
+ %16 = tt.splat %arg0 : !tt.ptr<bf16> -> tensor<1024x!tt.ptr<bf16>, #blocked> loc(#loc16)
36
+ %17 = tt.addptr %16, %15 : tensor<1024x!tt.ptr<bf16>, #blocked>, tensor<1024xi32, #blocked> loc(#loc16)
37
+ %18 = tt.load %17, %6 : tensor<1024x!tt.ptr<bf16>, #blocked> loc(#loc17)
38
+ %19 = arith.extf %18 : tensor<1024xbf16, #blocked> to tensor<1024xf32, #blocked> loc(#loc18)
39
+ %20 = tt.splat %arg1 : !tt.ptr<f32> -> tensor<1024x!tt.ptr<f32>, #blocked> loc(#loc19)
40
+ %21 = tt.addptr %20, %8 : tensor<1024x!tt.ptr<f32>, #blocked>, tensor<1024xi32, #blocked> loc(#loc19)
41
+ %22 = tt.load %21, %6 evictionPolicy = evict_last : tensor<1024x!tt.ptr<f32>, #blocked> loc(#loc20)
42
+ %23 = tt.splat %arg2 : !tt.ptr<bf16> -> tensor<1024x!tt.ptr<bf16>, #blocked> loc(#loc21)
43
+ %24 = tt.addptr %23, %7 : tensor<1024x!tt.ptr<bf16>, #blocked>, tensor<1024xi32, #blocked> loc(#loc21)
44
+ %25 = tt.load %24, %6 evictionPolicy = evict_last : tensor<1024x!tt.ptr<bf16>, #blocked> loc(#loc22)
45
+ %26 = arith.extf %25 : tensor<1024xbf16, #blocked> to tensor<1024xf32, #blocked> loc(#loc23)
46
+ %27 = tt.splat %arg3 : !tt.ptr<i64> -> tensor<1024x!tt.ptr<i64>, #blocked> loc(#loc24)
47
+ %28 = tt.addptr %27, %10 : tensor<1024x!tt.ptr<i64>, #blocked>, tensor<1024xi32, #blocked> loc(#loc24)
48
+ %29 = tt.load %28, %6 evictionPolicy = evict_last : tensor<1024x!tt.ptr<i64>, #blocked> loc(#loc25)
49
+ %30 = arith.addi %7, %cst_10 : tensor<1024xi32, #blocked> loc(#loc26)
50
+ %31 = arith.addi %30, %12 : tensor<1024xi32, #blocked> loc(#loc27)
51
+ %32 = arith.addi %7, %cst_4 : tensor<1024xi32, #blocked> loc(#loc28)
52
+ %33 = arith.divsi %32, %cst_8 : tensor<1024xi32, #blocked> loc(#loc29)
53
+ %34 = arith.muli %33, %cst_8 : tensor<1024xi32, #blocked> loc(#loc30)
54
+ %35 = arith.addi %31, %34 : tensor<1024xi32, #blocked> loc(#loc31)
55
+ %36 = arith.addi %35, %14 : tensor<1024xi32, #blocked> loc(#loc32)
56
+ %37 = tt.addptr %16, %36 : tensor<1024x!tt.ptr<bf16>, #blocked>, tensor<1024xi32, #blocked> loc(#loc33)
57
+ %38 = tt.load %37, %6 : tensor<1024x!tt.ptr<bf16>, #blocked> loc(#loc34)
58
+ %39 = arith.extf %38 : tensor<1024xbf16, #blocked> to tensor<1024xf32, #blocked> loc(#loc35)
59
+ %40 = arith.addi %8, %33 : tensor<1024xi32, #blocked> loc(#loc36)
60
+ %41 = tt.addptr %20, %40 : tensor<1024x!tt.ptr<f32>, #blocked>, tensor<1024xi32, #blocked> loc(#loc37)
61
+ %42 = tt.load %41, %6 evictionPolicy = evict_last : tensor<1024x!tt.ptr<f32>, #blocked> loc(#loc38)
62
+ %43 = tt.addptr %23, %32 : tensor<1024x!tt.ptr<bf16>, #blocked>, tensor<1024xi32, #blocked> loc(#loc39)
63
+ %44 = tt.load %43, %6 evictionPolicy = evict_last : tensor<1024x!tt.ptr<bf16>, #blocked> loc(#loc40)
64
+ %45 = arith.extf %44 : tensor<1024xbf16, #blocked> to tensor<1024xf32, #blocked> loc(#loc41)
65
+ %46 = arith.divf %22, %cst : tensor<1024xf32, #blocked> loc(#loc42)
66
+ %47 = arith.addf %46, %cst_0 : tensor<1024xf32, #blocked> loc(#loc43)
67
+ %48 = tt.extern_elementwise %47 {libname = "", libpath = "", pure = true, symbol = "__nv_rsqrtf"} : (tensor<1024xf32, #blocked>) -> tensor<1024xf32, #blocked> loc(#loc44)
68
+ %49 = arith.mulf %19, %48 : tensor<1024xf32, #blocked> loc(#loc45)
69
+ %50 = arith.mulf %49, %26 : tensor<1024xf32, #blocked> loc(#loc46)
70
+ %51 = arith.addi %29, %cst_2 : tensor<1024xi64, #blocked> loc(#loc47)
71
+ %52 = arith.cmpi slt, %29, %cst_3 : tensor<1024xi64, #blocked> loc(#loc48)
72
+ %53 = arith.select %52, %51, %29 : tensor<1024xi1, #blocked>, tensor<1024xi64, #blocked> loc(#loc49)
73
+ %54 = arith.cmpi sge, %53, %cst_3 : tensor<1024xi64, #blocked> loc(#loc50)
74
+ %55 = arith.cmpi slt, %53, %cst_2 : tensor<1024xi64, #blocked> loc(#loc51)
75
+ %56 = arith.andi %54, %55 : tensor<1024xi1, #blocked> loc(#loc52)
76
+ %57 = arith.xori %6, %cst_11 : tensor<1024xi1, #blocked> loc(#loc53)
77
+ %58 = arith.ori %56, %57 : tensor<1024xi1, #blocked> loc(#loc54)
78
+ tt.assert %58, "index out of bounds: 0 <= tmp16 < 40960" : tensor<1024xi1, #blocked> loc(#loc55)
79
+ %59 = arith.muli %53, %cst_1 : tensor<1024xi64, #blocked> loc(#loc56)
80
+ %60 = arith.extsi %7 : tensor<1024xi32, #blocked> to tensor<1024xi64, #blocked> loc(#loc57)
81
+ %61 = arith.addi %60, %59 : tensor<1024xi64, #blocked> loc(#loc57)
82
+ %62 = tt.splat %arg4 : !tt.ptr<bf16> -> tensor<1024x!tt.ptr<bf16>, #blocked> loc(#loc58)
83
+ %63 = tt.addptr %62, %61 : tensor<1024x!tt.ptr<bf16>, #blocked>, tensor<1024xi64, #blocked> loc(#loc58)
84
+ %64 = tt.load %63, %6 : tensor<1024x!tt.ptr<bf16>, #blocked> loc(#loc59)
85
+ %65 = arith.extf %64 : tensor<1024xbf16, #blocked> to tensor<1024xf32, #blocked> loc(#loc60)
86
+ %66 = arith.mulf %50, %65 : tensor<1024xf32, #blocked> loc(#loc61)
87
+ %67 = arith.divf %42, %cst : tensor<1024xf32, #blocked> loc(#loc62)
88
+ %68 = arith.addf %67, %cst_0 : tensor<1024xf32, #blocked> loc(#loc63)
89
+ %69 = tt.extern_elementwise %68 {libname = "", libpath = "", pure = true, symbol = "__nv_rsqrtf"} : (tensor<1024xf32, #blocked>) -> tensor<1024xf32, #blocked> loc(#loc64)
90
+ %70 = arith.mulf %39, %69 : tensor<1024xf32, #blocked> loc(#loc65)
91
+ %71 = arith.mulf %70, %45 : tensor<1024xf32, #blocked> loc(#loc66)
92
+ %72 = arith.extsi %32 : tensor<1024xi32, #blocked> to tensor<1024xi64, #blocked> loc(#loc67)
93
+ %73 = arith.addi %72, %59 : tensor<1024xi64, #blocked> loc(#loc67)
94
+ %74 = tt.addptr %62, %73 : tensor<1024x!tt.ptr<bf16>, #blocked>, tensor<1024xi64, #blocked> loc(#loc68)
95
+ %75 = tt.load %74, %6 : tensor<1024x!tt.ptr<bf16>, #blocked> loc(#loc69)
96
+ %76 = arith.extf %75 : tensor<1024xbf16, #blocked> to tensor<1024xf32, #blocked> loc(#loc70)
97
+ %77 = arith.mulf %71, %76 : tensor<1024xf32, #blocked> loc(#loc71)
98
+ %78 = arith.subf %66, %77 : tensor<1024xf32, #blocked> loc(#loc72)
99
+ %79 = arith.mulf %71, %65 : tensor<1024xf32, #blocked> loc(#loc73)
100
+ %80 = arith.mulf %50, %76 : tensor<1024xf32, #blocked> loc(#loc74)
101
+ %81 = arith.addf %79, %80 : tensor<1024xf32, #blocked> loc(#loc75)
102
+ %82 = arith.muli %8, %cst_8 : tensor<1024xi32, #blocked> loc(#loc76)
103
+ %83 = arith.addi %7, %82 : tensor<1024xi32, #blocked> loc(#loc77)
104
+ %84 = tt.splat %arg5 : !tt.ptr<bf16> -> tensor<1024x!tt.ptr<bf16>, #blocked> loc(#loc78)
105
+ %85 = tt.addptr %84, %83 : tensor<1024x!tt.ptr<bf16>, #blocked>, tensor<1024xi32, #blocked> loc(#loc78)
106
+ %86 = arith.truncf %78 : tensor<1024xf32, #blocked> to tensor<1024xbf16, #blocked> loc(#loc79)
107
+ tt.store %85, %86, %6 : tensor<1024x!tt.ptr<bf16>, #blocked> loc(#loc79)
108
+ %87 = tt.splat %arg6 : !tt.ptr<bf16> -> tensor<1024x!tt.ptr<bf16>, #blocked> loc(#loc80)
109
+ %88 = tt.addptr %87, %83 : tensor<1024x!tt.ptr<bf16>, #blocked>, tensor<1024xi32, #blocked> loc(#loc80)
110
+ %89 = arith.truncf %81 : tensor<1024xf32, #blocked> to tensor<1024xbf16, #blocked> loc(#loc81)
111
+ tt.store %88, %89, %6 : tensor<1024x!tt.ptr<bf16>, #blocked> loc(#loc81)
112
+ tt.return loc(#loc82)
113
+ } loc(#loc)
114
+ } loc(#loc)
115
+ #loc1 = loc(unknown)
116
+ #loc2 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":19:28)
117
+ #loc3 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":19:33)
118
+ #loc4 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":20:36)
119
+ #loc5 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":20:23)
120
+ #loc6 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":21:21)
121
+ #loc7 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":22:19)
122
+ #loc8 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":23:21)
123
+ #loc9 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":23:27)
124
+ #loc10 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":24:19)
125
+ #loc11 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:37)
126
+ #loc12 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:46)
127
+ #loc13 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:42)
128
+ #loc14 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:56)
129
+ #loc15 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:51)
130
+ #loc16 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:30)
131
+ #loc17 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:61)
132
+ #loc18 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:71)
133
+ #loc19 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":27:30)
134
+ #loc20 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":27:35)
135
+ #loc21 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":28:31)
136
+ #loc22 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":28:36)
137
+ #loc23 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":28:76)
138
+ #loc24 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":29:31)
139
+ #loc25 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":29:36)
140
+ #loc26 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:38)
141
+ #loc27 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:43)
142
+ #loc28 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:63)
143
+ #loc29 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:70)
144
+ #loc30 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:57)
145
+ #loc31 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:52)
146
+ #loc32 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:77)
147
+ #loc33 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:31)
148
+ #loc34 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:87)
149
+ #loc35 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:97)
150
+ #loc36 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":31:37)
151
+ #loc37 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":31:31)
152
+ #loc38 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":31:57)
153
+ #loc39 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":32:31)
154
+ #loc40 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":32:41)
155
+ #loc41 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":32:81)
156
+ #loc42 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":35:19)
157
+ #loc43 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":37:18)
158
+ #loc44 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":38:27)
159
+ #loc45 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":39:18)
160
+ #loc46 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":41:19)
161
+ #loc47 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":43:20)
162
+ #loc48 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":44:20)
163
+ #loc49 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":45:35)
164
+ #loc50 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:28)
165
+ #loc51 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:46)
166
+ #loc52 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:38)
167
+ #loc53 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:58)
168
+ #loc54 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:56)
169
+ #loc55 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:66)
170
+ #loc56 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:40)
171
+ #loc57 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:36)
172
+ #loc58 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:31)
173
+ #loc59 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:48)
174
+ #loc60 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:58)
175
+ #loc61 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":48:20)
176
+ #loc62 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":50:21)
177
+ #loc63 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":51:20)
178
+ #loc64 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":52:28)
179
+ #loc65 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":53:20)
180
+ #loc66 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":55:20)
181
+ #loc67 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":56:41)
182
+ #loc68 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":56:31)
183
+ #loc69 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":56:53)
184
+ #loc70 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":56:63)
185
+ #loc71 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":57:20)
186
+ #loc72 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":58:20)
187
+ #loc73 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":59:20)
188
+ #loc74 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":60:20)
189
+ #loc75 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":61:20)
190
+ #loc76 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":62:34)
191
+ #loc77 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":62:30)
192
+ #loc78 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":62:25)
193
+ #loc79 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":62:46)
194
+ #loc80 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":63:25)
195
+ #loc81 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":63:46)
196
+ #loc82 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":63:4)
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/5TS244IYEIF4EECWF7IIP4LY7WII7Y4ZJDYZBCSZG4CRQWGLCERQ/triton_poi_fused_add_mul_sub_6.ttir ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #loc = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0)
2
+ module {
3
+ tt.func public @triton_poi_fused_add_mul_sub_6(%arg0: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg1: !tt.ptr<f32> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg2: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg3: !tt.ptr<i64> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg4: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg5: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg6: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg7: i32 {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0)) attributes {noinline = false} {
4
+ %cst = arith.constant dense<0> : tensor<1024xi64> loc(#loc1)
5
+ %cst_0 = arith.constant dense<40960> : tensor<1024xi64> loc(#loc1)
6
+ %cst_1 = arith.constant dense<128> : tensor<1024xi64> loc(#loc1)
7
+ %cst_2 = arith.constant dense<true> : tensor<1024xi1> loc(#loc1)
8
+ %cst_3 = arith.constant dense<9.99999997E-7> : tensor<1024xf32> loc(#loc1)
9
+ %cst_4 = arith.constant dense<1.280000e+02> : tensor<1024xf32> loc(#loc1)
10
+ %cst_5 = arith.constant dense<5184> : tensor<1024xi32> loc(#loc1)
11
+ %cst_6 = arith.constant dense<7168> : tensor<1024xi32> loc(#loc1)
12
+ %cst_7 = arith.constant dense<128> : tensor<1024xi32> loc(#loc1)
13
+ %cst_8 = arith.constant dense<5120> : tensor<1024xi32> loc(#loc1)
14
+ %cst_9 = arith.constant dense<512> : tensor<1024xi32> loc(#loc1)
15
+ %cst_10 = arith.constant dense<8> : tensor<1024xi32> loc(#loc1)
16
+ %cst_11 = arith.constant dense<64> : tensor<1024xi32> loc(#loc1)
17
+ %c1024_i32 = arith.constant 1024 : i32 loc(#loc1)
18
+ %0 = tt.get_program_id x : i32 loc(#loc2)
19
+ %1 = arith.muli %0, %c1024_i32 : i32 loc(#loc3)
20
+ %2 = tt.make_range {end = 1024 : i32, start = 0 : i32} : tensor<1024xi32> loc(#loc4)
21
+ %3 = tt.splat %1 : i32 -> tensor<1024xi32> loc(#loc5)
22
+ %4 = arith.addi %3, %2 : tensor<1024xi32> loc(#loc5)
23
+ %5 = tt.splat %arg7 : i32 -> tensor<1024xi32> loc(#loc6)
24
+ %6 = arith.cmpi slt, %4, %5 : tensor<1024xi32> loc(#loc6)
25
+ %7 = arith.remsi %4, %cst_11 : tensor<1024xi32> loc(#loc7)
26
+ %8 = arith.divsi %4, %cst_11 : tensor<1024xi32> loc(#loc8)
27
+ %9 = arith.remsi %8, %cst_10 : tensor<1024xi32> loc(#loc9)
28
+ %10 = arith.divsi %4, %cst_9 : tensor<1024xi32> loc(#loc10)
29
+ %11 = arith.addi %7, %cst_8 : tensor<1024xi32> loc(#loc11)
30
+ %12 = arith.muli %9, %cst_7 : tensor<1024xi32> loc(#loc12)
31
+ %13 = arith.addi %11, %12 : tensor<1024xi32> loc(#loc13)
32
+ %14 = arith.muli %10, %cst_6 : tensor<1024xi32> loc(#loc14)
33
+ %15 = arith.addi %13, %14 : tensor<1024xi32> loc(#loc15)
34
+ %16 = tt.splat %arg0 : !tt.ptr<bf16> -> tensor<1024x!tt.ptr<bf16>> loc(#loc16)
35
+ %17 = tt.addptr %16, %15 : tensor<1024x!tt.ptr<bf16>>, tensor<1024xi32> loc(#loc16)
36
+ %18 = tt.load %17, %6 : tensor<1024x!tt.ptr<bf16>> loc(#loc17)
37
+ %19 = arith.extf %18 : tensor<1024xbf16> to tensor<1024xf32> loc(#loc18)
38
+ %20 = tt.splat %arg1 : !tt.ptr<f32> -> tensor<1024x!tt.ptr<f32>> loc(#loc19)
39
+ %21 = tt.addptr %20, %8 : tensor<1024x!tt.ptr<f32>>, tensor<1024xi32> loc(#loc19)
40
+ %22 = tt.load %21, %6 evictionPolicy = evict_last : tensor<1024x!tt.ptr<f32>> loc(#loc20)
41
+ %23 = tt.splat %arg2 : !tt.ptr<bf16> -> tensor<1024x!tt.ptr<bf16>> loc(#loc21)
42
+ %24 = tt.addptr %23, %7 : tensor<1024x!tt.ptr<bf16>>, tensor<1024xi32> loc(#loc21)
43
+ %25 = tt.load %24, %6 evictionPolicy = evict_last : tensor<1024x!tt.ptr<bf16>> loc(#loc22)
44
+ %26 = arith.extf %25 : tensor<1024xbf16> to tensor<1024xf32> loc(#loc23)
45
+ %27 = tt.splat %arg3 : !tt.ptr<i64> -> tensor<1024x!tt.ptr<i64>> loc(#loc24)
46
+ %28 = tt.addptr %27, %10 : tensor<1024x!tt.ptr<i64>>, tensor<1024xi32> loc(#loc24)
47
+ %29 = tt.load %28, %6 evictionPolicy = evict_last : tensor<1024x!tt.ptr<i64>> loc(#loc25)
48
+ %30 = arith.addi %7, %cst_5 : tensor<1024xi32> loc(#loc26)
49
+ %31 = arith.addi %30, %12 : tensor<1024xi32> loc(#loc27)
50
+ %32 = arith.addi %7, %cst_11 : tensor<1024xi32> loc(#loc28)
51
+ %33 = arith.divsi %32, %cst_7 : tensor<1024xi32> loc(#loc29)
52
+ %34 = arith.muli %33, %cst_7 : tensor<1024xi32> loc(#loc30)
53
+ %35 = arith.addi %31, %34 : tensor<1024xi32> loc(#loc31)
54
+ %36 = arith.addi %35, %14 : tensor<1024xi32> loc(#loc32)
55
+ %37 = tt.addptr %16, %36 : tensor<1024x!tt.ptr<bf16>>, tensor<1024xi32> loc(#loc33)
56
+ %38 = tt.load %37, %6 : tensor<1024x!tt.ptr<bf16>> loc(#loc34)
57
+ %39 = arith.extf %38 : tensor<1024xbf16> to tensor<1024xf32> loc(#loc35)
58
+ %40 = arith.addi %8, %33 : tensor<1024xi32> loc(#loc36)
59
+ %41 = tt.addptr %20, %40 : tensor<1024x!tt.ptr<f32>>, tensor<1024xi32> loc(#loc37)
60
+ %42 = tt.load %41, %6 evictionPolicy = evict_last : tensor<1024x!tt.ptr<f32>> loc(#loc38)
61
+ %43 = tt.addptr %23, %32 : tensor<1024x!tt.ptr<bf16>>, tensor<1024xi32> loc(#loc39)
62
+ %44 = tt.load %43, %6 evictionPolicy = evict_last : tensor<1024x!tt.ptr<bf16>> loc(#loc40)
63
+ %45 = arith.extf %44 : tensor<1024xbf16> to tensor<1024xf32> loc(#loc41)
64
+ %46 = arith.divf %22, %cst_4 : tensor<1024xf32> loc(#loc42)
65
+ %47 = arith.addf %46, %cst_3 : tensor<1024xf32> loc(#loc43)
66
+ %48 = tt.extern_elementwise %47 {libname = "", libpath = "", pure = true, symbol = "__nv_rsqrtf"} : (tensor<1024xf32>) -> tensor<1024xf32> loc(#loc44)
67
+ %49 = arith.mulf %19, %48 : tensor<1024xf32> loc(#loc45)
68
+ %50 = arith.mulf %49, %26 : tensor<1024xf32> loc(#loc46)
69
+ %51 = arith.addi %29, %cst_0 : tensor<1024xi64> loc(#loc47)
70
+ %52 = arith.cmpi slt, %29, %cst : tensor<1024xi64> loc(#loc48)
71
+ %53 = arith.select %52, %51, %29 : tensor<1024xi1>, tensor<1024xi64> loc(#loc49)
72
+ %54 = arith.cmpi sge, %53, %cst : tensor<1024xi64> loc(#loc50)
73
+ %55 = arith.cmpi slt, %53, %cst_0 : tensor<1024xi64> loc(#loc51)
74
+ %56 = arith.andi %54, %55 : tensor<1024xi1> loc(#loc52)
75
+ %57 = arith.xori %6, %cst_2 : tensor<1024xi1> loc(#loc53)
76
+ %58 = arith.ori %56, %57 : tensor<1024xi1> loc(#loc54)
77
+ tt.assert %58, "index out of bounds: 0 <= tmp16 < 40960" : tensor<1024xi1> loc(#loc55)
78
+ %59 = arith.muli %53, %cst_1 : tensor<1024xi64> loc(#loc56)
79
+ %60 = arith.extsi %7 : tensor<1024xi32> to tensor<1024xi64> loc(#loc57)
80
+ %61 = arith.addi %60, %59 : tensor<1024xi64> loc(#loc57)
81
+ %62 = tt.splat %arg4 : !tt.ptr<bf16> -> tensor<1024x!tt.ptr<bf16>> loc(#loc58)
82
+ %63 = tt.addptr %62, %61 : tensor<1024x!tt.ptr<bf16>>, tensor<1024xi64> loc(#loc58)
83
+ %64 = tt.load %63, %6 : tensor<1024x!tt.ptr<bf16>> loc(#loc59)
84
+ %65 = arith.extf %64 : tensor<1024xbf16> to tensor<1024xf32> loc(#loc60)
85
+ %66 = arith.mulf %50, %65 : tensor<1024xf32> loc(#loc61)
86
+ %67 = arith.divf %42, %cst_4 : tensor<1024xf32> loc(#loc62)
87
+ %68 = arith.addf %67, %cst_3 : tensor<1024xf32> loc(#loc63)
88
+ %69 = tt.extern_elementwise %68 {libname = "", libpath = "", pure = true, symbol = "__nv_rsqrtf"} : (tensor<1024xf32>) -> tensor<1024xf32> loc(#loc64)
89
+ %70 = arith.mulf %39, %69 : tensor<1024xf32> loc(#loc65)
90
+ %71 = arith.mulf %70, %45 : tensor<1024xf32> loc(#loc66)
91
+ %72 = arith.extsi %32 : tensor<1024xi32> to tensor<1024xi64> loc(#loc67)
92
+ %73 = arith.addi %72, %59 : tensor<1024xi64> loc(#loc67)
93
+ %74 = tt.addptr %62, %73 : tensor<1024x!tt.ptr<bf16>>, tensor<1024xi64> loc(#loc68)
94
+ %75 = tt.load %74, %6 : tensor<1024x!tt.ptr<bf16>> loc(#loc69)
95
+ %76 = arith.extf %75 : tensor<1024xbf16> to tensor<1024xf32> loc(#loc70)
96
+ %77 = arith.mulf %71, %76 : tensor<1024xf32> loc(#loc71)
97
+ %78 = arith.subf %66, %77 : tensor<1024xf32> loc(#loc72)
98
+ %79 = arith.mulf %71, %65 : tensor<1024xf32> loc(#loc73)
99
+ %80 = arith.mulf %50, %76 : tensor<1024xf32> loc(#loc74)
100
+ %81 = arith.addf %79, %80 : tensor<1024xf32> loc(#loc75)
101
+ %82 = arith.muli %8, %cst_7 : tensor<1024xi32> loc(#loc76)
102
+ %83 = arith.addi %7, %82 : tensor<1024xi32> loc(#loc77)
103
+ %84 = tt.splat %arg5 : !tt.ptr<bf16> -> tensor<1024x!tt.ptr<bf16>> loc(#loc78)
104
+ %85 = tt.addptr %84, %83 : tensor<1024x!tt.ptr<bf16>>, tensor<1024xi32> loc(#loc78)
105
+ %86 = arith.truncf %78 : tensor<1024xf32> to tensor<1024xbf16> loc(#loc79)
106
+ tt.store %85, %86, %6 : tensor<1024x!tt.ptr<bf16>> loc(#loc79)
107
+ %87 = tt.splat %arg6 : !tt.ptr<bf16> -> tensor<1024x!tt.ptr<bf16>> loc(#loc80)
108
+ %88 = tt.addptr %87, %83 : tensor<1024x!tt.ptr<bf16>>, tensor<1024xi32> loc(#loc80)
109
+ %89 = arith.truncf %81 : tensor<1024xf32> to tensor<1024xbf16> loc(#loc81)
110
+ tt.store %88, %89, %6 : tensor<1024x!tt.ptr<bf16>> loc(#loc81)
111
+ tt.return loc(#loc82)
112
+ } loc(#loc)
113
+ } loc(#loc)
114
+ #loc1 = loc(unknown)
115
+ #loc2 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":19:28)
116
+ #loc3 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":19:33)
117
+ #loc4 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":20:36)
118
+ #loc5 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":20:23)
119
+ #loc6 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":21:21)
120
+ #loc7 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":22:19)
121
+ #loc8 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":23:21)
122
+ #loc9 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":23:27)
123
+ #loc10 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":24:19)
124
+ #loc11 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:37)
125
+ #loc12 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:46)
126
+ #loc13 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:42)
127
+ #loc14 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:56)
128
+ #loc15 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:51)
129
+ #loc16 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:30)
130
+ #loc17 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:61)
131
+ #loc18 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:71)
132
+ #loc19 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":27:30)
133
+ #loc20 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":27:35)
134
+ #loc21 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":28:31)
135
+ #loc22 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":28:36)
136
+ #loc23 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":28:76)
137
+ #loc24 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":29:31)
138
+ #loc25 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":29:36)
139
+ #loc26 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:38)
140
+ #loc27 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:43)
141
+ #loc28 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:63)
142
+ #loc29 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:70)
143
+ #loc30 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:57)
144
+ #loc31 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:52)
145
+ #loc32 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:77)
146
+ #loc33 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:31)
147
+ #loc34 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:87)
148
+ #loc35 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:97)
149
+ #loc36 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":31:37)
150
+ #loc37 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":31:31)
151
+ #loc38 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":31:57)
152
+ #loc39 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":32:31)
153
+ #loc40 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":32:41)
154
+ #loc41 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":32:81)
155
+ #loc42 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":35:19)
156
+ #loc43 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":37:18)
157
+ #loc44 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":38:27)
158
+ #loc45 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":39:18)
159
+ #loc46 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":41:19)
160
+ #loc47 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":43:20)
161
+ #loc48 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":44:20)
162
+ #loc49 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":45:35)
163
+ #loc50 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:28)
164
+ #loc51 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:46)
165
+ #loc52 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:38)
166
+ #loc53 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:58)
167
+ #loc54 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:56)
168
+ #loc55 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:66)
169
+ #loc56 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:40)
170
+ #loc57 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:36)
171
+ #loc58 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:31)
172
+ #loc59 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:48)
173
+ #loc60 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:58)
174
+ #loc61 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":48:20)
175
+ #loc62 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":50:21)
176
+ #loc63 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":51:20)
177
+ #loc64 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":52:28)
178
+ #loc65 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":53:20)
179
+ #loc66 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":55:20)
180
+ #loc67 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":56:41)
181
+ #loc68 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":56:31)
182
+ #loc69 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":56:53)
183
+ #loc70 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":56:63)
184
+ #loc71 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":57:20)
185
+ #loc72 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":58:20)
186
+ #loc73 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":59:20)
187
+ #loc74 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":60:20)
188
+ #loc75 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":61:20)
189
+ #loc76 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":62:34)
190
+ #loc77 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":62:30)
191
+ #loc78 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":62:25)
192
+ #loc79 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":62:46)
193
+ #loc80 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":63:25)
194
+ #loc81 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":63:46)
195
+ #loc82 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":63:4)
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/7HDVHUPTEOX4FHYEQOTFHD5U573A7K7C62TYWKVZQQZ4QTR7356A/triton_red_fused__to_copy_mean_pow_4.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"hash": "f9c753d1f323afc29f0483a6538fb4eff60fabe2f6a78b2ab98433c84e3fdf7c", "target": {"backend": "cuda", "arch": 90, "warp_size": 32}, "num_warps": 8, "num_ctas": 1, "num_stages": 1, "num_buffers_warp_spec": 0, "num_consumer_groups": 0, "reg_dec_producer": 0, "reg_inc_consumer": 0, "maxnreg": null, "cluster_dims": [1, 1, 1], "ptx_version": null, "enable_fp_fusion": true, "launch_cooperative_grid": false, "supported_fp8_dtypes": ["fp8e4b15", "fp8e4nv", "fp8e5"], "deprecated_fp8_dtypes": ["fp8e4b15"], "default_dot_input_precision": "tf32", "allowed_dot_input_precisions": ["tf32", "tf32x3", "ieee"], "max_num_imprecise_acc_default": 1073741824, "extern_libs": [["libdevice", "/home/x/hfenv/lib/python3.12/site-packages/triton/backends/nvidia/lib/libdevice.10.bc"]], "debug": true, "backend_name": "cuda", "sanitize_overflow": false, "arch": "sm90", "triton_version": "3.3.1", "shared": 32, "tmem_size": 0, "global_scratch_size": 0, "global_scratch_align": 1, "name": "triton_red_fused__to_copy_mean_pow_4"}
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/7HDVHUPTEOX4FHYEQOTFHD5U573A7K7C62TYWKVZQQZ4QTR7356A/triton_red_fused__to_copy_mean_pow_4.llir ADDED
@@ -0,0 +1,143 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ; ModuleID = 'LLVMDialectModule'
2
+ source_filename = "LLVMDialectModule"
3
+ target datalayout = "e-p3:32:32-p4:32:32-p5:32:32-p6:32:32-i64:64-i128:128-v16:16-v32:32-n16:32:64"
4
+
5
+ @global_smem = external addrspace(3) global [0 x i8], align 16
6
+
7
+ define ptx_kernel void @triton_red_fused__to_copy_mean_pow_4(ptr addrspace(1) %0, ptr addrspace(1) %1, i32 %2, i32 %3, ptr addrspace(1) readnone captures(none) %4) local_unnamed_addr !dbg !6 {
8
+ %6 = tail call i32 @llvm.nvvm.read.ptx.sreg.ctaid.x(), !dbg !9
9
+ %7 = shl i32 %6, 3, !dbg !10
10
+ %8 = tail call i32 @llvm.nvvm.read.ptx.sreg.tid.x(), !dbg !11
11
+ %9 = lshr i32 %8, 5, !dbg !11
12
+ %10 = and i32 %9, 7, !dbg !11
13
+ %11 = and i32 %8, 7, !dbg !11
14
+ %12 = or disjoint i32 %10, %7, !dbg !12
15
+ %13 = or disjoint i32 %7, %11, !dbg !12
16
+ %14 = icmp slt i32 %12, %2, !dbg !13
17
+ %15 = icmp slt i32 %13, %2, !dbg !13
18
+ %16 = shl i32 %8, 2, !dbg !14
19
+ %17 = and i32 %16, 124, !dbg !14
20
+ %18 = sdiv i32 %12, 8, !dbg !15
21
+ %19 = mul i32 %18, 8, !dbg !16
22
+ %.decomposed = sub i32 %12, %19, !dbg !16
23
+ %20 = or disjoint i32 %17, 5120, !dbg !17
24
+ %21 = shl nsw i32 %.decomposed, 7, !dbg !18
25
+ %22 = add nsw i32 %20, %21, !dbg !19
26
+ %23 = mul i32 %18, 7168, !dbg !20
27
+ %24 = add i32 %22, %23, !dbg !21
28
+ %25 = sext i32 %24 to i64, !dbg !22
29
+ %26 = getelementptr bfloat, ptr addrspace(1) %0, i64 %25, !dbg !22
30
+ %27 = tail call { i32, i32 } asm sideeffect "mov.u32 $0, $2;\0A\09mov.u32 $1, $3;\0A\09@$5 ld.global.L1::evict_first.v2.b32 { $0, $1 }, [ $4 + 0 ];", "=r,=r,r,r,l,b"(i32 0, i32 0, ptr addrspace(1) %26, i1 %14) #3, !dbg !23
31
+ %28 = extractvalue { i32, i32 } %27, 0, !dbg !23
32
+ %29 = bitcast i32 %28 to <2 x bfloat>, !dbg !23
33
+ %30 = extractvalue { i32, i32 } %27, 1, !dbg !23
34
+ %31 = bitcast i32 %30 to <2 x bfloat>, !dbg !23
35
+ %32 = extractelement <2 x bfloat> %29, i64 0, !dbg !23
36
+ %33 = extractelement <2 x bfloat> %29, i64 1, !dbg !23
37
+ %34 = extractelement <2 x bfloat> %31, i64 0, !dbg !23
38
+ %35 = extractelement <2 x bfloat> %31, i64 1, !dbg !23
39
+ %36 = fpext bfloat %32 to float, !dbg !24
40
+ %37 = fpext bfloat %33 to float, !dbg !24
41
+ %38 = fpext bfloat %34 to float, !dbg !24
42
+ %39 = fpext bfloat %35 to float, !dbg !24
43
+ %40 = fmul float %36, %36, !dbg !25
44
+ %41 = fmul float %37, %37, !dbg !25
45
+ %42 = fmul float %38, %38, !dbg !25
46
+ %43 = fmul float %39, %39, !dbg !25
47
+ %44 = fadd float %40, %41, !dbg !26
48
+ %45 = fadd float %42, %44, !dbg !26
49
+ %46 = fadd float %43, %45, !dbg !26
50
+ %47 = select i1 %14, float %46, float 0.000000e+00, !dbg !26
51
+ %48 = bitcast float %47 to i32, !dbg !31
52
+ %49 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %48, i32 16, i32 31), !dbg !31
53
+ %50 = bitcast i32 %49 to float, !dbg !31
54
+ %51 = fadd float %47, %50, !dbg !26
55
+ %52 = bitcast float %51 to i32, !dbg !31
56
+ %53 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %52, i32 8, i32 31), !dbg !31
57
+ %54 = bitcast i32 %53 to float, !dbg !31
58
+ %55 = fadd float %51, %54, !dbg !26
59
+ %56 = bitcast float %55 to i32, !dbg !31
60
+ %57 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %56, i32 4, i32 31), !dbg !31
61
+ %58 = bitcast i32 %57 to float, !dbg !31
62
+ %59 = fadd float %55, %58, !dbg !26
63
+ %60 = bitcast float %59 to i32, !dbg !31
64
+ %61 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %60, i32 2, i32 31), !dbg !31
65
+ %62 = bitcast i32 %61 to float, !dbg !31
66
+ %63 = fadd float %59, %62, !dbg !26
67
+ %64 = bitcast float %63 to i32, !dbg !31
68
+ %65 = tail call i32 @llvm.nvvm.shfl.sync.bfly.i32(i32 -1, i32 %64, i32 1, i32 31), !dbg !31
69
+ %66 = bitcast i32 %65 to float, !dbg !31
70
+ %67 = fadd float %63, %66, !dbg !26
71
+ %68 = getelementptr inbounds nuw float, ptr addrspace(3) @global_smem, i32 %10, !dbg !32
72
+ %69 = bitcast float %67 to <1 x i32>, !dbg !32
73
+ tail call void asm sideeffect "@$2 st.shared.b32 [ $0 + 0 ], $1;", "r,r,b"(ptr addrspace(3) %68, <1 x i32> %69, i1 true) #3, !dbg !32
74
+ tail call void @llvm.nvvm.barrier0(), !dbg !32
75
+ %70 = getelementptr inbounds nuw float, ptr addrspace(3) @global_smem, i32 %11, !dbg !32
76
+ %71 = load i32, ptr addrspace(3) %70, align 4, !dbg !32
77
+ %72 = sext i32 %13 to i64, !dbg !33
78
+ %73 = getelementptr float, ptr addrspace(1) %1, i64 %72, !dbg !33
79
+ %74 = and i32 %8, 248, !dbg !34
80
+ %75 = icmp eq i32 %74, 0, !dbg !34
81
+ %76 = and i1 %75, %15, !dbg !34
82
+ tail call void asm sideeffect "@$2 st.global.b32 [ $1 + 0 ], { $0 };", "r,l,b"(i32 %71, ptr addrspace(1) %73, i1 %76) #3, !dbg !34
83
+ ret void, !dbg !35
84
+ }
85
+
86
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none)
87
+ declare noundef i32 @llvm.nvvm.read.ptx.sreg.ctaid.x() #0
88
+
89
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none)
90
+ declare noundef i32 @llvm.nvvm.read.ptx.sreg.tid.x() #0
91
+
92
+ ; Function Attrs: convergent nocallback nounwind memory(inaccessiblemem: readwrite)
93
+ declare i32 @llvm.nvvm.shfl.sync.bfly.i32(i32, i32, i32, i32) #1
94
+
95
+ ; Function Attrs: convergent nocallback nounwind
96
+ declare void @llvm.nvvm.barrier0() #2
97
+
98
+ attributes #0 = { mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none) }
99
+ attributes #1 = { convergent nocallback nounwind memory(inaccessiblemem: readwrite) }
100
+ attributes #2 = { convergent nocallback nounwind }
101
+ attributes #3 = { nounwind }
102
+
103
+ !llvm.module.flags = !{!0, !1}
104
+ !llvm.dbg.cu = !{!2}
105
+ !nvvm.annotations = !{!4}
106
+ !llvm.ident = !{!5}
107
+
108
+ !0 = !{i32 2, !"Debug Info Version", i32 3}
109
+ !1 = !{i32 4, !"nvvm-reflect-ftz", i32 1}
110
+ !2 = distinct !DICompileUnit(language: DW_LANG_C, file: !3, producer: "triton", isOptimized: true, runtimeVersion: 0, emissionKind: LineTablesOnly)
111
+ !3 = !DIFile(filename: "ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py", directory: "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty")
112
+ !4 = !{ptr @triton_red_fused__to_copy_mean_pow_4, !"reqntidx", i32 256}
113
+ !5 = !{!"clang version 3.8.0 (tags/RELEASE_380/final)"}
114
+ !6 = distinct !DISubprogram(name: "triton_red_fused__to_copy_mean_pow_4", linkageName: "triton_red_fused__to_copy_mean_pow_4", scope: !3, file: !3, line: 18, type: !7, scopeLine: 18, spFlags: DISPFlagDefinition | DISPFlagOptimized, unit: !2)
115
+ !7 = !DISubroutineType(cc: DW_CC_normal, types: !8)
116
+ !8 = !{}
117
+ !9 = !DILocation(line: 22, column: 28, scope: !6)
118
+ !10 = !DILocation(line: 22, column: 33, scope: !6)
119
+ !11 = !DILocation(line: 23, column: 44, scope: !6)
120
+ !12 = !DILocation(line: 23, column: 23, scope: !6)
121
+ !13 = !DILocation(line: 24, column: 21, scope: !6)
122
+ !14 = !DILocation(line: 25, column: 37, scope: !6)
123
+ !15 = !DILocation(line: 28, column: 19, scope: !6)
124
+ !16 = !DILocation(line: 27, column: 19, scope: !6)
125
+ !17 = !DILocation(line: 37, column: 41, scope: !6)
126
+ !18 = !DILocation(line: 37, column: 52, scope: !6)
127
+ !19 = !DILocation(line: 37, column: 48, scope: !6)
128
+ !20 = !DILocation(line: 37, column: 62, scope: !6)
129
+ !21 = !DILocation(line: 37, column: 57, scope: !6)
130
+ !22 = !DILocation(line: 37, column: 34, scope: !6)
131
+ !23 = !DILocation(line: 37, column: 67, scope: !6)
132
+ !24 = !DILocation(line: 37, column: 129, scope: !6)
133
+ !25 = !DILocation(line: 39, column: 22, scope: !6)
134
+ !26 = !DILocation(line: 256, column: 15, scope: !27, inlinedAt: !30)
135
+ !27 = distinct !DILexicalBlockFile(scope: !29, file: !28, discriminator: 0)
136
+ !28 = !DIFile(filename: "standard.py", directory: "/home/x/hfenv/lib/python3.12/site-packages/triton/language")
137
+ !29 = distinct !DILexicalBlockFile(scope: !6, file: !28, discriminator: 0)
138
+ !30 = !DILocation(line: 43, column: 25, scope: !6)
139
+ !31 = !DILocation(line: 286, column: 36, scope: !29, inlinedAt: !30)
140
+ !32 = !DILocation(line: 43, column: 28, scope: !6)
141
+ !33 = !DILocation(line: 44, column: 25, scope: !6)
142
+ !34 = !DILocation(line: 44, column: 36, scope: !6)
143
+ !35 = !DILocation(line: 44, column: 4, scope: !6)
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/7HDVHUPTEOX4FHYEQOTFHD5U573A7K7C62TYWKVZQQZ4QTR7356A/triton_red_fused__to_copy_mean_pow_4.ptx ADDED
@@ -0,0 +1,415 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ //
2
+ // Generated by LLVM NVPTX Back-End
3
+ //
4
+
5
+ .version 8.4
6
+ .target sm_90a
7
+ .address_size 64
8
+
9
+ // .globl triton_red_fused__to_copy_mean_pow_4 // -- Begin function triton_red_fused__to_copy_mean_pow_4
10
+ .extern .shared .align 16 .b8 global_smem[];
11
+ // @triton_red_fused__to_copy_mean_pow_4
12
+ .visible .entry triton_red_fused__to_copy_mean_pow_4(
13
+ .param .u64 .ptr .global .align 1 triton_red_fused__to_copy_mean_pow_4_param_0,
14
+ .param .u64 .ptr .global .align 1 triton_red_fused__to_copy_mean_pow_4_param_1,
15
+ .param .u32 triton_red_fused__to_copy_mean_pow_4_param_2,
16
+ .param .u32 triton_red_fused__to_copy_mean_pow_4_param_3,
17
+ .param .u64 .ptr .global .align 1 triton_red_fused__to_copy_mean_pow_4_param_4
18
+ )
19
+ .reqntid 256, 1, 1
20
+ {
21
+ .reg .pred %p<6>;
22
+ .reg .b16 %rs<5>;
23
+ .reg .b32 %r<43>;
24
+ .reg .f32 %f<20>;
25
+ .reg .b64 %rd<7>;
26
+ .loc 1 18 0 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:18:0
27
+ $L__func_begin0:
28
+ .loc 1 18 0 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:18:0
29
+
30
+ // %bb.0:
31
+ ld.param.u64 %rd3, [triton_red_fused__to_copy_mean_pow_4_param_0];
32
+ ld.param.u64 %rd4, [triton_red_fused__to_copy_mean_pow_4_param_1];
33
+ $L__tmp0:
34
+ .loc 1 22 28 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:22:28
35
+ mov.u32 %r8, %ctaid.x;
36
+ .loc 1 22 33 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:22:33
37
+ shl.b32 %r9, %r8, 3;
38
+ ld.param.u32 %r10, [triton_red_fused__to_copy_mean_pow_4_param_2];
39
+ .loc 1 23 44 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:23:44
40
+ mov.u32 %r11, %tid.x;
41
+ bfe.u32 %r12, %r11, 5, 3;
42
+ and.b32 %r13, %r11, 7;
43
+ .loc 1 23 23 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:23:23
44
+ or.b32 %r14, %r12, %r9;
45
+ or.b32 %r15, %r9, %r13;
46
+ .loc 1 24 21 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:24:21
47
+ setp.lt.s32 %p1, %r14, %r10;
48
+ setp.lt.s32 %p4, %r15, %r10;
49
+ .loc 1 25 37 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:25:37
50
+ shl.b32 %r16, %r11, 2;
51
+ and.b32 %r17, %r16, 124;
52
+ .loc 1 28 19 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:28:19
53
+ bfe.s32 %r18, %r8, 28, 1;
54
+ shr.u32 %r19, %r18, 29;
55
+ add.s32 %r20, %r14, %r19;
56
+ shr.u32 %r21, %r20, 3;
57
+ .loc 1 27 19 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:27:19
58
+ and.b32 %r22, %r20, 33554424;
59
+ sub.s32 %r23, %r14, %r22;
60
+ .loc 1 37 52 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:37:52
61
+ shl.b32 %r24, %r23, 7;
62
+ .loc 1 37 48 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:37:48
63
+ or.b32 %r25, %r24, %r17;
64
+ mad.lo.s32 %r26, %r21, 7168, %r25;
65
+ .loc 1 37 57 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:37:57
66
+ add.s32 %r27, %r26, 5120;
67
+ .loc 1 37 34 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:37:34
68
+ mul.wide.s32 %rd5, %r27, 2;
69
+ add.s64 %rd1, %rd3, %rd5;
70
+ mov.b32 %r3, 0;
71
+ .loc 1 37 67 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:37:67
72
+ // begin inline asm
73
+ mov.u32 %r1, %r3;
74
+ mov.u32 %r2, %r3;
75
+ @%p1 ld.global.L1::evict_first.v2.b32 { %r1, %r2 }, [ %rd1 + 0 ];
76
+ // end inline asm
77
+ mov.b32 {%rs1, %rs2}, %r1;
78
+ mov.b32 {%rs3, %rs4}, %r2;
79
+ .loc 1 37 129 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:37:129
80
+ cvt.f32.bf16 %f1, %rs1;
81
+ cvt.f32.bf16 %f2, %rs2;
82
+ cvt.f32.bf16 %f3, %rs3;
83
+ cvt.f32.bf16 %f4, %rs4;
84
+ .loc 1 39 22 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:39:22
85
+ mul.f32 %f5, %f2, %f2;
86
+ $L__tmp1:
87
+ .loc 2 256 15 // standard.py:256:15
88
+ fma.rn.f32 %f6, %f1, %f1, %f5;
89
+ fma.rn.f32 %f7, %f3, %f3, %f6;
90
+ fma.rn.f32 %f8, %f4, %f4, %f7;
91
+ selp.f32 %f9, %f8, 0f00000000, %p1;
92
+ .loc 2 286 36 // standard.py:286:36
93
+ mov.b32 %r28, %f9;
94
+ shfl.sync.bfly.b32 %r29, %r28, 16, 31, -1;
95
+ mov.b32 %f10, %r29;
96
+ .loc 2 256 15 // standard.py:256:15
97
+ add.f32 %f11, %f9, %f10;
98
+ .loc 2 286 36 // standard.py:286:36
99
+ mov.b32 %r30, %f11;
100
+ shfl.sync.bfly.b32 %r31, %r30, 8, 31, -1;
101
+ mov.b32 %f12, %r31;
102
+ .loc 2 256 15 // standard.py:256:15
103
+ add.f32 %f13, %f11, %f12;
104
+ .loc 2 286 36 // standard.py:286:36
105
+ mov.b32 %r32, %f13;
106
+ shfl.sync.bfly.b32 %r33, %r32, 4, 31, -1;
107
+ mov.b32 %f14, %r33;
108
+ .loc 2 256 15 // standard.py:256:15
109
+ add.f32 %f15, %f13, %f14;
110
+ .loc 2 286 36 // standard.py:286:36
111
+ mov.b32 %r34, %f15;
112
+ shfl.sync.bfly.b32 %r35, %r34, 2, 31, -1;
113
+ mov.b32 %f16, %r35;
114
+ .loc 2 256 15 // standard.py:256:15
115
+ add.f32 %f17, %f15, %f16;
116
+ .loc 2 286 36 // standard.py:286:36
117
+ mov.b32 %r36, %f17;
118
+ shfl.sync.bfly.b32 %r37, %r36, 1, 31, -1;
119
+ mov.b32 %f18, %r37;
120
+ .loc 2 256 15 // standard.py:256:15
121
+ add.f32 %f19, %f17, %f18;
122
+ $L__tmp2:
123
+ .loc 1 43 28 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:43:28
124
+ shl.b32 %r38, %r12, 2;
125
+ mov.u32 %r39, global_smem;
126
+ add.s32 %r5, %r39, %r38;
127
+ mov.b32 %r6, %f19;
128
+ mov.pred %p2, -1;
129
+ // begin inline asm
130
+ @%p2 st.shared.b32 [ %r5 + 0 ], %r6;
131
+ // end inline asm
132
+ bar.sync 0;
133
+ shl.b32 %r40, %r13, 2;
134
+ add.s32 %r41, %r39, %r40;
135
+ ld.shared.u32 %r7, [%r41];
136
+ .loc 1 44 25 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:44:25
137
+ mul.wide.s32 %rd6, %r15, 4;
138
+ add.s64 %rd2, %rd4, %rd6;
139
+ .loc 1 44 36 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:44:36
140
+ and.b32 %r42, %r11, 248;
141
+ setp.eq.s32 %p5, %r42, 0;
142
+ and.pred %p3, %p5, %p4;
143
+ // begin inline asm
144
+ @%p3 st.global.b32 [ %rd2 + 0 ], { %r7 };
145
+ // end inline asm
146
+ .loc 1 44 4 // ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py:44:4
147
+ ret;
148
+ $L__tmp3:
149
+ $L__func_end0:
150
+ // -- End function
151
+ }
152
+ .file 1 "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py"
153
+ .file 2 "/home/x/hfenv/lib/python3.12/site-packages/triton/language/standard.py"
154
+ .section .debug_abbrev
155
+ {
156
+ .b8 1 // Abbreviation Code
157
+ .b8 17 // DW_TAG_compile_unit
158
+ .b8 1 // DW_CHILDREN_yes
159
+ .b8 37 // DW_AT_producer
160
+ .b8 8 // DW_FORM_string
161
+ .b8 19 // DW_AT_language
162
+ .b8 5 // DW_FORM_data2
163
+ .b8 3 // DW_AT_name
164
+ .b8 8 // DW_FORM_string
165
+ .b8 16 // DW_AT_stmt_list
166
+ .b8 6 // DW_FORM_data4
167
+ .b8 27 // DW_AT_comp_dir
168
+ .b8 8 // DW_FORM_string
169
+ .b8 0 // EOM(1)
170
+ .b8 0 // EOM(2)
171
+ .b8 2 // Abbreviation Code
172
+ .b8 46 // DW_TAG_subprogram
173
+ .b8 0 // DW_CHILDREN_no
174
+ .b8 3 // DW_AT_name
175
+ .b8 8 // DW_FORM_string
176
+ .b8 32 // DW_AT_inline
177
+ .b8 11 // DW_FORM_data1
178
+ .b8 0 // EOM(1)
179
+ .b8 0 // EOM(2)
180
+ .b8 3 // Abbreviation Code
181
+ .b8 46 // DW_TAG_subprogram
182
+ .b8 1 // DW_CHILDREN_yes
183
+ .b8 17 // DW_AT_low_pc
184
+ .b8 1 // DW_FORM_addr
185
+ .b8 18 // DW_AT_high_pc
186
+ .b8 1 // DW_FORM_addr
187
+ .b8 49 // DW_AT_abstract_origin
188
+ .b8 19 // DW_FORM_ref4
189
+ .b8 0 // EOM(1)
190
+ .b8 0 // EOM(2)
191
+ .b8 4 // Abbreviation Code
192
+ .b8 29 // DW_TAG_inlined_subroutine
193
+ .b8 0 // DW_CHILDREN_no
194
+ .b8 49 // DW_AT_abstract_origin
195
+ .b8 19 // DW_FORM_ref4
196
+ .b8 17 // DW_AT_low_pc
197
+ .b8 1 // DW_FORM_addr
198
+ .b8 18 // DW_AT_high_pc
199
+ .b8 1 // DW_FORM_addr
200
+ .b8 88 // DW_AT_call_file
201
+ .b8 11 // DW_FORM_data1
202
+ .b8 89 // DW_AT_call_line
203
+ .b8 11 // DW_FORM_data1
204
+ .b8 87 // DW_AT_call_column
205
+ .b8 11 // DW_FORM_data1
206
+ .b8 0 // EOM(1)
207
+ .b8 0 // EOM(2)
208
+ .b8 0 // EOM(3)
209
+ }
210
+ .section .debug_info
211
+ {
212
+ .b32 241 // Length of Unit
213
+ .b8 2 // DWARF version number
214
+ .b8 0
215
+ .b32 .debug_abbrev // Offset Into Abbrev. Section
216
+ .b8 8 // Address Size (in bytes)
217
+ .b8 1 // Abbrev [1] 0xb:0xea DW_TAG_compile_unit
218
+ .b8 116 // DW_AT_producer
219
+ .b8 114
220
+ .b8 105
221
+ .b8 116
222
+ .b8 111
223
+ .b8 110
224
+ .b8 0
225
+ .b8 2 // DW_AT_language
226
+ .b8 0
227
+ .b8 99 // DW_AT_name
228
+ .b8 116
229
+ .b8 121
230
+ .b8 100
231
+ .b8 106
232
+ .b8 97
233
+ .b8 55
234
+ .b8 107
235
+ .b8 116
236
+ .b8 114
237
+ .b8 101
238
+ .b8 119
239
+ .b8 120
240
+ .b8 113
241
+ .b8 101
242
+ .b8 104
243
+ .b8 107
244
+ .b8 105
245
+ .b8 122
246
+ .b8 106
247
+ .b8 105
248
+ .b8 55
249
+ .b8 113
250
+ .b8 122
251
+ .b8 105
252
+ .b8 52
253
+ .b8 119
254
+ .b8 122
255
+ .b8 113
256
+ .b8 107
257
+ .b8 106
258
+ .b8 99
259
+ .b8 115
260
+ .b8 119
261
+ .b8 106
262
+ .b8 107
263
+ .b8 113
264
+ .b8 110
265
+ .b8 120
266
+ .b8 97
267
+ .b8 108
268
+ .b8 102
269
+ .b8 101
270
+ .b8 121
271
+ .b8 51
272
+ .b8 109
273
+ .b8 111
274
+ .b8 104
275
+ .b8 103
276
+ .b8 103
277
+ .b8 111
278
+ .b8 116
279
+ .b8 46
280
+ .b8 112
281
+ .b8 121
282
+ .b8 0
283
+ .b32 .debug_line // DW_AT_stmt_list
284
+ .b8 47 // DW_AT_comp_dir
285
+ .b8 104
286
+ .b8 111
287
+ .b8 109
288
+ .b8 101
289
+ .b8 47
290
+ .b8 120
291
+ .b8 47
292
+ .b8 46
293
+ .b8 99
294
+ .b8 97
295
+ .b8 99
296
+ .b8 104
297
+ .b8 101
298
+ .b8 47
299
+ .b8 118
300
+ .b8 108
301
+ .b8 108
302
+ .b8 109
303
+ .b8 47
304
+ .b8 116
305
+ .b8 111
306
+ .b8 114
307
+ .b8 99
308
+ .b8 104
309
+ .b8 95
310
+ .b8 99
311
+ .b8 111
312
+ .b8 109
313
+ .b8 112
314
+ .b8 105
315
+ .b8 108
316
+ .b8 101
317
+ .b8 95
318
+ .b8 99
319
+ .b8 97
320
+ .b8 99
321
+ .b8 104
322
+ .b8 101
323
+ .b8 47
324
+ .b8 57
325
+ .b8 48
326
+ .b8 98
327
+ .b8 52
328
+ .b8 53
329
+ .b8 98
330
+ .b8 99
331
+ .b8 101
332
+ .b8 48
333
+ .b8 50
334
+ .b8 47
335
+ .b8 114
336
+ .b8 97
337
+ .b8 110
338
+ .b8 107
339
+ .b8 95
340
+ .b8 48
341
+ .b8 95
342
+ .b8 48
343
+ .b8 47
344
+ .b8 105
345
+ .b8 110
346
+ .b8 100
347
+ .b8 117
348
+ .b8 99
349
+ .b8 116
350
+ .b8 111
351
+ .b8 114
352
+ .b8 95
353
+ .b8 99
354
+ .b8 97
355
+ .b8 99
356
+ .b8 104
357
+ .b8 101
358
+ .b8 47
359
+ .b8 116
360
+ .b8 121
361
+ .b8 0
362
+ .b8 2 // Abbrev [2] 0x9f:0x27 DW_TAG_subprogram
363
+ .b8 116 // DW_AT_name
364
+ .b8 114
365
+ .b8 105
366
+ .b8 116
367
+ .b8 111
368
+ .b8 110
369
+ .b8 95
370
+ .b8 114
371
+ .b8 101
372
+ .b8 100
373
+ .b8 95
374
+ .b8 102
375
+ .b8 117
376
+ .b8 115
377
+ .b8 101
378
+ .b8 100
379
+ .b8 95
380
+ .b8 95
381
+ .b8 116
382
+ .b8 111
383
+ .b8 95
384
+ .b8 99
385
+ .b8 111
386
+ .b8 112
387
+ .b8 121
388
+ .b8 95
389
+ .b8 109
390
+ .b8 101
391
+ .b8 97
392
+ .b8 110
393
+ .b8 95
394
+ .b8 112
395
+ .b8 111
396
+ .b8 119
397
+ .b8 95
398
+ .b8 52
399
+ .b8 0
400
+ .b8 1 // DW_AT_inline
401
+ .b8 3 // Abbrev [3] 0xc6:0x2e DW_TAG_subprogram
402
+ .b64 $L__func_begin0 // DW_AT_low_pc
403
+ .b64 $L__func_end0 // DW_AT_high_pc
404
+ .b32 159 // DW_AT_abstract_origin
405
+ .b8 4 // Abbrev [4] 0xdb:0x18 DW_TAG_inlined_subroutine
406
+ .b32 159 // DW_AT_abstract_origin
407
+ .b64 $L__tmp1 // DW_AT_low_pc
408
+ .b64 $L__tmp2 // DW_AT_high_pc
409
+ .b8 1 // DW_AT_call_file
410
+ .b8 43 // DW_AT_call_line
411
+ .b8 25 // DW_AT_call_column
412
+ .b8 0 // End Of Children Mark
413
+ .b8 0 // End Of Children Mark
414
+ }
415
+ .section .debug_macinfo { }
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/7HDVHUPTEOX4FHYEQOTFHD5U573A7K7C62TYWKVZQQZ4QTR7356A/triton_red_fused__to_copy_mean_pow_4.ttgir ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #blocked = #ttg.blocked<{sizePerThread = [1, 4], threadsPerWarp = [1, 32], warpsPerCTA = [8, 1], order = [1, 0]}>
2
+ #blocked1 = #ttg.blocked<{sizePerThread = [1, 1], threadsPerWarp = [8, 4], warpsPerCTA = [1, 8], order = [0, 1]}>
3
+ #loc = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":18:0)
4
+ #loc1 = loc(unknown)
5
+ #loc24 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":43:25)
6
+ #loc31 = loc(callsite(#loc1 at #loc24))
7
+ module attributes {"ttg.num-ctas" = 1 : i32, "ttg.num-warps" = 8 : i32, ttg.target = "cuda:90", "ttg.threads-per-warp" = 32 : i32} {
8
+ tt.func public @triton_red_fused__to_copy_mean_pow_4(%arg0: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":18:0), %arg1: !tt.ptr<f32> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":18:0), %arg2: i32 loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":18:0), %arg3: i32 {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":18:0)) attributes {noinline = false} {
9
+ %cst = arith.constant dense<0.000000e+00> : tensor<8x128xf32, #blocked> loc(#loc1)
10
+ %cst_0 = arith.constant dense<0.000000e+00> : tensor<8x128xbf16, #blocked> loc(#loc1)
11
+ %c8_i32 = arith.constant 8 : i32 loc(#loc1)
12
+ %cst_1 = arith.constant dense<8> : tensor<8x1xi32, #blocked> loc(#loc1)
13
+ %cst_2 = arith.constant dense<7168> : tensor<8x1xi32, #blocked> loc(#loc1)
14
+ %cst_3 = arith.constant dense<128> : tensor<8x1xi32, #blocked> loc(#loc1)
15
+ %cst_4 = arith.constant dense<5120> : tensor<1x128xi32, #blocked> loc(#loc1)
16
+ %cst_5 = arith.constant dense<128> : tensor<1x128xi32, #blocked> loc(#loc1)
17
+ %0 = tt.get_program_id x : i32 loc(#loc2)
18
+ %1 = arith.muli %0, %c8_i32 : i32 loc(#loc3)
19
+ %2 = tt.make_range {end = 8 : i32, start = 0 : i32} : tensor<8xi32, #ttg.slice<{dim = 1, parent = #blocked}>> loc(#loc4)
20
+ %3 = tt.make_range {end = 8 : i32, start = 0 : i32} : tensor<8xi32, #ttg.slice<{dim = 1, parent = #blocked1}>> loc(#loc4)
21
+ %4 = tt.expand_dims %2 {axis = 1 : i32} : tensor<8xi32, #ttg.slice<{dim = 1, parent = #blocked}>> -> tensor<8x1xi32, #blocked> loc(#loc4)
22
+ %5 = tt.expand_dims %3 {axis = 1 : i32} : tensor<8xi32, #ttg.slice<{dim = 1, parent = #blocked1}>> -> tensor<8x1xi32, #blocked1> loc(#loc4)
23
+ %6 = tt.splat %1 : i32 -> tensor<8x1xi32, #blocked> loc(#loc5)
24
+ %7 = tt.splat %1 : i32 -> tensor<8x1xi32, #blocked1> loc(#loc5)
25
+ %8 = arith.addi %6, %4 : tensor<8x1xi32, #blocked> loc(#loc5)
26
+ %9 = arith.addi %7, %5 : tensor<8x1xi32, #blocked1> loc(#loc5)
27
+ %10 = tt.splat %arg2 : i32 -> tensor<8x1xi32, #blocked> loc(#loc6)
28
+ %11 = tt.splat %arg2 : i32 -> tensor<8x1xi32, #blocked1> loc(#loc6)
29
+ %12 = arith.cmpi slt, %8, %10 : tensor<8x1xi32, #blocked> loc(#loc6)
30
+ %13 = arith.cmpi slt, %9, %11 : tensor<8x1xi32, #blocked1> loc(#loc6)
31
+ %14 = tt.make_range {end = 128 : i32, start = 0 : i32} : tensor<128xi32, #ttg.slice<{dim = 0, parent = #blocked}>> loc(#loc7)
32
+ %15 = tt.expand_dims %14 {axis = 0 : i32} : tensor<128xi32, #ttg.slice<{dim = 0, parent = #blocked}>> -> tensor<1x128xi32, #blocked> loc(#loc7)
33
+ %16 = arith.remsi %8, %cst_1 : tensor<8x1xi32, #blocked> loc(#loc8)
34
+ %17 = arith.divsi %8, %cst_1 : tensor<8x1xi32, #blocked> loc(#loc9)
35
+ %18 = arith.cmpi slt, %15, %cst_5 : tensor<1x128xi32, #blocked> loc(#loc10)
36
+ %19 = arith.addi %15, %cst_4 : tensor<1x128xi32, #blocked> loc(#loc11)
37
+ %20 = arith.muli %16, %cst_3 : tensor<8x1xi32, #blocked> loc(#loc12)
38
+ %21 = tt.broadcast %19 : tensor<1x128xi32, #blocked> -> tensor<8x128xi32, #blocked> loc(#loc13)
39
+ %22 = tt.broadcast %20 : tensor<8x1xi32, #blocked> -> tensor<8x128xi32, #blocked> loc(#loc13)
40
+ %23 = arith.addi %21, %22 : tensor<8x128xi32, #blocked> loc(#loc13)
41
+ %24 = arith.muli %17, %cst_2 : tensor<8x1xi32, #blocked> loc(#loc14)
42
+ %25 = tt.broadcast %24 : tensor<8x1xi32, #blocked> -> tensor<8x128xi32, #blocked> loc(#loc15)
43
+ %26 = arith.addi %23, %25 : tensor<8x128xi32, #blocked> loc(#loc15)
44
+ %27 = tt.splat %arg0 : !tt.ptr<bf16> -> tensor<8x128x!tt.ptr<bf16>, #blocked> loc(#loc16)
45
+ %28 = tt.addptr %27, %26 : tensor<8x128x!tt.ptr<bf16>, #blocked>, tensor<8x128xi32, #blocked> loc(#loc16)
46
+ %29 = tt.broadcast %18 : tensor<1x128xi1, #blocked> -> tensor<8x128xi1, #blocked> loc(#loc17)
47
+ %30 = tt.broadcast %12 : tensor<8x1xi1, #blocked> -> tensor<8x128xi1, #blocked> loc(#loc17)
48
+ %31 = arith.andi %29, %30 : tensor<8x128xi1, #blocked> loc(#loc17)
49
+ %32 = tt.load %28, %31, %cst_0 evictionPolicy = evict_first : tensor<8x128x!tt.ptr<bf16>, #blocked> loc(#loc18)
50
+ %33 = arith.extf %32 : tensor<8x128xbf16, #blocked> to tensor<8x128xf32, #blocked> loc(#loc19)
51
+ %34 = arith.mulf %33, %33 : tensor<8x128xf32, #blocked> loc(#loc20)
52
+ %35 = arith.addf %34, %cst : tensor<8x128xf32, #blocked> loc(#loc21)
53
+ %36 = arith.select %31, %35, %cst : tensor<8x128xi1, #blocked>, tensor<8x128xf32, #blocked> loc(#loc22)
54
+ %37 = "tt.reduce"(%36) <{axis = 1 : i32}> ({
55
+ ^bb0(%arg4: f32 loc(callsite(#loc1 at #loc24)), %arg5: f32 loc(callsite(#loc1 at #loc24))):
56
+ %42 = arith.addf %arg4, %arg5 : f32 loc(#loc33)
57
+ tt.reduce.return %42 : f32 loc(#loc30)
58
+ }) : (tensor<8x128xf32, #blocked>) -> tensor<8xf32, #ttg.slice<{dim = 1, parent = #blocked}>> loc(#loc30)
59
+ %38 = ttg.convert_layout %37 : tensor<8xf32, #ttg.slice<{dim = 1, parent = #blocked}>> -> tensor<8xf32, #ttg.slice<{dim = 1, parent = #blocked1}>> loc(#loc26)
60
+ %39 = tt.expand_dims %38 {axis = 1 : i32} : tensor<8xf32, #ttg.slice<{dim = 1, parent = #blocked1}>> -> tensor<8x1xf32, #blocked1> loc(#loc26)
61
+ %40 = tt.splat %arg1 : !tt.ptr<f32> -> tensor<8x1x!tt.ptr<f32>, #blocked1> loc(#loc27)
62
+ %41 = tt.addptr %40, %9 : tensor<8x1x!tt.ptr<f32>, #blocked1>, tensor<8x1xi32, #blocked1> loc(#loc27)
63
+ tt.store %41, %39, %13 : tensor<8x1x!tt.ptr<f32>, #blocked1> loc(#loc28)
64
+ tt.return loc(#loc29)
65
+ } loc(#loc)
66
+ } loc(#loc)
67
+ #loc2 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":22:28)
68
+ #loc3 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":22:33)
69
+ #loc4 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":23:44)
70
+ #loc5 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":23:23)
71
+ #loc6 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":24:21)
72
+ #loc7 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":25:37)
73
+ #loc8 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":27:19)
74
+ #loc9 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":28:19)
75
+ #loc10 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":33:29)
76
+ #loc11 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:41)
77
+ #loc12 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:52)
78
+ #loc13 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:48)
79
+ #loc14 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:62)
80
+ #loc15 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:57)
81
+ #loc16 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:34)
82
+ #loc17 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:77)
83
+ #loc18 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:67)
84
+ #loc19 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":37:129)
85
+ #loc20 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":39:22)
86
+ #loc21 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":41:23)
87
+ #loc22 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":42:48)
88
+ #loc23 = loc("/home/x/hfenv/lib/python3.12/site-packages/triton/language/standard.py":286:36)
89
+ #loc25 = loc("/home/x/hfenv/lib/python3.12/site-packages/triton/language/standard.py":256:15)
90
+ #loc26 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":43:28)
91
+ #loc27 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":44:25)
92
+ #loc28 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":44:36)
93
+ #loc29 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/ty/ctydja7ktrewxqehkizji7qzi4wzqkjcswjkqnxalfey3mohggot.py":44:4)
94
+ #loc30 = loc(callsite(#loc23 at #loc24))
95
+ #loc32 = loc(callsite(#loc25 at #loc23))
96
+ #loc33 = loc(callsite(#loc32 at #loc24))
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/CUASPVXQ7OI3JPFN55O2D57GL57SRLSB6LBTAARTQDPYHVXGYBSA/__grp__triton_poi_fused_add_mul_sub_6.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"child_paths": {"triton_poi_fused_add_mul_sub_6.ttir": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/CUASPVXQ7OI3JPFN55O2D57GL57SRLSB6LBTAARTQDPYHVXGYBSA/triton_poi_fused_add_mul_sub_6.ttir", "triton_poi_fused_add_mul_sub_6.ttgir": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/CUASPVXQ7OI3JPFN55O2D57GL57SRLSB6LBTAARTQDPYHVXGYBSA/triton_poi_fused_add_mul_sub_6.ttgir", "triton_poi_fused_add_mul_sub_6.llir": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/CUASPVXQ7OI3JPFN55O2D57GL57SRLSB6LBTAARTQDPYHVXGYBSA/triton_poi_fused_add_mul_sub_6.llir", "triton_poi_fused_add_mul_sub_6.ptx": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/CUASPVXQ7OI3JPFN55O2D57GL57SRLSB6LBTAARTQDPYHVXGYBSA/triton_poi_fused_add_mul_sub_6.ptx", "triton_poi_fused_add_mul_sub_6.cubin": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/CUASPVXQ7OI3JPFN55O2D57GL57SRLSB6LBTAARTQDPYHVXGYBSA/triton_poi_fused_add_mul_sub_6.cubin", "triton_poi_fused_add_mul_sub_6.json": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/CUASPVXQ7OI3JPFN55O2D57GL57SRLSB6LBTAARTQDPYHVXGYBSA/triton_poi_fused_add_mul_sub_6.json"}}
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/CUASPVXQ7OI3JPFN55O2D57GL57SRLSB6LBTAARTQDPYHVXGYBSA/triton_poi_fused_add_mul_sub_6.llir ADDED
@@ -0,0 +1,301 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ; ModuleID = 'LLVMDialectModule'
2
+ source_filename = "LLVMDialectModule"
3
+ target datalayout = "e-p3:32:32-p4:32:32-p5:32:32-p6:32:32-i64:64-i128:128-v16:16-v32:32-n16:32:64"
4
+
5
+ @assertFunc_0 = internal constant [8 x i8] c"unknown\00"
6
+ @assertFile_0 = internal constant [134 x i8] c"/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py\00"
7
+ @assertMessage_0 = internal constant [40 x i8] c"index out of bounds: 0 <= tmp16 < 40960\00"
8
+ @.str = private unnamed_addr constant [11 x i8] c"__CUDA_FTZ\00", align 1
9
+
10
+ ; Function Attrs: noreturn
11
+ declare !dbg !6 void @__assertfail(ptr, ptr, i32, ptr, i64) local_unnamed_addr #0
12
+
13
+ define ptx_kernel void @triton_poi_fused_add_mul_sub_6(ptr addrspace(1) %0, ptr addrspace(1) %1, ptr addrspace(1) %2, ptr addrspace(1) %3, ptr addrspace(1) %4, ptr addrspace(1) %5, ptr addrspace(1) %6, i32 %7, ptr addrspace(1) readnone captures(none) %8) local_unnamed_addr !dbg !10 {
14
+ %10 = tail call i32 @llvm.nvvm.read.ptx.sreg.ctaid.x(), !dbg !11
15
+ %11 = shl i32 %10, 9, !dbg !12
16
+ %12 = tail call i32 @llvm.nvvm.read.ptx.sreg.tid.x(), !dbg !13
17
+ %13 = shl i32 %12, 1, !dbg !13
18
+ %14 = and i32 %13, 510, !dbg !13
19
+ %15 = or disjoint i32 %14, %11, !dbg !14
20
+ %16 = or disjoint i32 %15, 1, !dbg !14
21
+ %17 = icmp slt i32 %15, %7, !dbg !15
22
+ %18 = sdiv i32 %15, 64, !dbg !16
23
+ %19 = mul i32 %18, 64, !dbg !17
24
+ %.decomposed = sub i32 %15, %19, !dbg !17
25
+ %20 = srem i32 %16, 64, !dbg !17
26
+ %21 = srem i32 %18, 8, !dbg !18
27
+ %22 = sdiv i32 %15, 512, !dbg !19
28
+ %23 = shl nsw i32 %21, 7, !dbg !20
29
+ %24 = mul i32 %22, 7168, !dbg !21
30
+ %25 = add i32 %23, %24, !dbg !22
31
+ %26 = add i32 %25, 5120, !dbg !23
32
+ %27 = add i32 %26, %.decomposed, !dbg !24
33
+ %28 = sext i32 %27 to i64, !dbg !25
34
+ %29 = getelementptr bfloat, ptr addrspace(1) %0, i64 %28, !dbg !25
35
+ %30 = tail call i32 asm sideeffect "mov.u32 $0, 0x0;\0A\09@$2 ld.global.b32 { $0 }, [ $1 + 0 ];", "=r,l,b"(ptr addrspace(1) %29, i1 %17) #5, !dbg !26
36
+ %31 = sext i32 %18 to i64, !dbg !27
37
+ %32 = getelementptr float, ptr addrspace(1) %1, i64 %31, !dbg !27
38
+ %33 = tail call i32 asm sideeffect "mov.u32 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b32 { $0 }, [ $1 + 0 ];", "=r,l,b"(ptr addrspace(1) %32, i1 %17) #5, !dbg !28
39
+ %34 = bitcast i32 %33 to float, !dbg !28
40
+ %35 = tail call i32 asm sideeffect "mov.u32 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b32 { $0 }, [ $1 + 0 ];", "=r,l,b"(ptr addrspace(1) %32, i1 %17) #5, !dbg !28
41
+ %36 = bitcast i32 %35 to float, !dbg !28
42
+ %37 = sext i32 %.decomposed to i64, !dbg !29
43
+ %38 = getelementptr bfloat, ptr addrspace(1) %2, i64 %37, !dbg !29
44
+ %39 = tail call i32 asm sideeffect "mov.u32 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b32 { $0 }, [ $1 + 0 ];", "=r,l,b"(ptr addrspace(1) %38, i1 %17) #5, !dbg !30
45
+ %40 = sext i32 %22 to i64, !dbg !31
46
+ %41 = getelementptr i64, ptr addrspace(1) %3, i64 %40, !dbg !31
47
+ %42 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %41, i1 %17) #5, !dbg !32
48
+ %43 = tail call i64 asm sideeffect "mov.u64 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b64 { $0 }, [ $1 + 0 ];", "=l,l,b"(ptr addrspace(1) %41, i1 %17) #5, !dbg !32
49
+ %44 = add nsw i32 %.decomposed, 64, !dbg !33
50
+ %45 = add i32 %25, 5184, !dbg !34
51
+ %46 = add nsw i32 %45, %.decomposed, !dbg !35
52
+ %47 = add nsw i32 %45, %20, !dbg !35
53
+ %48 = sext i32 %46 to i64, !dbg !36
54
+ %49 = getelementptr bfloat, ptr addrspace(1) %0, i64 %48, !dbg !36
55
+ %50 = sext i32 %47 to i64, !dbg !36
56
+ %51 = getelementptr bfloat, ptr addrspace(1) %0, i64 %50, !dbg !36
57
+ %52 = tail call i16 asm sideeffect "mov.u16 $0, 0x0;\0A\09@$2 ld.global.b16 { $0 }, [ $1 + 0 ];", "=c,l,b"(ptr addrspace(1) %49, i1 %17) #5, !dbg !37
58
+ %53 = tail call i16 asm sideeffect "mov.u16 $0, 0x0;\0A\09@$2 ld.global.b16 { $0 }, [ $1 + 0 ];", "=c,l,b"(ptr addrspace(1) %51, i1 %17) #5, !dbg !37
59
+ %54 = tail call i32 asm sideeffect "mov.u32 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b32 { $0 }, [ $1 + 0 ];", "=r,l,b"(ptr addrspace(1) %32, i1 %17) #5, !dbg !38
60
+ %55 = tail call i32 asm sideeffect "mov.u32 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b32 { $0 }, [ $1 + 0 ];", "=r,l,b"(ptr addrspace(1) %32, i1 %17) #5, !dbg !38
61
+ %56 = zext nneg i32 %44 to i64, !dbg !39
62
+ %57 = getelementptr bfloat, ptr addrspace(1) %2, i64 %56, !dbg !39
63
+ %58 = tail call i32 asm sideeffect "mov.u32 $0, 0x0;\0A\09@$2 ld.global.L1::evict_last.b32 { $0 }, [ $1 + 0 ];", "=r,l,b"(ptr addrspace(1) %57, i1 %17) #5, !dbg !40
64
+ %59 = tail call float @llvm.nvvm.div.full(float %34, float 1.280000e+02), !dbg !41
65
+ %60 = tail call float @llvm.nvvm.div.full(float %36, float 1.280000e+02), !dbg !41
66
+ %61 = fadd float %59, 0x3EB0C6F7A0000000, !dbg !42
67
+ %62 = fadd float %60, 0x3EB0C6F7A0000000, !dbg !42
68
+ %63 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #5, !dbg !43
69
+ %.not.i = icmp eq i32 %63, 0, !dbg !43
70
+ br i1 %.not.i, label %66, label %64, !dbg !43
71
+
72
+ 64: ; preds = %9
73
+ %65 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %61), !dbg !43
74
+ br label %__nv_rsqrtf.exit, !dbg !43
75
+
76
+ 66: ; preds = %9
77
+ %67 = tail call float @llvm.nvvm.rsqrt.approx.f(float %61), !dbg !43
78
+ br label %__nv_rsqrtf.exit, !dbg !43
79
+
80
+ __nv_rsqrtf.exit: ; preds = %64, %66
81
+ %.0.i = phi float [ %65, %64 ], [ %67, %66 ], !dbg !43
82
+ %68 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #5, !dbg !43
83
+ %.not.i2 = icmp eq i32 %68, 0, !dbg !43
84
+ br i1 %.not.i2, label %71, label %69, !dbg !43
85
+
86
+ 69: ; preds = %__nv_rsqrtf.exit
87
+ %70 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %62), !dbg !43
88
+ br label %__nv_rsqrtf.exit4, !dbg !43
89
+
90
+ 71: ; preds = %__nv_rsqrtf.exit
91
+ %72 = tail call float @llvm.nvvm.rsqrt.approx.f(float %62), !dbg !43
92
+ br label %__nv_rsqrtf.exit4, !dbg !43
93
+
94
+ __nv_rsqrtf.exit4: ; preds = %69, %71
95
+ %.0.i3 = phi float [ %70, %69 ], [ %72, %71 ], !dbg !43
96
+ %73 = add i64 %42, 40960, !dbg !44
97
+ %74 = icmp slt i64 %42, 0, !dbg !45
98
+ %75 = select i1 %74, i64 %73, i64 %42, !dbg !46
99
+ %76 = icmp ugt i64 %75, 40959, !dbg !47
100
+ %.not1 = and i1 %17, %76, !dbg !48
101
+ br i1 %.not1, label %77, label %78, !dbg !48
102
+
103
+ 77: ; preds = %__nv_rsqrtf.exit4
104
+ tail call void @__assertfail(ptr nonnull @assertMessage_0, ptr nonnull @assertFile_0, i32 46, ptr nonnull @assertFunc_0, i64 1), !dbg !48
105
+ unreachable, !dbg !48
106
+
107
+ 78: ; preds = %__nv_rsqrtf.exit4
108
+ %79 = bitcast i32 %55 to float, !dbg !38
109
+ %80 = bitcast i32 %54 to float, !dbg !38
110
+ tail call void @llvm.nvvm.barrier0(), !dbg !48
111
+ %.idx = shl i64 %75, 8, !dbg !49
112
+ %81 = getelementptr i8, ptr addrspace(1) %4, i64 %.idx, !dbg !49
113
+ %82 = getelementptr bfloat, ptr addrspace(1) %81, i64 %37, !dbg !49
114
+ %83 = tail call i32 asm sideeffect "mov.u32 $0, 0x0;\0A\09@$2 ld.global.b32 { $0 }, [ $1 + 0 ];", "=r,l,b"(ptr addrspace(1) %82, i1 %17) #5, !dbg !50
115
+ %84 = tail call float @llvm.nvvm.div.full(float %80, float 1.280000e+02), !dbg !51
116
+ %85 = tail call float @llvm.nvvm.div.full(float %79, float 1.280000e+02), !dbg !51
117
+ %86 = fadd float %84, 0x3EB0C6F7A0000000, !dbg !52
118
+ %87 = fadd float %85, 0x3EB0C6F7A0000000, !dbg !52
119
+ %88 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #5, !dbg !53
120
+ %.not.i5 = icmp eq i32 %88, 0, !dbg !53
121
+ br i1 %.not.i5, label %91, label %89, !dbg !53
122
+
123
+ 89: ; preds = %78
124
+ %90 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %86), !dbg !53
125
+ br label %__nv_rsqrtf.exit7, !dbg !53
126
+
127
+ 91: ; preds = %78
128
+ %92 = tail call float @llvm.nvvm.rsqrt.approx.f(float %86), !dbg !53
129
+ br label %__nv_rsqrtf.exit7, !dbg !53
130
+
131
+ __nv_rsqrtf.exit7: ; preds = %89, %91
132
+ %.0.i6 = phi float [ %90, %89 ], [ %92, %91 ], !dbg !53
133
+ %93 = tail call i32 @__nvvm_reflect(ptr nonnull @.str) #5, !dbg !53
134
+ %.not.i8 = icmp eq i32 %93, 0, !dbg !53
135
+ br i1 %.not.i8, label %96, label %94, !dbg !53
136
+
137
+ 94: ; preds = %__nv_rsqrtf.exit7
138
+ %95 = tail call float @llvm.nvvm.rsqrt.approx.ftz.f(float %87), !dbg !53
139
+ br label %__nv_rsqrtf.exit10, !dbg !53
140
+
141
+ 96: ; preds = %__nv_rsqrtf.exit7
142
+ %97 = tail call float @llvm.nvvm.rsqrt.approx.f(float %87), !dbg !53
143
+ br label %__nv_rsqrtf.exit10, !dbg !53
144
+
145
+ __nv_rsqrtf.exit10: ; preds = %94, %96
146
+ %.0.i9 = phi float [ %95, %94 ], [ %97, %96 ], !dbg !53
147
+ %98 = bitcast i32 %30 to <2 x bfloat>, !dbg !26
148
+ %99 = bitcast i32 %39 to <2 x bfloat>, !dbg !30
149
+ %100 = bitcast i32 %83 to <2 x bfloat>, !dbg !50
150
+ %101 = insertelement <2 x i16> poison, i16 %52, i64 0, !dbg !37
151
+ %102 = insertelement <2 x i16> %101, i16 %53, i64 1, !dbg !37
152
+ %103 = bitcast <2 x i16> %102 to <2 x bfloat>, !dbg !37
153
+ %104 = bitcast i32 %58 to <2 x bfloat>, !dbg !40
154
+ %105 = getelementptr bfloat, ptr addrspace(1) %81, i64 %56, !dbg !54
155
+ %106 = tail call i32 asm sideeffect "mov.u32 $0, 0x0;\0A\09@$2 ld.global.b32 { $0 }, [ $1 + 0 ];", "=r,l,b"(ptr addrspace(1) %105, i1 %17) #5, !dbg !55
156
+ %107 = bitcast i32 %106 to <2 x bfloat>, !dbg !55
157
+ %108 = shl i32 %18, 7, !dbg !56
158
+ %109 = add i32 %108, %.decomposed, !dbg !57
159
+ %110 = sext i32 %109 to i64, !dbg !58
160
+ %111 = getelementptr bfloat, ptr addrspace(1) %5, i64 %110, !dbg !58
161
+ %112 = fpext <2 x bfloat> %98 to <2 x float>, !dbg !59
162
+ %113 = insertelement <2 x float> poison, float %.0.i, i64 0, !dbg !60
163
+ %114 = insertelement <2 x float> %113, float %.0.i3, i64 1, !dbg !60
164
+ %115 = fmul <2 x float> %114, %112, !dbg !60
165
+ %116 = fpext <2 x bfloat> %99 to <2 x float>, !dbg !61
166
+ %117 = fmul <2 x float> %115, %116, !dbg !62
167
+ %118 = fpext <2 x bfloat> %100 to <2 x float>, !dbg !63
168
+ %119 = fmul <2 x float> %117, %118, !dbg !64
169
+ %120 = fpext <2 x bfloat> %103 to <2 x float>, !dbg !65
170
+ %121 = fpext <2 x bfloat> %104 to <2 x float>, !dbg !66
171
+ %122 = insertelement <2 x float> poison, float %.0.i6, i64 0, !dbg !67
172
+ %123 = insertelement <2 x float> %122, float %.0.i9, i64 1, !dbg !67
173
+ %124 = fmul <2 x float> %123, %120, !dbg !67
174
+ %125 = fmul <2 x float> %124, %121, !dbg !68
175
+ %126 = fpext <2 x bfloat> %107 to <2 x float>, !dbg !69
176
+ %127 = fmul <2 x float> %125, %126, !dbg !70
177
+ %128 = fsub <2 x float> %119, %127, !dbg !71
178
+ %129 = fptrunc <2 x float> %128 to <2 x bfloat>, !dbg !72
179
+ %130 = bitcast <2 x bfloat> %129 to i32, !dbg !72
180
+ tail call void asm sideeffect "@$2 st.global.b32 [ $1 + 0 ], { $0 };", "r,l,b"(i32 %130, ptr addrspace(1) %111, i1 %17) #5, !dbg !72
181
+ %131 = getelementptr bfloat, ptr addrspace(1) %6, i64 %110, !dbg !73
182
+ %132 = fmul <2 x float> %125, %118, !dbg !74
183
+ %133 = fmul <2 x float> %117, %126, !dbg !75
184
+ %134 = fadd <2 x float> %132, %133, !dbg !76
185
+ %135 = fptrunc <2 x float> %134 to <2 x bfloat>, !dbg !77
186
+ %136 = bitcast <2 x bfloat> %135 to i32, !dbg !77
187
+ tail call void asm sideeffect "@$2 st.global.b32 [ $1 + 0 ], { $0 };", "r,l,b"(i32 %136, ptr addrspace(1) %131, i1 %17) #5, !dbg !77
188
+ ret void, !dbg !78
189
+ }
190
+
191
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none)
192
+ declare noundef i32 @llvm.nvvm.read.ptx.sreg.ctaid.x() #1
193
+
194
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none)
195
+ declare noundef i32 @llvm.nvvm.read.ptx.sreg.tid.x() #1
196
+
197
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
198
+ declare float @llvm.nvvm.div.full(float, float) #2
199
+
200
+ ; Function Attrs: convergent nocallback nounwind
201
+ declare void @llvm.nvvm.barrier0() #3
202
+
203
+ declare i32 @__nvvm_reflect(ptr) local_unnamed_addr #4
204
+
205
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
206
+ declare float @llvm.nvvm.rsqrt.approx.ftz.f(float) #2
207
+
208
+ ; Function Attrs: mustprogress nocallback nofree nosync nounwind willreturn memory(none)
209
+ declare float @llvm.nvvm.rsqrt.approx.f(float) #2
210
+
211
+ attributes #0 = { noreturn }
212
+ attributes #1 = { mustprogress nocallback nofree nosync nounwind speculatable willreturn memory(none) }
213
+ attributes #2 = { mustprogress nocallback nofree nosync nounwind willreturn memory(none) }
214
+ attributes #3 = { convergent nocallback nounwind }
215
+ attributes #4 = { "disable-tail-calls"="false" "frame-pointer"="all" "less-precise-fpmad"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
216
+ attributes #5 = { nounwind }
217
+
218
+ !llvm.module.flags = !{!0, !1}
219
+ !llvm.dbg.cu = !{!2}
220
+ !nvvm.annotations = !{!4}
221
+ !llvm.ident = !{!5}
222
+
223
+ !0 = !{i32 2, !"Debug Info Version", i32 3}
224
+ !1 = !{i32 4, !"nvvm-reflect-ftz", i32 1}
225
+ !2 = distinct !DICompileUnit(language: DW_LANG_C, file: !3, producer: "triton", isOptimized: true, runtimeVersion: 0, emissionKind: LineTablesOnly)
226
+ !3 = !DIFile(filename: "cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py", directory: "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz")
227
+ !4 = !{ptr @triton_poi_fused_add_mul_sub_6, !"reqntidx", i32 256}
228
+ !5 = !{!"clang version 3.8.0 (tags/RELEASE_380/final)"}
229
+ !6 = !DISubprogram(name: "__assertfail", linkageName: "__assertfail", scope: !7, file: !7, type: !8, spFlags: DISPFlagOptimized)
230
+ !7 = !DIFile(filename: "<unknown>", directory: "")
231
+ !8 = !DISubroutineType(cc: DW_CC_normal, types: !9)
232
+ !9 = !{}
233
+ !10 = distinct !DISubprogram(name: "triton_poi_fused_add_mul_sub_6", linkageName: "triton_poi_fused_add_mul_sub_6", scope: !3, file: !3, line: 18, type: !8, scopeLine: 18, spFlags: DISPFlagDefinition | DISPFlagOptimized, unit: !2)
234
+ !11 = !DILocation(line: 19, column: 28, scope: !10)
235
+ !12 = !DILocation(line: 19, column: 33, scope: !10)
236
+ !13 = !DILocation(line: 20, column: 36, scope: !10)
237
+ !14 = !DILocation(line: 20, column: 23, scope: !10)
238
+ !15 = !DILocation(line: 21, column: 21, scope: !10)
239
+ !16 = !DILocation(line: 23, column: 21, scope: !10)
240
+ !17 = !DILocation(line: 22, column: 19, scope: !10)
241
+ !18 = !DILocation(line: 23, column: 27, scope: !10)
242
+ !19 = !DILocation(line: 24, column: 19, scope: !10)
243
+ !20 = !DILocation(line: 26, column: 46, scope: !10)
244
+ !21 = !DILocation(line: 26, column: 56, scope: !10)
245
+ !22 = !DILocation(line: 26, column: 37, scope: !10)
246
+ !23 = !DILocation(line: 26, column: 42, scope: !10)
247
+ !24 = !DILocation(line: 26, column: 51, scope: !10)
248
+ !25 = !DILocation(line: 26, column: 30, scope: !10)
249
+ !26 = !DILocation(line: 26, column: 61, scope: !10)
250
+ !27 = !DILocation(line: 27, column: 30, scope: !10)
251
+ !28 = !DILocation(line: 27, column: 35, scope: !10)
252
+ !29 = !DILocation(line: 28, column: 31, scope: !10)
253
+ !30 = !DILocation(line: 28, column: 36, scope: !10)
254
+ !31 = !DILocation(line: 29, column: 31, scope: !10)
255
+ !32 = !DILocation(line: 29, column: 36, scope: !10)
256
+ !33 = !DILocation(line: 30, column: 63, scope: !10)
257
+ !34 = !DILocation(line: 30, column: 43, scope: !10)
258
+ !35 = !DILocation(line: 30, column: 77, scope: !10)
259
+ !36 = !DILocation(line: 30, column: 31, scope: !10)
260
+ !37 = !DILocation(line: 30, column: 87, scope: !10)
261
+ !38 = !DILocation(line: 31, column: 57, scope: !10)
262
+ !39 = !DILocation(line: 32, column: 31, scope: !10)
263
+ !40 = !DILocation(line: 32, column: 41, scope: !10)
264
+ !41 = !DILocation(line: 35, column: 19, scope: !10)
265
+ !42 = !DILocation(line: 37, column: 18, scope: !10)
266
+ !43 = !DILocation(line: 38, column: 27, scope: !10)
267
+ !44 = !DILocation(line: 43, column: 20, scope: !10)
268
+ !45 = !DILocation(line: 44, column: 20, scope: !10)
269
+ !46 = !DILocation(line: 45, column: 35, scope: !10)
270
+ !47 = !DILocation(line: 46, column: 38, scope: !10)
271
+ !48 = !DILocation(line: 46, column: 66, scope: !10)
272
+ !49 = !DILocation(line: 47, column: 31, scope: !10)
273
+ !50 = !DILocation(line: 47, column: 48, scope: !10)
274
+ !51 = !DILocation(line: 50, column: 21, scope: !10)
275
+ !52 = !DILocation(line: 51, column: 20, scope: !10)
276
+ !53 = !DILocation(line: 52, column: 28, scope: !10)
277
+ !54 = !DILocation(line: 56, column: 31, scope: !10)
278
+ !55 = !DILocation(line: 56, column: 53, scope: !10)
279
+ !56 = !DILocation(line: 62, column: 34, scope: !10)
280
+ !57 = !DILocation(line: 62, column: 30, scope: !10)
281
+ !58 = !DILocation(line: 62, column: 25, scope: !10)
282
+ !59 = !DILocation(line: 26, column: 71, scope: !10)
283
+ !60 = !DILocation(line: 39, column: 18, scope: !10)
284
+ !61 = !DILocation(line: 28, column: 76, scope: !10)
285
+ !62 = !DILocation(line: 41, column: 19, scope: !10)
286
+ !63 = !DILocation(line: 47, column: 58, scope: !10)
287
+ !64 = !DILocation(line: 48, column: 20, scope: !10)
288
+ !65 = !DILocation(line: 30, column: 97, scope: !10)
289
+ !66 = !DILocation(line: 32, column: 81, scope: !10)
290
+ !67 = !DILocation(line: 53, column: 20, scope: !10)
291
+ !68 = !DILocation(line: 55, column: 20, scope: !10)
292
+ !69 = !DILocation(line: 56, column: 63, scope: !10)
293
+ !70 = !DILocation(line: 57, column: 20, scope: !10)
294
+ !71 = !DILocation(line: 58, column: 20, scope: !10)
295
+ !72 = !DILocation(line: 62, column: 46, scope: !10)
296
+ !73 = !DILocation(line: 63, column: 25, scope: !10)
297
+ !74 = !DILocation(line: 59, column: 20, scope: !10)
298
+ !75 = !DILocation(line: 60, column: 20, scope: !10)
299
+ !76 = !DILocation(line: 61, column: 20, scope: !10)
300
+ !77 = !DILocation(line: 63, column: 46, scope: !10)
301
+ !78 = !DILocation(line: 63, column: 4, scope: !10)
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/CUASPVXQ7OI3JPFN55O2D57GL57SRLSB6LBTAARTQDPYHVXGYBSA/triton_poi_fused_add_mul_sub_6.ptx ADDED
@@ -0,0 +1,516 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ //
2
+ // Generated by LLVM NVPTX Back-End
3
+ //
4
+
5
+ .version 8.4
6
+ .target sm_90a
7
+ .address_size 64
8
+
9
+ // .globl triton_poi_fused_add_mul_sub_6 // -- Begin function triton_poi_fused_add_mul_sub_6
10
+ .extern .func __assertfail
11
+ (
12
+ .param .b64 __assertfail_param_0,
13
+ .param .b64 __assertfail_param_1,
14
+ .param .b32 __assertfail_param_2,
15
+ .param .b64 __assertfail_param_3,
16
+ .param .b64 __assertfail_param_4
17
+ )
18
+ .noreturn;
19
+ .global .align 1 .b8 assertFunc_0[8] = {117, 110, 107, 110, 111, 119, 110};
20
+ .global .align 1 .b8 assertFile_0[134] = {47, 104, 111, 109, 101, 47, 120, 47, 46, 99, 97, 99, 104, 101, 47, 118, 108, 108, 109, 47, 116, 111, 114, 99, 104, 95, 99, 111, 109, 112, 105, 108, 101, 95, 99, 97, 99, 104, 101, 47, 57, 48, 98, 52, 53, 98, 99, 101, 48, 50, 47, 114, 97, 110, 107, 95, 48, 95, 48, 47, 105, 110, 100, 117, 99, 116, 111, 114, 95, 99, 97, 99, 104, 101, 47, 110, 122, 47, 99, 110, 122, 120, 102, 112, 115, 105, 99, 107, 50, 55, 54, 107, 101, 107, 109, 55, 116, 102, 111, 119, 55, 98, 108, 119, 52, 52, 113, 117, 54, 100, 54, 104, 118, 115, 101, 120, 55, 105, 99, 104, 117, 116, 101, 51, 106, 116, 119, 109, 51, 97, 46, 112, 121};
21
+ .global .align 1 .b8 assertMessage_0[40] = {105, 110, 100, 101, 120, 32, 111, 117, 116, 32, 111, 102, 32, 98, 111, 117, 110, 100, 115, 58, 32, 48, 32, 60, 61, 32, 116, 109, 112, 49, 54, 32, 60, 32, 52, 48, 57, 54, 48};
22
+ .global .align 1 .b8 _$_str[11] = {95, 95, 67, 85, 68, 65, 95, 70, 84, 90};
23
+ // @triton_poi_fused_add_mul_sub_6
24
+ .visible .entry triton_poi_fused_add_mul_sub_6(
25
+ .param .u64 .ptr .global .align 1 triton_poi_fused_add_mul_sub_6_param_0,
26
+ .param .u64 .ptr .global .align 1 triton_poi_fused_add_mul_sub_6_param_1,
27
+ .param .u64 .ptr .global .align 1 triton_poi_fused_add_mul_sub_6_param_2,
28
+ .param .u64 .ptr .global .align 1 triton_poi_fused_add_mul_sub_6_param_3,
29
+ .param .u64 .ptr .global .align 1 triton_poi_fused_add_mul_sub_6_param_4,
30
+ .param .u64 .ptr .global .align 1 triton_poi_fused_add_mul_sub_6_param_5,
31
+ .param .u64 .ptr .global .align 1 triton_poi_fused_add_mul_sub_6_param_6,
32
+ .param .u32 triton_poi_fused_add_mul_sub_6_param_7,
33
+ .param .u64 .ptr .global .align 1 triton_poi_fused_add_mul_sub_6_param_8
34
+ )
35
+ .reqntid 256, 1, 1
36
+ {
37
+ .reg .pred %p<20>;
38
+ .reg .b16 %rs<17>;
39
+ .reg .b32 %r<53>;
40
+ .reg .f32 %f<49>;
41
+ .reg .b64 %rd<47>;
42
+ .loc 1 18 0 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:18:0
43
+ $L__func_begin0:
44
+ .loc 1 18 0 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:18:0
45
+
46
+ // %bb.0: // %__nv_rsqrtf.exit
47
+ ld.param.u32 %r8, [triton_poi_fused_add_mul_sub_6_param_7];
48
+ ld.param.u64 %rd20, [triton_poi_fused_add_mul_sub_6_param_0];
49
+ ld.param.u64 %rd21, [triton_poi_fused_add_mul_sub_6_param_1];
50
+ $L__tmp0:
51
+ .loc 1 19 28 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:19:28
52
+ mov.u32 %r16, %ctaid.x;
53
+ .loc 1 19 33 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:19:33
54
+ shl.b32 %r17, %r16, 9;
55
+ ld.param.u64 %rd22, [triton_poi_fused_add_mul_sub_6_param_2];
56
+ ld.param.u64 %rd23, [triton_poi_fused_add_mul_sub_6_param_3];
57
+ .loc 1 20 36 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:20:36
58
+ mov.u32 %r18, %tid.x;
59
+ shl.b32 %r19, %r18, 1;
60
+ and.b32 %r20, %r19, 510;
61
+ .loc 1 20 23 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:20:23
62
+ or.b32 %r1, %r20, %r17;
63
+ or.b32 %r21, %r1, 1;
64
+ .loc 1 21 21 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:21:21
65
+ setp.ge.s32 %p12, %r1, %r8;
66
+ setp.lt.s32 %p11, %r1, %r8;
67
+ .loc 1 23 21 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:23:21
68
+ bfe.s32 %r22, %r16, 22, 1;
69
+ .loc 1 22 19 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:22:19
70
+ shr.u32 %r23, %r22, 26;
71
+ .loc 1 23 21 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:23:21
72
+ add.s32 %r24, %r1, %r23;
73
+ shr.s32 %r2, %r24, 6;
74
+ .loc 1 22 19 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:22:19
75
+ and.b32 %r25, %r24, -64;
76
+ sub.s32 %r26, %r1, %r25;
77
+ add.s32 %r27, %r21, %r23;
78
+ and.b32 %r28, %r27, -64;
79
+ sub.s32 %r29, %r21, %r28;
80
+ .loc 1 23 27 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:23:27
81
+ shr.u32 %r30, %r2, 29;
82
+ add.s32 %r31, %r2, %r30;
83
+ and.b32 %r32, %r31, 33554424;
84
+ sub.s32 %r33, %r2, %r32;
85
+ .loc 1 24 19 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:24:19
86
+ shr.u32 %r34, %r22, 23;
87
+ add.s32 %r35, %r1, %r34;
88
+ shr.s32 %r36, %r35, 9;
89
+ .loc 1 26 46 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:26:46
90
+ shl.b32 %r37, %r33, 7;
91
+ .loc 1 26 37 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:26:37
92
+ mad.lo.s32 %r38, %r36, 7168, %r37;
93
+ .loc 1 26 42 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:26:42
94
+ add.s32 %r39, %r38, %r26;
95
+ .loc 1 26 51 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:26:51
96
+ add.s32 %r40, %r39, 5120;
97
+ .loc 1 26 30 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:26:30
98
+ mul.wide.s32 %rd24, %r40, 2;
99
+ add.s64 %rd7, %rd20, %rd24;
100
+ .loc 1 26 61 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:26:61
101
+ // begin inline asm
102
+ mov.u32 %r9, 0x0;
103
+ @%p11 ld.global.b32 { %r9 }, [ %rd7 + 0 ];
104
+ // end inline asm
105
+ .loc 1 27 30 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:27:30
106
+ mul.wide.s32 %rd25, %r2, 4;
107
+ add.s64 %rd8, %rd21, %rd25;
108
+ .loc 1 27 35 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:27:35
109
+ // begin inline asm
110
+ mov.u32 %r10, 0x0;
111
+ @%p11 ld.global.L1::evict_last.b32 { %r10 }, [ %rd8 + 0 ];
112
+ // end inline asm
113
+ // begin inline asm
114
+ mov.u32 %r11, 0x0;
115
+ @%p11 ld.global.L1::evict_last.b32 { %r11 }, [ %rd8 + 0 ];
116
+ // end inline asm
117
+ .loc 1 28 31 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:28:31
118
+ mul.wide.s32 %rd26, %r26, 2;
119
+ add.s64 %rd10, %rd22, %rd26;
120
+ .loc 1 28 36 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:28:36
121
+ // begin inline asm
122
+ mov.u32 %r12, 0x0;
123
+ @%p11 ld.global.L1::evict_last.b32 { %r12 }, [ %rd10 + 0 ];
124
+ // end inline asm
125
+ .loc 1 29 31 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:29:31
126
+ mul.wide.s32 %rd27, %r36, 8;
127
+ add.s64 %rd12, %rd23, %rd27;
128
+ .loc 1 29 36 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:29:36
129
+ // begin inline asm
130
+ mov.u64 %rd11, 0x0;
131
+ @%p11 ld.global.L1::evict_last.b64 { %rd11 }, [ %rd12 + 0 ];
132
+ // end inline asm
133
+ // begin inline asm
134
+ mov.u64 %rd13, 0x0;
135
+ @%p11 ld.global.L1::evict_last.b64 { %rd13 }, [ %rd12 + 0 ];
136
+ // end inline asm
137
+ .loc 1 30 63 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:30:63
138
+ add.s32 %r41, %r26, 64;
139
+ .loc 1 30 43 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:30:43
140
+ add.s32 %r42, %r38, 5184;
141
+ .loc 1 30 77 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:30:77
142
+ add.s32 %r43, %r42, %r26;
143
+ add.s32 %r44, %r42, %r29;
144
+ .loc 1 30 31 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:30:31
145
+ mul.wide.s32 %rd28, %r43, 2;
146
+ add.s64 %rd15, %rd20, %rd28;
147
+ mul.wide.s32 %rd29, %r44, 2;
148
+ add.s64 %rd16, %rd20, %rd29;
149
+ .loc 1 30 87 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:30:87
150
+ // begin inline asm
151
+ mov.u16 %rs3, 0x0;
152
+ @%p11 ld.global.b16 { %rs3 }, [ %rd15 + 0 ];
153
+ // end inline asm
154
+ // begin inline asm
155
+ mov.u16 %rs4, 0x0;
156
+ @%p11 ld.global.b16 { %rs4 }, [ %rd16 + 0 ];
157
+ // end inline asm
158
+ .loc 1 31 57 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:31:57
159
+ // begin inline asm
160
+ mov.u32 %r13, 0x0;
161
+ @%p11 ld.global.L1::evict_last.b32 { %r13 }, [ %rd8 + 0 ];
162
+ // end inline asm
163
+ // begin inline asm
164
+ mov.u32 %r14, 0x0;
165
+ @%p11 ld.global.L1::evict_last.b32 { %r14 }, [ %rd8 + 0 ];
166
+ // end inline asm
167
+ .loc 1 32 31 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:32:31
168
+ mul.wide.u32 %rd30, %r41, 2;
169
+ add.s64 %rd19, %rd22, %rd30;
170
+ .loc 1 32 41 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:32:41
171
+ // begin inline asm
172
+ mov.u32 %r15, 0x0;
173
+ @%p11 ld.global.L1::evict_last.b32 { %r15 }, [ %rd19 + 0 ];
174
+ // end inline asm
175
+ .loc 1 43 20 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:43:20
176
+ add.s64 %rd31, %rd11, 40960;
177
+ .loc 1 44 20 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:44:20
178
+ setp.lt.s64 %p13, %rd11, 0;
179
+ .loc 1 45 35 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:45:35
180
+ selp.b64 %rd3, %rd31, %rd11, %p13;
181
+ .loc 1 46 38 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:46:38
182
+ setp.lt.u64 %p14, %rd3, 40960;
183
+ .loc 1 46 66 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:46:66
184
+ or.pred %p15, %p12, %p14;
185
+ @%p15 bra $L__BB0_2;
186
+ bra.uni $L__BB0_1;
187
+ $L__BB0_2: // %__nv_rsqrtf.exit7
188
+ .loc 1 0 66 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:0:66
189
+ ld.param.u64 %rd6, [triton_poi_fused_add_mul_sub_6_param_6];
190
+ ld.param.u64 %rd5, [triton_poi_fused_add_mul_sub_6_param_5];
191
+ ld.param.u64 %rd4, [triton_poi_fused_add_mul_sub_6_param_4];
192
+ mov.b32 %f3, %r10;
193
+ mov.b32 %f4, %r11;
194
+ cvt.s64.s32 %rd1, %r26;
195
+ cvt.u64.u32 %rd2, %r41;
196
+ mov.f32 %f5, 0f43000000;
197
+ div.full.f32 %f6, %f3, %f5;
198
+ div.full.f32 %f7, %f4, %f5;
199
+ add.f32 %f8, %f6, 0f358637BD;
200
+ add.f32 %f9, %f7, 0f358637BD;
201
+ rsqrt.approx.ftz.f32 %f1, %f8;
202
+ rsqrt.approx.ftz.f32 %f2, %f9;
203
+ cvt.u32.u64 %r49, %rd1;
204
+ .loc 1 31 57 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:31:57
205
+ mov.b32 %f10, %r14;
206
+ mov.b32 %f11, %r13;
207
+ .loc 1 46 66 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:46:66
208
+ bar.sync 0;
209
+ .loc 1 47 31 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:47:31
210
+ shl.b64 %rd36, %rd3, 8;
211
+ add.s64 %rd37, %rd4, %rd36;
212
+ shl.b64 %rd38, %rd1, 1;
213
+ add.s64 %rd32, %rd37, %rd38;
214
+ .loc 1 47 48 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:47:48
215
+ // begin inline asm
216
+ mov.u32 %r45, 0x0;
217
+ @%p11 ld.global.b32 { %r45 }, [ %rd32 + 0 ];
218
+ // end inline asm
219
+ .loc 1 50 21 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:50:21
220
+ div.full.f32 %f13, %f11, %f5;
221
+ div.full.f32 %f14, %f10, %f5;
222
+ .loc 1 51 20 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:51:20
223
+ add.f32 %f15, %f13, 0f358637BD;
224
+ add.f32 %f16, %f14, 0f358637BD;
225
+ .loc 1 52 28 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:52:28
226
+ rsqrt.approx.ftz.f32 %f17, %f15;
227
+ rsqrt.approx.ftz.f32 %f18, %f16;
228
+ .loc 1 30 87 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:30:87
229
+ mov.b32 %r50, {%rs3, %rs4};
230
+ .loc 1 56 31 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:56:31
231
+ shl.b64 %rd39, %rd2, 1;
232
+ add.s64 %rd33, %rd37, %rd39;
233
+ .loc 1 56 53 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:56:53
234
+ // begin inline asm
235
+ mov.u32 %r46, 0x0;
236
+ @%p11 ld.global.b32 { %r46 }, [ %rd33 + 0 ];
237
+ // end inline asm
238
+ .loc 1 62 34 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:62:34
239
+ shl.b32 %r51, %r2, 7;
240
+ .loc 1 62 30 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:62:30
241
+ add.s32 %r52, %r51, %r49;
242
+ .loc 1 62 25 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:62:25
243
+ mul.wide.s32 %rd40, %r52, 2;
244
+ add.s64 %rd34, %rd5, %rd40;
245
+ .loc 1 26 71 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:26:71
246
+ mov.b32 {%rs5, %rs6}, %r9;
247
+ cvt.f32.bf16 %f19, %rs6;
248
+ cvt.f32.bf16 %f20, %rs5;
249
+ .loc 1 39 18 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:39:18
250
+ mul.f32 %f21, %f1, %f20;
251
+ mul.f32 %f22, %f2, %f19;
252
+ .loc 1 28 76 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:28:76
253
+ mov.b32 {%rs7, %rs8}, %r12;
254
+ cvt.f32.bf16 %f23, %rs7;
255
+ cvt.f32.bf16 %f24, %rs8;
256
+ .loc 1 41 19 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:41:19
257
+ mul.f32 %f25, %f22, %f24;
258
+ mul.f32 %f26, %f21, %f23;
259
+ .loc 1 47 58 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:47:58
260
+ mov.b32 {%rs9, %rs10}, %r45;
261
+ cvt.f32.bf16 %f27, %rs10;
262
+ cvt.f32.bf16 %f28, %rs9;
263
+ .loc 1 30 97 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:30:97
264
+ mov.b32 {%rs11, %rs12}, %r50;
265
+ cvt.f32.bf16 %f29, %rs11;
266
+ cvt.f32.bf16 %f30, %rs12;
267
+ .loc 1 32 81 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:32:81
268
+ mov.b32 {%rs13, %rs14}, %r15;
269
+ cvt.f32.bf16 %f31, %rs14;
270
+ cvt.f32.bf16 %f32, %rs13;
271
+ .loc 1 53 20 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:53:20
272
+ mul.f32 %f33, %f18, %f30;
273
+ mul.f32 %f34, %f17, %f29;
274
+ .loc 1 55 20 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:55:20
275
+ mul.f32 %f35, %f34, %f32;
276
+ mul.f32 %f36, %f33, %f31;
277
+ .loc 1 56 63 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:56:63
278
+ mov.b32 {%rs15, %rs16}, %r46;
279
+ cvt.f32.bf16 %f37, %rs15;
280
+ cvt.f32.bf16 %f38, %rs16;
281
+ .loc 1 57 20 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:57:20
282
+ mul.f32 %f39, %f36, %f38;
283
+ mul.f32 %f40, %f35, %f37;
284
+ .loc 1 58 20 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:58:20
285
+ neg.f32 %f41, %f40;
286
+ fma.rn.f32 %f42, %f26, %f28, %f41;
287
+ neg.f32 %f43, %f39;
288
+ fma.rn.f32 %f44, %f25, %f27, %f43;
289
+ .loc 1 62 46 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:62:46
290
+ cvt.rn.bf16x2.f32 %r47, %f44, %f42;
291
+ // begin inline asm
292
+ @%p11 st.global.b32 [ %rd34 + 0 ], { %r47 };
293
+ // end inline asm
294
+ .loc 1 63 25 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:63:25
295
+ add.s64 %rd35, %rd6, %rd40;
296
+ .loc 1 60 20 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:60:20
297
+ mul.f32 %f45, %f25, %f38;
298
+ mul.f32 %f46, %f26, %f37;
299
+ .loc 1 61 20 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:61:20
300
+ fma.rn.f32 %f47, %f35, %f28, %f46;
301
+ fma.rn.f32 %f48, %f36, %f27, %f45;
302
+ .loc 1 63 46 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:63:46
303
+ cvt.rn.bf16x2.f32 %r48, %f48, %f47;
304
+ // begin inline asm
305
+ @%p11 st.global.b32 [ %rd35 + 0 ], { %r48 };
306
+ // end inline asm
307
+ .loc 1 63 4 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:63:4
308
+ ret;
309
+ $L__BB0_1:
310
+ .loc 1 46 66 // cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py:46:66
311
+ mov.u64 %rd41, assertMessage_0;
312
+ cvta.global.u64 %rd42, %rd41;
313
+ mov.u64 %rd43, assertFile_0;
314
+ cvta.global.u64 %rd44, %rd43;
315
+ mov.u64 %rd45, assertFunc_0;
316
+ cvta.global.u64 %rd46, %rd45;
317
+ { // callseq 8, 0
318
+ .param .b64 param0;
319
+ st.param.b64 [param0], %rd42;
320
+ .param .b64 param1;
321
+ st.param.b64 [param1], %rd44;
322
+ .param .b32 param2;
323
+ st.param.b32 [param2], 46;
324
+ .param .b64 param3;
325
+ st.param.b64 [param3], %rd46;
326
+ .param .b64 param4;
327
+ st.param.b64 [param4], 1;
328
+ call.uni
329
+ __assertfail,
330
+ (
331
+ param0,
332
+ param1,
333
+ param2,
334
+ param3,
335
+ param4
336
+ );
337
+ } // callseq 8
338
+ trap;
339
+ $L__tmp1:
340
+ $L__func_end0:
341
+ // -- End function
342
+ }
343
+ .file 1 "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py"
344
+ .section .debug_abbrev
345
+ {
346
+ .b8 1 // Abbreviation Code
347
+ .b8 17 // DW_TAG_compile_unit
348
+ .b8 0 // DW_CHILDREN_no
349
+ .b8 37 // DW_AT_producer
350
+ .b8 8 // DW_FORM_string
351
+ .b8 19 // DW_AT_language
352
+ .b8 5 // DW_FORM_data2
353
+ .b8 3 // DW_AT_name
354
+ .b8 8 // DW_FORM_string
355
+ .b8 16 // DW_AT_stmt_list
356
+ .b8 6 // DW_FORM_data4
357
+ .b8 27 // DW_AT_comp_dir
358
+ .b8 8 // DW_FORM_string
359
+ .b8 0 // EOM(1)
360
+ .b8 0 // EOM(2)
361
+ .b8 0 // EOM(3)
362
+ }
363
+ .section .debug_info
364
+ {
365
+ .b32 155 // Length of Unit
366
+ .b8 2 // DWARF version number
367
+ .b8 0
368
+ .b32 .debug_abbrev // Offset Into Abbrev. Section
369
+ .b8 8 // Address Size (in bytes)
370
+ .b8 1 // Abbrev [1] 0xb:0x94 DW_TAG_compile_unit
371
+ .b8 116 // DW_AT_producer
372
+ .b8 114
373
+ .b8 105
374
+ .b8 116
375
+ .b8 111
376
+ .b8 110
377
+ .b8 0
378
+ .b8 2 // DW_AT_language
379
+ .b8 0
380
+ .b8 99 // DW_AT_name
381
+ .b8 110
382
+ .b8 122
383
+ .b8 120
384
+ .b8 102
385
+ .b8 112
386
+ .b8 115
387
+ .b8 105
388
+ .b8 99
389
+ .b8 107
390
+ .b8 50
391
+ .b8 55
392
+ .b8 54
393
+ .b8 107
394
+ .b8 101
395
+ .b8 107
396
+ .b8 109
397
+ .b8 55
398
+ .b8 116
399
+ .b8 102
400
+ .b8 111
401
+ .b8 119
402
+ .b8 55
403
+ .b8 98
404
+ .b8 108
405
+ .b8 119
406
+ .b8 52
407
+ .b8 52
408
+ .b8 113
409
+ .b8 117
410
+ .b8 54
411
+ .b8 100
412
+ .b8 54
413
+ .b8 104
414
+ .b8 118
415
+ .b8 115
416
+ .b8 101
417
+ .b8 120
418
+ .b8 55
419
+ .b8 105
420
+ .b8 99
421
+ .b8 104
422
+ .b8 117
423
+ .b8 116
424
+ .b8 101
425
+ .b8 51
426
+ .b8 106
427
+ .b8 116
428
+ .b8 119
429
+ .b8 109
430
+ .b8 51
431
+ .b8 97
432
+ .b8 46
433
+ .b8 112
434
+ .b8 121
435
+ .b8 0
436
+ .b32 .debug_line // DW_AT_stmt_list
437
+ .b8 47 // DW_AT_comp_dir
438
+ .b8 104
439
+ .b8 111
440
+ .b8 109
441
+ .b8 101
442
+ .b8 47
443
+ .b8 120
444
+ .b8 47
445
+ .b8 46
446
+ .b8 99
447
+ .b8 97
448
+ .b8 99
449
+ .b8 104
450
+ .b8 101
451
+ .b8 47
452
+ .b8 118
453
+ .b8 108
454
+ .b8 108
455
+ .b8 109
456
+ .b8 47
457
+ .b8 116
458
+ .b8 111
459
+ .b8 114
460
+ .b8 99
461
+ .b8 104
462
+ .b8 95
463
+ .b8 99
464
+ .b8 111
465
+ .b8 109
466
+ .b8 112
467
+ .b8 105
468
+ .b8 108
469
+ .b8 101
470
+ .b8 95
471
+ .b8 99
472
+ .b8 97
473
+ .b8 99
474
+ .b8 104
475
+ .b8 101
476
+ .b8 47
477
+ .b8 57
478
+ .b8 48
479
+ .b8 98
480
+ .b8 52
481
+ .b8 53
482
+ .b8 98
483
+ .b8 99
484
+ .b8 101
485
+ .b8 48
486
+ .b8 50
487
+ .b8 47
488
+ .b8 114
489
+ .b8 97
490
+ .b8 110
491
+ .b8 107
492
+ .b8 95
493
+ .b8 48
494
+ .b8 95
495
+ .b8 48
496
+ .b8 47
497
+ .b8 105
498
+ .b8 110
499
+ .b8 100
500
+ .b8 117
501
+ .b8 99
502
+ .b8 116
503
+ .b8 111
504
+ .b8 114
505
+ .b8 95
506
+ .b8 99
507
+ .b8 97
508
+ .b8 99
509
+ .b8 104
510
+ .b8 101
511
+ .b8 47
512
+ .b8 110
513
+ .b8 122
514
+ .b8 0
515
+ }
516
+ .section .debug_macinfo { }
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/CUASPVXQ7OI3JPFN55O2D57GL57SRLSB6LBTAARTQDPYHVXGYBSA/triton_poi_fused_add_mul_sub_6.ttgir ADDED
@@ -0,0 +1,196 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #blocked = #ttg.blocked<{sizePerThread = [2], threadsPerWarp = [32], warpsPerCTA = [8], order = [0]}>
2
+ #loc = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0)
3
+ module attributes {"ttg.num-ctas" = 1 : i32, "ttg.num-warps" = 8 : i32, ttg.target = "cuda:90", "ttg.threads-per-warp" = 32 : i32} {
4
+ tt.func public @triton_poi_fused_add_mul_sub_6(%arg0: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg1: !tt.ptr<f32> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg2: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg3: !tt.ptr<i64> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg4: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg5: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg6: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg7: i32 {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0)) attributes {noinline = false} {
5
+ %cst = arith.constant dense<1.280000e+02> : tensor<512xf32, #blocked> loc(#loc1)
6
+ %cst_0 = arith.constant dense<9.99999997E-7> : tensor<512xf32, #blocked> loc(#loc1)
7
+ %cst_1 = arith.constant dense<128> : tensor<512xi64, #blocked> loc(#loc1)
8
+ %cst_2 = arith.constant dense<40960> : tensor<512xi64, #blocked> loc(#loc1)
9
+ %cst_3 = arith.constant dense<0> : tensor<512xi64, #blocked> loc(#loc1)
10
+ %c512_i32 = arith.constant 512 : i32 loc(#loc1)
11
+ %cst_4 = arith.constant dense<64> : tensor<512xi32, #blocked> loc(#loc1)
12
+ %cst_5 = arith.constant dense<8> : tensor<512xi32, #blocked> loc(#loc1)
13
+ %cst_6 = arith.constant dense<512> : tensor<512xi32, #blocked> loc(#loc1)
14
+ %cst_7 = arith.constant dense<5120> : tensor<512xi32, #blocked> loc(#loc1)
15
+ %cst_8 = arith.constant dense<128> : tensor<512xi32, #blocked> loc(#loc1)
16
+ %cst_9 = arith.constant dense<7168> : tensor<512xi32, #blocked> loc(#loc1)
17
+ %cst_10 = arith.constant dense<5184> : tensor<512xi32, #blocked> loc(#loc1)
18
+ %cst_11 = arith.constant dense<true> : tensor<512xi1, #blocked> loc(#loc1)
19
+ %0 = tt.get_program_id x : i32 loc(#loc2)
20
+ %1 = arith.muli %0, %c512_i32 : i32 loc(#loc3)
21
+ %2 = tt.make_range {end = 512 : i32, start = 0 : i32} : tensor<512xi32, #blocked> loc(#loc4)
22
+ %3 = tt.splat %1 : i32 -> tensor<512xi32, #blocked> loc(#loc5)
23
+ %4 = arith.addi %3, %2 : tensor<512xi32, #blocked> loc(#loc5)
24
+ %5 = tt.splat %arg7 : i32 -> tensor<512xi32, #blocked> loc(#loc6)
25
+ %6 = arith.cmpi slt, %4, %5 : tensor<512xi32, #blocked> loc(#loc6)
26
+ %7 = arith.remsi %4, %cst_4 : tensor<512xi32, #blocked> loc(#loc7)
27
+ %8 = arith.divsi %4, %cst_4 : tensor<512xi32, #blocked> loc(#loc8)
28
+ %9 = arith.remsi %8, %cst_5 : tensor<512xi32, #blocked> loc(#loc9)
29
+ %10 = arith.divsi %4, %cst_6 : tensor<512xi32, #blocked> loc(#loc10)
30
+ %11 = arith.addi %7, %cst_7 : tensor<512xi32, #blocked> loc(#loc11)
31
+ %12 = arith.muli %9, %cst_8 : tensor<512xi32, #blocked> loc(#loc12)
32
+ %13 = arith.addi %11, %12 : tensor<512xi32, #blocked> loc(#loc13)
33
+ %14 = arith.muli %10, %cst_9 : tensor<512xi32, #blocked> loc(#loc14)
34
+ %15 = arith.addi %13, %14 : tensor<512xi32, #blocked> loc(#loc15)
35
+ %16 = tt.splat %arg0 : !tt.ptr<bf16> -> tensor<512x!tt.ptr<bf16>, #blocked> loc(#loc16)
36
+ %17 = tt.addptr %16, %15 : tensor<512x!tt.ptr<bf16>, #blocked>, tensor<512xi32, #blocked> loc(#loc16)
37
+ %18 = tt.load %17, %6 : tensor<512x!tt.ptr<bf16>, #blocked> loc(#loc17)
38
+ %19 = arith.extf %18 : tensor<512xbf16, #blocked> to tensor<512xf32, #blocked> loc(#loc18)
39
+ %20 = tt.splat %arg1 : !tt.ptr<f32> -> tensor<512x!tt.ptr<f32>, #blocked> loc(#loc19)
40
+ %21 = tt.addptr %20, %8 : tensor<512x!tt.ptr<f32>, #blocked>, tensor<512xi32, #blocked> loc(#loc19)
41
+ %22 = tt.load %21, %6 evictionPolicy = evict_last : tensor<512x!tt.ptr<f32>, #blocked> loc(#loc20)
42
+ %23 = tt.splat %arg2 : !tt.ptr<bf16> -> tensor<512x!tt.ptr<bf16>, #blocked> loc(#loc21)
43
+ %24 = tt.addptr %23, %7 : tensor<512x!tt.ptr<bf16>, #blocked>, tensor<512xi32, #blocked> loc(#loc21)
44
+ %25 = tt.load %24, %6 evictionPolicy = evict_last : tensor<512x!tt.ptr<bf16>, #blocked> loc(#loc22)
45
+ %26 = arith.extf %25 : tensor<512xbf16, #blocked> to tensor<512xf32, #blocked> loc(#loc23)
46
+ %27 = tt.splat %arg3 : !tt.ptr<i64> -> tensor<512x!tt.ptr<i64>, #blocked> loc(#loc24)
47
+ %28 = tt.addptr %27, %10 : tensor<512x!tt.ptr<i64>, #blocked>, tensor<512xi32, #blocked> loc(#loc24)
48
+ %29 = tt.load %28, %6 evictionPolicy = evict_last : tensor<512x!tt.ptr<i64>, #blocked> loc(#loc25)
49
+ %30 = arith.addi %7, %cst_10 : tensor<512xi32, #blocked> loc(#loc26)
50
+ %31 = arith.addi %30, %12 : tensor<512xi32, #blocked> loc(#loc27)
51
+ %32 = arith.addi %7, %cst_4 : tensor<512xi32, #blocked> loc(#loc28)
52
+ %33 = arith.divsi %32, %cst_8 : tensor<512xi32, #blocked> loc(#loc29)
53
+ %34 = arith.muli %33, %cst_8 : tensor<512xi32, #blocked> loc(#loc30)
54
+ %35 = arith.addi %31, %34 : tensor<512xi32, #blocked> loc(#loc31)
55
+ %36 = arith.addi %35, %14 : tensor<512xi32, #blocked> loc(#loc32)
56
+ %37 = tt.addptr %16, %36 : tensor<512x!tt.ptr<bf16>, #blocked>, tensor<512xi32, #blocked> loc(#loc33)
57
+ %38 = tt.load %37, %6 : tensor<512x!tt.ptr<bf16>, #blocked> loc(#loc34)
58
+ %39 = arith.extf %38 : tensor<512xbf16, #blocked> to tensor<512xf32, #blocked> loc(#loc35)
59
+ %40 = arith.addi %8, %33 : tensor<512xi32, #blocked> loc(#loc36)
60
+ %41 = tt.addptr %20, %40 : tensor<512x!tt.ptr<f32>, #blocked>, tensor<512xi32, #blocked> loc(#loc37)
61
+ %42 = tt.load %41, %6 evictionPolicy = evict_last : tensor<512x!tt.ptr<f32>, #blocked> loc(#loc38)
62
+ %43 = tt.addptr %23, %32 : tensor<512x!tt.ptr<bf16>, #blocked>, tensor<512xi32, #blocked> loc(#loc39)
63
+ %44 = tt.load %43, %6 evictionPolicy = evict_last : tensor<512x!tt.ptr<bf16>, #blocked> loc(#loc40)
64
+ %45 = arith.extf %44 : tensor<512xbf16, #blocked> to tensor<512xf32, #blocked> loc(#loc41)
65
+ %46 = arith.divf %22, %cst : tensor<512xf32, #blocked> loc(#loc42)
66
+ %47 = arith.addf %46, %cst_0 : tensor<512xf32, #blocked> loc(#loc43)
67
+ %48 = tt.extern_elementwise %47 {libname = "", libpath = "", pure = true, symbol = "__nv_rsqrtf"} : (tensor<512xf32, #blocked>) -> tensor<512xf32, #blocked> loc(#loc44)
68
+ %49 = arith.mulf %19, %48 : tensor<512xf32, #blocked> loc(#loc45)
69
+ %50 = arith.mulf %49, %26 : tensor<512xf32, #blocked> loc(#loc46)
70
+ %51 = arith.addi %29, %cst_2 : tensor<512xi64, #blocked> loc(#loc47)
71
+ %52 = arith.cmpi slt, %29, %cst_3 : tensor<512xi64, #blocked> loc(#loc48)
72
+ %53 = arith.select %52, %51, %29 : tensor<512xi1, #blocked>, tensor<512xi64, #blocked> loc(#loc49)
73
+ %54 = arith.cmpi sge, %53, %cst_3 : tensor<512xi64, #blocked> loc(#loc50)
74
+ %55 = arith.cmpi slt, %53, %cst_2 : tensor<512xi64, #blocked> loc(#loc51)
75
+ %56 = arith.andi %54, %55 : tensor<512xi1, #blocked> loc(#loc52)
76
+ %57 = arith.xori %6, %cst_11 : tensor<512xi1, #blocked> loc(#loc53)
77
+ %58 = arith.ori %56, %57 : tensor<512xi1, #blocked> loc(#loc54)
78
+ tt.assert %58, "index out of bounds: 0 <= tmp16 < 40960" : tensor<512xi1, #blocked> loc(#loc55)
79
+ %59 = arith.muli %53, %cst_1 : tensor<512xi64, #blocked> loc(#loc56)
80
+ %60 = arith.extsi %7 : tensor<512xi32, #blocked> to tensor<512xi64, #blocked> loc(#loc57)
81
+ %61 = arith.addi %60, %59 : tensor<512xi64, #blocked> loc(#loc57)
82
+ %62 = tt.splat %arg4 : !tt.ptr<bf16> -> tensor<512x!tt.ptr<bf16>, #blocked> loc(#loc58)
83
+ %63 = tt.addptr %62, %61 : tensor<512x!tt.ptr<bf16>, #blocked>, tensor<512xi64, #blocked> loc(#loc58)
84
+ %64 = tt.load %63, %6 : tensor<512x!tt.ptr<bf16>, #blocked> loc(#loc59)
85
+ %65 = arith.extf %64 : tensor<512xbf16, #blocked> to tensor<512xf32, #blocked> loc(#loc60)
86
+ %66 = arith.mulf %50, %65 : tensor<512xf32, #blocked> loc(#loc61)
87
+ %67 = arith.divf %42, %cst : tensor<512xf32, #blocked> loc(#loc62)
88
+ %68 = arith.addf %67, %cst_0 : tensor<512xf32, #blocked> loc(#loc63)
89
+ %69 = tt.extern_elementwise %68 {libname = "", libpath = "", pure = true, symbol = "__nv_rsqrtf"} : (tensor<512xf32, #blocked>) -> tensor<512xf32, #blocked> loc(#loc64)
90
+ %70 = arith.mulf %39, %69 : tensor<512xf32, #blocked> loc(#loc65)
91
+ %71 = arith.mulf %70, %45 : tensor<512xf32, #blocked> loc(#loc66)
92
+ %72 = arith.extsi %32 : tensor<512xi32, #blocked> to tensor<512xi64, #blocked> loc(#loc67)
93
+ %73 = arith.addi %72, %59 : tensor<512xi64, #blocked> loc(#loc67)
94
+ %74 = tt.addptr %62, %73 : tensor<512x!tt.ptr<bf16>, #blocked>, tensor<512xi64, #blocked> loc(#loc68)
95
+ %75 = tt.load %74, %6 : tensor<512x!tt.ptr<bf16>, #blocked> loc(#loc69)
96
+ %76 = arith.extf %75 : tensor<512xbf16, #blocked> to tensor<512xf32, #blocked> loc(#loc70)
97
+ %77 = arith.mulf %71, %76 : tensor<512xf32, #blocked> loc(#loc71)
98
+ %78 = arith.subf %66, %77 : tensor<512xf32, #blocked> loc(#loc72)
99
+ %79 = arith.mulf %71, %65 : tensor<512xf32, #blocked> loc(#loc73)
100
+ %80 = arith.mulf %50, %76 : tensor<512xf32, #blocked> loc(#loc74)
101
+ %81 = arith.addf %79, %80 : tensor<512xf32, #blocked> loc(#loc75)
102
+ %82 = arith.muli %8, %cst_8 : tensor<512xi32, #blocked> loc(#loc76)
103
+ %83 = arith.addi %7, %82 : tensor<512xi32, #blocked> loc(#loc77)
104
+ %84 = tt.splat %arg5 : !tt.ptr<bf16> -> tensor<512x!tt.ptr<bf16>, #blocked> loc(#loc78)
105
+ %85 = tt.addptr %84, %83 : tensor<512x!tt.ptr<bf16>, #blocked>, tensor<512xi32, #blocked> loc(#loc78)
106
+ %86 = arith.truncf %78 : tensor<512xf32, #blocked> to tensor<512xbf16, #blocked> loc(#loc79)
107
+ tt.store %85, %86, %6 : tensor<512x!tt.ptr<bf16>, #blocked> loc(#loc79)
108
+ %87 = tt.splat %arg6 : !tt.ptr<bf16> -> tensor<512x!tt.ptr<bf16>, #blocked> loc(#loc80)
109
+ %88 = tt.addptr %87, %83 : tensor<512x!tt.ptr<bf16>, #blocked>, tensor<512xi32, #blocked> loc(#loc80)
110
+ %89 = arith.truncf %81 : tensor<512xf32, #blocked> to tensor<512xbf16, #blocked> loc(#loc81)
111
+ tt.store %88, %89, %6 : tensor<512x!tt.ptr<bf16>, #blocked> loc(#loc81)
112
+ tt.return loc(#loc82)
113
+ } loc(#loc)
114
+ } loc(#loc)
115
+ #loc1 = loc(unknown)
116
+ #loc2 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":19:28)
117
+ #loc3 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":19:33)
118
+ #loc4 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":20:36)
119
+ #loc5 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":20:23)
120
+ #loc6 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":21:21)
121
+ #loc7 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":22:19)
122
+ #loc8 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":23:21)
123
+ #loc9 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":23:27)
124
+ #loc10 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":24:19)
125
+ #loc11 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:37)
126
+ #loc12 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:46)
127
+ #loc13 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:42)
128
+ #loc14 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:56)
129
+ #loc15 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:51)
130
+ #loc16 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:30)
131
+ #loc17 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:61)
132
+ #loc18 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:71)
133
+ #loc19 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":27:30)
134
+ #loc20 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":27:35)
135
+ #loc21 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":28:31)
136
+ #loc22 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":28:36)
137
+ #loc23 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":28:76)
138
+ #loc24 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":29:31)
139
+ #loc25 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":29:36)
140
+ #loc26 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:38)
141
+ #loc27 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:43)
142
+ #loc28 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:63)
143
+ #loc29 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:70)
144
+ #loc30 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:57)
145
+ #loc31 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:52)
146
+ #loc32 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:77)
147
+ #loc33 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:31)
148
+ #loc34 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:87)
149
+ #loc35 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:97)
150
+ #loc36 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":31:37)
151
+ #loc37 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":31:31)
152
+ #loc38 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":31:57)
153
+ #loc39 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":32:31)
154
+ #loc40 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":32:41)
155
+ #loc41 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":32:81)
156
+ #loc42 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":35:19)
157
+ #loc43 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":37:18)
158
+ #loc44 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":38:27)
159
+ #loc45 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":39:18)
160
+ #loc46 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":41:19)
161
+ #loc47 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":43:20)
162
+ #loc48 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":44:20)
163
+ #loc49 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":45:35)
164
+ #loc50 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:28)
165
+ #loc51 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:46)
166
+ #loc52 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:38)
167
+ #loc53 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:58)
168
+ #loc54 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:56)
169
+ #loc55 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:66)
170
+ #loc56 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:40)
171
+ #loc57 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:36)
172
+ #loc58 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:31)
173
+ #loc59 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:48)
174
+ #loc60 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:58)
175
+ #loc61 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":48:20)
176
+ #loc62 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":50:21)
177
+ #loc63 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":51:20)
178
+ #loc64 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":52:28)
179
+ #loc65 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":53:20)
180
+ #loc66 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":55:20)
181
+ #loc67 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":56:41)
182
+ #loc68 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":56:31)
183
+ #loc69 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":56:53)
184
+ #loc70 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":56:63)
185
+ #loc71 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":57:20)
186
+ #loc72 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":58:20)
187
+ #loc73 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":59:20)
188
+ #loc74 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":60:20)
189
+ #loc75 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":61:20)
190
+ #loc76 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":62:34)
191
+ #loc77 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":62:30)
192
+ #loc78 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":62:25)
193
+ #loc79 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":62:46)
194
+ #loc80 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":63:25)
195
+ #loc81 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":63:46)
196
+ #loc82 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":63:4)
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/CUASPVXQ7OI3JPFN55O2D57GL57SRLSB6LBTAARTQDPYHVXGYBSA/triton_poi_fused_add_mul_sub_6.ttir ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #loc = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0)
2
+ module {
3
+ tt.func public @triton_poi_fused_add_mul_sub_6(%arg0: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg1: !tt.ptr<f32> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg2: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg3: !tt.ptr<i64> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg4: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg5: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg6: !tt.ptr<bf16> {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0), %arg7: i32 {tt.divisibility = 16 : i32} loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":18:0)) attributes {noinline = false} {
4
+ %cst = arith.constant dense<0> : tensor<512xi64> loc(#loc1)
5
+ %cst_0 = arith.constant dense<40960> : tensor<512xi64> loc(#loc1)
6
+ %cst_1 = arith.constant dense<128> : tensor<512xi64> loc(#loc1)
7
+ %cst_2 = arith.constant dense<true> : tensor<512xi1> loc(#loc1)
8
+ %cst_3 = arith.constant dense<9.99999997E-7> : tensor<512xf32> loc(#loc1)
9
+ %cst_4 = arith.constant dense<1.280000e+02> : tensor<512xf32> loc(#loc1)
10
+ %cst_5 = arith.constant dense<5184> : tensor<512xi32> loc(#loc1)
11
+ %cst_6 = arith.constant dense<7168> : tensor<512xi32> loc(#loc1)
12
+ %cst_7 = arith.constant dense<128> : tensor<512xi32> loc(#loc1)
13
+ %cst_8 = arith.constant dense<5120> : tensor<512xi32> loc(#loc1)
14
+ %cst_9 = arith.constant dense<512> : tensor<512xi32> loc(#loc1)
15
+ %cst_10 = arith.constant dense<8> : tensor<512xi32> loc(#loc1)
16
+ %cst_11 = arith.constant dense<64> : tensor<512xi32> loc(#loc1)
17
+ %c512_i32 = arith.constant 512 : i32 loc(#loc1)
18
+ %0 = tt.get_program_id x : i32 loc(#loc2)
19
+ %1 = arith.muli %0, %c512_i32 : i32 loc(#loc3)
20
+ %2 = tt.make_range {end = 512 : i32, start = 0 : i32} : tensor<512xi32> loc(#loc4)
21
+ %3 = tt.splat %1 : i32 -> tensor<512xi32> loc(#loc5)
22
+ %4 = arith.addi %3, %2 : tensor<512xi32> loc(#loc5)
23
+ %5 = tt.splat %arg7 : i32 -> tensor<512xi32> loc(#loc6)
24
+ %6 = arith.cmpi slt, %4, %5 : tensor<512xi32> loc(#loc6)
25
+ %7 = arith.remsi %4, %cst_11 : tensor<512xi32> loc(#loc7)
26
+ %8 = arith.divsi %4, %cst_11 : tensor<512xi32> loc(#loc8)
27
+ %9 = arith.remsi %8, %cst_10 : tensor<512xi32> loc(#loc9)
28
+ %10 = arith.divsi %4, %cst_9 : tensor<512xi32> loc(#loc10)
29
+ %11 = arith.addi %7, %cst_8 : tensor<512xi32> loc(#loc11)
30
+ %12 = arith.muli %9, %cst_7 : tensor<512xi32> loc(#loc12)
31
+ %13 = arith.addi %11, %12 : tensor<512xi32> loc(#loc13)
32
+ %14 = arith.muli %10, %cst_6 : tensor<512xi32> loc(#loc14)
33
+ %15 = arith.addi %13, %14 : tensor<512xi32> loc(#loc15)
34
+ %16 = tt.splat %arg0 : !tt.ptr<bf16> -> tensor<512x!tt.ptr<bf16>> loc(#loc16)
35
+ %17 = tt.addptr %16, %15 : tensor<512x!tt.ptr<bf16>>, tensor<512xi32> loc(#loc16)
36
+ %18 = tt.load %17, %6 : tensor<512x!tt.ptr<bf16>> loc(#loc17)
37
+ %19 = arith.extf %18 : tensor<512xbf16> to tensor<512xf32> loc(#loc18)
38
+ %20 = tt.splat %arg1 : !tt.ptr<f32> -> tensor<512x!tt.ptr<f32>> loc(#loc19)
39
+ %21 = tt.addptr %20, %8 : tensor<512x!tt.ptr<f32>>, tensor<512xi32> loc(#loc19)
40
+ %22 = tt.load %21, %6 evictionPolicy = evict_last : tensor<512x!tt.ptr<f32>> loc(#loc20)
41
+ %23 = tt.splat %arg2 : !tt.ptr<bf16> -> tensor<512x!tt.ptr<bf16>> loc(#loc21)
42
+ %24 = tt.addptr %23, %7 : tensor<512x!tt.ptr<bf16>>, tensor<512xi32> loc(#loc21)
43
+ %25 = tt.load %24, %6 evictionPolicy = evict_last : tensor<512x!tt.ptr<bf16>> loc(#loc22)
44
+ %26 = arith.extf %25 : tensor<512xbf16> to tensor<512xf32> loc(#loc23)
45
+ %27 = tt.splat %arg3 : !tt.ptr<i64> -> tensor<512x!tt.ptr<i64>> loc(#loc24)
46
+ %28 = tt.addptr %27, %10 : tensor<512x!tt.ptr<i64>>, tensor<512xi32> loc(#loc24)
47
+ %29 = tt.load %28, %6 evictionPolicy = evict_last : tensor<512x!tt.ptr<i64>> loc(#loc25)
48
+ %30 = arith.addi %7, %cst_5 : tensor<512xi32> loc(#loc26)
49
+ %31 = arith.addi %30, %12 : tensor<512xi32> loc(#loc27)
50
+ %32 = arith.addi %7, %cst_11 : tensor<512xi32> loc(#loc28)
51
+ %33 = arith.divsi %32, %cst_7 : tensor<512xi32> loc(#loc29)
52
+ %34 = arith.muli %33, %cst_7 : tensor<512xi32> loc(#loc30)
53
+ %35 = arith.addi %31, %34 : tensor<512xi32> loc(#loc31)
54
+ %36 = arith.addi %35, %14 : tensor<512xi32> loc(#loc32)
55
+ %37 = tt.addptr %16, %36 : tensor<512x!tt.ptr<bf16>>, tensor<512xi32> loc(#loc33)
56
+ %38 = tt.load %37, %6 : tensor<512x!tt.ptr<bf16>> loc(#loc34)
57
+ %39 = arith.extf %38 : tensor<512xbf16> to tensor<512xf32> loc(#loc35)
58
+ %40 = arith.addi %8, %33 : tensor<512xi32> loc(#loc36)
59
+ %41 = tt.addptr %20, %40 : tensor<512x!tt.ptr<f32>>, tensor<512xi32> loc(#loc37)
60
+ %42 = tt.load %41, %6 evictionPolicy = evict_last : tensor<512x!tt.ptr<f32>> loc(#loc38)
61
+ %43 = tt.addptr %23, %32 : tensor<512x!tt.ptr<bf16>>, tensor<512xi32> loc(#loc39)
62
+ %44 = tt.load %43, %6 evictionPolicy = evict_last : tensor<512x!tt.ptr<bf16>> loc(#loc40)
63
+ %45 = arith.extf %44 : tensor<512xbf16> to tensor<512xf32> loc(#loc41)
64
+ %46 = arith.divf %22, %cst_4 : tensor<512xf32> loc(#loc42)
65
+ %47 = arith.addf %46, %cst_3 : tensor<512xf32> loc(#loc43)
66
+ %48 = tt.extern_elementwise %47 {libname = "", libpath = "", pure = true, symbol = "__nv_rsqrtf"} : (tensor<512xf32>) -> tensor<512xf32> loc(#loc44)
67
+ %49 = arith.mulf %19, %48 : tensor<512xf32> loc(#loc45)
68
+ %50 = arith.mulf %49, %26 : tensor<512xf32> loc(#loc46)
69
+ %51 = arith.addi %29, %cst_0 : tensor<512xi64> loc(#loc47)
70
+ %52 = arith.cmpi slt, %29, %cst : tensor<512xi64> loc(#loc48)
71
+ %53 = arith.select %52, %51, %29 : tensor<512xi1>, tensor<512xi64> loc(#loc49)
72
+ %54 = arith.cmpi sge, %53, %cst : tensor<512xi64> loc(#loc50)
73
+ %55 = arith.cmpi slt, %53, %cst_0 : tensor<512xi64> loc(#loc51)
74
+ %56 = arith.andi %54, %55 : tensor<512xi1> loc(#loc52)
75
+ %57 = arith.xori %6, %cst_2 : tensor<512xi1> loc(#loc53)
76
+ %58 = arith.ori %56, %57 : tensor<512xi1> loc(#loc54)
77
+ tt.assert %58, "index out of bounds: 0 <= tmp16 < 40960" : tensor<512xi1> loc(#loc55)
78
+ %59 = arith.muli %53, %cst_1 : tensor<512xi64> loc(#loc56)
79
+ %60 = arith.extsi %7 : tensor<512xi32> to tensor<512xi64> loc(#loc57)
80
+ %61 = arith.addi %60, %59 : tensor<512xi64> loc(#loc57)
81
+ %62 = tt.splat %arg4 : !tt.ptr<bf16> -> tensor<512x!tt.ptr<bf16>> loc(#loc58)
82
+ %63 = tt.addptr %62, %61 : tensor<512x!tt.ptr<bf16>>, tensor<512xi64> loc(#loc58)
83
+ %64 = tt.load %63, %6 : tensor<512x!tt.ptr<bf16>> loc(#loc59)
84
+ %65 = arith.extf %64 : tensor<512xbf16> to tensor<512xf32> loc(#loc60)
85
+ %66 = arith.mulf %50, %65 : tensor<512xf32> loc(#loc61)
86
+ %67 = arith.divf %42, %cst_4 : tensor<512xf32> loc(#loc62)
87
+ %68 = arith.addf %67, %cst_3 : tensor<512xf32> loc(#loc63)
88
+ %69 = tt.extern_elementwise %68 {libname = "", libpath = "", pure = true, symbol = "__nv_rsqrtf"} : (tensor<512xf32>) -> tensor<512xf32> loc(#loc64)
89
+ %70 = arith.mulf %39, %69 : tensor<512xf32> loc(#loc65)
90
+ %71 = arith.mulf %70, %45 : tensor<512xf32> loc(#loc66)
91
+ %72 = arith.extsi %32 : tensor<512xi32> to tensor<512xi64> loc(#loc67)
92
+ %73 = arith.addi %72, %59 : tensor<512xi64> loc(#loc67)
93
+ %74 = tt.addptr %62, %73 : tensor<512x!tt.ptr<bf16>>, tensor<512xi64> loc(#loc68)
94
+ %75 = tt.load %74, %6 : tensor<512x!tt.ptr<bf16>> loc(#loc69)
95
+ %76 = arith.extf %75 : tensor<512xbf16> to tensor<512xf32> loc(#loc70)
96
+ %77 = arith.mulf %71, %76 : tensor<512xf32> loc(#loc71)
97
+ %78 = arith.subf %66, %77 : tensor<512xf32> loc(#loc72)
98
+ %79 = arith.mulf %71, %65 : tensor<512xf32> loc(#loc73)
99
+ %80 = arith.mulf %50, %76 : tensor<512xf32> loc(#loc74)
100
+ %81 = arith.addf %79, %80 : tensor<512xf32> loc(#loc75)
101
+ %82 = arith.muli %8, %cst_7 : tensor<512xi32> loc(#loc76)
102
+ %83 = arith.addi %7, %82 : tensor<512xi32> loc(#loc77)
103
+ %84 = tt.splat %arg5 : !tt.ptr<bf16> -> tensor<512x!tt.ptr<bf16>> loc(#loc78)
104
+ %85 = tt.addptr %84, %83 : tensor<512x!tt.ptr<bf16>>, tensor<512xi32> loc(#loc78)
105
+ %86 = arith.truncf %78 : tensor<512xf32> to tensor<512xbf16> loc(#loc79)
106
+ tt.store %85, %86, %6 : tensor<512x!tt.ptr<bf16>> loc(#loc79)
107
+ %87 = tt.splat %arg6 : !tt.ptr<bf16> -> tensor<512x!tt.ptr<bf16>> loc(#loc80)
108
+ %88 = tt.addptr %87, %83 : tensor<512x!tt.ptr<bf16>>, tensor<512xi32> loc(#loc80)
109
+ %89 = arith.truncf %81 : tensor<512xf32> to tensor<512xbf16> loc(#loc81)
110
+ tt.store %88, %89, %6 : tensor<512x!tt.ptr<bf16>> loc(#loc81)
111
+ tt.return loc(#loc82)
112
+ } loc(#loc)
113
+ } loc(#loc)
114
+ #loc1 = loc(unknown)
115
+ #loc2 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":19:28)
116
+ #loc3 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":19:33)
117
+ #loc4 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":20:36)
118
+ #loc5 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":20:23)
119
+ #loc6 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":21:21)
120
+ #loc7 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":22:19)
121
+ #loc8 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":23:21)
122
+ #loc9 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":23:27)
123
+ #loc10 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":24:19)
124
+ #loc11 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:37)
125
+ #loc12 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:46)
126
+ #loc13 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:42)
127
+ #loc14 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:56)
128
+ #loc15 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:51)
129
+ #loc16 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:30)
130
+ #loc17 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:61)
131
+ #loc18 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":26:71)
132
+ #loc19 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":27:30)
133
+ #loc20 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":27:35)
134
+ #loc21 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":28:31)
135
+ #loc22 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":28:36)
136
+ #loc23 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":28:76)
137
+ #loc24 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":29:31)
138
+ #loc25 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":29:36)
139
+ #loc26 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:38)
140
+ #loc27 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:43)
141
+ #loc28 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:63)
142
+ #loc29 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:70)
143
+ #loc30 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:57)
144
+ #loc31 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:52)
145
+ #loc32 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:77)
146
+ #loc33 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:31)
147
+ #loc34 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:87)
148
+ #loc35 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":30:97)
149
+ #loc36 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":31:37)
150
+ #loc37 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":31:31)
151
+ #loc38 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":31:57)
152
+ #loc39 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":32:31)
153
+ #loc40 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":32:41)
154
+ #loc41 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":32:81)
155
+ #loc42 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":35:19)
156
+ #loc43 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":37:18)
157
+ #loc44 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":38:27)
158
+ #loc45 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":39:18)
159
+ #loc46 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":41:19)
160
+ #loc47 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":43:20)
161
+ #loc48 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":44:20)
162
+ #loc49 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":45:35)
163
+ #loc50 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:28)
164
+ #loc51 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:46)
165
+ #loc52 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:38)
166
+ #loc53 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:58)
167
+ #loc54 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:56)
168
+ #loc55 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":46:66)
169
+ #loc56 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:40)
170
+ #loc57 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:36)
171
+ #loc58 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:31)
172
+ #loc59 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:48)
173
+ #loc60 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":47:58)
174
+ #loc61 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":48:20)
175
+ #loc62 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":50:21)
176
+ #loc63 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":51:20)
177
+ #loc64 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":52:28)
178
+ #loc65 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":53:20)
179
+ #loc66 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":55:20)
180
+ #loc67 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":56:41)
181
+ #loc68 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":56:31)
182
+ #loc69 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":56:53)
183
+ #loc70 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":56:63)
184
+ #loc71 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":57:20)
185
+ #loc72 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":58:20)
186
+ #loc73 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":59:20)
187
+ #loc74 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":60:20)
188
+ #loc75 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":61:20)
189
+ #loc76 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":62:34)
190
+ #loc77 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":62:30)
191
+ #loc78 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":62:25)
192
+ #loc79 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":62:46)
193
+ #loc80 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":63:25)
194
+ #loc81 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":63:46)
195
+ #loc82 = loc("/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/inductor_cache/nz/cnzxfpsick276kekm7tfow7blw44qu6d6hvsex7ichute3jtwm3a.py":63:4)
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/D6B7FVPYPXXUJ4C3PXEACCPD4ZWL5EA6LGYT663ZACLUNJHJ5LAQ/__grp__triton_poi_fused_add_mul_sub_5.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"child_paths": {"triton_poi_fused_add_mul_sub_5.ttir": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/D6B7FVPYPXXUJ4C3PXEACCPD4ZWL5EA6LGYT663ZACLUNJHJ5LAQ/triton_poi_fused_add_mul_sub_5.ttir", "triton_poi_fused_add_mul_sub_5.ttgir": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/D6B7FVPYPXXUJ4C3PXEACCPD4ZWL5EA6LGYT663ZACLUNJHJ5LAQ/triton_poi_fused_add_mul_sub_5.ttgir", "triton_poi_fused_add_mul_sub_5.llir": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/D6B7FVPYPXXUJ4C3PXEACCPD4ZWL5EA6LGYT663ZACLUNJHJ5LAQ/triton_poi_fused_add_mul_sub_5.llir", "triton_poi_fused_add_mul_sub_5.ptx": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/D6B7FVPYPXXUJ4C3PXEACCPD4ZWL5EA6LGYT663ZACLUNJHJ5LAQ/triton_poi_fused_add_mul_sub_5.ptx", "triton_poi_fused_add_mul_sub_5.cubin": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/D6B7FVPYPXXUJ4C3PXEACCPD4ZWL5EA6LGYT663ZACLUNJHJ5LAQ/triton_poi_fused_add_mul_sub_5.cubin", "triton_poi_fused_add_mul_sub_5.json": "/home/x/.cache/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/D6B7FVPYPXXUJ4C3PXEACCPD4ZWL5EA6LGYT663ZACLUNJHJ5LAQ/triton_poi_fused_add_mul_sub_5.json"}}
platform/aiml/models/vllm/torch_compile_cache/90b45bce02/rank_0_0/triton_cache/D6B7FVPYPXXUJ4C3PXEACCPD4ZWL5EA6LGYT663ZACLUNJHJ5LAQ/triton_poi_fused_add_mul_sub_5.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"hash": "1f83f2d5f87def44f05b7dc80109e3e66cbe901e59b13f7b79009746a4e9eac1", "target": {"backend": "cuda", "arch": 90, "warp_size": 32}, "num_warps": 8, "num_ctas": 1, "num_stages": 1, "num_buffers_warp_spec": 0, "num_consumer_groups": 0, "reg_dec_producer": 0, "reg_inc_consumer": 0, "maxnreg": null, "cluster_dims": [1, 1, 1], "ptx_version": null, "enable_fp_fusion": true, "launch_cooperative_grid": false, "supported_fp8_dtypes": ["fp8e4b15", "fp8e4nv", "fp8e5"], "deprecated_fp8_dtypes": ["fp8e4b15"], "default_dot_input_precision": "tf32", "allowed_dot_input_precisions": ["tf32", "tf32x3", "ieee"], "max_num_imprecise_acc_default": 1073741824, "extern_libs": [["libdevice", "/home/x/hfenv/lib/python3.12/site-packages/triton/backends/nvidia/lib/libdevice.10.bc"]], "debug": true, "backend_name": "cuda", "sanitize_overflow": false, "arch": "sm90", "triton_version": "3.3.1", "shared": 0, "tmem_size": 0, "global_scratch_size": 0, "global_scratch_align": 1, "name": "triton_poi_fused_add_mul_sub_5"}