@@ -39,29 +39,29 @@ define void @sqrt() {
39
39
call <4 x bfloat> @llvm.sqrt.v4bf16 (<4 x bfloat> undef )
40
40
call <8 x bfloat> @llvm.sqrt.v8bf16 (<8 x bfloat> undef )
41
41
call <16 x bfloat> @llvm.sqrt.v16bf16 (<16 x bfloat> undef )
42
- call <vscale x 2 x bfloat> @llvm.sqrt.nvx2bf16 (<vscale x 2 x bfloat> undef )
43
- call <vscale x 4 x bfloat> @llvm.sqrt.nvx4bf16 (<vscale x 4 x bfloat> undef )
44
- call <vscale x 8 x bfloat> @llvm.sqrt.nvx8bf16 (<vscale x 8 x bfloat> undef )
45
- call <vscale x 16 x bfloat> @llvm.sqrt.nvx16bf16 (<vscale x 16 x bfloat> undef )
42
+ call <vscale x 2 x bfloat> @llvm.sqrt.nxv2bf16 (<vscale x 2 x bfloat> undef )
43
+ call <vscale x 4 x bfloat> @llvm.sqrt.nxv4bf16 (<vscale x 4 x bfloat> undef )
44
+ call <vscale x 8 x bfloat> @llvm.sqrt.nxv8bf16 (<vscale x 8 x bfloat> undef )
45
+ call <vscale x 16 x bfloat> @llvm.sqrt.nxv16bf16 (<vscale x 16 x bfloat> undef )
46
46
call float @llvm.sqrt.f32 (float undef )
47
47
call <2 x float > @llvm.sqrt.v2f32 (<2 x float > undef )
48
48
call <4 x float > @llvm.sqrt.v4f32 (<4 x float > undef )
49
49
call <8 x float > @llvm.sqrt.v8f32 (<8 x float > undef )
50
50
call <16 x float > @llvm.sqrt.v16f32 (<16 x float > undef )
51
- call <vscale x 1 x float > @llvm.sqrt.nvx1f32 (<vscale x 1 x float > undef )
52
- call <vscale x 2 x float > @llvm.sqrt.nvx2f32 (<vscale x 2 x float > undef )
53
- call <vscale x 4 x float > @llvm.sqrt.nvx4f32 (<vscale x 4 x float > undef )
54
- call <vscale x 8 x float > @llvm.sqrt.nvx8f32 (<vscale x 8 x float > undef )
55
- call <vscale x 16 x float > @llvm.sqrt.nvx16f32 (<vscale x 16 x float > undef )
51
+ call <vscale x 1 x float > @llvm.sqrt.nxv1f32 (<vscale x 1 x float > undef )
52
+ call <vscale x 2 x float > @llvm.sqrt.nxv2f32 (<vscale x 2 x float > undef )
53
+ call <vscale x 4 x float > @llvm.sqrt.nxv4f32 (<vscale x 4 x float > undef )
54
+ call <vscale x 8 x float > @llvm.sqrt.nxv8f32 (<vscale x 8 x float > undef )
55
+ call <vscale x 16 x float > @llvm.sqrt.nxv16f32 (<vscale x 16 x float > undef )
56
56
call double @llvm.sqrt.f64 (double undef )
57
57
call <2 x double > @llvm.sqrt.v2f64 (<2 x double > undef )
58
58
call <4 x double > @llvm.sqrt.v4f64 (<4 x double > undef )
59
59
call <8 x double > @llvm.sqrt.v8f64 (<8 x double > undef )
60
60
call <16 x double > @llvm.sqrt.v16f64 (<16 x double > undef )
61
- call <vscale x 1 x double > @llvm.sqrt.nvx1f64 (<vscale x 1 x double > undef )
62
- call <vscale x 2 x double > @llvm.sqrt.nvx2f64 (<vscale x 2 x double > undef )
63
- call <vscale x 4 x double > @llvm.sqrt.nvx4f64 (<vscale x 4 x double > undef )
64
- call <vscale x 8 x double > @llvm.sqrt.nvx8f64 (<vscale x 8 x double > undef )
61
+ call <vscale x 1 x double > @llvm.sqrt.nxv1f64 (<vscale x 1 x double > undef )
62
+ call <vscale x 2 x double > @llvm.sqrt.nxv2f64 (<vscale x 2 x double > undef )
63
+ call <vscale x 4 x double > @llvm.sqrt.nxv4f64 (<vscale x 4 x double > undef )
64
+ call <vscale x 8 x double > @llvm.sqrt.nxv8f64 (<vscale x 8 x double > undef )
65
65
ret void
66
66
}
67
67
@@ -83,10 +83,10 @@ define void @sqrt_f16() {
83
83
call <4 x half > @llvm.sqrt.v4f16 (<4 x half > undef )
84
84
call <8 x half > @llvm.sqrt.v8f16 (<8 x half > undef )
85
85
call <16 x half > @llvm.sqrt.v16f16 (<16 x half > undef )
86
- call <vscale x 2 x half > @llvm.sqrt.nvx2f16 (<vscale x 2 x half > undef )
87
- call <vscale x 4 x half > @llvm.sqrt.nvx4f16 (<vscale x 4 x half > undef )
88
- call <vscale x 8 x half > @llvm.sqrt.nvx8f16 (<vscale x 8 x half > undef )
89
- call <vscale x 16 x half > @llvm.sqrt.nvx16f16 (<vscale x 16 x half > undef )
86
+ call <vscale x 2 x half > @llvm.sqrt.nxv2f16 (<vscale x 2 x half > undef )
87
+ call <vscale x 4 x half > @llvm.sqrt.nxv4f16 (<vscale x 4 x half > undef )
88
+ call <vscale x 8 x half > @llvm.sqrt.nxv8f16 (<vscale x 8 x half > undef )
89
+ call <vscale x 16 x half > @llvm.sqrt.nxv16f16 (<vscale x 16 x half > undef )
90
90
ret void
91
91
}
92
92
@@ -128,30 +128,30 @@ define void @pow() {
128
128
call <4 x bfloat> @llvm.pow.v4bf16 (<4 x bfloat> undef , <4 x bfloat> undef )
129
129
call <8 x bfloat> @llvm.pow.v8bf16 (<8 x bfloat> undef , <8 x bfloat> undef )
130
130
call <16 x bfloat> @llvm.pow.v16bf16 (<16 x bfloat> undef , <16 x bfloat> undef )
131
- call <vscale x 1 x bfloat> @llvm.pow.nvx1bf16 (<vscale x 1 x bfloat> undef , <vscale x 1 x bfloat> undef )
132
- call <vscale x 2 x bfloat> @llvm.pow.nvx2bf16 (<vscale x 2 x bfloat> undef , <vscale x 2 x bfloat> undef )
133
- call <vscale x 4 x bfloat> @llvm.pow.nvx4bf16 (<vscale x 4 x bfloat> undef , <vscale x 4 x bfloat> undef )
134
- call <vscale x 8 x bfloat> @llvm.pow.nvx8bf16 (<vscale x 8 x bfloat> undef , <vscale x 8 x bfloat> undef )
135
- call <vscale x 16 x bfloat> @llvm.pow.nvx16bf16 (<vscale x 16 x bfloat> undef , <vscale x 16 x bfloat> undef )
131
+ call <vscale x 1 x bfloat> @llvm.pow.nxv1bf16 (<vscale x 1 x bfloat> undef , <vscale x 1 x bfloat> undef )
132
+ call <vscale x 2 x bfloat> @llvm.pow.nxv2bf16 (<vscale x 2 x bfloat> undef , <vscale x 2 x bfloat> undef )
133
+ call <vscale x 4 x bfloat> @llvm.pow.nxv4bf16 (<vscale x 4 x bfloat> undef , <vscale x 4 x bfloat> undef )
134
+ call <vscale x 8 x bfloat> @llvm.pow.nxv8bf16 (<vscale x 8 x bfloat> undef , <vscale x 8 x bfloat> undef )
135
+ call <vscale x 16 x bfloat> @llvm.pow.nxv16bf16 (<vscale x 16 x bfloat> undef , <vscale x 16 x bfloat> undef )
136
136
call float @llvm.pow.f32 (float undef , float undef )
137
137
call <2 x float > @llvm.pow.v2f32 (<2 x float > undef , <2 x float > undef )
138
138
call <4 x float > @llvm.pow.v4f32 (<4 x float > undef , <4 x float > undef )
139
139
call <8 x float > @llvm.pow.v8f32 (<8 x float > undef , <8 x float > undef )
140
140
call <16 x float > @llvm.pow.v16f32 (<16 x float > undef , <16 x float > undef )
141
- call <vscale x 1 x float > @llvm.pow.nvx1f32 (<vscale x 1 x float > undef , <vscale x 1 x float > undef )
142
- call <vscale x 2 x float > @llvm.pow.nvx2f32 (<vscale x 2 x float > undef , <vscale x 2 x float > undef )
143
- call <vscale x 4 x float > @llvm.pow.nvx4f32 (<vscale x 4 x float > undef , <vscale x 4 x float > undef )
144
- call <vscale x 8 x float > @llvm.pow.nvx8f32 (<vscale x 8 x float > undef , <vscale x 8 x float > undef )
145
- call <vscale x 16 x float > @llvm.pow.nvx16f32 (<vscale x 16 x float > undef , <vscale x 16 x float > undef )
141
+ call <vscale x 1 x float > @llvm.pow.nxv1f32 (<vscale x 1 x float > undef , <vscale x 1 x float > undef )
142
+ call <vscale x 2 x float > @llvm.pow.nxv2f32 (<vscale x 2 x float > undef , <vscale x 2 x float > undef )
143
+ call <vscale x 4 x float > @llvm.pow.nxv4f32 (<vscale x 4 x float > undef , <vscale x 4 x float > undef )
144
+ call <vscale x 8 x float > @llvm.pow.nxv8f32 (<vscale x 8 x float > undef , <vscale x 8 x float > undef )
145
+ call <vscale x 16 x float > @llvm.pow.nxv16f32 (<vscale x 16 x float > undef , <vscale x 16 x float > undef )
146
146
call double @llvm.pow.f64 (double undef , double undef )
147
147
call <2 x double > @llvm.pow.v2f64 (<2 x double > undef , <2 x double > undef )
148
148
call <4 x double > @llvm.pow.v4f64 (<4 x double > undef , <4 x double > undef )
149
149
call <8 x double > @llvm.pow.v8f64 (<8 x double > undef , <8 x double > undef )
150
150
call <16 x double > @llvm.pow.v16f64 (<16 x double > undef , <16 x double > undef )
151
- call <vscale x 1 x double > @llvm.pow.nvx1f64 (<vscale x 1 x double > undef , <vscale x 1 x double > undef )
152
- call <vscale x 2 x double > @llvm.pow.nvx2f64 (<vscale x 2 x double > undef , <vscale x 2 x double > undef )
153
- call <vscale x 4 x double > @llvm.pow.nvx4f64 (<vscale x 4 x double > undef , <vscale x 4 x double > undef )
154
- call <vscale x 8 x double > @llvm.pow.nvx8f64 (<vscale x 8 x double > undef , <vscale x 8 x double > undef )
151
+ call <vscale x 1 x double > @llvm.pow.nxv1f64 (<vscale x 1 x double > undef , <vscale x 1 x double > undef )
152
+ call <vscale x 2 x double > @llvm.pow.nxv2f64 (<vscale x 2 x double > undef , <vscale x 2 x double > undef )
153
+ call <vscale x 4 x double > @llvm.pow.nxv4f64 (<vscale x 4 x double > undef , <vscale x 4 x double > undef )
154
+ call <vscale x 8 x double > @llvm.pow.nxv8f64 (<vscale x 8 x double > undef , <vscale x 8 x double > undef )
155
155
ret void
156
156
}
157
157
@@ -187,10 +187,10 @@ define void @pow_f16() {
187
187
call <4 x half > @llvm.pow.v4f16 (<4 x half > undef , <4 x half > undef )
188
188
call <8 x half > @llvm.pow.v8f16 (<8 x half > undef , <8 x half > undef )
189
189
call <16 x half > @llvm.pow.v16f16 (<16 x half > undef , <16 x half > undef )
190
- call <vscale x 1 x half > @llvm.pow.nvx1f16 (<vscale x 1 x half > undef , <vscale x 1 x half > undef )
191
- call <vscale x 2 x half > @llvm.pow.nvx2f16 (<vscale x 2 x half > undef , <vscale x 2 x half > undef )
192
- call <vscale x 4 x half > @llvm.pow.nvx4f16 (<vscale x 4 x half > undef , <vscale x 4 x half > undef )
193
- call <vscale x 8 x half > @llvm.pow.nvx8f16 (<vscale x 8 x half > undef , <vscale x 8 x half > undef )
194
- call <vscale x 16 x half > @llvm.pow.nvx16f16 (<vscale x 16 x half > undef , <vscale x 16 x half > undef )
190
+ call <vscale x 1 x half > @llvm.pow.nxv1f16 (<vscale x 1 x half > undef , <vscale x 1 x half > undef )
191
+ call <vscale x 2 x half > @llvm.pow.nxv2f16 (<vscale x 2 x half > undef , <vscale x 2 x half > undef )
192
+ call <vscale x 4 x half > @llvm.pow.nxv4f16 (<vscale x 4 x half > undef , <vscale x 4 x half > undef )
193
+ call <vscale x 8 x half > @llvm.pow.nxv8f16 (<vscale x 8 x half > undef , <vscale x 8 x half > undef )
194
+ call <vscale x 16 x half > @llvm.pow.nxv16f16 (<vscale x 16 x half > undef , <vscale x 16 x half > undef )
195
195
ret void
196
196
}
0 commit comments