3
3
;
4
4
; Forked from llvm/test/CodeGen/X86/f16c-intrinsics.ll
5
5
;
6
- ; Handled by visitInstruction :
6
+ ; Handled by handleSSEVectorConvertIntrinsicByProp :
7
7
; - llvm.x86.vcvtps2ph.128/256
8
8
9
9
target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128-n8:16:32:64-S128"
@@ -14,15 +14,11 @@ define <8 x i16> @test_x86_vcvtps2ph_128(<4 x float> %a0) #0 {
14
14
; CHECK-SAME: <4 x float> [[A0:%.*]]) #[[ATTR0:[0-9]+]] {
15
15
; CHECK-NEXT: [[TMP1:%.*]] = load <4 x i32>, ptr @__msan_param_tls, align 8
16
16
; CHECK-NEXT: call void @llvm.donothing()
17
- ; CHECK-NEXT: [[TMP2:%.*]] = bitcast <4 x i32> [[TMP1]] to i128
18
- ; CHECK-NEXT: [[_MSCMP:%.*]] = icmp ne i128 [[TMP2]], 0
19
- ; CHECK-NEXT: br i1 [[_MSCMP]], label %[[BB3:.*]], label %[[BB4:.*]], !prof [[PROF1:![0-9]+]]
20
- ; CHECK: [[BB3]]:
21
- ; CHECK-NEXT: call void @__msan_warning_noreturn() #[[ATTR4:[0-9]+]]
22
- ; CHECK-NEXT: unreachable
23
- ; CHECK: [[BB4]]:
17
+ ; CHECK-NEXT: [[TMP2:%.*]] = icmp ne <4 x i32> [[TMP1]], zeroinitializer
18
+ ; CHECK-NEXT: [[TMP3:%.*]] = sext <4 x i1> [[TMP2]] to <4 x i16>
19
+ ; CHECK-NEXT: [[TMP11:%.*]] = shufflevector <4 x i16> [[TMP3]], <4 x i16> zeroinitializer, <8 x i32> <i32 0, i32 1, i32 2, i32 3, i32 4, i32 5, i32 6, i32 7>
24
20
; CHECK-NEXT: [[RES:%.*]] = call <8 x i16> @llvm.x86.vcvtps2ph.128(<4 x float> [[A0]], i32 0)
25
- ; CHECK-NEXT: store <8 x i16> zeroinitializer , ptr @__msan_retval_tls, align 8
21
+ ; CHECK-NEXT: store <8 x i16> [[TMP11]] , ptr @__msan_retval_tls, align 8
26
22
; CHECK-NEXT: ret <8 x i16> [[RES]]
27
23
;
28
24
%res = call <8 x i16 > @llvm.x86.vcvtps2ph.128 (<4 x float > %a0 , i32 0 ) ; <<8 x i16>> [#uses=1]
@@ -35,15 +31,10 @@ define <8 x i16> @test_x86_vcvtps2ph_256(<8 x float> %a0) #0 {
35
31
; CHECK-SAME: <8 x float> [[A0:%.*]]) #[[ATTR0]] {
36
32
; CHECK-NEXT: [[TMP1:%.*]] = load <8 x i32>, ptr @__msan_param_tls, align 8
37
33
; CHECK-NEXT: call void @llvm.donothing()
38
- ; CHECK-NEXT: [[TMP2:%.*]] = bitcast <8 x i32> [[TMP1]] to i256
39
- ; CHECK-NEXT: [[_MSCMP:%.*]] = icmp ne i256 [[TMP2]], 0
40
- ; CHECK-NEXT: br i1 [[_MSCMP]], label %[[BB3:.*]], label %[[BB4:.*]], !prof [[PROF1]]
41
- ; CHECK: [[BB3]]:
42
- ; CHECK-NEXT: call void @__msan_warning_noreturn() #[[ATTR4]]
43
- ; CHECK-NEXT: unreachable
44
- ; CHECK: [[BB4]]:
34
+ ; CHECK-NEXT: [[TMP2:%.*]] = icmp ne <8 x i32> [[TMP1]], zeroinitializer
35
+ ; CHECK-NEXT: [[TMP3:%.*]] = sext <8 x i1> [[TMP2]] to <8 x i16>
45
36
; CHECK-NEXT: [[RES:%.*]] = call <8 x i16> @llvm.x86.vcvtps2ph.256(<8 x float> [[A0]], i32 0)
46
- ; CHECK-NEXT: store <8 x i16> zeroinitializer , ptr @__msan_retval_tls, align 8
37
+ ; CHECK-NEXT: store <8 x i16> [[TMP3]] , ptr @__msan_retval_tls, align 8
47
38
; CHECK-NEXT: ret <8 x i16> [[RES]]
48
39
;
49
40
%res = call <8 x i16 > @llvm.x86.vcvtps2ph.256 (<8 x float > %a0 , i32 0 ) ; <<8 x i16>> [#uses=1]
@@ -59,24 +50,19 @@ define void @test_x86_vcvtps2ph_256_m(ptr nocapture %d, <8 x float> %a) nounwind
59
50
; CHECK-NEXT: [[TMP17:%.*]] = load <8 x i32>, ptr inttoptr (i64 add (i64 ptrtoint (ptr @__msan_param_tls to i64), i64 8) to ptr), align 8
60
51
; CHECK-NEXT: [[TMP18:%.*]] = load i64, ptr @__msan_param_tls, align 8
61
52
; CHECK-NEXT: call void @llvm.donothing()
62
- ; CHECK-NEXT: [[TMP4:%.*]] = bitcast <8 x i32> [[TMP17]] to i256
63
- ; CHECK-NEXT: [[_MSCMP:%.*]] = icmp ne i256 [[TMP4]], 0
64
- ; CHECK-NEXT: br i1 [[_MSCMP]], label %[[BB3:.*]], label %[[BB4:.*]], !prof [[PROF1]]
65
- ; CHECK: [[BB3]]:
66
- ; CHECK-NEXT: call void @__msan_warning_noreturn() #[[ATTR4]]
67
- ; CHECK-NEXT: unreachable
68
- ; CHECK: [[BB4]]:
53
+ ; CHECK-NEXT: [[TMP20:%.*]] = icmp ne <8 x i32> [[TMP17]], zeroinitializer
54
+ ; CHECK-NEXT: [[TMP21:%.*]] = sext <8 x i1> [[TMP20]] to <8 x i16>
69
55
; CHECK-NEXT: [[TMP0:%.*]] = tail call <8 x i16> @llvm.x86.vcvtps2ph.256(<8 x float> [[A]], i32 3)
70
56
; CHECK-NEXT: [[_MSCMP1:%.*]] = icmp ne i64 [[TMP18]], 0
71
- ; CHECK-NEXT: br i1 [[_MSCMP1]], label %[[BB6 :.*]], label %[[BB7 :.*]], !prof [[PROF1]]
72
- ; CHECK: [[BB6 ]]:
73
- ; CHECK-NEXT: call void @__msan_warning_noreturn() #[[ATTR4]]
57
+ ; CHECK-NEXT: br i1 [[_MSCMP1]], label %[[BB5 :.*]], label %[[BB6 :.*]], !prof [[PROF1:![0-9]+ ]]
58
+ ; CHECK: [[BB5 ]]:
59
+ ; CHECK-NEXT: call void @__msan_warning_noreturn() #[[ATTR4:[0-9]+ ]]
74
60
; CHECK-NEXT: unreachable
75
- ; CHECK: [[BB7 ]]:
61
+ ; CHECK: [[BB6 ]]:
76
62
; CHECK-NEXT: [[TMP1:%.*]] = ptrtoint ptr [[D]] to i64
77
63
; CHECK-NEXT: [[TMP2:%.*]] = xor i64 [[TMP1]], 87960930222080
78
64
; CHECK-NEXT: [[TMP3:%.*]] = inttoptr i64 [[TMP2]] to ptr
79
- ; CHECK-NEXT: store <8 x i16> zeroinitializer , ptr [[TMP3]], align 16
65
+ ; CHECK-NEXT: store <8 x i16> [[TMP21]] , ptr [[TMP3]], align 16
80
66
; CHECK-NEXT: store <8 x i16> [[TMP0]], ptr [[D]], align 16
81
67
; CHECK-NEXT: ret void
82
68
;
@@ -93,14 +79,11 @@ define void @test_x86_vcvtps2ph_128_m(ptr nocapture %d, <4 x float> %a) nounwind
93
79
; CHECK-NEXT: [[TMP9:%.*]] = load <4 x i32>, ptr inttoptr (i64 add (i64 ptrtoint (ptr @__msan_param_tls to i64), i64 8) to ptr), align 8
94
80
; CHECK-NEXT: [[TMP10:%.*]] = load i64, ptr @__msan_param_tls, align 8
95
81
; CHECK-NEXT: call void @llvm.donothing()
96
- ; CHECK-NEXT: [[TMP5:%.*]] = bitcast <4 x i32> [[TMP9]] to i128
97
- ; CHECK-NEXT: [[_MSCMP:%.*]] = icmp ne i128 [[TMP5]], 0
98
- ; CHECK-NEXT: br i1 [[_MSCMP]], label %[[BB3:.*]], label %[[BB4:.*]], !prof [[PROF1]]
99
- ; CHECK: [[BB3]]:
100
- ; CHECK-NEXT: call void @__msan_warning_noreturn() #[[ATTR4]]
101
- ; CHECK-NEXT: unreachable
102
- ; CHECK: [[BB4]]:
82
+ ; CHECK-NEXT: [[TMP12:%.*]] = icmp ne <4 x i32> [[TMP9]], zeroinitializer
83
+ ; CHECK-NEXT: [[TMP13:%.*]] = sext <4 x i1> [[TMP12]] to <4 x i16>
84
+ ; CHECK-NEXT: [[TMP11:%.*]] = shufflevector <4 x i16> [[TMP13]], <4 x i16> zeroinitializer, <8 x i32> <i32 0, i32 1, i32 2, i32 3, i32 4, i32 5, i32 6, i32 7>
103
85
; CHECK-NEXT: [[TMP0:%.*]] = tail call <8 x i16> @llvm.x86.vcvtps2ph.128(<4 x float> [[A]], i32 3)
86
+ ; CHECK-NEXT: [[_MSPROP:%.*]] = shufflevector <8 x i16> [[TMP11]], <8 x i16> splat (i16 -1), <4 x i32> <i32 0, i32 1, i32 2, i32 3>
104
87
; CHECK-NEXT: [[TMP1:%.*]] = shufflevector <8 x i16> [[TMP0]], <8 x i16> poison, <4 x i32> <i32 0, i32 1, i32 2, i32 3>
105
88
; CHECK-NEXT: [[_MSCMP1:%.*]] = icmp ne i64 [[TMP10]], 0
106
89
; CHECK-NEXT: br i1 [[_MSCMP1]], label %[[BB7:.*]], label %[[BB8:.*]], !prof [[PROF1]]
@@ -111,7 +94,7 @@ define void @test_x86_vcvtps2ph_128_m(ptr nocapture %d, <4 x float> %a) nounwind
111
94
; CHECK-NEXT: [[TMP2:%.*]] = ptrtoint ptr [[D]] to i64
112
95
; CHECK-NEXT: [[TMP3:%.*]] = xor i64 [[TMP2]], 87960930222080
113
96
; CHECK-NEXT: [[TMP4:%.*]] = inttoptr i64 [[TMP3]] to ptr
114
- ; CHECK-NEXT: store <4 x i16> zeroinitializer , ptr [[TMP4]], align 8
97
+ ; CHECK-NEXT: store <4 x i16> [[_MSPROP]] , ptr [[TMP4]], align 8
115
98
; CHECK-NEXT: store <4 x i16> [[TMP1]], ptr [[D]], align 8
116
99
; CHECK-NEXT: ret void
117
100
;
@@ -129,26 +112,24 @@ define void @test_x86_vcvtps2ph_128_m2(ptr nocapture %hf4x16, <4 x float> %f4X86
129
112
; CHECK-NEXT: [[TMP0:%.*]] = load <4 x i32>, ptr inttoptr (i64 add (i64 ptrtoint (ptr @__msan_param_tls to i64), i64 8) to ptr), align 8
130
113
; CHECK-NEXT: [[TMP1:%.*]] = load i64, ptr @__msan_param_tls, align 8
131
114
; CHECK-NEXT: call void @llvm.donothing()
132
- ; CHECK-NEXT: [[TMP2:%.*]] = bitcast <4 x i32> [[TMP0]] to i128
133
- ; CHECK-NEXT: [[_MSCMP:%.*]] = icmp ne i128 [[TMP2]], 0
134
- ; CHECK-NEXT: br i1 [[_MSCMP]], label %[[BB3:.*]], label %[[BB4:.*]], !prof [[PROF1]]
135
- ; CHECK: [[BB3]]:
136
- ; CHECK-NEXT: call void @__msan_warning_noreturn() #[[ATTR4]]
137
- ; CHECK-NEXT: unreachable
138
- ; CHECK: [[BB4]]:
115
+ ; CHECK-NEXT: [[TMP2:%.*]] = icmp ne <4 x i32> [[TMP0]], zeroinitializer
116
+ ; CHECK-NEXT: [[TMP3:%.*]] = sext <4 x i1> [[TMP2]] to <4 x i16>
117
+ ; CHECK-NEXT: [[TMP14:%.*]] = shufflevector <4 x i16> [[TMP3]], <4 x i16> zeroinitializer, <8 x i32> <i32 0, i32 1, i32 2, i32 3, i32 4, i32 5, i32 6, i32 7>
139
118
; CHECK-NEXT: [[TMP11:%.*]] = tail call <8 x i16> @llvm.x86.vcvtps2ph.128(<4 x float> [[F4X86]], i32 3)
119
+ ; CHECK-NEXT: [[TMP13:%.*]] = bitcast <8 x i16> [[TMP14]] to <2 x i64>
140
120
; CHECK-NEXT: [[TMP12:%.*]] = bitcast <8 x i16> [[TMP11]] to <2 x double>
121
+ ; CHECK-NEXT: [[_MSPROP:%.*]] = extractelement <2 x i64> [[TMP13]], i32 0
141
122
; CHECK-NEXT: [[VECEXT:%.*]] = extractelement <2 x double> [[TMP12]], i32 0
142
123
; CHECK-NEXT: [[_MSCMP1:%.*]] = icmp ne i64 [[TMP1]], 0
143
- ; CHECK-NEXT: br i1 [[_MSCMP1]], label %[[BB7 :.*]], label %[[BB8 :.*]], !prof [[PROF1]]
144
- ; CHECK: [[BB7 ]]:
124
+ ; CHECK-NEXT: br i1 [[_MSCMP1]], label %[[BB8 :.*]], label %[[BB9 :.*]], !prof [[PROF1]]
125
+ ; CHECK: [[BB8 ]]:
145
126
; CHECK-NEXT: call void @__msan_warning_noreturn() #[[ATTR4]]
146
127
; CHECK-NEXT: unreachable
147
- ; CHECK: [[BB8 ]]:
128
+ ; CHECK: [[BB9 ]]:
148
129
; CHECK-NEXT: [[TMP15:%.*]] = ptrtoint ptr [[HF4X16]] to i64
149
130
; CHECK-NEXT: [[TMP16:%.*]] = xor i64 [[TMP15]], 87960930222080
150
131
; CHECK-NEXT: [[TMP17:%.*]] = inttoptr i64 [[TMP16]] to ptr
151
- ; CHECK-NEXT: store i64 0 , ptr [[TMP17]], align 8
132
+ ; CHECK-NEXT: store i64 [[_MSPROP]] , ptr [[TMP17]], align 8
152
133
; CHECK-NEXT: store double [[VECEXT]], ptr [[HF4X16]], align 8
153
134
; CHECK-NEXT: ret void
154
135
;
@@ -167,27 +148,25 @@ define void @test_x86_vcvtps2ph_128_m3(ptr nocapture %hf4x16, <4 x float> %f4X86
167
148
; CHECK-NEXT: [[TMP0:%.*]] = load <4 x i32>, ptr inttoptr (i64 add (i64 ptrtoint (ptr @__msan_param_tls to i64), i64 8) to ptr), align 8
168
149
; CHECK-NEXT: [[TMP1:%.*]] = load i64, ptr @__msan_param_tls, align 8
169
150
; CHECK-NEXT: call void @llvm.donothing()
170
- ; CHECK-NEXT: [[TMP2:%.*]] = bitcast <4 x i32> [[TMP0]] to i128
171
- ; CHECK-NEXT: [[_MSCMP:%.*]] = icmp ne i128 [[TMP2]], 0
172
- ; CHECK-NEXT: br i1 [[_MSCMP]], label %[[BB3:.*]], label %[[BB4:.*]], !prof [[PROF1]]
173
- ; CHECK: [[BB3]]:
174
- ; CHECK-NEXT: call void @__msan_warning_noreturn() #[[ATTR4]]
175
- ; CHECK-NEXT: unreachable
176
- ; CHECK: [[BB4]]:
151
+ ; CHECK-NEXT: [[TMP2:%.*]] = icmp ne <4 x i32> [[TMP0]], zeroinitializer
152
+ ; CHECK-NEXT: [[TMP3:%.*]] = sext <4 x i1> [[TMP2]] to <4 x i16>
153
+ ; CHECK-NEXT: [[TMP13:%.*]] = shufflevector <4 x i16> [[TMP3]], <4 x i16> zeroinitializer, <8 x i32> <i32 0, i32 1, i32 2, i32 3, i32 4, i32 5, i32 6, i32 7>
177
154
; CHECK-NEXT: [[TMP11:%.*]] = tail call <8 x i16> @llvm.x86.vcvtps2ph.128(<4 x float> [[F4X86]], i32 3)
178
- ; CHECK-NEXT: [[TMP12:%.*]] = bitcast <8 x i16> [[TMP11]] to <2 x i64>
155
+ ; CHECK-NEXT: [[TMP12:%.*]] = bitcast <8 x i16> [[TMP13]] to <2 x i64>
156
+ ; CHECK-NEXT: [[TMP14:%.*]] = bitcast <8 x i16> [[TMP11]] to <2 x i64>
179
157
; CHECK-NEXT: [[VECEXT:%.*]] = extractelement <2 x i64> [[TMP12]], i32 0
158
+ ; CHECK-NEXT: [[VECEXT1:%.*]] = extractelement <2 x i64> [[TMP14]], i32 0
180
159
; CHECK-NEXT: [[_MSCMP1:%.*]] = icmp ne i64 [[TMP1]], 0
181
- ; CHECK-NEXT: br i1 [[_MSCMP1]], label %[[BB7 :.*]], label %[[BB8 :.*]], !prof [[PROF1]]
182
- ; CHECK: [[BB7 ]]:
160
+ ; CHECK-NEXT: br i1 [[_MSCMP1]], label %[[BB8 :.*]], label %[[BB9 :.*]], !prof [[PROF1]]
161
+ ; CHECK: [[BB8 ]]:
183
162
; CHECK-NEXT: call void @__msan_warning_noreturn() #[[ATTR4]]
184
163
; CHECK-NEXT: unreachable
185
- ; CHECK: [[BB8 ]]:
164
+ ; CHECK: [[BB9 ]]:
186
165
; CHECK-NEXT: [[TMP15:%.*]] = ptrtoint ptr [[HF4X16]] to i64
187
166
; CHECK-NEXT: [[TMP16:%.*]] = xor i64 [[TMP15]], 87960930222080
188
167
; CHECK-NEXT: [[TMP17:%.*]] = inttoptr i64 [[TMP16]] to ptr
189
- ; CHECK-NEXT: store i64 0 , ptr [[TMP17]], align 8
190
- ; CHECK-NEXT: store i64 [[VECEXT ]], ptr [[HF4X16]], align 8
168
+ ; CHECK-NEXT: store i64 [[VECEXT]] , ptr [[TMP17]], align 8
169
+ ; CHECK-NEXT: store i64 [[VECEXT1 ]], ptr [[HF4X16]], align 8
191
170
; CHECK-NEXT: ret void
192
171
;
193
172
entry:
@@ -199,6 +178,3 @@ entry:
199
178
}
200
179
201
180
attributes #0 = { sanitize_memory }
202
- ;.
203
- ; CHECK: [[PROF1]] = !{!"branch_weights", i32 1, i32 1048575}
204
- ;.
0 commit comments