@@ -1052,31 +1052,15 @@ define <4 x float> @PR34724_add_v4f32_0u23(<4 x float> %0, <4 x float> %1) {
1052
1052
}
1053
1053
1054
1054
define <4 x float > @PR34724_add_v4f32_01u3 (<4 x float > %0 , <4 x float > %1 ) {
1055
- ; SSE-SLOW-LABEL: PR34724_add_v4f32_01u3:
1056
- ; SSE-SLOW: # %bb.0:
1057
- ; SSE-SLOW-NEXT: haddps %xmm0, %xmm0
1058
- ; SSE-SLOW-NEXT: movsldup {{.*#+}} xmm2 = xmm1[0,0,2,2]
1059
- ; SSE-SLOW-NEXT: addps %xmm1, %xmm2
1060
- ; SSE-SLOW-NEXT: shufps {{.*#+}} xmm0 = xmm0[0,1],xmm2[2,3]
1061
- ; SSE-SLOW-NEXT: retq
1062
- ;
1063
- ; SSE-FAST-LABEL: PR34724_add_v4f32_01u3:
1064
- ; SSE-FAST: # %bb.0:
1065
- ; SSE-FAST-NEXT: haddps %xmm1, %xmm0
1066
- ; SSE-FAST-NEXT: retq
1067
- ;
1068
- ; AVX-SLOW-LABEL: PR34724_add_v4f32_01u3:
1069
- ; AVX-SLOW: # %bb.0:
1070
- ; AVX-SLOW-NEXT: vhaddps %xmm0, %xmm0, %xmm0
1071
- ; AVX-SLOW-NEXT: vmovsldup {{.*#+}} xmm2 = xmm1[0,0,2,2]
1072
- ; AVX-SLOW-NEXT: vaddps %xmm1, %xmm2, %xmm1
1073
- ; AVX-SLOW-NEXT: vblendps {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,3]
1074
- ; AVX-SLOW-NEXT: retq
1055
+ ; SSE-LABEL: PR34724_add_v4f32_01u3:
1056
+ ; SSE: # %bb.0:
1057
+ ; SSE-NEXT: haddps %xmm1, %xmm0
1058
+ ; SSE-NEXT: retq
1075
1059
;
1076
- ; AVX-FAST- LABEL: PR34724_add_v4f32_01u3:
1077
- ; AVX-FAST : # %bb.0:
1078
- ; AVX-FAST- NEXT: vhaddps %xmm1, %xmm0, %xmm0
1079
- ; AVX-FAST- NEXT: retq
1060
+ ; AVX-LABEL: PR34724_add_v4f32_01u3:
1061
+ ; AVX: # %bb.0:
1062
+ ; AVX-NEXT: vhaddps %xmm1, %xmm0, %xmm0
1063
+ ; AVX-NEXT: retq
1080
1064
%3 = shufflevector <4 x float > %0 , <4 x float > undef , <2 x i32 > <i32 0 , i32 2 >
1081
1065
%4 = shufflevector <4 x float > %0 , <4 x float > undef , <2 x i32 > <i32 1 , i32 3 >
1082
1066
%5 = fadd <2 x float > %3 , %4
@@ -1088,31 +1072,15 @@ define <4 x float> @PR34724_add_v4f32_01u3(<4 x float> %0, <4 x float> %1) {
1088
1072
}
1089
1073
1090
1074
define <4 x float > @PR34724_add_v4f32_012u (<4 x float > %0 , <4 x float > %1 ) {
1091
- ; SSE-SLOW-LABEL: PR34724_add_v4f32_012u:
1092
- ; SSE-SLOW: # %bb.0:
1093
- ; SSE-SLOW-NEXT: haddps %xmm0, %xmm0
1094
- ; SSE-SLOW-NEXT: movshdup {{.*#+}} xmm2 = xmm1[1,1,3,3]
1095
- ; SSE-SLOW-NEXT: addps %xmm1, %xmm2
1096
- ; SSE-SLOW-NEXT: movlhps {{.*#+}} xmm0 = xmm0[0],xmm2[0]
1097
- ; SSE-SLOW-NEXT: retq
1098
- ;
1099
- ; SSE-FAST-LABEL: PR34724_add_v4f32_012u:
1100
- ; SSE-FAST: # %bb.0:
1101
- ; SSE-FAST-NEXT: haddps %xmm1, %xmm0
1102
- ; SSE-FAST-NEXT: retq
1103
- ;
1104
- ; AVX-SLOW-LABEL: PR34724_add_v4f32_012u:
1105
- ; AVX-SLOW: # %bb.0:
1106
- ; AVX-SLOW-NEXT: vhaddps %xmm0, %xmm0, %xmm0
1107
- ; AVX-SLOW-NEXT: vmovshdup {{.*#+}} xmm2 = xmm1[1,1,3,3]
1108
- ; AVX-SLOW-NEXT: vaddps %xmm1, %xmm2, %xmm1
1109
- ; AVX-SLOW-NEXT: vmovlhps {{.*#+}} xmm0 = xmm0[0],xmm1[0]
1110
- ; AVX-SLOW-NEXT: retq
1075
+ ; SSE-LABEL: PR34724_add_v4f32_012u:
1076
+ ; SSE: # %bb.0:
1077
+ ; SSE-NEXT: haddps %xmm1, %xmm0
1078
+ ; SSE-NEXT: retq
1111
1079
;
1112
- ; AVX-FAST- LABEL: PR34724_add_v4f32_012u:
1113
- ; AVX-FAST : # %bb.0:
1114
- ; AVX-FAST- NEXT: vhaddps %xmm1, %xmm0, %xmm0
1115
- ; AVX-FAST- NEXT: retq
1080
+ ; AVX-LABEL: PR34724_add_v4f32_012u:
1081
+ ; AVX: # %bb.0:
1082
+ ; AVX-NEXT: vhaddps %xmm1, %xmm0, %xmm0
1083
+ ; AVX-NEXT: retq
1116
1084
%3 = shufflevector <4 x float > %0 , <4 x float > undef , <2 x i32 > <i32 0 , i32 2 >
1117
1085
%4 = shufflevector <4 x float > %0 , <4 x float > undef , <2 x i32 > <i32 1 , i32 3 >
1118
1086
%5 = fadd <2 x float > %3 , %4
0 commit comments