@@ -347,21 +347,21 @@ void test_builtin_elementwise_max(float f1, float f2, double d1, double d2,
347
347
// CHECK-LABEL: define void @test_builtin_elementwise_max(
348
348
// CHECK: [[F1:%.+]] = load float, ptr %f1.addr, align 4
349
349
// CHECK-NEXT: [[F2:%.+]] = load float, ptr %f2.addr, align 4
350
- // CHECK-NEXT: call float @llvm.maxnum.f32(float [[F1]], float [[F2]])
350
+ // CHECK-NEXT: call nsz float @llvm.maxnum.f32(float [[F1]], float [[F2]])
351
351
f1 = __builtin_elementwise_max (f1 , f2 );
352
352
353
353
// CHECK: [[D1:%.+]] = load double, ptr %d1.addr, align 8
354
354
// CHECK-NEXT: [[D2:%.+]] = load double, ptr %d2.addr, align 8
355
- // CHECK-NEXT: call double @llvm.maxnum.f64(double [[D1]], double [[D2]])
355
+ // CHECK-NEXT: call nsz double @llvm.maxnum.f64(double [[D1]], double [[D2]])
356
356
d1 = __builtin_elementwise_max (d1 , d2 );
357
357
358
358
// CHECK: [[D2:%.+]] = load double, ptr %d2.addr, align 8
359
- // CHECK-NEXT: call double @llvm.maxnum.f64(double 2.000000e+01, double [[D2]])
359
+ // CHECK-NEXT: call nsz double @llvm.maxnum.f64(double 2.000000e+01, double [[D2]])
360
360
d1 = __builtin_elementwise_max (20.0 , d2 );
361
361
362
362
// CHECK: [[VF1:%.+]] = load <4 x float>, ptr %vf1.addr, align 16
363
363
// CHECK-NEXT: [[VF2:%.+]] = load <4 x float>, ptr %vf2.addr, align 16
364
- // CHECK-NEXT: call <4 x float> @llvm.maxnum.v4f32(<4 x float> [[VF1]], <4 x float> [[VF2]])
364
+ // CHECK-NEXT: call nsz <4 x float> @llvm.maxnum.v4f32(<4 x float> [[VF1]], <4 x float> [[VF2]])
365
365
vf1 = __builtin_elementwise_max (vf1 , vf2 );
366
366
367
367
// CHECK: [[I1:%.+]] = load i64, ptr %i1.addr, align 8
@@ -404,13 +404,13 @@ void test_builtin_elementwise_max(float f1, float f2, double d1, double d2,
404
404
405
405
// CHECK: [[CVF1:%.+]] = load <4 x float>, ptr %cvf1, align 16
406
406
// CHECK-NEXT: [[VF2:%.+]] = load <4 x float>, ptr %vf2.addr, align 16
407
- // CHECK-NEXT: call <4 x float> @llvm.maxnum.v4f32(<4 x float> [[CVF1]], <4 x float> [[VF2]])
407
+ // CHECK-NEXT: call nsz <4 x float> @llvm.maxnum.v4f32(<4 x float> [[CVF1]], <4 x float> [[VF2]])
408
408
const float4 cvf1 = vf1 ;
409
409
vf1 = __builtin_elementwise_max (cvf1 , vf2 );
410
410
411
411
// CHECK: [[VF2:%.+]] = load <4 x float>, ptr %vf2.addr, align 16
412
412
// CHECK-NEXT: [[CVF1:%.+]] = load <4 x float>, ptr %cvf1, align 16
413
- // CHECK-NEXT: call <4 x float> @llvm.maxnum.v4f32(<4 x float> [[VF2]], <4 x float> [[CVF1]])
413
+ // CHECK-NEXT: call nsz <4 x float> @llvm.maxnum.v4f32(<4 x float> [[VF2]], <4 x float> [[CVF1]])
414
414
vf1 = __builtin_elementwise_max (vf2 , cvf1 );
415
415
416
416
// CHECK: [[IAS1:%.+]] = load i32, ptr addrspace(1) @int_as_one, align 4
@@ -431,21 +431,21 @@ void test_builtin_elementwise_min(float f1, float f2, double d1, double d2,
431
431
// CHECK-LABEL: define void @test_builtin_elementwise_min(
432
432
// CHECK: [[F1:%.+]] = load float, ptr %f1.addr, align 4
433
433
// CHECK-NEXT: [[F2:%.+]] = load float, ptr %f2.addr, align 4
434
- // CHECK-NEXT: call float @llvm.minnum.f32(float [[F1]], float [[F2]])
434
+ // CHECK-NEXT: call nsz float @llvm.minnum.f32(float [[F1]], float [[F2]])
435
435
f1 = __builtin_elementwise_min (f1 , f2 );
436
436
437
437
// CHECK: [[D1:%.+]] = load double, ptr %d1.addr, align 8
438
438
// CHECK-NEXT: [[D2:%.+]] = load double, ptr %d2.addr, align 8
439
- // CHECK-NEXT: call double @llvm.minnum.f64(double [[D1]], double [[D2]])
439
+ // CHECK-NEXT: call nsz double @llvm.minnum.f64(double [[D1]], double [[D2]])
440
440
d1 = __builtin_elementwise_min (d1 , d2 );
441
441
442
442
// CHECK: [[D1:%.+]] = load double, ptr %d1.addr, align 8
443
- // CHECK-NEXT: call double @llvm.minnum.f64(double [[D1]], double 2.000000e+00)
443
+ // CHECK-NEXT: call nsz double @llvm.minnum.f64(double [[D1]], double 2.000000e+00)
444
444
d1 = __builtin_elementwise_min (d1 , 2.0 );
445
445
446
446
// CHECK: [[VF1:%.+]] = load <4 x float>, ptr %vf1.addr, align 16
447
447
// CHECK-NEXT: [[VF2:%.+]] = load <4 x float>, ptr %vf2.addr, align 16
448
- // CHECK-NEXT: call <4 x float> @llvm.minnum.v4f32(<4 x float> [[VF1]], <4 x float> [[VF2]])
448
+ // CHECK-NEXT: call nsz <4 x float> @llvm.minnum.v4f32(<4 x float> [[VF1]], <4 x float> [[VF2]])
449
449
vf1 = __builtin_elementwise_min (vf1 , vf2 );
450
450
451
451
// CHECK: [[I1:%.+]] = load i64, ptr %i1.addr, align 8
@@ -495,13 +495,13 @@ void test_builtin_elementwise_min(float f1, float f2, double d1, double d2,
495
495
496
496
// CHECK: [[CVF1:%.+]] = load <4 x float>, ptr %cvf1, align 16
497
497
// CHECK-NEXT: [[VF2:%.+]] = load <4 x float>, ptr %vf2.addr, align 16
498
- // CHECK-NEXT: call <4 x float> @llvm.minnum.v4f32(<4 x float> [[CVF1]], <4 x float> [[VF2]])
498
+ // CHECK-NEXT: call nsz <4 x float> @llvm.minnum.v4f32(<4 x float> [[CVF1]], <4 x float> [[VF2]])
499
499
const float4 cvf1 = vf1 ;
500
500
vf1 = __builtin_elementwise_min (cvf1 , vf2 );
501
501
502
502
// CHECK: [[VF2:%.+]] = load <4 x float>, ptr %vf2.addr, align 16
503
503
// CHECK-NEXT: [[CVF1:%.+]] = load <4 x float>, ptr %cvf1, align 16
504
- // CHECK-NEXT: call <4 x float> @llvm.minnum.v4f32(<4 x float> [[VF2]], <4 x float> [[CVF1]])
504
+ // CHECK-NEXT: call nsz <4 x float> @llvm.minnum.v4f32(<4 x float> [[VF2]], <4 x float> [[CVF1]])
505
505
vf1 = __builtin_elementwise_min (vf2 , cvf1 );
506
506
507
507
// CHECK: [[IAS1:%.+]] = load i32, ptr addrspace(1) @int_as_one, align 4
0 commit comments