Skip to content

Commit a9adcef

Browse files
authored
[flang] fix ppc test broken after #74709 (#74826)
1 parent 32ec462 commit a9adcef

File tree

1 file changed

+48
-48
lines changed

1 file changed

+48
-48
lines changed

flang/test/Lower/PowerPC/ppc-vec-store.f90

Lines changed: 48 additions & 48 deletions
Original file line numberDiff line numberDiff line change
@@ -89,10 +89,10 @@ subroutine vec_st_vi4i4via4(arg1, arg2, arg3, i)
8989

9090
! LLVMIR: %[[i:.*]] = load i32, ptr %3, align 4
9191
! LLVMIR: %[[iext:.*]] = sext i32 %[[i]] to i64
92-
! LLVMIR: %[[iextsub:.*]] = sub i64 %[[iext]], 1
93-
! LLVMIR: %[[iextmul:.*]] = mul i64 %[[iextsub]], 1
94-
! LLVMIR: %[[iextmul2:.*]] = mul i64 %[[iextmul]], 1
95-
! LLVMIR: %[[iextadd:.*]] = add i64 %[[iextmul2]], 0
92+
! LLVMIR: %[[iextsub:.*]] = sub nsw i64 %[[iext]], 1
93+
! LLVMIR: %[[iextmul:.*]] = mul nsw i64 %[[iextsub]], 1
94+
! LLVMIR: %[[iextmul2:.*]] = mul nsw i64 %[[iextmul]], 1
95+
! LLVMIR: %[[iextadd:.*]] = add nsw i64 %[[iextmul2]], 0
9696
! LLVMIR: %[[gep1:.*]] = getelementptr <4 x i32>, ptr %2, i64 %[[iextadd]]
9797
! LLVMIR: %[[arg1:.*]] = load <4 x i32>, ptr %0, align 16
9898
! LLVMIR: %[[arg2:.*]] = load i32, ptr %1, align 4
@@ -206,10 +206,10 @@ subroutine vec_ste_vi4i4ia4(arg1, arg2, arg3, i)
206206

207207
! LLVMIR: %[[i:.*]] = load i32, ptr %3, align 4
208208
! LLVMIR: %[[iext:.*]] = sext i32 %[[i]] to i64
209-
! LLVMIR: %[[isub:.*]] = sub i64 %[[iext]], 1
210-
! LLVMIR: %[[imul1:.*]] = mul i64 %[[isub]], 1
211-
! LLVMIR: %[[imul2:.*]] = mul i64 %[[imul1]], 1
212-
! LLVMIR: %[[iadd:.*]] = add i64 %[[imul2]], 0
209+
! LLVMIR: %[[isub:.*]] = sub nsw i64 %[[iext]], 1
210+
! LLVMIR: %[[imul1:.*]] = mul nsw i64 %[[isub]], 1
211+
! LLVMIR: %[[imul2:.*]] = mul nsw i64 %[[imul1]], 1
212+
! LLVMIR: %[[iadd:.*]] = add nsw i64 %[[imul2]], 0
213213
! LLVMIR: %[[gep1:.*]] = getelementptr i32, ptr %2, i64 %[[iadd]]
214214
! LLVMIR: %[[arg1:.*]] = load <4 x i32>, ptr %0, align 16
215215
! LLVMIR: %[[arg2:.*]] = load i32, ptr %1, align 4
@@ -244,10 +244,10 @@ subroutine vec_stxv_test_vi4i8ia4(arg1, arg2, arg3, i)
244244

245245
! LLVMIR: %[[i:.*]] = load i32, ptr %3, align 4
246246
! LLVMIR: %[[iext:.*]] = sext i32 %[[i]] to i64
247-
! LLVMIR: %[[isub:.*]] = sub i64 %[[iext]], 1
248-
! LLVMIR: %[[imul1:.*]] = mul i64 %[[isub]], 1
249-
! LLVMIR: %[[imul2:.*]] = mul i64 %[[imul1]], 1
250-
! LLVMIR: %[[iadd:.*]] = add i64 %[[imul2]], 0
247+
! LLVMIR: %[[isub:.*]] = sub nsw i64 %[[iext]], 1
248+
! LLVMIR: %[[imul1:.*]] = mul nsw i64 %[[isub]], 1
249+
! LLVMIR: %[[imul2:.*]] = mul nsw i64 %[[imul1]], 1
250+
! LLVMIR: %[[iadd:.*]] = add nsw i64 %[[imul2]], 0
251251
! LLVMIR: %[[gep1:.*]] = getelementptr i32, ptr %2, i64 %[[iadd]]
252252
! LLVMIR: %[[arg1:.*]] = load <4 x i32>, ptr %0, align 16
253253
! LLVMIR: %[[arg2:.*]] = load i64, ptr %1, align 8
@@ -278,10 +278,10 @@ subroutine vec_stxv_test_vi4i4vai4(arg1, arg2, arg3, i)
278278

279279
! LLVMIR: %[[i:.*]] = load i32, ptr %3, align 4
280280
! LLVMIR: %[[iext:.*]] = sext i32 %[[i]] to i64
281-
! LLVMIR: %[[isub:.*]] = sub i64 %[[iext]], 1
282-
! LLVMIR: %[[imul1:.*]] = mul i64 %[[isub]], 1
283-
! LLVMIR: %[[imul2:.*]] = mul i64 %[[imul1]], 1
284-
! LLVMIR: %[[iadd:.*]] = add i64 %[[imul2]], 0
281+
! LLVMIR: %[[isub:.*]] = sub nsw i64 %[[iext]], 1
282+
! LLVMIR: %[[imul1:.*]] = mul nsw i64 %[[isub]], 1
283+
! LLVMIR: %[[imul2:.*]] = mul nsw i64 %[[imul1]], 1
284+
! LLVMIR: %[[iadd:.*]] = add nsw i64 %[[imul2]], 0
285285
! LLVMIR: %[[gep1:.*]] = getelementptr <4 x i32>, ptr %2, i64 %[[iadd]]
286286
! LLVMIR: %[[arg1:.*]] = load <4 x i32>, ptr %0, align 16
287287
! LLVMIR: %[[arg2:.*]] = load i32, ptr %1, align 4
@@ -317,10 +317,10 @@ subroutine vec_xst_test_vi4i8ia4(arg1, arg2, arg3, i)
317317

318318
! LLVMIR: %[[i:.*]] = load i32, ptr %3, align 4
319319
! LLVMIR: %[[iext:.*]] = sext i32 %[[i]] to i64
320-
! LLVMIR: %[[isub:.*]] = sub i64 %[[iext]], 1
321-
! LLVMIR: %[[imul1:.*]] = mul i64 %[[isub]], 1
322-
! LLVMIR: %[[imul2:.*]] = mul i64 %[[imul1]], 1
323-
! LLVMIR: %[[iadd:.*]] = add i64 %[[imul2]], 0
320+
! LLVMIR: %[[isub:.*]] = sub nsw i64 %[[iext]], 1
321+
! LLVMIR: %[[imul1:.*]] = mul nsw i64 %[[isub]], 1
322+
! LLVMIR: %[[imul2:.*]] = mul nsw i64 %[[imul1]], 1
323+
! LLVMIR: %[[iadd:.*]] = add nsw i64 %[[imul2]], 0
324324
! LLVMIR: %[[gep1:.*]] = getelementptr i32, ptr %2, i64 %[[iadd]]
325325
! LLVMIR: %[[arg1:.*]] = load <4 x i32>, ptr %0, align 16
326326
! LLVMIR: %[[arg2:.*]] = load i64, ptr %1, align 8
@@ -351,10 +351,10 @@ subroutine vec_xst_test_vi4i4vai4(arg1, arg2, arg3, i)
351351

352352
! LLVMIR: %[[i:.*]] = load i32, ptr %3, align 4
353353
! LLVMIR: %[[iext:.*]] = sext i32 %[[i]] to i64
354-
! LLVMIR: %[[isub:.*]] = sub i64 %[[iext]], 1
355-
! LLVMIR: %[[imul1:.*]] = mul i64 %[[isub]], 1
356-
! LLVMIR: %[[imul2:.*]] = mul i64 %[[imul1]], 1
357-
! LLVMIR: %[[iadd:.*]] = add i64 %[[imul2]], 0
354+
! LLVMIR: %[[isub:.*]] = sub nsw i64 %[[iext]], 1
355+
! LLVMIR: %[[imul1:.*]] = mul nsw i64 %[[isub]], 1
356+
! LLVMIR: %[[imul2:.*]] = mul nsw i64 %[[imul1]], 1
357+
! LLVMIR: %[[iadd:.*]] = add nsw i64 %[[imul2]], 0
358358
! LLVMIR: %[[gep1:.*]] = getelementptr <4 x i32>, ptr %2, i64 %[[iadd]]
359359
! LLVMIR: %[[arg1:.*]] = load <4 x i32>, ptr %0, align 16
360360
! LLVMIR: %[[arg2:.*]] = load i32, ptr %1, align 4
@@ -390,10 +390,10 @@ subroutine vec_xst_be_test_vi4i8ia4(arg1, arg2, arg3, i)
390390

391391
! LLVMIR: %[[i:.*]] = load i32, ptr %3, align 4
392392
! LLVMIR: %[[iext:.*]] = sext i32 %[[i]] to i64
393-
! LLVMIR: %[[isub:.*]] = sub i64 %[[iext]], 1
394-
! LLVMIR: %[[imul1:.*]] = mul i64 %[[isub]], 1
395-
! LLVMIR: %[[imul2:.*]] = mul i64 %[[imul1]], 1
396-
! LLVMIR: %[[iadd:.*]] = add i64 %[[imul2]], 0
393+
! LLVMIR: %[[isub:.*]] = sub nsw i64 %[[iext]], 1
394+
! LLVMIR: %[[imul1:.*]] = mul nsw i64 %[[isub]], 1
395+
! LLVMIR: %[[imul2:.*]] = mul nsw i64 %[[imul1]], 1
396+
! LLVMIR: %[[iadd:.*]] = add nsw i64 %[[imul2]], 0
397397
! LLVMIR: %[[gep1:.*]] = getelementptr i32, ptr %2, i64 %[[iadd]]
398398
! LLVMIR: %[[arg1:.*]] = load <4 x i32>, ptr %0, align 16
399399
! LLVMIR: %[[arg2:.*]] = load i64, ptr %1, align 8
@@ -426,10 +426,10 @@ subroutine vec_xst_be_test_vi4i4vai4(arg1, arg2, arg3, i)
426426

427427
! LLVMIR: %[[i:.*]] = load i32, ptr %3, align 4
428428
! LLVMIR: %[[iext:.*]] = sext i32 %[[i]] to i64
429-
! LLVMIR: %[[isub:.*]] = sub i64 %[[iext]], 1
430-
! LLVMIR: %[[imul1:.*]] = mul i64 %[[isub]], 1
431-
! LLVMIR: %[[imul2:.*]] = mul i64 %[[imul1]], 1
432-
! LLVMIR: %[[iadd:.*]] = add i64 %[[imul2]], 0
429+
! LLVMIR: %[[isub:.*]] = sub nsw i64 %[[iext]], 1
430+
! LLVMIR: %[[imul1:.*]] = mul nsw i64 %[[isub]], 1
431+
! LLVMIR: %[[imul2:.*]] = mul nsw i64 %[[imul1]], 1
432+
! LLVMIR: %[[iadd:.*]] = add nsw i64 %[[imul2]], 0
433433
! LLVMIR: %[[gep1:.*]] = getelementptr <4 x i32>, ptr %2, i64 %[[iadd]]
434434
! LLVMIR: %[[arg1:.*]] = load <4 x i32>, ptr %0, align 16
435435
! LLVMIR: %[[arg2:.*]] = load i32, ptr %1, align 4
@@ -467,10 +467,10 @@ subroutine vec_xstd2_test_vi4i8ia4(arg1, arg2, arg3, i)
467467

468468
! LLVMIR: %[[i:.*]] = load i32, ptr %3, align 4
469469
! LLVMIR: %[[iext:.*]] = sext i32 %[[i]] to i64
470-
! LLVMIR: %[[isub:.*]] = sub i64 %[[iext]], 1
471-
! LLVMIR: %[[imul1:.*]] = mul i64 %[[isub]], 1
472-
! LLVMIR: %[[imul2:.*]] = mul i64 %[[imul1]], 1
473-
! LLVMIR: %[[iadd:.*]] = add i64 %[[imul2]], 0
470+
! LLVMIR: %[[isub:.*]] = sub nsw i64 %[[iext]], 1
471+
! LLVMIR: %[[imul1:.*]] = mul nsw i64 %[[isub]], 1
472+
! LLVMIR: %[[imul2:.*]] = mul nsw i64 %[[imul1]], 1
473+
! LLVMIR: %[[iadd:.*]] = add nsw i64 %[[imul2]], 0
474474
! LLVMIR: %[[gep1:.*]] = getelementptr i32, ptr %2, i64 %[[iadd]]
475475
! LLVMIR: %[[arg1:.*]] = load <4 x i32>, ptr %0, align 16
476476
! LLVMIR: %[[arg2:.*]] = load i64, ptr %1, align 8
@@ -503,10 +503,10 @@ subroutine vec_xstd2_test_vi4i4vai4(arg1, arg2, arg3, i)
503503

504504
! LLVMIR: %[[i:.*]] = load i32, ptr %3, align 4
505505
! LLVMIR: %[[iext:.*]] = sext i32 %[[i]] to i64
506-
! LLVMIR: %[[isub:.*]] = sub i64 %[[iext]], 1
507-
! LLVMIR: %[[imul1:.*]] = mul i64 %[[isub]], 1
508-
! LLVMIR: %[[imul2:.*]] = mul i64 %[[imul1]], 1
509-
! LLVMIR: %[[iadd:.*]] = add i64 %[[imul2]], 0
506+
! LLVMIR: %[[isub:.*]] = sub nsw i64 %[[iext]], 1
507+
! LLVMIR: %[[imul1:.*]] = mul nsw i64 %[[isub]], 1
508+
! LLVMIR: %[[imul2:.*]] = mul nsw i64 %[[imul1]], 1
509+
! LLVMIR: %[[iadd:.*]] = add nsw i64 %[[imul2]], 0
510510
! LLVMIR: %[[gep1:.*]] = getelementptr <4 x i32>, ptr %2, i64 %[[iadd]]
511511
! LLVMIR: %[[arg1:.*]] = load <4 x i32>, ptr %0, align 16
512512
! LLVMIR: %[[arg2:.*]] = load i32, ptr %1, align 4
@@ -543,10 +543,10 @@ subroutine vec_xstw4_test_vi4i8ia4(arg1, arg2, arg3, i)
543543

544544
! LLVMIR: %[[i:.*]] = load i32, ptr %3, align 4
545545
! LLVMIR: %[[iext:.*]] = sext i32 %[[i]] to i64
546-
! LLVMIR: %[[isub:.*]] = sub i64 %[[iext]], 1
547-
! LLVMIR: %[[imul1:.*]] = mul i64 %[[isub]], 1
548-
! LLVMIR: %[[imul2:.*]] = mul i64 %[[imul1]], 1
549-
! LLVMIR: %[[iadd:.*]] = add i64 %[[imul2]], 0
546+
! LLVMIR: %[[isub:.*]] = sub nsw i64 %[[iext]], 1
547+
! LLVMIR: %[[imul1:.*]] = mul nsw i64 %[[isub]], 1
548+
! LLVMIR: %[[imul2:.*]] = mul nsw i64 %[[imul1]], 1
549+
! LLVMIR: %[[iadd:.*]] = add nsw i64 %[[imul2]], 0
550550
! LLVMIR: %[[gep1:.*]] = getelementptr i32, ptr %2, i64 %[[iadd]]
551551
! LLVMIR: %[[arg1:.*]] = load <4 x i32>, ptr %0, align 16
552552
! LLVMIR: %[[arg2:.*]] = load i64, ptr %1, align 8
@@ -578,10 +578,10 @@ subroutine vec_xstw4_test_vi4i4vai4(arg1, arg2, arg3, i)
578578

579579
! LLVMIR: %[[i:.*]] = load i32, ptr %3, align 4
580580
! LLVMIR: %[[iext:.*]] = sext i32 %[[i]] to i64
581-
! LLVMIR: %[[isub:.*]] = sub i64 %[[iext]], 1
582-
! LLVMIR: %[[imul1:.*]] = mul i64 %[[isub]], 1
583-
! LLVMIR: %[[imul2:.*]] = mul i64 %[[imul1]], 1
584-
! LLVMIR: %[[iadd:.*]] = add i64 %[[imul2]], 0
581+
! LLVMIR: %[[isub:.*]] = sub nsw i64 %[[iext]], 1
582+
! LLVMIR: %[[imul1:.*]] = mul nsw i64 %[[isub]], 1
583+
! LLVMIR: %[[imul2:.*]] = mul nsw i64 %[[imul1]], 1
584+
! LLVMIR: %[[iadd:.*]] = add nsw i64 %[[imul2]], 0
585585
! LLVMIR: %[[gep1:.*]] = getelementptr <4 x i32>, ptr %2, i64 %[[iadd]]
586586
! LLVMIR: %[[arg1:.*]] = load <4 x i32>, ptr %0, align 16
587587
! LLVMIR: %[[arg2:.*]] = load i32, ptr %1, align 4

0 commit comments

Comments
 (0)