|
| 1 | +# NOTE: Assertions have been autogenerated by utils/update_mir_test_checks.py UTC_ARGS: --version 5 |
| 2 | +# RUN: llc -O0 -mtriple=aarch64 -verify-machineinstrs -run-pass=legalizer -global-isel-abort=0 -pass-remarks-missed='gisel.*' -o - %s 2> %t.err | FileCheck %s |
| 3 | +# RUN: FileCheck -check-prefix=ERR %s < %t.err |
| 4 | + |
| 5 | +# ERR: remark: <unknown>:0:0: unable to legalize instruction: %{{[0-9]+}}:_(s128) = G_LOAD %{{[0-9]+}}:_(p0) :: (load (<2 x s63>)) (in function: load-narrow-scalar-high-bits) |
| 6 | + |
| 7 | +# FIXME: Scalarized stores for non-byte-sized vector elements store incorrect partial values. |
| 8 | +--- |
| 9 | +name: store-narrow-non-byte-sized |
| 10 | +tracksRegLiveness: true |
| 11 | +body: | |
| 12 | + bb.1: |
| 13 | + liveins: $x8 |
| 14 | + ; CHECK-LABEL: name: store-narrow-non-byte-sized |
| 15 | + ; CHECK: liveins: $x8 |
| 16 | + ; CHECK-NEXT: {{ $}} |
| 17 | + ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(p0) = COPY $x8 |
| 18 | + ; CHECK-NEXT: [[C:%[0-9]+]]:_(s32) = G_CONSTANT i32 256 |
| 19 | + ; CHECK-NEXT: [[COPY1:%[0-9]+]]:_(s32) = COPY [[C]](s32) |
| 20 | + ; CHECK-NEXT: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 511 |
| 21 | + ; CHECK-NEXT: [[AND:%[0-9]+]]:_(s32) = G_AND [[COPY1]], [[C1]] |
| 22 | + ; CHECK-NEXT: [[TRUNC:%[0-9]+]]:_(s16) = G_TRUNC [[AND]](s32) |
| 23 | + ; CHECK-NEXT: G_STORE [[TRUNC]](s16), [[COPY]](p0) :: (store (s16), align 16) |
| 24 | + ; CHECK-NEXT: [[C2:%[0-9]+]]:_(s64) = G_CONSTANT i64 1 |
| 25 | + ; CHECK-NEXT: [[PTR_ADD:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C2]](s64) |
| 26 | + ; CHECK-NEXT: [[C3:%[0-9]+]]:_(s32) = G_CONSTANT i32 257 |
| 27 | + ; CHECK-NEXT: [[COPY2:%[0-9]+]]:_(s32) = COPY [[C3]](s32) |
| 28 | + ; CHECK-NEXT: [[COPY3:%[0-9]+]]:_(s32) = COPY [[C1]](s32) |
| 29 | + ; CHECK-NEXT: [[AND1:%[0-9]+]]:_(s32) = G_AND [[COPY2]], [[COPY3]] |
| 30 | + ; CHECK-NEXT: [[TRUNC1:%[0-9]+]]:_(s16) = G_TRUNC [[AND1]](s32) |
| 31 | + ; CHECK-NEXT: G_STORE [[TRUNC1]](s16), [[PTR_ADD]](p0) :: (store (s16) into unknown-address + 1, align 1) |
| 32 | + ; CHECK-NEXT: [[C4:%[0-9]+]]:_(s64) = G_CONSTANT i64 2 |
| 33 | + ; CHECK-NEXT: [[PTR_ADD1:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C4]](s64) |
| 34 | + ; CHECK-NEXT: [[COPY4:%[0-9]+]]:_(s32) = COPY [[C]](s32) |
| 35 | + ; CHECK-NEXT: [[COPY5:%[0-9]+]]:_(s32) = COPY [[C1]](s32) |
| 36 | + ; CHECK-NEXT: [[AND2:%[0-9]+]]:_(s32) = G_AND [[COPY4]], [[COPY5]] |
| 37 | + ; CHECK-NEXT: [[TRUNC2:%[0-9]+]]:_(s16) = G_TRUNC [[AND2]](s32) |
| 38 | + ; CHECK-NEXT: G_STORE [[TRUNC2]](s16), [[PTR_ADD1]](p0) :: (store (s16) into unknown-address + 2) |
| 39 | + ; CHECK-NEXT: [[C5:%[0-9]+]]:_(s64) = G_CONSTANT i64 3 |
| 40 | + ; CHECK-NEXT: [[PTR_ADD2:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C5]](s64) |
| 41 | + ; CHECK-NEXT: [[COPY6:%[0-9]+]]:_(s32) = COPY [[C3]](s32) |
| 42 | + ; CHECK-NEXT: [[COPY7:%[0-9]+]]:_(s32) = COPY [[C1]](s32) |
| 43 | + ; CHECK-NEXT: [[AND3:%[0-9]+]]:_(s32) = G_AND [[COPY6]], [[COPY7]] |
| 44 | + ; CHECK-NEXT: [[TRUNC3:%[0-9]+]]:_(s16) = G_TRUNC [[AND3]](s32) |
| 45 | + ; CHECK-NEXT: G_STORE [[TRUNC3]](s16), [[PTR_ADD2]](p0) :: (store (s16) into unknown-address + 3, align 1) |
| 46 | + ; CHECK-NEXT: [[C6:%[0-9]+]]:_(s64) = G_CONSTANT i64 4 |
| 47 | + ; CHECK-NEXT: [[PTR_ADD3:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C6]](s64) |
| 48 | + ; CHECK-NEXT: [[COPY8:%[0-9]+]]:_(s32) = COPY [[C]](s32) |
| 49 | + ; CHECK-NEXT: [[COPY9:%[0-9]+]]:_(s32) = COPY [[C1]](s32) |
| 50 | + ; CHECK-NEXT: [[AND4:%[0-9]+]]:_(s32) = G_AND [[COPY8]], [[COPY9]] |
| 51 | + ; CHECK-NEXT: [[TRUNC4:%[0-9]+]]:_(s16) = G_TRUNC [[AND4]](s32) |
| 52 | + ; CHECK-NEXT: G_STORE [[TRUNC4]](s16), [[PTR_ADD3]](p0) :: (store (s16) into unknown-address + 4, align 4) |
| 53 | + ; CHECK-NEXT: [[C7:%[0-9]+]]:_(s64) = G_CONSTANT i64 5 |
| 54 | + ; CHECK-NEXT: [[PTR_ADD4:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C7]](s64) |
| 55 | + ; CHECK-NEXT: [[COPY10:%[0-9]+]]:_(s32) = COPY [[C3]](s32) |
| 56 | + ; CHECK-NEXT: [[COPY11:%[0-9]+]]:_(s32) = COPY [[C1]](s32) |
| 57 | + ; CHECK-NEXT: [[AND5:%[0-9]+]]:_(s32) = G_AND [[COPY10]], [[COPY11]] |
| 58 | + ; CHECK-NEXT: [[TRUNC5:%[0-9]+]]:_(s16) = G_TRUNC [[AND5]](s32) |
| 59 | + ; CHECK-NEXT: G_STORE [[TRUNC5]](s16), [[PTR_ADD4]](p0) :: (store (s16) into unknown-address + 5, align 1) |
| 60 | + ; CHECK-NEXT: [[C8:%[0-9]+]]:_(s64) = G_CONSTANT i64 6 |
| 61 | + ; CHECK-NEXT: [[PTR_ADD5:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C8]](s64) |
| 62 | + ; CHECK-NEXT: [[COPY12:%[0-9]+]]:_(s32) = COPY [[C1]](s32) |
| 63 | + ; CHECK-NEXT: [[AND6:%[0-9]+]]:_(s32) = G_AND [[C]], [[COPY12]] |
| 64 | + ; CHECK-NEXT: [[TRUNC6:%[0-9]+]]:_(s16) = G_TRUNC [[AND6]](s32) |
| 65 | + ; CHECK-NEXT: G_STORE [[TRUNC6]](s16), [[PTR_ADD5]](p0) :: (store (s16) into unknown-address + 6) |
| 66 | + ; CHECK-NEXT: [[C9:%[0-9]+]]:_(s64) = G_CONSTANT i64 7 |
| 67 | + ; CHECK-NEXT: [[PTR_ADD6:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C9]](s64) |
| 68 | + ; CHECK-NEXT: [[COPY13:%[0-9]+]]:_(s32) = COPY [[C1]](s32) |
| 69 | + ; CHECK-NEXT: [[AND7:%[0-9]+]]:_(s32) = G_AND [[C3]], [[COPY13]] |
| 70 | + ; CHECK-NEXT: [[TRUNC7:%[0-9]+]]:_(s16) = G_TRUNC [[AND7]](s32) |
| 71 | + ; CHECK-NEXT: G_STORE [[TRUNC7]](s16), [[PTR_ADD6]](p0) :: (store (s16) into unknown-address + 7, align 1) |
| 72 | + ; CHECK-NEXT: RET_ReallyLR |
| 73 | + %0:_(p0) = COPY $x8 |
| 74 | + %1:_(s9) = G_CONSTANT i9 256 |
| 75 | + %2:_(s9) = G_CONSTANT i9 257 |
| 76 | + %3:_(<8 x s9>) = G_BUILD_VECTOR %1(s9), %2(s9), %1(s9), %2(s9), %1(s9), %2(s9), %1(s9), %2(s9) |
| 77 | + G_STORE %3(<8 x s9>), %0(p0) :: (store (<8 x s9>), align 16) |
| 78 | + RET_ReallyLR |
| 79 | +... |
| 80 | + |
| 81 | +# FIXME: Vector stores only sometimes act as per-lane truncating stores (see e.g. PR#121169). |
| 82 | +--- |
| 83 | +name: store-narrow-per-lane-trunc |
| 84 | +tracksRegLiveness: true |
| 85 | +body: | |
| 86 | + bb.1: |
| 87 | + liveins: $x8 |
| 88 | + ; CHECK-LABEL: name: store-narrow-per-lane-trunc |
| 89 | + ; CHECK: liveins: $x8 |
| 90 | + ; CHECK-NEXT: {{ $}} |
| 91 | + ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(p0) = COPY $x8 |
| 92 | + ; CHECK-NEXT: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 42 |
| 93 | + ; CHECK-NEXT: [[BUILD_VECTOR:%[0-9]+]]:_(<2 x s64>) = G_BUILD_VECTOR [[C]](s64), [[C]](s64) |
| 94 | + ; CHECK-NEXT: [[BUILD_VECTOR1:%[0-9]+]]:_(<2 x s64>) = G_BUILD_VECTOR [[C]](s64), [[C]](s64) |
| 95 | + ; CHECK-NEXT: G_STORE [[BUILD_VECTOR]](<2 x s64>), [[COPY]](p0) :: (store (<2 x s64>)) |
| 96 | + ; CHECK-NEXT: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 16 |
| 97 | + ; CHECK-NEXT: [[PTR_ADD:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C1]](s64) |
| 98 | + ; CHECK-NEXT: G_STORE [[BUILD_VECTOR1]](<2 x s64>), [[PTR_ADD]](p0) :: (store (<2 x s64>) into unknown-address + 16) |
| 99 | + ; CHECK-NEXT: RET_ReallyLR |
| 100 | + %0:_(p0) = COPY $x8 |
| 101 | + %1:_(s64) = G_CONSTANT i64 42 |
| 102 | + %2:_(<4 x s64>) = G_BUILD_VECTOR %1(s64), %1(s64), %1(s64), %1(s64) |
| 103 | + G_STORE %2(<4 x s64>), %0(p0) :: (store (<4 x s63>), align 16) |
| 104 | + RET_ReallyLR |
| 105 | +... |
| 106 | + |
| 107 | +# FIXME: Clarify behavior of stores between scalar and vector types in documentation. Should we consider this malformed? |
| 108 | +--- |
| 109 | +name: store-narrow-vector-high-bits |
| 110 | +tracksRegLiveness: true |
| 111 | +body: | |
| 112 | + bb.1: |
| 113 | + liveins: $x8 |
| 114 | + ; CHECK-LABEL: name: store-narrow-vector-high-bits |
| 115 | + ; CHECK: liveins: $x8 |
| 116 | + ; CHECK-NEXT: {{ $}} |
| 117 | + ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(p0) = COPY $x8 |
| 118 | + ; CHECK-NEXT: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 42 |
| 119 | + ; CHECK-NEXT: [[BUILD_VECTOR:%[0-9]+]]:_(<2 x s64>) = G_BUILD_VECTOR [[C]](s64), [[C]](s64) |
| 120 | + ; CHECK-NEXT: [[BUILD_VECTOR1:%[0-9]+]]:_(<2 x s64>) = G_BUILD_VECTOR [[C]](s64), [[C]](s64) |
| 121 | + ; CHECK-NEXT: G_STORE [[BUILD_VECTOR]](<2 x s64>), [[COPY]](p0) :: (store (<2 x s64>)) |
| 122 | + ; CHECK-NEXT: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 16 |
| 123 | + ; CHECK-NEXT: [[PTR_ADD:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C1]](s64) |
| 124 | + ; CHECK-NEXT: G_STORE [[BUILD_VECTOR1]](<2 x s64>), [[PTR_ADD]](p0) :: (store (<2 x s64>) into unknown-address + 16) |
| 125 | + ; CHECK-NEXT: RET_ReallyLR |
| 126 | + %0:_(p0) = COPY $x8 |
| 127 | + %1:_(s64) = G_CONSTANT i64 42 |
| 128 | + %2:_(<4 x s64>) = G_BUILD_VECTOR %1(s64), %1(s64), %1(s64), %1(s64) |
| 129 | + G_STORE %2(<4 x s64>), %0(p0) :: (store (s252), align 16) |
| 130 | + RET_ReallyLR |
| 131 | +... |
| 132 | +--- |
| 133 | +name: store-narrow-scalar-high-bits |
| 134 | +tracksRegLiveness: true |
| 135 | +body: | |
| 136 | + bb.1: |
| 137 | + liveins: $x8 |
| 138 | + ; CHECK-LABEL: name: store-narrow-scalar-high-bits |
| 139 | + ; CHECK: liveins: $x8 |
| 140 | + ; CHECK-NEXT: {{ $}} |
| 141 | + ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(p0) = COPY $x8 |
| 142 | + ; CHECK-NEXT: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 42 |
| 143 | + ; CHECK-NEXT: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 0 |
| 144 | + ; CHECK-NEXT: G_STORE [[C]](s64), [[COPY]](p0) :: (store (s64), align 16) |
| 145 | + ; CHECK-NEXT: [[C2:%[0-9]+]]:_(s64) = G_CONSTANT i64 8 |
| 146 | + ; CHECK-NEXT: [[PTR_ADD:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C2]](s64) |
| 147 | + ; CHECK-NEXT: G_STORE [[C1]](s64), [[PTR_ADD]](p0) :: (store (s64) into unknown-address + 8) |
| 148 | + ; CHECK-NEXT: RET_ReallyLR |
| 149 | + %0:_(p0) = COPY $x8 |
| 150 | + %1:_(s128) = G_CONSTANT i128 42 |
| 151 | + G_STORE %1(s128), %0(p0) :: (store (<2 x s63>), align 16) |
| 152 | + RET_ReallyLR |
| 153 | +... |
| 154 | + |
| 155 | + |
| 156 | +# FIXME: Scalarized loads for non-byte-sized vector elements load incorrect partial values. |
| 157 | +--- |
| 158 | +name: load-narrow-non-byte-sized |
| 159 | +tracksRegLiveness: true |
| 160 | +body: | |
| 161 | + bb.1: |
| 162 | + liveins: $x8 |
| 163 | + ; CHECK-LABEL: name: load-narrow-non-byte-sized |
| 164 | + ; CHECK: liveins: $x8 |
| 165 | + ; CHECK-NEXT: {{ $}} |
| 166 | + ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(p0) = COPY $x8 |
| 167 | + ; CHECK-NEXT: [[LOAD:%[0-9]+]]:_(s16) = G_LOAD [[COPY]](p0) :: (load (s16), align 16) |
| 168 | + ; CHECK-NEXT: [[ASSERT_ZEXT:%[0-9]+]]:_(s16) = G_ASSERT_ZEXT [[LOAD]], 9 |
| 169 | + ; CHECK-NEXT: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 1 |
| 170 | + ; CHECK-NEXT: [[PTR_ADD:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C]](s64) |
| 171 | + ; CHECK-NEXT: [[LOAD1:%[0-9]+]]:_(s16) = G_LOAD [[PTR_ADD]](p0) :: (load (s16) from unknown-address + 1, align 1) |
| 172 | + ; CHECK-NEXT: [[ASSERT_ZEXT1:%[0-9]+]]:_(s16) = G_ASSERT_ZEXT [[LOAD1]], 9 |
| 173 | + ; CHECK-NEXT: [[C1:%[0-9]+]]:_(s64) = G_CONSTANT i64 2 |
| 174 | + ; CHECK-NEXT: [[PTR_ADD1:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C1]](s64) |
| 175 | + ; CHECK-NEXT: [[LOAD2:%[0-9]+]]:_(s16) = G_LOAD [[PTR_ADD1]](p0) :: (load (s16) from unknown-address + 2) |
| 176 | + ; CHECK-NEXT: [[ASSERT_ZEXT2:%[0-9]+]]:_(s16) = G_ASSERT_ZEXT [[LOAD2]], 9 |
| 177 | + ; CHECK-NEXT: [[C2:%[0-9]+]]:_(s64) = G_CONSTANT i64 3 |
| 178 | + ; CHECK-NEXT: [[PTR_ADD2:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C2]](s64) |
| 179 | + ; CHECK-NEXT: [[LOAD3:%[0-9]+]]:_(s16) = G_LOAD [[PTR_ADD2]](p0) :: (load (s16) from unknown-address + 3, align 1) |
| 180 | + ; CHECK-NEXT: [[ASSERT_ZEXT3:%[0-9]+]]:_(s16) = G_ASSERT_ZEXT [[LOAD3]], 9 |
| 181 | + ; CHECK-NEXT: [[C3:%[0-9]+]]:_(s64) = G_CONSTANT i64 4 |
| 182 | + ; CHECK-NEXT: [[PTR_ADD3:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C3]](s64) |
| 183 | + ; CHECK-NEXT: [[LOAD4:%[0-9]+]]:_(s16) = G_LOAD [[PTR_ADD3]](p0) :: (load (s16) from unknown-address + 4, align 4) |
| 184 | + ; CHECK-NEXT: [[ASSERT_ZEXT4:%[0-9]+]]:_(s16) = G_ASSERT_ZEXT [[LOAD4]], 9 |
| 185 | + ; CHECK-NEXT: [[C4:%[0-9]+]]:_(s64) = G_CONSTANT i64 5 |
| 186 | + ; CHECK-NEXT: [[PTR_ADD4:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C4]](s64) |
| 187 | + ; CHECK-NEXT: [[LOAD5:%[0-9]+]]:_(s16) = G_LOAD [[PTR_ADD4]](p0) :: (load (s16) from unknown-address + 5, align 1) |
| 188 | + ; CHECK-NEXT: [[ASSERT_ZEXT5:%[0-9]+]]:_(s16) = G_ASSERT_ZEXT [[LOAD5]], 9 |
| 189 | + ; CHECK-NEXT: [[C5:%[0-9]+]]:_(s64) = G_CONSTANT i64 6 |
| 190 | + ; CHECK-NEXT: [[PTR_ADD5:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C5]](s64) |
| 191 | + ; CHECK-NEXT: [[LOAD6:%[0-9]+]]:_(s16) = G_LOAD [[PTR_ADD5]](p0) :: (load (s16) from unknown-address + 6) |
| 192 | + ; CHECK-NEXT: [[ASSERT_ZEXT6:%[0-9]+]]:_(s16) = G_ASSERT_ZEXT [[LOAD6]], 9 |
| 193 | + ; CHECK-NEXT: [[C6:%[0-9]+]]:_(s64) = G_CONSTANT i64 7 |
| 194 | + ; CHECK-NEXT: [[PTR_ADD6:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C6]](s64) |
| 195 | + ; CHECK-NEXT: [[LOAD7:%[0-9]+]]:_(s16) = G_LOAD [[PTR_ADD6]](p0) :: (load (s16) from unknown-address + 7, align 1) |
| 196 | + ; CHECK-NEXT: [[ASSERT_ZEXT7:%[0-9]+]]:_(s16) = G_ASSERT_ZEXT [[LOAD7]], 9 |
| 197 | + ; CHECK-NEXT: [[BUILD_VECTOR:%[0-9]+]]:_(<8 x s16>) = G_BUILD_VECTOR [[ASSERT_ZEXT]](s16), [[ASSERT_ZEXT1]](s16), [[ASSERT_ZEXT2]](s16), [[ASSERT_ZEXT3]](s16), [[ASSERT_ZEXT4]](s16), [[ASSERT_ZEXT5]](s16), [[ASSERT_ZEXT6]](s16), [[ASSERT_ZEXT7]](s16) |
| 198 | + ; CHECK-NEXT: [[C7:%[0-9]+]]:_(s16) = G_CONSTANT i16 511 |
| 199 | + ; CHECK-NEXT: [[BUILD_VECTOR1:%[0-9]+]]:_(<8 x s16>) = G_BUILD_VECTOR [[C7]](s16), [[C7]](s16), [[C7]](s16), [[C7]](s16), [[C7]](s16), [[C7]](s16), [[C7]](s16), [[C7]](s16) |
| 200 | + ; CHECK-NEXT: [[AND:%[0-9]+]]:_(<8 x s16>) = G_AND [[BUILD_VECTOR]], [[BUILD_VECTOR1]] |
| 201 | + ; CHECK-NEXT: $q0 = COPY [[AND]](<8 x s16>) |
| 202 | + ; CHECK-NEXT: RET_ReallyLR implicit $q0 |
| 203 | + %0:_(p0) = COPY $x8 |
| 204 | + %2:_(<8 x s9>) = G_LOAD %0(p0) :: (load (<8 x s9>), align 16) |
| 205 | + %3:_(<8 x s16>) = G_ZEXT %2(<8 x s9>) |
| 206 | + $q0 = COPY %3(<8 x s16>) |
| 207 | + RET_ReallyLR implicit $q0 |
| 208 | +... |
| 209 | + |
| 210 | +# FIXME: Vector stores sometimes act as per-lane truncating stores (see PR#121169). If we want to keep these semantics we should change the semantics of G_LOAD to behave as a per-lane extending load. |
| 211 | +--- |
| 212 | +name: load-narrow-per-lane-ext |
| 213 | +tracksRegLiveness: true |
| 214 | +body: | |
| 215 | + bb.1: |
| 216 | + liveins: $x8 |
| 217 | + ; CHECK-LABEL: name: load-narrow-per-lane-ext |
| 218 | + ; CHECK: liveins: $x8 |
| 219 | + ; CHECK-NEXT: {{ $}} |
| 220 | + ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(p0) = COPY $x8 |
| 221 | + ; CHECK-NEXT: [[LOAD:%[0-9]+]]:_(<2 x s64>) = G_LOAD [[COPY]](p0) :: (load (<2 x s64>)) |
| 222 | + ; CHECK-NEXT: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 16 |
| 223 | + ; CHECK-NEXT: [[PTR_ADD:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C]](s64) |
| 224 | + ; CHECK-NEXT: [[LOAD1:%[0-9]+]]:_(<2 x s64>) = G_LOAD [[PTR_ADD]](p0) :: (load (<2 x s64>) from unknown-address + 16) |
| 225 | + ; CHECK-NEXT: $q0 = COPY [[LOAD]](<2 x s64>) |
| 226 | + ; CHECK-NEXT: $q1 = COPY [[LOAD1]](<2 x s64>) |
| 227 | + ; CHECK-NEXT: RET_ReallyLR implicit $q0, implicit $q1 |
| 228 | + %0:_(p0) = COPY $x8 |
| 229 | + %2:_(<4 x s64>) = G_LOAD %0(p0) :: (load (<4 x s63>), align 16) |
| 230 | + %3:_(<2 x s64>), %4:_(<2 x s64>) = G_UNMERGE_VALUES %2(<4 x s64>) |
| 231 | + $q0 = COPY %3(<2 x s64>) |
| 232 | + $q1 = COPY %4(<2 x s64>) |
| 233 | + RET_ReallyLR implicit $q0, implicit $q1 |
| 234 | +... |
| 235 | + |
| 236 | +# FIXME: Clarify behavior of loads between scalar and vector types in documentation. Should we consider this malformed? |
| 237 | +--- |
| 238 | +name: load-narrow-vector-high-bits |
| 239 | +tracksRegLiveness: true |
| 240 | +body: | |
| 241 | + bb.1: |
| 242 | + liveins: $x8 |
| 243 | + ; CHECK-LABEL: name: load-narrow-vector-high-bits |
| 244 | + ; CHECK: liveins: $x8 |
| 245 | + ; CHECK-NEXT: {{ $}} |
| 246 | + ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(p0) = COPY $x8 |
| 247 | + ; CHECK-NEXT: [[LOAD:%[0-9]+]]:_(<2 x s64>) = G_LOAD [[COPY]](p0) :: (load (<2 x s64>)) |
| 248 | + ; CHECK-NEXT: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 16 |
| 249 | + ; CHECK-NEXT: [[PTR_ADD:%[0-9]+]]:_(p0) = G_PTR_ADD [[COPY]], [[C]](s64) |
| 250 | + ; CHECK-NEXT: [[LOAD1:%[0-9]+]]:_(<2 x s64>) = G_LOAD [[PTR_ADD]](p0) :: (load (<2 x s64>) from unknown-address + 16) |
| 251 | + ; CHECK-NEXT: $q0 = COPY [[LOAD]](<2 x s64>) |
| 252 | + ; CHECK-NEXT: $q1 = COPY [[LOAD1]](<2 x s64>) |
| 253 | + ; CHECK-NEXT: RET_ReallyLR implicit $q0, implicit $q1 |
| 254 | + %0:_(p0) = COPY $x8 |
| 255 | + %2:_(<4 x s64>) = G_LOAD %0(p0) :: (load (s252), align 16) |
| 256 | + %3:_(<2 x s64>), %4:_(<2 x s64>) = G_UNMERGE_VALUES %2(<4 x s64>) |
| 257 | + $q0 = COPY %3(<2 x s64>) |
| 258 | + $q1 = COPY %4(<2 x s64>) |
| 259 | + RET_ReallyLR implicit $q0, implicit $q1 |
| 260 | +... |
| 261 | +--- |
| 262 | +name: load-narrow-scalar-high-bits |
| 263 | +tracksRegLiveness: true |
| 264 | +body: | |
| 265 | + bb.1: |
| 266 | + liveins: $x8 |
| 267 | + ; CHECK-LABEL: name: load-narrow-scalar-high-bits |
| 268 | + ; CHECK: liveins: $x8 |
| 269 | + ; CHECK-NEXT: {{ $}} |
| 270 | + ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(p0) = COPY $x8 |
| 271 | + ; CHECK-NEXT: [[LOAD:%[0-9]+]]:_(s128) = G_LOAD [[COPY]](p0) :: (load (<2 x s63>)) |
| 272 | + ; CHECK-NEXT: [[UV:%[0-9]+]]:_(s64), [[UV1:%[0-9]+]]:_(s64) = G_UNMERGE_VALUES [[LOAD]](s128) |
| 273 | + ; CHECK-NEXT: $x0 = COPY [[UV]](s64) |
| 274 | + ; CHECK-NEXT: $x1 = COPY [[UV1]](s64) |
| 275 | + ; CHECK-NEXT: RET_ReallyLR implicit $x0, implicit $x1 |
| 276 | + %0:_(p0) = COPY $x8 |
| 277 | + %2:_(s128) = G_LOAD %0(p0) :: (load (<2 x s63>), align 16) |
| 278 | + %3:_(s64), %4:_(s64) = G_UNMERGE_VALUES %2(s128) |
| 279 | + $x0 = COPY %3(s64) |
| 280 | + $x1 = COPY %4(s64) |
| 281 | + RET_ReallyLR implicit $x0, implicit $x1 |
| 282 | +... |
0 commit comments