|
1 | 1 | ; NOTE: Assertions have been autogenerated by utils/update_mir_test_checks.py
|
2 |
| -; RUN: llc -O0 -mtriple=aarch64-linux-gnu -global-isel -stop-after=irtranslator %s -o - | FileCheck %s |
| 2 | +; RUN: llc -O0 -mtriple=aarch64-linux-gnu -mattr=+sve -global-isel -stop-after=irtranslator -aarch64-enable-gisel-sve=1 %s -o - | FileCheck %s |
3 | 3 |
|
4 | 4 | define i32 @extract_v4i32_vector_insert_const(<4 x i32> %a, <2 x i32> %b, i32 %c) {
|
5 | 5 | ; CHECK-LABEL: name: extract_v4i32_vector_insert_const
|
@@ -58,21 +58,149 @@ entry:
|
58 | 58 | ret i32 %d
|
59 | 59 | }
|
60 | 60 |
|
61 |
| -define i32 @extract_v4i32_vector_extract_const(<4 x i32> %a, <2 x i32> %b, i32 %c) { |
| 61 | +define i32 @extract_v4i32_vector_extract_const(<vscale x 4 x i32> %a, i32 %c, ptr %p) { |
62 | 62 | ; CHECK-LABEL: name: extract_v4i32_vector_extract_const
|
63 | 63 | ; CHECK: bb.1.entry:
|
64 |
| - ; CHECK-NEXT: liveins: $d1, $q0, $w0 |
| 64 | + ; CHECK-NEXT: liveins: $w0, $x1, $z0 |
| 65 | + ; CHECK-NEXT: {{ $}} |
| 66 | + ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(<vscale x 4 x s32>) = COPY $z0 |
| 67 | + ; CHECK-NEXT: [[COPY1:%[0-9]+]]:_(s32) = COPY $w0 |
| 68 | + ; CHECK-NEXT: [[COPY2:%[0-9]+]]:_(p0) = COPY $x1 |
| 69 | + ; CHECK-NEXT: [[C:%[0-9]+]]:_(s32) = G_CONSTANT i32 1 |
| 70 | + ; CHECK-NEXT: [[EXTRACT_SUBVECTOR:%[0-9]+]]:_(<vscale x 4 x s32>) = G_EXTRACT_SUBVECTOR [[COPY]](<vscale x 4 x s32>), 0 |
| 71 | + ; CHECK-NEXT: G_STORE [[EXTRACT_SUBVECTOR]](<vscale x 4 x s32>), [[COPY2]](p0) :: (store (<vscale x 4 x s32>) into %ir.p) |
| 72 | + ; CHECK-NEXT: $w0 = COPY [[C]](s32) |
| 73 | + ; CHECK-NEXT: RET_ReallyLR implicit $w0 |
| 74 | +entry: |
| 75 | + %vector = call <vscale x 4 x i32> @llvm.vector.extract(<vscale x 4 x i32> %a, i64 0) |
| 76 | + store <vscale x 4 x i32> %vector, ptr %p, align 16 |
| 77 | + ret i32 1 |
| 78 | +} |
| 79 | + |
| 80 | +define i32 @extract_v4i32_vector_insert_const_vscale(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b, i32 %c, ptr %p) { |
| 81 | + ; CHECK-LABEL: name: extract_v4i32_vector_insert_const_vscale |
| 82 | + ; CHECK: bb.1.entry: |
| 83 | + ; CHECK-NEXT: liveins: $w0, $x1, $z0, $z1 |
| 84 | + ; CHECK-NEXT: {{ $}} |
| 85 | + ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(<vscale x 4 x s32>) = COPY $z0 |
| 86 | + ; CHECK-NEXT: [[COPY1:%[0-9]+]]:_(<vscale x 4 x s32>) = COPY $z1 |
| 87 | + ; CHECK-NEXT: [[COPY2:%[0-9]+]]:_(s32) = COPY $w0 |
| 88 | + ; CHECK-NEXT: [[COPY3:%[0-9]+]]:_(p0) = COPY $x1 |
| 89 | + ; CHECK-NEXT: [[C:%[0-9]+]]:_(s32) = G_CONSTANT i32 1 |
| 90 | + ; CHECK-NEXT: [[INSERT_SUBVECTOR:%[0-9]+]]:_(<vscale x 4 x s32>) = G_INSERT_SUBVECTOR [[COPY]], [[COPY1]](<vscale x 4 x s32>), 0 |
| 91 | + ; CHECK-NEXT: G_STORE [[INSERT_SUBVECTOR]](<vscale x 4 x s32>), [[COPY3]](p0) :: (store (<vscale x 4 x s32>) into %ir.p) |
| 92 | + ; CHECK-NEXT: $w0 = COPY [[C]](s32) |
| 93 | + ; CHECK-NEXT: RET_ReallyLR implicit $w0 |
| 94 | +entry: |
| 95 | + %vector = call <vscale x 4 x i32> @llvm.vector.insert.nxv4i32.v4i32(<vscale x 4 x i32> %a, <vscale x 4 x i32> %b, i64 0) |
| 96 | + store <vscale x 4 x i32> %vector, ptr %p, align 16 |
| 97 | + ret i32 1 |
| 98 | +} |
| 99 | + |
| 100 | +define i32 @extract_v4i32_vector_extract_const_illegal_fixed(<4 x i32> %a, ptr %p) { |
| 101 | + ; CHECK-LABEL: name: extract_v4i32_vector_extract_const_illegal_fixed |
| 102 | + ; CHECK: bb.1.entry: |
| 103 | + ; CHECK-NEXT: liveins: $q0, $x0 |
| 104 | + ; CHECK-NEXT: {{ $}} |
| 105 | + ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(<4 x s32>) = COPY $q0 |
| 106 | + ; CHECK-NEXT: [[COPY1:%[0-9]+]]:_(p0) = COPY $x0 |
| 107 | + ; CHECK-NEXT: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0 |
| 108 | + ; CHECK-NEXT: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 1 |
| 109 | + ; CHECK-NEXT: [[EVEC:%[0-9]+]]:_(s32) = G_EXTRACT_VECTOR_ELT [[COPY]](<4 x s32>), [[C]](s64) |
| 110 | + ; CHECK-NEXT: G_STORE [[EVEC]](s32), [[COPY1]](p0) :: (store (s32) into %ir.p, align 16) |
| 111 | + ; CHECK-NEXT: $w0 = COPY [[C1]](s32) |
| 112 | + ; CHECK-NEXT: RET_ReallyLR implicit $w0 |
| 113 | +entry: |
| 114 | + %vector = call <1 x i32> @llvm.vector.extract(<4 x i32> %a, i64 0) |
| 115 | + store <1 x i32> %vector, ptr %p, align 16 |
| 116 | + ret i32 1 |
| 117 | +} |
| 118 | + |
| 119 | +define i32 @extract_v4i32_vector_extract_const_illegal_scalable(<vscale x 4 x i32> %a, ptr %p) { |
| 120 | + ; CHECK-LABEL: name: extract_v4i32_vector_extract_const_illegal_scalable |
| 121 | + ; CHECK: bb.1.entry: |
| 122 | + ; CHECK-NEXT: liveins: $x0, $z0 |
| 123 | + ; CHECK-NEXT: {{ $}} |
| 124 | + ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(<vscale x 4 x s32>) = COPY $z0 |
| 125 | + ; CHECK-NEXT: [[COPY1:%[0-9]+]]:_(p0) = COPY $x0 |
| 126 | + ; CHECK-NEXT: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0 |
| 127 | + ; CHECK-NEXT: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 1 |
| 128 | + ; CHECK-NEXT: [[VSCALE:%[0-9]+]]:_(s64) = G_VSCALE i64 1 |
| 129 | + ; CHECK-NEXT: [[MUL:%[0-9]+]]:_(s64) = G_MUL [[VSCALE]], [[C]] |
| 130 | + ; CHECK-NEXT: [[EVEC:%[0-9]+]]:_(s32) = G_EXTRACT_VECTOR_ELT [[COPY]](<vscale x 4 x s32>), [[MUL]](s64) |
| 131 | + ; CHECK-NEXT: G_STORE [[EVEC]](s32), [[COPY1]](p0) :: (store (s32) into %ir.p, align 16) |
| 132 | + ; CHECK-NEXT: $w0 = COPY [[C1]](s32) |
| 133 | + ; CHECK-NEXT: RET_ReallyLR implicit $w0 |
| 134 | +entry: |
| 135 | + %vector = call <1 x i32> @llvm.vector.extract(<vscale x 4 x i32> %a, i64 0) |
| 136 | + store <1 x i32> %vector, ptr %p, align 16 |
| 137 | + ret i32 1 |
| 138 | +} |
| 139 | + |
| 140 | +define i32 @extract_v4i32_vector_insert_const_illegal_scalable(<vscale x 4 x i32> %a, <1 x i32> %b, i32 %c, ptr %p) { |
| 141 | + ; CHECK-LABEL: name: extract_v4i32_vector_insert_const_illegal_scalable |
| 142 | + ; CHECK: bb.1.entry: |
| 143 | + ; CHECK-NEXT: liveins: $d1, $w0, $x1, $z0 |
| 144 | + ; CHECK-NEXT: {{ $}} |
| 145 | + ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(<vscale x 4 x s32>) = COPY $z0 |
| 146 | + ; CHECK-NEXT: [[COPY1:%[0-9]+]]:_(<2 x s32>) = COPY $d1 |
| 147 | + ; CHECK-NEXT: [[UV:%[0-9]+]]:_(s32), [[UV1:%[0-9]+]]:_(s32) = G_UNMERGE_VALUES [[COPY1]](<2 x s32>) |
| 148 | + ; CHECK-NEXT: [[COPY2:%[0-9]+]]:_(s32) = COPY $w0 |
| 149 | + ; CHECK-NEXT: [[COPY3:%[0-9]+]]:_(p0) = COPY $x1 |
| 150 | + ; CHECK-NEXT: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0 |
| 151 | + ; CHECK-NEXT: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 1 |
| 152 | + ; CHECK-NEXT: [[VSCALE:%[0-9]+]]:_(s64) = G_VSCALE i64 1 |
| 153 | + ; CHECK-NEXT: [[MUL:%[0-9]+]]:_(s64) = G_MUL [[VSCALE]], [[C]] |
| 154 | + ; CHECK-NEXT: [[IVEC:%[0-9]+]]:_(<vscale x 4 x s32>) = G_INSERT_VECTOR_ELT [[COPY]], [[UV]](s32), [[MUL]](s64) |
| 155 | + ; CHECK-NEXT: G_STORE [[IVEC]](<vscale x 4 x s32>), [[COPY3]](p0) :: (store (<vscale x 4 x s32>) into %ir.p) |
| 156 | + ; CHECK-NEXT: $w0 = COPY [[C1]](s32) |
| 157 | + ; CHECK-NEXT: RET_ReallyLR implicit $w0 |
| 158 | +entry: |
| 159 | + %vector = call <vscale x 4 x i32> @llvm.vector.insert.nxv4i32.v4i32(<vscale x 4 x i32> %a, <1 x i32> %b, i64 0) |
| 160 | + store <vscale x 4 x i32> %vector, ptr %p, align 16 |
| 161 | + ret i32 1 |
| 162 | +} |
| 163 | + |
| 164 | +define i32 @extract_v4i32_vector_insert_const_fixed(<4 x i32> %a, <1 x i32> %b, i32 %c, ptr %p) { |
| 165 | + ; CHECK-LABEL: name: extract_v4i32_vector_insert_const_fixed |
| 166 | + ; CHECK: bb.1.entry: |
| 167 | + ; CHECK-NEXT: liveins: $d1, $q0, $w0, $x1 |
65 | 168 | ; CHECK-NEXT: {{ $}}
|
66 | 169 | ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(<4 x s32>) = COPY $q0
|
67 | 170 | ; CHECK-NEXT: [[COPY1:%[0-9]+]]:_(<2 x s32>) = COPY $d1
|
| 171 | + ; CHECK-NEXT: [[UV:%[0-9]+]]:_(s32), [[UV1:%[0-9]+]]:_(s32) = G_UNMERGE_VALUES [[COPY1]](<2 x s32>) |
68 | 172 | ; CHECK-NEXT: [[COPY2:%[0-9]+]]:_(s32) = COPY $w0
|
| 173 | + ; CHECK-NEXT: [[COPY3:%[0-9]+]]:_(p0) = COPY $x1 |
69 | 174 | ; CHECK-NEXT: [[C:%[0-9]+]]:_(s64) = G_CONSTANT i64 0
|
70 |
| - ; CHECK-NEXT: [[EXTRACT_SUBVECTOR:%[0-9]+]]:_(<4 x s32>) = G_EXTRACT_SUBVECTOR [[COPY]](<4 x s32>), 0 |
71 |
| - ; CHECK-NEXT: [[EVEC:%[0-9]+]]:_(s32) = G_EXTRACT_VECTOR_ELT [[EXTRACT_SUBVECTOR]](<4 x s32>), [[C]](s64) |
72 |
| - ; CHECK-NEXT: $w0 = COPY [[EVEC]](s32) |
| 175 | + ; CHECK-NEXT: [[C1:%[0-9]+]]:_(s32) = G_CONSTANT i32 1 |
| 176 | + ; CHECK-NEXT: [[IVEC:%[0-9]+]]:_(<4 x s32>) = G_INSERT_VECTOR_ELT [[COPY]], [[UV]](s32), [[C]](s64) |
| 177 | + ; CHECK-NEXT: G_STORE [[IVEC]](<4 x s32>), [[COPY3]](p0) :: (store (<4 x s32>) into %ir.p) |
| 178 | + ; CHECK-NEXT: $w0 = COPY [[C1]](s32) |
73 | 179 | ; CHECK-NEXT: RET_ReallyLR implicit $w0
|
74 | 180 | entry:
|
75 |
| - %vector = call <4 x i32> @llvm.vector.extract.v2i32.v4i32(<4 x i32> %a, i64 0) |
76 |
| - %d = extractelement <4 x i32> %vector, i32 0 |
77 |
| - ret i32 %d |
| 181 | + %vector = call <4 x i32> @llvm.vector.insert.v4i32.v4i32(<4 x i32> %a, <1 x i32> %b, i64 0) |
| 182 | + store <4 x i32> %vector, ptr %p, align 16 |
| 183 | + ret i32 1 |
| 184 | +} |
| 185 | + |
| 186 | +define i32 @extract_v4i32_vector_insert_const_fixed_illegal(<1 x i32> %a, <1 x i32> %b, i32 %c, ptr %p) { |
| 187 | + ; CHECK-LABEL: name: extract_v4i32_vector_insert_const_fixed_illegal |
| 188 | + ; CHECK: bb.1.entry: |
| 189 | + ; CHECK-NEXT: liveins: $d0, $d1, $w0, $x1 |
| 190 | + ; CHECK-NEXT: {{ $}} |
| 191 | + ; CHECK-NEXT: [[COPY:%[0-9]+]]:_(<2 x s32>) = COPY $d0 |
| 192 | + ; CHECK-NEXT: [[UV:%[0-9]+]]:_(s32), [[UV1:%[0-9]+]]:_(s32) = G_UNMERGE_VALUES [[COPY]](<2 x s32>) |
| 193 | + ; CHECK-NEXT: [[COPY1:%[0-9]+]]:_(<2 x s32>) = COPY $d1 |
| 194 | + ; CHECK-NEXT: [[UV2:%[0-9]+]]:_(s32), [[UV3:%[0-9]+]]:_(s32) = G_UNMERGE_VALUES [[COPY1]](<2 x s32>) |
| 195 | + ; CHECK-NEXT: [[COPY2:%[0-9]+]]:_(s32) = COPY $w0 |
| 196 | + ; CHECK-NEXT: [[COPY3:%[0-9]+]]:_(p0) = COPY $x1 |
| 197 | + ; CHECK-NEXT: [[C:%[0-9]+]]:_(s32) = G_CONSTANT i32 1 |
| 198 | + ; CHECK-NEXT: [[COPY4:%[0-9]+]]:_(s32) = COPY [[UV]](s32) |
| 199 | + ; CHECK-NEXT: G_STORE [[COPY4]](s32), [[COPY3]](p0) :: (store (s32) into %ir.p, align 16) |
| 200 | + ; CHECK-NEXT: $w0 = COPY [[C]](s32) |
| 201 | + ; CHECK-NEXT: RET_ReallyLR implicit $w0 |
| 202 | +entry: |
| 203 | + %vector = call <1 x i32> @llvm.vector.insert.v1i32.v4i32(<1 x i32> %a, <1 x i32> %b, i64 0) |
| 204 | + store <1 x i32> %vector, ptr %p, align 16 |
| 205 | + ret i32 1 |
78 | 206 | }
|
0 commit comments