@@ -27,13 +27,13 @@ define void @widen_3xv4i16(ptr %x, ptr %z) {
27
27
; CHECK-NEXT: vsetivli zero, 4, e16, mf2, ta, ma
28
28
; CHECK-NEXT: vle16.v v8, (a0)
29
29
; CHECK-NEXT: addi a2, a0, 8
30
- ; CHECK-NEXT: vle16.v v10 , (a2)
30
+ ; CHECK-NEXT: vle16.v v9 , (a2)
31
31
; CHECK-NEXT: addi a0, a0, 16
32
- ; CHECK-NEXT: vle16.v v12 , (a0)
33
- ; CHECK-NEXT: vsetivli zero, 8, e16, m2 , tu, ma
34
- ; CHECK-NEXT: vslideup.vi v8, v10 , 4
32
+ ; CHECK-NEXT: vle16.v v10 , (a0)
33
+ ; CHECK-NEXT: vsetivli zero, 8, e16, m1 , tu, ma
34
+ ; CHECK-NEXT: vslideup.vi v8, v9 , 4
35
35
; CHECK-NEXT: vsetivli zero, 12, e16, m2, tu, ma
36
- ; CHECK-NEXT: vslideup.vi v8, v12 , 8
36
+ ; CHECK-NEXT: vslideup.vi v8, v10 , 8
37
37
; CHECK-NEXT: vse16.v v8, (a1)
38
38
; CHECK-NEXT: ret
39
39
%a = load <4 x i16 >, ptr %x
@@ -75,17 +75,17 @@ define void @widen_4xv4i16_unaligned(ptr %x, ptr %z) {
75
75
; CHECK-NO-MISALIGN-NEXT: vsetivli zero, 8, e8, mf2, ta, ma
76
76
; CHECK-NO-MISALIGN-NEXT: vle8.v v8, (a0)
77
77
; CHECK-NO-MISALIGN-NEXT: addi a2, a0, 8
78
- ; CHECK-NO-MISALIGN-NEXT: vle8.v v10 , (a2)
78
+ ; CHECK-NO-MISALIGN-NEXT: vle8.v v9 , (a2)
79
79
; CHECK-NO-MISALIGN-NEXT: addi a2, a0, 16
80
- ; CHECK-NO-MISALIGN-NEXT: vle8.v v12 , (a2)
80
+ ; CHECK-NO-MISALIGN-NEXT: vle8.v v10 , (a2)
81
81
; CHECK-NO-MISALIGN-NEXT: addi a0, a0, 24
82
- ; CHECK-NO-MISALIGN-NEXT: vle8.v v14 , (a0)
83
- ; CHECK-NO-MISALIGN-NEXT: vsetivli zero, 8 , e16, m2 , tu, ma
84
- ; CHECK-NO-MISALIGN-NEXT: vslideup.vi v8, v10 , 4
82
+ ; CHECK-NO-MISALIGN-NEXT: vle8.v v12 , (a0)
83
+ ; CHECK-NO-MISALIGN-NEXT: vsetvli zero, zero , e16, m1 , tu, ma
84
+ ; CHECK-NO-MISALIGN-NEXT: vslideup.vi v8, v9 , 4
85
85
; CHECK-NO-MISALIGN-NEXT: vsetivli zero, 12, e16, m2, tu, ma
86
- ; CHECK-NO-MISALIGN-NEXT: vslideup.vi v8, v12 , 8
86
+ ; CHECK-NO-MISALIGN-NEXT: vslideup.vi v8, v10 , 8
87
87
; CHECK-NO-MISALIGN-NEXT: vsetivli zero, 16, e16, m2, ta, ma
88
- ; CHECK-NO-MISALIGN-NEXT: vslideup.vi v8, v14 , 12
88
+ ; CHECK-NO-MISALIGN-NEXT: vslideup.vi v8, v12 , 12
89
89
; CHECK-NO-MISALIGN-NEXT: vse16.v v8, (a1)
90
90
; CHECK-NO-MISALIGN-NEXT: ret
91
91
;
@@ -188,17 +188,17 @@ define void @strided_constant_mismatch_4xv4i16(ptr %x, ptr %z) {
188
188
; CHECK-NEXT: vsetivli zero, 4, e16, mf2, ta, ma
189
189
; CHECK-NEXT: vle16.v v8, (a0)
190
190
; CHECK-NEXT: addi a2, a0, 2
191
- ; CHECK-NEXT: vle16.v v10 , (a2)
191
+ ; CHECK-NEXT: vle16.v v9 , (a2)
192
192
; CHECK-NEXT: addi a2, a0, 6
193
- ; CHECK-NEXT: vle16.v v12 , (a2)
193
+ ; CHECK-NEXT: vle16.v v10 , (a2)
194
194
; CHECK-NEXT: addi a0, a0, 8
195
- ; CHECK-NEXT: vle16.v v14 , (a0)
196
- ; CHECK-NEXT: vsetivli zero, 8, e16, m2 , tu, ma
197
- ; CHECK-NEXT: vslideup.vi v8, v10 , 4
195
+ ; CHECK-NEXT: vle16.v v12 , (a0)
196
+ ; CHECK-NEXT: vsetivli zero, 8, e16, m1 , tu, ma
197
+ ; CHECK-NEXT: vslideup.vi v8, v9 , 4
198
198
; CHECK-NEXT: vsetivli zero, 12, e16, m2, tu, ma
199
- ; CHECK-NEXT: vslideup.vi v8, v12 , 8
199
+ ; CHECK-NEXT: vslideup.vi v8, v10 , 8
200
200
; CHECK-NEXT: vsetivli zero, 16, e16, m2, ta, ma
201
- ; CHECK-NEXT: vslideup.vi v8, v14 , 12
201
+ ; CHECK-NEXT: vslideup.vi v8, v12 , 12
202
202
; CHECK-NEXT: vse16.v v8, (a1)
203
203
; CHECK-NEXT: ret
204
204
%a = load <4 x i16 >, ptr %x
@@ -258,17 +258,17 @@ define void @strided_runtime_mismatch_4xv4i16(ptr %x, ptr %z, i64 %s, i64 %t) {
258
258
; RV32-NEXT: vsetivli zero, 4, e16, mf2, ta, ma
259
259
; RV32-NEXT: vle16.v v8, (a0)
260
260
; RV32-NEXT: add a0, a0, a2
261
- ; RV32-NEXT: vle16.v v10 , (a0)
261
+ ; RV32-NEXT: vle16.v v9 , (a0)
262
262
; RV32-NEXT: add a0, a0, a4
263
- ; RV32-NEXT: vle16.v v12 , (a0)
263
+ ; RV32-NEXT: vle16.v v10 , (a0)
264
264
; RV32-NEXT: add a0, a0, a2
265
- ; RV32-NEXT: vle16.v v14 , (a0)
266
- ; RV32-NEXT: vsetivli zero, 8, e16, m2 , tu, ma
267
- ; RV32-NEXT: vslideup.vi v8, v10 , 4
265
+ ; RV32-NEXT: vle16.v v12 , (a0)
266
+ ; RV32-NEXT: vsetivli zero, 8, e16, m1 , tu, ma
267
+ ; RV32-NEXT: vslideup.vi v8, v9 , 4
268
268
; RV32-NEXT: vsetivli zero, 12, e16, m2, tu, ma
269
- ; RV32-NEXT: vslideup.vi v8, v12 , 8
269
+ ; RV32-NEXT: vslideup.vi v8, v10 , 8
270
270
; RV32-NEXT: vsetivli zero, 16, e16, m2, ta, ma
271
- ; RV32-NEXT: vslideup.vi v8, v14 , 12
271
+ ; RV32-NEXT: vslideup.vi v8, v12 , 12
272
272
; RV32-NEXT: vse16.v v8, (a1)
273
273
; RV32-NEXT: ret
274
274
;
@@ -277,17 +277,17 @@ define void @strided_runtime_mismatch_4xv4i16(ptr %x, ptr %z, i64 %s, i64 %t) {
277
277
; RV64-NEXT: vsetivli zero, 4, e16, mf2, ta, ma
278
278
; RV64-NEXT: vle16.v v8, (a0)
279
279
; RV64-NEXT: add a0, a0, a2
280
- ; RV64-NEXT: vle16.v v10 , (a0)
280
+ ; RV64-NEXT: vle16.v v9 , (a0)
281
281
; RV64-NEXT: add a0, a0, a3
282
- ; RV64-NEXT: vle16.v v12 , (a0)
282
+ ; RV64-NEXT: vle16.v v10 , (a0)
283
283
; RV64-NEXT: add a0, a0, a2
284
- ; RV64-NEXT: vle16.v v14 , (a0)
285
- ; RV64-NEXT: vsetivli zero, 8, e16, m2 , tu, ma
286
- ; RV64-NEXT: vslideup.vi v8, v10 , 4
284
+ ; RV64-NEXT: vle16.v v12 , (a0)
285
+ ; RV64-NEXT: vsetivli zero, 8, e16, m1 , tu, ma
286
+ ; RV64-NEXT: vslideup.vi v8, v9 , 4
287
287
; RV64-NEXT: vsetivli zero, 12, e16, m2, tu, ma
288
- ; RV64-NEXT: vslideup.vi v8, v12 , 8
288
+ ; RV64-NEXT: vslideup.vi v8, v10 , 8
289
289
; RV64-NEXT: vsetivli zero, 16, e16, m2, ta, ma
290
- ; RV64-NEXT: vslideup.vi v8, v14 , 12
290
+ ; RV64-NEXT: vslideup.vi v8, v12 , 12
291
291
; RV64-NEXT: vse16.v v8, (a1)
292
292
; RV64-NEXT: ret
293
293
;
@@ -296,17 +296,17 @@ define void @strided_runtime_mismatch_4xv4i16(ptr %x, ptr %z, i64 %s, i64 %t) {
296
296
; ZVE64F-NEXT: vsetivli zero, 4, e16, mf2, ta, ma
297
297
; ZVE64F-NEXT: vle16.v v8, (a0)
298
298
; ZVE64F-NEXT: add a0, a0, a2
299
- ; ZVE64F-NEXT: vle16.v v10 , (a0)
299
+ ; ZVE64F-NEXT: vle16.v v9 , (a0)
300
300
; ZVE64F-NEXT: add a0, a0, a3
301
- ; ZVE64F-NEXT: vle16.v v12 , (a0)
301
+ ; ZVE64F-NEXT: vle16.v v10 , (a0)
302
302
; ZVE64F-NEXT: add a0, a0, a2
303
- ; ZVE64F-NEXT: vle16.v v14 , (a0)
304
- ; ZVE64F-NEXT: vsetivli zero, 8, e16, m2 , tu, ma
305
- ; ZVE64F-NEXT: vslideup.vi v8, v10 , 4
303
+ ; ZVE64F-NEXT: vle16.v v12 , (a0)
304
+ ; ZVE64F-NEXT: vsetivli zero, 8, e16, m1 , tu, ma
305
+ ; ZVE64F-NEXT: vslideup.vi v8, v9 , 4
306
306
; ZVE64F-NEXT: vsetivli zero, 12, e16, m2, tu, ma
307
- ; ZVE64F-NEXT: vslideup.vi v8, v12 , 8
307
+ ; ZVE64F-NEXT: vslideup.vi v8, v10 , 8
308
308
; ZVE64F-NEXT: vsetivli zero, 16, e16, m2, ta, ma
309
- ; ZVE64F-NEXT: vslideup.vi v8, v14 , 12
309
+ ; ZVE64F-NEXT: vslideup.vi v8, v12 , 12
310
310
; ZVE64F-NEXT: vse16.v v8, (a1)
311
311
; ZVE64F-NEXT: ret
312
312
%a = load <4 x i16 >, ptr %x
0 commit comments