@@ -29,7 +29,7 @@ define i32 @vqdot_vv(<16 x i8> %a, <16 x i8> %b) {
29
29
entry:
30
30
%a.sext = sext <16 x i8 > %a to <16 x i32 >
31
31
%b.sext = sext <16 x i8 > %b to <16 x i32 >
32
- %mul = mul nuw nsw <16 x i32 > %a.sext , %b.sext
32
+ %mul = mul <16 x i32 > %a.sext , %b.sext
33
33
%res = tail call i32 @llvm.vector.reduce.add.v16i32 (<16 x i32 > %mul )
34
34
ret i32 %res
35
35
}
@@ -48,7 +48,7 @@ define i32 @vqdot_vx_constant(<16 x i8> %a) {
48
48
; CHECK-NEXT: ret
49
49
entry:
50
50
%a.sext = sext <16 x i8 > %a to <16 x i32 >
51
- %mul = mul nuw nsw <16 x i32 > %a.sext , splat (i32 23 )
51
+ %mul = mul <16 x i32 > %a.sext , splat (i32 23 )
52
52
%res = tail call i32 @llvm.vector.reduce.add.v16i32 (<16 x i32 > %mul )
53
53
ret i32 %res
54
54
}
@@ -67,7 +67,7 @@ define i32 @vqdot_vx_constant_swapped(<16 x i8> %a) {
67
67
; CHECK-NEXT: ret
68
68
entry:
69
69
%a.sext = sext <16 x i8 > %a to <16 x i32 >
70
- %mul = mul nuw nsw <16 x i32 > splat (i32 23 ), %a.sext
70
+ %mul = mul <16 x i32 > splat (i32 23 ), %a.sext
71
71
%res = tail call i32 @llvm.vector.reduce.add.v16i32 (<16 x i32 > %mul )
72
72
ret i32 %res
73
73
}
@@ -97,7 +97,7 @@ define i32 @vqdotu_vv(<16 x i8> %a, <16 x i8> %b) {
97
97
entry:
98
98
%a.zext = zext <16 x i8 > %a to <16 x i32 >
99
99
%b.zext = zext <16 x i8 > %b to <16 x i32 >
100
- %mul = mul nuw nsw <16 x i32 > %a.zext , %b.zext
100
+ %mul = mul <16 x i32 > %a.zext , %b.zext
101
101
%res = tail call i32 @llvm.vector.reduce.add.v16i32 (<16 x i32 > %mul )
102
102
ret i32 %res
103
103
}
@@ -116,7 +116,7 @@ define i32 @vqdotu_vx_constant(<16 x i8> %a) {
116
116
; CHECK-NEXT: ret
117
117
entry:
118
118
%a.zext = zext <16 x i8 > %a to <16 x i32 >
119
- %mul = mul nuw nsw <16 x i32 > %a.zext , splat (i32 123 )
119
+ %mul = mul <16 x i32 > %a.zext , splat (i32 123 )
120
120
%res = tail call i32 @llvm.vector.reduce.add.v16i32 (<16 x i32 > %mul )
121
121
ret i32 %res
122
122
}
@@ -146,7 +146,7 @@ define i32 @vqdotsu_vv(<16 x i8> %a, <16 x i8> %b) {
146
146
entry:
147
147
%a.sext = sext <16 x i8 > %a to <16 x i32 >
148
148
%b.zext = zext <16 x i8 > %b to <16 x i32 >
149
- %mul = mul nuw nsw <16 x i32 > %a.sext , %b.zext
149
+ %mul = mul <16 x i32 > %a.sext , %b.zext
150
150
%res = tail call i32 @llvm.vector.reduce.add.v16i32 (<16 x i32 > %mul )
151
151
ret i32 %res
152
152
}
@@ -176,7 +176,7 @@ define i32 @vqdotsu_vv_swapped(<16 x i8> %a, <16 x i8> %b) {
176
176
entry:
177
177
%a.sext = sext <16 x i8 > %a to <16 x i32 >
178
178
%b.zext = zext <16 x i8 > %b to <16 x i32 >
179
- %mul = mul nuw nsw <16 x i32 > %b.zext , %a.sext
179
+ %mul = mul <16 x i32 > %b.zext , %a.sext
180
180
%res = tail call i32 @llvm.vector.reduce.add.v16i32 (<16 x i32 > %mul )
181
181
ret i32 %res
182
182
}
@@ -195,7 +195,7 @@ define i32 @vdotqsu_vx_constant(<16 x i8> %a) {
195
195
; CHECK-NEXT: ret
196
196
entry:
197
197
%a.sext = sext <16 x i8 > %a to <16 x i32 >
198
- %mul = mul nuw nsw <16 x i32 > %a.sext , splat (i32 123 )
198
+ %mul = mul <16 x i32 > %a.sext , splat (i32 123 )
199
199
%res = tail call i32 @llvm.vector.reduce.add.v16i32 (<16 x i32 > %mul )
200
200
ret i32 %res
201
201
}
@@ -215,7 +215,7 @@ define i32 @vdotqus_vx_constant(<16 x i8> %a) {
215
215
; CHECK-NEXT: ret
216
216
entry:
217
217
%a.zext = zext <16 x i8 > %a to <16 x i32 >
218
- %mul = mul nuw nsw <16 x i32 > %a.zext , splat (i32 -23 )
218
+ %mul = mul <16 x i32 > %a.zext , splat (i32 -23 )
219
219
%res = tail call i32 @llvm.vector.reduce.add.v16i32 (<16 x i32 > %mul )
220
220
ret i32 %res
221
221
}
@@ -326,7 +326,7 @@ define i32 @vqdot_vv_accum(<16 x i8> %a, <16 x i8> %b, <16 x i32> %x) {
326
326
entry:
327
327
%a.sext = sext <16 x i8 > %a to <16 x i32 >
328
328
%b.sext = sext <16 x i8 > %b to <16 x i32 >
329
- %mul = mul nuw nsw <16 x i32 > %a.sext , %b.sext
329
+ %mul = mul <16 x i32 > %a.sext , %b.sext
330
330
%add = add <16 x i32 > %mul , %x
331
331
%sum = tail call i32 @llvm.vector.reduce.add.v16i32 (<16 x i32 > %add )
332
332
ret i32 %sum
@@ -360,7 +360,7 @@ define i32 @vqdotu_vv_accum(<16 x i8> %a, <16 x i8> %b, <16 x i32> %x) {
360
360
entry:
361
361
%a.zext = zext <16 x i8 > %a to <16 x i32 >
362
362
%b.zext = zext <16 x i8 > %b to <16 x i32 >
363
- %mul = mul nuw nsw <16 x i32 > %a.zext , %b.zext
363
+ %mul = mul <16 x i32 > %a.zext , %b.zext
364
364
%add = add <16 x i32 > %mul , %x
365
365
%sum = tail call i32 @llvm.vector.reduce.add.v16i32 (<16 x i32 > %add )
366
366
ret i32 %sum
@@ -394,7 +394,7 @@ define i32 @vqdotsu_vv_accum(<16 x i8> %a, <16 x i8> %b, <16 x i32> %x) {
394
394
entry:
395
395
%a.sext = sext <16 x i8 > %a to <16 x i32 >
396
396
%b.zext = zext <16 x i8 > %b to <16 x i32 >
397
- %mul = mul nuw nsw <16 x i32 > %a.sext , %b.zext
397
+ %mul = mul <16 x i32 > %a.sext , %b.zext
398
398
%add = add <16 x i32 > %mul , %x
399
399
%sum = tail call i32 @llvm.vector.reduce.add.v16i32 (<16 x i32 > %add )
400
400
ret i32 %sum
@@ -425,7 +425,7 @@ define i32 @vqdot_vv_scalar_add(<16 x i8> %a, <16 x i8> %b, i32 %x) {
425
425
entry:
426
426
%a.sext = sext <16 x i8 > %a to <16 x i32 >
427
427
%b.sext = sext <16 x i8 > %b to <16 x i32 >
428
- %mul = mul nuw nsw <16 x i32 > %a.sext , %b.sext
428
+ %mul = mul <16 x i32 > %a.sext , %b.sext
429
429
%sum = tail call i32 @llvm.vector.reduce.add.v16i32 (<16 x i32 > %mul )
430
430
%add = add i32 %sum , %x
431
431
ret i32 %add
@@ -456,7 +456,7 @@ define i32 @vqdotu_vv_scalar_add(<16 x i8> %a, <16 x i8> %b, i32 %x) {
456
456
entry:
457
457
%a.zext = zext <16 x i8 > %a to <16 x i32 >
458
458
%b.zext = zext <16 x i8 > %b to <16 x i32 >
459
- %mul = mul nuw nsw <16 x i32 > %a.zext , %b.zext
459
+ %mul = mul <16 x i32 > %a.zext , %b.zext
460
460
%sum = tail call i32 @llvm.vector.reduce.add.v16i32 (<16 x i32 > %mul )
461
461
%add = add i32 %sum , %x
462
462
ret i32 %add
@@ -487,7 +487,7 @@ define i32 @vqdotsu_vv_scalar_add(<16 x i8> %a, <16 x i8> %b, i32 %x) {
487
487
entry:
488
488
%a.sext = sext <16 x i8 > %a to <16 x i32 >
489
489
%b.zext = zext <16 x i8 > %b to <16 x i32 >
490
- %mul = mul nuw nsw <16 x i32 > %a.sext , %b.zext
490
+ %mul = mul <16 x i32 > %a.sext , %b.zext
491
491
%sum = tail call i32 @llvm.vector.reduce.add.v16i32 (<16 x i32 > %mul )
492
492
%add = add i32 %sum , %x
493
493
ret i32 %add
@@ -522,10 +522,10 @@ define i32 @vqdot_vv_split(<16 x i8> %a, <16 x i8> %b, <16 x i8> %c, <16 x i8> %
522
522
entry:
523
523
%a.sext = sext <16 x i8 > %a to <16 x i32 >
524
524
%b.sext = sext <16 x i8 > %b to <16 x i32 >
525
- %mul = mul nuw nsw <16 x i32 > %a.sext , %b.sext
525
+ %mul = mul <16 x i32 > %a.sext , %b.sext
526
526
%c.sext = sext <16 x i8 > %c to <16 x i32 >
527
527
%d.sext = sext <16 x i8 > %d to <16 x i32 >
528
- %mul2 = mul nuw nsw <16 x i32 > %c.sext , %d.sext
528
+ %mul2 = mul <16 x i32 > %c.sext , %d.sext
529
529
%add = add <16 x i32 > %mul , %mul2
530
530
%sum = tail call i32 @llvm.vector.reduce.add.v16i32 (<16 x i32 > %add )
531
531
ret i32 %sum
@@ -554,7 +554,7 @@ define <4 x i32> @vqdot_vv_partial_reduce(<16 x i8> %a, <16 x i8> %b) {
554
554
entry:
555
555
%a.sext = sext <16 x i8 > %a to <16 x i32 >
556
556
%b.sext = sext <16 x i8 > %b to <16 x i32 >
557
- %mul = mul nuw nsw <16 x i32 > %a.sext , %b.sext
557
+ %mul = mul <16 x i32 > %a.sext , %b.sext
558
558
%res = call <4 x i32 > @llvm.experimental.vector.partial.reduce.add (<4 x i32 > zeroinitializer , <16 x i32 > %mul )
559
559
ret <4 x i32 > %res
560
560
}
@@ -583,7 +583,7 @@ define <4 x i32> @vqdot_vv_partial_reduce2(<16 x i8> %a, <16 x i8> %b, <4 x i32>
583
583
entry:
584
584
%a.sext = sext <16 x i8 > %a to <16 x i32 >
585
585
%b.sext = sext <16 x i8 > %b to <16 x i32 >
586
- %mul = mul nuw nsw <16 x i32 > %a.sext , %b.sext
586
+ %mul = mul <16 x i32 > %a.sext , %b.sext
587
587
%res = call <4 x i32 > @llvm.experimental.vector.partial.reduce.add (<4 x i32 > %accum , <16 x i32 > %mul )
588
588
ret <4 x i32 > %res
589
589
}
@@ -599,7 +599,7 @@ define <16 x i32> @vqdot_vv_partial_reduce3(<16 x i8> %a, <16 x i8> %b) {
599
599
entry:
600
600
%a.sext = sext <16 x i8 > %a to <16 x i32 >
601
601
%b.sext = sext <16 x i8 > %b to <16 x i32 >
602
- %mul = mul nuw nsw <16 x i32 > %a.sext , %b.sext
602
+ %mul = mul <16 x i32 > %a.sext , %b.sext
603
603
%res = call <16 x i32 > @llvm.experimental.vector.partial.reduce.add.nvx8i32.nvx16i32.nvx16i32 (<16 x i32 > %mul , <16 x i32 > zeroinitializer )
604
604
ret <16 x i32 > %res
605
605
}
0 commit comments