|
@@ -367,12 +367,12 @@ define <4 x i64> @test_mm256_broadcastsi128_si256_mem(<2 x i64>* %p0) {
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vbroadcastf128 {{.*#+}} ymm0 = mem[0,1,0,1]
|
|
; X86-NEXT: vbroadcastf128 {{.*#+}} ymm0 = mem[0,1,0,1]
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_broadcastsi128_si256_mem:
|
|
; X64-LABEL: test_mm256_broadcastsi128_si256_mem:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vbroadcastf128 {{.*#+}} ymm0 = mem[0,1,0,1]
|
|
; X64-NEXT: vbroadcastf128 {{.*#+}} ymm0 = mem[0,1,0,1]
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%a0 = load <2 x i64>, <2 x i64>* %p0
|
|
%a0 = load <2 x i64>, <2 x i64>* %p0
|
|
%res = shufflevector <2 x i64> %a0, <2 x i64> undef, <4 x i32> <i32 0, i32 1, i32 0, i32 1>
|
|
%res = shufflevector <2 x i64> %a0, <2 x i64> undef, <4 x i32> <i32 0, i32 1, i32 0, i32 1>
|
|
ret <4 x i64> %res
|
|
ret <4 x i64> %res
|
|
@@ -766,7 +766,7 @@ define <2 x i64> @test_mm_i32gather_epi32(i32 *%a0, <2 x i64> %a1) {
|
|
; X86-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vpgatherdd %xmm2, (%eax,%xmm0,2), %xmm1
|
|
; X86-NEXT: vpgatherdd %xmm2, (%eax,%xmm0,2), %xmm1
|
|
; X86-NEXT: vmovdqa %xmm1, %xmm0
|
|
; X86-NEXT: vmovdqa %xmm1, %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_i32gather_epi32:
|
|
; X64-LABEL: test_mm_i32gather_epi32:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
@@ -774,7 +774,7 @@ define <2 x i64> @test_mm_i32gather_epi32(i32 *%a0, <2 x i64> %a1) {
|
|
; X64-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vpgatherdd %xmm2, (%rdi,%xmm0,2), %xmm1
|
|
; X64-NEXT: vpgatherdd %xmm2, (%rdi,%xmm0,2), %xmm1
|
|
; X64-NEXT: vmovdqa %xmm1, %xmm0
|
|
; X64-NEXT: vmovdqa %xmm1, %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast i32 *%a0 to i8*
|
|
%arg0 = bitcast i32 *%a0 to i8*
|
|
%arg1 = bitcast <2 x i64> %a1 to <4 x i32>
|
|
%arg1 = bitcast <2 x i64> %a1 to <4 x i32>
|
|
%mask = bitcast <2 x i64> <i64 -1, i64 -1> to <4 x i32>
|
|
%mask = bitcast <2 x i64> <i64 -1, i64 -1> to <4 x i32>
|
|
@@ -789,12 +789,12 @@ define <2 x i64> @test_mm_mask_i32gather_epi32(<2 x i64> %a0, i32 *%a1, <2 x i64
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vpgatherdd %xmm2, (%eax,%xmm1,2), %xmm0
|
|
; X86-NEXT: vpgatherdd %xmm2, (%eax,%xmm1,2), %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_mask_i32gather_epi32:
|
|
; X64-LABEL: test_mm_mask_i32gather_epi32:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vpgatherdd %xmm2, (%rdi,%xmm1,2), %xmm0
|
|
; X64-NEXT: vpgatherdd %xmm2, (%rdi,%xmm1,2), %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast <2 x i64> %a0 to <4 x i32>
|
|
%arg0 = bitcast <2 x i64> %a0 to <4 x i32>
|
|
%arg1 = bitcast i32 *%a1 to i8*
|
|
%arg1 = bitcast i32 *%a1 to i8*
|
|
%arg2 = bitcast <2 x i64> %a2 to <4 x i32>
|
|
%arg2 = bitcast <2 x i64> %a2 to <4 x i32>
|
|
@@ -812,7 +812,7 @@ define <4 x i64> @test_mm256_i32gather_epi32(i32 *%a0, <4 x i64> %a1) {
|
|
; X86-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vpgatherdd %ymm2, (%eax,%ymm0,2), %ymm1
|
|
; X86-NEXT: vpgatherdd %ymm2, (%eax,%ymm0,2), %ymm1
|
|
; X86-NEXT: vmovdqa %ymm1, %ymm0
|
|
; X86-NEXT: vmovdqa %ymm1, %ymm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_i32gather_epi32:
|
|
; X64-LABEL: test_mm256_i32gather_epi32:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
@@ -820,7 +820,7 @@ define <4 x i64> @test_mm256_i32gather_epi32(i32 *%a0, <4 x i64> %a1) {
|
|
; X64-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vpgatherdd %ymm2, (%rdi,%ymm0,2), %ymm1
|
|
; X64-NEXT: vpgatherdd %ymm2, (%rdi,%ymm0,2), %ymm1
|
|
; X64-NEXT: vmovdqa %ymm1, %ymm0
|
|
; X64-NEXT: vmovdqa %ymm1, %ymm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast i32 *%a0 to i8*
|
|
%arg0 = bitcast i32 *%a0 to i8*
|
|
%arg1 = bitcast <4 x i64> %a1 to <8 x i32>
|
|
%arg1 = bitcast <4 x i64> %a1 to <8 x i32>
|
|
%mask = bitcast <4 x i64> <i64 -1, i64 -1, i64 -1, i64 -1> to <8 x i32>
|
|
%mask = bitcast <4 x i64> <i64 -1, i64 -1, i64 -1, i64 -1> to <8 x i32>
|
|
@@ -835,12 +835,12 @@ define <4 x i64> @test_mm256_mask_i32gather_epi32(<4 x i64> %a0, i32 *%a1, <4 x
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vpgatherdd %ymm2, (%eax,%ymm1,2), %ymm0
|
|
; X86-NEXT: vpgatherdd %ymm2, (%eax,%ymm1,2), %ymm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_mask_i32gather_epi32:
|
|
; X64-LABEL: test_mm256_mask_i32gather_epi32:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vpgatherdd %ymm2, (%rdi,%ymm1,2), %ymm0
|
|
; X64-NEXT: vpgatherdd %ymm2, (%rdi,%ymm1,2), %ymm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast <4 x i64> %a0 to <8 x i32>
|
|
%arg0 = bitcast <4 x i64> %a0 to <8 x i32>
|
|
%arg1 = bitcast i32 *%a1 to i8*
|
|
%arg1 = bitcast i32 *%a1 to i8*
|
|
%arg2 = bitcast <4 x i64> %a2 to <8 x i32>
|
|
%arg2 = bitcast <4 x i64> %a2 to <8 x i32>
|
|
@@ -858,7 +858,7 @@ define <2 x i64> @test_mm_i32gather_epi64(i64 *%a0, <2 x i64> %a1) {
|
|
; X86-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vpgatherdq %xmm2, (%eax,%xmm0,2), %xmm1
|
|
; X86-NEXT: vpgatherdq %xmm2, (%eax,%xmm0,2), %xmm1
|
|
; X86-NEXT: vmovdqa %xmm1, %xmm0
|
|
; X86-NEXT: vmovdqa %xmm1, %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_i32gather_epi64:
|
|
; X64-LABEL: test_mm_i32gather_epi64:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
@@ -866,7 +866,7 @@ define <2 x i64> @test_mm_i32gather_epi64(i64 *%a0, <2 x i64> %a1) {
|
|
; X64-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vpgatherdq %xmm2, (%rdi,%xmm0,2), %xmm1
|
|
; X64-NEXT: vpgatherdq %xmm2, (%rdi,%xmm0,2), %xmm1
|
|
; X64-NEXT: vmovdqa %xmm1, %xmm0
|
|
; X64-NEXT: vmovdqa %xmm1, %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast i64 *%a0 to i8*
|
|
%arg0 = bitcast i64 *%a0 to i8*
|
|
%arg1 = bitcast <2 x i64> %a1 to <4 x i32>
|
|
%arg1 = bitcast <2 x i64> %a1 to <4 x i32>
|
|
%res = call <2 x i64> @llvm.x86.avx2.gather.d.q(<2 x i64> undef, i8* %arg0, <4 x i32> %arg1, <2 x i64> <i64 -1, i64 -1>, i8 2)
|
|
%res = call <2 x i64> @llvm.x86.avx2.gather.d.q(<2 x i64> undef, i8* %arg0, <4 x i32> %arg1, <2 x i64> <i64 -1, i64 -1>, i8 2)
|
|
@@ -879,12 +879,12 @@ define <2 x i64> @test_mm_mask_i32gather_epi64(<2 x i64> %a0, i64 *%a1, <2 x i64
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vpgatherdq %xmm2, (%eax,%xmm1,2), %xmm0
|
|
; X86-NEXT: vpgatherdq %xmm2, (%eax,%xmm1,2), %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_mask_i32gather_epi64:
|
|
; X64-LABEL: test_mm_mask_i32gather_epi64:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vpgatherdq %xmm2, (%rdi,%xmm1,2), %xmm0
|
|
; X64-NEXT: vpgatherdq %xmm2, (%rdi,%xmm1,2), %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg1 = bitcast i64 *%a1 to i8*
|
|
%arg1 = bitcast i64 *%a1 to i8*
|
|
%arg2 = bitcast <2 x i64> %a2 to <4 x i32>
|
|
%arg2 = bitcast <2 x i64> %a2 to <4 x i32>
|
|
%res = call <2 x i64> @llvm.x86.avx2.gather.d.q(<2 x i64> %a0, i8* %arg1, <4 x i32> %arg2, <2 x i64> %a3, i8 2)
|
|
%res = call <2 x i64> @llvm.x86.avx2.gather.d.q(<2 x i64> %a0, i8* %arg1, <4 x i32> %arg2, <2 x i64> %a3, i8 2)
|
|
@@ -899,7 +899,7 @@ define <4 x i64> @test_mm256_i32gather_epi64(i64 *%a0, <2 x i64> %a1) {
|
|
; X86-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vpgatherdq %ymm2, (%eax,%xmm0,2), %ymm1
|
|
; X86-NEXT: vpgatherdq %ymm2, (%eax,%xmm0,2), %ymm1
|
|
; X86-NEXT: vmovdqa %ymm1, %ymm0
|
|
; X86-NEXT: vmovdqa %ymm1, %ymm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_i32gather_epi64:
|
|
; X64-LABEL: test_mm256_i32gather_epi64:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
@@ -907,7 +907,7 @@ define <4 x i64> @test_mm256_i32gather_epi64(i64 *%a0, <2 x i64> %a1) {
|
|
; X64-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vpgatherdq %ymm2, (%rdi,%xmm0,2), %ymm1
|
|
; X64-NEXT: vpgatherdq %ymm2, (%rdi,%xmm0,2), %ymm1
|
|
; X64-NEXT: vmovdqa %ymm1, %ymm0
|
|
; X64-NEXT: vmovdqa %ymm1, %ymm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast i64 *%a0 to i8*
|
|
%arg0 = bitcast i64 *%a0 to i8*
|
|
%arg1 = bitcast <2 x i64> %a1 to <4 x i32>
|
|
%arg1 = bitcast <2 x i64> %a1 to <4 x i32>
|
|
%res = call <4 x i64> @llvm.x86.avx2.gather.d.q.256(<4 x i64> undef, i8* %arg0, <4 x i32> %arg1, <4 x i64> <i64 -1, i64 -1, i64 -1, i64 -1>, i8 2)
|
|
%res = call <4 x i64> @llvm.x86.avx2.gather.d.q.256(<4 x i64> undef, i8* %arg0, <4 x i32> %arg1, <4 x i64> <i64 -1, i64 -1, i64 -1, i64 -1>, i8 2)
|
|
@@ -920,12 +920,12 @@ define <4 x i64> @test_mm256_mask_i32gather_epi64(<4 x i64> %a0, i64 *%a1, <2 x
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vpgatherdq %ymm2, (%eax,%xmm1,2), %ymm0
|
|
; X86-NEXT: vpgatherdq %ymm2, (%eax,%xmm1,2), %ymm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_mask_i32gather_epi64:
|
|
; X64-LABEL: test_mm256_mask_i32gather_epi64:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vpgatherdq %ymm2, (%rdi,%xmm1,2), %ymm0
|
|
; X64-NEXT: vpgatherdq %ymm2, (%rdi,%xmm1,2), %ymm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg1 = bitcast i64 *%a1 to i8*
|
|
%arg1 = bitcast i64 *%a1 to i8*
|
|
%arg2 = bitcast <2 x i64> %a2 to <4 x i32>
|
|
%arg2 = bitcast <2 x i64> %a2 to <4 x i32>
|
|
%res = call <4 x i64> @llvm.x86.avx2.gather.d.q.256(<4 x i64> %a0, i8* %arg1, <4 x i32> %arg2, <4 x i64> %a3, i8 2)
|
|
%res = call <4 x i64> @llvm.x86.avx2.gather.d.q.256(<4 x i64> %a0, i8* %arg1, <4 x i32> %arg2, <4 x i64> %a3, i8 2)
|
|
@@ -940,7 +940,7 @@ define <2 x double> @test_mm_i32gather_pd(double *%a0, <2 x i64> %a1) {
|
|
; X86-NEXT: vxorpd %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vxorpd %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vgatherdpd %xmm2, (%eax,%xmm0,2), %xmm1
|
|
; X86-NEXT: vgatherdpd %xmm2, (%eax,%xmm0,2), %xmm1
|
|
; X86-NEXT: vmovapd %xmm1, %xmm0
|
|
; X86-NEXT: vmovapd %xmm1, %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_i32gather_pd:
|
|
; X64-LABEL: test_mm_i32gather_pd:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
@@ -948,7 +948,7 @@ define <2 x double> @test_mm_i32gather_pd(double *%a0, <2 x i64> %a1) {
|
|
; X64-NEXT: vxorpd %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vxorpd %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vgatherdpd %xmm2, (%rdi,%xmm0,2), %xmm1
|
|
; X64-NEXT: vgatherdpd %xmm2, (%rdi,%xmm0,2), %xmm1
|
|
; X64-NEXT: vmovapd %xmm1, %xmm0
|
|
; X64-NEXT: vmovapd %xmm1, %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast double *%a0 to i8*
|
|
%arg0 = bitcast double *%a0 to i8*
|
|
%arg1 = bitcast <2 x i64> %a1 to <4 x i32>
|
|
%arg1 = bitcast <2 x i64> %a1 to <4 x i32>
|
|
%cmp = fcmp oeq <2 x double> zeroinitializer, zeroinitializer
|
|
%cmp = fcmp oeq <2 x double> zeroinitializer, zeroinitializer
|
|
@@ -964,12 +964,12 @@ define <2 x double> @test_mm_mask_i32gather_pd(<2 x double> %a0, double *%a1, <2
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vgatherdpd %xmm2, (%eax,%xmm1,2), %xmm0
|
|
; X86-NEXT: vgatherdpd %xmm2, (%eax,%xmm1,2), %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_mask_i32gather_pd:
|
|
; X64-LABEL: test_mm_mask_i32gather_pd:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vgatherdpd %xmm2, (%rdi,%xmm1,2), %xmm0
|
|
; X64-NEXT: vgatherdpd %xmm2, (%rdi,%xmm1,2), %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg1 = bitcast double *%a1 to i8*
|
|
%arg1 = bitcast double *%a1 to i8*
|
|
%arg2 = bitcast <2 x i64> %a2 to <4 x i32>
|
|
%arg2 = bitcast <2 x i64> %a2 to <4 x i32>
|
|
%res = call <2 x double> @llvm.x86.avx2.gather.d.pd(<2 x double> %a0, i8* %arg1, <4 x i32> %arg2, <2 x double> %a3, i8 2)
|
|
%res = call <2 x double> @llvm.x86.avx2.gather.d.pd(<2 x double> %a0, i8* %arg1, <4 x i32> %arg2, <2 x double> %a3, i8 2)
|
|
@@ -984,7 +984,7 @@ define <4 x double> @test_mm256_i32gather_pd(double *%a0, <2 x i64> %a1) {
|
|
; X86-NEXT: vcmpeqpd %ymm1, %ymm1, %ymm2
|
|
; X86-NEXT: vcmpeqpd %ymm1, %ymm1, %ymm2
|
|
; X86-NEXT: vgatherdpd %ymm2, (%eax,%xmm0,2), %ymm1
|
|
; X86-NEXT: vgatherdpd %ymm2, (%eax,%xmm0,2), %ymm1
|
|
; X86-NEXT: vmovapd %ymm1, %ymm0
|
|
; X86-NEXT: vmovapd %ymm1, %ymm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_i32gather_pd:
|
|
; X64-LABEL: test_mm256_i32gather_pd:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
@@ -992,7 +992,7 @@ define <4 x double> @test_mm256_i32gather_pd(double *%a0, <2 x i64> %a1) {
|
|
; X64-NEXT: vcmpeqpd %ymm1, %ymm1, %ymm2
|
|
; X64-NEXT: vcmpeqpd %ymm1, %ymm1, %ymm2
|
|
; X64-NEXT: vgatherdpd %ymm2, (%rdi,%xmm0,2), %ymm1
|
|
; X64-NEXT: vgatherdpd %ymm2, (%rdi,%xmm0,2), %ymm1
|
|
; X64-NEXT: vmovapd %ymm1, %ymm0
|
|
; X64-NEXT: vmovapd %ymm1, %ymm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast double *%a0 to i8*
|
|
%arg0 = bitcast double *%a0 to i8*
|
|
%arg1 = bitcast <2 x i64> %a1 to <4 x i32>
|
|
%arg1 = bitcast <2 x i64> %a1 to <4 x i32>
|
|
%mask = call <4 x double> @llvm.x86.avx.cmp.pd.256(<4 x double> zeroinitializer, <4 x double> zeroinitializer, i8 0)
|
|
%mask = call <4 x double> @llvm.x86.avx.cmp.pd.256(<4 x double> zeroinitializer, <4 x double> zeroinitializer, i8 0)
|
|
@@ -1006,12 +1006,12 @@ define <4 x double> @test_mm256_mask_i32gather_pd(<4 x double> %a0, double *%a1,
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vgatherdpd %ymm2, (%eax,%xmm1,2), %ymm0
|
|
; X86-NEXT: vgatherdpd %ymm2, (%eax,%xmm1,2), %ymm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_mask_i32gather_pd:
|
|
; X64-LABEL: test_mm256_mask_i32gather_pd:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vgatherdpd %ymm2, (%rdi,%xmm1,2), %ymm0
|
|
; X64-NEXT: vgatherdpd %ymm2, (%rdi,%xmm1,2), %ymm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg1 = bitcast double *%a1 to i8*
|
|
%arg1 = bitcast double *%a1 to i8*
|
|
%arg2 = bitcast <2 x i64> %a2 to <4 x i32>
|
|
%arg2 = bitcast <2 x i64> %a2 to <4 x i32>
|
|
%res = call <4 x double> @llvm.x86.avx2.gather.d.pd.256(<4 x double> %a0, i8* %arg1, <4 x i32> %arg2, <4 x double> %a3, i8 2)
|
|
%res = call <4 x double> @llvm.x86.avx2.gather.d.pd.256(<4 x double> %a0, i8* %arg1, <4 x i32> %arg2, <4 x double> %a3, i8 2)
|
|
@@ -1026,7 +1026,7 @@ define <4 x float> @test_mm_i32gather_ps(float *%a0, <2 x i64> %a1) {
|
|
; X86-NEXT: vxorps %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vxorps %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vgatherdps %xmm2, (%eax,%xmm0,2), %xmm1
|
|
; X86-NEXT: vgatherdps %xmm2, (%eax,%xmm0,2), %xmm1
|
|
; X86-NEXT: vmovaps %xmm1, %xmm0
|
|
; X86-NEXT: vmovaps %xmm1, %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_i32gather_ps:
|
|
; X64-LABEL: test_mm_i32gather_ps:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
@@ -1034,7 +1034,7 @@ define <4 x float> @test_mm_i32gather_ps(float *%a0, <2 x i64> %a1) {
|
|
; X64-NEXT: vxorps %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vxorps %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vgatherdps %xmm2, (%rdi,%xmm0,2), %xmm1
|
|
; X64-NEXT: vgatherdps %xmm2, (%rdi,%xmm0,2), %xmm1
|
|
; X64-NEXT: vmovaps %xmm1, %xmm0
|
|
; X64-NEXT: vmovaps %xmm1, %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast float *%a0 to i8*
|
|
%arg0 = bitcast float *%a0 to i8*
|
|
%arg1 = bitcast <2 x i64> %a1 to <4 x i32>
|
|
%arg1 = bitcast <2 x i64> %a1 to <4 x i32>
|
|
%cmp = fcmp oeq <4 x float> zeroinitializer, zeroinitializer
|
|
%cmp = fcmp oeq <4 x float> zeroinitializer, zeroinitializer
|
|
@@ -1050,12 +1050,12 @@ define <4 x float> @test_mm_mask_i32gather_ps(<4 x float> %a0, float *%a1, <2 x
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vgatherdps %xmm2, (%eax,%xmm1,2), %xmm0
|
|
; X86-NEXT: vgatherdps %xmm2, (%eax,%xmm1,2), %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_mask_i32gather_ps:
|
|
; X64-LABEL: test_mm_mask_i32gather_ps:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vgatherdps %xmm2, (%rdi,%xmm1,2), %xmm0
|
|
; X64-NEXT: vgatherdps %xmm2, (%rdi,%xmm1,2), %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg1 = bitcast float *%a1 to i8*
|
|
%arg1 = bitcast float *%a1 to i8*
|
|
%arg2 = bitcast <2 x i64> %a2 to <4 x i32>
|
|
%arg2 = bitcast <2 x i64> %a2 to <4 x i32>
|
|
%call = call <4 x float> @llvm.x86.avx2.gather.d.ps(<4 x float> %a0, i8* %arg1, <4 x i32> %arg2, <4 x float> %a3, i8 2)
|
|
%call = call <4 x float> @llvm.x86.avx2.gather.d.ps(<4 x float> %a0, i8* %arg1, <4 x i32> %arg2, <4 x float> %a3, i8 2)
|
|
@@ -1070,7 +1070,7 @@ define <8 x float> @test_mm256_i32gather_ps(float *%a0, <4 x i64> %a1) {
|
|
; X86-NEXT: vcmpeqps %ymm1, %ymm1, %ymm2
|
|
; X86-NEXT: vcmpeqps %ymm1, %ymm1, %ymm2
|
|
; X86-NEXT: vgatherdps %ymm2, (%eax,%ymm0,2), %ymm1
|
|
; X86-NEXT: vgatherdps %ymm2, (%eax,%ymm0,2), %ymm1
|
|
; X86-NEXT: vmovaps %ymm1, %ymm0
|
|
; X86-NEXT: vmovaps %ymm1, %ymm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_i32gather_ps:
|
|
; X64-LABEL: test_mm256_i32gather_ps:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
@@ -1078,7 +1078,7 @@ define <8 x float> @test_mm256_i32gather_ps(float *%a0, <4 x i64> %a1) {
|
|
; X64-NEXT: vcmpeqps %ymm1, %ymm1, %ymm2
|
|
; X64-NEXT: vcmpeqps %ymm1, %ymm1, %ymm2
|
|
; X64-NEXT: vgatherdps %ymm2, (%rdi,%ymm0,2), %ymm1
|
|
; X64-NEXT: vgatherdps %ymm2, (%rdi,%ymm0,2), %ymm1
|
|
; X64-NEXT: vmovaps %ymm1, %ymm0
|
|
; X64-NEXT: vmovaps %ymm1, %ymm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast float *%a0 to i8*
|
|
%arg0 = bitcast float *%a0 to i8*
|
|
%arg1 = bitcast <4 x i64> %a1 to <8 x i32>
|
|
%arg1 = bitcast <4 x i64> %a1 to <8 x i32>
|
|
%mask = call <8 x float> @llvm.x86.avx.cmp.ps.256(<8 x float> zeroinitializer, <8 x float> zeroinitializer, i8 0)
|
|
%mask = call <8 x float> @llvm.x86.avx.cmp.ps.256(<8 x float> zeroinitializer, <8 x float> zeroinitializer, i8 0)
|
|
@@ -1092,12 +1092,12 @@ define <8 x float> @test_mm256_mask_i32gather_ps(<8 x float> %a0, float *%a1, <4
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vgatherdps %ymm2, (%eax,%ymm1,2), %ymm0
|
|
; X86-NEXT: vgatherdps %ymm2, (%eax,%ymm1,2), %ymm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_mask_i32gather_ps:
|
|
; X64-LABEL: test_mm256_mask_i32gather_ps:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vgatherdps %ymm2, (%rdi,%ymm1,2), %ymm0
|
|
; X64-NEXT: vgatherdps %ymm2, (%rdi,%ymm1,2), %ymm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg1 = bitcast float *%a1 to i8*
|
|
%arg1 = bitcast float *%a1 to i8*
|
|
%arg2 = bitcast <4 x i64> %a2 to <8 x i32>
|
|
%arg2 = bitcast <4 x i64> %a2 to <8 x i32>
|
|
%call = call <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float> %a0, i8* %arg1, <8 x i32> %arg2, <8 x float> %a3, i8 2)
|
|
%call = call <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float> %a0, i8* %arg1, <8 x i32> %arg2, <8 x float> %a3, i8 2)
|
|
@@ -1112,7 +1112,7 @@ define <2 x i64> @test_mm_i64gather_epi32(i32 *%a0, <2 x i64> %a1) {
|
|
; X86-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vpgatherqd %xmm2, (%eax,%xmm0,2), %xmm1
|
|
; X86-NEXT: vpgatherqd %xmm2, (%eax,%xmm0,2), %xmm1
|
|
; X86-NEXT: vmovdqa %xmm1, %xmm0
|
|
; X86-NEXT: vmovdqa %xmm1, %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_i64gather_epi32:
|
|
; X64-LABEL: test_mm_i64gather_epi32:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
@@ -1120,7 +1120,7 @@ define <2 x i64> @test_mm_i64gather_epi32(i32 *%a0, <2 x i64> %a1) {
|
|
; X64-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vpgatherqd %xmm2, (%rdi,%xmm0,2), %xmm1
|
|
; X64-NEXT: vpgatherqd %xmm2, (%rdi,%xmm0,2), %xmm1
|
|
; X64-NEXT: vmovdqa %xmm1, %xmm0
|
|
; X64-NEXT: vmovdqa %xmm1, %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast i32 *%a0 to i8*
|
|
%arg0 = bitcast i32 *%a0 to i8*
|
|
%mask = bitcast <2 x i64> <i64 -1, i64 -1> to <4 x i32>
|
|
%mask = bitcast <2 x i64> <i64 -1, i64 -1> to <4 x i32>
|
|
%call = call <4 x i32> @llvm.x86.avx2.gather.q.d(<4 x i32> undef, i8* %arg0, <2 x i64> %a1, <4 x i32> %mask, i8 2)
|
|
%call = call <4 x i32> @llvm.x86.avx2.gather.q.d(<4 x i32> undef, i8* %arg0, <2 x i64> %a1, <4 x i32> %mask, i8 2)
|
|
@@ -1134,12 +1134,12 @@ define <2 x i64> @test_mm_mask_i64gather_epi32(<2 x i64> %a0, i32 *%a1, <2 x i64
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vpgatherqd %xmm2, (%eax,%xmm1,2), %xmm0
|
|
; X86-NEXT: vpgatherqd %xmm2, (%eax,%xmm1,2), %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_mask_i64gather_epi32:
|
|
; X64-LABEL: test_mm_mask_i64gather_epi32:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vpgatherqd %xmm2, (%rdi,%xmm1,2), %xmm0
|
|
; X64-NEXT: vpgatherqd %xmm2, (%rdi,%xmm1,2), %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast <2 x i64> %a0 to <4 x i32>
|
|
%arg0 = bitcast <2 x i64> %a0 to <4 x i32>
|
|
%arg1 = bitcast i32 *%a1 to i8*
|
|
%arg1 = bitcast i32 *%a1 to i8*
|
|
%arg3 = bitcast <2 x i64> %a3 to <4 x i32>
|
|
%arg3 = bitcast <2 x i64> %a3 to <4 x i32>
|
|
@@ -1157,7 +1157,7 @@ define <2 x i64> @test_mm256_i64gather_epi32(i32 *%a0, <4 x i64> %a1) {
|
|
; X86-NEXT: vpgatherqd %xmm2, (%eax,%ymm0,2), %xmm1
|
|
; X86-NEXT: vpgatherqd %xmm2, (%eax,%ymm0,2), %xmm1
|
|
; X86-NEXT: vmovdqa %xmm1, %xmm0
|
|
; X86-NEXT: vmovdqa %xmm1, %xmm0
|
|
; X86-NEXT: vzeroupper
|
|
; X86-NEXT: vzeroupper
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_i64gather_epi32:
|
|
; X64-LABEL: test_mm256_i64gather_epi32:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
@@ -1166,7 +1166,7 @@ define <2 x i64> @test_mm256_i64gather_epi32(i32 *%a0, <4 x i64> %a1) {
|
|
; X64-NEXT: vpgatherqd %xmm2, (%rdi,%ymm0,2), %xmm1
|
|
; X64-NEXT: vpgatherqd %xmm2, (%rdi,%ymm0,2), %xmm1
|
|
; X64-NEXT: vmovdqa %xmm1, %xmm0
|
|
; X64-NEXT: vmovdqa %xmm1, %xmm0
|
|
; X64-NEXT: vzeroupper
|
|
; X64-NEXT: vzeroupper
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast i32 *%a0 to i8*
|
|
%arg0 = bitcast i32 *%a0 to i8*
|
|
%mask = bitcast <2 x i64> <i64 -1, i64 -1> to <4 x i32>
|
|
%mask = bitcast <2 x i64> <i64 -1, i64 -1> to <4 x i32>
|
|
%call = call <4 x i32> @llvm.x86.avx2.gather.q.d.256(<4 x i32> undef, i8* %arg0, <4 x i64> %a1, <4 x i32> %mask, i8 2)
|
|
%call = call <4 x i32> @llvm.x86.avx2.gather.q.d.256(<4 x i32> undef, i8* %arg0, <4 x i64> %a1, <4 x i32> %mask, i8 2)
|
|
@@ -1181,13 +1181,13 @@ define <2 x i64> @test_mm256_mask_i64gather_epi32(<2 x i64> %a0, i32 *%a1, <4 x
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vpgatherqd %xmm2, (%eax,%ymm1,2), %xmm0
|
|
; X86-NEXT: vpgatherqd %xmm2, (%eax,%ymm1,2), %xmm0
|
|
; X86-NEXT: vzeroupper
|
|
; X86-NEXT: vzeroupper
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_mask_i64gather_epi32:
|
|
; X64-LABEL: test_mm256_mask_i64gather_epi32:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vpgatherqd %xmm2, (%rdi,%ymm1,2), %xmm0
|
|
; X64-NEXT: vpgatherqd %xmm2, (%rdi,%ymm1,2), %xmm0
|
|
; X64-NEXT: vzeroupper
|
|
; X64-NEXT: vzeroupper
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast <2 x i64> %a0 to <4 x i32>
|
|
%arg0 = bitcast <2 x i64> %a0 to <4 x i32>
|
|
%arg1 = bitcast i32 *%a1 to i8*
|
|
%arg1 = bitcast i32 *%a1 to i8*
|
|
%arg3 = bitcast <2 x i64> %a3 to <4 x i32>
|
|
%arg3 = bitcast <2 x i64> %a3 to <4 x i32>
|
|
@@ -1204,7 +1204,7 @@ define <2 x i64> @test_mm_i64gather_epi64(i64 *%a0, <2 x i64> %a1) {
|
|
; X86-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vpgatherqq %xmm2, (%eax,%xmm0,2), %xmm1
|
|
; X86-NEXT: vpgatherqq %xmm2, (%eax,%xmm0,2), %xmm1
|
|
; X86-NEXT: vmovdqa %xmm1, %xmm0
|
|
; X86-NEXT: vmovdqa %xmm1, %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_i64gather_epi64:
|
|
; X64-LABEL: test_mm_i64gather_epi64:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
@@ -1212,7 +1212,7 @@ define <2 x i64> @test_mm_i64gather_epi64(i64 *%a0, <2 x i64> %a1) {
|
|
; X64-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vpgatherqq %xmm2, (%rdi,%xmm0,2), %xmm1
|
|
; X64-NEXT: vpgatherqq %xmm2, (%rdi,%xmm0,2), %xmm1
|
|
; X64-NEXT: vmovdqa %xmm1, %xmm0
|
|
; X64-NEXT: vmovdqa %xmm1, %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast i64 *%a0 to i8*
|
|
%arg0 = bitcast i64 *%a0 to i8*
|
|
%call = call <2 x i64> @llvm.x86.avx2.gather.q.q(<2 x i64> undef, i8* %arg0, <2 x i64> %a1, <2 x i64> <i64 -1, i64 -1>, i8 2)
|
|
%call = call <2 x i64> @llvm.x86.avx2.gather.q.q(<2 x i64> undef, i8* %arg0, <2 x i64> %a1, <2 x i64> <i64 -1, i64 -1>, i8 2)
|
|
ret <2 x i64> %call
|
|
ret <2 x i64> %call
|
|
@@ -1224,12 +1224,12 @@ define <2 x i64> @test_mm_mask_i64gather_epi64(<2 x i64> %a0, i64 *%a1, <2 x i64
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vpgatherqq %xmm2, (%eax,%xmm1,2), %xmm0
|
|
; X86-NEXT: vpgatherqq %xmm2, (%eax,%xmm1,2), %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_mask_i64gather_epi64:
|
|
; X64-LABEL: test_mm_mask_i64gather_epi64:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vpgatherqq %xmm2, (%rdi,%xmm1,2), %xmm0
|
|
; X64-NEXT: vpgatherqq %xmm2, (%rdi,%xmm1,2), %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg1 = bitcast i64 *%a1 to i8*
|
|
%arg1 = bitcast i64 *%a1 to i8*
|
|
%call = call <2 x i64> @llvm.x86.avx2.gather.q.q(<2 x i64> %a0, i8* %arg1, <2 x i64> %a2, <2 x i64> %a3, i8 2)
|
|
%call = call <2 x i64> @llvm.x86.avx2.gather.q.q(<2 x i64> %a0, i8* %arg1, <2 x i64> %a2, <2 x i64> %a3, i8 2)
|
|
ret <2 x i64> %call
|
|
ret <2 x i64> %call
|
|
@@ -1243,7 +1243,7 @@ define <4 x i64> @test_mm256_i64gather_epi64(i64 *%a0, <4 x i64> %a1) {
|
|
; X86-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vpgatherqq %ymm2, (%eax,%ymm0,2), %ymm1
|
|
; X86-NEXT: vpgatherqq %ymm2, (%eax,%ymm0,2), %ymm1
|
|
; X86-NEXT: vmovdqa %ymm1, %ymm0
|
|
; X86-NEXT: vmovdqa %ymm1, %ymm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_i64gather_epi64:
|
|
; X64-LABEL: test_mm256_i64gather_epi64:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
@@ -1251,7 +1251,7 @@ define <4 x i64> @test_mm256_i64gather_epi64(i64 *%a0, <4 x i64> %a1) {
|
|
; X64-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vpxor %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vpgatherqq %ymm2, (%rdi,%ymm0,2), %ymm1
|
|
; X64-NEXT: vpgatherqq %ymm2, (%rdi,%ymm0,2), %ymm1
|
|
; X64-NEXT: vmovdqa %ymm1, %ymm0
|
|
; X64-NEXT: vmovdqa %ymm1, %ymm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast i64 *%a0 to i8*
|
|
%arg0 = bitcast i64 *%a0 to i8*
|
|
%call = call <4 x i64> @llvm.x86.avx2.gather.q.q.256(<4 x i64> undef, i8* %arg0, <4 x i64> %a1, <4 x i64> <i64 -1, i64 -1, i64 -1, i64 -1>, i8 2)
|
|
%call = call <4 x i64> @llvm.x86.avx2.gather.q.q.256(<4 x i64> undef, i8* %arg0, <4 x i64> %a1, <4 x i64> <i64 -1, i64 -1, i64 -1, i64 -1>, i8 2)
|
|
ret <4 x i64> %call
|
|
ret <4 x i64> %call
|
|
@@ -1263,12 +1263,12 @@ define <4 x i64> @test_mm256_mask_i64gather_epi64(<4 x i64> %a0, i64 *%a1, <4 x
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vpgatherqq %ymm2, (%eax,%ymm1,2), %ymm0
|
|
; X86-NEXT: vpgatherqq %ymm2, (%eax,%ymm1,2), %ymm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_mask_i64gather_epi64:
|
|
; X64-LABEL: test_mm256_mask_i64gather_epi64:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vpgatherqq %ymm2, (%rdi,%ymm1,2), %ymm0
|
|
; X64-NEXT: vpgatherqq %ymm2, (%rdi,%ymm1,2), %ymm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg1 = bitcast i64 *%a1 to i8*
|
|
%arg1 = bitcast i64 *%a1 to i8*
|
|
%call = call <4 x i64> @llvm.x86.avx2.gather.q.q.256(<4 x i64> %a0, i8* %arg1, <4 x i64> %a2, <4 x i64> %a3, i8 2)
|
|
%call = call <4 x i64> @llvm.x86.avx2.gather.q.q.256(<4 x i64> %a0, i8* %arg1, <4 x i64> %a2, <4 x i64> %a3, i8 2)
|
|
ret <4 x i64> %call
|
|
ret <4 x i64> %call
|
|
@@ -1282,7 +1282,7 @@ define <2 x double> @test_mm_i64gather_pd(double *%a0, <2 x i64> %a1) {
|
|
; X86-NEXT: vxorpd %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vxorpd %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vgatherqpd %xmm2, (%eax,%xmm0,2), %xmm1
|
|
; X86-NEXT: vgatherqpd %xmm2, (%eax,%xmm0,2), %xmm1
|
|
; X86-NEXT: vmovapd %xmm1, %xmm0
|
|
; X86-NEXT: vmovapd %xmm1, %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_i64gather_pd:
|
|
; X64-LABEL: test_mm_i64gather_pd:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
@@ -1290,7 +1290,7 @@ define <2 x double> @test_mm_i64gather_pd(double *%a0, <2 x i64> %a1) {
|
|
; X64-NEXT: vxorpd %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vxorpd %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vgatherqpd %xmm2, (%rdi,%xmm0,2), %xmm1
|
|
; X64-NEXT: vgatherqpd %xmm2, (%rdi,%xmm0,2), %xmm1
|
|
; X64-NEXT: vmovapd %xmm1, %xmm0
|
|
; X64-NEXT: vmovapd %xmm1, %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast double *%a0 to i8*
|
|
%arg0 = bitcast double *%a0 to i8*
|
|
%cmp = fcmp oeq <2 x double> zeroinitializer, zeroinitializer
|
|
%cmp = fcmp oeq <2 x double> zeroinitializer, zeroinitializer
|
|
%sext = sext <2 x i1> %cmp to <2 x i64>
|
|
%sext = sext <2 x i1> %cmp to <2 x i64>
|
|
@@ -1305,12 +1305,12 @@ define <2 x double> @test_mm_mask_i64gather_pd(<2 x double> %a0, double *%a1, <2
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vgatherqpd %xmm2, (%eax,%xmm1,2), %xmm0
|
|
; X86-NEXT: vgatherqpd %xmm2, (%eax,%xmm1,2), %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_mask_i64gather_pd:
|
|
; X64-LABEL: test_mm_mask_i64gather_pd:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vgatherqpd %xmm2, (%rdi,%xmm1,2), %xmm0
|
|
; X64-NEXT: vgatherqpd %xmm2, (%rdi,%xmm1,2), %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg1 = bitcast double *%a1 to i8*
|
|
%arg1 = bitcast double *%a1 to i8*
|
|
%call = call <2 x double> @llvm.x86.avx2.gather.q.pd(<2 x double> %a0, i8* %arg1, <2 x i64> %a2, <2 x double> %a3, i8 2)
|
|
%call = call <2 x double> @llvm.x86.avx2.gather.q.pd(<2 x double> %a0, i8* %arg1, <2 x i64> %a2, <2 x double> %a3, i8 2)
|
|
ret <2 x double> %call
|
|
ret <2 x double> %call
|
|
@@ -1324,7 +1324,7 @@ define <4 x double> @test_mm256_i64gather_pd(double *%a0, <4 x i64> %a1) {
|
|
; X86-NEXT: vcmpeqpd %ymm1, %ymm1, %ymm2
|
|
; X86-NEXT: vcmpeqpd %ymm1, %ymm1, %ymm2
|
|
; X86-NEXT: vgatherqpd %ymm2, (%eax,%ymm0,2), %ymm1
|
|
; X86-NEXT: vgatherqpd %ymm2, (%eax,%ymm0,2), %ymm1
|
|
; X86-NEXT: vmovapd %ymm1, %ymm0
|
|
; X86-NEXT: vmovapd %ymm1, %ymm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_i64gather_pd:
|
|
; X64-LABEL: test_mm256_i64gather_pd:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
@@ -1332,7 +1332,7 @@ define <4 x double> @test_mm256_i64gather_pd(double *%a0, <4 x i64> %a1) {
|
|
; X64-NEXT: vcmpeqpd %ymm1, %ymm1, %ymm2
|
|
; X64-NEXT: vcmpeqpd %ymm1, %ymm1, %ymm2
|
|
; X64-NEXT: vgatherqpd %ymm2, (%rdi,%ymm0,2), %ymm1
|
|
; X64-NEXT: vgatherqpd %ymm2, (%rdi,%ymm0,2), %ymm1
|
|
; X64-NEXT: vmovapd %ymm1, %ymm0
|
|
; X64-NEXT: vmovapd %ymm1, %ymm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast double *%a0 to i8*
|
|
%arg0 = bitcast double *%a0 to i8*
|
|
%mask = call <4 x double> @llvm.x86.avx.cmp.pd.256(<4 x double> zeroinitializer, <4 x double> zeroinitializer, i8 0)
|
|
%mask = call <4 x double> @llvm.x86.avx.cmp.pd.256(<4 x double> zeroinitializer, <4 x double> zeroinitializer, i8 0)
|
|
%call = call <4 x double> @llvm.x86.avx2.gather.q.pd.256(<4 x double> undef, i8* %arg0, <4 x i64> %a1, <4 x double> %mask, i8 2)
|
|
%call = call <4 x double> @llvm.x86.avx2.gather.q.pd.256(<4 x double> undef, i8* %arg0, <4 x i64> %a1, <4 x double> %mask, i8 2)
|
|
@@ -1345,12 +1345,12 @@ define <4 x double> @test_mm256_mask_i64gather_pd(<4 x double> %a0, i64 *%a1, <4
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vgatherqpd %ymm2, (%eax,%ymm1,2), %ymm0
|
|
; X86-NEXT: vgatherqpd %ymm2, (%eax,%ymm1,2), %ymm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_mask_i64gather_pd:
|
|
; X64-LABEL: test_mm256_mask_i64gather_pd:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vgatherqpd %ymm2, (%rdi,%ymm1,2), %ymm0
|
|
; X64-NEXT: vgatherqpd %ymm2, (%rdi,%ymm1,2), %ymm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg1 = bitcast i64 *%a1 to i8*
|
|
%arg1 = bitcast i64 *%a1 to i8*
|
|
%call = call <4 x double> @llvm.x86.avx2.gather.q.pd.256(<4 x double> %a0, i8* %arg1, <4 x i64> %a2, <4 x double> %a3, i8 2)
|
|
%call = call <4 x double> @llvm.x86.avx2.gather.q.pd.256(<4 x double> %a0, i8* %arg1, <4 x i64> %a2, <4 x double> %a3, i8 2)
|
|
ret <4 x double> %call
|
|
ret <4 x double> %call
|
|
@@ -1364,7 +1364,7 @@ define <4 x float> @test_mm_i64gather_ps(float *%a0, <2 x i64> %a1) {
|
|
; X86-NEXT: vxorps %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vxorps %xmm1, %xmm1, %xmm1
|
|
; X86-NEXT: vgatherqps %xmm2, (%eax,%xmm0,2), %xmm1
|
|
; X86-NEXT: vgatherqps %xmm2, (%eax,%xmm0,2), %xmm1
|
|
; X86-NEXT: vmovaps %xmm1, %xmm0
|
|
; X86-NEXT: vmovaps %xmm1, %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_i64gather_ps:
|
|
; X64-LABEL: test_mm_i64gather_ps:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
@@ -1372,7 +1372,7 @@ define <4 x float> @test_mm_i64gather_ps(float *%a0, <2 x i64> %a1) {
|
|
; X64-NEXT: vxorps %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vxorps %xmm1, %xmm1, %xmm1
|
|
; X64-NEXT: vgatherqps %xmm2, (%rdi,%xmm0,2), %xmm1
|
|
; X64-NEXT: vgatherqps %xmm2, (%rdi,%xmm0,2), %xmm1
|
|
; X64-NEXT: vmovaps %xmm1, %xmm0
|
|
; X64-NEXT: vmovaps %xmm1, %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast float *%a0 to i8*
|
|
%arg0 = bitcast float *%a0 to i8*
|
|
%cmp = fcmp oeq <4 x float> zeroinitializer, zeroinitializer
|
|
%cmp = fcmp oeq <4 x float> zeroinitializer, zeroinitializer
|
|
%sext = sext <4 x i1> %cmp to <4 x i32>
|
|
%sext = sext <4 x i1> %cmp to <4 x i32>
|
|
@@ -1387,12 +1387,12 @@ define <4 x float> @test_mm_mask_i64gather_ps(<4 x float> %a0, float *%a1, <2 x
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vgatherqps %xmm2, (%eax,%xmm1,2), %xmm0
|
|
; X86-NEXT: vgatherqps %xmm2, (%eax,%xmm1,2), %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_mask_i64gather_ps:
|
|
; X64-LABEL: test_mm_mask_i64gather_ps:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vgatherqps %xmm2, (%rdi,%xmm1,2), %xmm0
|
|
; X64-NEXT: vgatherqps %xmm2, (%rdi,%xmm1,2), %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg1 = bitcast float *%a1 to i8*
|
|
%arg1 = bitcast float *%a1 to i8*
|
|
%call = call <4 x float> @llvm.x86.avx2.gather.q.ps(<4 x float> %a0, i8* %arg1, <2 x i64> %a2, <4 x float> %a3, i8 2)
|
|
%call = call <4 x float> @llvm.x86.avx2.gather.q.ps(<4 x float> %a0, i8* %arg1, <2 x i64> %a2, <4 x float> %a3, i8 2)
|
|
ret <4 x float> %call
|
|
ret <4 x float> %call
|
|
@@ -1407,7 +1407,7 @@ define <4 x float> @test_mm256_i64gather_ps(float *%a0, <4 x i64> %a1) {
|
|
; X86-NEXT: vgatherqps %xmm2, (%eax,%ymm0,2), %xmm1
|
|
; X86-NEXT: vgatherqps %xmm2, (%eax,%ymm0,2), %xmm1
|
|
; X86-NEXT: vmovaps %xmm1, %xmm0
|
|
; X86-NEXT: vmovaps %xmm1, %xmm0
|
|
; X86-NEXT: vzeroupper
|
|
; X86-NEXT: vzeroupper
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_i64gather_ps:
|
|
; X64-LABEL: test_mm256_i64gather_ps:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
@@ -1416,7 +1416,7 @@ define <4 x float> @test_mm256_i64gather_ps(float *%a0, <4 x i64> %a1) {
|
|
; X64-NEXT: vgatherqps %xmm2, (%rdi,%ymm0,2), %xmm1
|
|
; X64-NEXT: vgatherqps %xmm2, (%rdi,%ymm0,2), %xmm1
|
|
; X64-NEXT: vmovaps %xmm1, %xmm0
|
|
; X64-NEXT: vmovaps %xmm1, %xmm0
|
|
; X64-NEXT: vzeroupper
|
|
; X64-NEXT: vzeroupper
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast float *%a0 to i8*
|
|
%arg0 = bitcast float *%a0 to i8*
|
|
%cmp = fcmp oeq <4 x float> zeroinitializer, zeroinitializer
|
|
%cmp = fcmp oeq <4 x float> zeroinitializer, zeroinitializer
|
|
%sext = sext <4 x i1> %cmp to <4 x i32>
|
|
%sext = sext <4 x i1> %cmp to <4 x i32>
|
|
@@ -1432,13 +1432,13 @@ define <4 x float> @test_mm256_mask_i64gather_ps(<4 x float> %a0, float *%a1, <4
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vgatherqps %xmm2, (%eax,%ymm1,2), %xmm0
|
|
; X86-NEXT: vgatherqps %xmm2, (%eax,%ymm1,2), %xmm0
|
|
; X86-NEXT: vzeroupper
|
|
; X86-NEXT: vzeroupper
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_mask_i64gather_ps:
|
|
; X64-LABEL: test_mm256_mask_i64gather_ps:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vgatherqps %xmm2, (%rdi,%ymm1,2), %xmm0
|
|
; X64-NEXT: vgatherqps %xmm2, (%rdi,%ymm1,2), %xmm0
|
|
; X64-NEXT: vzeroupper
|
|
; X64-NEXT: vzeroupper
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg1 = bitcast float *%a1 to i8*
|
|
%arg1 = bitcast float *%a1 to i8*
|
|
%call = call <4 x float> @llvm.x86.avx2.gather.q.ps.256(<4 x float> %a0, i8* %arg1, <4 x i64> %a2, <4 x float> %a3, i8 2)
|
|
%call = call <4 x float> @llvm.x86.avx2.gather.q.ps.256(<4 x float> %a0, i8* %arg1, <4 x i64> %a2, <4 x float> %a3, i8 2)
|
|
ret <4 x float> %call
|
|
ret <4 x float> %call
|
|
@@ -1496,12 +1496,12 @@ define <2 x i64> @test_mm_maskload_epi32(i32* %a0, <2 x i64> %a1) nounwind {
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vpmaskmovd (%eax), %xmm0, %xmm0
|
|
; X86-NEXT: vpmaskmovd (%eax), %xmm0, %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_maskload_epi32:
|
|
; X64-LABEL: test_mm_maskload_epi32:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vpmaskmovd (%rdi), %xmm0, %xmm0
|
|
; X64-NEXT: vpmaskmovd (%rdi), %xmm0, %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast i32* %a0 to i8*
|
|
%arg0 = bitcast i32* %a0 to i8*
|
|
%arg1 = bitcast <2 x i64> %a1 to <4 x i32>
|
|
%arg1 = bitcast <2 x i64> %a1 to <4 x i32>
|
|
%call = call <4 x i32> @llvm.x86.avx2.maskload.d(i8* %arg0, <4 x i32> %arg1)
|
|
%call = call <4 x i32> @llvm.x86.avx2.maskload.d(i8* %arg0, <4 x i32> %arg1)
|
|
@@ -1515,12 +1515,12 @@ define <4 x i64> @test_mm256_maskload_epi32(i32* %a0, <4 x i64> %a1) nounwind {
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vpmaskmovd (%eax), %ymm0, %ymm0
|
|
; X86-NEXT: vpmaskmovd (%eax), %ymm0, %ymm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_maskload_epi32:
|
|
; X64-LABEL: test_mm256_maskload_epi32:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vpmaskmovd (%rdi), %ymm0, %ymm0
|
|
; X64-NEXT: vpmaskmovd (%rdi), %ymm0, %ymm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast i32* %a0 to i8*
|
|
%arg0 = bitcast i32* %a0 to i8*
|
|
%arg1 = bitcast <4 x i64> %a1 to <8 x i32>
|
|
%arg1 = bitcast <4 x i64> %a1 to <8 x i32>
|
|
%call = call <8 x i32> @llvm.x86.avx2.maskload.d.256(i8* %arg0, <8 x i32> %arg1)
|
|
%call = call <8 x i32> @llvm.x86.avx2.maskload.d.256(i8* %arg0, <8 x i32> %arg1)
|
|
@@ -1534,12 +1534,12 @@ define <2 x i64> @test_mm_maskload_epi64(i64* %a0, <2 x i64> %a1) nounwind {
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vpmaskmovq (%eax), %xmm0, %xmm0
|
|
; X86-NEXT: vpmaskmovq (%eax), %xmm0, %xmm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_maskload_epi64:
|
|
; X64-LABEL: test_mm_maskload_epi64:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vpmaskmovq (%rdi), %xmm0, %xmm0
|
|
; X64-NEXT: vpmaskmovq (%rdi), %xmm0, %xmm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast i64* %a0 to i8*
|
|
%arg0 = bitcast i64* %a0 to i8*
|
|
%res = call <2 x i64> @llvm.x86.avx2.maskload.q(i8* %arg0, <2 x i64> %a1)
|
|
%res = call <2 x i64> @llvm.x86.avx2.maskload.q(i8* %arg0, <2 x i64> %a1)
|
|
ret <2 x i64> %res
|
|
ret <2 x i64> %res
|
|
@@ -1551,12 +1551,12 @@ define <4 x i64> @test_mm256_maskload_epi64(i64* %a0, <4 x i64> %a1) nounwind {
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vpmaskmovq (%eax), %ymm0, %ymm0
|
|
; X86-NEXT: vpmaskmovq (%eax), %ymm0, %ymm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_maskload_epi64:
|
|
; X64-LABEL: test_mm256_maskload_epi64:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vpmaskmovq (%rdi), %ymm0, %ymm0
|
|
; X64-NEXT: vpmaskmovq (%rdi), %ymm0, %ymm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast i64* %a0 to i8*
|
|
%arg0 = bitcast i64* %a0 to i8*
|
|
%res = call <4 x i64> @llvm.x86.avx2.maskload.q.256(i8* %arg0, <4 x i64> %a1)
|
|
%res = call <4 x i64> @llvm.x86.avx2.maskload.q.256(i8* %arg0, <4 x i64> %a1)
|
|
ret <4 x i64> %res
|
|
ret <4 x i64> %res
|
|
@@ -1568,12 +1568,12 @@ define void @test_mm_maskstore_epi32(float* %a0, <2 x i64> %a1, <2 x i64> %a2) n
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vpmaskmovd %xmm1, %xmm0, (%eax)
|
|
; X86-NEXT: vpmaskmovd %xmm1, %xmm0, (%eax)
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_maskstore_epi32:
|
|
; X64-LABEL: test_mm_maskstore_epi32:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vpmaskmovd %xmm1, %xmm0, (%rdi)
|
|
; X64-NEXT: vpmaskmovd %xmm1, %xmm0, (%rdi)
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast float* %a0 to i8*
|
|
%arg0 = bitcast float* %a0 to i8*
|
|
%arg1 = bitcast <2 x i64> %a1 to <4 x i32>
|
|
%arg1 = bitcast <2 x i64> %a1 to <4 x i32>
|
|
%arg2 = bitcast <2 x i64> %a2 to <4 x i32>
|
|
%arg2 = bitcast <2 x i64> %a2 to <4 x i32>
|
|
@@ -1588,13 +1588,13 @@ define void @test_mm256_maskstore_epi32(float* %a0, <4 x i64> %a1, <4 x i64> %a2
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vpmaskmovd %ymm1, %ymm0, (%eax)
|
|
; X86-NEXT: vpmaskmovd %ymm1, %ymm0, (%eax)
|
|
; X86-NEXT: vzeroupper
|
|
; X86-NEXT: vzeroupper
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_maskstore_epi32:
|
|
; X64-LABEL: test_mm256_maskstore_epi32:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vpmaskmovd %ymm1, %ymm0, (%rdi)
|
|
; X64-NEXT: vpmaskmovd %ymm1, %ymm0, (%rdi)
|
|
; X64-NEXT: vzeroupper
|
|
; X64-NEXT: vzeroupper
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast float* %a0 to i8*
|
|
%arg0 = bitcast float* %a0 to i8*
|
|
%arg1 = bitcast <4 x i64> %a1 to <8 x i32>
|
|
%arg1 = bitcast <4 x i64> %a1 to <8 x i32>
|
|
%arg2 = bitcast <4 x i64> %a2 to <8 x i32>
|
|
%arg2 = bitcast <4 x i64> %a2 to <8 x i32>
|
|
@@ -1608,12 +1608,12 @@ define void @test_mm_maskstore_epi64(i64* %a0, <2 x i64> %a1, <2 x i64> %a2) nou
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vpmaskmovq %xmm1, %xmm0, (%eax)
|
|
; X86-NEXT: vpmaskmovq %xmm1, %xmm0, (%eax)
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm_maskstore_epi64:
|
|
; X64-LABEL: test_mm_maskstore_epi64:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vpmaskmovq %xmm1, %xmm0, (%rdi)
|
|
; X64-NEXT: vpmaskmovq %xmm1, %xmm0, (%rdi)
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast i64* %a0 to i8*
|
|
%arg0 = bitcast i64* %a0 to i8*
|
|
call void @llvm.x86.avx2.maskstore.q(i8* %arg0, <2 x i64> %a1, <2 x i64> %a2)
|
|
call void @llvm.x86.avx2.maskstore.q(i8* %arg0, <2 x i64> %a1, <2 x i64> %a2)
|
|
ret void
|
|
ret void
|
|
@@ -1626,13 +1626,13 @@ define void @test_mm256_maskstore_epi64(i64* %a0, <4 x i64> %a1, <4 x i64> %a2)
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vpmaskmovq %ymm1, %ymm0, (%eax)
|
|
; X86-NEXT: vpmaskmovq %ymm1, %ymm0, (%eax)
|
|
; X86-NEXT: vzeroupper
|
|
; X86-NEXT: vzeroupper
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_maskstore_epi64:
|
|
; X64-LABEL: test_mm256_maskstore_epi64:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vpmaskmovq %ymm1, %ymm0, (%rdi)
|
|
; X64-NEXT: vpmaskmovq %ymm1, %ymm0, (%rdi)
|
|
; X64-NEXT: vzeroupper
|
|
; X64-NEXT: vzeroupper
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast i64* %a0 to i8*
|
|
%arg0 = bitcast i64* %a0 to i8*
|
|
call void @llvm.x86.avx2.maskstore.q.256(i8* %arg0, <4 x i64> %a1, <4 x i64> %a2)
|
|
call void @llvm.x86.avx2.maskstore.q.256(i8* %arg0, <4 x i64> %a1, <4 x i64> %a2)
|
|
ret void
|
|
ret void
|
|
@@ -2465,12 +2465,12 @@ define <4 x i64> @test_mm256_stream_load_si256(<4 x i64> *%a0) {
|
|
; X86: # %bb.0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: vmovntdqa (%eax), %ymm0
|
|
; X86-NEXT: vmovntdqa (%eax), %ymm0
|
|
-; X86-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X86-NEXT: retl
|
|
;
|
|
;
|
|
; X64-LABEL: test_mm256_stream_load_si256:
|
|
; X64-LABEL: test_mm256_stream_load_si256:
|
|
; X64: # %bb.0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: vmovntdqa (%rdi), %ymm0
|
|
; X64-NEXT: vmovntdqa (%rdi), %ymm0
|
|
-; X64-NEXT: ret{{[l|q]}}
|
|
|
|
|
|
+; X64-NEXT: retq
|
|
%arg0 = bitcast <4 x i64> *%a0 to i8*
|
|
%arg0 = bitcast <4 x i64> *%a0 to i8*
|
|
%res = call <4 x i64> @llvm.x86.avx2.movntdqa(i8* %arg0)
|
|
%res = call <4 x i64> @llvm.x86.avx2.movntdqa(i8* %arg0)
|
|
ret <4 x i64> %res
|
|
ret <4 x i64> %res
|