| ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py |
| ; RUN: llc < %s -fast-isel -mtriple=i386-unknown-unknown -mattr=+sse4.2 | FileCheck %s --check-prefixes=CHECK,X86,SSE,X86-SSE |
| ; RUN: llc < %s -fast-isel -mtriple=i386-unknown-unknown -mattr=+avx | FileCheck %s --check-prefixes=CHECK,X86,AVX,X86-AVX,AVX1,X86-AVX1 |
| ; RUN: llc < %s -fast-isel -mtriple=i386-unknown-unknown -mattr=+avx512f,+avx512bw,+avx512dq,+avx512vl | FileCheck %s --check-prefixes=CHECK,X86,AVX,X86-AVX,AVX512,X86-AVX512 |
| ; RUN: llc < %s -fast-isel -mtriple=x86_64-unknown-unknown -mattr=+sse4.2 | FileCheck %s --check-prefixes=CHECK,X64,SSE,X64-SSE |
| ; RUN: llc < %s -fast-isel -mtriple=x86_64-unknown-unknown -mattr=+avx | FileCheck %s --check-prefixes=CHECK,X64,AVX,X64-AVX,AVX1,X64-AVX1 |
| ; RUN: llc < %s -fast-isel -mtriple=x86_64-unknown-unknown -mattr=+avx512f,+avx512bw,+avx512dq,+avx512vl | FileCheck %s --check-prefixes=CHECK,X64,AVX,X64-AVX,AVX512,X64-AVX512 |
| |
| ; NOTE: This should use IR equivalent to what is generated by clang/test/CodeGen/sse42-builtins.c |
| |
| define i32 @test_mm_cmpestra(<2 x i64> %a0, i32 %a1, <2 x i64> %a2, i32 %a3) nounwind { |
| ; X86-SSE-LABEL: test_mm_cmpestra: |
| ; X86-SSE: # %bb.0: |
| ; X86-SSE-NEXT: pushl %ebx |
| ; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-SSE-NEXT: xorl %ebx, %ebx |
| ; X86-SSE-NEXT: pcmpestri $7, %xmm1, %xmm0 |
| ; X86-SSE-NEXT: seta %bl |
| ; X86-SSE-NEXT: movl %ebx, %eax |
| ; X86-SSE-NEXT: popl %ebx |
| ; X86-SSE-NEXT: retl |
| ; |
| ; X86-AVX-LABEL: test_mm_cmpestra: |
| ; X86-AVX: # %bb.0: |
| ; X86-AVX-NEXT: pushl %ebx |
| ; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-AVX-NEXT: xorl %ebx, %ebx |
| ; X86-AVX-NEXT: vpcmpestri $7, %xmm1, %xmm0 |
| ; X86-AVX-NEXT: seta %bl |
| ; X86-AVX-NEXT: movl %ebx, %eax |
| ; X86-AVX-NEXT: popl %ebx |
| ; X86-AVX-NEXT: retl |
| ; |
| ; X64-SSE-LABEL: test_mm_cmpestra: |
| ; X64-SSE: # %bb.0: |
| ; X64-SSE-NEXT: xorl %r8d, %r8d |
| ; X64-SSE-NEXT: movl %edi, %eax |
| ; X64-SSE-NEXT: movl %esi, %edx |
| ; X64-SSE-NEXT: pcmpestri $7, %xmm1, %xmm0 |
| ; X64-SSE-NEXT: seta %r8b |
| ; X64-SSE-NEXT: movl %r8d, %eax |
| ; X64-SSE-NEXT: retq |
| ; |
| ; X64-AVX-LABEL: test_mm_cmpestra: |
| ; X64-AVX: # %bb.0: |
| ; X64-AVX-NEXT: xorl %r8d, %r8d |
| ; X64-AVX-NEXT: movl %edi, %eax |
| ; X64-AVX-NEXT: movl %esi, %edx |
| ; X64-AVX-NEXT: vpcmpestri $7, %xmm1, %xmm0 |
| ; X64-AVX-NEXT: seta %r8b |
| ; X64-AVX-NEXT: movl %r8d, %eax |
| ; X64-AVX-NEXT: retq |
| %arg0 = bitcast <2 x i64> %a0 to <16 x i8> |
| %arg2 = bitcast <2 x i64> %a2 to <16 x i8> |
| %res = call i32 @llvm.x86.sse42.pcmpestria128(<16 x i8> %arg0, i32 %a1, <16 x i8> %arg2, i32 %a3, i8 7) |
| ret i32 %res |
| } |
| declare i32 @llvm.x86.sse42.pcmpestria128(<16 x i8>, i32, <16 x i8>, i32, i8) nounwind readnone |
| |
| define i32 @test_mm_cmpestrc(<2 x i64> %a0, i32 %a1, <2 x i64> %a2, i32 %a3) nounwind { |
| ; X86-SSE-LABEL: test_mm_cmpestrc: |
| ; X86-SSE: # %bb.0: |
| ; X86-SSE-NEXT: pushl %ebx |
| ; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-SSE-NEXT: xorl %ebx, %ebx |
| ; X86-SSE-NEXT: pcmpestri $7, %xmm1, %xmm0 |
| ; X86-SSE-NEXT: setb %bl |
| ; X86-SSE-NEXT: movl %ebx, %eax |
| ; X86-SSE-NEXT: popl %ebx |
| ; X86-SSE-NEXT: retl |
| ; |
| ; X86-AVX-LABEL: test_mm_cmpestrc: |
| ; X86-AVX: # %bb.0: |
| ; X86-AVX-NEXT: pushl %ebx |
| ; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-AVX-NEXT: xorl %ebx, %ebx |
| ; X86-AVX-NEXT: vpcmpestri $7, %xmm1, %xmm0 |
| ; X86-AVX-NEXT: setb %bl |
| ; X86-AVX-NEXT: movl %ebx, %eax |
| ; X86-AVX-NEXT: popl %ebx |
| ; X86-AVX-NEXT: retl |
| ; |
| ; X64-SSE-LABEL: test_mm_cmpestrc: |
| ; X64-SSE: # %bb.0: |
| ; X64-SSE-NEXT: xorl %r8d, %r8d |
| ; X64-SSE-NEXT: movl %edi, %eax |
| ; X64-SSE-NEXT: movl %esi, %edx |
| ; X64-SSE-NEXT: pcmpestri $7, %xmm1, %xmm0 |
| ; X64-SSE-NEXT: setb %r8b |
| ; X64-SSE-NEXT: movl %r8d, %eax |
| ; X64-SSE-NEXT: retq |
| ; |
| ; X64-AVX-LABEL: test_mm_cmpestrc: |
| ; X64-AVX: # %bb.0: |
| ; X64-AVX-NEXT: xorl %r8d, %r8d |
| ; X64-AVX-NEXT: movl %edi, %eax |
| ; X64-AVX-NEXT: movl %esi, %edx |
| ; X64-AVX-NEXT: vpcmpestri $7, %xmm1, %xmm0 |
| ; X64-AVX-NEXT: setb %r8b |
| ; X64-AVX-NEXT: movl %r8d, %eax |
| ; X64-AVX-NEXT: retq |
| %arg0 = bitcast <2 x i64> %a0 to <16 x i8> |
| %arg2 = bitcast <2 x i64> %a2 to <16 x i8> |
| %res = call i32 @llvm.x86.sse42.pcmpestric128(<16 x i8> %arg0, i32 %a1, <16 x i8> %arg2, i32 %a3, i8 7) |
| ret i32 %res |
| } |
| declare i32 @llvm.x86.sse42.pcmpestric128(<16 x i8>, i32, <16 x i8>, i32, i8) nounwind readnone |
| |
| define i32 @test_mm_cmpestri(<2 x i64> %a0, i32 %a1, <2 x i64> %a2, i32 %a3) { |
| ; X86-SSE-LABEL: test_mm_cmpestri: |
| ; X86-SSE: # %bb.0: |
| ; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-SSE-NEXT: pcmpestri $7, %xmm1, %xmm0 |
| ; X86-SSE-NEXT: movl %ecx, %eax |
| ; X86-SSE-NEXT: retl |
| ; |
| ; X86-AVX-LABEL: test_mm_cmpestri: |
| ; X86-AVX: # %bb.0: |
| ; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-AVX-NEXT: vpcmpestri $7, %xmm1, %xmm0 |
| ; X86-AVX-NEXT: movl %ecx, %eax |
| ; X86-AVX-NEXT: retl |
| ; |
| ; X64-SSE-LABEL: test_mm_cmpestri: |
| ; X64-SSE: # %bb.0: |
| ; X64-SSE-NEXT: movl %edi, %eax |
| ; X64-SSE-NEXT: movl %esi, %edx |
| ; X64-SSE-NEXT: pcmpestri $7, %xmm1, %xmm0 |
| ; X64-SSE-NEXT: movl %ecx, %eax |
| ; X64-SSE-NEXT: retq |
| ; |
| ; X64-AVX-LABEL: test_mm_cmpestri: |
| ; X64-AVX: # %bb.0: |
| ; X64-AVX-NEXT: movl %edi, %eax |
| ; X64-AVX-NEXT: movl %esi, %edx |
| ; X64-AVX-NEXT: vpcmpestri $7, %xmm1, %xmm0 |
| ; X64-AVX-NEXT: movl %ecx, %eax |
| ; X64-AVX-NEXT: retq |
| %arg0 = bitcast <2 x i64> %a0 to <16 x i8> |
| %arg2 = bitcast <2 x i64> %a2 to <16 x i8> |
| %res = call i32 @llvm.x86.sse42.pcmpestri128(<16 x i8> %arg0, i32 %a1, <16 x i8> %arg2, i32 %a3, i8 7) |
| ret i32 %res |
| } |
| declare i32 @llvm.x86.sse42.pcmpestri128(<16 x i8>, i32, <16 x i8>, i32, i8) nounwind readnone |
| |
| define <2 x i64> @test_mm_cmpestrm(<2 x i64> %a0, i32 %a1, <2 x i64> %a2, i32 %a3) { |
| ; X86-SSE-LABEL: test_mm_cmpestrm: |
| ; X86-SSE: # %bb.0: |
| ; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-SSE-NEXT: pcmpestrm $7, %xmm1, %xmm0 |
| ; X86-SSE-NEXT: retl |
| ; |
| ; X86-AVX-LABEL: test_mm_cmpestrm: |
| ; X86-AVX: # %bb.0: |
| ; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-AVX-NEXT: vpcmpestrm $7, %xmm1, %xmm0 |
| ; X86-AVX-NEXT: retl |
| ; |
| ; X64-SSE-LABEL: test_mm_cmpestrm: |
| ; X64-SSE: # %bb.0: |
| ; X64-SSE-NEXT: movl %edi, %eax |
| ; X64-SSE-NEXT: movl %esi, %edx |
| ; X64-SSE-NEXT: pcmpestrm $7, %xmm1, %xmm0 |
| ; X64-SSE-NEXT: retq |
| ; |
| ; X64-AVX-LABEL: test_mm_cmpestrm: |
| ; X64-AVX: # %bb.0: |
| ; X64-AVX-NEXT: movl %edi, %eax |
| ; X64-AVX-NEXT: movl %esi, %edx |
| ; X64-AVX-NEXT: vpcmpestrm $7, %xmm1, %xmm0 |
| ; X64-AVX-NEXT: retq |
| %arg0 = bitcast <2 x i64> %a0 to <16 x i8> |
| %arg2 = bitcast <2 x i64> %a2 to <16 x i8> |
| %res = call <16 x i8> @llvm.x86.sse42.pcmpestrm128(<16 x i8> %arg0, i32 %a1, <16 x i8> %arg2, i32 %a3, i8 7) |
| %bc = bitcast <16 x i8> %res to <2 x i64> |
| ret <2 x i64> %bc |
| } |
| declare <16 x i8> @llvm.x86.sse42.pcmpestrm128(<16 x i8>, i32, <16 x i8>, i32, i8) nounwind readnone |
| |
| define i32 @test_mm_cmpestro(<2 x i64> %a0, i32 %a1, <2 x i64> %a2, i32 %a3) nounwind { |
| ; X86-SSE-LABEL: test_mm_cmpestro: |
| ; X86-SSE: # %bb.0: |
| ; X86-SSE-NEXT: pushl %ebx |
| ; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-SSE-NEXT: xorl %ebx, %ebx |
| ; X86-SSE-NEXT: pcmpestri $7, %xmm1, %xmm0 |
| ; X86-SSE-NEXT: seto %bl |
| ; X86-SSE-NEXT: movl %ebx, %eax |
| ; X86-SSE-NEXT: popl %ebx |
| ; X86-SSE-NEXT: retl |
| ; |
| ; X86-AVX-LABEL: test_mm_cmpestro: |
| ; X86-AVX: # %bb.0: |
| ; X86-AVX-NEXT: pushl %ebx |
| ; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-AVX-NEXT: xorl %ebx, %ebx |
| ; X86-AVX-NEXT: vpcmpestri $7, %xmm1, %xmm0 |
| ; X86-AVX-NEXT: seto %bl |
| ; X86-AVX-NEXT: movl %ebx, %eax |
| ; X86-AVX-NEXT: popl %ebx |
| ; X86-AVX-NEXT: retl |
| ; |
| ; X64-SSE-LABEL: test_mm_cmpestro: |
| ; X64-SSE: # %bb.0: |
| ; X64-SSE-NEXT: xorl %r8d, %r8d |
| ; X64-SSE-NEXT: movl %edi, %eax |
| ; X64-SSE-NEXT: movl %esi, %edx |
| ; X64-SSE-NEXT: pcmpestri $7, %xmm1, %xmm0 |
| ; X64-SSE-NEXT: seto %r8b |
| ; X64-SSE-NEXT: movl %r8d, %eax |
| ; X64-SSE-NEXT: retq |
| ; |
| ; X64-AVX-LABEL: test_mm_cmpestro: |
| ; X64-AVX: # %bb.0: |
| ; X64-AVX-NEXT: xorl %r8d, %r8d |
| ; X64-AVX-NEXT: movl %edi, %eax |
| ; X64-AVX-NEXT: movl %esi, %edx |
| ; X64-AVX-NEXT: vpcmpestri $7, %xmm1, %xmm0 |
| ; X64-AVX-NEXT: seto %r8b |
| ; X64-AVX-NEXT: movl %r8d, %eax |
| ; X64-AVX-NEXT: retq |
| %arg0 = bitcast <2 x i64> %a0 to <16 x i8> |
| %arg2 = bitcast <2 x i64> %a2 to <16 x i8> |
| %res = call i32 @llvm.x86.sse42.pcmpestrio128(<16 x i8> %arg0, i32 %a1, <16 x i8> %arg2, i32 %a3, i8 7) |
| ret i32 %res |
| } |
| declare i32 @llvm.x86.sse42.pcmpestrio128(<16 x i8>, i32, <16 x i8>, i32, i8) nounwind readnone |
| |
| define i32 @test_mm_cmpestrs(<2 x i64> %a0, i32 %a1, <2 x i64> %a2, i32 %a3) nounwind { |
| ; X86-SSE-LABEL: test_mm_cmpestrs: |
| ; X86-SSE: # %bb.0: |
| ; X86-SSE-NEXT: pushl %ebx |
| ; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-SSE-NEXT: xorl %ebx, %ebx |
| ; X86-SSE-NEXT: pcmpestri $7, %xmm1, %xmm0 |
| ; X86-SSE-NEXT: sets %bl |
| ; X86-SSE-NEXT: movl %ebx, %eax |
| ; X86-SSE-NEXT: popl %ebx |
| ; X86-SSE-NEXT: retl |
| ; |
| ; X86-AVX-LABEL: test_mm_cmpestrs: |
| ; X86-AVX: # %bb.0: |
| ; X86-AVX-NEXT: pushl %ebx |
| ; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-AVX-NEXT: xorl %ebx, %ebx |
| ; X86-AVX-NEXT: vpcmpestri $7, %xmm1, %xmm0 |
| ; X86-AVX-NEXT: sets %bl |
| ; X86-AVX-NEXT: movl %ebx, %eax |
| ; X86-AVX-NEXT: popl %ebx |
| ; X86-AVX-NEXT: retl |
| ; |
| ; X64-SSE-LABEL: test_mm_cmpestrs: |
| ; X64-SSE: # %bb.0: |
| ; X64-SSE-NEXT: xorl %r8d, %r8d |
| ; X64-SSE-NEXT: movl %edi, %eax |
| ; X64-SSE-NEXT: movl %esi, %edx |
| ; X64-SSE-NEXT: pcmpestri $7, %xmm1, %xmm0 |
| ; X64-SSE-NEXT: sets %r8b |
| ; X64-SSE-NEXT: movl %r8d, %eax |
| ; X64-SSE-NEXT: retq |
| ; |
| ; X64-AVX-LABEL: test_mm_cmpestrs: |
| ; X64-AVX: # %bb.0: |
| ; X64-AVX-NEXT: xorl %r8d, %r8d |
| ; X64-AVX-NEXT: movl %edi, %eax |
| ; X64-AVX-NEXT: movl %esi, %edx |
| ; X64-AVX-NEXT: vpcmpestri $7, %xmm1, %xmm0 |
| ; X64-AVX-NEXT: sets %r8b |
| ; X64-AVX-NEXT: movl %r8d, %eax |
| ; X64-AVX-NEXT: retq |
| %arg0 = bitcast <2 x i64> %a0 to <16 x i8> |
| %arg2 = bitcast <2 x i64> %a2 to <16 x i8> |
| %res = call i32 @llvm.x86.sse42.pcmpestris128(<16 x i8> %arg0, i32 %a1, <16 x i8> %arg2, i32 %a3, i8 7) |
| ret i32 %res |
| } |
| declare i32 @llvm.x86.sse42.pcmpestris128(<16 x i8>, i32, <16 x i8>, i32, i8) nounwind readnone |
| |
| define i32 @test_mm_cmpestrz(<2 x i64> %a0, i32 %a1, <2 x i64> %a2, i32 %a3) nounwind { |
| ; X86-SSE-LABEL: test_mm_cmpestrz: |
| ; X86-SSE: # %bb.0: |
| ; X86-SSE-NEXT: pushl %ebx |
| ; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-SSE-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-SSE-NEXT: xorl %ebx, %ebx |
| ; X86-SSE-NEXT: pcmpestri $7, %xmm1, %xmm0 |
| ; X86-SSE-NEXT: sete %bl |
| ; X86-SSE-NEXT: movl %ebx, %eax |
| ; X86-SSE-NEXT: popl %ebx |
| ; X86-SSE-NEXT: retl |
| ; |
| ; X86-AVX-LABEL: test_mm_cmpestrz: |
| ; X86-AVX: # %bb.0: |
| ; X86-AVX-NEXT: pushl %ebx |
| ; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-AVX-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-AVX-NEXT: xorl %ebx, %ebx |
| ; X86-AVX-NEXT: vpcmpestri $7, %xmm1, %xmm0 |
| ; X86-AVX-NEXT: sete %bl |
| ; X86-AVX-NEXT: movl %ebx, %eax |
| ; X86-AVX-NEXT: popl %ebx |
| ; X86-AVX-NEXT: retl |
| ; |
| ; X64-SSE-LABEL: test_mm_cmpestrz: |
| ; X64-SSE: # %bb.0: |
| ; X64-SSE-NEXT: xorl %r8d, %r8d |
| ; X64-SSE-NEXT: movl %edi, %eax |
| ; X64-SSE-NEXT: movl %esi, %edx |
| ; X64-SSE-NEXT: pcmpestri $7, %xmm1, %xmm0 |
| ; X64-SSE-NEXT: sete %r8b |
| ; X64-SSE-NEXT: movl %r8d, %eax |
| ; X64-SSE-NEXT: retq |
| ; |
| ; X64-AVX-LABEL: test_mm_cmpestrz: |
| ; X64-AVX: # %bb.0: |
| ; X64-AVX-NEXT: xorl %r8d, %r8d |
| ; X64-AVX-NEXT: movl %edi, %eax |
| ; X64-AVX-NEXT: movl %esi, %edx |
| ; X64-AVX-NEXT: vpcmpestri $7, %xmm1, %xmm0 |
| ; X64-AVX-NEXT: sete %r8b |
| ; X64-AVX-NEXT: movl %r8d, %eax |
| ; X64-AVX-NEXT: retq |
| %arg0 = bitcast <2 x i64> %a0 to <16 x i8> |
| %arg2 = bitcast <2 x i64> %a2 to <16 x i8> |
| %res = call i32 @llvm.x86.sse42.pcmpestriz128(<16 x i8> %arg0, i32 %a1, <16 x i8> %arg2, i32 %a3, i8 7) |
| ret i32 %res |
| } |
| declare i32 @llvm.x86.sse42.pcmpestriz128(<16 x i8>, i32, <16 x i8>, i32, i8) nounwind readnone |
| |
| define <2 x i64> @test_mm_cmpgt_epi64(<2 x i64> %a0, <2 x i64> %a1) { |
| ; SSE-LABEL: test_mm_cmpgt_epi64: |
| ; SSE: # %bb.0: |
| ; SSE-NEXT: pcmpgtq %xmm1, %xmm0 |
| ; SSE-NEXT: ret{{[l|q]}} |
| ; |
| ; AVX1-LABEL: test_mm_cmpgt_epi64: |
| ; AVX1: # %bb.0: |
| ; AVX1-NEXT: vpcmpgtq %xmm1, %xmm0, %xmm0 |
| ; AVX1-NEXT: ret{{[l|q]}} |
| ; |
| ; AVX512-LABEL: test_mm_cmpgt_epi64: |
| ; AVX512: # %bb.0: |
| ; AVX512-NEXT: vpcmpgtq %xmm1, %xmm0, %k0 |
| ; AVX512-NEXT: vpmovm2q %k0, %xmm0 |
| ; AVX512-NEXT: ret{{[l|q]}} |
| %cmp = icmp sgt <2 x i64> %a0, %a1 |
| %res = sext <2 x i1> %cmp to <2 x i64> |
| ret <2 x i64> %res |
| } |
| |
| define i32 @test_mm_cmpistra(<2 x i64> %a0, <2 x i64> %a1) { |
| ; SSE-LABEL: test_mm_cmpistra: |
| ; SSE: # %bb.0: |
| ; SSE-NEXT: xorl %eax, %eax |
| ; SSE-NEXT: pcmpistri $7, %xmm1, %xmm0 |
| ; SSE-NEXT: seta %al |
| ; SSE-NEXT: ret{{[l|q]}} |
| ; |
| ; AVX-LABEL: test_mm_cmpistra: |
| ; AVX: # %bb.0: |
| ; AVX-NEXT: xorl %eax, %eax |
| ; AVX-NEXT: vpcmpistri $7, %xmm1, %xmm0 |
| ; AVX-NEXT: seta %al |
| ; AVX-NEXT: ret{{[l|q]}} |
| %arg0 = bitcast <2 x i64> %a0 to <16 x i8> |
| %arg1 = bitcast <2 x i64> %a1 to <16 x i8> |
| %res = call i32 @llvm.x86.sse42.pcmpistria128(<16 x i8> %arg0, <16 x i8> %arg1, i8 7) |
| ret i32 %res |
| } |
| declare i32 @llvm.x86.sse42.pcmpistria128(<16 x i8>, <16 x i8>, i8) nounwind readnone |
| |
| define i32 @test_mm_cmpistrc(<2 x i64> %a0, <2 x i64> %a1) { |
| ; SSE-LABEL: test_mm_cmpistrc: |
| ; SSE: # %bb.0: |
| ; SSE-NEXT: xorl %eax, %eax |
| ; SSE-NEXT: pcmpistri $7, %xmm1, %xmm0 |
| ; SSE-NEXT: setb %al |
| ; SSE-NEXT: ret{{[l|q]}} |
| ; |
| ; AVX-LABEL: test_mm_cmpistrc: |
| ; AVX: # %bb.0: |
| ; AVX-NEXT: xorl %eax, %eax |
| ; AVX-NEXT: vpcmpistri $7, %xmm1, %xmm0 |
| ; AVX-NEXT: setb %al |
| ; AVX-NEXT: ret{{[l|q]}} |
| %arg0 = bitcast <2 x i64> %a0 to <16 x i8> |
| %arg1 = bitcast <2 x i64> %a1 to <16 x i8> |
| %res = call i32 @llvm.x86.sse42.pcmpistric128(<16 x i8> %arg0, <16 x i8> %arg1, i8 7) |
| ret i32 %res |
| } |
| declare i32 @llvm.x86.sse42.pcmpistric128(<16 x i8>, <16 x i8>, i8) nounwind readnone |
| |
| define i32 @test_mm_cmpistri(<2 x i64> %a0, <2 x i64> %a1) { |
| ; SSE-LABEL: test_mm_cmpistri: |
| ; SSE: # %bb.0: |
| ; SSE-NEXT: pcmpistri $7, %xmm1, %xmm0 |
| ; SSE-NEXT: movl %ecx, %eax |
| ; SSE-NEXT: ret{{[l|q]}} |
| ; |
| ; AVX-LABEL: test_mm_cmpistri: |
| ; AVX: # %bb.0: |
| ; AVX-NEXT: vpcmpistri $7, %xmm1, %xmm0 |
| ; AVX-NEXT: movl %ecx, %eax |
| ; AVX-NEXT: ret{{[l|q]}} |
| %arg0 = bitcast <2 x i64> %a0 to <16 x i8> |
| %arg1 = bitcast <2 x i64> %a1 to <16 x i8> |
| %res = call i32 @llvm.x86.sse42.pcmpistri128(<16 x i8> %arg0, <16 x i8> %arg1, i8 7) |
| ret i32 %res |
| } |
| declare i32 @llvm.x86.sse42.pcmpistri128(<16 x i8>, <16 x i8>, i8) nounwind readnone |
| |
| define <2 x i64> @test_mm_cmpistrm(<2 x i64> %a0, <2 x i64> %a1) { |
| ; SSE-LABEL: test_mm_cmpistrm: |
| ; SSE: # %bb.0: |
| ; SSE-NEXT: pcmpistrm $7, %xmm1, %xmm0 |
| ; SSE-NEXT: ret{{[l|q]}} |
| ; |
| ; AVX-LABEL: test_mm_cmpistrm: |
| ; AVX: # %bb.0: |
| ; AVX-NEXT: vpcmpistrm $7, %xmm1, %xmm0 |
| ; AVX-NEXT: ret{{[l|q]}} |
| %arg0 = bitcast <2 x i64> %a0 to <16 x i8> |
| %arg1 = bitcast <2 x i64> %a1 to <16 x i8> |
| %res = call <16 x i8> @llvm.x86.sse42.pcmpistrm128(<16 x i8> %arg0, <16 x i8> %arg1, i8 7) |
| %bc = bitcast <16 x i8> %res to <2 x i64> |
| ret <2 x i64> %bc |
| } |
| declare <16 x i8> @llvm.x86.sse42.pcmpistrm128(<16 x i8>, <16 x i8>, i8) nounwind readnone |
| |
| define i32 @test_mm_cmpistro(<2 x i64> %a0, <2 x i64> %a1) { |
| ; SSE-LABEL: test_mm_cmpistro: |
| ; SSE: # %bb.0: |
| ; SSE-NEXT: xorl %eax, %eax |
| ; SSE-NEXT: pcmpistri $7, %xmm1, %xmm0 |
| ; SSE-NEXT: seto %al |
| ; SSE-NEXT: ret{{[l|q]}} |
| ; |
| ; AVX-LABEL: test_mm_cmpistro: |
| ; AVX: # %bb.0: |
| ; AVX-NEXT: xorl %eax, %eax |
| ; AVX-NEXT: vpcmpistri $7, %xmm1, %xmm0 |
| ; AVX-NEXT: seto %al |
| ; AVX-NEXT: ret{{[l|q]}} |
| %arg0 = bitcast <2 x i64> %a0 to <16 x i8> |
| %arg1 = bitcast <2 x i64> %a1 to <16 x i8> |
| %res = call i32 @llvm.x86.sse42.pcmpistrio128(<16 x i8> %arg0, <16 x i8> %arg1, i8 7) |
| ret i32 %res |
| } |
| declare i32 @llvm.x86.sse42.pcmpistrio128(<16 x i8>, <16 x i8>, i8) nounwind readnone |
| |
| define i32 @test_mm_cmpistrs(<2 x i64> %a0, <2 x i64> %a1) { |
| ; SSE-LABEL: test_mm_cmpistrs: |
| ; SSE: # %bb.0: |
| ; SSE-NEXT: xorl %eax, %eax |
| ; SSE-NEXT: pcmpistri $7, %xmm1, %xmm0 |
| ; SSE-NEXT: sets %al |
| ; SSE-NEXT: ret{{[l|q]}} |
| ; |
| ; AVX-LABEL: test_mm_cmpistrs: |
| ; AVX: # %bb.0: |
| ; AVX-NEXT: xorl %eax, %eax |
| ; AVX-NEXT: vpcmpistri $7, %xmm1, %xmm0 |
| ; AVX-NEXT: sets %al |
| ; AVX-NEXT: ret{{[l|q]}} |
| %arg0 = bitcast <2 x i64> %a0 to <16 x i8> |
| %arg1 = bitcast <2 x i64> %a1 to <16 x i8> |
| %res = call i32 @llvm.x86.sse42.pcmpistris128(<16 x i8> %arg0, <16 x i8> %arg1, i8 7) |
| ret i32 %res |
| } |
| declare i32 @llvm.x86.sse42.pcmpistris128(<16 x i8>, <16 x i8>, i8) nounwind readnone |
| |
| define i32 @test_mm_cmpistrz(<2 x i64> %a0, <2 x i64> %a1) { |
| ; SSE-LABEL: test_mm_cmpistrz: |
| ; SSE: # %bb.0: |
| ; SSE-NEXT: xorl %eax, %eax |
| ; SSE-NEXT: pcmpistri $7, %xmm1, %xmm0 |
| ; SSE-NEXT: sete %al |
| ; SSE-NEXT: ret{{[l|q]}} |
| ; |
| ; AVX-LABEL: test_mm_cmpistrz: |
| ; AVX: # %bb.0: |
| ; AVX-NEXT: xorl %eax, %eax |
| ; AVX-NEXT: vpcmpistri $7, %xmm1, %xmm0 |
| ; AVX-NEXT: sete %al |
| ; AVX-NEXT: ret{{[l|q]}} |
| %arg0 = bitcast <2 x i64> %a0 to <16 x i8> |
| %arg1 = bitcast <2 x i64> %a1 to <16 x i8> |
| %res = call i32 @llvm.x86.sse42.pcmpistriz128(<16 x i8> %arg0, <16 x i8> %arg1, i8 7) |
| ret i32 %res |
| } |
| declare i32 @llvm.x86.sse42.pcmpistriz128(<16 x i8>, <16 x i8>, i8) nounwind readnone |
| |
| define i32 @test_mm_crc32_u8(i32 %a0, i8 %a1) { |
| ; X86-LABEL: test_mm_crc32_u8: |
| ; X86: # %bb.0: |
| ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-NEXT: crc32b {{[0-9]+}}(%esp), %eax |
| ; X86-NEXT: retl |
| ; |
| ; X64-LABEL: test_mm_crc32_u8: |
| ; X64: # %bb.0: |
| ; X64-NEXT: crc32b %sil, %edi |
| ; X64-NEXT: movl %edi, %eax |
| ; X64-NEXT: retq |
| %res = call i32 @llvm.x86.sse42.crc32.32.8(i32 %a0, i8 %a1) |
| ret i32 %res |
| } |
| declare i32 @llvm.x86.sse42.crc32.32.8(i32, i8) nounwind readnone |
| |
| define i32 @test_mm_crc32_u16(i32 %a0, i16 %a1) { |
| ; X86-LABEL: test_mm_crc32_u16: |
| ; X86: # %bb.0: |
| ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-NEXT: crc32w {{[0-9]+}}(%esp), %eax |
| ; X86-NEXT: retl |
| ; |
| ; X64-LABEL: test_mm_crc32_u16: |
| ; X64: # %bb.0: |
| ; X64-NEXT: crc32w %si, %edi |
| ; X64-NEXT: movl %edi, %eax |
| ; X64-NEXT: retq |
| %res = call i32 @llvm.x86.sse42.crc32.32.16(i32 %a0, i16 %a1) |
| ret i32 %res |
| } |
| declare i32 @llvm.x86.sse42.crc32.32.16(i32, i16) nounwind readnone |
| |
| define i32 @test_mm_crc32_u32(i32 %a0, i32 %a1) { |
| ; X86-LABEL: test_mm_crc32_u32: |
| ; X86: # %bb.0: |
| ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-NEXT: crc32l {{[0-9]+}}(%esp), %eax |
| ; X86-NEXT: retl |
| ; |
| ; X64-LABEL: test_mm_crc32_u32: |
| ; X64: # %bb.0: |
| ; X64-NEXT: crc32l %esi, %edi |
| ; X64-NEXT: movl %edi, %eax |
| ; X64-NEXT: retq |
| %res = call i32 @llvm.x86.sse42.crc32.32.32(i32 %a0, i32 %a1) |
| ret i32 %res |
| } |
| declare i32 @llvm.x86.sse42.crc32.32.32(i32, i32) nounwind readnone |