100 lines
		
	
	
		
			3.6 KiB
		
	
	
	
		
			LLVM
		
	
	
	
			
		
		
	
	
			100 lines
		
	
	
		
			3.6 KiB
		
	
	
	
		
			LLVM
		
	
	
	
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
 | 
						|
; RUN: llc < %s -mtriple=i686-unknown -mattr=+sse2 | FileCheck %s --check-prefix=X86
 | 
						|
; RUN: llc < %s -mtriple=x86_64-unknown -mattr=+sse2 | FileCheck %s --check-prefix=X64
 | 
						|
 | 
						|
; This test makes sure that the compiler does not crash with an
 | 
						|
; assertion failure when trying to fold a vector shift left
 | 
						|
; by immediate count if the type of the input vector is different
 | 
						|
; to the result type.
 | 
						|
;
 | 
						|
; This happens for example when lowering a shift left of a MVT::v16i8 vector.
 | 
						|
; This is custom lowered into the following sequence:
 | 
						|
;     count << 5
 | 
						|
;     A =  VSHLI(MVT::v8i16, r & (char16)15, 4)
 | 
						|
;     B = BITCAST MVT::v16i8, A
 | 
						|
;     VSELECT(r, B, count);
 | 
						|
;     count += count
 | 
						|
;     C = VSHLI(MVT::v8i16, r & (char16)63, 2)
 | 
						|
;     D = BITCAST MVT::v16i8, C
 | 
						|
;     r = VSELECT(r, C, count);
 | 
						|
;     count += count
 | 
						|
;     VSELECT(r, r+r, count);
 | 
						|
;     count = count << 5;
 | 
						|
;
 | 
						|
; Where 'r' is a vector of type MVT::v16i8, and
 | 
						|
; 'count' is the vector shift count.
 | 
						|
 | 
						|
define <16 x i8> @do_not_crash(i8*, i32*, i64*, i32, i64, i8) {
 | 
						|
; X86-LABEL: do_not_crash:
 | 
						|
; X86:       # %bb.0: # %entry
 | 
						|
; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax
 | 
						|
; X86-NEXT:    movl {{[0-9]+}}(%esp), %ecx
 | 
						|
; X86-NEXT:    movb %al, (%ecx)
 | 
						|
; X86-NEXT:    movd %eax, %xmm1
 | 
						|
; X86-NEXT:    psllq $56, %xmm1
 | 
						|
; X86-NEXT:    por {{\.?LCPI[0-9]+_[0-9]+}}, %xmm1
 | 
						|
; X86-NEXT:    pcmpeqd %xmm3, %xmm3
 | 
						|
; X86-NEXT:    psllw $5, %xmm1
 | 
						|
; X86-NEXT:    pxor %xmm2, %xmm2
 | 
						|
; X86-NEXT:    pxor %xmm0, %xmm0
 | 
						|
; X86-NEXT:    pcmpgtb %xmm1, %xmm0
 | 
						|
; X86-NEXT:    pxor %xmm0, %xmm3
 | 
						|
; X86-NEXT:    pand {{\.?LCPI[0-9]+_[0-9]+}}, %xmm0
 | 
						|
; X86-NEXT:    por %xmm3, %xmm0
 | 
						|
; X86-NEXT:    paddb %xmm1, %xmm1
 | 
						|
; X86-NEXT:    pxor %xmm3, %xmm3
 | 
						|
; X86-NEXT:    pcmpgtb %xmm1, %xmm3
 | 
						|
; X86-NEXT:    movdqa %xmm3, %xmm4
 | 
						|
; X86-NEXT:    pandn %xmm0, %xmm4
 | 
						|
; X86-NEXT:    psllw $2, %xmm0
 | 
						|
; X86-NEXT:    pand %xmm3, %xmm0
 | 
						|
; X86-NEXT:    pand {{\.?LCPI[0-9]+_[0-9]+}}, %xmm0
 | 
						|
; X86-NEXT:    por %xmm4, %xmm0
 | 
						|
; X86-NEXT:    paddb %xmm1, %xmm1
 | 
						|
; X86-NEXT:    pcmpgtb %xmm1, %xmm2
 | 
						|
; X86-NEXT:    movdqa %xmm2, %xmm1
 | 
						|
; X86-NEXT:    pandn %xmm0, %xmm1
 | 
						|
; X86-NEXT:    paddb %xmm0, %xmm0
 | 
						|
; X86-NEXT:    pand %xmm2, %xmm0
 | 
						|
; X86-NEXT:    por %xmm1, %xmm0
 | 
						|
; X86-NEXT:    retl
 | 
						|
;
 | 
						|
; X64-LABEL: do_not_crash:
 | 
						|
; X64:       # %bb.0: # %entry
 | 
						|
; X64-NEXT:    movb %r9b, (%rdi)
 | 
						|
; X64-NEXT:    movd %r9d, %xmm1
 | 
						|
; X64-NEXT:    psllq $56, %xmm1
 | 
						|
; X64-NEXT:    por {{\.?LCPI[0-9]+_[0-9]+}}(%rip), %xmm1
 | 
						|
; X64-NEXT:    pcmpeqd %xmm2, %xmm2
 | 
						|
; X64-NEXT:    psllw $5, %xmm1
 | 
						|
; X64-NEXT:    pxor %xmm3, %xmm3
 | 
						|
; X64-NEXT:    pxor %xmm0, %xmm0
 | 
						|
; X64-NEXT:    pcmpgtb %xmm1, %xmm0
 | 
						|
; X64-NEXT:    pxor %xmm0, %xmm2
 | 
						|
; X64-NEXT:    pand {{\.?LCPI[0-9]+_[0-9]+}}(%rip), %xmm0
 | 
						|
; X64-NEXT:    por %xmm2, %xmm0
 | 
						|
; X64-NEXT:    paddb %xmm1, %xmm1
 | 
						|
; X64-NEXT:    pxor %xmm2, %xmm2
 | 
						|
; X64-NEXT:    pcmpgtb %xmm1, %xmm2
 | 
						|
; X64-NEXT:    movdqa %xmm2, %xmm4
 | 
						|
; X64-NEXT:    pandn %xmm0, %xmm4
 | 
						|
; X64-NEXT:    psllw $2, %xmm0
 | 
						|
; X64-NEXT:    pand %xmm2, %xmm0
 | 
						|
; X64-NEXT:    pand {{\.?LCPI[0-9]+_[0-9]+}}(%rip), %xmm0
 | 
						|
; X64-NEXT:    por %xmm4, %xmm0
 | 
						|
; X64-NEXT:    paddb %xmm1, %xmm1
 | 
						|
; X64-NEXT:    pcmpgtb %xmm1, %xmm3
 | 
						|
; X64-NEXT:    movdqa %xmm3, %xmm1
 | 
						|
; X64-NEXT:    pandn %xmm0, %xmm1
 | 
						|
; X64-NEXT:    paddb %xmm0, %xmm0
 | 
						|
; X64-NEXT:    pand %xmm3, %xmm0
 | 
						|
; X64-NEXT:    por %xmm1, %xmm0
 | 
						|
; X64-NEXT:    retq
 | 
						|
entry:
 | 
						|
  store i8 %5, i8* %0
 | 
						|
  %L5 = load i8, i8* %0
 | 
						|
  %I8 = insertelement <16 x i8> <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1>, i8 %L5, i32 7
 | 
						|
  %B51 = shl <16 x i8> <i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1, i8 -1>, %I8
 | 
						|
  ret <16 x i8> %B51
 | 
						|
}
 |