forked from OSchip/llvm-project
				
			
		
			
				
	
	
		
			231 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			LLVM
		
	
	
	
			
		
		
	
	
			231 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			LLVM
		
	
	
	
; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
 | 
						|
; RUN: opt < %s -instcombine -S | FileCheck %s
 | 
						|
target datalayout = "e-m:e-i64:64-f80:128-n8:16:32:64-S128"
 | 
						|
 | 
						|
; Verify that instcombine is able to fold identity shuffles.
 | 
						|
 | 
						|
define <4 x float> @identity_test_vpermilvar_ps(<4 x float> %v) {
 | 
						|
; CHECK-LABEL: @identity_test_vpermilvar_ps(
 | 
						|
; CHECK-NEXT:    ret <4 x float> %v
 | 
						|
;
 | 
						|
  %a = tail call <4 x float> @llvm.x86.avx.vpermilvar.ps(<4 x float> %v, <4 x i32> <i32 0, i32 1, i32 2, i32 3>)
 | 
						|
  ret <4 x float> %a
 | 
						|
}
 | 
						|
 | 
						|
define <8 x float> @identity_test_vpermilvar_ps_256(<8 x float> %v) {
 | 
						|
; CHECK-LABEL: @identity_test_vpermilvar_ps_256(
 | 
						|
; CHECK-NEXT:    ret <8 x float> %v
 | 
						|
;
 | 
						|
  %a = tail call <8 x float> @llvm.x86.avx.vpermilvar.ps.256(<8 x float> %v, <8 x i32> <i32 0, i32 1, i32 2, i32 3, i32 4, i32 5, i32 6, i32 7>)
 | 
						|
  ret <8 x float> %a
 | 
						|
}
 | 
						|
 | 
						|
define <16 x float> @identity_test_vpermilvar_ps_512(<16 x float> %v) {
 | 
						|
; CHECK-LABEL: @identity_test_vpermilvar_ps_512(
 | 
						|
; CHECK-NEXT:    ret <16 x float> %v
 | 
						|
;
 | 
						|
  %a = tail call <16 x float> @llvm.x86.avx512.vpermilvar.ps.512(<16 x float> %v, <16 x i32> <i32 0, i32 1, i32 2, i32 3, i32 0, i32 1, i32 2, i32 3, i32 0, i32 1, i32 2, i32 3, i32 0, i32 1, i32 2, i32 3>)
 | 
						|
  ret <16 x float> %a
 | 
						|
}
 | 
						|
 | 
						|
define <2 x double> @identity_test_vpermilvar_pd(<2 x double> %v) {
 | 
						|
; CHECK-LABEL: @identity_test_vpermilvar_pd(
 | 
						|
; CHECK-NEXT:    ret <2 x double> %v
 | 
						|
;
 | 
						|
  %a = tail call <2 x double> @llvm.x86.avx.vpermilvar.pd(<2 x double> %v, <2 x i64> <i64 0, i64 2>)
 | 
						|
  ret <2 x double> %a
 | 
						|
}
 | 
						|
 | 
						|
define <4 x double> @identity_test_vpermilvar_pd_256(<4 x double> %v) {
 | 
						|
; CHECK-LABEL: @identity_test_vpermilvar_pd_256(
 | 
						|
; CHECK-NEXT:    ret <4 x double> %v
 | 
						|
;
 | 
						|
  %a = tail call <4 x double> @llvm.x86.avx.vpermilvar.pd.256(<4 x double> %v, <4 x i64> <i64 0, i64 2, i64 0, i64 2>)
 | 
						|
  ret <4 x double> %a
 | 
						|
}
 | 
						|
 | 
						|
define <8 x double> @identity_test_vpermilvar_pd_512(<8 x double> %v) {
 | 
						|
; CHECK-LABEL: @identity_test_vpermilvar_pd_512(
 | 
						|
; CHECK-NEXT:    ret <8 x double> %v
 | 
						|
;
 | 
						|
  %a = tail call <8 x double> @llvm.x86.avx512.vpermilvar.pd.512(<8 x double> %v, <8 x i64> <i64 0, i64 2, i64 0, i64 2, i64 0, i64 2, i64 0, i64 2>)
 | 
						|
  ret <8 x double> %a
 | 
						|
}
 | 
						|
 | 
						|
; Instcombine should be able to fold the following byte shuffle to a builtin shufflevector
 | 
						|
; with a shuffle mask of all zeroes.
 | 
						|
 | 
						|
define <4 x float> @zero_test_vpermilvar_ps_zero(<4 x float> %v) {
 | 
						|
; CHECK-LABEL: @zero_test_vpermilvar_ps_zero(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <4 x float> %v, <4 x float> undef, <4 x i32> zeroinitializer
 | 
						|
; CHECK-NEXT:    ret <4 x float> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <4 x float> @llvm.x86.avx.vpermilvar.ps(<4 x float> %v, <4 x i32> zeroinitializer)
 | 
						|
  ret <4 x float> %a
 | 
						|
}
 | 
						|
 | 
						|
define <8 x float> @zero_test_vpermilvar_ps_256_zero(<8 x float> %v) {
 | 
						|
; CHECK-LABEL: @zero_test_vpermilvar_ps_256_zero(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <8 x float> %v, <8 x float> undef, <8 x i32> <i32 0, i32 0, i32 0, i32 0, i32 4, i32 4, i32 4, i32 4>
 | 
						|
; CHECK-NEXT:    ret <8 x float> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <8 x float> @llvm.x86.avx.vpermilvar.ps.256(<8 x float> %v, <8 x i32> zeroinitializer)
 | 
						|
  ret <8 x float> %a
 | 
						|
}
 | 
						|
 | 
						|
define <16 x float> @zero_test_vpermilvar_ps_512_zero(<16 x float> %v) {
 | 
						|
; CHECK-LABEL: @zero_test_vpermilvar_ps_512_zero(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <16 x float> %v, <16 x float> undef, <16 x i32> <i32 0, i32 0, i32 0, i32 0, i32 4, i32 4, i32 4, i32 4, i32 8, i32 8, i32 8, i32 8, i32 12, i32 12, i32 12, i32 12>
 | 
						|
; CHECK-NEXT:    ret <16 x float> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <16 x float> @llvm.x86.avx512.vpermilvar.ps.512(<16 x float> %v, <16 x i32> zeroinitializer)
 | 
						|
  ret <16 x float> %a
 | 
						|
}
 | 
						|
 | 
						|
define <2 x double> @zero_test_vpermilvar_pd_zero(<2 x double> %v) {
 | 
						|
; CHECK-LABEL: @zero_test_vpermilvar_pd_zero(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <2 x double> %v, <2 x double> undef, <2 x i32> zeroinitializer
 | 
						|
; CHECK-NEXT:    ret <2 x double> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <2 x double> @llvm.x86.avx.vpermilvar.pd(<2 x double> %v, <2 x i64> zeroinitializer)
 | 
						|
  ret <2 x double> %a
 | 
						|
}
 | 
						|
 | 
						|
define <4 x double> @zero_test_vpermilvar_pd_256_zero(<4 x double> %v) {
 | 
						|
; CHECK-LABEL: @zero_test_vpermilvar_pd_256_zero(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <4 x double> %v, <4 x double> undef, <4 x i32> <i32 0, i32 0, i32 2, i32 2>
 | 
						|
; CHECK-NEXT:    ret <4 x double> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <4 x double> @llvm.x86.avx.vpermilvar.pd.256(<4 x double> %v, <4 x i64> zeroinitializer)
 | 
						|
  ret <4 x double> %a
 | 
						|
}
 | 
						|
 | 
						|
define <8 x double> @zero_test_vpermilvar_pd_512_zero(<8 x double> %v) {
 | 
						|
; CHECK-LABEL: @zero_test_vpermilvar_pd_512_zero(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <8 x double> %v, <8 x double> undef, <8 x i32> <i32 0, i32 0, i32 2, i32 2, i32 4, i32 4, i32 6, i32 6>
 | 
						|
; CHECK-NEXT:    ret <8 x double> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <8 x double> @llvm.x86.avx512.vpermilvar.pd.512(<8 x double> %v, <8 x i64> zeroinitializer)
 | 
						|
  ret <8 x double> %a
 | 
						|
}
 | 
						|
 | 
						|
; Verify that instcombine is able to fold constant shuffles.
 | 
						|
 | 
						|
define <4 x float> @test_vpermilvar_ps(<4 x float> %v) {
 | 
						|
; CHECK-LABEL: @test_vpermilvar_ps(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <4 x float> %v, <4 x float> undef, <4 x i32> <i32 3, i32 2, i32 1, i32 0>
 | 
						|
; CHECK-NEXT:    ret <4 x float> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <4 x float> @llvm.x86.avx.vpermilvar.ps(<4 x float> %v, <4 x i32> <i32 3, i32 2, i32 1, i32 0>)
 | 
						|
  ret <4 x float> %a
 | 
						|
}
 | 
						|
 | 
						|
define <8 x float> @test_vpermilvar_ps_256(<8 x float> %v) {
 | 
						|
; CHECK-LABEL: @test_vpermilvar_ps_256(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <8 x float> %v, <8 x float> undef, <8 x i32> <i32 3, i32 2, i32 1, i32 0, i32 7, i32 6, i32 5, i32 4>
 | 
						|
; CHECK-NEXT:    ret <8 x float> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <8 x float> @llvm.x86.avx.vpermilvar.ps.256(<8 x float> %v, <8 x i32> <i32 7, i32 6, i32 5, i32 4, i32 3, i32 2, i32 1, i32 0>)
 | 
						|
  ret <8 x float> %a
 | 
						|
}
 | 
						|
 | 
						|
define <16 x float> @test_vpermilvar_ps_512(<16 x float> %v) {
 | 
						|
; CHECK-LABEL: @test_vpermilvar_ps_512(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <16 x float> %v, <16 x float> undef, <16 x i32> <i32 3, i32 2, i32 1, i32 0, i32 7, i32 6, i32 5, i32 4, i32 11, i32 10, i32 9, i32 8, i32 15, i32 14, i32 13, i32 12>
 | 
						|
; CHECK-NEXT:    ret <16 x float> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <16 x float> @llvm.x86.avx512.vpermilvar.ps.512(<16 x float> %v, <16 x i32> <i32 15, i32 14, i32 13, i32 12, i32 11, i32 10, i32 9, i32 8, i32 7, i32 6, i32 5, i32 4, i32 3, i32 2, i32 1, i32 0>)
 | 
						|
  ret <16 x float> %a
 | 
						|
}
 | 
						|
 | 
						|
define <2 x double> @test_vpermilvar_pd(<2 x double> %v) {
 | 
						|
; CHECK-LABEL: @test_vpermilvar_pd(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <2 x double> %v, <2 x double> undef, <2 x i32> <i32 1, i32 0>
 | 
						|
; CHECK-NEXT:    ret <2 x double> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <2 x double> @llvm.x86.avx.vpermilvar.pd(<2 x double> %v, <2 x i64> <i64 2, i64 0>)
 | 
						|
  ret <2 x double> %a
 | 
						|
}
 | 
						|
 | 
						|
define <4 x double> @test_vpermilvar_pd_256(<4 x double> %v) {
 | 
						|
; CHECK-LABEL: @test_vpermilvar_pd_256(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <4 x double> %v, <4 x double> undef, <4 x i32> <i32 1, i32 0, i32 3, i32 2>
 | 
						|
; CHECK-NEXT:    ret <4 x double> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <4 x double> @llvm.x86.avx.vpermilvar.pd.256(<4 x double> %v, <4 x i64> <i64 3, i64 1, i64 2, i64 0>)
 | 
						|
  ret <4 x double> %a
 | 
						|
}
 | 
						|
 | 
						|
define <8 x double> @test_vpermilvar_pd_512(<8 x double> %v) {
 | 
						|
; CHECK-LABEL: @test_vpermilvar_pd_512(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <8 x double> %v, <8 x double> undef, <8 x i32> <i32 1, i32 0, i32 3, i32 2, i32 5, i32 4, i32 7, i32 6>
 | 
						|
; CHECK-NEXT:    ret <8 x double> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <8 x double> @llvm.x86.avx512.vpermilvar.pd.512(<8 x double> %v, <8 x i64> <i64 3, i64 1, i64 2, i64 0, i64 7, i64 5, i64 6, i64 4>)
 | 
						|
  ret <8 x double> %a
 | 
						|
}
 | 
						|
 | 
						|
; Verify that instcombine is able to fold constant shuffles with undef mask elements.
 | 
						|
 | 
						|
define <4 x float> @undef_test_vpermilvar_ps(<4 x float> %v) {
 | 
						|
; CHECK-LABEL: @undef_test_vpermilvar_ps(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <4 x float> %v, <4 x float> undef, <4 x i32> <i32 undef, i32 2, i32 1, i32 undef>
 | 
						|
; CHECK-NEXT:    ret <4 x float> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <4 x float> @llvm.x86.avx.vpermilvar.ps(<4 x float> %v, <4 x i32> <i32 undef, i32 2, i32 1, i32 undef>)
 | 
						|
  ret <4 x float> %a
 | 
						|
}
 | 
						|
 | 
						|
define <8 x float> @undef_test_vpermilvar_ps_256(<8 x float> %v) {
 | 
						|
; CHECK-LABEL: @undef_test_vpermilvar_ps_256(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <8 x float> %v, <8 x float> undef, <8 x i32> <i32 undef, i32 2, i32 1, i32 undef, i32 7, i32 6, i32 5, i32 4>
 | 
						|
; CHECK-NEXT:    ret <8 x float> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <8 x float> @llvm.x86.avx.vpermilvar.ps.256(<8 x float> %v, <8 x i32> <i32 undef, i32 6, i32 5, i32 undef, i32 3, i32 2, i32 1, i32 0>)
 | 
						|
  ret <8 x float> %a
 | 
						|
}
 | 
						|
 | 
						|
define <16 x float> @undef_test_vpermilvar_ps_512(<16 x float> %v) {
 | 
						|
; CHECK-LABEL: @undef_test_vpermilvar_ps_512(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <16 x float> %v, <16 x float> undef, <16 x i32> <i32 undef, i32 2, i32 1, i32 undef, i32 7, i32 6, i32 5, i32 4, i32 undef, i32 10, i32 9, i32 undef, i32 15, i32 14, i32 13, i32 12>
 | 
						|
; CHECK-NEXT:    ret <16 x float> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <16 x float> @llvm.x86.avx512.vpermilvar.ps.512(<16 x float> %v, <16 x i32> <i32 undef, i32 6, i32 5, i32 undef, i32 3, i32 2, i32 1, i32 0, i32 undef, i32 6, i32 5, i32 undef, i32 3, i32 2, i32 1, i32 0>)
 | 
						|
  ret <16 x float> %a
 | 
						|
}
 | 
						|
 | 
						|
define <2 x double> @undef_test_vpermilvar_pd(<2 x double> %v) {
 | 
						|
; CHECK-LABEL: @undef_test_vpermilvar_pd(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <2 x double> %v, <2 x double> undef, <2 x i32> <i32 undef, i32 0>
 | 
						|
; CHECK-NEXT:    ret <2 x double> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <2 x double> @llvm.x86.avx.vpermilvar.pd(<2 x double> %v, <2 x i64> <i64 undef, i64 0>)
 | 
						|
  ret <2 x double> %a
 | 
						|
}
 | 
						|
 | 
						|
define <4 x double> @undef_test_vpermilvar_pd_256(<4 x double> %v) {
 | 
						|
; CHECK-LABEL: @undef_test_vpermilvar_pd_256(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <4 x double> %v, <4 x double> undef, <4 x i32> <i32 undef, i32 0, i32 3, i32 undef>
 | 
						|
; CHECK-NEXT:    ret <4 x double> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <4 x double> @llvm.x86.avx.vpermilvar.pd.256(<4 x double> %v, <4 x i64> <i64 undef, i64 1, i64 2, i64 undef>)
 | 
						|
  ret <4 x double> %a
 | 
						|
}
 | 
						|
 | 
						|
define <8 x double> @undef_test_vpermilvar_pd_512(<8 x double> %v) {
 | 
						|
; CHECK-LABEL: @undef_test_vpermilvar_pd_512(
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = shufflevector <8 x double> %v, <8 x double> undef, <8 x i32> <i32 undef, i32 0, i32 3, i32 undef, i32 undef, i32 4, i32 7, i32 undef>
 | 
						|
; CHECK-NEXT:    ret <8 x double> [[TMP1]]
 | 
						|
;
 | 
						|
  %a = tail call <8 x double> @llvm.x86.avx512.vpermilvar.pd.512(<8 x double> %v, <8 x i64> <i64 undef, i64 1, i64 2, i64 undef, i64 undef, i64 1, i64 2, i64 undef>)
 | 
						|
  ret <8 x double> %a
 | 
						|
}
 | 
						|
 | 
						|
declare <2 x double> @llvm.x86.avx.vpermilvar.pd(<2 x double>, <2 x i64>)
 | 
						|
declare <4 x double> @llvm.x86.avx.vpermilvar.pd.256(<4 x double>, <4 x i64>)
 | 
						|
declare <8 x double> @llvm.x86.avx512.vpermilvar.pd.512(<8 x double>, <8 x i64>)
 | 
						|
 | 
						|
declare <4 x float> @llvm.x86.avx.vpermilvar.ps(<4 x float>, <4 x i32>)
 | 
						|
declare <8 x float> @llvm.x86.avx.vpermilvar.ps.256(<8 x float>, <8 x i32>)
 | 
						|
declare <16 x float> @llvm.x86.avx512.vpermilvar.ps.512(<16 x float>, <16 x i32>)
 |