forked from OSchip/llvm-project
				
			
		
			
				
	
	
		
			928 lines
		
	
	
		
			32 KiB
		
	
	
	
		
			LLVM
		
	
	
	
			
		
		
	
	
			928 lines
		
	
	
		
			32 KiB
		
	
	
	
		
			LLVM
		
	
	
	
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
 | 
						|
; RUN: llc < %s -mtriple=x86_64-unknown-linux-gnu | FileCheck %s
 | 
						|
 | 
						|
define i8* @test_memcpy1_generic(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memcpy1_generic:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memcpy_element_unordered_atomic_1
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i32(i8* align 4 %P, i8* align 4 %Q, i32 1024, i32 1)
 | 
						|
  ret i8* %P
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memcpy2_generic(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memcpy2_generic:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memcpy_element_unordered_atomic_2
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i32(i8* align 4 %P, i8* align 4 %Q, i32 1024, i32 2)
 | 
						|
  ret i8* %P
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memcpy4_generic(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memcpy4_generic:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memcpy_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i32(i8* align 4 %P, i8* align 4 %Q, i32 1024, i32 4)
 | 
						|
  ret i8* %P
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memcpy8(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memcpy8:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memcpy_element_unordered_atomic_8
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i32(i8* align 8 %P, i8* align 8 %Q, i32 1024, i32 8)
 | 
						|
  ret i8* %P
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memcpy16_generic(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memcpy16_generic:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memcpy_element_unordered_atomic_16
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i32(i8* align 16 %P, i8* align 16 %Q, i32 1024, i32 16)
 | 
						|
  ret i8* %P
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
}
 | 
						|
 | 
						|
define void @test_memcpy_args(i8** %Storage) {
 | 
						|
; CHECK-LABEL: test_memcpy_args:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rax
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    movq (%rdi), %rax
 | 
						|
; CHECK-NEXT:    movq 8(%rdi), %rsi
 | 
						|
; CHECK-NEXT:    movq %rax, %rdi
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memcpy_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    popq %rax
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  %Dst = load i8*, i8** %Storage
 | 
						|
  %Src.addr = getelementptr i8*, i8** %Storage, i64 1
 | 
						|
  %Src = load i8*, i8** %Src.addr
 | 
						|
 | 
						|
  ; 1st arg (%rdi)
 | 
						|
  ; 2nd arg (%rsi)
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
  call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i32(i8* align 4 %Dst, i8* align 4 %Src, i32 1024, i32 4)
 | 
						|
  ret void
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memmove1_generic(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memmove1_generic:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memmove_element_unordered_atomic_1
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i32(i8* align 4 %P, i8* align 4 %Q, i32 1024, i32 1)
 | 
						|
  ret i8* %P
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memmove2_generic(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memmove2_generic:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memmove_element_unordered_atomic_2
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i32(i8* align 4 %P, i8* align 4 %Q, i32 1024, i32 2)
 | 
						|
  ret i8* %P
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memmove4_generic(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memmove4_generic:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memmove_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i32(i8* align 4 %P, i8* align 4 %Q, i32 1024, i32 4)
 | 
						|
  ret i8* %P
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memmove8_generic(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memmove8_generic:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memmove_element_unordered_atomic_8
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i32(i8* align 8 %P, i8* align 8 %Q, i32 1024, i32 8)
 | 
						|
  ret i8* %P
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memmove16_generic(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memmove16_generic:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memmove_element_unordered_atomic_16
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i32(i8* align 16 %P, i8* align 16 %Q, i32 1024, i32 16)
 | 
						|
  ret i8* %P
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
}
 | 
						|
 | 
						|
define void @test_memmove_args(i8** %Storage) {
 | 
						|
; CHECK-LABEL: test_memmove_args:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rax
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    movq (%rdi), %rax
 | 
						|
; CHECK-NEXT:    movq 8(%rdi), %rsi
 | 
						|
; CHECK-NEXT:    movq %rax, %rdi
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memmove_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    popq %rax
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  %Dst = load i8*, i8** %Storage
 | 
						|
  %Src.addr = getelementptr i8*, i8** %Storage, i64 1
 | 
						|
  %Src = load i8*, i8** %Src.addr
 | 
						|
 | 
						|
  ; 1st arg (%rdi)
 | 
						|
  ; 2nd arg (%rsi)
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
  call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i32(i8* align 4 %Dst, i8* align 4 %Src, i32 1024, i32 4)
 | 
						|
  ret void
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset1_generic(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset1_generic:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_1
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 1 %P, i8 %V, i32 1024, i32 1)
 | 
						|
  ret i8* %P
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset2_generic(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset2_generic:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_2
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 2 %P, i8 %V, i32 1024, i32 2)
 | 
						|
  ret i8* %P
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset4_generic(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset4_generic:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 4 %P, i8 %V, i32 1024, i32 4)
 | 
						|
  ret i8* %P
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset8_generic(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset8_generic:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_8
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 8 %P, i8 %V, i32 1024, i32 8)
 | 
						|
  ret i8* %P
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset16_generic(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset16_generic:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_16
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 16 %P, i8 %V, i32 1024, i32 16)
 | 
						|
  ret i8* %P
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
}
 | 
						|
 | 
						|
define void @test_memset_args(i8** %Storage, i8* %V) {
 | 
						|
; CHECK-LABEL: test_memset_args:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rax
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    movq (%rdi), %rdi
 | 
						|
; CHECK-NEXT:    movzbl (%rsi), %esi
 | 
						|
; CHECK-NEXT:    movl $1024, %edx # imm = 0x400
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    popq %rax
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  %Dst = load i8*, i8** %Storage
 | 
						|
  %Val = load i8, i8* %V
 | 
						|
 | 
						|
  ; 1st arg (%rdi)
 | 
						|
  ; 2nd arg (%rsi)
 | 
						|
  ; 3rd arg (%edx) -- length
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 4 %Dst, i8 %Val, i32 1024, i32 4)
 | 
						|
  ret void
 | 
						|
}
 | 
						|
 | 
						|
;; Next batch of tests are cases where we could profitably lower to
 | 
						|
;; atomic loads and stores directly, just as we do for non-atomic ones for
 | 
						|
;; non element.unorderered.atomic variants.
 | 
						|
 | 
						|
define i8* @test_memcpy1_64(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memcpy1_64:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memcpy_element_unordered_atomic_1
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i32(i8* align 1 %P, i8* align 1 %Q, i32 64, i32 1)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
; Ensure align 16 generates vector load/stores even with small element size
 | 
						|
define i8* @test_memcpy1_64_align4(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memcpy1_64_align4:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memcpy_element_unordered_atomic_1
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i32(i8* align 4 %P, i8* align 4 %Q, i32 64, i32 1)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memcpy1_64_align8(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memcpy1_64_align8:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memcpy_element_unordered_atomic_1
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i32(i8* align 8 %P, i8* align 8 %Q, i32 64, i32 1)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
define i8* @test_memcpy1_64_align16(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memcpy1_64_align16:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memcpy_element_unordered_atomic_1
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i32(i8* align 16 %P, i8* align 16 %Q, i32 64, i32 1)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
; Make sure that different source & dest alignments are handled correctly.
 | 
						|
define i8* @test_memcpy1_64_diff_aligns(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memcpy1_64_diff_aligns:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memcpy_element_unordered_atomic_1
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i32(i8* align 1 %P, i8* align 16 %Q, i32 64, i32 1)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memcpy2_64(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memcpy2_64:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memcpy_element_unordered_atomic_2
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i32(i8* align 2 %P, i8* align 2 %Q, i32 64, i32 2)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memcpy4_64(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memcpy4_64:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memcpy_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i32(i8* align 4 %P, i8* align 4 %Q, i32 64, i32 4)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memcpy8_64(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memcpy8_64:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memcpy_element_unordered_atomic_8
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i32(i8* align 8 %P, i8* align 8 %Q, i32 64, i32 8)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memcpy16_64(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memcpy16_64:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memcpy_element_unordered_atomic_16
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i32(i8* align 16 %P, i8* align 16 %Q, i32 64, i32 16)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
; ==================================
 | 
						|
 | 
						|
define i8* @test_memmove1_64(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memmove1_64:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memmove_element_unordered_atomic_1
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i32(i8* align 1 %P, i8* align 1 %Q, i32 64, i32 1)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
; Ensure align 16 generates vector load/stores even with small element size
 | 
						|
define i8* @test_memmove1_64_align16(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memmove1_64_align16:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memmove_element_unordered_atomic_1
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i32(i8* align 16 %P, i8* align 16 %Q, i32 64, i32 1)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
; Make sure that different source & dest alignments are handled correctly.
 | 
						|
define i8* @test_memmove1_64_diff_aligns(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memmove1_64_diff_aligns:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memmove_element_unordered_atomic_1
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i32(i8* align 1 %P, i8* align 16 %Q, i32 64, i32 1)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memmove2_64(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memmove2_64:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memmove_element_unordered_atomic_2
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i32(i8* align 2 %P, i8* align 2 %Q, i32 64, i32 2)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memmove4_64(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memmove4_64:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memmove_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i32(i8* align 4 %P, i8* align 4 %Q, i32 64, i32 4)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memmove8_64(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memmove8_64:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memmove_element_unordered_atomic_8
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i32(i8* align 8 %P, i8* align 8 %Q, i32 64, i32 8)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memmove16_64(i8* %P, i8* %Q) {
 | 
						|
; CHECK-LABEL: test_memmove16_64:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memmove_element_unordered_atomic_16
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i32(i8* align 16 %P, i8* align 16 %Q, i32 64, i32 16)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
; ==============================
 | 
						|
 | 
						|
 | 
						|
define i8* @test_memset1_64(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset1_64:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_1
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 1 %P, i8 %V, i32 64, i32 1)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset1_64_align16(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset1_64_align16:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_1
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 16 %P, i8 %V, i32 64, i32 1)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset2_64(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset2_64:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_2
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 2 %P, i8 %V, i32 64, i32 2)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
;; Use the memset4 case to explore alignment and sizing requirements in the
 | 
						|
;; lowering
 | 
						|
 | 
						|
define i8* @test_memset4_64(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset4_64:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 4 %P, i8 %V, i32 64, i32 4)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset4_64_align8(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset4_64_align8:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 8 %P, i8 %V, i32 64, i32 4)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset4_64_align16(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset4_64_align16:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 16 %P, i8 %V, i32 64, i32 4)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset4_64_align64(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset4_64_align64:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 64 %P, i8 %V, i32 64, i32 4)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset4_4(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset4_4:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $4, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 4 %P, i8 %V, i32 4, i32 4)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset4_8(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset4_8:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $8, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 4 %P, i8 %V, i32 8, i32 4)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset4_8_align8(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset4_8_align8:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $8, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 8 %P, i8 %V, i32 8, i32 4)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset4_12(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset4_12:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $12, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 4 %P, i8 %V, i32 12, i32 4)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset4_16(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset4_16:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $16, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 4 %P, i8 %V, i32 16, i32 4)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset4_16_align16(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset4_16_align16:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $16, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 16 %P, i8 %V, i32 16, i32 4)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset4_60(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset4_60:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $60, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_4
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 4 %P, i8 %V, i32 60, i32 4)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset8_64(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset8_64:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_8
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 8 %P, i8 %V, i32 64, i32 8)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset16_64(i8* %P, i8 %V) {
 | 
						|
; CHECK-LABEL: test_memset16_64:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_16
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 16 %P, i8 %V, i32 64, i32 16)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
define i8* @test_memset16_64_zero(i8* %P) {
 | 
						|
; CHECK-LABEL: test_memset16_64_zero:
 | 
						|
; CHECK:       # %bb.0:
 | 
						|
; CHECK-NEXT:    pushq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 16
 | 
						|
; CHECK-NEXT:    .cfi_offset %rbx, -16
 | 
						|
; CHECK-NEXT:    movq %rdi, %rbx
 | 
						|
; CHECK-NEXT:    xorl %esi, %esi
 | 
						|
; CHECK-NEXT:    movl $64, %edx
 | 
						|
; CHECK-NEXT:    callq __llvm_memset_element_unordered_atomic_16
 | 
						|
; CHECK-NEXT:    movq %rbx, %rax
 | 
						|
; CHECK-NEXT:    popq %rbx
 | 
						|
; CHECK-NEXT:    .cfi_def_cfa_offset 8
 | 
						|
; CHECK-NEXT:    retq
 | 
						|
  call void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* align 16 %P, i8 0, i32 64, i32 16)
 | 
						|
  ret i8* %P
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
declare void @llvm.memcpy.element.unordered.atomic.p0i8.p0i8.i32(i8* nocapture, i8* nocapture, i32, i32) nounwind
 | 
						|
declare void @llvm.memmove.element.unordered.atomic.p0i8.p0i8.i32(i8* nocapture, i8* nocapture, i32, i32) nounwind
 | 
						|
declare void @llvm.memset.element.unordered.atomic.p0i8.i32(i8* nocapture, i8, i32, i32) nounwind
 |