120 lines
		
	
	
		
			3.5 KiB
		
	
	
	
		
			LLVM
		
	
	
	
			
		
		
	
	
			120 lines
		
	
	
		
			3.5 KiB
		
	
	
	
		
			LLVM
		
	
	
	
| ; RUN: opt -basicaa -print-memoryssa -verify-memoryssa -analyze < %s 2>&1 | FileCheck %s
 | |
| ; RUN: opt -aa-pipeline=basic-aa -passes='print<memoryssa>,verify<memoryssa>' -disable-output < %s 2>&1 | FileCheck %s
 | |
| ;
 | |
| ; Ensures that atomic loads count as MemoryDefs
 | |
| 
 | |
| ; CHECK-LABEL: define i32 @foo
 | |
| define i32 @foo(i32* %a, i32* %b) {
 | |
| ; CHECK: 1 = MemoryDef(liveOnEntry)
 | |
| ; CHECK-NEXT: store i32 4
 | |
|   store i32 4, i32* %a, align 4
 | |
| ; CHECK: 2 = MemoryDef(1)
 | |
| ; CHECK-NEXT: %1 = load atomic i32
 | |
|   %1 = load atomic i32, i32* %b acquire, align 4
 | |
| ; CHECK: MemoryUse(2)
 | |
| ; CHECK-NEXT: %2 = load i32
 | |
|   %2 = load i32, i32* %a, align 4
 | |
|   %3 = add i32 %1, %2
 | |
|   ret i32 %3
 | |
| }
 | |
| 
 | |
| ; CHECK-LABEL: define void @bar
 | |
| define void @bar(i32* %a) {
 | |
| ; CHECK: MemoryUse(liveOnEntry)
 | |
| ; CHECK-NEXT: load atomic i32, i32* %a unordered, align 4
 | |
|   load atomic i32, i32* %a unordered, align 4
 | |
| ; CHECK: 1 = MemoryDef(liveOnEntry)
 | |
| ; CHECK-NEXT: load atomic i32, i32* %a monotonic, align 4
 | |
|   load atomic i32, i32* %a monotonic, align 4
 | |
| ; CHECK: 2 = MemoryDef(1)
 | |
| ; CHECK-NEXT: load atomic i32, i32* %a acquire, align 4
 | |
|   load atomic i32, i32* %a acquire, align 4
 | |
| ; CHECK: 3 = MemoryDef(2)
 | |
| ; CHECK-NEXT: load atomic i32, i32* %a seq_cst, align 4
 | |
|   load atomic i32, i32* %a seq_cst, align 4
 | |
|   ret void
 | |
| }
 | |
| 
 | |
| ; CHECK-LABEL: define void @baz
 | |
| define void @baz(i32* %a) {
 | |
| ; CHECK: 1 = MemoryDef(liveOnEntry)
 | |
| ; CHECK-NEXT: %1 = load atomic i32
 | |
|   %1 = load atomic i32, i32* %a acquire, align 4
 | |
| ; CHECK: MemoryUse(1)
 | |
| ; CHECK-NEXT: %2 = load atomic i32, i32* %a unordered, align 4
 | |
|   %2 = load atomic i32, i32* %a unordered, align 4
 | |
| ; CHECK: 2 = MemoryDef(1)
 | |
| ; CHECK-NEXT: %3 = load atomic i32, i32* %a monotonic, align 4
 | |
|   %3 = load atomic i32, i32* %a monotonic, align 4
 | |
|   ret void
 | |
| }
 | |
| 
 | |
| ; CHECK-LABEL: define void @fences
 | |
| define void @fences(i32* %a) {
 | |
| ; CHECK: 1 = MemoryDef(liveOnEntry)
 | |
| ; CHECK-NEXT: fence acquire
 | |
|   fence acquire
 | |
| ; CHECK: MemoryUse(1)
 | |
| ; CHECK-NEXT: %1 = load i32, i32* %a
 | |
|   %1 = load i32, i32* %a
 | |
| 
 | |
| ; CHECK: 2 = MemoryDef(1)
 | |
| ; CHECK-NEXT: fence release
 | |
|   fence release
 | |
| ; CHECK: MemoryUse(2)
 | |
| ; CHECK-NEXT: %2 = load i32, i32* %a
 | |
|   %2 = load i32, i32* %a
 | |
| 
 | |
| ; CHECK: 3 = MemoryDef(2)
 | |
| ; CHECK-NEXT: fence acq_rel
 | |
|   fence acq_rel
 | |
| ; CHECK: MemoryUse(3)
 | |
| ; CHECK-NEXT: %3 = load i32, i32* %a
 | |
|   %3 = load i32, i32* %a
 | |
| 
 | |
| ; CHECK: 4 = MemoryDef(3)
 | |
| ; CHECK-NEXT: fence seq_cst
 | |
|   fence seq_cst
 | |
| ; CHECK: MemoryUse(4)
 | |
| ; CHECK-NEXT: %4 = load i32, i32* %a
 | |
|   %4 = load i32, i32* %a
 | |
|   ret void
 | |
| }
 | |
| 
 | |
| ; CHECK-LABEL: define void @seq_cst_clobber
 | |
| define void @seq_cst_clobber(i32* noalias %a, i32* noalias %b) {
 | |
| ; CHECK: 1 = MemoryDef(liveOnEntry)
 | |
| ; CHECK-NEXT: %1 = load atomic i32, i32* %a monotonic, align 4
 | |
|   load atomic i32, i32* %a monotonic, align 4
 | |
| 
 | |
| ; CHECK: 2 = MemoryDef(1)
 | |
| ; CHECK-NEXT: %2 = load atomic i32, i32* %a seq_cst, align 4
 | |
|   load atomic i32, i32* %a seq_cst, align 4
 | |
| 
 | |
| ; CHECK: 3 = MemoryDef(2)
 | |
| ; CHECK-NEXT: load atomic i32, i32* %a monotonic, align 4
 | |
|   load atomic i32, i32* %a monotonic, align 4
 | |
| 
 | |
|   ret void
 | |
| }
 | |
| 
 | |
| ; Ensure that AA hands us MRI_Mod on unreorderable atomic ops.
 | |
| ;
 | |
| ; This test is a bit implementation-specific. In particular, it depends on that
 | |
| ; we pass cmpxchg-load queries to AA, without trying to reason about them on
 | |
| ; our own.
 | |
| ;
 | |
| ; If AA gets more aggressive, we can find another way.
 | |
| ;
 | |
| ; CHECK-LABEL: define void @check_aa_is_sane
 | |
| define void @check_aa_is_sane(i32* noalias %a, i32* noalias %b) {
 | |
| ; CHECK: 1 = MemoryDef(liveOnEntry)
 | |
| ; CHECK-NEXT: cmpxchg i32* %a, i32 0, i32 1 acquire acquire
 | |
|   cmpxchg i32* %a, i32 0, i32 1 acquire acquire
 | |
| ; CHECK: MemoryUse(1)
 | |
| ; CHECK-NEXT: load i32, i32* %b, align 4
 | |
|   load i32, i32* %b, align 4
 | |
| 
 | |
|   ret void
 | |
| }
 |