530 lines
		
	
	
		
			21 KiB
		
	
	
	
		
			LLVM
		
	
	
	
			
		
		
	
	
			530 lines
		
	
	
		
			21 KiB
		
	
	
	
		
			LLVM
		
	
	
	
; NOTE: Assertions have been autogenerated by utils/update_test_checks.py UTC_ARGS: --function-signature
 | 
						|
; RUN: opt < %s -S -early-cse -earlycse-debug-hash | FileCheck %s --check-prefixes=CHECK,NO_ASSUME
 | 
						|
; RUN: opt < %s -S -early-cse --enable-knowledge-retention | FileCheck %s --check-prefixes=CHECK,USE_ASSUME
 | 
						|
; RUN: opt < %s -S -passes=early-cse | FileCheck %s --check-prefixes=CHECK,NO_ASSUME
 | 
						|
 | 
						|
declare {}* @llvm.invariant.start.p0i8(i64, i8* nocapture) nounwind readonly
 | 
						|
declare void @llvm.invariant.end.p0i8({}*, i64, i8* nocapture) nounwind
 | 
						|
 | 
						|
; Check that we do load-load forwarding over invariant.start, since it does not
 | 
						|
; clobber memory
 | 
						|
define i8 @test_bypass1(i8 *%P) {
 | 
						|
; NO_ASSUME-LABEL: define {{[^@]+}}@test_bypass1
 | 
						|
; NO_ASSUME-SAME: (i8* [[P:%.*]])
 | 
						|
; NO_ASSUME-NEXT:    [[V1:%.*]] = load i8, i8* [[P]], align 1
 | 
						|
; NO_ASSUME-NEXT:    [[I:%.*]] = call {}* @llvm.invariant.start.p0i8(i64 1, i8* [[P]])
 | 
						|
; NO_ASSUME-NEXT:    ret i8 0
 | 
						|
;
 | 
						|
; USE_ASSUME-LABEL: define {{[^@]+}}@test_bypass1
 | 
						|
; USE_ASSUME-SAME: (i8* [[P:%.*]])
 | 
						|
; USE_ASSUME-NEXT:    [[V1:%.*]] = load i8, i8* [[P]], align 1
 | 
						|
; USE_ASSUME-NEXT:    [[I:%.*]] = call {}* @llvm.invariant.start.p0i8(i64 1, i8* [[P]])
 | 
						|
; USE_ASSUME-NEXT:    call void @llvm.assume(i1 true) [ "dereferenceable"(i8* [[P]], i64 1), "nonnull"(i8* [[P]]) ]
 | 
						|
; USE_ASSUME-NEXT:    ret i8 0
 | 
						|
;
 | 
						|
 | 
						|
  %V1 = load i8, i8* %P
 | 
						|
  %i = call {}* @llvm.invariant.start.p0i8(i64 1, i8* %P)
 | 
						|
  %V2 = load i8, i8* %P
 | 
						|
  %Diff = sub i8 %V1, %V2
 | 
						|
  ret i8 %Diff
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
; Trivial Store->load forwarding over invariant.start
 | 
						|
define i8 @test_bypass2(i8 *%P) {
 | 
						|
; NO_ASSUME-LABEL: define {{[^@]+}}@test_bypass2
 | 
						|
; NO_ASSUME-SAME: (i8* [[P:%.*]])
 | 
						|
; NO_ASSUME-NEXT:    store i8 42, i8* [[P]], align 1
 | 
						|
; NO_ASSUME-NEXT:    [[I:%.*]] = call {}* @llvm.invariant.start.p0i8(i64 1, i8* [[P]])
 | 
						|
; NO_ASSUME-NEXT:    ret i8 42
 | 
						|
;
 | 
						|
; USE_ASSUME-LABEL: define {{[^@]+}}@test_bypass2
 | 
						|
; USE_ASSUME-SAME: (i8* [[P:%.*]])
 | 
						|
; USE_ASSUME-NEXT:    store i8 42, i8* [[P]], align 1
 | 
						|
; USE_ASSUME-NEXT:    [[I:%.*]] = call {}* @llvm.invariant.start.p0i8(i64 1, i8* [[P]])
 | 
						|
; USE_ASSUME-NEXT:    call void @llvm.assume(i1 true) [ "dereferenceable"(i8* [[P]], i64 1), "nonnull"(i8* [[P]]) ]
 | 
						|
; USE_ASSUME-NEXT:    ret i8 42
 | 
						|
;
 | 
						|
 | 
						|
  store i8 42, i8* %P
 | 
						|
  %i = call {}* @llvm.invariant.start.p0i8(i64 1, i8* %P)
 | 
						|
  %V1 = load i8, i8* %P
 | 
						|
  ret i8 %V1
 | 
						|
}
 | 
						|
 | 
						|
; We can DSE over invariant.start calls, since the first store to
 | 
						|
; %P is valid, and the second store is actually unreachable based on semantics
 | 
						|
; of invariant.start.
 | 
						|
define void @test_bypass3(i8* %P) {
 | 
						|
; NO_ASSUME-LABEL: define {{[^@]+}}@test_bypass3
 | 
						|
; NO_ASSUME-SAME: (i8* [[P:%.*]])
 | 
						|
; NO_ASSUME-NEXT:    [[I:%.*]] = call {}* @llvm.invariant.start.p0i8(i64 1, i8* [[P]])
 | 
						|
; NO_ASSUME-NEXT:    store i8 60, i8* [[P]], align 1
 | 
						|
; NO_ASSUME-NEXT:    ret void
 | 
						|
;
 | 
						|
; USE_ASSUME-LABEL: define {{[^@]+}}@test_bypass3
 | 
						|
; USE_ASSUME-SAME: (i8* [[P:%.*]])
 | 
						|
; USE_ASSUME-NEXT:    [[I:%.*]] = call {}* @llvm.invariant.start.p0i8(i64 1, i8* [[P]])
 | 
						|
; USE_ASSUME-NEXT:    call void @llvm.assume(i1 true) [ "dereferenceable"(i8* [[P]], i64 1), "nonnull"(i8* [[P]]) ]
 | 
						|
; USE_ASSUME-NEXT:    store i8 60, i8* [[P]], align 1
 | 
						|
; USE_ASSUME-NEXT:    ret void
 | 
						|
;
 | 
						|
 | 
						|
  store i8 50, i8* %P
 | 
						|
  %i = call {}* @llvm.invariant.start.p0i8(i64 1, i8* %P)
 | 
						|
  store i8 60, i8* %P
 | 
						|
  ret void
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
; FIXME: Now the first store can actually be eliminated, since there is no read within
 | 
						|
; the invariant region, between start and end.
 | 
						|
define void @test_bypass4(i8* %P) {
 | 
						|
; CHECK-LABEL: define {{[^@]+}}@test_bypass4
 | 
						|
; CHECK-SAME: (i8* [[P:%.*]])
 | 
						|
; CHECK-NEXT:    store i8 50, i8* [[P]], align 1
 | 
						|
; CHECK-NEXT:    [[I:%.*]] = call {}* @llvm.invariant.start.p0i8(i64 1, i8* [[P]])
 | 
						|
; CHECK-NEXT:    call void @llvm.invariant.end.p0i8({}* [[I]], i64 1, i8* [[P]])
 | 
						|
; CHECK-NEXT:    store i8 60, i8* [[P]], align 1
 | 
						|
; CHECK-NEXT:    ret void
 | 
						|
;
 | 
						|
 | 
						|
 | 
						|
  store i8 50, i8* %P
 | 
						|
  %i = call {}* @llvm.invariant.start.p0i8(i64 1, i8* %P)
 | 
						|
  call void @llvm.invariant.end.p0i8({}* %i, i64 1, i8* %P)
 | 
						|
  store i8 60, i8* %P
 | 
						|
  ret void
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
declare void @clobber()
 | 
						|
declare {}* @llvm.invariant.start.p0i32(i64 %size, i32* nocapture %ptr)
 | 
						|
declare void @llvm.invariant.end.p0i32({}*, i64, i32* nocapture) nounwind
 | 
						|
 | 
						|
define i32 @test_before_load(i32* %p) {
 | 
						|
; NO_ASSUME-LABEL: define {{[^@]+}}@test_before_load
 | 
						|
; NO_ASSUME-SAME: (i32* [[P:%.*]])
 | 
						|
; NO_ASSUME-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; NO_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; NO_ASSUME-NEXT:    call void @clobber()
 | 
						|
; NO_ASSUME-NEXT:    ret i32 0
 | 
						|
;
 | 
						|
; USE_ASSUME-LABEL: define {{[^@]+}}@test_before_load
 | 
						|
; USE_ASSUME-SAME: (i32* [[P:%.*]])
 | 
						|
; USE_ASSUME-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; USE_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; USE_ASSUME-NEXT:    call void @clobber()
 | 
						|
; USE_ASSUME-NEXT:    call void @llvm.assume(i1 true) [ "dereferenceable"(i32* [[P]], i64 4), "nonnull"(i32* [[P]]), "align"(i32* [[P]], i64 4) ]
 | 
						|
; USE_ASSUME-NEXT:    ret i32 0
 | 
						|
;
 | 
						|
  call {}* @llvm.invariant.start.p0i32(i64 4, i32* %p)
 | 
						|
  %v1 = load i32, i32* %p
 | 
						|
  call void @clobber()
 | 
						|
  %v2 = load i32, i32* %p
 | 
						|
  %sub = sub i32 %v1, %v2
 | 
						|
  ret i32 %sub
 | 
						|
}
 | 
						|
 | 
						|
define i32 @test_before_clobber(i32* %p) {
 | 
						|
; NO_ASSUME-LABEL: define {{[^@]+}}@test_before_clobber
 | 
						|
; NO_ASSUME-SAME: (i32* [[P:%.*]])
 | 
						|
; NO_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; NO_ASSUME-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; NO_ASSUME-NEXT:    call void @clobber()
 | 
						|
; NO_ASSUME-NEXT:    ret i32 0
 | 
						|
;
 | 
						|
; USE_ASSUME-LABEL: define {{[^@]+}}@test_before_clobber
 | 
						|
; USE_ASSUME-SAME: (i32* [[P:%.*]])
 | 
						|
; USE_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; USE_ASSUME-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; USE_ASSUME-NEXT:    call void @clobber()
 | 
						|
; USE_ASSUME-NEXT:    call void @llvm.assume(i1 true) [ "dereferenceable"(i32* [[P]], i64 4), "nonnull"(i32* [[P]]), "align"(i32* [[P]], i64 4) ]
 | 
						|
; USE_ASSUME-NEXT:    ret i32 0
 | 
						|
;
 | 
						|
  %v1 = load i32, i32* %p
 | 
						|
  call {}* @llvm.invariant.start.p0i32(i64 4, i32* %p)
 | 
						|
  call void @clobber()
 | 
						|
  %v2 = load i32, i32* %p
 | 
						|
  %sub = sub i32 %v1, %v2
 | 
						|
  ret i32 %sub
 | 
						|
}
 | 
						|
 | 
						|
define i32 @test_duplicate_scope(i32* %p) {
 | 
						|
; NO_ASSUME-LABEL: define {{[^@]+}}@test_duplicate_scope
 | 
						|
; NO_ASSUME-SAME: (i32* [[P:%.*]])
 | 
						|
; NO_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; NO_ASSUME-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; NO_ASSUME-NEXT:    call void @clobber()
 | 
						|
; NO_ASSUME-NEXT:    [[TMP2:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; NO_ASSUME-NEXT:    ret i32 0
 | 
						|
;
 | 
						|
; USE_ASSUME-LABEL: define {{[^@]+}}@test_duplicate_scope
 | 
						|
; USE_ASSUME-SAME: (i32* [[P:%.*]])
 | 
						|
; USE_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; USE_ASSUME-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; USE_ASSUME-NEXT:    call void @clobber()
 | 
						|
; USE_ASSUME-NEXT:    [[TMP2:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; USE_ASSUME-NEXT:    call void @llvm.assume(i1 true) [ "dereferenceable"(i32* [[P]], i64 4), "nonnull"(i32* [[P]]), "align"(i32* [[P]], i64 4) ]
 | 
						|
; USE_ASSUME-NEXT:    ret i32 0
 | 
						|
;
 | 
						|
  %v1 = load i32, i32* %p
 | 
						|
  call {}* @llvm.invariant.start.p0i32(i64 4, i32* %p)
 | 
						|
  call void @clobber()
 | 
						|
  call {}* @llvm.invariant.start.p0i32(i64 4, i32* %p)
 | 
						|
  %v2 = load i32, i32* %p
 | 
						|
  %sub = sub i32 %v1, %v2
 | 
						|
  ret i32 %sub
 | 
						|
}
 | 
						|
 | 
						|
define i32 @test_unanalzyable_load(i32* %p) {
 | 
						|
; NO_ASSUME-LABEL: define {{[^@]+}}@test_unanalzyable_load
 | 
						|
; NO_ASSUME-SAME: (i32* [[P:%.*]])
 | 
						|
; NO_ASSUME-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; NO_ASSUME-NEXT:    call void @clobber()
 | 
						|
; NO_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; NO_ASSUME-NEXT:    call void @clobber()
 | 
						|
; NO_ASSUME-NEXT:    ret i32 0
 | 
						|
;
 | 
						|
; USE_ASSUME-LABEL: define {{[^@]+}}@test_unanalzyable_load
 | 
						|
; USE_ASSUME-SAME: (i32* [[P:%.*]])
 | 
						|
; USE_ASSUME-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; USE_ASSUME-NEXT:    call void @clobber()
 | 
						|
; USE_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; USE_ASSUME-NEXT:    call void @clobber()
 | 
						|
; USE_ASSUME-NEXT:    call void @llvm.assume(i1 true) [ "dereferenceable"(i32* [[P]], i64 4), "nonnull"(i32* [[P]]), "align"(i32* [[P]], i64 4) ]
 | 
						|
; USE_ASSUME-NEXT:    ret i32 0
 | 
						|
;
 | 
						|
  call {}* @llvm.invariant.start.p0i32(i64 4, i32* %p)
 | 
						|
  call void @clobber()
 | 
						|
  %v1 = load i32, i32* %p
 | 
						|
  call void @clobber()
 | 
						|
  %v2 = load i32, i32* %p
 | 
						|
  %sub = sub i32 %v1, %v2
 | 
						|
  ret i32 %sub
 | 
						|
}
 | 
						|
 | 
						|
define i32 @test_negative_after_clobber(i32* %p) {
 | 
						|
; CHECK-LABEL: define {{[^@]+}}@test_negative_after_clobber
 | 
						|
; CHECK-SAME: (i32* [[P:%.*]])
 | 
						|
; CHECK-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; CHECK-NEXT:    call void @clobber()
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; CHECK-NEXT:    [[V2:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; CHECK-NEXT:    [[SUB:%.*]] = sub i32 [[V1]], [[V2]]
 | 
						|
; CHECK-NEXT:    ret i32 [[SUB]]
 | 
						|
;
 | 
						|
  %v1 = load i32, i32* %p
 | 
						|
  call void @clobber()
 | 
						|
  call {}* @llvm.invariant.start.p0i32(i64 4, i32* %p)
 | 
						|
  %v2 = load i32, i32* %p
 | 
						|
  %sub = sub i32 %v1, %v2
 | 
						|
  ret i32 %sub
 | 
						|
}
 | 
						|
 | 
						|
define i32 @test_merge(i32* %p, i1 %cnd) {
 | 
						|
; NO_ASSUME-LABEL: define {{[^@]+}}@test_merge
 | 
						|
; NO_ASSUME-SAME: (i32* [[P:%.*]], i1 [[CND:%.*]])
 | 
						|
; NO_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; NO_ASSUME-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; NO_ASSUME-NEXT:    br i1 [[CND]], label [[MERGE:%.*]], label [[TAKEN:%.*]]
 | 
						|
; NO_ASSUME:       taken:
 | 
						|
; NO_ASSUME-NEXT:    call void @clobber()
 | 
						|
; NO_ASSUME-NEXT:    br label [[MERGE]]
 | 
						|
; NO_ASSUME:       merge:
 | 
						|
; NO_ASSUME-NEXT:    ret i32 0
 | 
						|
;
 | 
						|
; USE_ASSUME-LABEL: define {{[^@]+}}@test_merge
 | 
						|
; USE_ASSUME-SAME: (i32* [[P:%.*]], i1 [[CND:%.*]])
 | 
						|
; USE_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; USE_ASSUME-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; USE_ASSUME-NEXT:    br i1 [[CND]], label [[MERGE:%.*]], label [[TAKEN:%.*]]
 | 
						|
; USE_ASSUME:       taken:
 | 
						|
; USE_ASSUME-NEXT:    call void @clobber()
 | 
						|
; USE_ASSUME-NEXT:    br label [[MERGE]]
 | 
						|
; USE_ASSUME:       merge:
 | 
						|
; USE_ASSUME-NEXT:    call void @llvm.assume(i1 true) [ "dereferenceable"(i32* [[P]], i64 4), "nonnull"(i32* [[P]]), "align"(i32* [[P]], i64 4) ]
 | 
						|
; USE_ASSUME-NEXT:    ret i32 0
 | 
						|
;
 | 
						|
  %v1 = load i32, i32* %p
 | 
						|
  call {}* @llvm.invariant.start.p0i32(i64 4, i32* %p)
 | 
						|
  br i1 %cnd, label %merge, label %taken
 | 
						|
 | 
						|
taken:
 | 
						|
  call void @clobber()
 | 
						|
  br label %merge
 | 
						|
merge:
 | 
						|
  %v2 = load i32, i32* %p
 | 
						|
  %sub = sub i32 %v1, %v2
 | 
						|
  ret i32 %sub
 | 
						|
}
 | 
						|
 | 
						|
define i32 @test_negative_after_mergeclobber(i32* %p, i1 %cnd) {
 | 
						|
; CHECK-LABEL: define {{[^@]+}}@test_negative_after_mergeclobber
 | 
						|
; CHECK-SAME: (i32* [[P:%.*]], i1 [[CND:%.*]])
 | 
						|
; CHECK-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; CHECK-NEXT:    br i1 [[CND]], label [[MERGE:%.*]], label [[TAKEN:%.*]]
 | 
						|
; CHECK:       taken:
 | 
						|
; CHECK-NEXT:    call void @clobber()
 | 
						|
; CHECK-NEXT:    br label [[MERGE]]
 | 
						|
; CHECK:       merge:
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; CHECK-NEXT:    [[V2:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; CHECK-NEXT:    [[SUB:%.*]] = sub i32 [[V1]], [[V2]]
 | 
						|
; CHECK-NEXT:    ret i32 [[SUB]]
 | 
						|
;
 | 
						|
  %v1 = load i32, i32* %p
 | 
						|
  br i1 %cnd, label %merge, label %taken
 | 
						|
 | 
						|
taken:
 | 
						|
  call void @clobber()
 | 
						|
  br label %merge
 | 
						|
merge:
 | 
						|
  call {}* @llvm.invariant.start.p0i32(i64 4, i32* %p)
 | 
						|
  %v2 = load i32, i32* %p
 | 
						|
  %sub = sub i32 %v1, %v2
 | 
						|
  ret i32 %sub
 | 
						|
}
 | 
						|
 | 
						|
; In theory, this version could work, but earlycse is incapable of
 | 
						|
; merging facts along distinct paths.
 | 
						|
define i32 @test_false_negative_merge(i32* %p, i1 %cnd) {
 | 
						|
; CHECK-LABEL: define {{[^@]+}}@test_false_negative_merge
 | 
						|
; CHECK-SAME: (i32* [[P:%.*]], i1 [[CND:%.*]])
 | 
						|
; CHECK-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; CHECK-NEXT:    br i1 [[CND]], label [[MERGE:%.*]], label [[TAKEN:%.*]]
 | 
						|
; CHECK:       taken:
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; CHECK-NEXT:    call void @clobber()
 | 
						|
; CHECK-NEXT:    br label [[MERGE]]
 | 
						|
; CHECK:       merge:
 | 
						|
; CHECK-NEXT:    [[V2:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; CHECK-NEXT:    [[SUB:%.*]] = sub i32 [[V1]], [[V2]]
 | 
						|
; CHECK-NEXT:    ret i32 [[SUB]]
 | 
						|
;
 | 
						|
  %v1 = load i32, i32* %p
 | 
						|
  br i1 %cnd, label %merge, label %taken
 | 
						|
 | 
						|
taken:
 | 
						|
  call {}* @llvm.invariant.start.p0i32(i64 4, i32* %p)
 | 
						|
  call void @clobber()
 | 
						|
  br label %merge
 | 
						|
merge:
 | 
						|
  %v2 = load i32, i32* %p
 | 
						|
  %sub = sub i32 %v1, %v2
 | 
						|
  ret i32 %sub
 | 
						|
}
 | 
						|
 | 
						|
define i32 @test_merge_unanalyzable_load(i32* %p, i1 %cnd) {
 | 
						|
; NO_ASSUME-LABEL: define {{[^@]+}}@test_merge_unanalyzable_load
 | 
						|
; NO_ASSUME-SAME: (i32* [[P:%.*]], i1 [[CND:%.*]])
 | 
						|
; NO_ASSUME-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; NO_ASSUME-NEXT:    call void @clobber()
 | 
						|
; NO_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; NO_ASSUME-NEXT:    br i1 [[CND]], label [[MERGE:%.*]], label [[TAKEN:%.*]]
 | 
						|
; NO_ASSUME:       taken:
 | 
						|
; NO_ASSUME-NEXT:    call void @clobber()
 | 
						|
; NO_ASSUME-NEXT:    br label [[MERGE]]
 | 
						|
; NO_ASSUME:       merge:
 | 
						|
; NO_ASSUME-NEXT:    ret i32 0
 | 
						|
;
 | 
						|
; USE_ASSUME-LABEL: define {{[^@]+}}@test_merge_unanalyzable_load
 | 
						|
; USE_ASSUME-SAME: (i32* [[P:%.*]], i1 [[CND:%.*]])
 | 
						|
; USE_ASSUME-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; USE_ASSUME-NEXT:    call void @clobber()
 | 
						|
; USE_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; USE_ASSUME-NEXT:    br i1 [[CND]], label [[MERGE:%.*]], label [[TAKEN:%.*]]
 | 
						|
; USE_ASSUME:       taken:
 | 
						|
; USE_ASSUME-NEXT:    call void @clobber()
 | 
						|
; USE_ASSUME-NEXT:    br label [[MERGE]]
 | 
						|
; USE_ASSUME:       merge:
 | 
						|
; USE_ASSUME-NEXT:    call void @llvm.assume(i1 true) [ "dereferenceable"(i32* [[P]], i64 4), "nonnull"(i32* [[P]]), "align"(i32* [[P]], i64 4) ]
 | 
						|
; USE_ASSUME-NEXT:    ret i32 0
 | 
						|
;
 | 
						|
  call {}* @llvm.invariant.start.p0i32(i64 4, i32* %p)
 | 
						|
  call void @clobber()
 | 
						|
  %v1 = load i32, i32* %p
 | 
						|
  br i1 %cnd, label %merge, label %taken
 | 
						|
 | 
						|
taken:
 | 
						|
  call void @clobber()
 | 
						|
  br label %merge
 | 
						|
merge:
 | 
						|
  %v2 = load i32, i32* %p
 | 
						|
  %sub = sub i32 %v1, %v2
 | 
						|
  ret i32 %sub
 | 
						|
}
 | 
						|
 | 
						|
define void @test_dse_before_load(i32* %p, i1 %cnd) {
 | 
						|
; NO_ASSUME-LABEL: define {{[^@]+}}@test_dse_before_load
 | 
						|
; NO_ASSUME-SAME: (i32* [[P:%.*]], i1 [[CND:%.*]])
 | 
						|
; NO_ASSUME-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; NO_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; NO_ASSUME-NEXT:    call void @clobber()
 | 
						|
; NO_ASSUME-NEXT:    ret void
 | 
						|
;
 | 
						|
; USE_ASSUME-LABEL: define {{[^@]+}}@test_dse_before_load
 | 
						|
; USE_ASSUME-SAME: (i32* [[P:%.*]], i1 [[CND:%.*]])
 | 
						|
; USE_ASSUME-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; USE_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; USE_ASSUME-NEXT:    call void @clobber()
 | 
						|
; USE_ASSUME-NEXT:    call void @llvm.assume(i1 true) [ "dereferenceable"(i32* [[P]], i64 4), "nonnull"(i32* [[P]]), "align"(i32* [[P]], i64 4) ]
 | 
						|
; USE_ASSUME-NEXT:    ret void
 | 
						|
;
 | 
						|
  call {}* @llvm.invariant.start.p0i32(i64 4, i32* %p)
 | 
						|
  %v1 = load i32, i32* %p
 | 
						|
  call void @clobber()
 | 
						|
  store i32 %v1, i32* %p
 | 
						|
  ret void
 | 
						|
}
 | 
						|
 | 
						|
define void @test_dse_after_load(i32* %p, i1 %cnd) {
 | 
						|
; NO_ASSUME-LABEL: define {{[^@]+}}@test_dse_after_load
 | 
						|
; NO_ASSUME-SAME: (i32* [[P:%.*]], i1 [[CND:%.*]])
 | 
						|
; NO_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; NO_ASSUME-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; NO_ASSUME-NEXT:    call void @clobber()
 | 
						|
; NO_ASSUME-NEXT:    ret void
 | 
						|
;
 | 
						|
; USE_ASSUME-LABEL: define {{[^@]+}}@test_dse_after_load
 | 
						|
; USE_ASSUME-SAME: (i32* [[P:%.*]], i1 [[CND:%.*]])
 | 
						|
; USE_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; USE_ASSUME-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; USE_ASSUME-NEXT:    call void @clobber()
 | 
						|
; USE_ASSUME-NEXT:    call void @llvm.assume(i1 true) [ "dereferenceable"(i32* [[P]], i64 4), "nonnull"(i32* [[P]]), "align"(i32* [[P]], i64 4) ]
 | 
						|
; USE_ASSUME-NEXT:    ret void
 | 
						|
;
 | 
						|
  %v1 = load i32, i32* %p
 | 
						|
  call {}* @llvm.invariant.start.p0i32(i64 4, i32* %p)
 | 
						|
  call void @clobber()
 | 
						|
  store i32 %v1, i32* %p
 | 
						|
  ret void
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
; In this case, we have a false negative since MemoryLocation is implicitly
 | 
						|
; typed due to the user of a Value to represent the address.  Note that other
 | 
						|
; passes will canonicalize away the bitcasts in this example.
 | 
						|
define i32 @test_false_negative_types(i32* %p) {
 | 
						|
; CHECK-LABEL: define {{[^@]+}}@test_false_negative_types
 | 
						|
; CHECK-SAME: (i32* [[P:%.*]])
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; CHECK-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; CHECK-NEXT:    call void @clobber()
 | 
						|
; CHECK-NEXT:    [[PF:%.*]] = bitcast i32* [[P]] to float*
 | 
						|
; CHECK-NEXT:    [[V2F:%.*]] = load float, float* [[PF]], align 4
 | 
						|
; CHECK-NEXT:    [[V2:%.*]] = bitcast float [[V2F]] to i32
 | 
						|
; CHECK-NEXT:    [[SUB:%.*]] = sub i32 [[V1]], [[V2]]
 | 
						|
; CHECK-NEXT:    ret i32 [[SUB]]
 | 
						|
;
 | 
						|
  call {}* @llvm.invariant.start.p0i32(i64 4, i32* %p)
 | 
						|
  %v1 = load i32, i32* %p
 | 
						|
  call void @clobber()
 | 
						|
  %pf = bitcast i32* %p to float*
 | 
						|
  %v2f = load float, float* %pf
 | 
						|
  %v2 = bitcast float %v2f to i32
 | 
						|
  %sub = sub i32 %v1, %v2
 | 
						|
  ret i32 %sub
 | 
						|
}
 | 
						|
 | 
						|
define i32 @test_negative_size1(i32* %p) {
 | 
						|
; CHECK-LABEL: define {{[^@]+}}@test_negative_size1
 | 
						|
; CHECK-SAME: (i32* [[P:%.*]])
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 3, i32* [[P]])
 | 
						|
; CHECK-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; CHECK-NEXT:    call void @clobber()
 | 
						|
; CHECK-NEXT:    [[V2:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; CHECK-NEXT:    [[SUB:%.*]] = sub i32 [[V1]], [[V2]]
 | 
						|
; CHECK-NEXT:    ret i32 [[SUB]]
 | 
						|
;
 | 
						|
  call {}* @llvm.invariant.start.p0i32(i64 3, i32* %p)
 | 
						|
  %v1 = load i32, i32* %p
 | 
						|
  call void @clobber()
 | 
						|
  %v2 = load i32, i32* %p
 | 
						|
  %sub = sub i32 %v1, %v2
 | 
						|
  ret i32 %sub
 | 
						|
}
 | 
						|
 | 
						|
define i32 @test_negative_size2(i32* %p) {
 | 
						|
; CHECK-LABEL: define {{[^@]+}}@test_negative_size2
 | 
						|
; CHECK-SAME: (i32* [[P:%.*]])
 | 
						|
; CHECK-NEXT:    [[TMP1:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 0, i32* [[P]])
 | 
						|
; CHECK-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; CHECK-NEXT:    call void @clobber()
 | 
						|
; CHECK-NEXT:    [[V2:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; CHECK-NEXT:    [[SUB:%.*]] = sub i32 [[V1]], [[V2]]
 | 
						|
; CHECK-NEXT:    ret i32 [[SUB]]
 | 
						|
;
 | 
						|
  call {}* @llvm.invariant.start.p0i32(i64 0, i32* %p)
 | 
						|
  %v1 = load i32, i32* %p
 | 
						|
  call void @clobber()
 | 
						|
  %v2 = load i32, i32* %p
 | 
						|
  %sub = sub i32 %v1, %v2
 | 
						|
  ret i32 %sub
 | 
						|
}
 | 
						|
 | 
						|
define i32 @test_negative_scope(i32* %p) {
 | 
						|
; CHECK-LABEL: define {{[^@]+}}@test_negative_scope
 | 
						|
; CHECK-SAME: (i32* [[P:%.*]])
 | 
						|
; CHECK-NEXT:    [[SCOPE:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; CHECK-NEXT:    call void @llvm.invariant.end.p0i32({}* [[SCOPE]], i64 4, i32* [[P]])
 | 
						|
; CHECK-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; CHECK-NEXT:    call void @clobber()
 | 
						|
; CHECK-NEXT:    [[V2:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; CHECK-NEXT:    [[SUB:%.*]] = sub i32 [[V1]], [[V2]]
 | 
						|
; CHECK-NEXT:    ret i32 [[SUB]]
 | 
						|
;
 | 
						|
  %scope = call {}* @llvm.invariant.start.p0i32(i64 4, i32* %p)
 | 
						|
  call void @llvm.invariant.end.p0i32({}* %scope, i64 4, i32* %p)
 | 
						|
  %v1 = load i32, i32* %p
 | 
						|
  call void @clobber()
 | 
						|
  %v2 = load i32, i32* %p
 | 
						|
  %sub = sub i32 %v1, %v2
 | 
						|
  ret i32 %sub
 | 
						|
}
 | 
						|
 | 
						|
define i32 @test_false_negative_scope(i32* %p) {
 | 
						|
; CHECK-LABEL: define {{[^@]+}}@test_false_negative_scope
 | 
						|
; CHECK-SAME: (i32* [[P:%.*]])
 | 
						|
; CHECK-NEXT:    [[SCOPE:%.*]] = call {}* @llvm.invariant.start.p0i32(i64 4, i32* [[P]])
 | 
						|
; CHECK-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; CHECK-NEXT:    call void @clobber()
 | 
						|
; CHECK-NEXT:    [[V2:%.*]] = load i32, i32* [[P]], align 4
 | 
						|
; CHECK-NEXT:    call void @llvm.invariant.end.p0i32({}* [[SCOPE]], i64 4, i32* [[P]])
 | 
						|
; CHECK-NEXT:    [[SUB:%.*]] = sub i32 [[V1]], [[V2]]
 | 
						|
; CHECK-NEXT:    ret i32 [[SUB]]
 | 
						|
;
 | 
						|
  %scope = call {}* @llvm.invariant.start.p0i32(i64 4, i32* %p)
 | 
						|
  %v1 = load i32, i32* %p
 | 
						|
  call void @clobber()
 | 
						|
  %v2 = load i32, i32* %p
 | 
						|
  call void @llvm.invariant.end.p0i32({}* %scope, i64 4, i32* %p)
 | 
						|
  %sub = sub i32 %v1, %v2
 | 
						|
  ret i32 %sub
 | 
						|
}
 | 
						|
 | 
						|
; Invariant load defact starts an invariant.start scope of the appropriate size
 | 
						|
define i32 @test_invariant_load_scope(i32* %p) {
 | 
						|
; NO_ASSUME-LABEL: define {{[^@]+}}@test_invariant_load_scope
 | 
						|
; NO_ASSUME-SAME: (i32* [[P:%.*]])
 | 
						|
; NO_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4, !invariant.load !0
 | 
						|
; NO_ASSUME-NEXT:    call void @clobber()
 | 
						|
; NO_ASSUME-NEXT:    ret i32 0
 | 
						|
;
 | 
						|
; USE_ASSUME-LABEL: define {{[^@]+}}@test_invariant_load_scope
 | 
						|
; USE_ASSUME-SAME: (i32* [[P:%.*]])
 | 
						|
; USE_ASSUME-NEXT:    [[V1:%.*]] = load i32, i32* [[P]], align 4, !invariant.load !0
 | 
						|
; USE_ASSUME-NEXT:    call void @clobber()
 | 
						|
; USE_ASSUME-NEXT:    call void @llvm.assume(i1 true) [ "dereferenceable"(i32* [[P]], i64 4), "nonnull"(i32* [[P]]), "align"(i32* [[P]], i64 4) ]
 | 
						|
; USE_ASSUME-NEXT:    ret i32 0
 | 
						|
;
 | 
						|
  %v1 = load i32, i32* %p, !invariant.load !{}
 | 
						|
  call void @clobber()
 | 
						|
  %v2 = load i32, i32* %p
 | 
						|
  %sub = sub i32 %v1, %v2
 | 
						|
  ret i32 %sub
 | 
						|
}
 | 
						|
 | 
						|
; USE_ASSUME: declare void @llvm.assume(i1 noundef)
 |