@@ -5088,7 +5088,7 @@ define amdgpu_kernel void @atomic_cmpxchg_i64_offset(ptr %out, i64 %in, i64 %old
5088
5088
; GFX12-NEXT: s_endpgm
5089
5089
entry:
5090
5090
%gep = getelementptr i64 , ptr %out , i64 4
5091
- %val = cmpxchg volatile ptr %gep , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst
5091
+ %val = cmpxchg volatile ptr %gep , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst , !noalias.addrspace !0
5092
5092
ret void
5093
5093
}
5094
5094
@@ -5145,7 +5145,7 @@ define amdgpu_kernel void @atomic_cmpxchg_i64_soffset(ptr %out, i64 %in, i64 %ol
5145
5145
; GFX12-NEXT: s_endpgm
5146
5146
entry:
5147
5147
%gep = getelementptr i64 , ptr %out , i64 9000
5148
- %val = cmpxchg volatile ptr %gep , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst
5148
+ %val = cmpxchg volatile ptr %gep , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst , !noalias.addrspace !0
5149
5149
ret void
5150
5150
}
5151
5151
@@ -5206,7 +5206,7 @@ define amdgpu_kernel void @atomic_cmpxchg_i64_ret_offset(ptr %out, ptr %out2, i6
5206
5206
; GFX12-NEXT: s_endpgm
5207
5207
entry:
5208
5208
%gep = getelementptr i64 , ptr %out , i64 4
5209
- %val = cmpxchg volatile ptr %gep , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst
5209
+ %val = cmpxchg volatile ptr %gep , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst , !noalias.addrspace !0
5210
5210
%extract0 = extractvalue { i64 , i1 } %val , 0
5211
5211
store i64 %extract0 , ptr %out2
5212
5212
ret void
@@ -5270,7 +5270,7 @@ define amdgpu_kernel void @atomic_cmpxchg_i64_addr64_offset(ptr %out, i64 %in, i
5270
5270
entry:
5271
5271
%ptr = getelementptr i64 , ptr %out , i64 %index
5272
5272
%gep = getelementptr i64 , ptr %ptr , i64 4
5273
- %val = cmpxchg volatile ptr %gep , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst
5273
+ %val = cmpxchg volatile ptr %gep , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst , !noalias.addrspace !0
5274
5274
ret void
5275
5275
}
5276
5276
@@ -5344,7 +5344,7 @@ define amdgpu_kernel void @atomic_cmpxchg_i64_ret_addr64_offset(ptr %out, ptr %o
5344
5344
entry:
5345
5345
%ptr = getelementptr i64 , ptr %out , i64 %index
5346
5346
%gep = getelementptr i64 , ptr %ptr , i64 4
5347
- %val = cmpxchg volatile ptr %gep , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst
5347
+ %val = cmpxchg volatile ptr %gep , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst , !noalias.addrspace !0
5348
5348
%extract0 = extractvalue { i64 , i1 } %val , 0
5349
5349
store i64 %extract0 , ptr %out2
5350
5350
ret void
@@ -5398,7 +5398,7 @@ define amdgpu_kernel void @atomic_cmpxchg_i64(ptr %out, i64 %in, i64 %old) {
5398
5398
; GFX12-NEXT: global_inv scope:SCOPE_DEV
5399
5399
; GFX12-NEXT: s_endpgm
5400
5400
entry:
5401
- %val = cmpxchg volatile ptr %out , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst
5401
+ %val = cmpxchg volatile ptr %out , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst , !noalias.addrspace !0
5402
5402
ret void
5403
5403
}
5404
5404
@@ -5454,7 +5454,7 @@ define amdgpu_kernel void @atomic_cmpxchg_i64_ret(ptr %out, ptr %out2, i64 %in,
5454
5454
; GFX12-NEXT: flat_store_b64 v[2:3], v[0:1]
5455
5455
; GFX12-NEXT: s_endpgm
5456
5456
entry:
5457
- %val = cmpxchg volatile ptr %out , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst
5457
+ %val = cmpxchg volatile ptr %out , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst , !noalias.addrspace !0
5458
5458
%extract0 = extractvalue { i64 , i1 } %val , 0
5459
5459
store i64 %extract0 , ptr %out2
5460
5460
ret void
@@ -5513,7 +5513,7 @@ define amdgpu_kernel void @atomic_cmpxchg_i64_addr64(ptr %out, i64 %in, i64 %ind
5513
5513
; GFX12-NEXT: s_endpgm
5514
5514
entry:
5515
5515
%ptr = getelementptr i64 , ptr %out , i64 %index
5516
- %val = cmpxchg volatile ptr %ptr , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst
5516
+ %val = cmpxchg volatile ptr %ptr , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst , !noalias.addrspace !0
5517
5517
ret void
5518
5518
}
5519
5519
@@ -5582,7 +5582,7 @@ define amdgpu_kernel void @atomic_cmpxchg_i64_ret_addr64(ptr %out, ptr %out2, i6
5582
5582
; GFX12-NEXT: s_endpgm
5583
5583
entry:
5584
5584
%ptr = getelementptr i64 , ptr %out , i64 %index
5585
- %val = cmpxchg volatile ptr %ptr , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst
5585
+ %val = cmpxchg volatile ptr %ptr , i64 %old , i64 %in syncscope("agent" ) seq_cst seq_cst , !noalias.addrspace !0
5586
5586
%extract0 = extractvalue { i64 , i1 } %val , 0
5587
5587
store i64 %extract0 , ptr %out2
5588
5588
ret void
@@ -5634,7 +5634,7 @@ define amdgpu_kernel void @atomic_load_f64_offset(ptr %in, ptr %out) {
5634
5634
; GFX12-NEXT: s_endpgm
5635
5635
entry:
5636
5636
%gep = getelementptr double , ptr %in , i64 4
5637
- %val = load atomic double , ptr %gep seq_cst , align 8
5637
+ %val = load atomic double , ptr %gep seq_cst , align 8 , !noalias.addrspace !0
5638
5638
store double %val , ptr %out
5639
5639
ret void
5640
5640
}
@@ -5680,7 +5680,7 @@ define amdgpu_kernel void @atomic_load_f64(ptr %in, ptr %out) {
5680
5680
; GFX12-NEXT: flat_store_b64 v[2:3], v[0:1]
5681
5681
; GFX12-NEXT: s_endpgm
5682
5682
entry:
5683
- %val = load atomic double , ptr %in syncscope("agent" ) seq_cst , align 8
5683
+ %val = load atomic double , ptr %in syncscope("agent" ) seq_cst , align 8 , !noalias.addrspace !0
5684
5684
store double %val , ptr %out
5685
5685
ret void
5686
5686
}
@@ -5745,7 +5745,7 @@ define amdgpu_kernel void @atomic_load_f64_addr64_offset(ptr %in, ptr %out, i64
5745
5745
entry:
5746
5746
%ptr = getelementptr double , ptr %in , i64 %index
5747
5747
%gep = getelementptr double , ptr %ptr , i64 4
5748
- %val = load atomic double , ptr %gep seq_cst , align 8
5748
+ %val = load atomic double , ptr %gep seq_cst , align 8 , !noalias.addrspace !0
5749
5749
store double %val , ptr %out
5750
5750
ret void
5751
5751
}
@@ -5805,7 +5805,7 @@ define amdgpu_kernel void @atomic_load_f64_addr64(ptr %in, ptr %out, i64 %index)
5805
5805
; GFX12-NEXT: s_endpgm
5806
5806
entry:
5807
5807
%ptr = getelementptr double , ptr %in , i64 %index
5808
- %val = load atomic double , ptr %ptr seq_cst , align 8
5808
+ %val = load atomic double , ptr %ptr seq_cst , align 8 , !noalias.addrspace !0
5809
5809
store double %val , ptr %out
5810
5810
ret void
5811
5811
}
@@ -5848,7 +5848,7 @@ define amdgpu_kernel void @atomic_store_f64_offset(double %in, ptr %out) {
5848
5848
; GFX12-NEXT: s_endpgm
5849
5849
entry:
5850
5850
%gep = getelementptr double , ptr %out , i64 4
5851
- store atomic double %in , ptr %gep seq_cst , align 8
5851
+ store atomic double %in , ptr %gep seq_cst , align 8 , !noalias.addrspace !0
5852
5852
ret void
5853
5853
}
5854
5854
@@ -5885,7 +5885,7 @@ define amdgpu_kernel void @atomic_store_f64(double %in, ptr %out) {
5885
5885
; GFX12-NEXT: flat_store_b64 v[2:3], v[0:1] scope:SCOPE_SYS
5886
5886
; GFX12-NEXT: s_endpgm
5887
5887
entry:
5888
- store atomic double %in , ptr %out seq_cst , align 8
5888
+ store atomic double %in , ptr %out seq_cst , align 8 , !noalias.addrspace !0
5889
5889
ret void
5890
5890
}
5891
5891
@@ -5941,7 +5941,7 @@ define amdgpu_kernel void @atomic_store_f64_addr64_offset(double %in, ptr %out,
5941
5941
entry:
5942
5942
%ptr = getelementptr double , ptr %out , i64 %index
5943
5943
%gep = getelementptr double , ptr %ptr , i64 4
5944
- store atomic double %in , ptr %gep seq_cst , align 8
5944
+ store atomic double %in , ptr %gep seq_cst , align 8 , !noalias.addrspace !0
5945
5945
ret void
5946
5946
}
5947
5947
@@ -5992,7 +5992,7 @@ define amdgpu_kernel void @atomic_store_f64_addr64(double %in, ptr %out, i64 %in
5992
5992
; GFX12-NEXT: s_endpgm
5993
5993
entry:
5994
5994
%ptr = getelementptr double , ptr %out , i64 %index
5995
- store atomic double %in , ptr %ptr seq_cst , align 8
5995
+ store atomic double %in , ptr %ptr seq_cst , align 8 , !noalias.addrspace !0
5996
5996
ret void
5997
5997
}
5998
5998
0 commit comments