mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-15 20:29:48 +00:00
7c9c6ed761
Essentially the same as the GEP change in r230786. A similar migration script can be used to update test cases, though a few more test case improvements/changes were required this time around: (r229269-r229278) import fileinput import sys import re pat = re.compile(r"((?:=|:|^)\s*load (?:atomic )?(?:volatile )?(.*?))(| addrspace\(\d+\) *)\*($| *(?:%|@|null|undef|blockaddress|getelementptr|addrspacecast|bitcast|inttoptr|\[\[[a-zA-Z]|\{\{).*$)") for line in sys.stdin: sys.stdout.write(re.sub(pat, r"\1, \2\3*\4", line)) Reviewers: rafael, dexonsmith, grosser Differential Revision: http://reviews.llvm.org/D7649 git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@230794 91177308-0d34-0410-b5e6-96231b3b80d8
237 lines
9.9 KiB
LLVM
237 lines
9.9 KiB
LLVM
; RUN: llc -march=amdgcn -mcpu=bonaire -verify-machineinstrs -enable-misched -enable-aa-sched-mi < %s | FileCheck -check-prefix=FUNC -check-prefix=CI %s
|
|
|
|
declare void @llvm.SI.tbuffer.store.i32(<16 x i8>, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32)
|
|
declare void @llvm.SI.tbuffer.store.v4i32(<16 x i8>, <4 x i32>, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32, i32)
|
|
declare void @llvm.AMDGPU.barrier.local() #2
|
|
|
|
|
|
@stored_lds_ptr = addrspace(3) global i32 addrspace(3)* undef, align 4
|
|
@stored_constant_ptr = addrspace(3) global i32 addrspace(2)* undef, align 8
|
|
@stored_global_ptr = addrspace(3) global i32 addrspace(1)* undef, align 8
|
|
|
|
; FUNC-LABEL: @reorder_local_load_global_store_local_load
|
|
; CI: ds_read_b32 {{v[0-9]+}}, {{v[0-9]+}} offset:4
|
|
; CI-NEXT: ds_read_b32 {{v[0-9]+}}, {{v[0-9]+}} offset:8
|
|
; CI: buffer_store_dword
|
|
define void @reorder_local_load_global_store_local_load(i32 addrspace(1)* %out, i32 addrspace(1)* %gptr) #0 {
|
|
%ptr0 = load i32 addrspace(3)*, i32 addrspace(3)* addrspace(3)* @stored_lds_ptr, align 4
|
|
|
|
%ptr1 = getelementptr inbounds i32, i32 addrspace(3)* %ptr0, i32 1
|
|
%ptr2 = getelementptr inbounds i32, i32 addrspace(3)* %ptr0, i32 2
|
|
|
|
%tmp1 = load i32, i32 addrspace(3)* %ptr1, align 4
|
|
store i32 99, i32 addrspace(1)* %gptr, align 4
|
|
%tmp2 = load i32, i32 addrspace(3)* %ptr2, align 4
|
|
|
|
%add = add nsw i32 %tmp1, %tmp2
|
|
|
|
store i32 %add, i32 addrspace(1)* %out, align 4
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: @no_reorder_local_load_volatile_global_store_local_load
|
|
; CI: ds_read_b32 {{v[0-9]+}}, {{v[0-9]+}} offset:4
|
|
; CI: buffer_store_dword
|
|
; CI: ds_read_b32 {{v[0-9]+}}, {{v[0-9]+}} offset:8
|
|
define void @no_reorder_local_load_volatile_global_store_local_load(i32 addrspace(1)* %out, i32 addrspace(1)* %gptr) #0 {
|
|
%ptr0 = load i32 addrspace(3)*, i32 addrspace(3)* addrspace(3)* @stored_lds_ptr, align 4
|
|
|
|
%ptr1 = getelementptr inbounds i32, i32 addrspace(3)* %ptr0, i32 1
|
|
%ptr2 = getelementptr inbounds i32, i32 addrspace(3)* %ptr0, i32 2
|
|
|
|
%tmp1 = load i32, i32 addrspace(3)* %ptr1, align 4
|
|
store volatile i32 99, i32 addrspace(1)* %gptr, align 4
|
|
%tmp2 = load i32, i32 addrspace(3)* %ptr2, align 4
|
|
|
|
%add = add nsw i32 %tmp1, %tmp2
|
|
|
|
store i32 %add, i32 addrspace(1)* %out, align 4
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: @no_reorder_barrier_local_load_global_store_local_load
|
|
; CI: ds_read_b32 {{v[0-9]+}}, {{v[0-9]+}} offset:4
|
|
; CI: ds_read_b32 {{v[0-9]+}}, {{v[0-9]+}} offset:8
|
|
; CI: buffer_store_dword
|
|
define void @no_reorder_barrier_local_load_global_store_local_load(i32 addrspace(1)* %out, i32 addrspace(1)* %gptr) #0 {
|
|
%ptr0 = load i32 addrspace(3)*, i32 addrspace(3)* addrspace(3)* @stored_lds_ptr, align 4
|
|
|
|
%ptr1 = getelementptr inbounds i32, i32 addrspace(3)* %ptr0, i32 1
|
|
%ptr2 = getelementptr inbounds i32, i32 addrspace(3)* %ptr0, i32 2
|
|
|
|
%tmp1 = load i32, i32 addrspace(3)* %ptr1, align 4
|
|
store i32 99, i32 addrspace(1)* %gptr, align 4
|
|
call void @llvm.AMDGPU.barrier.local() #2
|
|
%tmp2 = load i32, i32 addrspace(3)* %ptr2, align 4
|
|
|
|
%add = add nsw i32 %tmp1, %tmp2
|
|
|
|
store i32 %add, i32 addrspace(1)* %out, align 4
|
|
ret void
|
|
}
|
|
|
|
; Technically we could reorder these, but just comparing the
|
|
; instruction type of the load is insufficient.
|
|
|
|
; FUNC-LABEL: @no_reorder_constant_load_global_store_constant_load
|
|
; CI: buffer_load_dword
|
|
; CI: buffer_store_dword
|
|
; CI: buffer_load_dword
|
|
; CI: buffer_store_dword
|
|
define void @no_reorder_constant_load_global_store_constant_load(i32 addrspace(1)* %out, i32 addrspace(1)* %gptr) #0 {
|
|
%ptr0 = load i32 addrspace(2)*, i32 addrspace(2)* addrspace(3)* @stored_constant_ptr, align 8
|
|
|
|
%ptr1 = getelementptr inbounds i32, i32 addrspace(2)* %ptr0, i64 1
|
|
%ptr2 = getelementptr inbounds i32, i32 addrspace(2)* %ptr0, i64 2
|
|
|
|
%tmp1 = load i32, i32 addrspace(2)* %ptr1, align 4
|
|
store i32 99, i32 addrspace(1)* %gptr, align 4
|
|
%tmp2 = load i32, i32 addrspace(2)* %ptr2, align 4
|
|
|
|
%add = add nsw i32 %tmp1, %tmp2
|
|
|
|
store i32 %add, i32 addrspace(1)* %out, align 4
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: @reorder_constant_load_local_store_constant_load
|
|
; CI: buffer_load_dword
|
|
; CI: buffer_load_dword
|
|
; CI: ds_write_b32
|
|
; CI: buffer_store_dword
|
|
define void @reorder_constant_load_local_store_constant_load(i32 addrspace(1)* %out, i32 addrspace(3)* %lptr) #0 {
|
|
%ptr0 = load i32 addrspace(2)*, i32 addrspace(2)* addrspace(3)* @stored_constant_ptr, align 8
|
|
|
|
%ptr1 = getelementptr inbounds i32, i32 addrspace(2)* %ptr0, i64 1
|
|
%ptr2 = getelementptr inbounds i32, i32 addrspace(2)* %ptr0, i64 2
|
|
|
|
%tmp1 = load i32, i32 addrspace(2)* %ptr1, align 4
|
|
store i32 99, i32 addrspace(3)* %lptr, align 4
|
|
%tmp2 = load i32, i32 addrspace(2)* %ptr2, align 4
|
|
|
|
%add = add nsw i32 %tmp1, %tmp2
|
|
|
|
store i32 %add, i32 addrspace(1)* %out, align 4
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: @reorder_smrd_load_local_store_smrd_load
|
|
; CI: s_load_dword
|
|
; CI: s_load_dword
|
|
; CI: s_load_dword
|
|
; CI: ds_write_b32
|
|
; CI: buffer_store_dword
|
|
define void @reorder_smrd_load_local_store_smrd_load(i32 addrspace(1)* %out, i32 addrspace(3)* noalias %lptr, i32 addrspace(2)* %ptr0) #0 {
|
|
%ptr1 = getelementptr inbounds i32, i32 addrspace(2)* %ptr0, i64 1
|
|
%ptr2 = getelementptr inbounds i32, i32 addrspace(2)* %ptr0, i64 2
|
|
|
|
%tmp1 = load i32, i32 addrspace(2)* %ptr1, align 4
|
|
store i32 99, i32 addrspace(3)* %lptr, align 4
|
|
%tmp2 = load i32, i32 addrspace(2)* %ptr2, align 4
|
|
|
|
%add = add nsw i32 %tmp1, %tmp2
|
|
|
|
store i32 %add, i32 addrspace(1)* %out, align 4
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: @reorder_global_load_local_store_global_load
|
|
; CI: buffer_load_dword
|
|
; CI: buffer_load_dword
|
|
; CI: ds_write_b32
|
|
; CI: buffer_store_dword
|
|
define void @reorder_global_load_local_store_global_load(i32 addrspace(1)* %out, i32 addrspace(3)* %lptr, i32 addrspace(1)* %ptr0) #0 {
|
|
%ptr1 = getelementptr inbounds i32, i32 addrspace(1)* %ptr0, i64 1
|
|
%ptr2 = getelementptr inbounds i32, i32 addrspace(1)* %ptr0, i64 2
|
|
|
|
%tmp1 = load i32, i32 addrspace(1)* %ptr1, align 4
|
|
store i32 99, i32 addrspace(3)* %lptr, align 4
|
|
%tmp2 = load i32, i32 addrspace(1)* %ptr2, align 4
|
|
|
|
%add = add nsw i32 %tmp1, %tmp2
|
|
|
|
store i32 %add, i32 addrspace(1)* %out, align 4
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: @reorder_local_offsets
|
|
; CI: ds_write_b32 {{v[0-9]+}}, {{v[0-9]+}} offset:12
|
|
; CI: ds_read_b32 {{v[0-9]+}}, {{v[0-9]+}} offset:400
|
|
; CI: ds_read_b32 {{v[0-9]+}}, {{v[0-9]+}} offset:404
|
|
; CI: ds_write_b32 {{v[0-9]+}}, {{v[0-9]+}} offset:400
|
|
; CI: ds_write_b32 {{v[0-9]+}}, {{v[0-9]+}} offset:404
|
|
; CI: buffer_store_dword
|
|
; CI: s_endpgm
|
|
define void @reorder_local_offsets(i32 addrspace(1)* nocapture %out, i32 addrspace(1)* noalias nocapture readnone %gptr, i32 addrspace(3)* noalias nocapture %ptr0) #0 {
|
|
%ptr1 = getelementptr inbounds i32, i32 addrspace(3)* %ptr0, i32 3
|
|
%ptr2 = getelementptr inbounds i32, i32 addrspace(3)* %ptr0, i32 100
|
|
%ptr3 = getelementptr inbounds i32, i32 addrspace(3)* %ptr0, i32 101
|
|
|
|
store i32 123, i32 addrspace(3)* %ptr1, align 4
|
|
%tmp1 = load i32, i32 addrspace(3)* %ptr2, align 4
|
|
%tmp2 = load i32, i32 addrspace(3)* %ptr3, align 4
|
|
store i32 123, i32 addrspace(3)* %ptr2, align 4
|
|
%tmp3 = load i32, i32 addrspace(3)* %ptr1, align 4
|
|
store i32 789, i32 addrspace(3)* %ptr3, align 4
|
|
|
|
%add.0 = add nsw i32 %tmp2, %tmp1
|
|
%add.1 = add nsw i32 %add.0, %tmp3
|
|
store i32 %add.1, i32 addrspace(1)* %out, align 4
|
|
ret void
|
|
}
|
|
|
|
; FUNC-LABEL: @reorder_global_offsets
|
|
; CI: buffer_store_dword {{v[0-9]+}}, {{s\[[0-9]+:[0-9]+\]}}, 0 offset:12
|
|
; CI: buffer_load_dword {{v[0-9]+}}, {{s\[[0-9]+:[0-9]+\]}}, 0 offset:400
|
|
; CI: buffer_load_dword {{v[0-9]+}}, {{s\[[0-9]+:[0-9]+\]}}, 0 offset:404
|
|
; CI: buffer_store_dword {{v[0-9]+}}, {{s\[[0-9]+:[0-9]+\]}}, 0 offset:400
|
|
; CI: buffer_store_dword {{v[0-9]+}}, {{s\[[0-9]+:[0-9]+\]}}, 0 offset:404
|
|
; CI: buffer_store_dword
|
|
; CI: s_endpgm
|
|
define void @reorder_global_offsets(i32 addrspace(1)* nocapture %out, i32 addrspace(1)* noalias nocapture readnone %gptr, i32 addrspace(1)* noalias nocapture %ptr0) #0 {
|
|
%ptr1 = getelementptr inbounds i32, i32 addrspace(1)* %ptr0, i32 3
|
|
%ptr2 = getelementptr inbounds i32, i32 addrspace(1)* %ptr0, i32 100
|
|
%ptr3 = getelementptr inbounds i32, i32 addrspace(1)* %ptr0, i32 101
|
|
|
|
store i32 123, i32 addrspace(1)* %ptr1, align 4
|
|
%tmp1 = load i32, i32 addrspace(1)* %ptr2, align 4
|
|
%tmp2 = load i32, i32 addrspace(1)* %ptr3, align 4
|
|
store i32 123, i32 addrspace(1)* %ptr2, align 4
|
|
%tmp3 = load i32, i32 addrspace(1)* %ptr1, align 4
|
|
store i32 789, i32 addrspace(1)* %ptr3, align 4
|
|
|
|
%add.0 = add nsw i32 %tmp2, %tmp1
|
|
%add.1 = add nsw i32 %add.0, %tmp3
|
|
store i32 %add.1, i32 addrspace(1)* %out, align 4
|
|
ret void
|
|
}
|
|
|
|
; XFUNC-LABEL: @reorder_local_load_tbuffer_store_local_load
|
|
; XCI: ds_read_b32 {{v[0-9]+}}, {{v[0-9]+}}, 0x4
|
|
; XCI: TBUFFER_STORE_FORMAT
|
|
; XCI: ds_read_b32 {{v[0-9]+}}, {{v[0-9]+}}, 0x8
|
|
; define void @reorder_local_load_tbuffer_store_local_load(i32 addrspace(1)* %out, i32 %a1, i32 %vaddr) #1 {
|
|
; %ptr0 = load i32 addrspace(3)*, i32 addrspace(3)* addrspace(3)* @stored_lds_ptr, align 4
|
|
|
|
; %ptr1 = getelementptr inbounds i32, i32 addrspace(3)* %ptr0, i32 1
|
|
; %ptr2 = getelementptr inbounds i32, i32 addrspace(3)* %ptr0, i32 2
|
|
|
|
; %tmp1 = load i32, i32 addrspace(3)* %ptr1, align 4
|
|
|
|
; %vdata = insertelement <4 x i32> undef, i32 %a1, i32 0
|
|
; call void @llvm.SI.tbuffer.store.v4i32(<16 x i8> undef, <4 x i32> %vdata,
|
|
; i32 4, i32 %vaddr, i32 0, i32 32, i32 14, i32 4, i32 1, i32 0, i32 1,
|
|
; i32 1, i32 0)
|
|
|
|
; %tmp2 = load i32, i32 addrspace(3)* %ptr2, align 4
|
|
|
|
; %add = add nsw i32 %tmp1, %tmp2
|
|
|
|
; store i32 %add, i32 addrspace(1)* %out, align 4
|
|
; ret void
|
|
; }
|
|
|
|
attributes #0 = { nounwind "less-precise-fpmad"="false" "no-frame-pointer-elim"="true" "no-frame-pointer-elim-non-leaf" "no-infs-fp-math"="true" "no-nans-fp-math"="true" "stack-protector-buffer-size"="8" "unsafe-fp-math"="true" "use-soft-float"="false" }
|
|
attributes #1 = { "ShaderType"="1" nounwind "less-precise-fpmad"="false" "no-frame-pointer-elim"="true" "no-frame-pointer-elim-non-leaf" "no-infs-fp-math"="true" "no-nans-fp-math"="true" "stack-protector-buffer-size"="8" "unsafe-fp-math"="true" "use-soft-float"="false" }
|
|
attributes #2 = { nounwind noduplicate }
|