mirror of
https://github.com/c64scene-ar/llvm-6502.git
synced 2024-12-17 18:31:04 +00:00
1f996fa36b
This is equivalent to the AMDGPUTargetMachine now, but it is the starting point for separating R600 and GCN functionality into separate targets. It is recommened that users start using the gcn triple for GCN-based GPUs, because using the r600 triple for these GPUs will be deprecated in the future. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@225277 91177308-0d34-0410-b5e6-96231b3b80d8
273 lines
12 KiB
LLVM
273 lines
12 KiB
LLVM
; RUN: llc -march=amdgcn -mcpu=bonaire -verify-machineinstrs -mattr=+load-store-opt -enable-misched < %s | FileCheck -check-prefix=SI %s
|
|
|
|
@lds = addrspace(3) global [512 x float] undef, align 4
|
|
@lds.f64 = addrspace(3) global [512 x double] undef, align 8
|
|
|
|
|
|
; SI-LABEL: @simple_read2st64_f32_0_1
|
|
; SI: ds_read2st64_b32 v{{\[}}[[LO_VREG:[0-9]+]]:[[HI_VREG:[0-9]+]]{{\]}}, v{{[0-9]+}} offset0:0 offset1:1
|
|
; SI: s_waitcnt lgkmcnt(0)
|
|
; SI: v_add_f32_e32 [[RESULT:v[0-9]+]], v[[HI_VREG]], v[[LO_VREG]]
|
|
; SI: buffer_store_dword [[RESULT]]
|
|
; SI: s_endpgm
|
|
define void @simple_read2st64_f32_0_1(float addrspace(1)* %out) #0 {
|
|
%x.i = tail call i32 @llvm.r600.read.tidig.x() #1
|
|
%arrayidx0 = getelementptr inbounds [512 x float] addrspace(3)* @lds, i32 0, i32 %x.i
|
|
%val0 = load float addrspace(3)* %arrayidx0, align 4
|
|
%add.x = add nsw i32 %x.i, 64
|
|
%arrayidx1 = getelementptr inbounds [512 x float] addrspace(3)* @lds, i32 0, i32 %add.x
|
|
%val1 = load float addrspace(3)* %arrayidx1, align 4
|
|
%sum = fadd float %val0, %val1
|
|
%out.gep = getelementptr inbounds float addrspace(1)* %out, i32 %x.i
|
|
store float %sum, float addrspace(1)* %out.gep, align 4
|
|
ret void
|
|
}
|
|
|
|
; SI-LABEL: @simple_read2st64_f32_1_2
|
|
; SI: ds_read2st64_b32 v{{\[}}[[LO_VREG:[0-9]+]]:[[HI_VREG:[0-9]+]]{{\]}}, v{{[0-9]+}} offset0:1 offset1:2
|
|
; SI: s_waitcnt lgkmcnt(0)
|
|
; SI: v_add_f32_e32 [[RESULT:v[0-9]+]], v[[HI_VREG]], v[[LO_VREG]]
|
|
; SI: buffer_store_dword [[RESULT]]
|
|
; SI: s_endpgm
|
|
define void @simple_read2st64_f32_1_2(float addrspace(1)* %out, float addrspace(3)* %lds) #0 {
|
|
%x.i = tail call i32 @llvm.r600.read.tidig.x() #1
|
|
%add.x.0 = add nsw i32 %x.i, 64
|
|
%arrayidx0 = getelementptr inbounds float addrspace(3)* %lds, i32 %add.x.0
|
|
%val0 = load float addrspace(3)* %arrayidx0, align 4
|
|
%add.x.1 = add nsw i32 %x.i, 128
|
|
%arrayidx1 = getelementptr inbounds float addrspace(3)* %lds, i32 %add.x.1
|
|
%val1 = load float addrspace(3)* %arrayidx1, align 4
|
|
%sum = fadd float %val0, %val1
|
|
%out.gep = getelementptr inbounds float addrspace(1)* %out, i32 %x.i
|
|
store float %sum, float addrspace(1)* %out.gep, align 4
|
|
ret void
|
|
}
|
|
|
|
; SI-LABEL: @simple_read2st64_f32_max_offset
|
|
; SI: ds_read2st64_b32 v{{\[}}[[LO_VREG:[0-9]+]]:[[HI_VREG:[0-9]+]]{{\]}}, v{{[0-9]+}} offset0:1 offset1:255
|
|
; SI: s_waitcnt lgkmcnt(0)
|
|
; SI: v_add_f32_e32 [[RESULT:v[0-9]+]], v[[HI_VREG]], v[[LO_VREG]]
|
|
; SI: buffer_store_dword [[RESULT]]
|
|
; SI: s_endpgm
|
|
define void @simple_read2st64_f32_max_offset(float addrspace(1)* %out, float addrspace(3)* %lds) #0 {
|
|
%x.i = tail call i32 @llvm.r600.read.tidig.x() #1
|
|
%add.x.0 = add nsw i32 %x.i, 64
|
|
%arrayidx0 = getelementptr inbounds float addrspace(3)* %lds, i32 %add.x.0
|
|
%val0 = load float addrspace(3)* %arrayidx0, align 4
|
|
%add.x.1 = add nsw i32 %x.i, 16320
|
|
%arrayidx1 = getelementptr inbounds float addrspace(3)* %lds, i32 %add.x.1
|
|
%val1 = load float addrspace(3)* %arrayidx1, align 4
|
|
%sum = fadd float %val0, %val1
|
|
%out.gep = getelementptr inbounds float addrspace(1)* %out, i32 %x.i
|
|
store float %sum, float addrspace(1)* %out.gep, align 4
|
|
ret void
|
|
}
|
|
|
|
; SI-LABEL: @simple_read2st64_f32_over_max_offset
|
|
; SI-NOT: ds_read2st64_b32
|
|
; SI: ds_read_b32 {{v[0-9]+}}, {{v[0-9]+}} offset:256
|
|
; SI: v_add_i32_e32 [[BIGADD:v[0-9]+]], 0x10000, {{v[0-9]+}}
|
|
; SI: ds_read_b32 {{v[0-9]+}}, [[BIGADD]]
|
|
; SI: s_endpgm
|
|
define void @simple_read2st64_f32_over_max_offset(float addrspace(1)* %out, float addrspace(3)* %lds) #0 {
|
|
%x.i = tail call i32 @llvm.r600.read.tidig.x() #1
|
|
%add.x.0 = add nsw i32 %x.i, 64
|
|
%arrayidx0 = getelementptr inbounds float addrspace(3)* %lds, i32 %add.x.0
|
|
%val0 = load float addrspace(3)* %arrayidx0, align 4
|
|
%add.x.1 = add nsw i32 %x.i, 16384
|
|
%arrayidx1 = getelementptr inbounds float addrspace(3)* %lds, i32 %add.x.1
|
|
%val1 = load float addrspace(3)* %arrayidx1, align 4
|
|
%sum = fadd float %val0, %val1
|
|
%out.gep = getelementptr inbounds float addrspace(1)* %out, i32 %x.i
|
|
store float %sum, float addrspace(1)* %out.gep, align 4
|
|
ret void
|
|
}
|
|
|
|
; SI-LABEL: @odd_invalid_read2st64_f32_0
|
|
; SI-NOT: ds_read2st64_b32
|
|
; SI: s_endpgm
|
|
define void @odd_invalid_read2st64_f32_0(float addrspace(1)* %out) #0 {
|
|
%x.i = tail call i32 @llvm.r600.read.tidig.x() #1
|
|
%arrayidx0 = getelementptr inbounds [512 x float] addrspace(3)* @lds, i32 0, i32 %x.i
|
|
%val0 = load float addrspace(3)* %arrayidx0, align 4
|
|
%add.x = add nsw i32 %x.i, 63
|
|
%arrayidx1 = getelementptr inbounds [512 x float] addrspace(3)* @lds, i32 0, i32 %add.x
|
|
%val1 = load float addrspace(3)* %arrayidx1, align 4
|
|
%sum = fadd float %val0, %val1
|
|
%out.gep = getelementptr inbounds float addrspace(1)* %out, i32 %x.i
|
|
store float %sum, float addrspace(1)* %out.gep, align 4
|
|
ret void
|
|
}
|
|
|
|
; SI-LABEL: @odd_invalid_read2st64_f32_1
|
|
; SI-NOT: ds_read2st64_b32
|
|
; SI: s_endpgm
|
|
define void @odd_invalid_read2st64_f32_1(float addrspace(1)* %out) #0 {
|
|
%x.i = tail call i32 @llvm.r600.read.tidig.x() #1
|
|
%add.x.0 = add nsw i32 %x.i, 64
|
|
%arrayidx0 = getelementptr inbounds [512 x float] addrspace(3)* @lds, i32 0, i32 %add.x.0
|
|
%val0 = load float addrspace(3)* %arrayidx0, align 4
|
|
%add.x.1 = add nsw i32 %x.i, 127
|
|
%arrayidx1 = getelementptr inbounds [512 x float] addrspace(3)* @lds, i32 0, i32 %add.x.1
|
|
%val1 = load float addrspace(3)* %arrayidx1, align 4
|
|
%sum = fadd float %val0, %val1
|
|
%out.gep = getelementptr inbounds float addrspace(1)* %out, i32 %x.i
|
|
store float %sum, float addrspace(1)* %out.gep, align 4
|
|
ret void
|
|
}
|
|
|
|
; SI-LABEL: @simple_read2st64_f64_0_1
|
|
; SI: ds_read2st64_b64 v{{\[}}[[LO_VREG:[0-9]+]]:[[HI_VREG:[0-9]+]]{{\]}}, v{{[0-9]+}} offset0:0 offset1:1
|
|
; SI: s_waitcnt lgkmcnt(0)
|
|
; SI: v_add_f64 [[RESULT:v\[[0-9]+:[0-9]+\]]], v{{\[}}[[LO_VREG]]:{{[0-9]+\]}}, v{{\[[0-9]+}}:[[HI_VREG]]{{\]}}
|
|
; SI: buffer_store_dwordx2 [[RESULT]]
|
|
; SI: s_endpgm
|
|
define void @simple_read2st64_f64_0_1(double addrspace(1)* %out) #0 {
|
|
%x.i = tail call i32 @llvm.r600.read.tidig.x() #1
|
|
%arrayidx0 = getelementptr inbounds [512 x double] addrspace(3)* @lds.f64, i32 0, i32 %x.i
|
|
%val0 = load double addrspace(3)* %arrayidx0, align 8
|
|
%add.x = add nsw i32 %x.i, 64
|
|
%arrayidx1 = getelementptr inbounds [512 x double] addrspace(3)* @lds.f64, i32 0, i32 %add.x
|
|
%val1 = load double addrspace(3)* %arrayidx1, align 8
|
|
%sum = fadd double %val0, %val1
|
|
%out.gep = getelementptr inbounds double addrspace(1)* %out, i32 %x.i
|
|
store double %sum, double addrspace(1)* %out.gep, align 8
|
|
ret void
|
|
}
|
|
|
|
; SI-LABEL: @simple_read2st64_f64_1_2
|
|
; SI: ds_read2st64_b64 v{{\[}}[[LO_VREG:[0-9]+]]:[[HI_VREG:[0-9]+]]{{\]}}, v{{[0-9]+}} offset0:1 offset1:2
|
|
; SI: s_waitcnt lgkmcnt(0)
|
|
; SI: v_add_f64 [[RESULT:v\[[0-9]+:[0-9]+\]]], v{{\[}}[[LO_VREG]]:{{[0-9]+\]}}, v{{\[[0-9]+}}:[[HI_VREG]]{{\]}}
|
|
; SI: buffer_store_dwordx2 [[RESULT]]
|
|
; SI: s_endpgm
|
|
define void @simple_read2st64_f64_1_2(double addrspace(1)* %out, double addrspace(3)* %lds) #0 {
|
|
%x.i = tail call i32 @llvm.r600.read.tidig.x() #1
|
|
%add.x.0 = add nsw i32 %x.i, 64
|
|
%arrayidx0 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x.0
|
|
%val0 = load double addrspace(3)* %arrayidx0, align 8
|
|
%add.x.1 = add nsw i32 %x.i, 128
|
|
%arrayidx1 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x.1
|
|
%val1 = load double addrspace(3)* %arrayidx1, align 8
|
|
%sum = fadd double %val0, %val1
|
|
%out.gep = getelementptr inbounds double addrspace(1)* %out, i32 %x.i
|
|
store double %sum, double addrspace(1)* %out.gep, align 8
|
|
ret void
|
|
}
|
|
|
|
; Alignment only
|
|
|
|
; SI-LABEL: @misaligned_read2st64_f64
|
|
; SI: ds_read2_b32 v{{\[[0-9]+:[0-9]+\]}}, {{v[0-9]+}} offset0:0 offset1:1
|
|
; SI: ds_read2_b32 v{{\[[0-9]+:[0-9]+\]}}, {{v[0-9]+}} offset0:128 offset1:129
|
|
; SI: s_endpgm
|
|
define void @misaligned_read2st64_f64(double addrspace(1)* %out, double addrspace(3)* %lds) #0 {
|
|
%x.i = tail call i32 @llvm.r600.read.tidig.x() #1
|
|
%arrayidx0 = getelementptr inbounds double addrspace(3)* %lds, i32 %x.i
|
|
%val0 = load double addrspace(3)* %arrayidx0, align 4
|
|
%add.x = add nsw i32 %x.i, 64
|
|
%arrayidx1 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x
|
|
%val1 = load double addrspace(3)* %arrayidx1, align 4
|
|
%sum = fadd double %val0, %val1
|
|
%out.gep = getelementptr inbounds double addrspace(1)* %out, i32 %x.i
|
|
store double %sum, double addrspace(1)* %out.gep, align 4
|
|
ret void
|
|
}
|
|
|
|
; The maximum is not the usual 0xff because 0xff * 8 * 64 > 0xffff
|
|
; SI-LABEL: @simple_read2st64_f64_max_offset
|
|
; SI: ds_read2st64_b64 v{{\[}}[[LO_VREG:[0-9]+]]:[[HI_VREG:[0-9]+]]{{\]}}, v{{[0-9]+}} offset0:4 offset1:127
|
|
; SI: s_waitcnt lgkmcnt(0)
|
|
; SI: v_add_f64 [[RESULT:v\[[0-9]+:[0-9]+\]]], v{{\[}}[[LO_VREG]]:{{[0-9]+\]}}, v{{\[[0-9]+}}:[[HI_VREG]]{{\]}}
|
|
; SI: buffer_store_dwordx2 [[RESULT]]
|
|
; SI: s_endpgm
|
|
define void @simple_read2st64_f64_max_offset(double addrspace(1)* %out, double addrspace(3)* %lds) #0 {
|
|
%x.i = tail call i32 @llvm.r600.read.tidig.x() #1
|
|
%add.x.0 = add nsw i32 %x.i, 256
|
|
%arrayidx0 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x.0
|
|
%val0 = load double addrspace(3)* %arrayidx0, align 8
|
|
%add.x.1 = add nsw i32 %x.i, 8128
|
|
%arrayidx1 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x.1
|
|
%val1 = load double addrspace(3)* %arrayidx1, align 8
|
|
%sum = fadd double %val0, %val1
|
|
%out.gep = getelementptr inbounds double addrspace(1)* %out, i32 %x.i
|
|
store double %sum, double addrspace(1)* %out.gep, align 8
|
|
ret void
|
|
}
|
|
|
|
; SI-LABEL: @simple_read2st64_f64_over_max_offset
|
|
; SI-NOT: ds_read2st64_b64
|
|
; SI: ds_read_b64 {{v\[[0-9]+:[0-9]+\]}}, {{v[0-9]+}} offset:512
|
|
; SI: v_add_i32_e32 [[BIGADD:v[0-9]+]], 0x10000, {{v[0-9]+}}
|
|
; SI: ds_read_b64 {{v\[[0-9]+:[0-9]+\]}}, [[BIGADD]]
|
|
; SI: s_endpgm
|
|
define void @simple_read2st64_f64_over_max_offset(double addrspace(1)* %out, double addrspace(3)* %lds) #0 {
|
|
%x.i = tail call i32 @llvm.r600.read.tidig.x() #1
|
|
%add.x.0 = add nsw i32 %x.i, 64
|
|
%arrayidx0 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x.0
|
|
%val0 = load double addrspace(3)* %arrayidx0, align 8
|
|
%add.x.1 = add nsw i32 %x.i, 8192
|
|
%arrayidx1 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x.1
|
|
%val1 = load double addrspace(3)* %arrayidx1, align 8
|
|
%sum = fadd double %val0, %val1
|
|
%out.gep = getelementptr inbounds double addrspace(1)* %out, i32 %x.i
|
|
store double %sum, double addrspace(1)* %out.gep, align 8
|
|
ret void
|
|
}
|
|
|
|
; SI-LABEL: @invalid_read2st64_f64_odd_offset
|
|
; SI-NOT: ds_read2st64_b64
|
|
; SI: s_endpgm
|
|
define void @invalid_read2st64_f64_odd_offset(double addrspace(1)* %out, double addrspace(3)* %lds) #0 {
|
|
%x.i = tail call i32 @llvm.r600.read.tidig.x() #1
|
|
%add.x.0 = add nsw i32 %x.i, 64
|
|
%arrayidx0 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x.0
|
|
%val0 = load double addrspace(3)* %arrayidx0, align 8
|
|
%add.x.1 = add nsw i32 %x.i, 8129
|
|
%arrayidx1 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x.1
|
|
%val1 = load double addrspace(3)* %arrayidx1, align 8
|
|
%sum = fadd double %val0, %val1
|
|
%out.gep = getelementptr inbounds double addrspace(1)* %out, i32 %x.i
|
|
store double %sum, double addrspace(1)* %out.gep, align 8
|
|
ret void
|
|
}
|
|
|
|
; The stride of 8 elements is 8 * 8 bytes. We need to make sure the
|
|
; stride in elements, not bytes, is a multiple of 64.
|
|
|
|
; SI-LABEL: @byte_size_only_divisible_64_read2_f64
|
|
; SI-NOT: ds_read2st_b64
|
|
; SI: ds_read2_b64 v{{\[[0-9]+:[0-9]+\]}}, v{{[0-9]+}} offset0:0 offset1:8
|
|
; SI: s_endpgm
|
|
define void @byte_size_only_divisible_64_read2_f64(double addrspace(1)* %out, double addrspace(3)* %lds) #0 {
|
|
%x.i = tail call i32 @llvm.r600.read.tidig.x() #1
|
|
%arrayidx0 = getelementptr inbounds double addrspace(3)* %lds, i32 %x.i
|
|
%val0 = load double addrspace(3)* %arrayidx0, align 8
|
|
%add.x = add nsw i32 %x.i, 8
|
|
%arrayidx1 = getelementptr inbounds double addrspace(3)* %lds, i32 %add.x
|
|
%val1 = load double addrspace(3)* %arrayidx1, align 8
|
|
%sum = fadd double %val0, %val1
|
|
%out.gep = getelementptr inbounds double addrspace(1)* %out, i32 %x.i
|
|
store double %sum, double addrspace(1)* %out.gep, align 4
|
|
ret void
|
|
}
|
|
|
|
; Function Attrs: nounwind readnone
|
|
declare i32 @llvm.r600.read.tgid.x() #1
|
|
|
|
; Function Attrs: nounwind readnone
|
|
declare i32 @llvm.r600.read.tgid.y() #1
|
|
|
|
; Function Attrs: nounwind readnone
|
|
declare i32 @llvm.r600.read.tidig.x() #1
|
|
|
|
; Function Attrs: nounwind readnone
|
|
declare i32 @llvm.r600.read.tidig.y() #1
|
|
|
|
; Function Attrs: noduplicate nounwind
|
|
declare void @llvm.AMDGPU.barrier.local() #2
|
|
|
|
attributes #0 = { nounwind "less-precise-fpmad"="false" "no-frame-pointer-elim"="false" "no-infs-fp-math"="false" "no-nans-fp-math"="false" "no-realign-stack" "stack-protector-buffer-size"="8" "unsafe-fp-math"="false" "use-soft-float"="false" }
|
|
attributes #1 = { nounwind readnone }
|
|
attributes #2 = { noduplicate nounwind }
|