llvm-6502/lib/Target/SystemZ
Richard Sandiford b594c4c873 [SystemZ] Rework handling of constant PC-relative operands
The GNU assembler treats things like:

        brasl   %r14, 100

in the same way as:

        brasl   %r14, .+100

rather than as a branch to absolute address 100.  We implemented this in
LLVM by creating an immediate operand rather than the usual expr operand,
and by handling immediate operands specially in the code emitter.
This was undesirable for (at least) three reasons:

- the specialness of immediate operands was exposed to the backend MC code,
  rather than being limited to the assembler parser.

- in disassembly, an immediate operand really is an absolute address.
  (Note that this means reassembling printed disassembly can't recreate
  the original code.)

- it would interfere with any assembly manipulation that we might
  try in future.  E.g. operations like branch shortening can change
  the relative position of instructions, but any code that updates
  sym+offset addresses wouldn't update an immediate "100" operand
  in the same way as an explicit ".+100" operand.

This patch changes the implementation so that the assembler creates
a "." label for immediate PC-relative operands, so that the operand
to the MCInst is always the absolute address.  The patch also adds
some error checking of the offset.


git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@181773 91177308-0d34-0410-b5e6-96231b3b80d8
2013-05-14 09:47:26 +00:00
..
AsmParser [SystemZ] Rework handling of constant PC-relative operands 2013-05-14 09:47:26 +00:00
InstPrinter
MCTargetDesc [SystemZ] Rework handling of constant PC-relative operands 2013-05-14 09:47:26 +00:00
TargetInfo
CMakeLists.txt
LLVMBuild.txt
Makefile
README.txt
SystemZ.h
SystemZ.td
SystemZAsmPrinter.cpp
SystemZAsmPrinter.h
SystemZCallingConv.cpp
SystemZCallingConv.h
SystemZCallingConv.td
SystemZConstantPoolValue.cpp
SystemZConstantPoolValue.h
SystemZFrameLowering.cpp Change getFrameMoves to return a const reference. 2013-05-11 02:38:11 +00:00
SystemZFrameLowering.h
SystemZInstrBuilder.h
SystemZInstrFormats.td [SystemZ] Match operands to fields by name rather than by order 2013-05-14 09:28:21 +00:00
SystemZInstrFP.td [SystemZ] Match operands to fields by name rather than by order 2013-05-14 09:28:21 +00:00
SystemZInstrInfo.cpp
SystemZInstrInfo.h
SystemZInstrInfo.td [SystemZ] Remove bogus isAsmParserOnly 2013-05-14 09:38:07 +00:00
SystemZISelDAGToDAG.cpp
SystemZISelLowering.cpp
SystemZISelLowering.h
SystemZMachineFunctionInfo.h
SystemZMCInstLower.cpp
SystemZMCInstLower.h
SystemZOperands.td [SystemZ] Rework handling of constant PC-relative operands 2013-05-14 09:47:26 +00:00
SystemZOperators.td
SystemZPatterns.td
SystemZRegisterInfo.cpp
SystemZRegisterInfo.h
SystemZRegisterInfo.td
SystemZSubtarget.cpp
SystemZSubtarget.h
SystemZTargetMachine.cpp Remove the MachineMove class. 2013-05-13 01:16:13 +00:00
SystemZTargetMachine.h

//===---------------------------------------------------------------------===//
// Random notes about and ideas for the SystemZ backend.
//===---------------------------------------------------------------------===//

The initial backend is deliberately restricted to z10.  We should add support
for later architectures at some point.

--

SystemZDAGToDAGISel::SelectInlineAsmMemoryOperand() is passed "m" for all
inline asm memory constraints; it doesn't get to see the original constraint.
This means that it must conservatively treat all inline asm constraints
as the most restricted type, "R".

--

If an inline asm ties an i32 "r" result to an i64 input, the input
will be treated as an i32, leaving the upper bits uninitialised.
For example:

define void @f4(i32 *%dst) {
  %val = call i32 asm "blah $0", "=r,0" (i64 103)
  store i32 %val, i32 *%dst
  ret void
}

from CodeGen/SystemZ/asm-09.ll will use LHI rather than LGHI.
to load 103.  This seems to be a general target-independent problem.

--

The tuning of the choice between Load Address (LA) and addition in
SystemZISelDAGToDAG.cpp is suspect.  It should be tweaked based on
performance measurements.

--

There is no scheduling support.

--

We don't use the Branch on Count or Branch on Index families of instruction.

--

We don't use the condition code results of anything except comparisons.

Implementing this may need something more finely grained than the z_cmp
and z_ucmp that we have now.  It might (or might not) also be useful to
have a mask of "don't care" values in conditional branches.  For example,
integer comparisons never set CC to 3, so the bottom bit of the CC mask
isn't particularly relevant.  JNLH and JE are equally good for testing
equality after an integer comparison, etc.

--

We don't optimize string and block memory operations.

--

We don't take full advantage of builtins like fabsl because the calling
conventions require f128s to be returned by invisible reference.

--

DAGCombiner can detect integer absolute, but there's not yet an associated
ISD opcode.  We could add one and implement it using Load Positive.
Negated absolutes could use Load Negative.

--

DAGCombiner doesn't yet fold truncations of extended loads.  Functions like:

    unsigned long f (unsigned long x, unsigned short *y)
    {
      return (x << 32) | *y;
    }

therefore end up as:

        sllg    %r2, %r2, 32
        llgh    %r0, 0(%r3)
        lr      %r2, %r0
        br      %r14

but truncating the load would give:

        sllg    %r2, %r2, 32
        lh      %r2, 0(%r3)
        br      %r14

--

Functions like:

define i64 @f1(i64 %a) {
  %and = and i64 %a, 1
  ret i64 %and
}

ought to be implemented as:

        lhi     %r0, 1
        ngr     %r2, %r0
        br      %r14

but two-address optimisations reverse the order of the AND and force:

        lhi     %r0, 1
        ngr     %r0, %r2
        lgr     %r2, %r0
        br      %r14

CodeGen/SystemZ/and-04.ll has several examples of this.

--

Out-of-range displacements are usually handled by loading the full
address into a register.  In many cases it would be better to create
an anchor point instead.  E.g. for:

define void @f4a(i128 *%aptr, i64 %base) {
  %addr = add i64 %base, 524288
  %bptr = inttoptr i64 %addr to i128 *
  %a = load volatile i128 *%aptr
  %b = load i128 *%bptr
  %add = add i128 %a, %b
  store i128 %add, i128 *%aptr
  ret void
}

(from CodeGen/SystemZ/int-add-08.ll) we load %base+524288 and %base+524296
into separate registers, rather than using %base+524288 as a base for both.

--

Dynamic stack allocations round the size to 8 bytes and then allocate
that rounded amount.  It would be simpler to subtract the unrounded
size from the copy of the stack pointer and then align the result.
See CodeGen/SystemZ/alloca-01.ll for an example.

--

Atomic loads and stores use the default compare-and-swap based implementation.
This is probably much too conservative in practice, and the overhead is
especially bad for 8- and 16-bit accesses.