mirror of
https://gitlab.com/freepascal.org/fpc/source.git
synced 2025-04-19 23:19:24 +02:00
Don't reuse registers on ARM 64 bits shift
git-svn-id: trunk@26669 -
This commit is contained in:
parent
cdc5e9a73a
commit
80d2ee8ad0
@ -438,7 +438,8 @@ implementation
|
||||
|
||||
function tarmshlshrnode.first_shlshr64bitint: tnode;
|
||||
begin
|
||||
if GenerateThumbCode or GenerateThumb2Code then
|
||||
if GenerateThumbCode or GenerateThumb2Code then//or
|
||||
// (right.nodetype <> ordconstn) then
|
||||
result:=inherited
|
||||
else
|
||||
result := nil;
|
||||
@ -446,70 +447,81 @@ implementation
|
||||
|
||||
procedure tarmshlshrnode.second_64bit;
|
||||
var
|
||||
hreg64hi,hreg64lo,shiftreg:Tregister;
|
||||
v : TConstExprInt;
|
||||
l1,l2,l3:Tasmlabel;
|
||||
so: tshifterop;
|
||||
lreg, resreg: TRegister64;
|
||||
|
||||
procedure emit_instr(p: tai);
|
||||
begin
|
||||
current_asmdata.CurrAsmList.concat(p);
|
||||
end;
|
||||
|
||||
{Reg1 gets shifted and moved into reg2, and is set to zero afterwards}
|
||||
procedure shift_more_than_32(reg1, reg2: TRegister; shiftval: Byte ; sm: TShiftMode);
|
||||
begin
|
||||
shifterop_reset(so); so.shiftimm:=shiftval - 32; so.shiftmode:=sm;
|
||||
emit_instr(taicpu.op_reg_reg_shifterop(A_MOV, reg2, reg1, so));
|
||||
emit_instr(taicpu.op_reg_const(A_MOV, reg1, 0));
|
||||
end;
|
||||
|
||||
procedure shift_less_than_32(reg1, reg2: TRegister; shiftval: Byte; shiftright: boolean);
|
||||
begin
|
||||
shifterop_reset(so); so.shiftimm:=shiftval;
|
||||
if shiftright then so.shiftmode:=SM_LSR else so.shiftmode:=SM_LSL;
|
||||
emit_instr(taicpu.op_reg_reg_shifterop(A_MOV, reg1, reg1, so));
|
||||
|
||||
if shiftright then so.shiftmode:=SM_LSL else so.shiftmode:=SM_LSR;
|
||||
so.shiftimm:=32-shiftval;
|
||||
emit_instr(taicpu.op_reg_reg_reg_shifterop(A_ORR, reg1, reg1, reg2, so));
|
||||
|
||||
if shiftright then so.shiftmode:=SM_LSR else so.shiftmode:=SM_LSL;
|
||||
so.shiftimm:=shiftval;
|
||||
emit_instr(taicpu.op_reg_reg_shifterop(A_MOV, reg2, reg2, so));
|
||||
end;
|
||||
|
||||
procedure shift_by_variable(reg1, reg2, shiftval: TRegister; shiftright: boolean);
|
||||
var
|
||||
shiftval2:TRegister;
|
||||
{This code is build like it gets called with sm=SM_LSR all the time, for SM_LSL dst* and src* have to be reversed}
|
||||
procedure shift_less_than_32(srchi, srclo, dsthi, dstlo: TRegister; shiftval: Byte; sm: TShiftMode);
|
||||
begin
|
||||
shifterop_reset(so);
|
||||
|
||||
so.shiftimm:=shiftval;
|
||||
so.shiftmode:=sm;
|
||||
emit_instr(taicpu.op_reg_reg_shifterop(A_MOV, dstlo, srclo, so));
|
||||
emit_instr(taicpu.op_reg_reg_shifterop(A_MOV, dsthi, srchi, so));
|
||||
|
||||
if sm = SM_LSR then so.shiftmode:=SM_LSL else so.shiftmode:=SM_LSR;
|
||||
so.shiftimm:=32-shiftval;
|
||||
emit_instr(taicpu.op_reg_reg_reg_shifterop(A_ORR, dstlo, dstlo, srchi, so));
|
||||
|
||||
end;
|
||||
|
||||
{This code is build like it gets called with sm=SM_LSR all the time, for SM_LSL dst* and src* have to be reversed
|
||||
This will generate
|
||||
mov shiftval1, shiftval
|
||||
cmp shiftval1, #64
|
||||
movcs shiftval1, #64
|
||||
rsb shiftval2, shiftval1, #32
|
||||
mov dstlo, srclo, lsr shiftval1
|
||||
mov dsthi, srchi, lsr shiftval1
|
||||
orr dstlo, srchi, lsl shiftval2
|
||||
subs shiftval2, shiftval1, #32
|
||||
movpl dstlo, srchi, lsr shiftval2
|
||||
}
|
||||
procedure shift_by_variable(srchi, srclo, dsthi, dstlo, shiftval: TRegister; sm: TShiftMode);
|
||||
var
|
||||
shiftval1,shiftval2:TRegister;
|
||||
begin
|
||||
shifterop_reset(so);
|
||||
shiftval1:=cg.getintregister(current_asmdata.CurrAsmList,OS_INT);
|
||||
shiftval2:=cg.getintregister(current_asmdata.CurrAsmList,OS_INT);
|
||||
|
||||
cg.a_load_reg_reg(current_asmdata.CurrAsmList, OS_INT, OS_INT, shiftval, shiftval1);
|
||||
|
||||
{The ARM barrel shifter only considers the lower 8 bits of a register for the shift}
|
||||
cg.a_reg_alloc(current_asmdata.CurrAsmList,NR_DEFAULTFLAGS);
|
||||
|
||||
{Do we shift more than 32 bits?}
|
||||
emit_instr(setoppostfix(taicpu.op_reg_reg_const(A_RSB, shiftval2, shiftval, 32), PF_S));
|
||||
|
||||
{This part cares for 32 bits and more}
|
||||
emit_instr(setcondition(taicpu.op_reg_reg_const(A_SUB, shiftval2, shiftval, 32), C_MI));
|
||||
if shiftright then so.shiftmode:=SM_LSR else so.shiftmode:=SM_LSL;
|
||||
so.rs:=shiftval2;
|
||||
emit_instr(setcondition(taicpu.op_reg_reg_shifterop(A_MOV, reg2, reg1, so), C_MI));
|
||||
|
||||
{Less than 32 bits}
|
||||
so.rs:=shiftval;
|
||||
emit_instr(setcondition(taicpu.op_reg_reg_shifterop(A_MOV, reg2, reg2, so), C_PL));
|
||||
if shiftright then so.shiftmode:=SM_LSL else so.shiftmode:=SM_LSR;
|
||||
so.rs:=shiftval2;
|
||||
emit_instr(setcondition(taicpu.op_reg_reg_reg_shifterop(A_ORR, reg2, reg2, reg1, so), C_PL));
|
||||
|
||||
emit_instr(taicpu.op_reg_const(A_CMP, shiftval1, 64));
|
||||
emit_instr(setcondition(taicpu.op_reg_const(A_MOV, shiftval1, 64), C_CS));
|
||||
cg.a_reg_dealloc(current_asmdata.CurrAsmList,NR_DEFAULTFLAGS);
|
||||
|
||||
{Final adjustments}
|
||||
if shiftright then so.shiftmode:=SM_LSR else so.shiftmode:=SM_LSL;
|
||||
so.rs:=shiftval;
|
||||
emit_instr(taicpu.op_reg_reg_shifterop(A_MOV, reg1, reg1, so));
|
||||
{Calculate how much the upper register needs to be shifted left}
|
||||
emit_instr(taicpu.op_reg_reg_const(A_RSB, shiftval2, shiftval1, 32));
|
||||
|
||||
so.shiftmode:=sm;
|
||||
so.rs:=shiftval1;
|
||||
|
||||
{Shift and zerofill the hi+lo register}
|
||||
emit_instr(taicpu.op_reg_reg_shifterop(A_MOV, dstlo, srclo, so));
|
||||
emit_instr(taicpu.op_reg_reg_shifterop(A_MOV, dsthi, srchi, so));
|
||||
|
||||
{Fold in the lower 32-shiftval bits}
|
||||
if sm = SM_LSR then so.shiftmode:=SM_LSL else so.shiftmode:=SM_LSR;
|
||||
so.rs:=shiftval2;
|
||||
emit_instr(taicpu.op_reg_reg_reg_shifterop(A_ORR, dstlo, dstlo, srchi, so));
|
||||
|
||||
cg.a_reg_alloc(current_asmdata.CurrAsmList,NR_DEFAULTFLAGS);
|
||||
emit_instr(setoppostfix(taicpu.op_reg_reg_const(A_SUB, shiftval2, shiftval1, 32), PF_S));
|
||||
|
||||
so.shiftmode:=sm;
|
||||
emit_instr(setcondition(taicpu.op_reg_reg_shifterop(A_MOV, dstlo, srchi, so), C_PL));
|
||||
cg.a_reg_dealloc(current_asmdata.CurrAsmList,NR_DEFAULTFLAGS);
|
||||
end;
|
||||
|
||||
begin
|
||||
@ -520,13 +532,18 @@ implementation
|
||||
end;
|
||||
|
||||
location_reset(location,LOC_REGISTER,def_cgsize(resultdef));
|
||||
location.register64.reghi:=cg.getintregister(current_asmdata.CurrAsmList,OS_INT);
|
||||
location.register64.reglo:=cg.getintregister(current_asmdata.CurrAsmList,OS_INT);
|
||||
|
||||
|
||||
{ load left operator in a register }
|
||||
hlcg.location_force_reg(current_asmdata.CurrAsmList,left.location,left.resultdef,resultdef,false);
|
||||
hreg64hi:=left.location.register64.reghi;
|
||||
hreg64lo:=left.location.register64.reglo;
|
||||
location.register64.reghi:=hreg64hi;
|
||||
location.register64.reglo:=hreg64lo;
|
||||
if not(left.location.loc in [LOC_CREGISTER,LOC_REGISTER]) or
|
||||
(left.location.size<>OS_64) then
|
||||
hlcg.location_force_reg(current_asmdata.CurrAsmList,left.location,left.resultdef,resultdef,true);
|
||||
|
||||
lreg := left.location.register64;
|
||||
resreg := location.register64;
|
||||
shifterop_reset(so);
|
||||
|
||||
{ shifting by a constant directly coded: }
|
||||
if (right.nodetype=ordconstn) then
|
||||
@ -538,8 +555,8 @@ implementation
|
||||
begin
|
||||
{Shift left by one by 2 simple 32bit additions}
|
||||
cg.a_reg_alloc(current_asmdata.CurrAsmList,NR_DEFAULTFLAGS);
|
||||
emit_instr(setoppostfix(taicpu.op_reg_reg_reg(A_ADD, hreg64lo, hreg64lo, hreg64lo), PF_S));
|
||||
emit_instr(taicpu.op_reg_reg_reg(A_ADC, hreg64hi, hreg64hi, hreg64hi));
|
||||
emit_instr(setoppostfix(taicpu.op_reg_reg_reg(A_ADD, resreg.reglo, lreg.reglo, lreg.reglo), PF_S));
|
||||
emit_instr(taicpu.op_reg_reg_reg(A_ADC, resreg.reghi, lreg.reghi, lreg.reghi));
|
||||
cg.a_reg_dealloc(current_asmdata.CurrAsmList,NR_DEFAULTFLAGS);
|
||||
end
|
||||
else
|
||||
@ -547,42 +564,41 @@ implementation
|
||||
{Shift right by first shifting hi by one and then using RRX (rotate right extended), which rotates through the carry}
|
||||
shifterop_reset(so); so.shiftmode:=SM_LSR; so.shiftimm:=1;
|
||||
cg.a_reg_alloc(current_asmdata.CurrAsmList,NR_DEFAULTFLAGS);
|
||||
emit_instr(setoppostfix(taicpu.op_reg_reg_shifterop(A_MOV, hreg64hi, hreg64hi, so), PF_S));
|
||||
emit_instr(setoppostfix(taicpu.op_reg_reg_shifterop(A_MOV, resreg.reghi, lreg.reghi, so), PF_S));
|
||||
so.shiftmode:=SM_RRX; so.shiftimm:=0; {RRX does NOT have a shift amount}
|
||||
emit_instr(taicpu.op_reg_reg_shifterop(A_MOV, hreg64lo, hreg64lo, so));
|
||||
emit_instr(taicpu.op_reg_reg_shifterop(A_MOV, resreg.reglo, lreg.reglo, so));
|
||||
cg.a_reg_dealloc(current_asmdata.CurrAsmList,NR_DEFAULTFLAGS);
|
||||
end
|
||||
{A 32bit shift just replaces a register and clears the other}
|
||||
else if v = 32 then
|
||||
{Clear one register and use the cg to generate a normal 32-bit shift}
|
||||
else if v >= 32 then
|
||||
if nodetype=shln then
|
||||
begin
|
||||
if nodetype=shln then
|
||||
emit_instr(taicpu.op_reg_const(A_MOV, hreg64hi, 0))
|
||||
else
|
||||
emit_instr(taicpu.op_reg_const(A_MOV, hreg64lo, 0));
|
||||
location.register64.reghi:=hreg64lo;
|
||||
location.register64.reglo:=hreg64hi;
|
||||
emit_instr(taicpu.op_reg_const(A_MOV, resreg.reglo, 0));
|
||||
cg.a_op_const_reg_reg(current_asmdata.CurrAsmList,OP_SHL,OS_32,v.uvalue-32,lreg.reglo,resreg.reghi);
|
||||
end
|
||||
{Shift LESS than 32}
|
||||
else
|
||||
begin
|
||||
emit_instr(taicpu.op_reg_const(A_MOV, resreg.reghi, 0));
|
||||
cg.a_op_const_reg_reg(current_asmdata.CurrAsmList,OP_SHR,OS_32,v.uvalue-32,lreg.reghi,resreg.reglo);
|
||||
end
|
||||
{Shift LESS than 32, thats the tricky one}
|
||||
else if (v < 32) and (v > 1) then
|
||||
if nodetype=shln then
|
||||
shift_less_than_32(hreg64hi, hreg64lo, v.uvalue, false)
|
||||
shift_less_than_32(lreg.reglo, lreg.reghi, resreg.reglo, resreg.reghi, v.uvalue, SM_LSL)
|
||||
else
|
||||
shift_less_than_32(hreg64lo, hreg64hi, v.uvalue, true)
|
||||
{More than 32}
|
||||
else if v > 32 then
|
||||
if nodetype=shln then
|
||||
shift_more_than_32(hreg64lo, hreg64hi, v.uvalue, SM_LSL)
|
||||
else
|
||||
shift_more_than_32(hreg64hi, hreg64lo, v.uvalue, SM_LSR);
|
||||
shift_less_than_32(lreg.reghi, lreg.reglo, resreg.reghi, resreg.reglo, v.uvalue, SM_LSR);
|
||||
end
|
||||
else
|
||||
begin
|
||||
{ force right operators in a register }
|
||||
hlcg.location_force_reg(current_asmdata.CurrAsmList,right.location,right.resultdef,resultdef,false);
|
||||
{ force right operator into a register }
|
||||
if not(right.location.loc in [LOC_CREGISTER,LOC_REGISTER]) or
|
||||
(right.location.size<>OS_32) then
|
||||
hlcg.location_force_reg(current_asmdata.CurrAsmList,right.location,right.resultdef,u32inttype,true);
|
||||
|
||||
if nodetype = shln then
|
||||
shift_by_variable(hreg64lo,hreg64hi,right.location.register, false)
|
||||
shift_by_variable(lreg.reglo, lreg.reghi, resreg.reglo, resreg.reghi, right.location.register, SM_LSL)
|
||||
else
|
||||
shift_by_variable(hreg64hi,hreg64lo,right.location.register, true);
|
||||
shift_by_variable(lreg.reghi, lreg.reglo, resreg.reghi, resreg.reglo, right.location.register, SM_LSR);
|
||||
end;
|
||||
end;
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user