fpc/compiler/i8086/cgcpu.pas

3301 lines
139 KiB
ObjectPascal

{
Copyright (c) 1998-2002 by Florian Klaempfl
This unit implements the code generator for the i8086
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 2 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program; if not, write to the Free Software
Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
****************************************************************************
}
unit cgcpu;
{$i fpcdefs.inc}
interface
uses
globtype,
cgbase,cgobj,cg64f32,cgx86,
aasmbase,aasmtai,aasmdata,aasmcpu,
cpubase,parabase,cgutils,
symconst,symdef
;
type
{ tcg8086 }
tcg8086 = class(tcgx86)
procedure init_register_allocators;override;
procedure do_register_allocation(list:TAsmList;headertai:tai);override;
procedure a_call_name(list : TAsmList;const s : string; weak: boolean);override;
procedure a_call_name_far(list : TAsmList;const s : string; weak: boolean);
procedure a_call_name_static(list : TAsmList;const s : string);override;
procedure a_call_name_static_far(list : TAsmList;const s : string);
procedure a_op_const_reg(list : TAsmList; Op: TOpCG; size: TCGSize; a: tcgint; reg: TRegister); override;
procedure a_op_const_ref(list : TAsmList; Op: TOpCG; size: TCGSize; a: tcgint; const ref: TReference); override;
procedure a_op_reg_reg(list : TAsmList; Op: TOpCG; size: TCGSize; src, dst: TRegister); override;
procedure a_op_ref_reg(list : TAsmList; Op: TOpCG; size: TCGSize; const ref: TReference; reg: TRegister); override;
procedure a_op_reg_ref(list : TAsmList; Op: TOpCG; size: TCGSize;reg: TRegister; const ref: TReference); override;
procedure a_op_ref(list : TAsmList; Op: TOpCG; size: TCGSize; const ref: TReference); override;
procedure push_const(list:TAsmList;size:tcgsize;a:tcgint);
{ passing parameter using push instead of mov }
procedure a_load_reg_cgpara(list : TAsmList;size : tcgsize;r : tregister;const cgpara : tcgpara);override;
procedure a_load_const_cgpara(list : TAsmList;size : tcgsize;a : tcgint;const cgpara : tcgpara);override;
procedure a_load_ref_cgpara(list : TAsmList;size : tcgsize;const r : treference;const cgpara : tcgpara);override;
procedure a_loadaddr_ref_cgpara(list : TAsmList;const r : treference;const cgpara : tcgpara);override;
{ move instructions }
procedure a_load_const_reg(list : TAsmList; tosize: tcgsize; a : tcgint;reg : tregister);override;
procedure a_load_const_ref(list : TAsmList; tosize: tcgsize; a : tcgint;const ref : treference);override;
procedure a_load_reg_ref(list : TAsmList;fromsize,tosize: tcgsize; reg : tregister;const ref : treference);override;
{ use a_load_ref_reg_internal() instead }
//procedure a_load_ref_reg(list : TAsmList;fromsize,tosize: tcgsize;const ref : treference;reg : tregister);override;
procedure a_load_reg_reg(list : TAsmList;fromsize,tosize: tcgsize;reg1,reg2 : tregister);override;
{ comparison operations }
procedure a_cmp_const_reg_label(list : TAsmList;size : tcgsize;cmp_op : topcmp;a : tcgint;reg : tregister;
l : tasmlabel);override;
procedure a_cmp_const_ref_label(list : TAsmList;size : tcgsize;cmp_op : topcmp;a : tcgint;const ref : treference;
l : tasmlabel);override;
procedure a_cmp_reg_reg_label(list : TAsmList;size : tcgsize;cmp_op : topcmp;reg1,reg2 : tregister;l : tasmlabel); override;
procedure a_cmp_ref_reg_label(list : TAsmList;size : tcgsize;cmp_op : topcmp;const ref: treference; reg : tregister; l : tasmlabel); override;
procedure a_cmp_reg_ref_label(list : TAsmList;size : tcgsize;cmp_op : topcmp;reg : tregister; const ref: treference; l : tasmlabel); override;
procedure gen_cmp32_jmp1(list: TAsmList; cmp_op: topcmp; l_skip, l_target: TAsmLabel);
procedure gen_cmp32_jmp2(list: TAsmList; cmp_op: topcmp; l_skip, l_target: TAsmLabel);
procedure g_flags2reg(list: TAsmList; size: TCgSize; const f: tresflags; reg: TRegister);override;
procedure g_flags2ref(list: TAsmList; size: TCgSize; const f: tresflags; const ref: TReference);override;
procedure g_stackpointer_alloc(list : TAsmList;localsize: longint);override;
procedure g_proc_exit(list : TAsmList;parasize:longint;nostackframe:boolean);override;
procedure g_copyvaluepara_openarray(list : TAsmList;const ref:treference;const lenloc:tlocation;elesize:tcgint;destreg:tregister);
procedure g_releasevaluepara_openarray(list : TAsmList;const l:tlocation);
procedure g_adjust_self_value(list:TAsmList;procdef: tprocdef;ioffset: tcgint);override;
procedure get_32bit_ops(op: TOpCG; out op1,op2: TAsmOp);
protected
procedure a_load_ref_reg_internal(list : TAsmList;fromsize,tosize: tcgsize;const ref : treference;reg : tregister;isdirect:boolean);override;
end;
tcg64f8086 = class(tcg64f32)
procedure a_op64_ref_reg(list : TAsmList;op:TOpCG;size : tcgsize;const ref : treference;reg : tregister64);override;
procedure a_op64_reg_ref(list : TAsmList;op:TOpCG;size : tcgsize;reg : tregister64; const ref: treference);override;
procedure a_op64_reg_reg(list : TAsmList;op:TOpCG;size : tcgsize;regsrc,regdst : tregister64);override;
procedure a_op64_const_reg(list : TAsmList;op:TOpCG;size : tcgsize;value : int64;reg : tregister64);override;
procedure a_op64_const_ref(list : TAsmList;op:TOpCG;size : tcgsize;value : int64;const ref : treference);override;
procedure a_op64_ref(list : TAsmList;op:TOpCG;size : tcgsize;const ref : treference);override;
private
procedure get_64bit_ops(op:TOpCG;var op1,op2:TAsmOp);
end;
procedure create_codegen;
implementation
uses
globals,verbose,systems,cutils,
paramgr,procinfo,fmodule,
rgcpu,rgx86,cpuinfo,
symtype,symsym,symcpu,
tgobj,
hlcgobj;
{ Range check must be disabled explicitly as the code uses
implicit typecast to aint troughout }
{$R-}
function use_push(const cgpara:tcgpara):boolean;
begin
result:=(not paramanager.use_fixed_stack) and
assigned(cgpara.location) and
(cgpara.location^.loc=LOC_REFERENCE) and
(cgpara.location^.reference.index=NR_STACK_POINTER_REG);
end;
procedure tcg8086.init_register_allocators;
begin
inherited init_register_allocators;
if cs_create_pic in current_settings.moduleswitches then
rg[R_INTREGISTER]:=trgintcpu.create(R_INTREGISTER,R_SUBWHOLE,[RS_AX,RS_DX,RS_CX,RS_SI,RS_DI],first_int_imreg,[RS_BP])
else
if (cs_useebp in current_settings.optimizerswitches) and assigned(current_procinfo) and (current_procinfo.framepointer<>NR_BP) then
rg[R_INTREGISTER]:=trgintcpu.create(R_INTREGISTER,R_SUBWHOLE,[RS_AX,RS_DX,RS_CX,RS_BX,RS_SI,RS_DI,RS_BP],first_int_imreg,[])
else
rg[R_INTREGISTER]:=trgintcpu.create(R_INTREGISTER,R_SUBWHOLE,[RS_AX,RS_DX,RS_CX,RS_BX,RS_SI,RS_DI],first_int_imreg,[RS_BP]);
rg[R_MMXREGISTER]:=trgcpu.create(R_MMXREGISTER,R_SUBNONE,[RS_XMM0,RS_XMM1,RS_XMM2,RS_XMM3,RS_XMM4,RS_XMM5,RS_XMM6,RS_XMM7],first_mm_imreg,[]);
rg[R_MMREGISTER]:=trgcpu.create(R_MMREGISTER,R_SUBWHOLE,[RS_XMM0,RS_XMM1,RS_XMM2,RS_XMM3,RS_XMM4,RS_XMM5,RS_XMM6,RS_XMM7],first_mm_imreg,[]);
rgfpu:=Trgx86fpu.create;
end;
procedure tcg8086.do_register_allocation(list:TAsmList;headertai:tai);
begin
if (tf_pic_uses_got in target_info.flags) and (pi_needs_got in current_procinfo.flags) then
begin
if getsupreg(current_procinfo.got) < first_int_imreg then
include(rg[R_INTREGISTER].used_in_proc,getsupreg(current_procinfo.got));
end;
inherited do_register_allocation(list,headertai);
end;
procedure tcg8086.a_call_name(list: TAsmList; const s: string; weak: boolean);
begin
if current_settings.x86memorymodel in x86_far_code_models then
a_call_name_far(list,s,weak)
else
a_call_name_near(list,s,weak);
end;
procedure tcg8086.a_call_name_far(list: TAsmList; const s: string;
weak: boolean);
var
sym : tasmsymbol;
begin
if not(weak) then
sym:=current_asmdata.RefAsmSymbol(s,AT_FUNCTION)
else
sym:=current_asmdata.WeakRefAsmSymbol(s,AT_FUNCTION);
list.concat(taicpu.op_sym(A_CALL,S_FAR,sym));
end;
procedure tcg8086.a_call_name_static(list: TAsmList; const s: string);
begin
if current_settings.x86memorymodel in x86_far_code_models then
a_call_name_static_far(list,s)
else
a_call_name_static_near(list,s);
end;
procedure tcg8086.a_call_name_static_far(list: TAsmList; const s: string);
var
sym : tasmsymbol;
begin
sym:=current_asmdata.RefAsmSymbol(s,AT_FUNCTION);
list.concat(taicpu.op_sym(A_CALL,S_FAR,sym));
end;
procedure tcg8086.a_op_const_reg(list: TAsmList; Op: TOpCG; size: TCGSize;
a: tcgint; reg: TRegister);
type
trox32method=(rm_unspecified,rm_unrolledleftloop,rm_unrolledrightloop,
rm_loopleft,rm_loopright,rm_fast_386);
var
tmpreg: tregister;
op1, op2: TAsmOp;
ax_subreg: tregister;
hl_loop_start: tasmlabel;
ai: taicpu;
use_loop, use_186_fast_shift, use_8086_fast_shift,
use_386_fast_shift: Boolean;
rox32method: trox32method=rm_unspecified;
i: Integer;
rol_amount, ror_amount: TCGInt;
begin
optimize_op_const(size, op, a);
check_register_size(size,reg);
if size in [OS_64, OS_S64] then
internalerror(2013030904);
if size in [OS_32, OS_S32] then
begin
case op of
OP_NONE:
begin
{ Opcode is optimized away }
end;
OP_MOVE:
begin
{ Optimized, replaced with a simple load }
a_load_const_reg(list,size,a,reg);
end;
OP_ADD, OP_SUB:
begin
get_32bit_ops(op, op1, op2);
{ Optimization when the low 16-bits of the constant are 0 }
if aint(a and $FFFF) = 0 then
begin
{ use a_op_const_reg to allow the use of inc/dec }
a_op_const_reg(list,op,OS_16,aint(a shr 16),GetNextReg(reg));
end
else
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(op1,S_W,aint(a and $FFFF),reg));
list.concat(taicpu.op_const_reg(op2,S_W,aint(a shr 16),GetNextReg(reg)));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
end;
OP_AND, OP_OR, OP_XOR:
begin
{ low word operation }
if aint(a and $FFFF) = aint(0) then
begin
case op of
OP_AND:
a_load_const_reg(list,OS_16,aint(0),reg);
OP_OR,OP_XOR:
{do nothing};
else
InternalError(2013100704);
end;
end
else if aint(a and $FFFF) = aint($FFFF) then
begin
case op of
OP_AND:
{do nothing};
OP_OR:
a_load_const_reg(list,OS_16,aint($FFFF),reg);
OP_XOR:
list.concat(taicpu.op_reg(A_NOT,S_W,reg));
else
InternalError(2013100705);
end;
end
else
a_op_const_reg(list,op,OS_16,aint(a and $FFFF),reg);
{ high word operation }
if aint(a shr 16) = aint(0) then
begin
case op of
OP_AND:
a_load_const_reg(list,OS_16,aint(0),GetNextReg(reg));
OP_OR,OP_XOR:
{do nothing};
else
InternalError(2013100706);
end;
end
else if aint(a shr 16) = aint($FFFF) then
begin
case op of
OP_AND:
{do nothing};
OP_OR:
a_load_const_reg(list,OS_16,aint($FFFF),GetNextReg(reg));
OP_XOR:
list.concat(taicpu.op_reg(A_NOT,S_W,GetNextReg(reg)));
else
InternalError(2013100707);
end;
end
else
a_op_const_reg(list,op,OS_16,aint(a shr 16),GetNextReg(reg));
end;
OP_SHR,OP_SHL,OP_SAR:
begin
a:=a and 31;
{ for shl with const >= 16, we can just move the low register
to the high reg, then zero the low register, then do the
remaining part of the shift (by const-16) in 16 bit on the
high register. the same thing applies to shr with low and high
reversed. sar is exactly like shr, except that instead of
zeroing the high register, we sar it by 15. }
if a>=16 then
case op of
OP_SHR:
begin
a_load_reg_reg(list,OS_16,OS_16,GetNextReg(reg),reg);
a_load_const_reg(list,OS_16,0,GetNextReg(reg));
a_op_const_reg(list,OP_SHR,OS_16,a-16,reg);
end;
OP_SHL:
begin
a_load_reg_reg(list,OS_16,OS_16,reg,GetNextReg(reg));
a_load_const_reg(list,OS_16,0,reg);
a_op_const_reg(list,OP_SHL,OS_16,a-16,GetNextReg(reg));
end;
OP_SAR:
begin
if a=31 then
begin
a_op_const_reg(list,OP_SAR,OS_16,15,GetNextReg(reg));
a_load_reg_reg(list,OS_16,OS_16,GetNextReg(reg),reg);
end
else
begin
a_load_reg_reg(list,OS_16,OS_16,GetNextReg(reg),reg);
a_op_const_reg(list,OP_SAR,OS_16,15,GetNextReg(reg));
a_op_const_reg(list,OP_SAR,OS_16,a-16,reg);
end;
end;
else
internalerror(2013060201);
end
else if a<>0 then
begin
use_loop:=a>2;
use_386_fast_shift:=(current_settings.cputype>=cpu_386) and (a>1);
use_186_fast_shift:=not use_386_fast_shift
and (current_settings.cputype>=cpu_186) and (a>2)
and not (cs_opt_size in current_settings.optimizerswitches);
use_8086_fast_shift:=(current_settings.cputype<cpu_186) and (a>2)
and not (cs_opt_size in current_settings.optimizerswitches);
if use_386_fast_shift then
begin
case op of
OP_SHR:
begin
list.concat(taicpu.op_const_reg_reg(A_SHRD,S_W,a,GetNextReg(reg),reg));
list.concat(taicpu.op_const_reg(A_SHR,S_W,a,GetNextReg(reg)));
end;
OP_SAR:
begin
list.concat(taicpu.op_const_reg_reg(A_SHRD,S_W,a,GetNextReg(reg),reg));
list.concat(taicpu.op_const_reg(A_SAR,S_W,a,GetNextReg(reg)));
end;
OP_SHL:
begin
list.concat(taicpu.op_const_reg_reg(A_SHLD,S_W,a,reg,GetNextReg(reg)));
list.concat(taicpu.op_const_reg(A_SHL,S_W,a,reg));
end;
else
internalerror(2017040401);
end;
end
else if use_186_fast_shift then
begin
tmpreg:=getintregister(list,OS_16);
case op of
OP_SHR:
begin
a_load_reg_reg(list,OS_16,OS_16,GetNextReg(reg),tmpreg);
list.concat(taicpu.op_const_reg(A_SHR,S_W,a,GetNextReg(reg)));
list.concat(taicpu.op_const_reg(A_SHR,S_W,a,reg));
list.concat(taicpu.op_const_reg(A_SHL,S_W,16-a,tmpreg));
list.concat(taicpu.op_reg_reg(A_OR,S_W,tmpreg,reg));
end;
OP_SAR:
begin
a_load_reg_reg(list,OS_16,OS_16,GetNextReg(reg),tmpreg);
list.concat(taicpu.op_const_reg(A_SAR,S_W,a,GetNextReg(reg)));
list.concat(taicpu.op_const_reg(A_SHR,S_W,a,reg));
list.concat(taicpu.op_const_reg(A_SHL,S_W,16-a,tmpreg));
list.concat(taicpu.op_reg_reg(A_OR,S_W,tmpreg,reg));
end;
OP_SHL:
begin
a_load_reg_reg(list,OS_16,OS_16,reg,tmpreg);
list.concat(taicpu.op_const_reg(A_SHL,S_W,a,reg));
list.concat(taicpu.op_const_reg(A_SHL,S_W,a,GetNextReg(reg)));
list.concat(taicpu.op_const_reg(A_SHR,S_W,16-a,tmpreg));
list.concat(taicpu.op_reg_reg(A_OR,S_W,tmpreg,GetNextReg(reg)));
end;
else
internalerror(2017040301);
end;
end
else if use_8086_fast_shift then
begin
getcpuregister(list,NR_CX);
a_load_const_reg(list,OS_8,a,NR_CL);
tmpreg:=getintregister(list,OS_16);
case op of
OP_SHR:
begin
a_load_reg_reg(list,OS_16,OS_16,GetNextReg(reg),tmpreg);
list.concat(taicpu.op_reg_reg(A_SHR,S_W,NR_CL,GetNextReg(reg)));
list.concat(taicpu.op_reg_reg(A_SHR,S_W,NR_CL,reg));
if a<>8 then
a_load_const_reg(list,OS_8,16-a,NR_CL);
list.concat(taicpu.op_reg_reg(A_SHL,S_W,NR_CL,tmpreg));
list.concat(taicpu.op_reg_reg(A_OR,S_W,tmpreg,reg));
end;
OP_SAR:
begin
a_load_reg_reg(list,OS_16,OS_16,GetNextReg(reg),tmpreg);
list.concat(taicpu.op_reg_reg(A_SAR,S_W,NR_CL,GetNextReg(reg)));
list.concat(taicpu.op_reg_reg(A_SHR,S_W,NR_CL,reg));
if a<>8 then
a_load_const_reg(list,OS_8,16-a,NR_CL);
list.concat(taicpu.op_reg_reg(A_SHL,S_W,NR_CL,tmpreg));
list.concat(taicpu.op_reg_reg(A_OR,S_W,tmpreg,reg));
end;
OP_SHL:
begin
a_load_reg_reg(list,OS_16,OS_16,reg,tmpreg);
list.concat(taicpu.op_reg_reg(A_SHL,S_W,NR_CL,reg));
list.concat(taicpu.op_reg_reg(A_SHL,S_W,NR_CL,GetNextReg(reg)));
if a<>8 then
a_load_const_reg(list,OS_8,16-a,NR_CL);
list.concat(taicpu.op_reg_reg(A_SHR,S_W,NR_CL,tmpreg));
list.concat(taicpu.op_reg_reg(A_OR,S_W,tmpreg,GetNextReg(reg)));
end;
else
internalerror(2017040302);
end;
ungetcpuregister(list,NR_CX);
end
else if use_loop then
begin
getcpuregister(list,NR_CX);
a_load_const_reg(list,OS_16,a,NR_CX);
current_asmdata.getjumplabel(hl_loop_start);
a_label(list,hl_loop_start);
case op of
OP_SHR:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_SHR,S_W,1,GetNextReg(reg)));
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,reg));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_SAR:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_SAR,S_W,1,GetNextReg(reg)));
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,reg));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_SHL:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_SHL,S_W,1,reg));
list.concat(taicpu.op_const_reg(A_RCL,S_W,1,GetNextReg(reg)));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
else
internalerror(2013030903);
end;
ai:=Taicpu.Op_Sym(A_LOOP,S_W,hl_loop_start);
ai.is_jmp:=true;
list.concat(ai);
ungetcpuregister(list,NR_CX);
end
else
begin
for i:=1 to a do
begin
case op of
OP_SHR:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_SHR,S_W,1,GetNextReg(reg)));
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,reg));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_SAR:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_SAR,S_W,1,GetNextReg(reg)));
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,reg));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_SHL:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_SHL,S_W,1,reg));
list.concat(taicpu.op_const_reg(A_RCL,S_W,1,GetNextReg(reg)));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
else
internalerror(2013030907);
end;
end;
end;
end;
end;
OP_ROL,OP_ROR:
begin
a:=a and 31;
if a=0 then
exit;
if op=OP_ROL then
begin
rol_amount:=a;
ror_amount:=32-a;
end
else
begin
rol_amount:=32-a;
ror_amount:=a;
end;
case rol_amount of
1,17:
rox32method:=rm_unrolledleftloop;
2,18:
if current_settings.cputype>=cpu_386 then
rox32method:=rm_fast_386
else if not (cs_opt_size in current_settings.optimizerswitches) then
rox32method:=rm_unrolledleftloop
else
rox32method:=rm_loopleft;
3..8,19..24:
if current_settings.cputype>=cpu_386 then
rox32method:=rm_fast_386
else
rox32method:=rm_loopleft;
15,31:
rox32method:=rm_unrolledrightloop;
14,30:
if current_settings.cputype>=cpu_386 then
rox32method:=rm_fast_386
else if not (cs_opt_size in current_settings.optimizerswitches) then
rox32method:=rm_unrolledrightloop
else
{ the left loop has a smaller size }
rox32method:=rm_loopleft;
9..13,25..29:
if current_settings.cputype>=cpu_386 then
rox32method:=rm_fast_386
else if not (cs_opt_size in current_settings.optimizerswitches) then
rox32method:=rm_loopright
else
{ the left loop has a smaller size }
rox32method:=rm_loopleft;
16:
rox32method:=rm_unrolledleftloop;
else
internalerror(2017040601);
end;
case rox32method of
rm_unrolledleftloop:
begin
if rol_amount>=16 then
begin
list.Concat(taicpu.op_reg_reg(A_XCHG,S_W,reg,GetNextReg(reg)));
dec(rol_amount,16);
end;
for i:=1 to rol_amount do
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.Concat(taicpu.op_const_reg(A_SHL,S_W,1,GetNextReg(reg)));
list.Concat(taicpu.op_const_reg(A_RCL,S_W,1,reg));
list.Concat(taicpu.op_const_reg(A_ADC,S_W,0,GetNextReg(reg)));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
end;
rm_unrolledrightloop:
begin
if ror_amount>=16 then
begin
list.Concat(taicpu.op_reg_reg(A_XCHG,S_W,reg,GetNextReg(reg)));
dec(ror_amount,16);
end;
tmpreg:=getintregister(list,OS_16);
for i:=1 to ror_amount do
begin
a_load_reg_reg(list,OS_16,OS_16,reg,tmpreg);
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.Concat(taicpu.op_const_reg(A_SHR,S_W,1,tmpreg));
list.Concat(taicpu.op_const_reg(A_RCR,S_W,1,GetNextReg(reg)));
list.Concat(taicpu.op_const_reg(A_RCR,S_W,1,reg));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
end;
rm_loopleft:
begin
if (rol_amount>=16) and not (cs_opt_size in current_settings.optimizerswitches) then
begin
list.Concat(taicpu.op_reg_reg(A_XCHG,S_W,reg,GetNextReg(reg)));
dec(rol_amount,16);
if rol_amount=0 then
exit;
end;
getcpuregister(list,NR_CX);
a_load_const_reg(list,OS_16,rol_amount,NR_CX);
current_asmdata.getjumplabel(hl_loop_start);
a_label(list,hl_loop_start);
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.Concat(taicpu.op_const_reg(A_SHL,S_W,1,GetNextReg(reg)));
list.Concat(taicpu.op_const_reg(A_RCL,S_W,1,reg));
list.Concat(taicpu.op_const_reg(A_ADC,S_W,0,GetNextReg(reg)));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
ai:=Taicpu.Op_Sym(A_LOOP,S_W,hl_loop_start);
ai.is_jmp:=true;
list.concat(ai);
ungetcpuregister(list,NR_CX);
end;
rm_loopright:
begin
if (ror_amount>=16) and not (cs_opt_size in current_settings.optimizerswitches) then
begin
list.Concat(taicpu.op_reg_reg(A_XCHG,S_W,reg,GetNextReg(reg)));
dec(ror_amount,16);
if ror_amount=0 then
exit;
end;
getcpuregister(list,NR_CX);
a_load_const_reg(list,OS_16,ror_amount,NR_CX);
current_asmdata.getjumplabel(hl_loop_start);
a_label(list,hl_loop_start);
tmpreg:=getintregister(list,OS_16);
a_load_reg_reg(list,OS_16,OS_16,reg,tmpreg);
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.Concat(taicpu.op_const_reg(A_SHR,S_W,1,tmpreg));
list.Concat(taicpu.op_const_reg(A_RCR,S_W,1,GetNextReg(reg)));
list.Concat(taicpu.op_const_reg(A_RCR,S_W,1,reg));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
ai:=Taicpu.Op_Sym(A_LOOP,S_W,hl_loop_start);
ai.is_jmp:=true;
list.concat(ai);
ungetcpuregister(list,NR_CX);
end;
rm_fast_386:
begin
tmpreg:=getintregister(list,OS_16);
a_load_reg_reg(list,OS_16,OS_16,GetNextReg(reg),tmpreg);
if op=OP_ROL then
begin
list.Concat(taicpu.op_const_reg_reg(A_SHLD,S_W,rol_amount,reg,GetNextReg(reg)));
list.Concat(taicpu.op_const_reg_reg(A_SHLD,S_W,rol_amount,tmpreg,reg));
end
else
begin
list.Concat(taicpu.op_const_reg_reg(A_SHRD,S_W,ror_amount,reg,GetNextReg(reg)));
list.Concat(taicpu.op_const_reg_reg(A_SHRD,S_W,ror_amount,tmpreg,reg));
end;
end;
else
internalerror(2017040602);
end;
end;
else
begin
tmpreg:=getintregister(list,size);
a_load_const_reg(list,size,a,tmpreg);
a_op_reg_reg(list,op,size,tmpreg,reg);
end;
end;
end
else
begin
{ size <= 16-bit }
{ 8086 doesn't support 'imul reg,const', so we handle it here }
if (current_settings.cputype<cpu_186) and (op in [OP_MUL,OP_IMUL]) then
begin
if op = OP_IMUL then
begin
if size in [OS_16,OS_S16] then
ax_subreg := NR_AX
else
if size in [OS_8,OS_S8] then
ax_subreg := NR_AL
else
internalerror(2013050102);
getcpuregister(list,NR_AX);
a_load_const_reg(list,size,a,ax_subreg);
if size in [OS_16,OS_S16] then
getcpuregister(list,NR_DX);
{ prefer MUL over IMUL when overflow checking is off, }
{ because it's faster on the 8086 & 8088 }
if not (cs_check_overflow in current_settings.localswitches) then
list.concat(taicpu.op_reg(A_MUL,TCgSize2OpSize[size],reg))
else
list.concat(taicpu.op_reg(A_IMUL,TCgSize2OpSize[size],reg));
if size in [OS_16,OS_S16] then
ungetcpuregister(list,NR_DX);
a_load_reg_reg(list,size,size,ax_subreg,reg);
ungetcpuregister(list,NR_AX);
exit;
end
else
{ OP_MUL should be handled specifically in the code }
{ generator because of the silly register usage restraints }
internalerror(2001092208);
end
else
inherited a_op_const_reg(list, Op, size, a, reg);
end;
end;
procedure tcg8086.a_op_const_ref(list: TAsmList; Op: TOpCG; size: TCGSize; a: tcgint; const ref: TReference);
var
tmpref: treference;
op1,op2: TAsmOp;
tmpreg: TRegister;
begin
optimize_op_const(size, op, a);
tmpref:=ref;
make_simple_ref(list,tmpref);
if size in [OS_64, OS_S64] then
internalerror(2013050801);
if size in [OS_32, OS_S32] then
begin
case Op of
OP_NONE :
begin
{ Opcode is optimized away }
end;
OP_MOVE :
begin
{ Optimized, replaced with a simple load }
a_load_const_ref(list,size,a,ref);
end;
OP_ADD, OP_SUB:
begin
get_32bit_ops(op, op1, op2);
{ Optimization when the low 16-bits of the constant are 0 }
if aint(a and $FFFF) = 0 then
begin
inc(tmpref.offset, 2);
{ use a_op_const_ref to allow the use of inc/dec }
a_op_const_ref(list,op,OS_16,aint(a shr 16),tmpref);
end
else
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_ref(op1,S_W,aint(a and $FFFF),tmpref));
inc(tmpref.offset, 2);
list.concat(taicpu.op_const_ref(op2,S_W,aint(a shr 16),tmpref));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
end;
OP_AND, OP_OR, OP_XOR:
begin
{ low word operation }
if aint(a and $FFFF) = aint(0) then
begin
case op of
OP_AND:
a_load_const_ref(list,OS_16,aint(0),ref);
OP_OR,OP_XOR:
{do nothing};
else
InternalError(2013100708);
end;
end
else if aint(a and $FFFF) = aint($FFFF) then
begin
case op of
OP_AND:
{do nothing};
OP_OR:
a_load_const_ref(list,OS_16,aint($FFFF),tmpref);
OP_XOR:
list.concat(taicpu.op_ref(A_NOT,S_W,tmpref));
else
InternalError(2013100709);
end;
end
else
a_op_const_ref(list,op,OS_16,aint(a and $FFFF),tmpref);
{ high word operation }
inc(tmpref.offset, 2);
if aint(a shr 16) = aint(0) then
begin
case op of
OP_AND:
a_load_const_ref(list,OS_16,aint(0),tmpref);
OP_OR,OP_XOR:
{do nothing};
else
InternalError(2013100710);
end;
end
else if aint(a shr 16) = aint($FFFF) then
begin
case op of
OP_AND:
{do nothing};
OP_OR:
a_load_const_ref(list,OS_16,aint($FFFF),tmpref);
OP_XOR:
list.concat(taicpu.op_ref(A_NOT,S_W,tmpref));
else
InternalError(2013100711);
end;
end
else
a_op_const_ref(list,op,OS_16,aint(a shr 16),tmpref);
end;
OP_SHR,OP_SHL,OP_SAR:
begin
a:=a and 31;
if a=1 then
begin
case op of
OP_SHR:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
inc(tmpref.offset, 2);
list.concat(taicpu.op_const_ref(A_SHR,S_W,1,tmpref));
dec(tmpref.offset, 2);
list.concat(taicpu.op_const_ref(A_RCR,S_W,1,tmpref));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_SAR:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
inc(tmpref.offset, 2);
list.concat(taicpu.op_const_ref(A_SAR,S_W,1,tmpref));
dec(tmpref.offset, 2);
list.concat(taicpu.op_const_ref(A_RCR,S_W,1,tmpref));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_SHL:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_ref(A_SHL,S_W,1,tmpref));
inc(tmpref.offset, 2);
list.concat(taicpu.op_const_ref(A_RCL,S_W,1,tmpref));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
else
internalerror(2015042501);
end;
end
else
begin
tmpreg:=getintregister(list,size);
a_load_ref_reg(list,size,size,ref,tmpreg);
a_op_const_reg(list,Op,size,a,tmpreg);
a_load_reg_ref(list,size,size,tmpreg,ref);
end;
end;
OP_ROL,OP_ROR:
begin
tmpreg:=getintregister(list,size);
a_load_ref_reg(list,size,size,ref,tmpreg);
a_op_const_reg(list,Op,size,a,tmpreg);
a_load_reg_ref(list,size,size,tmpreg,ref);
end;
else
internalerror(2013050802);
end;
end
else
inherited a_op_const_ref(list,Op,size,a,tmpref);
end;
procedure tcg8086.a_op_reg_reg(list: TAsmList; Op: TOpCG; size: TCGSize;
src, dst: TRegister);
var
op1, op2: TAsmOp;
hl_skip, hl_loop_start: TAsmLabel;
ai: taicpu;
tmpreg: TRegister;
begin
check_register_size(size,src);
check_register_size(size,dst);
if size in [OS_64, OS_S64] then
internalerror(2013030902);
if size in [OS_32, OS_S32] then
begin
case op of
OP_NEG:
begin
if src<>dst then
a_load_reg_reg(list,size,size,src,dst);
list.concat(taicpu.op_reg(A_NOT, S_W, GetNextReg(dst)));
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_reg(A_NEG, S_W, dst));
list.concat(taicpu.op_const_reg(A_SBB, S_W,-1, GetNextReg(dst)));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_NOT:
begin
if src<>dst then
a_load_reg_reg(list,size,size,src,dst);
list.concat(taicpu.op_reg(A_NOT, S_W, dst));
list.concat(taicpu.op_reg(A_NOT, S_W, GetNextReg(dst)));
end;
OP_ADD,OP_SUB,OP_XOR,OP_OR,OP_AND:
begin
get_32bit_ops(op, op1, op2);
if op in [OP_ADD,OP_SUB] then
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_reg_reg(op1, S_W, src, dst));
list.concat(taicpu.op_reg_reg(op2, S_W, GetNextReg(src), GetNextReg(dst)));
if op in [OP_ADD,OP_SUB] then
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_SHR,OP_SHL,OP_SAR:
begin
getcpuregister(list,NR_CX);
a_load_reg_reg(list,size,OS_16,src,NR_CX);
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_AND,S_W,$1f,NR_CX));
current_asmdata.getjumplabel(hl_skip);
ai:=Taicpu.Op_Sym(A_Jcc,S_NO,hl_skip);
ai.SetCondition(C_Z);
ai.is_jmp:=true;
list.concat(ai);
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
current_asmdata.getjumplabel(hl_loop_start);
a_label(list,hl_loop_start);
case op of
OP_SHR:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_SHR,S_W,1,GetNextReg(dst)));
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,dst));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_SAR:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_SAR,S_W,1,GetNextReg(dst)));
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,dst));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_SHL:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_SHL,S_W,1,dst));
list.concat(taicpu.op_const_reg(A_RCL,S_W,1,GetNextReg(dst)));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
else
internalerror(2013030908);
end;
ai:=Taicpu.Op_Sym(A_LOOP,S_W,hl_loop_start);
ai.is_jmp:=true;
list.concat(ai);
a_label(list,hl_skip);
ungetcpuregister(list,NR_CX);
end;
OP_ROL,OP_ROR:
begin
getcpuregister(list,NR_CX);
a_load_reg_reg(list,size,OS_16,src,NR_CX);
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_AND,S_W,$1f,NR_CX));
current_asmdata.getjumplabel(hl_skip);
ai:=Taicpu.Op_Sym(A_Jcc,S_NO,hl_skip);
ai.SetCondition(C_Z);
ai.is_jmp:=true;
list.concat(ai);
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
current_asmdata.getjumplabel(hl_loop_start);
a_label(list,hl_loop_start);
case op of
OP_ROL:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.Concat(taicpu.op_const_reg(A_SHL,S_W,1,GetNextReg(dst)));
list.Concat(taicpu.op_const_reg(A_RCL,S_W,1,dst));
list.Concat(taicpu.op_const_reg(A_ADC,S_W,0,GetNextReg(dst)));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_ROR:
begin
tmpreg:=getintregister(list,OS_16);
a_load_reg_reg(list,OS_16,OS_16,dst,tmpreg);
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.Concat(taicpu.op_const_reg(A_SHR,S_W,1,tmpreg));
list.Concat(taicpu.op_const_reg(A_RCR,S_W,1,GetNextReg(dst)));
list.Concat(taicpu.op_const_reg(A_RCR,S_W,1,dst));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
else
internalerror(2017042502);
end;
ai:=Taicpu.Op_Sym(A_LOOP,S_W,hl_loop_start);
ai.is_jmp:=true;
list.concat(ai);
a_label(list,hl_skip);
ungetcpuregister(list,NR_CX);
end;
else
internalerror(2013030905);
end;
end
else
inherited a_op_reg_reg(list, Op, size, src, dst);
end;
procedure tcg8086.a_op_ref_reg(list: TAsmList; Op: TOpCG; size: TCGSize; const ref: TReference; reg: TRegister);
var
tmpref : treference;
op1, op2: TAsmOp;
begin
tmpref:=ref;
make_simple_ref(list,tmpref);
check_register_size(size,reg);
if size in [OS_64, OS_S64] then
internalerror(2013030906);
if size in [OS_32, OS_S32] then
begin
case op of
OP_ADD,OP_SUB,OP_XOR,OP_OR,OP_AND:
begin
get_32bit_ops(op, op1, op2);
if op in [OP_ADD,OP_SUB] then
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_ref_reg(op1, S_W, tmpref, reg));
inc(tmpref.offset, 2);
list.concat(taicpu.op_ref_reg(op2, S_W, tmpref, GetNextReg(reg)));
if op in [OP_ADD,OP_SUB] then
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
else
internalerror(2013050701);
end;
end
else
inherited a_op_ref_reg(list,Op,size,tmpref,reg);
end;
procedure tcg8086.a_op_reg_ref(list: TAsmList; Op: TOpCG; size: TCGSize; reg: TRegister; const ref: TReference);
var
tmpref: treference;
op1,op2: TAsmOp;
hl_skip, hl_loop_start: TAsmLabel;
ai: taicpu;
tmpreg: TRegister;
begin
tmpref:=ref;
make_simple_ref(list,tmpref);
if not (op in [OP_SHR,OP_SHL,OP_SAR,OP_ROL,OP_ROR]) then
check_register_size(size,reg);
if size in [OS_64, OS_S64] then
internalerror(2013050803);
if size in [OS_32, OS_S32] then
begin
case op of
OP_NEG,OP_NOT:
inherited;
OP_IMUL:
begin
{ this one needs a load/imul/store, which is the default }
inherited a_op_ref_reg(list,op,size,tmpref,reg);
end;
OP_MUL,OP_DIV,OP_IDIV:
{ special stuff, needs separate handling inside code }
{ generator }
internalerror(2001092304);
OP_ADD,OP_SUB,OP_XOR,OP_OR,OP_AND:
begin
get_32bit_ops(op, op1, op2);
if op in [OP_ADD,OP_SUB] then
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_reg_ref(op1, S_W, reg, tmpref));
inc(tmpref.offset, 2);
list.concat(taicpu.op_reg_ref(op2, S_W, GetNextReg(reg), tmpref));
if op in [OP_ADD,OP_SUB] then
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_SHR,OP_SHL,OP_SAR:
begin
getcpuregister(list,NR_CX);
a_load_reg_reg(list,size,OS_16,reg,NR_CX);
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_AND,S_W,$1f,NR_CX));
current_asmdata.getjumplabel(hl_skip);
ai:=Taicpu.Op_Sym(A_Jcc,S_NO,hl_skip);
ai.SetCondition(C_Z);
ai.is_jmp:=true;
list.concat(ai);
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
current_asmdata.getjumplabel(hl_loop_start);
a_label(list,hl_loop_start);
case op of
OP_SHR:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
inc(tmpref.offset, 2);
list.concat(taicpu.op_const_ref(A_SHR,S_W,1,tmpref));
dec(tmpref.offset, 2);
list.concat(taicpu.op_const_ref(A_RCR,S_W,1,tmpref));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_SAR:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
inc(tmpref.offset, 2);
list.concat(taicpu.op_const_ref(A_SAR,S_W,1,tmpref));
dec(tmpref.offset, 2);
list.concat(taicpu.op_const_ref(A_RCR,S_W,1,tmpref));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_SHL:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_ref(A_SHL,S_W,1,tmpref));
inc(tmpref.offset, 2);
list.concat(taicpu.op_const_ref(A_RCL,S_W,1,tmpref));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
else
internalerror(2013030909);
end;
ai:=Taicpu.Op_Sym(A_LOOP,S_W,hl_loop_start);
ai.is_jmp:=true;
list.concat(ai);
a_label(list,hl_skip);
ungetcpuregister(list,NR_CX);
end;
OP_ROL,OP_ROR:
begin
tmpreg:=getintregister(list,size);
a_load_ref_reg(list,size,size,ref,tmpreg);
a_op_reg_reg(list,Op,size,reg,tmpreg);
a_load_reg_ref(list,size,size,tmpreg,ref);
end;
else
internalerror(2013050804);
end;
end
else
inherited a_op_reg_ref(list,Op,size,reg,tmpref);
end;
procedure tcg8086.a_op_ref(list: TAsmList; Op: TOpCG; size: TCGSize; const ref: TReference);
var
tmpref: treference;
begin
tmpref:=ref;
make_simple_ref(list,tmpref);
if size in [OS_64, OS_S64] then
internalerror(2013050805);
if size in [OS_32, OS_S32] then
begin
case op of
OP_NEG:
begin
inc(tmpref.offset, 2);
list.concat(taicpu.op_ref(A_NOT, S_W, tmpref));
dec(tmpref.offset, 2);
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_ref(A_NEG, S_W, tmpref));
inc(tmpref.offset, 2);
list.concat(taicpu.op_const_ref(A_SBB, S_W,-1, tmpref));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_NOT:
begin
list.concat(taicpu.op_ref(A_NOT, S_W, tmpref));
inc(tmpref.offset, 2);
list.concat(taicpu.op_ref(A_NOT, S_W, tmpref));
end;
else
internalerror(2020050709);
end;
end
else
inherited;
end;
procedure tcg8086.push_const(list: TAsmList; size: tcgsize; a: tcgint);
var
tmpreg: TRegister;
begin
if not (size in [OS_16,OS_S16]) then
internalerror(2013043001);
if current_settings.cputype < cpu_186 then
begin
tmpreg:=getintregister(list,size);
a_load_const_reg(list,size,a,tmpreg);
list.concat(taicpu.op_reg(A_PUSH,S_W,tmpreg));
end
else
list.concat(taicpu.op_const(A_PUSH,TCGSize2OpSize[size],a));
end;
procedure tcg8086.a_load_reg_cgpara(list : TAsmList;size : tcgsize;r : tregister;const cgpara : tcgpara);
procedure load_para_loc(r : TRegister;paraloc : PCGParaLocation);
var
ref : treference;
begin
paramanager.allocparaloc(list,paraloc);
case paraloc^.loc of
LOC_REGISTER,LOC_CREGISTER:
a_load_reg_reg(list,paraloc^.size,paraloc^.size,r,paraloc^.register);
LOC_REFERENCE,LOC_CREFERENCE:
begin
reference_reset_base(ref,paraloc^.reference.index,paraloc^.reference.offset,ctempposinvalid,2,[]);
a_load_reg_ref(list,paraloc^.size,paraloc^.size,r,ref);
end;
else
internalerror(2002071006);
end;
end;
var
pushsize,pushsize2 : tcgsize;
begin
check_register_size(size,r);
if use_push(cgpara) then
begin
if tcgsize2size[cgpara.Size] > 2 then
begin
if tcgsize2size[cgpara.Size] <> 4 then
internalerror(2013031101);
if cgpara.location^.Next = nil then
begin
if tcgsize2size[cgpara.location^.size] <> 4 then
internalerror(2013031103);
end
else
begin
if tcgsize2size[cgpara.location^.size] <> 2 then
internalerror(2013031104);
if tcgsize2size[cgpara.location^.Next^.size] <> 2 then
internalerror(2013031105);
if cgpara.location^.Next^.Next <> nil then
internalerror(2013031106);
end;
if tcgsize2size[cgpara.size]>cgpara.alignment then
pushsize:=cgpara.size
else
pushsize:=int_cgsize(cgpara.alignment);
pushsize2 := int_cgsize(tcgsize2size[pushsize] - 2);
list.concat(taicpu.op_reg(A_PUSH,TCgsize2opsize[pushsize2],makeregsize(list,GetNextReg(r),pushsize2)));
list.concat(taicpu.op_reg(A_PUSH,S_W,makeregsize(list,r,OS_16)));
end
else
begin
cgpara.check_simple_location;
if tcgsize2size[cgpara.location^.size]>cgpara.alignment then
pushsize:=cgpara.location^.size
else
pushsize:=int_cgsize(cgpara.alignment);
list.concat(taicpu.op_reg(A_PUSH,TCgsize2opsize[pushsize],makeregsize(list,r,pushsize)));
end;
end
else
begin
if tcgsize2size[cgpara.Size]=4 then
begin
if (cgpara.location^.Next=nil) or
(tcgsize2size[cgpara.location^.size]<>2) or
(tcgsize2size[cgpara.location^.Next^.size]<>2) or
(cgpara.location^.Next^.Next<>nil) or
(cgpara.location^.shiftval<>0) then
internalerror(2013031102);
load_para_loc(r,cgpara.Location);
load_para_loc(GetNextReg(r),cgpara.Location^.Next);
end
else
inherited a_load_reg_cgpara(list,size,r,cgpara);
end;
end;
procedure tcg8086.a_load_const_cgpara(list : TAsmList;size : tcgsize;a : tcgint;const cgpara : tcgpara);
var
pushsize : tcgsize;
begin
if use_push(cgpara) then
begin
if tcgsize2size[cgpara.Size] > 2 then
begin
if tcgsize2size[cgpara.Size] <> 4 then
internalerror(2013031107);
if cgpara.location^.Next = nil then
begin
if tcgsize2size[cgpara.location^.size] <> 4 then
internalerror(2013031108);
end
else
begin
if tcgsize2size[cgpara.location^.size] <> 2 then
internalerror(2013031109);
if tcgsize2size[cgpara.location^.Next^.size] <> 2 then
internalerror(2013031110);
if cgpara.location^.Next^.Next <> nil then
internalerror(2013031111);
end;
if (cgpara.alignment <> 4) and (cgpara.alignment <> 2) then
internalerror(2013031112);
push_const(list,OS_16,a shr 16);
push_const(list,OS_16,a and $FFFF);
end
else
begin
cgpara.check_simple_location;
if tcgsize2size[cgpara.location^.size]>cgpara.alignment then
pushsize:=cgpara.location^.size
else
pushsize:=int_cgsize(cgpara.alignment);
push_const(list,pushsize,a);
end;
end
else if (tcgsize2size[cgpara.Size]>2) and
(cgpara.location^.loc in [LOC_REGISTER,LOC_CREGISTER]) and
(cgpara.location^.Next<>nil) then
begin
if (tcgsize2size[cgpara.Size]<>4) or
(tcgsize2size[cgpara.location^.Size]<>2) or
not (cgpara.location^.Next^.Loc in [LOC_REGISTER,LOC_CREGISTER]) or
(tcgsize2size[cgpara.location^.Next^.Size]<>2) or
(cgpara.location^.Next^.Next<>nil) then
internalerror(2018041801);
a_load_const_reg(list,cgpara.location^.size,a and $FFFF,cgpara.location^.register);
a_load_const_reg(list,cgpara.location^.Next^.size,a shr 16,cgpara.location^.Next^.register);
end
else
inherited a_load_const_cgpara(list,size,a,cgpara);
end;
procedure tcg8086.a_load_ref_cgpara(list : TAsmList;size : tcgsize;const r : treference;const cgpara : tcgpara);
procedure pushdata(paraloc:pcgparalocation;ofs:tcgint);
var
pushsize : tcgsize;
opsize : topsize;
tmpreg : tregister;
href,tmpref: treference;
begin
if not assigned(paraloc) then
exit;
if (paraloc^.loc<>LOC_REFERENCE) or
(paraloc^.reference.index<>NR_STACK_POINTER_REG) or
(tcgsize2size[paraloc^.size]>4) then
internalerror(200501162);
{ Pushes are needed in reverse order, add the size of the
current location to the offset where to load from. This
prevents wrong calculations for the last location when
the size is not a power of 2 }
if assigned(paraloc^.next) then
pushdata(paraloc^.next,ofs+tcgsize2size[paraloc^.size]);
{ Push the data starting at ofs }
href:=r;
inc(href.offset,ofs);
if tcgsize2size[paraloc^.size]>cgpara.alignment then
pushsize:=paraloc^.size
else
pushsize:=int_cgsize(cgpara.alignment);
opsize:=TCgsize2opsize[pushsize];
{ for go32v2 we obtain OS_F32,
but pushs is not valid, we need pushl }
if opsize=S_FS then
opsize:=S_W;
if tcgsize2size[paraloc^.size]<cgpara.alignment then
begin
tmpreg:=getintregister(list,pushsize);
a_load_ref_reg(list,paraloc^.size,pushsize,href,tmpreg);
list.concat(taicpu.op_reg(A_PUSH,opsize,tmpreg));
end
else
begin
make_simple_ref(list,href);
if tcgsize2size[pushsize] > 2 then
begin
tmpref := href;
Inc(tmpref.offset, 2);
list.concat(taicpu.op_ref(A_PUSH,TCgsize2opsize[int_cgsize(tcgsize2size[pushsize]-2)],tmpref));
end;
list.concat(taicpu.op_ref(A_PUSH,opsize,href));
end;
end;
var
len : tcgint;
href : treference;
begin
{ cgpara.size=OS_NO requires a copy on the stack }
if use_push(cgpara) then
begin
{ Record copy? }
if (cgpara.size in [OS_NO,OS_F64]) or (size=OS_NO) then
begin
cgpara.check_simple_location;
len:=align(cgpara.intsize,cgpara.alignment);
g_stackpointer_alloc(list,len);
reference_reset_base(href,NR_STACK_POINTER_REG,0,ctempposinvalid,4,[]);
g_concatcopy(list,r,href,len);
end
else
begin
if tcgsize2size[cgpara.size]<>tcgsize2size[size] then
internalerror(200501161);
{ We need to push the data in reverse order,
therefor we use a recursive algorithm }
pushdata(cgpara.location,0);
end
end
else
inherited a_load_ref_cgpara(list,size,r,cgpara);
end;
procedure tcg8086.a_loadaddr_ref_cgpara(list : TAsmList;const r : treference;const cgpara : tcgpara);
var
tmpreg : tregister;
tmpref : treference;
begin
with r do
begin
if use_push(cgpara) then
begin
if tcgsize2size[cgpara.Size] > 2 then
begin
if tcgsize2size[cgpara.Size] <> 4 then
internalerror(2014032401);
if cgpara.location^.Next = nil then
begin
if tcgsize2size[cgpara.location^.size] <> 4 then
internalerror(2014032404);
end
else
begin
if tcgsize2size[cgpara.location^.size] <> 2 then
internalerror(2014032405);
if tcgsize2size[cgpara.location^.Next^.size] <> 2 then
internalerror(2014032406);
if cgpara.location^.Next^.Next <> nil then
internalerror(2014032407);
end;
if cgpara.alignment > 4 then
internalerror(2014032408);
if segment<>NR_NO then
begin
list.concat(Taicpu.op_reg(A_PUSH,S_W,segment));
tmpref:=r;
tmpref.segment:=NR_NO;
tmpreg:=getaddressregister(list);
a_loadaddr_ref_reg(list,tmpref,tmpreg);
list.concat(taicpu.op_reg(A_PUSH,S_W,tmpreg));
end
else
begin
if (base=NR_NO) and (index=NR_NO) then
begin
if assigned(symbol) then
begin
tmpref:=r;
tmpref.refaddr:=addr_seg;
tmpref.offset:=0;
if current_settings.cputype < cpu_186 then
begin
tmpreg:=getaddressregister(list);
a_load_ref_reg(list,OS_16,OS_16,tmpref,tmpreg);
list.concat(taicpu.op_reg(A_PUSH,S_W,tmpreg));
end
else
list.concat(Taicpu.Op_ref(A_PUSH,S_W,tmpref));
if current_settings.cputype < cpu_186 then
begin
tmpreg:=getaddressregister(list);
a_loadaddr_ref_reg(list,r,tmpreg);
list.concat(taicpu.op_reg(A_PUSH,S_W,tmpreg));
end
else
list.concat(Taicpu.Op_sym_ofs(A_PUSH,S_W,symbol,offset));
end
else
internalerror(2014032402);
end
else if assigned(symbol) then
begin
reference_reset_symbol(tmpref,r.symbol,0,r.alignment,r.volatility);
tmpref.refaddr:=addr_seg;
if current_settings.cputype < cpu_186 then
begin
tmpreg:=getaddressregister(list);
a_load_ref_reg(list,OS_16,OS_16,tmpref,tmpreg);
list.concat(taicpu.op_reg(A_PUSH,S_W,tmpreg));
end
else
list.concat(Taicpu.Op_ref(A_PUSH,S_W,tmpref));
tmpreg:=getaddressregister(list);
a_loadaddr_ref_reg(list,r,tmpreg);
list.concat(taicpu.op_reg(A_PUSH,S_W,tmpreg));
end
else if (base=NR_BP) or (base=NR_SP) then
begin
list.concat(Taicpu.op_reg(A_PUSH,S_W,NR_SS));
tmpreg:=getaddressregister(list);
a_loadaddr_ref_reg(list,r,tmpreg);
list.concat(taicpu.op_reg(A_PUSH,S_W,tmpreg));
end
else
begin
list.concat(Taicpu.op_reg(A_PUSH,S_W,NR_DS));
tmpreg:=getaddressregister(list);
a_loadaddr_ref_reg(list,r,tmpreg);
list.concat(taicpu.op_reg(A_PUSH,S_W,tmpreg));
end;
end;
end
else
begin
cgpara.check_simple_location;
tmpref:=r;
tmpref.segment:=NR_NO;
with tmpref do
begin
if (base=NR_NO) and (index=NR_NO) then
begin
if assigned(symbol) then
begin
if current_settings.cputype < cpu_186 then
begin
tmpreg:=getaddressregister(list);
a_loadaddr_ref_reg(list,tmpref,tmpreg);
list.concat(taicpu.op_reg(A_PUSH,S_W,tmpreg));
end
else
list.concat(Taicpu.Op_sym_ofs(A_PUSH,S_W,symbol,offset));
end
else
push_const(list,OS_16,offset);
end
else if (base=NR_NO) and (index<>NR_NO) and
(offset=0) and (scalefactor=0) and (symbol=nil) then
list.concat(Taicpu.Op_reg(A_PUSH,S_W,index))
else if (base<>NR_NO) and (index=NR_NO) and
(offset=0) and (symbol=nil) then
list.concat(Taicpu.Op_reg(A_PUSH,S_W,base))
else
begin
tmpreg:=getaddressregister(list);
a_loadaddr_ref_reg(list,tmpref,tmpreg);
list.concat(taicpu.op_reg(A_PUSH,S_W,tmpreg));
end;
end;
end;
end
else
inherited a_loadaddr_ref_cgpara(list,r,cgpara);
end;
end;
procedure tcg8086.a_load_const_reg(list : TAsmList; tosize: tcgsize; a : tcgint;reg : tregister);
begin
check_register_size(tosize,reg);
if tosize in [OS_S32,OS_32] then
begin
list.concat(taicpu.op_const_reg(A_MOV,S_W,longint(a and $ffff),reg));
list.concat(taicpu.op_const_reg(A_MOV,S_W,longint(a shr 16),GetNextReg(reg)));
end
else
list.concat(taicpu.op_const_reg(A_MOV,TCGSize2OpSize[tosize],a,reg));
end;
procedure tcg8086.a_load_const_ref(list : TAsmList; tosize: tcgsize; a : tcgint;const ref : treference);
var
tmpref : treference;
begin
tmpref:=ref;
make_simple_ref(list,tmpref);
if tosize in [OS_S32,OS_32] then
begin
a_load_const_ref(list,OS_16,longint(a and $ffff),tmpref);
inc(tmpref.offset,2);
a_load_const_ref(list,OS_16,longint(a shr 16),tmpref);
end
else
list.concat(taicpu.op_const_ref(A_MOV,TCGSize2OpSize[tosize],a,tmpref));
end;
procedure tcg8086.a_load_reg_ref(list : TAsmList;fromsize,tosize: tcgsize; reg : tregister;const ref : treference);
var
tmpreg : tregister;
tmpref : treference;
begin
tmpref:=ref;
make_simple_ref(list,tmpref);
check_register_size(fromsize,reg);
case tosize of
OS_8,OS_S8:
if fromsize in [OS_8,OS_S8] then
list.concat(taicpu.op_reg_ref(A_MOV, S_B, reg, tmpref))
else
internalerror(2013030310);
OS_16,OS_S16:
case fromsize of
OS_8,OS_S8:
begin
tmpreg:=getintregister(list,tosize);
a_load_reg_reg(list,fromsize,tosize,reg,tmpreg);
a_load_reg_ref(list,tosize,tosize,tmpreg,tmpref);
end;
OS_16,OS_S16:
begin
list.concat(taicpu.op_reg_ref(A_MOV, S_W, reg, tmpref));
end;
else
internalerror(2013030312);
end;
OS_32,OS_S32:
case fromsize of
OS_8,OS_S8,OS_S16:
begin
tmpreg:=getintregister(list,tosize);
a_load_reg_reg(list,fromsize,tosize,reg,tmpreg);
a_load_reg_ref(list,tosize,tosize,tmpreg,tmpref);
end;
OS_16:
begin
{ Preload the ref base to reduce spilling }
if (tmpref.base<>NR_NO) and
(tmpref.index<>NR_NO) and
(getsupreg(tmpref.base)>=first_int_imreg) then
begin
tmpreg:=getaddressregister(list);
a_load_reg_reg(list,OS_ADDR,OS_ADDR,tmpref.base,tmpreg);
tmpref.base:=tmpreg;
end;
list.concat(taicpu.op_reg_ref(A_MOV, S_W, reg, tmpref));
inc(tmpref.offset, 2);
list.concat(taicpu.op_const_ref(A_MOV, S_W, 0, tmpref));
end;
OS_32,OS_S32:
begin
{ Preload the ref base to a new register to reduce spilling
Also preload if the first source reg is used as base or index
to prevent overwriting }
if ((tmpref.base<>NR_NO) and
(tmpref.index<>NR_NO) and
(getsupreg(tmpref.base)>=first_int_imreg)) or
(tmpref.base=reg) or
(tmpref.index=reg) then
begin
tmpreg:=getaddressregister(list);
a_load_reg_reg(list,OS_ADDR,OS_ADDR,tmpref.base,tmpreg);
tmpref.base:=tmpreg;
if tmpref.index=reg then
begin
list.concat(taicpu.op_ref_reg(A_LEA, S_W, tmpref, tmpref.base));
tmpref.index:=NR_NO;
tmpref.offset:=0;
tmpref.scalefactor:=0;
end;
end;
list.concat(taicpu.op_reg_ref(A_MOV, S_W, reg, tmpref));
inc(tmpref.offset, 2);
list.concat(taicpu.op_reg_ref(A_MOV, S_W, GetNextReg(reg), tmpref));
end;
else
internalerror(2013030313);
end;
else
internalerror(2013030311);
end;
end;
procedure tcg8086.a_load_ref_reg_internal(list : TAsmList;fromsize,tosize: tcgsize;const ref : treference;reg : tregister;isdirect:boolean);
procedure add_mov(instr: Taicpu);
begin
{ Notify the register allocator that we have written a move instruction so
it can try to eliminate it. }
if (instr.oper[0]^.reg<>current_procinfo.framepointer) and (instr.oper[0]^.reg<>NR_STACK_POINTER_REG) then
add_move_instruction(instr);
list.concat(instr);
end;
var
tmpref : treference;
tmpreg : tregister;
begin
tmpref:=ref;
make_simple_ref(list,tmpref,isdirect);
check_register_size(tosize,reg);
if (tcgsize2size[fromsize]>32) or (tcgsize2size[tosize]>32) or (fromsize=OS_NO) or (tosize=OS_NO) then
internalerror(2011021302);
{ if tcgsize2size[tosize]<=tcgsize2size[fromsize] then
fromsize:=tosize;}
case tosize of
OS_8,OS_S8:
if fromsize in [OS_8,OS_S8] then
list.concat(taicpu.op_ref_reg(A_MOV, S_B, tmpref, reg))
else
internalerror(2013030210);
OS_16,OS_S16:
case fromsize of
OS_8:
begin
if current_settings.cputype>=cpu_386 then
list.concat(taicpu.op_ref_reg(A_MOVZX, S_BW, tmpref, reg))
else
begin
reg := makeregsize(list, reg, OS_8);
list.concat(taicpu.op_ref_reg(A_MOV, S_B, tmpref, reg));
setsubreg(reg, R_SUBH);
list.concat(taicpu.op_const_reg(A_MOV, S_B, 0, reg));
makeregsize(list, reg, OS_16);
end;
end;
OS_S8:
begin
if current_settings.cputype>=cpu_386 then
list.concat(taicpu.op_ref_reg(A_MOVSX, S_BW, tmpref, reg))
else
begin
getcpuregister(list, NR_AX);
list.concat(taicpu.op_ref_reg(A_MOV, S_B, tmpref, NR_AL));
list.concat(taicpu.op_none(A_CBW));
ungetcpuregister(list, NR_AX);
add_mov(taicpu.op_reg_reg(A_MOV, S_W, NR_AX, reg));
end;
end;
OS_16,OS_S16:
list.concat(taicpu.op_ref_reg(A_MOV, S_W, tmpref, reg));
else
internalerror(2013030212);
end;
OS_32,OS_S32:
case fromsize of
OS_8:
begin
list.concat(taicpu.op_const_reg(A_MOV,S_W,0,GetNextReg(reg)));
if current_settings.cputype>=cpu_386 then
list.concat(taicpu.op_ref_reg(A_MOVZX, S_BW, tmpref, reg))
else
begin
reg := makeregsize(list, reg, OS_8);
list.concat(taicpu.op_ref_reg(A_MOV, S_B, tmpref, reg));
setsubreg(reg, R_SUBH);
list.concat(taicpu.op_const_reg(A_MOV, S_B, 0, reg));
makeregsize(list, reg, OS_16);
end;
end;
OS_S8:
begin
getcpuregister(list, NR_AX);
list.concat(taicpu.op_ref_reg(A_MOV, S_B, tmpref, NR_AL));
getcpuregister(list, NR_DX);
list.concat(taicpu.op_none(A_CBW));
list.concat(taicpu.op_none(A_CWD));
ungetcpuregister(list, NR_AX);
add_mov(taicpu.op_reg_reg(A_MOV, S_W, NR_AX, reg));
ungetcpuregister(list, NR_DX);
add_mov(taicpu.op_reg_reg(A_MOV, S_W, NR_DX, GetNextReg(reg)));
end;
OS_16:
begin
list.concat(taicpu.op_ref_reg(A_MOV, S_W, tmpref, reg));
list.concat(taicpu.op_const_reg(A_MOV,S_W,0,GetNextReg(reg)));
end;
OS_S16:
begin
getcpuregister(list, NR_AX);
list.concat(taicpu.op_ref_reg(A_MOV, S_W, tmpref, NR_AX));
getcpuregister(list, NR_DX);
list.concat(taicpu.op_none(A_CWD));
ungetcpuregister(list, NR_AX);
add_mov(taicpu.op_reg_reg(A_MOV, S_W, NR_AX, reg));
ungetcpuregister(list, NR_DX);
add_mov(taicpu.op_reg_reg(A_MOV, S_W, NR_DX, GetNextReg(reg)));
end;
OS_32,OS_S32:
begin
{ Preload the ref base to a new register to reduce spilling
Also preload if the first target reg is used as base or index
to prevent overwriting }
if ((tmpref.base<>NR_NO) and
(tmpref.index<>NR_NO) and
(getsupreg(tmpref.base)>=first_int_imreg)) or
(tmpref.base=reg) or
(tmpref.index=reg) then
begin
tmpreg:=getaddressregister(list);
a_load_reg_reg(list,OS_ADDR,OS_ADDR,tmpref.base,tmpreg);
tmpref.base:=tmpreg;
if tmpref.index=reg then
begin
list.concat(taicpu.op_ref_reg(A_LEA, S_W, tmpref, tmpref.base));
tmpref.index:=NR_NO;
tmpref.offset:=0;
tmpref.scalefactor:=0;
end;
end;
list.concat(taicpu.op_ref_reg(A_MOV, S_W, tmpref, reg));
inc(tmpref.offset, 2);
list.concat(taicpu.op_ref_reg(A_MOV, S_W, tmpref, GetNextReg(reg)));
end;
else
internalerror(2013030213);
end;
else
internalerror(2013030211);
end;
end;
procedure tcg8086.a_load_reg_reg(list : TAsmList;fromsize,tosize: tcgsize;reg1,reg2 : tregister);
procedure add_mov(instr: Taicpu);
begin
{ Notify the register allocator that we have written a move instruction so
it can try to eliminate it. }
if (instr.oper[0]^.reg<>current_procinfo.framepointer) and (instr.oper[0]^.reg<>NR_STACK_POINTER_REG) then
add_move_instruction(instr);
list.concat(instr);
end;
begin
check_register_size(fromsize,reg1);
check_register_size(tosize,reg2);
if tcgsize2size[fromsize]>tcgsize2size[tosize] then
begin
if tosize in [OS_32, OS_S32] then
internalerror(2013031801);
reg1:=makeregsize(list,reg1,tosize);
fromsize:=tosize;
end;
if (reg1<>reg2) or (fromsize<>tosize) then
begin
case tosize of
OS_8,OS_S8:
if fromsize in [OS_8,OS_S8] then
begin
if reg1<>reg2 then
add_mov(taicpu.op_reg_reg(A_MOV, S_B, reg1, reg2));
end
else
internalerror(2013030203);
OS_16,OS_S16:
case fromsize of
OS_8:
begin
if current_settings.cputype>=cpu_386 then
add_mov(taicpu.op_reg_reg(A_MOVZX, S_BW, reg1, reg2))
else
begin
reg2 := makeregsize(list, reg2, OS_8);
if reg1<>reg2 then
add_mov(taicpu.op_reg_reg(A_MOV, S_B, reg1, reg2));
setsubreg(reg2,R_SUBH);
list.concat(taicpu.op_const_reg(A_MOV, S_B, 0, reg2));
makeregsize(list, reg2, OS_16);
end;
end;
OS_S8:
begin
if current_settings.cputype>=cpu_386 then
add_mov(taicpu.op_reg_reg(A_MOVSX, S_BW, reg1, reg2))
else
begin
getcpuregister(list, NR_AX);
add_mov(taicpu.op_reg_reg(A_MOV, S_B, reg1, NR_AL));
list.concat(taicpu.op_none(A_CBW));
ungetcpuregister(list, NR_AX);
add_mov(taicpu.op_reg_reg(A_MOV, S_W, NR_AX, reg2));
end;
end;
OS_16,OS_S16:
begin
if reg1<>reg2 then
add_mov(taicpu.op_reg_reg(A_MOV, S_W, reg1, reg2));
end
else
internalerror(2013030205);
end;
OS_32,OS_S32:
case fromsize of
OS_8:
begin
list.concat(taicpu.op_const_reg(A_MOV, S_W, 0, GetNextReg(reg2)));
if current_settings.cputype>=cpu_386 then
add_mov(taicpu.op_reg_reg(A_MOVZX, S_BW, reg1, reg2))
else
begin
reg2 := makeregsize(list, reg2, OS_8);
if reg1<>reg2 then
add_mov(taicpu.op_reg_reg(A_MOV, S_B, reg1, reg2));
setsubreg(reg2,R_SUBH);
list.concat(taicpu.op_const_reg(A_MOV, S_B, 0, reg2));
makeregsize(list, reg2, OS_16);
end;
end;
OS_S8:
begin
getcpuregister(list, NR_AX);
add_mov(taicpu.op_reg_reg(A_MOV, S_B, reg1, NR_AL));
getcpuregister(list, NR_DX);
list.concat(taicpu.op_none(A_CBW));
list.concat(taicpu.op_none(A_CWD));
ungetcpuregister(list, NR_AX);
add_mov(taicpu.op_reg_reg(A_MOV, S_W, NR_AX, reg2));
ungetcpuregister(list, NR_DX);
add_mov(taicpu.op_reg_reg(A_MOV, S_W, NR_DX, GetNextReg(reg2)));
end;
OS_16:
begin
if reg1<>reg2 then
add_mov(taicpu.op_reg_reg(A_MOV, S_W, reg1, reg2));
list.concat(taicpu.op_const_reg(A_MOV,S_W,0,GetNextReg(reg2)));
end;
OS_S16:
begin
getcpuregister(list, NR_AX);
add_mov(taicpu.op_reg_reg(A_MOV, S_W, reg1, NR_AX));
getcpuregister(list, NR_DX);
list.concat(taicpu.op_none(A_CWD));
ungetcpuregister(list, NR_AX);
if reg1<>reg2 then
add_mov(taicpu.op_reg_reg(A_MOV, S_W, NR_AX, reg2));
ungetcpuregister(list, NR_DX);
add_mov(taicpu.op_reg_reg(A_MOV, S_W, NR_DX, GetNextReg(reg2)));
end;
OS_32,OS_S32:
begin
if reg1<>reg2 then
begin
add_mov(taicpu.op_reg_reg(A_MOV, S_W, reg1, reg2));
add_mov(taicpu.op_reg_reg(A_MOV, S_W, GetNextReg(reg1), GetNextReg(reg2)));
end;
end;
else
internalerror(2013030206);
end;
else
internalerror(2013030204);
end;
end;
end;
procedure tcg8086.a_cmp_const_reg_label(list: TAsmList; size: tcgsize; cmp_op: topcmp; a: tcgint; reg: tregister; l: tasmlabel);
var
hl_skip: TAsmLabel;
begin
if size in [OS_32, OS_S32] then
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
if (longint(a shr 16) = 0) then
list.concat(taicpu.op_reg_reg(A_TEST,S_W,GetNextReg(reg),GetNextReg(reg)))
else
list.concat(taicpu.op_const_reg(A_CMP,S_W,longint(a shr 16),GetNextReg(reg)));
current_asmdata.getjumplabel(hl_skip);
gen_cmp32_jmp1(list, cmp_op, hl_skip, l);
if (longint(a and $ffff) = 0) then
list.concat(taicpu.op_reg_reg(A_TEST,S_W,reg,reg))
else
list.concat(taicpu.op_const_reg(A_CMP,S_W,longint(a and $ffff),reg));
gen_cmp32_jmp2(list, cmp_op, hl_skip, l);
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
a_label(list,hl_skip);
end
else
inherited a_cmp_const_reg_label(list, size, cmp_op, a, reg, l);
end;
procedure tcg8086.a_cmp_const_ref_label(list: TAsmList; size: tcgsize; cmp_op: topcmp; a: tcgint; const ref: treference; l: tasmlabel);
var
tmpref: treference;
hl_skip: TAsmLabel;
begin
if size in [OS_32, OS_S32] then
begin
tmpref:=ref;
make_simple_ref(list,tmpref);
inc(tmpref.offset,2);
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_ref(A_CMP,S_W,longint(a shr 16),tmpref));
current_asmdata.getjumplabel(hl_skip);
gen_cmp32_jmp1(list, cmp_op, hl_skip, l);
dec(tmpref.offset,2);
list.concat(taicpu.op_const_ref(A_CMP,S_W,longint(a and $ffff),tmpref));
gen_cmp32_jmp2(list, cmp_op, hl_skip, l);
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
a_label(list,hl_skip);
end
else
inherited a_cmp_const_ref_label(list, size, cmp_op, a, ref, l);
end;
procedure tcg8086.a_cmp_reg_reg_label(list: TAsmList; size: tcgsize; cmp_op: topcmp; reg1, reg2: tregister; l: tasmlabel);
var
hl_skip: TAsmLabel;
begin
if size in [OS_32, OS_S32] then
begin
check_register_size(size,reg1);
check_register_size(size,reg2);
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_reg_reg(A_CMP,S_W,GetNextReg(reg1),GetNextReg(reg2)));
current_asmdata.getjumplabel(hl_skip);
gen_cmp32_jmp1(list, cmp_op, hl_skip, l);
list.concat(taicpu.op_reg_reg(A_CMP,S_W,reg1,reg2));
gen_cmp32_jmp2(list, cmp_op, hl_skip, l);
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
a_label(list,hl_skip);
end
else
inherited a_cmp_reg_reg_label(list, size, cmp_op, reg1, reg2, l);
end;
procedure tcg8086.a_cmp_ref_reg_label(list: TAsmList; size: tcgsize; cmp_op: topcmp; const ref: treference; reg: tregister; l: tasmlabel);
var
tmpref: treference;
hl_skip: TAsmLabel;
begin
if size in [OS_32, OS_S32] then
begin
tmpref:=ref;
make_simple_ref(list,tmpref);
check_register_size(size,reg);
inc(tmpref.offset,2);
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_ref_reg(A_CMP,S_W,tmpref,GetNextReg(reg)));
current_asmdata.getjumplabel(hl_skip);
gen_cmp32_jmp1(list, cmp_op, hl_skip, l);
dec(tmpref.offset,2);
list.concat(taicpu.op_ref_reg(A_CMP,S_W,tmpref,reg));
gen_cmp32_jmp2(list, cmp_op, hl_skip, l);
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
a_label(list,hl_skip);
end
else
inherited a_cmp_ref_reg_label(list, size, cmp_op, ref, reg, l);
end;
procedure tcg8086.a_cmp_reg_ref_label(list: TAsmList; size: tcgsize; cmp_op: topcmp; reg: tregister; const ref: treference; l: tasmlabel);
var
tmpref: treference;
hl_skip: TAsmLabel;
begin
if size in [OS_32, OS_S32] then
begin
tmpref:=ref;
make_simple_ref(list,tmpref);
check_register_size(size,reg);
inc(tmpref.offset,2);
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_reg_ref(A_CMP,S_W,GetNextReg(reg),tmpref));
current_asmdata.getjumplabel(hl_skip);
gen_cmp32_jmp1(list, cmp_op, hl_skip, l);
dec(tmpref.offset,2);
list.concat(taicpu.op_reg_ref(A_CMP,S_W,reg,tmpref));
gen_cmp32_jmp2(list, cmp_op, hl_skip, l);
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
a_label(list,hl_skip);
end
else
inherited a_cmp_reg_ref_label(list, size, cmp_op, reg, ref, l);
end;
procedure tcg8086.gen_cmp32_jmp1(list: TAsmList; cmp_op: topcmp; l_skip, l_target: TAsmLabel);
begin
case cmp_op of
OC_EQ:
a_jmp_cond(list, OC_NE, l_skip);
OC_NE:
a_jmp_cond(list, OC_NE, l_target);
OC_GT,OC_GTE:
begin
a_jmp_cond(list, OC_GT, l_target);
a_jmp_cond(list, OC_LT, l_skip);
end;
OC_LT,OC_LTE:
begin
a_jmp_cond(list, OC_LT, l_target);
a_jmp_cond(list, OC_GT, l_skip);
end;
OC_B,OC_BE:
begin
a_jmp_cond(list, OC_B, l_target);
a_jmp_cond(list, OC_A, l_skip);
end;
OC_A,OC_AE:
begin
a_jmp_cond(list, OC_A, l_target);
a_jmp_cond(list, OC_B, l_skip);
end;
else
internalerror(2014010321);
end;
end;
procedure tcg8086.gen_cmp32_jmp2(list: TAsmList; cmp_op: topcmp; l_skip, l_target: TAsmLabel);
begin
case cmp_op of
OC_EQ:
a_jmp_cond(list, OC_EQ, l_target);
OC_GT:
a_jmp_cond(list, OC_A, l_target);
OC_LT:
a_jmp_cond(list, OC_B, l_target);
OC_GTE:
a_jmp_cond(list, OC_AE, l_target);
OC_LTE:
a_jmp_cond(list, OC_BE, l_target);
OC_NE:
a_jmp_cond(list, OC_NE, l_target);
OC_BE:
a_jmp_cond(list, OC_BE, l_target);
OC_B:
a_jmp_cond(list, OC_B, l_target);
OC_AE:
a_jmp_cond(list, OC_AE, l_target);
OC_A:
a_jmp_cond(list, OC_A, l_target);
else
internalerror(2014010322);
end;
end;
procedure tcg8086.g_flags2reg(list: TAsmList; size: TCgSize; const f: tresflags; reg: TRegister);
var
ai : taicpu;
hreg16 : tregister;
hl_skip: TAsmLabel;
invf: TResFlags;
tmpsize: TCgSize;
tmpopsize: topsize;
begin
{ optimized case for the carry flag, using ADC/RCL }
if f in [F_C,F_B,F_FB] then
begin
case size of
OS_8,OS_S8:
begin
tmpsize:=OS_8;
tmpopsize:=S_B;
end;
OS_16,OS_S16,OS_32,OS_S32:
begin
tmpsize:=OS_16;
tmpopsize:=S_W;
end;
else
internalerror(2013123101);
end;
list.concat(Taicpu.op_const_reg(A_MOV, tmpopsize, 0, reg));
hl_skip:=nil;
if f=F_FB then
begin
current_asmdata.getjumplabel(hl_skip);
ai:=Taicpu.op_sym(A_Jcc,S_NO,hl_skip);
ai.SetCondition(C_P);
ai.is_jmp:=true;
list.concat(ai);
end;
{ RCL is faster than ADC on 8086/8088. On the 80286, it is
equally fast and it also has the same size. In these cases,
we still prefer it over ADC, because it's a better choice in
case the register is spilled. }
if (cs_opt_size in current_settings.optimizerswitches) or
(current_settings.optimizecputype<=cpu_286) then
list.concat(Taicpu.op_const_reg(A_RCL, tmpopsize, 1, reg))
else
{ ADC is much faster on the 386. }
list.concat(Taicpu.op_reg_reg(A_ADC, tmpopsize, reg, reg));
if f=F_FB then
a_label(list,hl_skip);
a_load_reg_reg(list,tmpsize,size,reg,reg);
end
{ optimized case for the inverted carry flag, using SBB }
else if f in [F_NC,F_AE,F_FAE] then
begin
case size of
OS_8,OS_S8:
begin
tmpsize:=OS_8;
list.concat(Taicpu.op_const_reg(A_MOV, S_B, 1, reg));
list.concat(Taicpu.op_const_reg(A_SBB, S_B, 0, reg));
end;
OS_16,OS_S16,OS_32,OS_S32:
begin
tmpsize:=OS_16;
list.concat(Taicpu.op_const_reg(A_MOV, S_W, 1, reg));
list.concat(Taicpu.op_const_reg(A_SBB, S_W, 0, reg));
end;
else
internalerror(2013123102);
end;
a_load_reg_reg(list,tmpsize,size,reg,reg);
end
else
begin
invf := f;
inverse_flags(invf);
case size of
OS_8,OS_S8:
begin
tmpsize:=OS_8;
list.concat(Taicpu.op_const_reg(A_MOV, S_B, 0, reg));
end;
OS_16,OS_S16,OS_32,OS_S32:
begin
tmpsize:=OS_16;
list.concat(Taicpu.op_const_reg(A_MOV, S_W, 0, reg));
end;
else
internalerror(2013123103);
end;
current_asmdata.getjumplabel(hl_skip);
{ we can't just forward invf to a_jmp_flags for FA,FAE,FB and FBE, because
in the case of NaNs:
not(F_FA )<>F_FBE
not(F_FAE)<>F_FB
not(F_FB )<>F_FAE
not(F_FBE)<>F_FA
}
case f of
F_FA:
invf:=FPUFlags2Flags[invf];
F_FAE,F_FB:
{ F_FAE and F_FB are handled above, using ADC/RCL/SBB }
internalerror(2015102101);
F_FBE:
begin
ai:=Taicpu.op_sym(A_Jcc,S_NO,hl_skip);
ai.SetCondition(C_P);
ai.is_jmp:=true;
list.concat(ai);
invf:=FPUFlags2Flags[invf];
end;
else
;
end;
a_jmp_flags(list,invf,hl_skip);
{ 16-bit INC is shorter than 8-bit }
hreg16:=makeregsize(list,reg,OS_16);
list.concat(Taicpu.op_reg(A_INC, S_W, hreg16));
makeregsize(list,hreg16,tmpsize);
a_label(list,hl_skip);
a_load_reg_reg(list,tmpsize,size,reg,reg);
end;
end;
procedure tcg8086.g_flags2ref(list: TAsmList; size: TCgSize; const f: tresflags; const ref: TReference);
var
tmpreg : tregister;
tmpregsize: TCgSize;
tmpref: treference;
begin
if size in [OS_8,OS_S8,OS_16,OS_S16] then
tmpregsize:=size
else
tmpregsize:=OS_16;
tmpreg:=getintregister(list,tmpregsize);
g_flags2reg(list,tmpregsize,f,tmpreg);
tmpref:=ref;
make_simple_ref(list,tmpref);
if size in [OS_64,OS_S64] then
begin
a_load_reg_ref(list,tmpregsize,OS_32,tmpreg,tmpref);
inc(tmpref.offset,4);
a_load_const_ref(list,OS_32,0,tmpref);
end
else
a_load_reg_ref(list,tmpregsize,size,tmpreg,tmpref);
end;
procedure tcg8086.g_stackpointer_alloc(list : TAsmList;localsize: longint);
begin
if cs_check_stack in current_settings.localswitches then
begin
cg.getcpuregister(list,NR_AX);
cg.a_load_const_reg(list,OS_16, localsize,NR_AX);
cg.a_call_name(list,'FPC_STACKCHECK_I8086',false);
cg.ungetcpuregister(list, NR_AX);
end;
if localsize>0 then
list.concat(Taicpu.Op_const_reg(A_SUB,S_W,localsize,NR_STACK_POINTER_REG));
end;
procedure tcg8086.g_proc_exit(list : TAsmList;parasize:longint;nostackframe:boolean);
var
stacksize : longint;
ret_instr: TAsmOp;
sp_moved : boolean;
procedure maybe_move_sp;
var
ref : treference;
begin
if sp_moved then
exit;
if not(pi_has_open_array_parameter in current_procinfo.flags) then
exit;
{ Restore SP position before SP change }
if current_settings.x86memorymodel=mm_huge then
stacksize:=stacksize + 2;
reference_reset_base(ref,NR_BP,-stacksize,ctempposinvalid,2,[]);
list.concat(Taicpu.op_ref_reg(A_LEA,S_W,ref,NR_SP));
sp_moved:=true;
end;
begin
if is_proc_far(current_procinfo.procdef) then
ret_instr:=A_RETF
else
ret_instr:=A_RET;
{ MMX needs to call EMMS }
if assigned(rg[R_MMXREGISTER]) and
(rg[R_MMXREGISTER].uses_registers) then
list.concat(Taicpu.op_none(A_EMMS,S_NO));
sp_moved:=false;
{ remove stackframe }
if not nostackframe then
begin
stacksize:=current_procinfo.calc_stackframe_size;
if (target_info.stackalign>4) and
((stacksize <> 0) or
(pi_do_call in current_procinfo.flags) or
{ can't detect if a call in this case -> use nostackframe }
{ if you (think you) know what you are doing }
(po_assembler in current_procinfo.procdef.procoptions)) then
stacksize := align(stacksize+sizeof(aint),target_info.stackalign) - sizeof(aint);
if (po_exports in current_procinfo.procdef.procoptions) and
(target_info.system=system_i8086_win16) then
begin
maybe_move_sp;
list.concat(Taicpu.Op_reg(A_POP,S_W,NR_DI));
list.concat(Taicpu.Op_reg(A_POP,S_W,NR_SI));
end;
if ((current_settings.x86memorymodel=mm_huge) and
not (po_interrupt in current_procinfo.procdef.procoptions)) or
((po_exports in current_procinfo.procdef.procoptions) and
(target_info.system=system_i8086_win16)) then
begin
maybe_move_sp;
list.concat(Taicpu.Op_reg(A_POP,S_W,NR_DS));
end;
if (current_procinfo.framepointer=NR_STACK_POINTER_REG) then
begin
if (stacksize<>0) then
cg.a_op_const_reg(list,OP_ADD,OS_ADDR,stacksize,current_procinfo.framepointer);
end
else
begin
generate_leave(list);
if ((ts_x86_far_procs_push_odd_bp in current_settings.targetswitches) or
((po_exports in current_procinfo.procdef.procoptions) and
(target_info.system=system_i8086_win16))) and
is_proc_far(current_procinfo.procdef) then
cg.a_op_const_reg(list,OP_SUB,OS_ADDR,1,current_procinfo.framepointer);
end;
list.concat(tai_regalloc.dealloc(current_procinfo.framepointer,nil));
end;
{ return from interrupt }
if po_interrupt in current_procinfo.procdef.procoptions then
begin
list.concat(Taicpu.Op_reg(A_POP,S_W,NR_ES));
list.concat(Taicpu.Op_reg(A_POP,S_W,NR_DS));
list.concat(Taicpu.Op_reg(A_POP,S_W,NR_DI));
list.concat(Taicpu.Op_reg(A_POP,S_W,NR_SI));
list.concat(Taicpu.Op_reg(A_POP,S_W,NR_DX));
list.concat(Taicpu.Op_reg(A_POP,S_W,NR_CX));
list.concat(Taicpu.Op_reg(A_POP,S_W,NR_BX));
list.concat(Taicpu.Op_reg(A_POP,S_W,NR_AX));
list.concat(Taicpu.Op_none(A_IRET,S_NO));
end
{ Routines with the poclearstack flag set use only a ret }
else if (current_procinfo.procdef.proccalloption in clearstack_pocalls) and
(not paramanager.use_fixed_stack) then
begin
{ complex return values are removed from stack in C code PM }
{ but not on win32 }
{ and not for safecall with hidden exceptions, because the result }
{ wich contains the exception is passed in EAX }
if (target_info.system <> system_i386_win32) and
not ((current_procinfo.procdef.proccalloption = pocall_safecall) and
(tf_safecall_exceptions in target_info.flags)) and
paramanager.ret_in_param(current_procinfo.procdef.returndef,
current_procinfo.procdef) then
list.concat(Taicpu.Op_const(ret_instr,S_W,sizeof(aint)))
else
list.concat(Taicpu.Op_none(ret_instr,S_NO));
end
{ ... also routines with parasize=0 }
else if (parasize=0) then
list.concat(Taicpu.Op_none(ret_instr,S_NO))
else
begin
{ parameters are limited to 65535 bytes because ret allows only imm16 }
if (parasize>65535) then
CGMessage(cg_e_parasize_too_big);
list.concat(Taicpu.Op_const(ret_instr,S_W,parasize));
end;
end;
procedure tcg8086.g_copyvaluepara_openarray(list : TAsmList;const ref:treference;const lenloc:tlocation;elesize:tcgint;destreg:tregister);
var
power : longint;
opsize : topsize;
saved_ds: Boolean;
begin
{ get stack space }
getcpuregister(list,NR_DI);
a_load_loc_reg(list,OS_INT,lenloc,NR_DI);
list.concat(Taicpu.op_reg(A_INC,S_W,NR_DI));
{ Now DI contains (high+1). }
include(current_procinfo.flags, pi_has_open_array_parameter);
{ special case handling for elesize=2:
set CX = (high+1) instead of CX = (high+1)*elesize.
This allows us to avoid the SHR later. }
if elesize=2 then
begin
{ Now DI contains (high+1). Copy it to CX for later use. }
getcpuregister(list,NR_CX);
list.concat(Taicpu.op_reg_reg(A_MOV,S_W,NR_DI,NR_CX));
end;
{ DI := DI * elesize }
if (elesize<>1) then
begin
if ispowerof2(elesize, power) then
a_op_const_reg(list,OP_SHL,OS_16,power,NR_DI)
else
a_op_const_reg(list,OP_IMUL,OS_16,elesize,NR_DI);
end;
if elesize<>2 then
begin
{ Now DI contains (high+1)*elesize. Copy it to CX for later use. }
getcpuregister(list,NR_CX);
list.concat(Taicpu.op_reg_reg(A_MOV,S_W,NR_DI,NR_CX));
end;
{ If we were probing pages, EDI=(size mod pagesize) and ESP is decremented
by (size div pagesize)*pagesize, otherwise EDI=size.
Either way, subtracting EDI from ESP will set ESP to desired final value. }
list.concat(Taicpu.op_reg_reg(A_SUB,S_W,NR_DI,NR_SP));
{ align stack on 2 bytes }
list.concat(Taicpu.op_const_reg(A_AND,S_W,aint($fffe),NR_SP));
{ load destination, don't use a_load_reg_reg, that will add a move instruction
that can confuse the reg allocator }
list.concat(Taicpu.Op_reg_reg(A_MOV,S_W,NR_SP,NR_DI));
{$ifdef volatile_es}
list.concat(taicpu.op_reg(A_PUSH,S_W,NR_SS));
list.concat(taicpu.op_reg(A_POP,S_W,NR_ES));
{$endif volatile_es}
{ Allocate SI and load it with source }
getcpuregister(list,NR_SI);
if ((ref.segment=NR_NO) and (segment_regs_equal(NR_SS,NR_DS) or (ref.base<>NR_BP))) or
(is_segment_reg(ref.segment) and segment_regs_equal(ref.segment,NR_DS)) then
begin
hlcg.a_loadaddr_ref_reg(list,voidnearpointertype,voidnearpointertype,ref,NR_SI);
saved_ds:=false;
end
else
begin
hlcg.a_loadaddr_ref_reg(list,voidnearpointertype,voidnearpointertype,ref,NR_SI);
list.concat(taicpu.op_reg(A_PUSH,S_W,NR_DS));
saved_ds:=true;
if ref.segment<>NR_NO then
list.concat(taicpu.op_reg(A_PUSH,S_W,ref.segment))
else if ref.base=NR_BP then
list.concat(taicpu.op_reg(A_PUSH,S_W,NR_SS))
else
internalerror(2014040403);
list.concat(taicpu.op_reg(A_POP,S_W,NR_DS));
end;
{ calculate size }
opsize:=S_B;
if elesize=2 then
begin
opsize:=S_W;
{ CX is already number of words, so no need to SHL/SHR }
end
else if (elesize and 1)=0 then
begin
opsize:=S_W;
{ CX is number of bytes, convert to words }
list.concat(Taicpu.op_const_reg(A_SHR,S_W,1,NR_CX))
end;
if ts_cld in current_settings.targetswitches then
list.concat(Taicpu.op_none(A_CLD,S_NO));
if (opsize=S_B) and not (cs_opt_size in current_settings.optimizerswitches) then
begin
{ SHR CX,1 moves the lowest (odd/even) bit to the carry flag }
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(Taicpu.op_const_reg(A_SHR,S_W,1,NR_CX));
list.concat(Taicpu.op_none(A_REP,S_NO));
list.concat(Taicpu.op_none(A_MOVSW,S_NO));
{ ADC CX,CX will set CX to 1 if the number of bytes was odd }
list.concat(Taicpu.op_reg_reg(A_ADC,S_W,NR_CX,NR_CX));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
list.concat(Taicpu.op_none(A_REP,S_NO));
list.concat(Taicpu.op_none(A_MOVSB,S_NO));
end
else
begin
list.concat(Taicpu.op_none(A_REP,S_NO));
case opsize of
S_B : list.concat(Taicpu.Op_none(A_MOVSB,S_NO));
S_W : list.concat(Taicpu.Op_none(A_MOVSW,S_NO));
else
internalerror(2019051019);
end;
end;
ungetcpuregister(list,NR_DI);
ungetcpuregister(list,NR_CX);
ungetcpuregister(list,NR_SI);
if saved_ds then
list.concat(taicpu.op_reg(A_POP,S_W,NR_DS));
{ patch the new address, but don't use a_load_reg_reg, that will add a move instruction
that can confuse the reg allocator }
list.concat(Taicpu.Op_reg_reg(A_MOV,S_W,NR_SP,destreg));
if current_settings.x86memorymodel in x86_far_data_models then
list.concat(Taicpu.Op_reg_reg(A_MOV,S_W,NR_SS,GetNextReg(destreg)));
end;
procedure tcg8086.g_releasevaluepara_openarray(list : TAsmList;const l:tlocation);
begin
{ Nothing to do }
end;
procedure tcg8086.get_32bit_ops(op: TOpCG; out op1, op2: TAsmOp);
begin
case op of
OP_ADD :
begin
op1:=A_ADD;
op2:=A_ADC;
end;
OP_SUB :
begin
op1:=A_SUB;
op2:=A_SBB;
end;
OP_XOR :
begin
op1:=A_XOR;
op2:=A_XOR;
end;
OP_OR :
begin
op1:=A_OR;
op2:=A_OR;
end;
OP_AND :
begin
op1:=A_AND;
op2:=A_AND;
end;
else
internalerror(200203241);
end;
end;
procedure tcg8086.g_adjust_self_value(list:TAsmList;procdef: tprocdef;ioffset: tcgint);
var
hsym : tsym;
href : treference;
paraloc : Pcgparalocation;
return_address_size: Integer;
begin
if current_settings.x86memorymodel in x86_far_code_models then
return_address_size:=4
else
return_address_size:=2;
{ calculate the parameter info for the procdef }
procdef.init_paraloc_info(callerside);
hsym:=tsym(procdef.parast.Find('self'));
if not(assigned(hsym) and
(hsym.typ=paravarsym)) then
internalerror(2003052502);
paraloc:=tparavarsym(hsym).paraloc[callerside].location;
with paraloc^ do
begin
case loc of
LOC_REGISTER:
a_op_const_reg(list,OP_SUB,size,ioffset,register);
LOC_REFERENCE:
begin
{ offset in the wrapper needs to be adjusted for the stored
return address }
if (reference.index<>NR_BP) and (reference.index<>NR_BX) and (reference.index<>NR_DI)
and (reference.index<>NR_SI) then
begin
list.concat(taicpu.op_reg(A_PUSH,S_W,NR_DI));
list.concat(taicpu.op_reg_reg(A_MOV,S_W,reference.index,NR_DI));
if reference.index=NR_SP then
reference_reset_base(href,NR_DI,reference.offset+return_address_size+2,ctempposinvalid,sizeof(pint),[])
else
reference_reset_base(href,NR_DI,reference.offset+return_address_size,ctempposinvalid,sizeof(pint),[]);
href.segment:=NR_SS;
a_op_const_ref(list,OP_SUB,size,ioffset,href);
list.concat(taicpu.op_reg(A_POP,S_W,NR_DI));
end
else
begin
reference_reset_base(href,reference.index,reference.offset+return_address_size,ctempposinvalid,sizeof(pint),[]);
href.segment:=NR_SS;
a_op_const_ref(list,OP_SUB,size,ioffset,href);
end;
end
else
internalerror(2003091802);
end;
paraloc:=next;
end;
end;
{ ************* 64bit operations ************ }
procedure tcg64f8086.get_64bit_ops(op:TOpCG;var op1,op2:TAsmOp);
begin
case op of
OP_ADD :
begin
op1:=A_ADD;
op2:=A_ADC;
end;
OP_SUB :
begin
op1:=A_SUB;
op2:=A_SBB;
end;
OP_XOR :
begin
op1:=A_XOR;
op2:=A_XOR;
end;
OP_OR :
begin
op1:=A_OR;
op2:=A_OR;
end;
OP_AND :
begin
op1:=A_AND;
op2:=A_AND;
end;
else
internalerror(2002032403);
end;
end;
procedure tcg64f8086.a_op64_ref_reg(list : TAsmList;op:TOpCG;size : tcgsize;const ref : treference;reg : tregister64);
var
op1,op2 : TAsmOp;
tempref : treference;
begin
if not(op in [OP_NEG,OP_NOT]) then
begin
get_64bit_ops(op,op1,op2);
tempref:=ref;
tcgx86(cg).make_simple_ref(list,tempref);
if op in [OP_ADD,OP_SUB] then
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_ref_reg(op1,S_W,tempref,reg.reglo));
inc(tempref.offset,2);
list.concat(taicpu.op_ref_reg(op2,S_W,tempref,cg.GetNextReg(reg.reglo)));
inc(tempref.offset,2);
list.concat(taicpu.op_ref_reg(op2,S_W,tempref,reg.reghi));
inc(tempref.offset,2);
list.concat(taicpu.op_ref_reg(op2,S_W,tempref,cg.GetNextReg(reg.reghi)));
if op in [OP_ADD,OP_SUB] then
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end
else
begin
a_load64_ref_reg(list,ref,reg);
a_op64_reg_reg(list,op,size,reg,reg);
end;
end;
procedure tcg64f8086.a_op64_reg_ref(list : TAsmList;op:TOpCG;size : tcgsize;reg : tregister64; const ref: treference);
var
op1,op2 : TAsmOp;
tempref : treference;
begin
case op of
OP_NOT,OP_NEG:
inherited;
else
begin
get_64bit_ops(op,op1,op2);
tempref:=ref;
tcgx86(cg).make_simple_ref(list,tempref);
if op in [OP_ADD,OP_SUB] then
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_reg_ref(op1,S_W,reg.reglo,tempref));
inc(tempref.offset,2);
list.concat(taicpu.op_reg_ref(op2,S_W,cg.GetNextReg(reg.reglo),tempref));
inc(tempref.offset,2);
list.concat(taicpu.op_reg_ref(op2,S_W,reg.reghi,tempref));
inc(tempref.offset,2);
list.concat(taicpu.op_reg_ref(op2,S_W,cg.GetNextReg(reg.reghi),tempref));
if op in [OP_ADD,OP_SUB] then
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
end;
end;
procedure tcg64f8086.a_op64_reg_reg(list : TAsmList;op:TOpCG;size : tcgsize;regsrc,regdst : tregister64);
var
op1,op2 : TAsmOp;
l2, l3: TAsmLabel;
ai: taicpu;
begin
case op of
OP_NEG :
begin
if (regsrc.reglo<>regdst.reglo) then
a_load64_reg_reg(list,regsrc,regdst);
cg.a_op_reg_reg(list,OP_NOT,OS_32,regdst.reghi,regdst.reghi);
list.concat(taicpu.op_reg(A_NOT,S_W,cg.GetNextReg(regdst.reglo)));
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_reg(A_NEG,S_W,regdst.reglo));
list.concat(taicpu.op_const_reg(A_SBB,S_W,-1,cg.GetNextReg(regdst.reglo)));
list.concat(taicpu.op_const_reg(A_SBB,S_W,-1,regdst.reghi));
list.concat(taicpu.op_const_reg(A_SBB,S_W,-1,cg.GetNextReg(regdst.reghi)));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
exit;
end;
OP_NOT :
begin
if (regsrc.reglo<>regdst.reglo) then
a_load64_reg_reg(list,regsrc,regdst);
cg.a_op_reg_reg(list,OP_NOT,OS_32,regdst.reglo,regdst.reglo);
cg.a_op_reg_reg(list,OP_NOT,OS_32,regdst.reghi,regdst.reghi);
exit;
end;
OP_SHR,OP_SHL,OP_SAR:
begin
{ load right operators in a register }
cg.getcpuregister(list,NR_CX);
cg.a_load_reg_reg(list,OS_16,OS_16,regsrc.reglo,NR_CX);
current_asmdata.getjumplabel(l2);
current_asmdata.getjumplabel(l3);
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_AND,S_W,63,NR_CX));
cg.a_jmp_flags(list,F_E,l3);
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
cg.a_label(list,l2);
case op of
OP_SHL:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_SHL,S_W,1,regdst.reglo));
list.concat(taicpu.op_const_reg(A_RCL,S_W,1,cg.GetNextReg(regdst.reglo)));
list.concat(taicpu.op_const_reg(A_RCL,S_W,1,regdst.reghi));
list.concat(taicpu.op_const_reg(A_RCL,S_W,1,cg.GetNextReg(regdst.reghi)));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_SHR,OP_SAR:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
cg.a_op_const_reg(list,op,OS_16,1,cg.GetNextReg(regdst.reghi));
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,regdst.reghi));
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,cg.GetNextReg(regdst.reglo)));
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,regdst.reglo));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
else
internalerror(2019051018);
end;
ai:=Taicpu.Op_Sym(A_LOOP,S_W,l2);
ai.is_jmp := True;
list.Concat(ai);
cg.a_label(list,l3);
cg.ungetcpuregister(list,NR_CX);
exit;
end;
else
;
end;
get_64bit_ops(op,op1,op2);
if op in [OP_ADD,OP_SUB] then
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_reg_reg(op1,S_W,regsrc.reglo,regdst.reglo));
list.concat(taicpu.op_reg_reg(op2,S_W,cg.GetNextReg(regsrc.reglo),cg.GetNextReg(regdst.reglo)));
list.concat(taicpu.op_reg_reg(op2,S_W,regsrc.reghi,regdst.reghi));
list.concat(taicpu.op_reg_reg(op2,S_W,cg.GetNextReg(regsrc.reghi),cg.GetNextReg(regdst.reghi)));
if op in [OP_ADD,OP_SUB] then
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
procedure tcg64f8086.a_op64_const_reg(list : TAsmList;op:TOpCG;size : tcgsize;value : int64;reg : tregister64);
var
op1,op2 : TAsmOp;
loop_start: TAsmLabel;
ai: taicpu;
begin
case op of
OP_AND,OP_OR,OP_XOR:
begin
cg.a_op_const_reg(list,op,OS_32,tcgint(lo(value)),reg.reglo);
cg.a_op_const_reg(list,op,OS_32,tcgint(hi(value)),reg.reghi);
end;
OP_ADD, OP_SUB:
begin
get_64bit_ops(op,op1,op2);
if (value and $ffffffffffff) = 0 then
begin
{ use a_op_const_reg to allow the use of inc/dec }
cg.a_op_const_reg(list,op,OS_16,aint((value shr 48) and $ffff),cg.GetNextReg(reg.reghi));
end
// can't use a_op_const_ref because this may use dec/inc
else if (value and $ffffffff) = 0 then
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(op1,S_W,aint((value shr 32) and $ffff),reg.reghi));
list.concat(taicpu.op_const_reg(op2,S_W,aint((value shr 48) and $ffff),cg.GetNextReg(reg.reghi)));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end
else if (value and $ffff) = 0 then
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(op1,S_W,aint((value shr 16) and $ffff),cg.GetNextReg(reg.reglo)));
list.concat(taicpu.op_const_reg(op2,S_W,aint((value shr 32) and $ffff),reg.reghi));
list.concat(taicpu.op_const_reg(op2,S_W,aint((value shr 48) and $ffff),cg.GetNextReg(reg.reghi)));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end
else
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(op1,S_W,aint(value and $ffff),reg.reglo));
list.concat(taicpu.op_const_reg(op2,S_W,aint((value shr 16) and $ffff),cg.GetNextReg(reg.reglo)));
list.concat(taicpu.op_const_reg(op2,S_W,aint((value shr 32) and $ffff),reg.reghi));
list.concat(taicpu.op_const_reg(op2,S_W,aint((value shr 48) and $ffff),cg.GetNextReg(reg.reghi)));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
end;
OP_SHR,OP_SHL,OP_SAR:
begin
value:=value and 63;
case value of
0:
{ ultra hyper fast shift by 0 };
1:
case op of
OP_SHL:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_SHL,S_W,1,reg.reglo));
list.concat(taicpu.op_const_reg(A_RCL,S_W,1,cg.GetNextReg(reg.reglo)));
list.concat(taicpu.op_const_reg(A_RCL,S_W,1,reg.reghi));
list.concat(taicpu.op_const_reg(A_RCL,S_W,1,cg.GetNextReg(reg.reghi)));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_SHR,OP_SAR:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
cg.a_op_const_reg(list,op,OS_16,1,cg.GetNextReg(reg.reghi));
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,reg.reghi));
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,cg.GetNextReg(reg.reglo)));
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,reg.reglo));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
else
internalerror(2019051017);
end;
2..15:
begin
cg.getcpuregister(list,NR_CX);
cg.a_load_const_reg(list,OS_16,value,NR_CX);
current_asmdata.getjumplabel(loop_start);
cg.a_label(list,loop_start);
case op of
OP_SHL:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_SHL,S_W,1,reg.reglo));
list.concat(taicpu.op_const_reg(A_RCL,S_W,1,cg.GetNextReg(reg.reglo)));
list.concat(taicpu.op_const_reg(A_RCL,S_W,1,reg.reghi));
list.concat(taicpu.op_const_reg(A_RCL,S_W,1,cg.GetNextReg(reg.reghi)));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_SHR,OP_SAR:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
cg.a_op_const_reg(list,op,OS_16,1,cg.GetNextReg(reg.reghi));
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,reg.reghi));
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,cg.GetNextReg(reg.reglo)));
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,reg.reglo));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
else
internalerror(2019051010);
end;
ai:=Taicpu.Op_Sym(A_LOOP,S_W,loop_start);
ai.is_jmp := True;
list.Concat(ai);
cg.ungetcpuregister(list,NR_CX);
end;
16,17:
begin
case op of
OP_SHL:
begin
cg.a_load_reg_reg(list,OS_16,OS_16,reg.reghi,cg.GetNextReg(reg.reghi));
cg.a_load_reg_reg(list,OS_16,OS_16,cg.GetNextReg(reg.reglo),reg.reghi);
cg.a_load_reg_reg(list,OS_16,OS_16,reg.reglo,cg.GetNextReg(reg.reglo));
cg.a_op_reg_reg(list,OP_XOR,OS_16,reg.reglo,reg.reglo);
end;
OP_SHR:
begin
cg.a_load_reg_reg(list,OS_16,OS_16,cg.GetNextReg(reg.reglo),reg.reglo);
cg.a_load_reg_reg(list,OS_16,OS_16,reg.reghi,cg.GetNextReg(reg.reglo));
cg.a_load_reg_reg(list,OS_16,OS_16,cg.GetNextReg(reg.reghi),reg.reghi);
cg.a_op_reg_reg(list,OP_XOR,OS_16,cg.GetNextReg(reg.reghi),cg.GetNextReg(reg.reghi));
end;
OP_SAR:
begin
cg.a_load_reg_reg(list,OS_16,OS_16,cg.GetNextReg(reg.reglo),reg.reglo);
cg.a_load_reg_reg(list,OS_16,OS_16,reg.reghi,cg.GetNextReg(reg.reglo));
cg.a_load_reg_reg(list,OS_16,OS_16,cg.GetNextReg(reg.reghi),reg.reghi);
cg.a_op_const_reg(list,OP_SAR,OS_16,15,cg.GetNextReg(reg.reghi));
end;
else
internalerror(2019051011);
end;
if value=17 then
case op of
OP_SHL:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_SHL,S_W,1,cg.GetNextReg(reg.reglo)));
list.concat(taicpu.op_const_reg(A_RCL,S_W,1,reg.reghi));
list.concat(taicpu.op_const_reg(A_RCL,S_W,1,cg.GetNextReg(reg.reghi)));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_SHR,OP_SAR:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
cg.a_op_const_reg(list,op,OS_16,1,reg.reghi);
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,cg.GetNextReg(reg.reglo)));
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,reg.reglo));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
else
internalerror(2019051012);
end;
end;
18..31:
begin
case op of
OP_SHL:
begin
cg.a_load_reg_reg(list,OS_16,OS_16,reg.reghi,cg.GetNextReg(reg.reghi));
cg.a_load_reg_reg(list,OS_16,OS_16,cg.GetNextReg(reg.reglo),reg.reghi);
cg.a_load_reg_reg(list,OS_16,OS_16,reg.reglo,cg.GetNextReg(reg.reglo));
cg.a_op_reg_reg(list,OP_XOR,OS_16,reg.reglo,reg.reglo);
end;
OP_SHR:
begin
cg.a_load_reg_reg(list,OS_16,OS_16,cg.GetNextReg(reg.reglo),reg.reglo);
cg.a_load_reg_reg(list,OS_16,OS_16,reg.reghi,cg.GetNextReg(reg.reglo));
cg.a_load_reg_reg(list,OS_16,OS_16,cg.GetNextReg(reg.reghi),reg.reghi);
cg.a_op_reg_reg(list,OP_XOR,OS_16,cg.GetNextReg(reg.reghi),cg.GetNextReg(reg.reghi));
end;
OP_SAR:
begin
cg.a_load_reg_reg(list,OS_16,OS_16,cg.GetNextReg(reg.reglo),reg.reglo);
cg.a_load_reg_reg(list,OS_16,OS_16,reg.reghi,cg.GetNextReg(reg.reglo));
cg.a_load_reg_reg(list,OS_16,OS_16,cg.GetNextReg(reg.reghi),reg.reghi);
cg.a_op_const_reg(list,OP_SAR,OS_16,15,cg.GetNextReg(reg.reghi));
end;
else
internalerror(2019051013);
end;
cg.getcpuregister(list,NR_CX);
cg.a_load_const_reg(list,OS_16,value-16,NR_CX);
current_asmdata.getjumplabel(loop_start);
cg.a_label(list,loop_start);
case op of
OP_SHL:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_reg(A_SHL,S_W,1,cg.GetNextReg(reg.reglo)));
list.concat(taicpu.op_const_reg(A_RCL,S_W,1,reg.reghi));
list.concat(taicpu.op_const_reg(A_RCL,S_W,1,cg.GetNextReg(reg.reghi)));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
OP_SHR,OP_SAR:
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
cg.a_op_const_reg(list,op,OS_16,1,reg.reghi);
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,cg.GetNextReg(reg.reglo)));
list.concat(taicpu.op_const_reg(A_RCR,S_W,1,reg.reglo));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
else
internalerror(2019051014);
end;
ai:=Taicpu.Op_Sym(A_LOOP,S_W,loop_start);
ai.is_jmp := True;
list.Concat(ai);
cg.ungetcpuregister(list,NR_CX);
end;
32..47:
case op of
OP_SHL:
begin
cg.a_op_const_reg_reg(list,OP_SHL,OS_32,value-32,reg.reglo,reg.reghi);
cg.a_op_reg_reg(list,OP_XOR,OS_16,reg.reglo,reg.reglo);
cg.a_op_reg_reg(list,OP_XOR,OS_16,cg.GetNextReg(reg.reglo),cg.GetNextReg(reg.reglo));
end;
OP_SHR:
begin
cg.a_op_const_reg_reg(list,OP_SHR,OS_32,value-32,reg.reghi,reg.reglo);
cg.a_op_reg_reg(list,OP_XOR,OS_16,reg.reghi,reg.reghi);
cg.a_op_reg_reg(list,OP_XOR,OS_16,cg.GetNextReg(reg.reghi),cg.GetNextReg(reg.reghi));
end;
OP_SAR:
begin
cg.a_op_const_reg_reg(list,OP_SAR,OS_32,value-32,reg.reghi,reg.reglo);
cg.a_op_const_reg_reg(list,OP_SAR,OS_16,15-(value-32),cg.GetNextReg(reg.reglo),reg.reghi);
cg.a_load_reg_reg(list,OS_16,OS_16,reg.reghi,cg.GetNextReg(reg.reghi));
end;
else
internalerror(2019051015);
end;
48..63:
case op of
OP_SHL:
begin
cg.a_load_reg_reg(list,OS_16,OS_16,reg.reglo,cg.GetNextReg(reg.reghi));
cg.a_op_reg_reg(list,OP_XOR,OS_16,reg.reglo,reg.reglo);
cg.a_op_reg_reg(list,OP_XOR,OS_16,cg.GetNextReg(reg.reglo),cg.GetNextReg(reg.reglo));
cg.a_op_reg_reg(list,OP_XOR,OS_16,reg.reghi,reg.reghi);
cg.a_op_const_reg(list,OP_SHL,OS_16,value-48,cg.GetNextReg(reg.reghi));
end;
OP_SHR:
begin
cg.a_load_reg_reg(list,OS_16,OS_16,cg.GetNextReg(reg.reghi),reg.reglo);
cg.a_op_reg_reg(list,OP_XOR,OS_16,cg.GetNextReg(reg.reghi),cg.GetNextReg(reg.reghi));
cg.a_op_reg_reg(list,OP_XOR,OS_16,reg.reghi,reg.reghi);
cg.a_op_reg_reg(list,OP_XOR,OS_16,cg.GetNextReg(reg.reglo),cg.GetNextReg(reg.reglo));
cg.a_op_const_reg(list,OP_SHR,OS_16,value-48,reg.reglo);
end;
OP_SAR:
if value=63 then
begin
cg.a_op_const_reg(list,OP_SAR,OS_16,15,cg.GetNextReg(reg.reghi));
cg.a_load_reg_reg(list,OS_16,OS_16,cg.GetNextReg(reg.reghi),reg.reghi);
cg.a_load_reg_reg(list,OS_16,OS_16,cg.GetNextReg(reg.reghi),cg.GetNextReg(reg.reglo));
cg.a_load_reg_reg(list,OS_16,OS_16,cg.GetNextReg(reg.reghi),reg.reglo);
end
else
begin
cg.a_op_const_reg_reg(list,OP_SAR,OS_16,value-48,cg.GetNextReg(reg.reghi),reg.reglo);
cg.a_op_const_reg_reg(list,OP_SAR,OS_16,15-(value-48),reg.reglo,cg.GetNextReg(reg.reglo));
cg.a_load_reg_reg(list,OS_16,OS_16,cg.GetNextReg(reg.reglo),reg.reghi);
cg.a_load_reg_reg(list,OS_16,OS_16,cg.GetNextReg(reg.reglo),cg.GetNextReg(reg.reghi));
end;
else
internalerror(2019051016);
end;
end;
end;
else
internalerror(200204021);
end;
end;
procedure tcg64f8086.a_op64_const_ref(list : TAsmList;op:TOpCG;size : tcgsize;value : int64;const ref : treference);
var
op1,op2 : TAsmOp;
tempref : treference;
begin
tempref:=ref;
tcgx86(cg).make_simple_ref(list,tempref);
case op of
OP_AND,OP_OR,OP_XOR:
begin
cg.a_op_const_ref(list,op,OS_32,tcgint(lo(value)),tempref);
inc(tempref.offset,4);
cg.a_op_const_ref(list,op,OS_32,tcgint(hi(value)),tempref);
end;
OP_ADD, OP_SUB:
begin
get_64bit_ops(op,op1,op2);
if (value and $ffffffffffff) = 0 then
begin
inc(tempref.offset,6);
{ use a_op_const_ref to allow the use of inc/dec }
cg.a_op_const_ref(list,op,OS_16,aint((value shr 48) and $ffff),tempref);
end
// can't use a_op_const_ref because this may use dec/inc
else if (value and $ffffffff) = 0 then
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
inc(tempref.offset,4);
list.concat(taicpu.op_const_ref(op1,S_W,aint((value shr 32) and $ffff),tempref));
inc(tempref.offset,2);
list.concat(taicpu.op_const_ref(op2,S_W,aint((value shr 48) and $ffff),tempref));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end
else if (value and $ffff) = 0 then
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
inc(tempref.offset,2);
list.concat(taicpu.op_const_ref(op1,S_W,aint((value shr 16) and $ffff),tempref));
inc(tempref.offset,2);
list.concat(taicpu.op_const_ref(op2,S_W,aint((value shr 32) and $ffff),tempref));
inc(tempref.offset,2);
list.concat(taicpu.op_const_ref(op2,S_W,aint((value shr 48) and $ffff),tempref));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end
else
begin
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_const_ref(op1,S_W,aint(value and $ffff),tempref));
inc(tempref.offset,2);
list.concat(taicpu.op_const_ref(op2,S_W,aint((value shr 16) and $ffff),tempref));
inc(tempref.offset,2);
list.concat(taicpu.op_const_ref(op2,S_W,aint((value shr 32) and $ffff),tempref));
inc(tempref.offset,2);
list.concat(taicpu.op_const_ref(op2,S_W,aint((value shr 48) and $ffff),tempref));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
end;
else
internalerror(200204022);
end;
end;
procedure tcg64f8086.a_op64_ref(list : TAsmList;op:TOpCG;size : tcgsize;const ref : treference);
var
tempref: treference;
begin
tempref:=ref;
tcgx86(cg).make_simple_ref(list,tempref);
case op of
OP_NOT:
begin
cg.a_op_ref(list,op,OS_32,tempref);
inc(tempref.offset,4);
cg.a_op_ref(list,op,OS_32,tempref);
end;
OP_NEG :
begin
inc(tempref.offset,6);
cg.a_op_ref(list,OP_NOT,OS_16,tempref);
dec(tempref.offset,2);
cg.a_op_ref(list,OP_NOT,OS_16,tempref);
dec(tempref.offset,2);
cg.a_op_ref(list,OP_NOT,OS_16,tempref);
dec(tempref.offset,2);
cg.a_reg_alloc(list,NR_DEFAULTFLAGS);
list.concat(taicpu.op_ref(A_NEG,S_W,tempref));
inc(tempref.offset,2);
list.concat(taicpu.op_const_ref(A_SBB,S_W,-1,tempref));
inc(tempref.offset,2);
list.concat(taicpu.op_const_ref(A_SBB,S_W,-1,tempref));
inc(tempref.offset,2);
list.concat(taicpu.op_const_ref(A_SBB,S_W,-1,tempref));
cg.a_reg_dealloc(list,NR_DEFAULTFLAGS);
end;
else
inherited;
end;
end;
procedure create_codegen;
begin
cg := tcg8086.create;
cg64 := tcg64f8086.create;
end;
end.