Retro68/gcc/libgcc/config/rl78/mulsi3.S

324 lines
4.8 KiB
ArmAsm
Raw Normal View History

2018-12-28 15:30:48 +00:00
; Copyright (C) 2011-2018 Free Software Foundation, Inc.
2012-03-27 23:13:14 +00:00
; Contributed by Red Hat.
;
; This file is free software; you can redistribute it and/or modify it
; under the terms of the GNU General Public License as published by the
; Free Software Foundation; either version 3, or (at your option) any
; later version.
;
; This file is distributed in the hope that it will be useful, but
; WITHOUT ANY WARRANTY; without even the implied warranty of
; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
; General Public License for more details.
;
; Under Section 7 of GPL version 3, you are granted additional
; permissions described in the GCC Runtime Library Exception, version
; 3.1, as published by the Free Software Foundation.
;
; You should have received a copy of the GNU General Public License and
; a copy of the GCC Runtime Library Exception along with this program;
; see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
; <http://www.gnu.org/licenses/>.
;; 32x32=32 multiply
2014-09-21 17:33:12 +00:00
#include "vregs.h"
2012-03-27 23:13:14 +00:00
;----------------------------------------------------------------------
; Register use:
; RB0 RB1 RB2
; AX op2L res32L res32H
; BC op2H (resH) op1
; DE count (resL-tmp)
; HL [sp+4]
2015-08-28 15:33:40 +00:00
; Register use (G10):
;
; AX op2L
; BC op2H
; DE count
; HL [sp+4]
; r8/r9 res32L
; r10/r11 (resH)
; r12/r13 (resL-tmp)
; r16/r17 res32H
; r18/r19 op1
START_FUNC ___mulsi3
2012-03-27 23:13:14 +00:00
;; A is at [sp+4]
;; B is at [sp+8]
;; result is in R8..R11
2015-08-28 15:33:40 +00:00
#ifdef __RL78_G10__
movw ax, r16
push ax
movw ax, r18
push ax
#else
2012-03-27 23:13:14 +00:00
sel rb2
push ax
push bc
sel rb0
2015-08-28 15:33:40 +00:00
#endif
2012-03-27 23:13:14 +00:00
clrw ax
movw r8, ax
movw r16, ax
2014-09-21 17:33:12 +00:00
movw ax, [sp+14]
2012-03-27 23:13:14 +00:00
cmpw ax, #0
bz $1f
cmpw ax, #0xffff
bnz $2f
2014-09-21 17:33:12 +00:00
movw ax, [sp+8]
2015-08-28 15:33:40 +00:00
#ifdef __RL78_G10__
push bc
movw bc, r8
xchw ax, bc
subw ax, bc
movw r8, ax
movw ax, bc
pop bc
#else
2012-03-27 23:13:14 +00:00
sel rb1
subw ax, r_0
sel rb0
2015-08-28 15:33:40 +00:00
#endif
2012-03-27 23:13:14 +00:00
br $1f
2:
movw bc, ax
2014-09-21 17:33:12 +00:00
movw ax, [sp+8]
2012-03-27 23:13:14 +00:00
cmpw ax, #0
skz
call !.Lmul_hi
1:
2014-09-21 17:33:12 +00:00
movw ax, [sp+10]
2012-03-27 23:13:14 +00:00
cmpw ax, #0
bz $1f
cmpw ax, #0xffff
bnz $2f
2014-09-21 17:33:12 +00:00
movw ax, [sp+12]
2015-08-28 15:33:40 +00:00
#ifdef __RL78_G10__
push bc
movw bc, r8
xchw ax, bc
subw ax, bc
movw r8, ax
movw ax, bc
pop bc
#else
2012-03-27 23:13:14 +00:00
sel rb1
subw ax, r_0
sel rb0
2015-08-28 15:33:40 +00:00
#endif
2012-03-27 23:13:14 +00:00
br $1f
2:
movw bc, ax
2014-09-21 17:33:12 +00:00
movw ax, [sp+12]
2012-03-27 23:13:14 +00:00
cmpw ax, #0
skz
call !.Lmul_hi
1:
movw ax, r8
movw r16, ax
clrw ax
movw r8, ax
;; now do R16:R8 += op1L * op2L
;; op1 is in AX.0 (needs to shrw)
;; op2 is in BC.2 and BC.1 (bc can shlw/rolcw)
;; res is in AX.2 and AX.1 (needs to addw)
2014-09-21 17:33:12 +00:00
movw ax, [sp+8]
2012-03-27 23:13:14 +00:00
movw r10, ax ; BC.1
2014-09-21 17:33:12 +00:00
movw ax, [sp+12]
2012-03-27 23:13:14 +00:00
cmpw ax, r10
bc $.Lmul_hisi_top
movw bc, r10
movw r10, ax
movw ax, bc
.Lmul_hisi_top:
movw bc, #0
.Lmul_hisi_loop:
shrw ax, 1
2015-08-28 15:33:40 +00:00
#ifdef __RL78_G10__
push ax
bnc $.Lmul_hisi_no_add_g10
movw ax, r8
addw ax, r10
movw r8, ax
sknc
incw r16
movw ax, r16
addw ax, r_2
movw r16, ax
.Lmul_hisi_no_add_g10:
movw ax, r10
shlw ax, 1
movw r10, ax
pop ax
#else
2012-03-27 23:13:14 +00:00
bnc $.Lmul_hisi_no_add
sel rb1
addw ax, bc
sel rb2
sknc
incw ax
addw ax, r_2
2015-08-28 15:33:40 +00:00
.Lmul_hisi_no_add:
2012-03-27 23:13:14 +00:00
sel rb1
shlw bc, 1
sel rb0
2015-08-28 15:33:40 +00:00
#endif
2012-03-27 23:13:14 +00:00
rolwc bc, 1
cmpw ax, #0
bz $.Lmul_hisi_done
shrw ax, 1
2015-08-28 15:33:40 +00:00
#ifdef __RL78_G10__
push ax
bnc $.Lmul_hisi_no_add2_g10
movw ax, r8
addw ax, r10
movw r8, ax
movw ax, r16
sknc
incw ax
addw ax, r_2
movw r16, ax
.Lmul_hisi_no_add2_g10:
movw ax, r10
shlw ax, 1
movw r10, ax
pop ax
#else
2012-03-27 23:13:14 +00:00
bnc $.Lmul_hisi_no_add2
sel rb1
addw ax, bc
sel rb2
sknc
incw ax
addw ax, r_2
.Lmul_hisi_no_add2:
sel rb1
shlw bc, 1
sel rb0
2015-08-28 15:33:40 +00:00
#endif
2012-03-27 23:13:14 +00:00
rolwc bc, 1
cmpw ax, #0
bnz $.Lmul_hisi_loop
.Lmul_hisi_done:
movw ax, r16
movw r10, ax
2015-08-28 15:33:40 +00:00
#ifdef __RL78_G10__
pop ax
movw r18, ax
pop ax
movw r16, ax
#else
2012-03-27 23:13:14 +00:00
sel rb2
pop bc
pop ax
sel rb0
2015-08-28 15:33:40 +00:00
#endif
2012-03-27 23:13:14 +00:00
ret
2015-08-28 15:33:40 +00:00
END_FUNC ___mulsi3
2012-03-27 23:13:14 +00:00
;----------------------------------------------------------------------
2015-08-28 15:33:40 +00:00
START_FUNC ___mulhi3
2014-09-21 17:33:12 +00:00
movw r8, #0
movw ax, [sp+6]
movw bc, ax
movw ax, [sp+4]
2012-03-27 23:13:14 +00:00
;; R8 += AX * BC
.Lmul_hi:
cmpw ax, bc
skc
xchw ax, bc
br $.Lmul_hi_loop
.Lmul_hi_top:
2015-08-28 15:33:40 +00:00
#ifdef __RL78_G10__
push ax
movw ax, r8
addw ax, r_2
movw r8, ax
pop ax
#else
2012-03-27 23:13:14 +00:00
sel rb1
addw ax, r_2
sel rb0
2015-08-28 15:33:40 +00:00
#endif
2012-03-27 23:13:14 +00:00
.Lmul_hi_no_add:
shlw bc, 1
.Lmul_hi_loop:
shrw ax, 1
bc $.Lmul_hi_top
cmpw ax, #0
bz $.Lmul_hi_done
shlw bc, 1
shrw ax, 1
bc $.Lmul_hi_top
cmpw ax, #0
bnz $.Lmul_hi_no_add
.Lmul_hi_done:
ret
2015-08-28 15:33:40 +00:00
END_FUNC ___mulhi3
;;; --------------------------------------
#ifdef __RL78_G10__
START_FUNC ___mulqi3
mov a, [sp+4]
mov r9, a
mov a, [sp+6]
mov r10, a
mov a, #9
mov r11, a
clrb a
mov r8, a
.L2:
cmp0 r10
skz
dec r11
sknz
ret
mov a, r10
and a, #1
mov r12, a
cmp0 r12
sknz
br !!.L3
mov a, r9
mov l, a
mov a, r8
add a, l
mov r8, a
.L3:
mov a, r9
add a, a
mov r9, a
mov a, r10
shr a, 1
mov r10, a
br !!.L2
END_FUNC ___mulqi3
#endif
2012-03-27 23:13:14 +00:00