/*      $NetBSD: rmixl_spl.S,v 1.7 2021/09/01 14:17:46 andvar Exp $     */

/*-
* Copyright (c) 2009, 2010 The NetBSD Foundation, Inc.
* All rights reserved.
*
* This code is derived from software contributed to The NetBSD Foundation
* by Matt Thomas <[email protected]>.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
*    notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
*    notice, this list of conditions and the following disclaimer in the
*    documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
* POSSIBILITY OF SUCH DAMAGE.
*/

#include "opt_cputype.h"        /* which mips CPU levels do we support? */

#include <sys/cdefs.h>

#include <mips/asm.h>
#include <mips/cpuregs.h>

RCSID("$NetBSD: rmixl_spl.S,v 1.7 2021/09/01 14:17:46 andvar Exp $");

#include "assym.h"

#define MAP_SCALESHIFT          3
#define RMIXL_SOFT_INT_MASK_1   (MIPS_SOFT_INT_MASK_1 >> 8)
#define RMIXL_SOFT_INT_MASK     (MIPS_SOFT_INT_MASK >> 8)
#define RMIXL_INT_MASK_5        (MIPS_INT_MASK_5 >> 8)
#define RMIXL_EIRR_PRESERVE     (RMIXL_INT_MASK_5 | RMIXL_SOFT_INT_MASK)
#define RMIXL_INT_MASK_1        (MIPS_INT_MASK_1 >> 8)
#define RMIXL_INT_MASK_5        (MIPS_INT_MASK_5 >> 8)

       .set noreorder

/*
* Array of mask of bits to set in the EIMR when we go to a
* given hardware interrupt priority level.
* The softint bits in [IPL_NONE] and [IPL_SOFTCLOCK] should stay constant
* Hard intr bits are managed by rmixl_vec_establish and rmixl_vec_disestablish.
*/
       .data
       .globl  _C_LABEL(ipl_eimr_map)
       .type   _C_LABEL(ipl_eimr_map),@object
       .p2align MAP_SCALESHIFT
_C_LABEL(ipl_eimr_map):
       .dword  RMIXL_SOFT_INT_MASK     /* IPL_NONE */
       .dword  RMIXL_SOFT_INT_MASK_1   /* IPL_SOFT{CLOCK,BIO} */
       .dword  0                       /* IPL_SOFT{NET,SERIAL} */
       .dword  0                       /* IPL_VM */
       .dword  0                       /* IPL_SCHED */
       .dword  0                       /* IPL_DDB */
       .dword  0                       /* IPL_HIGH */

       .text

/*
* initialize cp0 interrupt control for this cpu
* - set STATUS[IE]
* - clear EIRR and EIMR
* on return, all interrupts are disabled by EIMR
*
* henceforth STATUS[IE] is expected to remain normally set
* but may be cleared and restored for temporary interrupt disablement
*
* call before the first call to spl0 on this cpu
*/
LEAF_NOPROFILE(rmixl_spl_init_cpu)
       mfc0    t0, MIPS_COP_0_STATUS           # get STATUS
       ori     t0, MIPS_SR_INT_IE              # set IE
       mtc0    zero, MIPS_COP_0_STATUS         ## disable all ints in STATUS
       COP0_SYNC
       dmtc0   zero, MIPS_COP_0_EIMR           ##  "       "   "   "  EIMR
       COP0_SYNC
       dmtc0   zero, MIPS_COP_0_EIRR           ## clear EIRR
       COP0_SYNC
       mtc0    t0, MIPS_COP_0_STATUS           ## set STATUS | IE
       JR_HB_RA
END(rmixl_spl_init_cpu)

/*
* RMIXL processor interrupt control
*
* Used as building blocks for spl(9) kernel interface.
*/

_splraise:
       /*
        * a0 = EIMR bits requested to be set for this IPL
        * a1 = this IPL (IPL_*)
        * Can only use a0-a3 and v0-v1
        * old IPL is returned in v0
        */
       dmfc0   a2, MIPS_COP_0_EIMR             # save EIMR
       dmtc0   zero, MIPS_COP_0_EIMR           ## disable all interrupts
       COP0_SYNC
       PTR_L   a3, L_CPU(MIPS_CURLWP)          ##
       INT_L   v0, CPU_INFO_CPL(a3)            ## get current IPL from cpu_info
       sltu    v1, a1, v0                      ## newipl < curipl
       bnez    v1, 1f                          ## yes, don't change.
        nop
       INT_S   a1, CPU_INFO_CPL(a3)            ## save IPL in cpu_info
       dmtc0   a0, MIPS_COP_0_EIMR             ## set new EIMR
       JR_HB_RA
1:
       dmtc0   a2, MIPS_COP_0_EIMR             ## restore saved EIMR
       JR_HB_RA

STATIC_LEAF(_splsw_splx)
STATIC_XLEAF(_splsw_splx_noprof)                # does not get mcount hooks
       PTR_LA  v1, _C_LABEL(ipl_eimr_map)      # get address of table
       sll     a2, a0, MAP_SCALESHIFT          # convert IPL to array offset
       PTR_ADDU v1, a2                         # add to table addr
       REG_L   v1, (v1)                        # load EIMR bits for this IPL

       dmtc0   zero, MIPS_COP_0_EIMR           ## disable all interrupts
       COP0_SYNC
       PTR_L   a3, L_CPU(MIPS_CURLWP)          ## get cpu_info
       INT_S   a0, CPU_INFO_CPL(a3)            ## save IPL in cpu_info
       dmtc0   v1, MIPS_COP_0_EIMR             ## set new EIMR
       JR_HB_RA
END(_splsw_splx)

STATIC_LEAF(_splsw_spl0)
       REG_L   v1, _C_LABEL(ipl_eimr_map) + 8*IPL_NONE
       dmtc0   zero, MIPS_COP_0_EIMR           ## disable all interrupts
       COP0_SYNC
       PTR_L   a3, L_CPU(MIPS_CURLWP)          ## get cpu_info
#if IPL_NONE == 0
       INT_S   zero, CPU_INFO_CPL(a3)          ## save IPL in cpu_info
#else
#error IPL_NONE != 0
#endif
       dmtc0   v1, MIPS_COP_0_EIMR             ## set new EIMR
       JR_HB_RA
END(_splsw_spl0)

STATIC_LEAF(_splsw_setsoftintr)
       dsrl    a0, 8                           # convert CAUSE bit to EIRR bit
       and     a0, RMIXL_SOFT_INT_MASK         # restrict to softint bits
       dmfc0   v1, MIPS_COP_0_EIMR             # save EIMR register
       dmtc0   zero, MIPS_COP_0_EIMR           ## disable all interrupts
       COP0_SYNC
       dmfc0   v0, MIPS_COP_0_EIRR             ## load EIRR
       and     v0, RMIXL_EIRR_PRESERVE         ## preserve clock & softints
       or      v0, a0                          ## set new softint bit
       dmtc0   v0, MIPS_COP_0_EIRR             ## store EIRR
       COP0_SYNC
       dmtc0   v1, MIPS_COP_0_EIMR             ## restore EIMR
       JR_HB_RA
END(_splsw_setsoftintr)

STATIC_LEAF(_splsw_clrsoftintr)
       dsrl    a0, 8                           # convert CAUSE bit to EIRR bit
       and     a0, RMIXL_SOFT_INT_MASK         # restrict to softint bits
       xor     a0, RMIXL_EIRR_PRESERVE         # clear from preserve mask
       dmfc0   v1, MIPS_COP_0_EIMR             # save EIMR register
       dmtc0   zero, MIPS_COP_0_EIMR           ## disable all interrupts
       COP0_SYNC
       dmfc0   v0, MIPS_COP_0_EIRR             ## load EIRR
       and     v0, a0                          ## apply preserve mask
       dmtc0   v0, MIPS_COP_0_EIRR             ## store EIRR
       COP0_SYNC
       dmtc0   v1, MIPS_COP_0_EIMR             ## restore EIMR
       JR_HB_RA
END(_splsw_clrsoftintr)

STATIC_LEAF(_splsw_splraise)
       move    a1, a0
       PTR_LA  v1, _C_LABEL(ipl_eimr_map)
       sll     a2, a0, MAP_SCALESHIFT
       PTR_ADDU v1, a2
       REG_L   a0, (v1)
       b       _splraise
        nop
END(_splsw_splraise)

STATIC_LEAF(_splsw_splhigh)
STATIC_XLEAF(_splsw_splhigh_noprof)
       dmtc0   zero, MIPS_COP_0_EIMR           ## disable all interrupts
       COP0_SYNC
       PTR_L   a3, L_CPU(MIPS_CURLWP)          ## get cpu_info from curlwp
       li      a1, IPL_HIGH                    ##
       INT_L   v0, CPU_INFO_CPL(a3)            ## old IPL for return value
       INT_S   a1, CPU_INFO_CPL(a3)            ## save IPL in cpu_info
                                               ## interrupts remain disabled!
       j       ra                              # return
        nop
END(_splsw_splhigh)

STATIC_LEAF(_splsw_splddb)
       REG_L   a0, _C_LABEL(ipl_eimr_map) + 8*IPL_DDB
       li      a1, IPL_DDB
       b       _splraise
        nop
END(_splsw_splddb)

STATIC_LEAF(_splsw_splsched)
       REG_L   a0, _C_LABEL(ipl_eimr_map) + 8*IPL_SCHED
       li      a1, IPL_SCHED
       b       _splraise
        nop
END(_splsw_splsched)

STATIC_LEAF(_splsw_splvm)
       REG_L   a0, _C_LABEL(ipl_eimr_map) + 8*IPL_VM
       li      a1, IPL_VM
       b       _splraise
        nop
END(_splsw_splvm)

STATIC_LEAF(_splsw_splsoftserial)
       REG_L   a0, _C_LABEL(ipl_eimr_map) + 8*IPL_SOFTSERIAL
       li      a1, IPL_SOFTSERIAL
       b       _splraise
        nop
END(_splsw_splsoftserial)

STATIC_LEAF(_splsw_splsoftnet)
       REG_L   a0, _C_LABEL(ipl_eimr_map) + 8*IPL_SOFTNET
       li      a1, IPL_SOFTNET
       b       _splraise
        nop
END(_splsw_splsoftnet)

STATIC_LEAF(_splsw_splsoftbio)
       REG_L   a0, _C_LABEL(ipl_eimr_map) + 8*IPL_SOFTBIO
       li      a1, IPL_SOFTBIO
       b       _splraise
        nop
END(_splsw_splsoftbio)

STATIC_LEAF(_splsw_splsoftclock)
       REG_L   a0, _C_LABEL(ipl_eimr_map) + 8*IPL_SOFTCLOCK
       li      a1, IPL_SOFTCLOCK
       b       _splraise
        nop
END(_splsw_splsoftclock)

STATIC_LEAF(_splsw_splintr)
       dmfc0   ta1, MIPS_COP_0_EIRR            # get active interrupts
                                               # restrict to hard int bits:
       and     v1, ta1, RMIXL_SOFT_INT_MASK    #  v1 = ta1 & ~RMIXL_SOFT_INT_MASK
       xor     v1, ta1                         #   "       "

       li      v0, IPL_NONE
       PTR_LA  ta3, _C_LABEL(ipl_eimr_map) + 8*IPL_VM
       REG_L   ta2, -8(ta3)                    # load 'enabled' bits for IPL_SOFTSERIAL
       and     v1, ta2                         # apply to pending bits
       beq     v1, zero, 4f                    # if nothing pending...
        nop                                    # ... return IPL_NONE

       li      v0, IPL_VM                      # ipl=IPL_VM
1:
       REG_L   ta2, (ta3)                      # load 'enabled' bits for ipl
       and     ta2, v1                         # any match to pending intrs?
       beq     ta2, zero, 2f                   #  no, return ipl
        PTR_ADDI ta3, 1 << MAP_SCALESHIFT      #  point to next entry
       addiu   v0, 1                           # ipl++
       move    v1, ta2                         # update highest pending
       b       1b                              # loop
        nop

2:
       /*
        * Emulate the CP0_SR 'IM' bits in 'pending'
        * - if clock intr is requested, set MIPS_INT_MASK_5
        * - if other HW intr is requested, set MIPS_INT_MASK_1 as summary bit
        *   the RMI evbmips_iointr function will sort through
        *   individual EIRR requests
        */
       li      t2, RMIXL_INT_MASK_5            # load RMIXL_INT_MASK_5
       and     t1, v1, t2                      # save count/compare intr request value
       nor     t0, zero, t2                    # invert the mask
       and     v1, t0                          # v1 &= ~RMIXL_INT_MASK_5
       beq     v1, zero, 3f                    # no non-clock intrs? skip ahead
        li     v1, RMIXL_INT_MASK_1            # use INT_MASK_1 as 'summary' bit
                                               #  for non-clock hw intrs
3:
       or      v1, t1                          # combine clock and non-clock-summary
       sll     v1, MIPS_INT_MASK_SHIFT         # shift to emulate COP0_SR 'IM' bits
4:
       INT_S   v1, (a0)                        # set a (fake) new pending mask
       j       ra                              # and return highest ipl pending
        nop
END(_splsw_splintr)

STATIC_LEAF(_splsw_splcheck)
#ifdef PARANOIA
       PTR_L   t0, L_CPU(MIPS_CURLWP)
       INT_L   t1, CPU_INFO_CPL(t0)            # get current priority level

       dmfc0   t0, MIPS_COP_0_EIMR             # get current EIMR

       PTR_LA  t2, _C_LABEL(ipl_eimr_map)
       sll     t1, MAP_SCALESHIFT              # shift cpl to array index
       PTR_ADDU t2, t1
       REG_L   t3, (t2)                        # load value
1:      bne     t0, t3, 1b                      # loop forever if not equal
        nop
#endif /* PARANOIA */
       j       ra
        nop
END(_splsw_splcheck)

       .rdata
       .globl _C_LABEL(rmixl_splsw)
_C_LABEL(rmixl_splsw):
       PTR_WORD _C_LABEL(_splsw_splhigh)
       PTR_WORD _C_LABEL(_splsw_splsched)
       PTR_WORD _C_LABEL(_splsw_splvm)
       PTR_WORD _C_LABEL(_splsw_splsoftserial)
       PTR_WORD _C_LABEL(_splsw_splsoftnet)
       PTR_WORD _C_LABEL(_splsw_splsoftbio)
       PTR_WORD _C_LABEL(_splsw_splsoftclock)
       PTR_WORD _C_LABEL(_splsw_splraise)
       PTR_WORD _C_LABEL(_splsw_spl0)
       PTR_WORD _C_LABEL(_splsw_splx)
       PTR_WORD _C_LABEL(_splsw_splhigh_noprof)
       PTR_WORD _C_LABEL(_splsw_splx_noprof)
       PTR_WORD _C_LABEL(_splsw_setsoftintr)
       PTR_WORD _C_LABEL(_splsw_clrsoftintr)
       PTR_WORD _C_LABEL(_splsw_splintr)
       PTR_WORD _C_LABEL(_splsw_splcheck)