/*      $NetBSD: cache.S,v 1.2 2018/09/07 17:30:32 jmcneill Exp $       */

/*-
* Copyright (c) 2014 Robin Randhawa
* Copyright (c) 2015 The FreeBSD Foundation
* All rights reserved.
*
* Portions of this software were developed by Andrew Turner
* under sponsorship from the FreeBSD Foundation
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
*    notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
*    notice, this list of conditions and the following disclaimer in the
*    documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*
* $FreeBSD: head/sys/arm64/arm64/cpufunc_asm.S 313347 2017-02-06 17:50:09Z andrew $
*/

#include <aarch64/asm.h>

#define SCTLR_M (1<<0)
#define SCTLR_C (1<<2)

       .text
       .align  2

/*
* Macro to handle the cache. This takes the start address in x0, length
* in x1. It will corrupt x0, x1, x2, and x3.
*/
macro cache_handle_range dcop = 0, ic = 0, icop = 0
if \ic == 0
       mrs     x3, ctr_el0
       ubfx    x3, x3, #16, #4         /* x3 = D cache shift */
       mov     x2, #4                  /* size of word */
       lsl     x3, x2, x3              /* x3 = D cache line size */
else
       mrs     x3, ctr_el0
       ubfx    x2, x3, #16, #4         /* x2 = D cache shift */
       and     x3, x3, #15             /* x3 = I cache shift */
       cmp     x3, x2
       bcs     1f
       mov     x3, x2
1:                                      /* x3 = MAX(IcacheShift,DcacheShift) */
       mov     x2, #4                  /* size of word */
       lsl     x3, x2, x3              /* x3 = cache line size */
endif
       sub     x4, x3, #1              /* Get the address mask */
       and     x2, x0, x4              /* Get the low bits of the address */
       add     x1, x1, x2              /* Add these to the size */
       bic     x0, x0, x4              /* Clear the low bit of the address */
1:
       dc      \dcop, x0
       dsb     ish
if \ic != 0
       ic      \icop, x0
       dsb     ish
endif
       add     x0, x0, x3              /* Move to the next line */
       subs    x1, x1, x3              /* Reduce the size */
       b.hi    1b                      /* Check if we are done */
if \ic != 0
       isb
endif
       ret
endm


/*
* void aarch64_dcache_wbinv_range(vaddr_t, vsize_t)
*/
ENTRY(aarch64_dcache_wbinv_range)
       cache_handle_range      dcop = civac
END(aarch64_dcache_wbinv_range)

/*
* void aarch64_icache_inv_all(void)
*/
ENTRY(aarch64_icache_inv_all)
       dsb     ish
       ic      ialluis
       dsb     ish
       isb
       ret
END(aarch64_icache_inv_all)

/*
* void aarch64_exec_kernel(paddr_t entry, paddr_t dtb)
*/
ENTRY(aarch64_exec_kernel)
       mov     x20, x0 /* kernel entry point */
       mov     x21, x1 /* dtb address */

       mrs     x0, CurrentEL
       lsr     x0, x0, #2
       cmp     x0, #0x2
       b.eq    1f

       /* Disable MMU and dcache, CurrentEL = EL1 */
       mrs     x0, sctlr_el1
       bic     x0, x0, #SCTLR_M
       bic     x0, x0, #SCTLR_C
       msr     sctlr_el1, x0
       isb
       b       2f
1:
       /* Disable MMU and dcache, CurrentEL = EL2 */
       mrs     x0, sctlr_el2
       bic     x0, x0, #SCTLR_M
       bic     x0, x0, #SCTLR_C
       msr     sctlr_el2, x0
       isb
2:

       /* Jump to kernel */
       mov     x0, x21
       mov     x1, xzr
       mov     x2, xzr
       mov     x3, xzr
       br      x20

END(aarch64_exec_kernel)