/*      $NetBSD: pmap.h,v 1.39 2023/12/15 09:42:33 rin Exp $    */

/*-
* Copyright (C) 1995, 1996 Wolfgang Solfrank.
* Copyright (C) 1995, 1996 TooLs GmbH.
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
*    notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
*    notice, this list of conditions and the following disclaimer in the
*    documentation and/or other materials provided with the distribution.
* 3. All advertising materials mentioning features or use of this software
*    must display the following acknowledgement:
*      This product includes software developed by TooLs GmbH.
* 4. The name of TooLs GmbH may not be used to endorse or promote products
*    derived from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY TOOLS GMBH ``AS IS'' AND ANY EXPRESS OR
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
* IN NO EVENT SHALL TOOLS GMBH BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
* PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS;
* OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
* WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR
* OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF
* ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/

#ifndef _POWERPC_OEA_PMAP_H_
#define _POWERPC_OEA_PMAP_H_

#ifdef _LOCORE
#error use assym.h instead
#endif

#ifdef _MODULE
#error this file should not be included by loadable kernel modules
#endif

#ifdef _KERNEL_OPT
#include "opt_ppcarch.h"
#include "opt_modular.h"
#endif
#include <powerpc/oea/pte.h>

#define __HAVE_PMAP_PV_TRACK
#include <uvm/pmap/pmap_pvt.h>

/*
* Pmap stuff
*/
struct pmap {
#ifdef PPC_OEA64
       struct steg *pm_steg_table;             /* segment table pointer */
       /* XXX need way to track exec pages */
#endif

#if defined(PPC_OEA) || defined (PPC_OEA64_BRIDGE)
       register_t pm_sr[16];                   /* segments used in this pmap */
       int pm_exec[16];                        /* counts of exec mappings */
#endif
       register_t pm_vsid;                     /* VSID bits */
       int pm_refs;                            /* ref count */
       struct pmap_statistics pm_stats;        /* pmap statistics */
       unsigned int pm_evictions;              /* pvo's not in page table */

#ifdef PPC_OEA64
       unsigned int pm_ste_evictions;
#endif
};

struct pmap_ops {
       int (*pmapop_pte_spill)(struct pmap *, vaddr_t, bool);
       void (*pmapop_real_memory)(paddr_t *, psize_t *);
       void (*pmapop_init)(void);
       void (*pmapop_virtual_space)(vaddr_t *, vaddr_t *);
       pmap_t (*pmapop_create)(void);
       void (*pmapop_reference)(pmap_t);
       void (*pmapop_destroy)(pmap_t);
       void (*pmapop_copy)(pmap_t, pmap_t, vaddr_t, vsize_t, vaddr_t);
       void (*pmapop_update)(pmap_t);
       int (*pmapop_enter)(pmap_t, vaddr_t, paddr_t, vm_prot_t, u_int);
       void (*pmapop_remove)(pmap_t, vaddr_t, vaddr_t);
       void (*pmapop_kenter_pa)(vaddr_t, paddr_t, vm_prot_t, u_int);
       void (*pmapop_kremove)(vaddr_t, vsize_t);
       bool (*pmapop_extract)(pmap_t, vaddr_t, paddr_t *);

       void (*pmapop_protect)(pmap_t, vaddr_t, vaddr_t, vm_prot_t);
       void (*pmapop_unwire)(pmap_t, vaddr_t);
       void (*pmapop_page_protect)(struct vm_page *, vm_prot_t);
       void (*pmapop_pv_protect)(paddr_t, vm_prot_t);
       bool (*pmapop_query_bit)(struct vm_page *, int);
       bool (*pmapop_clear_bit)(struct vm_page *, int);

       void (*pmapop_activate)(struct lwp *);
       void (*pmapop_deactivate)(struct lwp *);

       void (*pmapop_pinit)(pmap_t);
       void (*pmapop_procwr)(struct proc *, vaddr_t, size_t);

       void (*pmapop_pte_print)(volatile struct pte *);
       void (*pmapop_pteg_check)(void);
       void (*pmapop_print_mmuregs)(void);
       void (*pmapop_print_pte)(pmap_t, vaddr_t);
       void (*pmapop_pteg_dist)(void);
       void (*pmapop_pvo_verify)(void);
       vaddr_t (*pmapop_steal_memory)(vsize_t, vaddr_t *, vaddr_t *);
       void (*pmapop_bootstrap)(paddr_t, paddr_t);
       void (*pmapop_bootstrap1)(paddr_t, paddr_t);
       void (*pmapop_bootstrap2)(void);
};

#ifdef  _KERNEL
#include <sys/cdefs.h>
__BEGIN_DECLS
#include <sys/param.h>
#include <sys/systm.h>

/*
* For OEA and OEA64_BRIDGE, we guarantee that pa below USER_ADDR
* (== 3GB < VM_MIN_KERNEL_ADDRESS) is direct-mapped.
*/
#if defined(PPC_OEA) || defined(PPC_OEA64_BRIDGE)
#define PMAP_DIRECT_MAPPED_SR   (USER_SR - 1)
#define PMAP_DIRECT_MAPPED_LEN \
   ((vaddr_t)SEGMENT_LENGTH * (PMAP_DIRECT_MAPPED_SR + 1))
#endif

#if defined (PPC_OEA) || defined (PPC_OEA64_BRIDGE)
extern register_t iosrtable[];
#endif
extern int pmap_use_altivec;

#define pmap_clear_modify(pg)           (pmap_clear_bit((pg), PTE_CHG))
#define pmap_clear_reference(pg)        (pmap_clear_bit((pg), PTE_REF))
#define pmap_is_modified(pg)            (pmap_query_bit((pg), PTE_CHG))
#define pmap_is_referenced(pg)          (pmap_query_bit((pg), PTE_REF))

#define pmap_resident_count(pmap)       ((pmap)->pm_stats.resident_count)
#define pmap_wired_count(pmap)          ((pmap)->pm_stats.wired_count)

/* ARGSUSED */
static __inline bool
pmap_remove_all(struct pmap *pmap)
{
       /* Nothing. */
       return false;
}

#if (defined(PPC_OEA) + defined(PPC_OEA64) + defined(PPC_OEA64_BRIDGE)) != 1
#define PMAP_NEEDS_FIXUP
#endif

extern volatile struct pteg *pmap_pteg_table;
extern unsigned int pmap_pteg_cnt;
extern unsigned int pmap_pteg_mask;

void pmap_bootstrap(vaddr_t, vaddr_t);
void pmap_bootstrap1(vaddr_t, vaddr_t);
void pmap_bootstrap2(void);
bool pmap_extract(pmap_t, vaddr_t, paddr_t *);
bool pmap_query_bit(struct vm_page *, int);
bool pmap_clear_bit(struct vm_page *, int);
void pmap_real_memory(paddr_t *, psize_t *);
void pmap_procwr(struct proc *, vaddr_t, size_t);
int pmap_pte_spill(pmap_t, vaddr_t, bool);
int pmap_ste_spill(pmap_t, vaddr_t, bool);
void pmap_pinit(pmap_t);

#ifdef PPC_OEA601
bool    pmap_extract_ioseg601(vaddr_t, paddr_t *);
#endif /* PPC_OEA601 */
#ifdef PPC_OEA
bool    pmap_extract_battable(vaddr_t, paddr_t *);
#endif /* PPC_OEA */

u_int powerpc_mmap_flags(paddr_t);
#define POWERPC_MMAP_FLAG_MASK  0xf
#define POWERPC_MMAP_FLAG_PREFETCHABLE  0x1
#define POWERPC_MMAP_FLAG_CACHEABLE     0x2

#define pmap_phys_address(ppn)          (ppn & ~POWERPC_MMAP_FLAG_MASK)
#define pmap_mmap_flags(ppn)            powerpc_mmap_flags(ppn)

static __inline paddr_t vtophys (vaddr_t);

/*
* Alternate mapping hooks for pool pages.  Avoids thrashing the TLB.
*
* Note: This won't work if we have more memory than can be direct-mapped
* VA==PA all at once.  But pmap_copy_page() and pmap_zero_page() will have
* this problem, too.
*/
#if !defined(PPC_OEA64)
#define PMAP_MAP_POOLPAGE(pa)   (pa)
#define PMAP_UNMAP_POOLPAGE(pa) (pa)
#define POOL_VTOPHYS(va)        vtophys((vaddr_t) va)

#define PMAP_ALLOC_POOLPAGE(flags)      pmap_alloc_poolpage(flags)
struct vm_page *pmap_alloc_poolpage(int);
#endif

static __inline paddr_t
vtophys(vaddr_t va)
{
       paddr_t pa;

       if (pmap_extract(pmap_kernel(), va, &pa))
               return pa;
       KASSERTMSG(0, "vtophys: pmap_extract of %#"PRIxVADDR" failed", va);
       return (paddr_t) -1;
}


#ifdef PMAP_NEEDS_FIXUP
extern const struct pmap_ops *pmapops;
extern const struct pmap_ops pmap32_ops;
extern const struct pmap_ops pmap64_ops;
extern const struct pmap_ops pmap64bridge_ops;

static __inline void
pmap_setup32(void)
{
       pmapops = &pmap32_ops;
}

static __inline void
pmap_setup64(void)
{
       pmapops = &pmap64_ops;
}

static __inline void
pmap_setup64bridge(void)
{
       pmapops = &pmap64bridge_ops;
}
#endif

bool pmap_pageidlezero (paddr_t);
void pmap_syncicache (paddr_t, psize_t);
#ifdef PPC_OEA64
vaddr_t pmap_setusr (vaddr_t);
vaddr_t pmap_unsetusr (void);
#endif

#ifdef PPC_OEA64_BRIDGE
int pmap_setup_segment0_map(int use_large_pages, ...);
#endif

#define PMAP_MD_PREFETCHABLE            0x2000000
#define PMAP_STEAL_MEMORY
#define PMAP_NEED_PROCWR

void pmap_zero_page(paddr_t);
void pmap_copy_page(paddr_t, paddr_t);

LIST_HEAD(pvo_head, pvo_entry);

#define __HAVE_VM_PAGE_MD

struct pmap_page {
       unsigned int pp_attrs;
       struct pvo_head pp_pvoh;
#ifdef MODULAR
       uintptr_t pp_dummy[3];
#endif
};

struct vm_page_md {
       struct pmap_page mdpg_pp;
#define mdpg_attrs      mdpg_pp.pp_attrs
#define mdpg_pvoh       mdpg_pp.pp_pvoh
#ifdef MODULAR
#define mdpg_dummy      mdpg_pp.pp_dummy
#endif
};

#define VM_MDPAGE_INIT(pg) do {                 \
       (pg)->mdpage.mdpg_attrs = 0;            \
       LIST_INIT(&(pg)->mdpage.mdpg_pvoh);     \
} while (/*CONSTCOND*/0)

__END_DECLS
#endif  /* _KERNEL */

#endif  /* _POWERPC_OEA_PMAP_H_ */