/*      $NetBSD: pmap.h,v 1.41 2022/05/26 05:34:04 skrll Exp $  */

/*      $OpenBSD: pmap.h,v 1.35 2007/12/14 18:32:23 deraadt Exp $       */

/*
* Copyright (c) 2002-2004 Michael Shalayeff
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
*    notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
*    notice, this list of conditions and the following disclaimer in the
*    documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
* IN NO EVENT SHALL THE AUTHOR OR HIS RELATIVES BE LIABLE FOR ANY DIRECT,
* INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
* SERVICES; LOSS OF MIND, USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
* STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING
* IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
* THE POSSIBILITY OF SUCH DAMAGE.
*/

/*
*      Pmap header for hppa.
*/

#ifndef _HPPA_PMAP_H_
#define _HPPA_PMAP_H_

#ifdef _KERNEL_OPT
#include "opt_cputype.h"
#endif

#include <sys/rwlock.h>
#include <machine/pte.h>
#include <machine/cpufunc.h>

#include <uvm/uvm_pglist.h>
#include <uvm/uvm_object.h>

#ifdef  _KERNEL

#define PMAP_NEED_PROCWR

struct pmap {
       struct uvm_object pm_obj;       /* object (lck by object lock) */
#define pm_lock pm_obj.vmobjlock
       krwlock_t       pm_obj_lock;    /* lock for pm_obj */
       struct vm_page  *pm_ptphint;
       struct vm_page  *pm_pdir_pg;    /* vm_page for pdir */
       volatile uint32_t *pm_pdir;     /* page dir (read-only after create) */
       pa_space_t      pm_space;       /* space id (read-only after create) */
       u_int           pm_pid;         /* prot id (read-only after create) */

       struct pmap_statistics  pm_stats;
};

#define PVF_MOD         PTE_PROT(TLB_DIRTY)     /* pg/mp is modified */
#define PVF_REF         PTE_PROT(TLB_REFTRAP)   /* pg/mp (inv) is referenced */
#define PVF_WRITE       PTE_PROT(TLB_WRITE)     /* pg/mp is writable */
#define PVF_EXEC        PTE_PROT(TLB_EXECUTE)   /* pg/mp is executable */
#define PVF_UNCACHEABLE PTE_PROT(TLB_UNCACHEABLE)       /* pg/mp is uncacheable */

#define HPPA_MAX_PID    0xfffa
#define HPPA_SID_MAX    0x7ffd

#define PMAP_DIRECTMAP  0x10000000 /* kenter_pa */

/*
* DON'T CHANGE THIS - this is assumed in lots of places.
*/
#define HPPA_SID_KERNEL 0
#define HPPA_PID_KERNEL 2

struct pv_entry {                       /* locked by its list's pvh_lock */
       struct pv_entry *pv_next;
       struct pmap     *pv_pmap;       /* the pmap */
       vaddr_t         pv_va;          /* the virtual address + flags */
#define PV_VAMASK       (~(PAGE_SIZE - 1))
#define PV_KENTER       __BIT(0)

       struct vm_page  *pv_ptp;        /* the vm_page of the PTP */
};

extern int pmap_hptsize;
extern struct pdc_hwtlb pdc_hwtlb;

/*
* pool quickmaps
*/
static inline vaddr_t hppa_map_poolpage(paddr_t pa)
{
       return (vaddr_t)pa;
}

static inline paddr_t hppa_unmap_poolpage(vaddr_t va)
{
       pdcache(HPPA_SID_KERNEL, va, PAGE_SIZE);

#if defined(HP8000_CPU) || defined(HP8200_CPU) || \
   defined(HP8500_CPU) || defined(HP8600_CPU)
       pdtlb(HPPA_SID_KERNEL, va);
#endif

       return (paddr_t)va;
}

#define PMAP_MAP_POOLPAGE(pa)   hppa_map_poolpage(pa)
#define PMAP_UNMAP_POOLPAGE(va) hppa_unmap_poolpage(va)

/*
* according to the parisc manual aliased va's should be
* different by high 12 bits only.
*/
#define PMAP_PREFER(o,h,s,td)   pmap_prefer((o), (h), (td))

static inline void
pmap_prefer(vaddr_t fo, vaddr_t *va, int td)
{
       vaddr_t newva;

       newva = (*va & HPPA_PGAMASK) | (fo & HPPA_PGAOFF);
       if (td) {
               if (newva > *va)
                       newva -= HPPA_PGALIAS;
       } else {
               if (newva < *va)
                       newva += HPPA_PGALIAS;
       }
       *va = newva;
}

#define pmap_sid2pid(s)                 (((s) + 1) << 1)
#define pmap_resident_count(pmap)       ((pmap)->pm_stats.resident_count)
#define pmap_wired_count(pmap)          ((pmap)->pm_stats.wired_count)
#define pmap_update(p)

#define pmap_copy(dpmap,spmap,da,len,sa)

#define pmap_clear_modify(pg)   pmap_changebit(pg, 0, PTE_PROT(TLB_DIRTY))
#define pmap_clear_reference(pg) \
                               pmap_changebit(pg, PTE_PROT(TLB_REFTRAP), 0)
#define pmap_is_modified(pg)    pmap_testbit(pg, PTE_PROT(TLB_DIRTY))
#define pmap_is_referenced(pg)  pmap_testbit(pg, PTE_PROT(TLB_REFTRAP))
#define pmap_phys_address(ppn)  ((ppn) << PAGE_SHIFT)

void    pmap_activate(struct lwp *);

void pmap_bootstrap(vaddr_t);
bool pmap_changebit(struct vm_page *, u_int, u_int);
bool pmap_testbit(struct vm_page *, u_int);
void pmap_write_protect(struct pmap *, vaddr_t, vaddr_t, vm_prot_t);
void pmap_remove(struct pmap *pmap, vaddr_t sva, vaddr_t eva);
void pmap_page_remove(struct vm_page *pg);

void pmap_procwr(struct proc *, vaddr_t, size_t);

static inline void
pmap_deactivate(struct lwp *l)
{
       /* Nothing. */
}

static inline bool
pmap_remove_all(struct pmap *pmap)
{
       /* Nothing. */
       return false;
}

static inline int
pmap_prot(struct pmap *pmap, int prot)
{
       extern u_int hppa_prot[];
       return (hppa_prot[prot] | (pmap == pmap_kernel() ? 0 : TLB_USER));
}

static inline void
pmap_page_protect(struct vm_page *pg, vm_prot_t prot)
{
       if ((prot & UVM_PROT_WRITE) == 0) {
               if (prot & (UVM_PROT_RX))
                       pmap_changebit(pg, 0, PTE_PROT(TLB_WRITE));
               else
                       pmap_page_remove(pg);
       }
}

static inline void
pmap_protect(struct pmap *pmap, vaddr_t sva, vaddr_t eva, vm_prot_t prot)
{
       if ((prot & UVM_PROT_WRITE) == 0) {
               if (prot & (UVM_PROT_RX))
                       pmap_write_protect(pmap, sva, eva, prot);
               else
                       pmap_remove(pmap, sva, eva);
       }
}

#define pmap_sid(pmap, va) \
       ((((va) & 0xc0000000) != 0xc0000000) ? \
        (pmap)->pm_space : HPPA_SID_KERNEL)

#define __HAVE_VM_PAGE_MD

struct pv_entry;

struct vm_page_md {
       struct pv_entry *pvh_list;      /* head of list */
       u_int           pvh_attrs;      /* to preserve ref/mod */
};

#define VM_MDPAGE_INIT(pg) \
do {                                                                    \
       (pg)->mdpage.pvh_list = NULL;                                   \
       (pg)->mdpage.pvh_attrs = 0;                                     \
} while (0)

#endif /* _KERNEL */

#endif /* _HPPA_PMAP_H_ */