/*      $NetBSD: vmpagemd.h,v 1.17 2020/12/20 16:38:26 skrll Exp $      */

/*-
* Copyright (c) 2011 The NetBSD Foundation, Inc.
* All rights reserved.
*
* This code is derived from software contributed to The NetBSD Foundation
* by Raytheon BBN Technologies Corp and Defense Advanced Research Projects
* Agency and which was developed by Matt Thomas of 3am Software Foundry.
*
* This material is based upon work supported by the Defense Advanced Research
* Projects Agency and Space and Naval Warfare Systems Center, Pacific, under
* Contract No. N66001-09-C-2073.
* Approved for Public Release, Distribution Unlimited
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
*    notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
*    notice, this list of conditions and the following disclaimer in the
*    documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
* POSSIBILITY OF SUCH DAMAGE.
*/

#ifndef _UVM_PMAP_VMPAGEMD_H_
#define _UVM_PMAP_VMPAGEMD_H_

#ifdef _LOCORE
#error use assym.h instead
#endif

//#ifdef _MODULE
//#error this file should not be included by loadable kernel modules
//#endif

#ifdef _KERNEL_OPT
#include "opt_modular.h"
#include "opt_multiprocessor.h"
#endif

#include <sys/atomic.h>
#include <sys/mutex.h>

#define __HAVE_VM_PAGE_MD

typedef struct pv_entry {
       struct pv_entry *pv_next;
       struct pmap *pv_pmap;
       vaddr_t pv_va;
#define PV_KENTER               __BIT(0)
} *pv_entry_t;

#define PV_ISKENTER_P(pv)       (((pv->pv_va) & PV_KENTER) != 0)

#ifndef _MODULE

#define VM_PAGEMD_VMPAGE        __BIT(0)        /* page is vm managed */
#define VM_PAGEMD_REFERENCED    __BIT(1)        /* page has been referenced */
#define VM_PAGEMD_MODIFIED      __BIT(2)        /* page has been modified */
#define VM_PAGEMD_POOLPAGE      __BIT(3)        /* page is used as a poolpage */
#define VM_PAGEMD_EXECPAGE      __BIT(4)        /* page is exec mapped */
#ifdef PMAP_VIRTUAL_CACHE_ALIASES
#define VM_PAGEMD_UNCACHED      __BIT(5)        /* page is mapped uncached */
#endif

#define VM_PAGEMD_VMPAGE_P(mdpg)        (((mdpg)->mdpg_attrs & VM_PAGEMD_VMPAGE) != 0)
#define VM_PAGEMD_REFERENCED_P(mdpg)    (((mdpg)->mdpg_attrs & VM_PAGEMD_REFERENCED) != 0)
#define VM_PAGEMD_MODIFIED_P(mdpg)      (((mdpg)->mdpg_attrs & VM_PAGEMD_MODIFIED) != 0)
#define VM_PAGEMD_POOLPAGE_P(mdpg)      (((mdpg)->mdpg_attrs & VM_PAGEMD_POOLPAGE) != 0)
#define VM_PAGEMD_EXECPAGE_P(mdpg)      (((mdpg)->mdpg_attrs & VM_PAGEMD_EXECPAGE) != 0)
#ifdef PMAP_VIRTUAL_CACHE_ALIASES
#define VM_PAGEMD_CACHED_P(mdpg)        (((mdpg)->mdpg_attrs & VM_PAGEMD_UNCACHED) == 0)
#define VM_PAGEMD_UNCACHED_P(mdpg)      (((mdpg)->mdpg_attrs & VM_PAGEMD_UNCACHED) != 0)
#endif

#endif /* !_MODULE */

struct vm_page_md {
       volatile unsigned long mdpg_attrs;      /* page attributes */
       struct pv_entry mdpg_first;             /* pv_entry first */
#if defined(MULTIPROCESSOR) || defined(MODULAR) || defined(_MODULE)
       kmutex_t *mdpg_lock;                    /* pv list lock */
#endif
};

#ifndef _MODULE
#if defined(MULTIPROCESSOR) || defined(MODULAR)
#define VM_PAGEMD_PVLIST_LOCK_INIT(mdpg)        (mdpg)->mdpg_lock = NULL
#else
#define VM_PAGEMD_PVLIST_LOCK_INIT(mdpg)        __nothing
#endif /* MULTIPROCESSOR || MODULAR */

#define VM_PAGEMD_PVLIST_LOCK(mdpg)     pmap_pvlist_lock(mdpg, 1)
#define VM_PAGEMD_PVLIST_READLOCK(mdpg) pmap_pvlist_lock(mdpg, 0)
#define VM_PAGEMD_PVLIST_UNLOCK(mdpg)   pmap_pvlist_unlock(mdpg)
#define VM_PAGEMD_PVLIST_LOCKED_P(mdpg) pmap_pvlist_locked_p(mdpg)
#define VM_PAGEMD_PVLIST_GEN(mdpg)      ((mdpg)->mdpg_attrs >> 16)

#define VM_PAGEMD_PVLIST_EMPTY_P(mdpg)  ((mdpg)->mdpg_first.pv_pmap == NULL)

#ifdef _KERNEL
#if defined(MULTIPROCESSOR) || defined(MODULAR)
kmutex_t *pmap_pvlist_lock_addr(struct vm_page_md *);
#else
extern kmutex_t pmap_pvlist_mutex;
static __inline kmutex_t *
pmap_pvlist_lock_addr(struct vm_page_md *mdpg)
{
       return &pmap_pvlist_mutex;
}
#endif

static __inline uintptr_t
pmap_pvlist_lock(struct vm_page_md *mdpg, uintptr_t increment)
{
       mutex_spin_enter(pmap_pvlist_lock_addr(mdpg));
       const uintptr_t gen = VM_PAGEMD_PVLIST_GEN(mdpg);
       mdpg->mdpg_attrs += increment << 16;
       return gen;
}

static __inline uintptr_t
pmap_pvlist_unlock(struct vm_page_md *mdpg)
{
       const uintptr_t gen = VM_PAGEMD_PVLIST_GEN(mdpg);
       mutex_spin_exit(pmap_pvlist_lock_addr(mdpg));
       return gen;
}

static __inline bool
pmap_pvlist_locked_p(struct vm_page_md *mdpg)
{

       return mutex_owned(pmap_pvlist_lock_addr(mdpg));
}
#endif /* _KERNEL */

#define VM_MDPAGE_INIT(pg)                                              \
do {                                                                    \
       (pg)->mdpage.mdpg_first.pv_next = NULL;                         \
       (pg)->mdpage.mdpg_first.pv_pmap = NULL;                         \
       (pg)->mdpage.mdpg_first.pv_va = VM_PAGE_TO_PHYS(pg);            \
       (pg)->mdpage.mdpg_attrs = VM_PAGEMD_VMPAGE;                     \
       VM_PAGEMD_PVLIST_LOCK_INIT(&(pg)->mdpage);                      \
} while (/* CONSTCOND */ 0)

#endif /* _MODULE */

#endif /* _UVM_PMAP_VMPAGEMD_H_ */