/*-
* Copyright (c) 2020 Mindaugas Rasiukevicius <rmind at noxt eu>
* Copyright (c) 2009-2025 The NetBSD Foundation, Inc.
* All rights reserved.
*
* This material is based upon work partially supported by The
* NetBSD Foundation under a contract with Mindaugas Rasiukevicius.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
*    notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
*    notice, this list of conditions and the following disclaimer in the
*    documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
* POSSIBILITY OF SUCH DAMAGE.
*/

/*
* NPF ruleset module.
*/

#ifdef _KERNEL
#include <sys/cdefs.h>
__KERNEL_RCSID(0, "$NetBSD: npf_ruleset.c,v 1.55 2025/06/01 00:54:36 joe Exp $");

#include <sys/param.h>
#include <sys/types.h>

#include <sys/atomic.h>
#include <sys/kmem.h>
#include <sys/queue.h>
#include <sys/mbuf.h>
#include <sys/types.h>
#include <sys/kauth.h>

#include <net/bpf.h>
#include <net/bpfjit.h>
#include <net/pfil.h>
#include <net/if.h>
#endif

#include "npf_impl.h"

struct npf_ruleset {
       /*
        * - List of all rules.
        * - Dynamic (i.e. named) rules.
        * - G/C list for convenience.
        */
       LIST_HEAD(, npf_rule)   rs_all;
       LIST_HEAD(, npf_rule)   rs_dynamic;
       LIST_HEAD(, npf_rule)   rs_gc;

       /* Unique ID counter. */
       uint64_t                rs_idcnt;

       /* Number of array slots and active rules. */
       unsigned                rs_slots;
       unsigned                rs_nitems;

       /* Array of ordered rules. */
       npf_rule_t *            rs_rules[];
};

struct npf_rule {
       /* Attributes, interface and skip slot. */
       uint32_t                r_attr;
       unsigned                r_ifid;
       unsigned                r_skip_to;

       /* Code to process, if any. */
       int                     r_type;
       bpfjit_func_t           r_jcode;
       void *                  r_code;
       unsigned                r_clen;

       /* NAT policy (optional), rule procedure and subset. */
       npf_natpolicy_t *       r_natp;
       npf_rproc_t *           r_rproc;

       union {
               /*
                * Dynamic group: rule subset and a group list entry.
                */
               struct {
                       npf_rule_t *            r_subset;
                       LIST_ENTRY(npf_rule)    r_dentry;
               };

               /*
                * Dynamic rule: priority, parent group and next rule.
                */
               struct {
                       int                     r_priority;
                       npf_rule_t *            r_parent;
                       npf_rule_t *            r_next;
               };
       };

       /* Rule ID, name and the optional key. */
       uint64_t                r_id;
       char                    r_name[NPF_RULE_MAXNAMELEN];
       uint8_t                 r_key[NPF_RULE_MAXKEYLEN];

       /* All-list entry and the auxiliary info. */
       LIST_ENTRY(npf_rule)    r_aentry;
       nvlist_t *              r_info;
       size_t                  r_info_len;

       rid_t uid;
       rid_t gid;
};

#define SKIPTO_ADJ_FLAG         (1U << 31)
#define SKIPTO_MASK             (SKIPTO_ADJ_FLAG - 1)

static nvlist_t *       npf_rule_export(npf_t *, const npf_rule_t *);

/*
* Private attributes - must be in the NPF_RULE_PRIVMASK range.
*/
#define NPF_RULE_KEEPNAT        (0x01000000 & NPF_RULE_PRIVMASK)

#define NPF_DYNAMIC_GROUP_P(attr) \
   (((attr) & NPF_DYNAMIC_GROUP) == NPF_DYNAMIC_GROUP)

#define NPF_DYNAMIC_RULE_P(attr) \
   (((attr) & NPF_DYNAMIC_GROUP) == NPF_RULE_DYNAMIC)

npf_ruleset_t *
npf_ruleset_create(size_t slots)
{
       size_t len = offsetof(npf_ruleset_t, rs_rules[slots]);
       npf_ruleset_t *rlset;

       rlset = kmem_zalloc(len, KM_SLEEP);
       LIST_INIT(&rlset->rs_dynamic);
       LIST_INIT(&rlset->rs_all);
       LIST_INIT(&rlset->rs_gc);
       rlset->rs_slots = slots;

       return rlset;
}

void
npf_ruleset_destroy(npf_ruleset_t *rlset)
{
       size_t len = offsetof(npf_ruleset_t, rs_rules[rlset->rs_slots]);
       npf_rule_t *rl;

       while ((rl = LIST_FIRST(&rlset->rs_all)) != NULL) {
               if (NPF_DYNAMIC_GROUP_P(rl->r_attr)) {
                       /*
                        * Note: r_subset may point to the rules which
                        * were inherited by a new ruleset.
                        */
                       rl->r_subset = NULL;
                       LIST_REMOVE(rl, r_dentry);
               }
               if (NPF_DYNAMIC_RULE_P(rl->r_attr)) {
                       /* Not removing from r_subset, see above. */
                       KASSERT(rl->r_parent != NULL);
               }
               LIST_REMOVE(rl, r_aentry);
               npf_rule_free(rl);
       }
       KASSERT(LIST_EMPTY(&rlset->rs_dynamic));

       npf_ruleset_gc(rlset);
       KASSERT(LIST_EMPTY(&rlset->rs_gc));
       kmem_free(rlset, len);
}

/*
* npf_ruleset_insert: insert the rule into the specified ruleset.
*/
void
npf_ruleset_insert(npf_ruleset_t *rlset, npf_rule_t *rl)
{
       unsigned n = rlset->rs_nitems;

       KASSERT(n < rlset->rs_slots);

       LIST_INSERT_HEAD(&rlset->rs_all, rl, r_aentry);
       if (NPF_DYNAMIC_GROUP_P(rl->r_attr)) {
               LIST_INSERT_HEAD(&rlset->rs_dynamic, rl, r_dentry);
       } else {
               KASSERTMSG(rl->r_parent == NULL, "cannot be dynamic rule");
               rl->r_attr &= ~NPF_RULE_DYNAMIC;
       }

       rlset->rs_rules[n] = rl;
       rlset->rs_nitems++;
       rl->r_id = ++rlset->rs_idcnt;

       if (rl->r_skip_to < ++n) {
               rl->r_skip_to = SKIPTO_ADJ_FLAG | n;
       }
}

npf_rule_t *
npf_ruleset_lookup(npf_ruleset_t *rlset, const char *name)
{
       npf_rule_t *rl;

       LIST_FOREACH(rl, &rlset->rs_dynamic, r_dentry) {
               KASSERT(NPF_DYNAMIC_GROUP_P(rl->r_attr));
               if (strncmp(rl->r_name, name, NPF_RULE_MAXNAMELEN) == 0)
                       break;
       }
       return rl;
}

/*
* npf_ruleset_add: insert dynamic rule into the (active) ruleset.
*/
int
npf_ruleset_add(npf_ruleset_t *rlset, const char *rname, npf_rule_t *rl)
{
       npf_rule_t *rg, *it, *target;
       int priocmd;

       if (!NPF_DYNAMIC_RULE_P(rl->r_attr)) {
               return EINVAL;
       }
       rg = npf_ruleset_lookup(rlset, rname);
       if (rg == NULL) {
               return ESRCH;
       }

       /* Dynamic rule - assign a unique ID and save the parent. */
       rl->r_id = ++rlset->rs_idcnt;
       rl->r_parent = rg;

       /*
        * Rule priority: (highest) 1, 2 ... n (lowest).
        * Negative priority indicates an operation and is reset to zero.
        */
       if ((priocmd = rl->r_priority) < 0) {
               rl->r_priority = 0;
       }

       /*
        * WARNING: once rg->subset or target->r_next of an *active*
        * rule is set, then our rule becomes globally visible and active.
        * Must issue a load fence to ensure rl->r_next visibility first.
        */
       switch (priocmd) {
       case NPF_PRI_LAST:
       default:
               target = NULL;
               it = rg->r_subset;
               while (it && it->r_priority <= rl->r_priority) {
                       target = it;
                       it = it->r_next;
               }
               if (target) {
                       atomic_store_relaxed(&rl->r_next, target->r_next);
                       membar_producer();
                       atomic_store_relaxed(&target->r_next, rl);
                       break;
               }
               /* FALLTHROUGH */

       case NPF_PRI_FIRST:
               atomic_store_relaxed(&rl->r_next, rg->r_subset);
               membar_producer();
               atomic_store_relaxed(&rg->r_subset, rl);
               break;
       }

       /* Finally, add into the all-list. */
       LIST_INSERT_HEAD(&rlset->rs_all, rl, r_aentry);
       return 0;
}

static void
npf_ruleset_unlink(npf_rule_t *rl, npf_rule_t *prev)
{
       KASSERT(NPF_DYNAMIC_RULE_P(rl->r_attr));
       if (prev) {
               prev->r_next = rl->r_next;
       } else {
               npf_rule_t *rg = rl->r_parent;
               rg->r_subset = rl->r_next;
       }
       LIST_REMOVE(rl, r_aentry);
}

/*
* npf_ruleset_remove: remove the dynamic rule given the rule ID.
*/
int
npf_ruleset_remove(npf_ruleset_t *rlset, const char *rname, uint64_t id)
{
       npf_rule_t *rg, *prev = NULL;

       if ((rg = npf_ruleset_lookup(rlset, rname)) == NULL) {
               return ESRCH;
       }
       for (npf_rule_t *rl = rg->r_subset; rl; rl = rl->r_next) {
               KASSERT(rl->r_parent == rg);
               KASSERT(NPF_DYNAMIC_RULE_P(rl->r_attr));

               /* Compare ID.  On match, remove and return. */
               if (rl->r_id == id) {
                       npf_ruleset_unlink(rl, prev);
                       LIST_INSERT_HEAD(&rlset->rs_gc, rl, r_aentry);
                       return 0;
               }
               prev = rl;
       }
       return ENOENT;
}

/*
* npf_ruleset_remkey: remove the dynamic rule given the rule key.
*/
int
npf_ruleset_remkey(npf_ruleset_t *rlset, const char *rname,
   const void *key, size_t len)
{
       npf_rule_t *rg, *rlast = NULL, *prev = NULL, *lastprev = NULL;

       KASSERT(len && len <= NPF_RULE_MAXKEYLEN);

       if ((rg = npf_ruleset_lookup(rlset, rname)) == NULL) {
               return ESRCH;
       }

       /* Compare the key and find the last in the list. */
       for (npf_rule_t *rl = rg->r_subset; rl; rl = rl->r_next) {
               KASSERT(rl->r_parent == rg);
               KASSERT(NPF_DYNAMIC_RULE_P(rl->r_attr));
               if (memcmp(rl->r_key, key, len) == 0) {
                       lastprev = prev;
                       rlast = rl;
               }
               prev = rl;
       }
       if (!rlast) {
               return ENOENT;
       }
       npf_ruleset_unlink(rlast, lastprev);
       LIST_INSERT_HEAD(&rlset->rs_gc, rlast, r_aentry);
       return 0;
}

/*
* npf_ruleset_list: serialise and return the dynamic rules.
*/
int
npf_ruleset_list(npf_t *npf, npf_ruleset_t *rlset, const char *rname,
   nvlist_t *rlset_nvl)
{
       const npf_rule_t *rg;

       KASSERT(npf_config_locked_p(npf));

       if ((rg = npf_ruleset_lookup(rlset, rname)) == NULL) {
               return ESRCH;
       }
       for (const npf_rule_t *rl = rg->r_subset; rl; rl = rl->r_next) {
               nvlist_t *rule;

               KASSERT(rl->r_parent == rg);
               KASSERT(NPF_DYNAMIC_RULE_P(rl->r_attr));

               if ((rule = npf_rule_export(npf, rl)) == NULL) {
                       return ENOMEM;
               }
               nvlist_append_nvlist_array(rlset_nvl, "rules", rule);
               nvlist_destroy(rule);
       }
       return 0;
}

/*
* npf_ruleset_flush: flush the dynamic rules in the ruleset by inserting
* them into the G/C list.
*/
int
npf_ruleset_flush(npf_ruleset_t *rlset, const char *rname)
{
       npf_rule_t *rg, *rl;

       if ((rg = npf_ruleset_lookup(rlset, rname)) == NULL) {
               return ESRCH;
       }

       rl = atomic_swap_ptr(&rg->r_subset, NULL);
       membar_producer();

       while (rl) {
               KASSERT(NPF_DYNAMIC_RULE_P(rl->r_attr));
               KASSERT(rl->r_parent == rg);

               LIST_REMOVE(rl, r_aentry);
               LIST_INSERT_HEAD(&rlset->rs_gc, rl, r_aentry);
               rl = rl->r_next;
       }
       rlset->rs_idcnt = 0;
       return 0;
}

/*
* npf_ruleset_gc: destroy the rules in G/C list.
*/
void
npf_ruleset_gc(npf_ruleset_t *rlset)
{
       npf_rule_t *rl;

       while ((rl = LIST_FIRST(&rlset->rs_gc)) != NULL) {
               LIST_REMOVE(rl, r_aentry);
               npf_rule_free(rl);
       }
}

/*
* npf_ruleset_export: serialise and return the static rules.
*/
int
npf_ruleset_export(npf_t *npf, const npf_ruleset_t *rlset,
   const char *key, nvlist_t *npf_nv)
{
       const unsigned nitems = rlset->rs_nitems;
       unsigned n = 0;
       int error = 0;

       KASSERT(npf_config_locked_p(npf));

       while (n < nitems) {
               const npf_rule_t *rl = rlset->rs_rules[n];
               const npf_natpolicy_t *natp = rl->r_natp;
               nvlist_t *rule;

               rule = npf_rule_export(npf, rl);
               if (!rule) {
                       error = ENOMEM;
                       break;
               }
               if (natp && (error = npf_natpolicy_export(natp, rule)) != 0) {
                       nvlist_destroy(rule);
                       break;
               }
               nvlist_append_nvlist_array(npf_nv, key, rule);
               nvlist_destroy(rule);
               n++;
       }
       return error;
}

/*
* npf_ruleset_reload: prepare the new ruleset by scanning the active
* ruleset and: 1) sharing the dynamic rules 2) sharing NAT policies.
*
* => The active (old) ruleset should be exclusively locked.
*/
void
npf_ruleset_reload(npf_t *npf, npf_ruleset_t *newset,
   npf_ruleset_t *oldset, bool load)
{
       npf_rule_t *rg, *rl;
       uint64_t nid = 0;

       KASSERT(npf_config_locked_p(npf));

       /*
        * Scan the dynamic rules and share (migrate) if needed.
        */
       LIST_FOREACH(rg, &newset->rs_dynamic, r_dentry) {
               npf_rule_t *active_rgroup;

               /* Look for a dynamic ruleset group with such name. */
               active_rgroup = npf_ruleset_lookup(oldset, rg->r_name);
               if (active_rgroup == NULL) {
                       continue;
               }

               /*
                * ATOMICITY: Copy the head pointer of the linked-list,
                * but do not remove the rules from the active r_subset.
                * This is necessary because the rules are still active
                * and therefore are accessible for inspection via the
                * old ruleset.
                */
               rg->r_subset = active_rgroup->r_subset;

               /*
                * We can safely migrate to the new all-rule list and
                * reset the parent rule, though.
                */
               for (rl = rg->r_subset; rl; rl = rl->r_next) {
                       KASSERT(NPF_DYNAMIC_RULE_P(rl->r_attr));
                       LIST_REMOVE(rl, r_aentry);
                       LIST_INSERT_HEAD(&newset->rs_all, rl, r_aentry);

                       KASSERT(rl->r_parent == active_rgroup);
                       rl->r_parent = rg;
               }
       }

       /*
        * If performing the load of connections then NAT policies might
        * already have translated connections associated with them and
        * we should not share or inherit anything.
        */
       if (load)
               return;

       /*
        * Scan all rules in the new ruleset and inherit the active NAT
        * policies if they are the same.  Also, assign a unique ID for
        * each policy here.
        */
       LIST_FOREACH(rl, &newset->rs_all, r_aentry) {
               npf_natpolicy_t *np;
               npf_rule_t *actrl;

               /* Does the rule have a NAT policy associated? */
               if ((np = rl->r_natp) == NULL) {
                       continue;
               }

               /* Does it match with any policy in the active ruleset? */
               LIST_FOREACH(actrl, &oldset->rs_all, r_aentry) {
                       if (!actrl->r_natp)
                               continue;
                       if ((actrl->r_attr & NPF_RULE_KEEPNAT) != 0)
                               continue;
                       if (npf_natpolicy_cmp(actrl->r_natp, np))
                               break;
               }
               if (!actrl) {
                       /* No: just set the ID and continue. */
                       npf_nat_setid(np, ++nid);
                       continue;
               }

               /* Yes: inherit the matching NAT policy. */
               rl->r_natp = actrl->r_natp;
               npf_nat_setid(rl->r_natp, ++nid);

               /*
                * Finally, mark the active rule to not destroy its NAT
                * policy later as we inherited it (but the rule must be
                * kept active for now).  Destroy the new/unused policy.
                */
               actrl->r_attr |= NPF_RULE_KEEPNAT;
               npf_natpolicy_destroy(np);
       }

       /* Inherit the ID counter. */
       newset->rs_idcnt = oldset->rs_idcnt;
}

/*
* npf_ruleset_findnat: find a NAT policy in the ruleset by a given ID.
*/
npf_natpolicy_t *
npf_ruleset_findnat(npf_ruleset_t *rlset, uint64_t id)
{
       npf_rule_t *rl;

       LIST_FOREACH(rl, &rlset->rs_all, r_aentry) {
               npf_natpolicy_t *np = rl->r_natp;
               if (np && npf_nat_getid(np) == id) {
                       return np;
               }
       }
       return NULL;
}

/*
* npf_ruleset_freealg: inspect the ruleset and disassociate specified
* ALG from all NAT entries using it.
*/
void
npf_ruleset_freealg(npf_ruleset_t *rlset, npf_alg_t *alg)
{
       npf_rule_t *rl;
       npf_natpolicy_t *np;

       LIST_FOREACH(rl, &rlset->rs_all, r_aentry) {
               if ((np = rl->r_natp) != NULL) {
                       npf_nat_freealg(np, alg);
               }
       }
}

/*
* npf_rule_alloc: allocate a rule and initialise it.
*/
npf_rule_t *
npf_rule_alloc(npf_t *npf, const nvlist_t *rule)
{
       npf_rule_t *rl;
       const char *rname;
       const void *key, *info;
       size_t len;

       /* Allocate a rule structure and keep the information. */
       rl = kmem_zalloc(sizeof(npf_rule_t), KM_SLEEP);
       info = dnvlist_get_binary(rule, "info", &rl->r_info_len, NULL, 0);
       if (info) {
               rl->r_info = kmem_alloc(rl->r_info_len, KM_SLEEP);
               memcpy(rl->r_info, info, rl->r_info_len);
       }
       rl->r_natp = NULL;

       /* Name (optional) */
       if ((rname = dnvlist_get_string(rule, "name", NULL)) != NULL) {
               strlcpy(rl->r_name, rname, NPF_RULE_MAXNAMELEN);
       } else {
               rl->r_name[0] = '\0';
       }

       /* Attributes, priority and interface ID (optional). */
       rl->r_attr = dnvlist_get_number(rule, "attr", 0);
       rl->r_attr &= ~NPF_RULE_PRIVMASK;

       if (NPF_DYNAMIC_RULE_P(rl->r_attr)) {
               /* Priority of the dynamic rule. */
               rl->r_priority = (int)dnvlist_get_number(rule, "prio", 0);
       } else {
               /* The skip-to index.  No need to validate it. */
               rl->r_skip_to = dnvlist_get_number(rule, "skip-to", 0);
       }

       /* Interface name; register and get the npf-if-id. */
       if ((rname = dnvlist_get_string(rule, "ifname", NULL)) != NULL) {
               if ((rl->r_ifid = npf_ifmap_register(npf, rname)) == 0) {
                       kmem_free(rl, sizeof(npf_rule_t));
                       return NULL;
               }
       } else {
               rl->r_ifid = 0;
       }

       /* Key (optional). */
       if ((key = dnvlist_get_binary(rule, "key", &len, NULL, 0)) != NULL) {
               if (len > NPF_RULE_MAXKEYLEN) {
                       kmem_free(rl, sizeof(npf_rule_t));
                       return NULL;
               }
               memcpy(rl->r_key, key, len);
       }

       /* no gid/uid set yet */
       rl->gid.op = rl->uid.op = NPF_OP_NONE;
       return rl;
}

static void
npf_rid_export(nvlist_t *rl, struct r_id rid, const char *name)
{
       uint64_t uid_element[3] = { rid.id[0], rid.id[1], rid.op };
       nvlist_add_number_array(rl, name, uid_element, 3);
}

static nvlist_t *
npf_rule_export(npf_t *npf, const npf_rule_t *rl)
{
       nvlist_t *rule = nvlist_create(0);
       unsigned skip_to = 0;
       npf_rproc_t *rp;

       nvlist_add_number(rule, "attr", rl->r_attr);
       nvlist_add_number(rule, "prio", rl->r_priority);
       if ((rl->r_skip_to & SKIPTO_ADJ_FLAG) == 0) {
               skip_to = rl->r_skip_to & SKIPTO_MASK;
       }
       nvlist_add_number(rule, "skip-to", skip_to);
       nvlist_add_number(rule, "code-type", rl->r_type);
       if (rl->r_code) {
               nvlist_add_binary(rule, "code", rl->r_code, rl->r_clen);
       }
       if (rl->r_ifid) {
               char ifname[IFNAMSIZ];
               npf_ifmap_copyname(npf, rl->r_ifid, ifname, sizeof(ifname));
               nvlist_add_string(rule, "ifname", ifname);
       }
       nvlist_add_number(rule, "id", rl->r_id);

       if (rl->r_name[0]) {
               nvlist_add_string(rule, "name", rl->r_name);
       }
       if (NPF_DYNAMIC_RULE_P(rl->r_attr)) {
               nvlist_add_binary(rule, "key", rl->r_key, NPF_RULE_MAXKEYLEN);
       }
       if (rl->r_info) {
               nvlist_add_binary(rule, "info", rl->r_info, rl->r_info_len);
       }
       if (rl->uid.op != NPF_OP_NONE) {
               npf_rid_export(rule, rl->uid, "r_user");
       }
       if (rl->gid.op != NPF_OP_NONE) {
               npf_rid_export(rule, rl->gid, "r_group");
       }
       if ((rp = npf_rule_getrproc(rl)) != NULL) {
               const char *rname = npf_rproc_getname(rp);
               nvlist_add_string(rule, "rproc", rname);
               npf_rproc_release(rp);
       }
       return rule;
}

/*
* npf_rule_setcode: assign filter code to the rule.
*
* => The code must be validated by the caller.
* => JIT compilation may be performed here.
*/
void
npf_rule_setcode(npf_rule_t *rl, const int type, void *code, size_t size)
{
       KASSERT(type == NPF_CODE_BPF);

       rl->r_type = type;
       rl->r_code = code;
       rl->r_clen = size;
       rl->r_jcode = npf_bpf_compile(code, size);
}

void
npf_rule_setrid(const nvlist_t *req, npf_rule_t *rl, const char *name)
{
       size_t nitems;
       rid_t id;
       const uint64_t *rid = nvlist_get_number_array(req, name, &nitems);
       KASSERT(nitems == 3);

       id.id[0] = (uint32_t)rid[0];
       id.id[1] = (uint32_t)rid[1];
       id.op = (uint8_t)rid[2];

       if (!strcmp(name, "r_user"))
               rl->uid = id;
       else if (!strcmp(name, "r_group"))
               rl->gid = id;
}

/*
* npf_rule_setrproc: assign a rule procedure and hold a reference on it.
*/
void
npf_rule_setrproc(npf_rule_t *rl, npf_rproc_t *rp)
{
       npf_rproc_acquire(rp);
       rl->r_rproc = rp;
}

/*
* npf_rule_free: free the specified rule.
*/
void
npf_rule_free(npf_rule_t *rl)
{
       npf_natpolicy_t *np = rl->r_natp;
       npf_rproc_t *rp = rl->r_rproc;

       if (np && (rl->r_attr & NPF_RULE_KEEPNAT) == 0) {
               /* Destroy the NAT policy. */
               npf_natpolicy_destroy(np);
       }
       if (rp) {
               /* Release rule procedure. */
               npf_rproc_release(rp);
       }
       if (rl->r_code) {
               /* Free byte-code. */
               kmem_free(rl->r_code, rl->r_clen);
       }
       if (rl->r_jcode) {
               /* Free JIT code. */
               bpf_jit_freecode(rl->r_jcode);
       }
       if (rl->r_info) {
               kmem_free(rl->r_info, rl->r_info_len);
       }
       kmem_free(rl, sizeof(npf_rule_t));
}

/*
* npf_rule_getid: return the unique ID of a rule.
* npf_rule_getrproc: acquire a reference and return rule procedure, if any.
* npf_rule_getnat: get NAT policy assigned to the rule.
*/

uint64_t
npf_rule_getid(const npf_rule_t *rl)
{
       KASSERT(NPF_DYNAMIC_RULE_P(rl->r_attr));
       return rl->r_id;
}

npf_rproc_t *
npf_rule_getrproc(const npf_rule_t *rl)
{
       npf_rproc_t *rp = rl->r_rproc;

       if (rp) {
               npf_rproc_acquire(rp);
       }
       return rp;
}

npf_natpolicy_t *
npf_rule_getnat(const npf_rule_t *rl)
{
       return rl->r_natp;
}

/*
* npf_rule_setnat: assign NAT policy to the rule and insert into the
* NAT policy list in the ruleset.
*/
void
npf_rule_setnat(npf_rule_t *rl, npf_natpolicy_t *np)
{
       KASSERT(rl->r_natp == NULL);
       rl->r_natp = np;
}

/*
* npf_rule_inspect: match the interface, direction and run the filter code.
* Returns true if rule matches and false otherwise.
*/
static inline bool
npf_rule_inspect(const npf_rule_t *rl, bpf_args_t *bc_args,
   const int di_mask, const unsigned ifid)
{
       /* Match the interface. */
       if (rl->r_ifid && rl->r_ifid != ifid) {
               return false;
       }

       /* Match the direction. */
       if ((rl->r_attr & NPF_RULE_DIMASK) != NPF_RULE_DIMASK) {
               if ((rl->r_attr & di_mask) == 0)
                       return false;
       }

       /* Any code? */
       if (!rl->r_code) {
               KASSERT(rl->r_jcode == NULL);
               return true;
       }
       KASSERT(rl->r_type == NPF_CODE_BPF);
       return npf_bpf_filter(bc_args, rl->r_code, rl->r_jcode) != 0;
}

/*
* npf_rule_reinspect: re-inspect the dynamic rule by iterating its list.
* This is only for the dynamic rules.  Subrules cannot have nested rules.
*/
static inline npf_rule_t *
npf_rule_reinspect(const npf_rule_t *rg, bpf_args_t *bc_args,
   const int di_mask, const unsigned ifid)
{
       npf_rule_t *final_rl = NULL, *rl;

       KASSERT(NPF_DYNAMIC_GROUP_P(rg->r_attr));

       rl = atomic_load_relaxed(&rg->r_subset);
       for (; rl; rl = atomic_load_relaxed(&rl->r_next)) {
               KASSERT(!final_rl || rl->r_priority >= final_rl->r_priority);
               if (!npf_rule_inspect(rl, bc_args, di_mask, ifid)) {
                       continue;
               }
               if (rl->r_attr & NPF_RULE_FINAL) {
                       return rl;
               }
               final_rl = rl;
       }
       return final_rl;
}

/*
* npf_ruleset_inspect: inspect the packet against the given ruleset.
*
* Loop through the rules in the set and run the byte-code of each rule
* against the packet (nbuf chain).  If sub-ruleset is found, inspect it.
*/
npf_rule_t *
npf_ruleset_inspect(npf_cache_t *npc, const npf_ruleset_t *rlset,
   const int di, const int layer)
{
       nbuf_t *nbuf = npc->npc_nbuf;
       const int di_mask = (di & PFIL_IN) ? NPF_RULE_IN : NPF_RULE_OUT;
       const unsigned nitems = rlset->rs_nitems;
       const unsigned ifid = nbuf->nb_ifid;
       npf_rule_t *final_rl = NULL;
       bpf_args_t bc_args;
       unsigned n = 0;

       KASSERT(((di & PFIL_IN) != 0) ^ ((di & PFIL_OUT) != 0));

       /*
        * Prepare the external memory store and the arguments for
        * the BPF programs to be executed.  Reset mbuf before taking
        * any pointers for the BPF.
        */
       uint32_t bc_words[NPF_BPF_NWORDS];

       nbuf_reset(nbuf);
       npf_bpf_prepare(npc, &bc_args, bc_words);

       while (n < nitems) {
               npf_rule_t *rl = rlset->rs_rules[n];
               const unsigned skip_to = rl->r_skip_to & SKIPTO_MASK;
               const uint32_t attr = rl->r_attr;

               KASSERT(!nbuf_flag_p(nbuf, NBUF_DATAREF_RESET));
               KASSERT(n < skip_to);

               /* Group is a barrier: return a matching if found any. */
               if ((attr & NPF_DYNAMIC_GROUP) == NPF_RULE_GROUP && final_rl) {
                       break;
               }

               /* Main inspection of the rule. */
               if (!npf_rule_inspect(rl, &bc_args, di_mask, ifid)) {
                       n = skip_to;
                       continue;
               }

               if (NPF_DYNAMIC_GROUP_P(attr)) {
                       /*
                        * If this is a dynamic rule, re-inspect the subrules.
                        * If it has any matching rule, then it is final.
                        */
                       rl = npf_rule_reinspect(rl, &bc_args, di_mask, ifid);
                       if (rl != NULL) {
                               final_rl = rl;
                               break;
                       }
               } else if ((attr & NPF_RULE_GROUP) == 0) {
                       /*
                        * Groups themselves are not matching.
                        */
                       final_rl = rl;
               }

               /* Set the matching rule and check for "final". */
               if (attr & NPF_RULE_FINAL) {
                       break;
               }
               n++;
       }

       KASSERT(!nbuf_flag_p(nbuf, NBUF_DATAREF_RESET));
       return final_rl;
}

/*
* just exchange the flag attributes for pass/block for the diff protocols.
* for passing, we set the STATEFULNESS for TCP connection establishment
* if ret == 0, it is for a pass to be changed to block
* non-zero ret indicates a block to pass
* when we change to block, we assume the default RST rerturn for TCP
* when we change to pass, we ensure no bit field for RST for tcp and ICMP for udp
* finally change the ret condition too
*/
int
npf_rule_reverse(npf_cache_t *npc, npf_match_info_t *mi, int ret)
{
       KASSERT(npf_iscached(npc, NPC_LAYER4));
       switch(npc->npc_proto) {
               case IPPROTO_TCP:
                       if (ret == 0) /* switch pass to block */ {
                               mi->mi_retfl &= !(NPF_RULE_PASS | NPF_RULE_STATEFUL |
                                       NPF_RULE_GSTATEFUL);
                               mi->mi_retfl |= NPF_RULE_RETRST;
                       }
                       else /* block to pass */ {
                               mi->mi_retfl &= !(NPF_RULE_RETRST);
                               mi->mi_retfl |= (NPF_RULE_PASS | NPF_RULE_STATEFUL |
                                       NPF_RULE_GSTATEFUL);
                       }
                       break;
               case IPPROTO_UDP:
                       if (ret == 0) /* pass to block */ {
                               mi->mi_retfl &= !(NPF_RULE_PASS);
                               mi->mi_retfl |= NPF_RULE_RETICMP;
                       }
                       else /* block to pass */ {
                               mi->mi_retfl &= !(NPF_RULE_RETICMP);
                               mi->mi_retfl |= NPF_RULE_PASS;
                       }
                       break;
       }

       return (ret == 0) ? ENETUNREACH : 0;
}

/* only perform uid/gid checks when set */
int
npf_rule_match_rid(npf_rule_t *rl, npf_cache_t *npc, int dir)
{
       uint32_t sock_gid, sock_uid;
       bool uid_matched = false, gid_matched = false;

       if (rl->gid.op == NPF_OP_NONE && rl->uid.op == NPF_OP_NONE)
               return -1; /* quickly return if packet has nothing to do with rids */

       KASSERT(npf_iscached(npc, NPC_IP46));
       KASSERT(npf_iscached(npc, NPC_LAYER4));

       if (rl->gid.op != NPF_OP_NONE) {
               if (npf_socket_lookup_rid(npc, kauth_cred_getegid, &sock_gid, dir) == -1)
                       return ENOTCONN;

               gid_matched |= npf_match_rid(&rl->gid, sock_gid);
       }
       if (rl->uid.op != NPF_OP_NONE) {
               if (npf_socket_lookup_rid(npc, kauth_cred_geteuid, &sock_uid, dir) == -1)
                       return ENOTCONN;

               uid_matched |= npf_match_rid(&rl->uid, sock_uid);
       }

       /* if both uid and gid are set on rule, both must be matching to agree */
       if (rl->gid.op && rl->uid.op)
               return gid_matched && uid_matched;
       else
               return gid_matched || uid_matched;
}

/*
* npf_rule_conclude: return decision and the flags for conclusion.
*
* => Returns ENETUNREACH if "block" and 0 if "pass".
*/
int
npf_rule_conclude(const npf_rule_t *rl, npf_match_info_t *mi)
{
       /* If not passing - drop the packet. */
       mi->mi_retfl = rl->r_attr;
       mi->mi_rid = rl->r_id;
       return (rl->r_attr & NPF_RULE_PASS) ? 0 : ENETUNREACH;
}


#if defined(DDB) || defined(_NPF_TESTING)

void
npf_ruleset_dump(npf_t *npf, const char *name)
{
       npf_ruleset_t *rlset = npf_config_ruleset(npf);
       npf_rule_t *rg, *rl;

       LIST_FOREACH(rg, &rlset->rs_dynamic, r_dentry) {
               printf("ruleset '%s':\n", rg->r_name);
               for (rl = rg->r_subset; rl; rl = rl->r_next) {
                       printf("\tid %"PRIu64", key: ", rl->r_id);
                       for (unsigned i = 0; i < NPF_RULE_MAXKEYLEN; i++)
                               printf("%x", rl->r_key[i]);
                       printf("\n");
               }
       }
}

#endif