/*-
* Copyright (c) 2014 The NetBSD Foundation, Inc.
* All rights reserved.
*
* This code is derived from software contributed to The NetBSD Foundation
* by Mindaugas Rasiukevicius.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
* ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
* TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
* POSSIBILITY OF SUCH DAMAGE.
*/
/*
* The packet queue (pktqueue) interface is a lockless IP input queue
* which also abstracts and handles network ISR scheduling. It provides
* a mechanism to enable receiver-side packet steering (RPS).
*/
struct pktqueue {
/*
* The lock used for a barrier mechanism. The barrier counter,
* as well as the drop counter, are managed atomically though.
* Ensure this group is in a separate cache line.
*/
union {
struct {
kmutex_t pq_lock;
volatile u_int pq_barrier;
};
uint8_t _pad[COHERENCY_UNIT];
};
/* The size of the queue, counters and the interrupt handler. */
u_int pq_maxlen;
percpu_t * pq_counters;
void * pq_sih;
/* Special marker value used by pktq_barrier() mechanism. */
#define PKTQ_MARKER ((void *)(~0ULL))
/*
* This is a list of all pktqueues. This list is used by
* pktq_ifdetach() to issue a barrier on every pktqueue.
*
* The r/w lock is acquired for writing in pktq_create() and
* pktq_destroy(), and for reading in pktq_ifdetach().
*
* This list is not performance critical, and will seldom be
* accessed.
*/
static LIST_HEAD(, pktqueue) pktqueue_list __read_mostly;
static krwlock_t pktqueue_list_lock __read_mostly;
static once_t pktqueue_list_init_once __read_mostly;
static int
pktqueue_list_init(void)
{
LIST_INIT(&pktqueue_list);
rw_init(&pktqueue_list_lock);
return 0;
}
/*
* As long as preemption is disabled, the xcall to swap percpu
* buffers can't complete, so it is safe to read the pointer.
*/
KASSERT(kpreempt_disabled());
/*
* - pktq_inc_counter: increment the counter given an ID.
* - pktq_collect_counts: handler to sum up the counts from each CPU.
* - pktq_getcount: return the effective count given an ID.
*/
static uint32_t
pktq_rps_hash_toeplitz(const struct mbuf *m)
{
struct ip *ip;
/*
* Disable UDP port - IP fragments aren't currently being handled
* and so we end up with a mix of 2-tuple and 4-tuple
* traffic.
*/
const u_int flag = RSS_TOEPLITZ_USE_TCP_PORT;
/* glance IP version */
if ((m->m_flags & M_PKTHDR) == 0)
return 0;
ip = mtod(m, struct ip *);
if (ip->ip_v == IPVERSION) {
if (__predict_false(m->m_len < sizeof(struct ip)))
return 0;
return rss_toeplitz_hash_from_mbuf_ipv4(m, flag);
} else if (ip->ip_v == 6) {
if (__predict_false(m->m_len < sizeof(struct ip6_hdr)))
return 0;
return rss_toeplitz_hash_from_mbuf_ipv6(m, flag);
}
return 0;
}
/*
* toeplitz without curcpu.
* Generally, this has better performance than toeplitz.
*/
static uint32_t
pktq_rps_hash_toeplitz_othercpus(const struct mbuf *m)
{
uint32_t hash;
/*
* pktq_enqueue: inject the packet into the end of the queue.
*
* => Must be called from the interrupt or with the preemption disabled.
* => Consumes the packet and returns true on success.
* => Returns false on failure; caller is responsible to free the packet.
*/
bool
pktq_enqueue(pktqueue_t *pq, struct mbuf *m, const u_int hash __unused)
{
#if defined(_RUMPKERNEL) || defined(_RUMP_NATIVE_ABI)
struct cpu_info *ci = curcpu();
#else
struct cpu_info *ci = cpu_lookup(hash % ncpu);
#endif
/*
* pktq_dequeue: take a packet from the queue.
*
* => Must be called with preemption disabled.
* => Must ensure there are not concurrent dequeue calls.
*/
struct mbuf *
pktq_dequeue(pktqueue_t *pq)
{
struct cpu_info *ci = curcpu();
struct mbuf *m;
KASSERT(kpreempt_disabled());
m = pcq_get(pktq_pcq(pq, ci));
if (__predict_false(m == PKTQ_MARKER)) {
/* Note the marker entry. */
atomic_inc_uint(&pq->pq_barrier);
/* Get the next queue entry. */
m = pcq_get(pktq_pcq(pq, ci));
/*
* There can only be one barrier operation pending
* on a pktqueue at any given time, so we can assert
* that the next item is not a marker.
*/
KASSERT(m != PKTQ_MARKER);
}
if (__predict_true(m != NULL)) {
pktq_inc_count(pq, PQCNT_DEQUEUE);
}
return m;
}
/*
* pktq_barrier: waits for a grace period when all packets enqueued at
* the moment of calling this routine will be processed. This is used
* to ensure that e.g. packets referencing some interface were drained.
*/
void
pktq_barrier(pktqueue_t *pq)
{
CPU_INFO_ITERATOR cii;
struct cpu_info *ci;
u_int pending = 0;
/* If the queue is empty - nothing to do. */
if (pcq_peek(q) == NULL) {
continue;
}
/* Otherwise, put the marker and entry. */
while (!pcq_put(q, PKTQ_MARKER)) {
kpause("pktqsync", false, 1, NULL);
}
kpreempt_disable();
softint_schedule_cpu(pq->pq_sih, ci);
kpreempt_enable();
pending++;
}
/* Wait for each queue to process the markers. */
while (pq->pq_barrier != pending) {
kpause("pktqsync", false, 1, NULL);
}
pq->pq_barrier = 0;
mutex_exit(&pq->pq_lock);
}
/*
* pktq_ifdetach: issue a barrier on all pktqueues when a network
* interface is detached.
*/
void
pktq_ifdetach(void)
{
pktqueue_t *pq;
/* Just in case no pktqueues have been created yet... */
RUN_ONCE(&pktqueue_list_init_once, pktqueue_list_init);
/*
* pktq_flush: free mbufs in all queues.
*
* => The caller must ensure there are no concurrent writers or flush calls.
*/
void
pktq_flush(pktqueue_t *pq)
{
CPU_INFO_ITERATOR cii;
struct cpu_info *ci;
struct mbuf *m, *m0 = NULL;
ASSERT_SLEEPABLE();
/*
* Run a dummy softint at IPL_SOFTNET on all CPUs to ensure that any
* already running handler for this pktqueue is no longer running.
*/
xc_barrier(XC_HIGHPRI_IPL(IPL_SOFTNET));
/*
* Acquire the barrier lock. While the caller ensures that
* no explicit pktq_barrier() calls will be issued, this holds
* off any implicit pktq_barrier() calls that would happen
* as the result of pktq_ifdetach().
*/
mutex_enter(&pq->pq_lock);
/*
* Pull the packets off the pcq and chain them into
* a list to be freed later.
*/
while ((m = pcq_get(q)) != NULL) {
pktq_inc_count(pq, PQCNT_DEQUEUE);
m->m_nextpkt = m0;
m0 = m;
}
}
mutex_exit(&pq->pq_lock);
/* Free the packets now that the critical section is over. */
while ((m = m0) != NULL) {
m0 = m->m_nextpkt;
m_freem(m);
}
}
/*
* pktq_set_maxlen: create per-CPU queues using a new size and replace
* the existing queues without losing any packets.
*
* XXX ncpu must remain stable throughout.
*/
int
pktq_set_maxlen(pktqueue_t *pq, size_t maxlen)
{
const u_int slotbytes = ncpu * sizeof(pcq_t *);
pcq_t **qs;
if (!maxlen || maxlen > PCQ_MAXLEN)
return EINVAL;
if (pq->pq_maxlen == maxlen)
return 0;
/* First, allocate the new queues. */
qs = kmem_zalloc(slotbytes, KM_SLEEP);
for (u_int i = 0; i < ncpu; i++) {
qs[i] = pcq_create(maxlen, KM_SLEEP);
}
/*
* Issue an xcall to replace the queue pointers on each CPU.
* This implies all the necessary memory barriers.
*/
mutex_enter(&pq->pq_lock);
xc_wait(xc_broadcast(XC_HIGHPRI, pktq_set_maxlen_cpu, pq, qs));
pq->pq_maxlen = maxlen;
mutex_exit(&pq->pq_lock);
/*
* At this point, the new packets are flowing into the new
* queues. However, the old queues may have some packets
* present which are no longer being processed. We are going
* to re-enqueue them. This may change the order of packet
* arrival, but it is not considered an issue.
*
* There may be in-flight interrupts calling pktq_dequeue()
* which reference the old queues. Issue a barrier to ensure
* that we are going to be the only pcq_get() callers on the
* old queues.
*/
pktq_barrier(pq);
for (u_int i = 0; i < ncpu; i++) {
struct pcq *q;
struct mbuf *m;
static int
sysctl_pktq_nitems(SYSCTLFN_ARGS)
{
return sysctl_pktq_count(SYSCTLFN_CALL(rnode), PKTQ_NITEMS);
}
static int
sysctl_pktq_drops(SYSCTLFN_ARGS)
{
return sysctl_pktq_count(SYSCTLFN_CALL(rnode), PKTQ_DROPS);
}
/*
* pktqueue_sysctl_setup: set up the sysctl nodes for a pktqueue
* using standardized names at the specified parent node and
* node ID (or CTL_CREATE).
*/
void
pktq_sysctl_setup(pktqueue_t * const pq, struct sysctllog ** const clog,
const struct sysctlnode * const parent_node, const int qid)
{
const struct sysctlnode *rnode = parent_node, *cnode;