2017-04-21 10:43:26 +00:00
|
|
|
/*
|
2019-06-25 11:12:58 +00:00
|
|
|
* SPDX-License-Identifier: BSD-3-Clause
|
|
|
|
* Copyright 2015 Intel Corporation.
|
|
|
|
* Copyright 2010-2011 Dmitry Vyukov
|
2017-04-21 10:43:26 +00:00
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef LTHREAD_QUEUE_H_
|
|
|
|
#define LTHREAD_QUEUE_H_
|
|
|
|
|
2018-05-15 09:49:22 +00:00
|
|
|
#ifdef __cplusplus
|
|
|
|
extern "C" {
|
|
|
|
#endif
|
|
|
|
|
2017-04-21 10:43:26 +00:00
|
|
|
#include <string.h>
|
|
|
|
|
|
|
|
#include <rte_prefetch.h>
|
|
|
|
#include <rte_per_lcore.h>
|
|
|
|
|
|
|
|
#include "lthread_int.h"
|
|
|
|
#include "lthread.h"
|
|
|
|
#include "lthread_diag.h"
|
|
|
|
#include "lthread_pool.h"
|
|
|
|
|
|
|
|
struct lthread_queue;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This file implements an unbounded FIFO queue based on a lock free
|
|
|
|
* linked list.
|
|
|
|
*
|
|
|
|
* The queue is non-intrusive in that it uses intermediate nodes, and does
|
|
|
|
* not require these nodes to be inserted into the object being placed
|
|
|
|
* in the queue.
|
|
|
|
*
|
|
|
|
* This is slightly more efficient than the very similar queue in lthread_pool
|
|
|
|
* in that it does not have to swing a stub node as the queue becomes empty.
|
|
|
|
*
|
|
|
|
* The queue access functions allocate and free intermediate node
|
|
|
|
* transparently from/to a per scheduler pool ( see lthread_pool.h ).
|
|
|
|
*
|
|
|
|
* The queue provides both MPSC and SPSC insert methods
|
|
|
|
*/
|
|
|
|
|
|
|
|
/*
|
|
|
|
* define a queue of lthread nodes
|
|
|
|
*/
|
|
|
|
struct lthread_queue {
|
|
|
|
struct qnode *head;
|
|
|
|
struct qnode *tail __rte_cache_aligned;
|
|
|
|
struct lthread_queue *p;
|
|
|
|
char name[LT_MAX_NAME_SIZE];
|
|
|
|
|
|
|
|
DIAG_COUNT_DEFINE(rd);
|
|
|
|
DIAG_COUNT_DEFINE(wr);
|
|
|
|
DIAG_COUNT_DEFINE(size);
|
|
|
|
|
|
|
|
} __rte_cache_aligned;
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
static inline struct lthread_queue *
|
|
|
|
_lthread_queue_create(const char *name)
|
|
|
|
{
|
|
|
|
struct qnode *stub;
|
|
|
|
struct lthread_queue *new_queue;
|
|
|
|
|
|
|
|
new_queue = rte_malloc_socket(NULL, sizeof(struct lthread_queue),
|
|
|
|
RTE_CACHE_LINE_SIZE,
|
|
|
|
rte_socket_id());
|
|
|
|
if (new_queue == NULL)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
/* allocated stub node */
|
|
|
|
stub = _qnode_alloc();
|
|
|
|
RTE_ASSERT(stub);
|
|
|
|
|
|
|
|
if (name != NULL)
|
|
|
|
strncpy(new_queue->name, name, sizeof(new_queue->name));
|
|
|
|
new_queue->name[sizeof(new_queue->name)-1] = 0;
|
|
|
|
|
|
|
|
/* initialize queue as empty */
|
|
|
|
stub->next = NULL;
|
|
|
|
new_queue->head = stub;
|
|
|
|
new_queue->tail = stub;
|
|
|
|
|
|
|
|
DIAG_COUNT_INIT(new_queue, rd);
|
|
|
|
DIAG_COUNT_INIT(new_queue, wr);
|
|
|
|
DIAG_COUNT_INIT(new_queue, size);
|
|
|
|
|
|
|
|
return new_queue;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Return true if the queue is empty
|
|
|
|
*/
|
2018-05-15 09:49:22 +00:00
|
|
|
static __rte_always_inline int
|
2017-04-21 10:43:26 +00:00
|
|
|
_lthread_queue_empty(struct lthread_queue *q)
|
|
|
|
{
|
|
|
|
return q->tail == q->head;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Destroy a queue
|
|
|
|
* fail if queue is not empty
|
|
|
|
*/
|
|
|
|
static inline int _lthread_queue_destroy(struct lthread_queue *q)
|
|
|
|
{
|
|
|
|
if (q == NULL)
|
|
|
|
return -1;
|
|
|
|
|
|
|
|
if (!_lthread_queue_empty(q))
|
|
|
|
return -1;
|
|
|
|
|
|
|
|
_qnode_free(q->head);
|
|
|
|
rte_free(q);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
RTE_DECLARE_PER_LCORE(struct lthread_sched *, this_sched);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Insert a node into a queue
|
|
|
|
* this implementation is multi producer safe
|
|
|
|
*/
|
2018-05-15 09:49:22 +00:00
|
|
|
static __rte_always_inline struct qnode *
|
2017-04-21 10:43:26 +00:00
|
|
|
_lthread_queue_insert_mp(struct lthread_queue
|
|
|
|
*q, void *data)
|
|
|
|
{
|
|
|
|
struct qnode *prev;
|
|
|
|
struct qnode *n = _qnode_alloc();
|
|
|
|
|
|
|
|
if (n == NULL)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
/* set object in node */
|
|
|
|
n->data = data;
|
|
|
|
n->next = NULL;
|
|
|
|
|
|
|
|
/* this is an MPSC method, perform a locked update */
|
|
|
|
prev = n;
|
|
|
|
prev =
|
|
|
|
(struct qnode *)__sync_lock_test_and_set((uint64_t *) &(q)->head,
|
|
|
|
(uint64_t) prev);
|
|
|
|
/* there is a window of inconsistency until prev next is set,
|
|
|
|
* which is why remove must retry
|
|
|
|
*/
|
|
|
|
prev->next = n;
|
|
|
|
|
|
|
|
DIAG_COUNT_INC(q, wr);
|
|
|
|
DIAG_COUNT_INC(q, size);
|
|
|
|
|
|
|
|
return n;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Insert an node into a queue in single producer mode
|
|
|
|
* this implementation is NOT mult producer safe
|
|
|
|
*/
|
2018-05-15 09:49:22 +00:00
|
|
|
static __rte_always_inline struct qnode *
|
2017-04-21 10:43:26 +00:00
|
|
|
_lthread_queue_insert_sp(struct lthread_queue
|
|
|
|
*q, void *data)
|
|
|
|
{
|
|
|
|
/* allocate a queue node */
|
|
|
|
struct qnode *prev;
|
|
|
|
struct qnode *n = _qnode_alloc();
|
|
|
|
|
|
|
|
if (n == NULL)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
/* set data in node */
|
|
|
|
n->data = data;
|
|
|
|
n->next = NULL;
|
|
|
|
|
|
|
|
/* this is an SPSC method, no need for locked exchange operation */
|
|
|
|
prev = q->head;
|
|
|
|
prev->next = q->head = n;
|
|
|
|
|
|
|
|
DIAG_COUNT_INC(q, wr);
|
|
|
|
DIAG_COUNT_INC(q, size);
|
|
|
|
|
|
|
|
return n;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Remove a node from a queue
|
|
|
|
*/
|
2018-05-15 09:49:22 +00:00
|
|
|
static __rte_always_inline void *
|
2017-04-21 10:43:26 +00:00
|
|
|
_lthread_queue_poll(struct lthread_queue *q)
|
|
|
|
{
|
|
|
|
void *data = NULL;
|
|
|
|
struct qnode *tail = q->tail;
|
|
|
|
struct qnode *next = (struct qnode *)tail->next;
|
|
|
|
/*
|
|
|
|
* There is a small window of inconsistency between producer and
|
|
|
|
* consumer whereby the queue may appear empty if consumer and
|
|
|
|
* producer access it at the same time.
|
|
|
|
* The consumer must handle this by retrying
|
|
|
|
*/
|
|
|
|
|
|
|
|
if (likely(next != NULL)) {
|
|
|
|
q->tail = next;
|
|
|
|
tail->data = next->data;
|
|
|
|
data = tail->data;
|
|
|
|
|
|
|
|
/* free the node */
|
|
|
|
_qnode_free(tail);
|
|
|
|
|
|
|
|
DIAG_COUNT_INC(q, rd);
|
|
|
|
DIAG_COUNT_DEC(q, size);
|
|
|
|
return data;
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Remove a node from a queue
|
|
|
|
*/
|
2018-05-15 09:49:22 +00:00
|
|
|
static __rte_always_inline void *
|
2017-04-21 10:43:26 +00:00
|
|
|
_lthread_queue_remove(struct lthread_queue *q)
|
|
|
|
{
|
|
|
|
void *data = NULL;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* There is a small window of inconsistency between producer and
|
|
|
|
* consumer whereby the queue may appear empty if consumer and
|
|
|
|
* producer access it at the same time. We handle this by retrying
|
|
|
|
*/
|
|
|
|
do {
|
|
|
|
data = _lthread_queue_poll(q);
|
|
|
|
|
|
|
|
if (likely(data != NULL)) {
|
|
|
|
|
|
|
|
DIAG_COUNT_INC(q, rd);
|
|
|
|
DIAG_COUNT_DEC(q, size);
|
|
|
|
return data;
|
|
|
|
}
|
|
|
|
rte_compiler_barrier();
|
|
|
|
} while (unlikely(!_lthread_queue_empty(q)));
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2018-05-15 09:49:22 +00:00
|
|
|
#ifdef __cplusplus
|
|
|
|
}
|
|
|
|
#endif
|
2017-04-21 10:43:26 +00:00
|
|
|
|
|
|
|
#endif /* LTHREAD_QUEUE_H_ */
|