/* SPDX-License-Identifier: BSD-3-Clause * Copyright(c) 2019 Intel Corporation */ #ifndef _RTE_STACK_LF_C11_H_ #define _RTE_STACK_LF_C11_H_ #include #include static __rte_always_inline unsigned int __rte_stack_lf_count(struct rte_stack *s) { /* stack_lf_push() and stack_lf_pop() do not update the list's contents * and stack_lf->len atomically, which can cause the list to appear * shorter than it actually is if this function is called while other * threads are modifying the list. * * However, given the inherently approximate nature of the get_count * callback -- even if the list and its size were updated atomically, * the size could change between when get_count executes and when the * value is returned to the caller -- this is acceptable. * * The stack_lf->len updates are placed such that the list may appear to * have fewer elements than it does, but will never appear to have more * elements. If the mempool is near-empty to the point that this is a * concern, the user should consider increasing the mempool size. */ return (unsigned int)__atomic_load_n(&s->stack_lf.used.len, __ATOMIC_RELAXED); } static __rte_always_inline void __rte_stack_lf_push_elems(struct rte_stack_lf_list *list, struct rte_stack_lf_elem *first, struct rte_stack_lf_elem *last, unsigned int num) { struct rte_stack_lf_head old_head; int success; old_head = list->head; do { struct rte_stack_lf_head new_head; /* Use an acquire fence to establish a synchronized-with * relationship between the list->head load and store-release * operations (as part of the rte_atomic128_cmp_exchange()). */ __atomic_thread_fence(__ATOMIC_ACQUIRE); /* Swing the top pointer to the first element in the list and * make the last element point to the old top. */ new_head.top = first; new_head.cnt = old_head.cnt + 1; last->next = old_head.top; /* Use the release memmodel to ensure the writes to the LF LIFO * elements are visible before the head pointer write. */ success = rte_atomic128_cmp_exchange( (rte_int128_t *)&list->head, (rte_int128_t *)&old_head, (rte_int128_t *)&new_head, 1, __ATOMIC_RELEASE, __ATOMIC_RELAXED); } while (success == 0); /* Ensure the stack modifications are not reordered with respect * to the LIFO len update. */ __atomic_add_fetch(&list->len, num, __ATOMIC_RELEASE); } static __rte_always_inline struct rte_stack_lf_elem * __rte_stack_lf_pop_elems(struct rte_stack_lf_list *list, unsigned int num, void **obj_table, struct rte_stack_lf_elem **last) { struct rte_stack_lf_head old_head; uint64_t len; int success; /* Reserve num elements, if available */ len = __atomic_load_n(&list->len, __ATOMIC_ACQUIRE); while (1) { /* Does the list contain enough elements? */ if (unlikely(len < num)) return NULL; /* len is updated on failure */ if (__atomic_compare_exchange_n(&list->len, &len, len - num, 0, __ATOMIC_ACQUIRE, __ATOMIC_ACQUIRE)) break; } /* If a torn read occurs, the CAS will fail and set old_head to the * correct/latest value. */ old_head = list->head; /* Pop num elements */ do { struct rte_stack_lf_head new_head; struct rte_stack_lf_elem *tmp; unsigned int i; /* Use the acquire memmodel to ensure the reads to the LF LIFO * elements are properly ordered with respect to the head * pointer read. */ __atomic_thread_fence(__ATOMIC_ACQUIRE); rte_prefetch0(old_head.top); tmp = old_head.top; /* Traverse the list to find the new head. A next pointer will * either point to another element or NULL; if a thread * encounters a pointer that has already been popped, the CAS * will fail. */ for (i = 0; i < num && tmp != NULL; i++) { rte_prefetch0(tmp->next); if (obj_table) obj_table[i] = tmp->data; if (last) *last = tmp; tmp = tmp->next; } /* If NULL was encountered, the list was modified while * traversing it. Retry. */ if (i != num) continue; new_head.top = tmp; new_head.cnt = old_head.cnt + 1; success = rte_atomic128_cmp_exchange( (rte_int128_t *)&list->head, (rte_int128_t *)&old_head, (rte_int128_t *)&new_head, 1, __ATOMIC_RELEASE, __ATOMIC_RELAXED); } while (success == 0); return old_head.top; } #endif /* _RTE_STACK_LF_C11_H_ */