| /f-stack/app/redis-5.0.5/deps/jemalloc/test/include/test/ |
| H A D | mq.h | 43 ql_new(&mq->msgs); \ 44 mq->count = 0; \ 49 mtx_fini(&mq->lock); \ 55 mtx_lock(&mq->lock); \ 56 count = mq->count; \ 57 mtx_unlock(&mq->lock); \ 64 mtx_lock(&mq->lock); \ 68 mq->count--; \ 70 mtx_unlock(&mq->lock); \ 102 mtx_lock(&mq->lock); \ [all …]
|
| /f-stack/app/redis-5.0.5/deps/jemalloc/test/unit/ |
| H A D | mq.c | 13 mq_t mq; in mq_gen() local 18 assert_ptr_null(mq_tryget(&mq), in mq_gen() 21 mq_put(&mq, &msg); in mq_gen() 25 mq_put(&mq, &msg); in mq_gen() 28 mq_fini(&mq); in mq_gen() 34 mq_t *mq = (mq_t *)arg; in thd_receiver_start() local 38 mq_msg_t *msg = mq_get(mq); in thd_receiver_start() 47 mq_t *mq = (mq_t *)arg; in thd_sender_start() local 56 mq_put(mq, msg); in thd_sender_start() 62 mq_t mq; in TEST_BEGIN() local [all …]
|
| /f-stack/dpdk/drivers/net/memif/ |
| H A D | rte_eth_memif.c | 267 rte_mbuf_refcnt_update(mq->buffers[mq->last_tail & mask], -1); in memif_free_stored_mbufs() 268 rte_pktmbuf_free_seg(mq->buffers[mq->last_tail & mask]); in memif_free_stored_mbufs() 269 mq->last_tail++; in memif_free_stored_mbufs() 1085 mq->region = 0; in memif_init_queues() 1108 mq->region = 0; in memif_init_queues() 1307 mq->type = in memif_tx_queue_setup() 1309 mq->n_pkts = 0; in memif_tx_queue_setup() 1310 mq->n_bytes = 0; in memif_tx_queue_setup() 1337 mq->n_pkts = 0; in memif_rx_queue_setup() 1353 if (!mq) in memif_queue_release() [all …]
|
| H A D | memif_socket.c | 295 struct memif_queue *mq; in memif_msg_receive_add_ring() local 320 mq->intr_handle.fd = fd; in memif_msg_receive_add_ring() 322 mq->region = ar->region; in memif_msg_receive_add_ring() 323 mq->ring_offset = ar->offset; in memif_msg_receive_add_ring() 445 struct memif_queue *mq; in memif_msg_enq_add_ring() local 456 e->fd = mq->intr_handle.fd; in memif_msg_enq_add_ring() 459 ar->region = mq->region; in memif_msg_enq_add_ring() 523 struct memif_queue *mq; in memif_disconnect() local 591 close(mq->intr_handle.fd); in memif_disconnect() 592 mq->intr_handle.fd = -1; in memif_disconnect() [all …]
|
| /f-stack/freebsd/kern/ |
| H A D | uipc_mqueue.c | 1613 knlist_init_mtx(&mq->mq_rsel.si_note, &mq->mq_mutex); in mqueue_alloc() 1614 knlist_init_mtx(&mq->mq_wsel.si_note, &mq->mq_mutex); in mqueue_alloc() 1616 return (mq); in mqueue_alloc() 1769 while (mq->mq_curmsgs >= mq->mq_maxmsg && error == 0) { in _mqueue_send() 1775 error = msleep(&mq->mq_senders, &mq->mq_mutex, in _mqueue_send() 1781 if (mq->mq_curmsgs >= mq->mq_maxmsg) { in _mqueue_send() 1931 error = msleep(&mq->mq_receivers, &mq->mq_mutex, in _mqueue_recv() 1950 if (mq->mq_notifier != NULL && mq->mq_receivers == 0 && in _mqueue_recv() 2028 mq = NULL; in kern_kmq_open() 2517 if (mq->mq_curmsgs < mq->mq_maxmsg) in mqf_poll() [all …]
|
| /f-stack/tools/compat/include/sys/ |
| H A D | mbuf.h | 1226 STAILQ_INIT(&mq->mq_head); in mbufq_init() 1227 mq->mq_maxlen = maxlen; in mbufq_init() 1228 mq->mq_len = 0; in mbufq_init() 1238 mq->mq_len = 0; in mbufq_flush() 1247 n = mbufq_flush(mq); in mbufq_drain() 1272 return (mq->mq_len >= mq->mq_maxlen); in mbufq_full() 1279 return (mq->mq_len); in mbufq_len() 1286 if (mbufq_full(mq)) in mbufq_enqueue() 1289 mq->mq_len++; in mbufq_enqueue() 1302 mq->mq_len--; in mbufq_dequeue() [all …]
|
| /f-stack/freebsd/netpfil/ipfw/ |
| H A D | dn_sched_prio.c | 133 if (q->mq.head == NULL) { in prio_dequeue() 171 if (q->mq.head == NULL) in prio_new_queue() 186 oldq->mq.tail->m_nextpkt = q->mq.head; in prio_new_queue() 187 oldq->mq.tail = q->mq.tail; in prio_new_queue() 192 q->mq.tail = q->mq.head = NULL; in prio_new_queue()
|
| H A D | dn_sched_fq_codel.h | 62 struct mq mq; /* list of packets */ member 141 struct mbuf *m = q->mq.head; in fq_codel_extract_head() 145 q->mq.head = m->m_nextpkt; in fq_codel_extract_head()
|
| H A D | ip_dn_private.h | 90 struct mq { /* a basic queue of packets*/ struct 166 struct mq pending; 196 struct mq mq; member 260 struct mq mq; /* packets queue */ member 457 mq_append(struct mq *q, struct mbuf *m) in mq_append()
|
| H A D | dn_sched_wf2q.c | 144 if (m != q->mq.head) { in wf2qp_enqueue() 147 if (m != q->mq.head) /* queue was already busy */ in wf2qp_enqueue() 249 if (q->mq.head == 0) { /* not backlogged any more. */ in wf2qp_dequeue() 253 uint64_t len = q->mq.head->m_pkthdr.len; in wf2qp_dequeue() 312 if (_q->mq.head != NULL) { in wf2qp_new_queue() 313 wf2qp_enqueue(_q->_si, _q, _q->mq.head); in wf2qp_new_queue() 339 if (q->mq.head == NULL) { in wf2qp_free_queue()
|
| H A D | dn_sched.h | 171 struct mbuf *m = q->mq.head; in dn_dequeue() 179 q->mq.head = m->m_nextpkt; in dn_dequeue() 180 q->mq.count--; in dn_dequeue()
|
| H A D | ip_dn_io.c | 271 mq_append(struct mq *q, struct mbuf *m) in mq_append() 535 mq_append(&q->mq, m); in dn_enqueue() 556 transmit_event(struct mq *q, struct delay_line *dline, uint64_t now) in transmit_event() 562 while ((m = dline->mq.head) != NULL) { in transmit_event() 566 dline->mq.head = m->m_nextpkt; in transmit_event() 567 dline->mq.count--; in transmit_event() 605 serve_sched(struct mq *q, struct dn_sch_inst *si, uint64_t now) in serve_sched() 607 struct mq def_q; in serve_sched() 610 int delay_line_idle = (si->dline.mq.head == NULL); in serve_sched() 636 mq_append(&si->dline.mq, m); in serve_sched() [all …]
|
| H A D | dn_sched_rr.c | 157 if (m != q->mq.head) { in rr_enqueue() 160 if (m != q->mq.head) in rr_enqueue() 186 struct mbuf *m = rrq->q.mq.head; in rr_dequeue() 274 if (_q->mq.head != NULL) { in rr_new_queue()
|
| H A D | dn_aqm_codel.c | 195 struct mbuf *m = q->mq.head; in codel_extract_head() 199 q->mq.head = m->m_nextpkt; in codel_extract_head() 267 mq_append(&q->mq, m); in aqm_codel_enqueue()
|
| H A D | dn_sched_fq_pie.c | 108 struct mq mq; /* list of packets */ member 341 struct mbuf *m = q->mq.head; in fq_pie_extract_head() 345 q->mq.head = m->m_nextpkt; in fq_pie_extract_head() 745 mq_append(&q->mq, m); in pie_enqueue() 762 struct mbuf *m = q->mq.head; in pie_drop_head() 766 q->mq.head = m->m_nextpkt; in pie_drop_head()
|
| H A D | dn_sched_fifo.c | 98 dn_free_pkts(q->mq.head); in fifo_free_sched()
|
| /f-stack/freebsd/sys/ |
| H A D | mbuf.h | 1489 STAILQ_INIT(&mq->mq_head); in mbufq_init() 1490 mq->mq_maxlen = maxlen; in mbufq_init() 1491 mq->mq_len = 0; in mbufq_init() 1501 mq->mq_len = 0; in mbufq_flush() 1510 n = mbufq_flush(mq); in mbufq_drain() 1535 return (mq->mq_maxlen > 0 && mq->mq_len >= mq->mq_maxlen); in mbufq_full() 1542 return (mq->mq_len); in mbufq_len() 1549 if (mbufq_full(mq)) in mbufq_enqueue() 1552 mq->mq_len++; in mbufq_enqueue() 1565 mq->mq_len--; in mbufq_dequeue() [all …]
|
| /f-stack/freebsd/netpfil/ipfw/test/ |
| H A D | test_dn_sched.c | 40 if (q->mq.head) in dn_delete_queue() 41 dn_free_pkts(q->mq.head); in dn_delete_queue() 61 mq_append(&q->mq, m); in dn_enqueue()
|
| H A D | main.c | 142 struct mq *mq = (struct mq *)si; in default_enqueue() local 146 if (mq->head == NULL) in default_enqueue() 147 mq->head = m; in default_enqueue() 149 mq->tail->m_nextpkt = m; in default_enqueue() 150 mq->tail = m; in default_enqueue() 157 struct mq *mq = (struct mq *)si; in default_dequeue() local 160 if ((m = mq->head)) { in default_dequeue() 161 m = mq->head; in default_dequeue() 162 mq->head = m->m_nextpkt; in default_dequeue()
|
| /f-stack/dpdk/doc/guides/vdpadevs/features/ |
| H A D | mlx5.ini | 16 mq = Y key 18 proto mq = Y
|
| H A D | default.ini | 25 mq = key 34 proto mq =
|
| /f-stack/freebsd/contrib/device-tree/Bindings/powerpc/4xx/ |
| H A D | ppc440spe-adma.txt | 82 - compatible : "ibm,mq-440spe"; 87 MQ0: mq { 88 compatible = "ibm,mq-440spe";
|
| /f-stack/freebsd/netinet6/ |
| H A D | mld6.c | 433 while ((m = mbufq_dequeue(mq)) != NULL) { in mld_dispatch_queue() 1978 struct mbufq *mq; in mld_initial_join() local 2068 mq = &inm->in6m_scq; in mld_initial_join() 2069 mbufq_drain(mq); in mld_initial_join() 2459 m0 = mbufq_last(mq); in mld_v2_enqueue_group_record() 2471 if (mbufq_full(mq)) { in mld_v2_enqueue_group_record() 2584 mbufq_enqueue(mq, m); in mld_v2_enqueue_group_record() 2600 if (mbufq_full(mq)) { in mld_v2_enqueue_group_record() 2659 mbufq_enqueue(mq, m); in mld_v2_enqueue_group_record() 2745 m0 = mbufq_last(mq); in mld_v2_enqueue_filter_change() [all …]
|
| /f-stack/freebsd/netinet/ |
| H A D | igmp.c | 555 while ((m = mbufq_dequeue(mq)) != NULL) { in igmp_dispatch_queue() 2385 struct mbufq *mq; in igmp_initial_join() local 2459 mq = &inm->inm_scq; in igmp_initial_join() 2460 mbufq_drain(mq); in igmp_initial_join() 2806 m0 = mbufq_last(mq); in igmp_v3_enqueue_group_record() 2817 if (mbufq_full(mq)) { in igmp_v3_enqueue_group_record() 2930 mbufq_enqueue(mq, m); in igmp_v3_enqueue_group_record() 2946 if (mbufq_full(mq)) { in igmp_v3_enqueue_group_record() 3009 mbufq_enqueue(mq, m); in igmp_v3_enqueue_group_record() 3093 m0 = mbufq_last(mq); in igmp_v3_enqueue_filter_change() [all …]
|
| /f-stack/dpdk/doc/guides/vdpadevs/ |
| H A D | features_overview.rst | 54 mq 81 proto mq
|