Searched refs:rx_nb_avail (Results 1 – 11 of 11) sorted by relevance
189 rxq->rx_nb_avail = 0; in reset_rx_queue()292 if (rxq->rx_nb_avail == 0) in release_rxq_mbufs()294 for (i = 0; i < rxq->rx_nb_avail; i++) { in release_rxq_mbufs()300 rxq->rx_nb_avail = 0; in release_rxq_mbufs()1791 nb_pkts = (uint16_t)RTE_MIN(nb_pkts, rxq->rx_nb_avail); in iavf_rx_fill_from_stage()1796 rxq->rx_nb_avail = (uint16_t)(rxq->rx_nb_avail - nb_pkts); in iavf_rx_fill_from_stage()1861 if (rxq->rx_nb_avail) in rx_recv_pkts()1869 rxq->rx_nb_avail = nb_rx; in rx_recv_pkts()1878 rxq->rx_nb_avail = 0; in rx_recv_pkts()1894 if (rxq->rx_nb_avail) in rx_recv_pkts()
176 uint16_t rx_nb_avail; /* number of staged packets ready */ member
111 uint16_t rx_nb_avail; /**< nr of staged pkts ready to ret to app */ member
1638 nb_pkts = (uint16_t)RTE_MIN(nb_pkts, rxq->rx_nb_avail); in ixgbe_rx_fill_from_stage()1645 rxq->rx_nb_avail = (uint16_t)(rxq->rx_nb_avail - nb_pkts); in ixgbe_rx_fill_from_stage()1659 if (rxq->rx_nb_avail) in rx_recv_pkts()1667 rxq->rx_nb_avail = nb_rx; in rx_recv_pkts()1688 rxq->rx_nb_avail = 0; in rx_recv_pkts()1706 if (rxq->rx_nb_avail) in rx_recv_pkts()2816 if (rxq->rx_nb_avail) { in ixgbe_rx_queue_release_mbufs()2817 for (i = 0; i < rxq->rx_nb_avail; ++i) { in ixgbe_rx_queue_release_mbufs()2823 rxq->rx_nb_avail = 0; in ixgbe_rx_queue_release_mbufs()2931 rxq->rx_nb_avail = 0; in ixgbe_reset_rx_queue()
511 nb_pkts = (uint16_t)RTE_MIN(nb_pkts, rxq->rx_nb_avail); in i40e_rx_fill_from_stage()516 rxq->rx_nb_avail = (uint16_t)(rxq->rx_nb_avail - nb_pkts); in i40e_rx_fill_from_stage()582 if (rxq->rx_nb_avail) in rx_recv_pkts()587 rxq->rx_nb_avail = nb_rx; in rx_recv_pkts()598 rxq->rx_nb_avail = 0; in rx_recv_pkts()610 if (rxq->rx_nb_avail) in rx_recv_pkts()2439 if (rxq->rx_nb_avail == 0) in i40e_rx_queue_release_mbufs()2441 for (i = 0; i < rxq->rx_nb_avail; i++) { in i40e_rx_queue_release_mbufs()2447 rxq->rx_nb_avail = 0; in i40e_rx_queue_release_mbufs()2477 rxq->rx_nb_avail = 0; in i40e_reset_rx_queue()
97 uint16_t rx_nb_avail; /**< number of staged packets ready */ member
408 if (rxq->rx_nb_avail == 0) in _ice_rx_queue_release_mbufs()410 for (i = 0; i < rxq->rx_nb_avail; i++) in _ice_rx_queue_release_mbufs()413 rxq->rx_nb_avail = 0; in _ice_rx_queue_release_mbufs()518 rxq->rx_nb_avail = 0; in ice_reset_rx_queue()1577 nb_pkts = (uint16_t)RTE_MIN(nb_pkts, rxq->rx_nb_avail); in ice_rx_fill_from_stage()1582 rxq->rx_nb_avail = (uint16_t)(rxq->rx_nb_avail - nb_pkts); in ice_rx_fill_from_stage()1647 if (rxq->rx_nb_avail) in rx_recv_pkts()1652 rxq->rx_nb_avail = nb_rx; in rx_recv_pkts()1665 rxq->rx_nb_avail = 0; in rx_recv_pkts()1677 if (rxq->rx_nb_avail) in rx_recv_pkts()
64 uint16_t rx_nb_avail; /**< number of staged packets ready */ member
326 rxq->rx_nb_avail = 0; in reset_rx_queue()
1259 nb_pkts = (uint16_t)RTE_MIN(nb_pkts, rxq->rx_nb_avail); in txgbe_rx_fill_from_stage()1266 rxq->rx_nb_avail = (uint16_t)(rxq->rx_nb_avail - nb_pkts); in txgbe_rx_fill_from_stage()1281 if (rxq->rx_nb_avail) in txgbe_rx_recv_pkts()1289 rxq->rx_nb_avail = nb_rx; in txgbe_rx_recv_pkts()1310 rxq->rx_nb_avail = 0; in txgbe_rx_recv_pkts()1327 if (rxq->rx_nb_avail) in txgbe_rx_recv_pkts()2340 if (rxq->rx_nb_avail) { in txgbe_rx_queue_release_mbufs()2341 for (i = 0; i < rxq->rx_nb_avail; ++i) { in txgbe_rx_queue_release_mbufs()2347 rxq->rx_nb_avail = 0; in txgbe_rx_queue_release_mbufs()2454 rxq->rx_nb_avail = 0; in txgbe_reset_rx_queue()
293 uint16_t rx_nb_avail; /**< nr of staged pkts ready to ret to app */ member