| /dpdk/lib/ip_frag/ |
| H A D | rte_ipv4_fragmentation.c | 24 const struct rte_ipv4_hdr *src, uint16_t header_len, in __fill_ipv4hdr_frag() argument 27 rte_memcpy(dst, src, header_len); in __fill_ipv4hdr_frag() 92 if (unlikely(pkt_in->data_len < header_len) || in rte_ipv4_fragment_packet() 93 unlikely(mtu_size < header_len)) in rte_ipv4_fragment_packet() 111 (uint16_t)(pkt_in->pkt_len - header_len))) in rte_ipv4_fragment_packet() 115 in_seg_data_pos = header_len; in rte_ipv4_fragment_packet() 133 out_pkt->data_len = header_len; in rte_ipv4_fragment_packet() 134 out_pkt->pkt_len = header_len; in rte_ipv4_fragment_packet() 185 __fill_ipv4hdr_frag(out_hdr, in_hdr, header_len, in rte_ipv4_fragment_packet() 190 out_pkt->pkt_len - header_len); in rte_ipv4_fragment_packet() [all …]
|
| /dpdk/drivers/net/sfc/ |
| H A D | sfc_tso.c | 56 sfc_tso_prepare_header(uint8_t *tsoh, size_t header_len, in sfc_tso_prepare_header() argument 61 size_t bytes_left = header_len; in sfc_tso_prepare_header() 105 size_t header_len = m->l2_len + m->l3_len + m->l4_len; in sfc_efx_tso_do() local 118 if (m->data_len < header_len) { in sfc_efx_tso_do() 125 if (unlikely(header_len > SFC_TSOH_STD_LEN)) in sfc_efx_tso_do() 129 sfc_tso_prepare_header(tsoh, header_len, in_seg, in_off); in sfc_efx_tso_do() 133 if (m->data_len == header_len) { in sfc_efx_tso_do() 137 *in_off = header_len; in sfc_efx_tso_do() 172 efx_tx_qdesc_dma_create(txq->common, header_paddr, header_len, in sfc_efx_tso_do() 175 *pkt_len -= header_len; in sfc_efx_tso_do()
|
| H A D | sfc_ef100_tx.c | 106 size_t header_len = ((m->ol_flags & RTE_MBUF_F_TX_TUNNEL_MASK) ? in sfc_ef100_tx_prepare_pkt_tso() local 109 size_t payload_len = m->pkt_len - header_len; in sfc_ef100_tx_prepare_pkt_tso() 143 if (unlikely((header_len > txq->tso_max_header_len) | in sfc_ef100_tx_prepare_pkt_tso() 147 (m->pkt_len == header_len))) in sfc_ef100_tx_prepare_pkt_tso() 452 size_t header_len, size_t payload_len, in sfc_ef100_tx_qdesc_tso_create() argument 479 ESF_GZ_TX_TSO_HDR_LEN_W, header_len >> 1, in sfc_ef100_tx_qdesc_tso_create() 599 size_t header_len; in sfc_ef100_xmit_tso_pkt() local 613 header_len = tcph_off + m_seg->l4_len; in sfc_ef100_xmit_tso_pkt() 623 remaining_hdr_len = header_len; in sfc_ef100_xmit_tso_pkt() 665 sfc_ef100_tx_qdesc_tso_create(*m, nb_hdr_descs, nb_pld_descs, header_len, in sfc_ef100_xmit_tso_pkt() [all …]
|
| H A D | sfc_dp_tx.h | 252 unsigned int header_len = tcph_off + m->l4_len; in sfc_dp_tx_prepare_pkt() local 253 unsigned int header_len_remaining = header_len; in sfc_dp_tx_prepare_pkt() 328 unlikely(header_len > tso_bounce_buffer_len)) { in sfc_dp_tx_prepare_pkt()
|
| H A D | sfc_ef10_tx.c | 378 size_t header_len = tcph_off + m_seg->l4_len; in sfc_ef10_xmit_tso_pkt() local 425 if (rte_pktmbuf_data_len(m_seg) >= header_len) { in sfc_ef10_xmit_tso_pkt() 428 if (rte_pktmbuf_data_len(m_seg) == header_len) { in sfc_ef10_xmit_tso_pkt() 448 in_off = header_len; in sfc_ef10_xmit_tso_pkt() 461 if (unlikely(header_len > SFC_TSOH_STD_LEN)) in sfc_ef10_xmit_tso_pkt() 466 copied_segs = sfc_tso_prepare_header(hdr_addr, header_len, in sfc_ef10_xmit_tso_pkt() 524 sfc_ef10_tx_qdesc_dma_create(hdr_iova, header_len, false, in sfc_ef10_xmit_tso_pkt()
|
| H A D | sfc_tso.h | 73 unsigned int sfc_tso_prepare_header(uint8_t *tsoh, size_t header_len,
|
| /dpdk/drivers/net/ena/base/ |
| H A D | ena_eth_com.c | 92 u16 header_len) in ena_com_write_header_to_bounce() argument 105 if (unlikely((header_offset + header_len) > llq_info->desc_list_entry_size)) { in ena_com_write_header_to_bounce() 117 memcpy(bounce_buffer + header_offset, header_src, header_len); in ena_com_write_header_to_bounce() 393 u16 header_len = ena_tx_ctx->header_len; in ena_com_prepare_tx() local 410 if (unlikely(header_len > io_sq->tx_max_header_size)) { in ena_com_prepare_tx() 413 header_len, io_sq->tx_max_header_size); in ena_com_prepare_tx() 424 rc = ena_com_write_header_to_bounce(io_sq, buffer_to_push, header_len); in ena_com_prepare_tx() 436 if (unlikely(!num_bufs && !header_len)) { in ena_com_prepare_tx() 454 desc->buff_addr_hi_hdr_sz |= ((u32)header_len << in ena_com_prepare_tx()
|
| H A D | ena_eth_com.h | 30 u16 header_len; member
|
| /dpdk/drivers/net/enic/ |
| H A D | enic_rxtx.c | 435 uint16_t header_len; in enic_prep_pkts() local 437 header_len = m->l2_len + m->l3_len + m->l4_len; in enic_prep_pkts() 438 if (m->tso_segsz + header_len > ENIC_TX_MAX_PKT_SIZE) { in enic_prep_pkts() 486 uint16_t header_len; in enic_xmit_pkts() local 531 header_len = 0; in enic_xmit_pkts() 534 header_len = tx_pkt->l2_len + tx_pkt->l3_len + in enic_xmit_pkts() 538 if (unlikely(header_len == 0 || ((tx_pkt->tso_segsz + in enic_xmit_pkts() 539 header_len) > ENIC_TX_MAX_PKT_SIZE))) { in enic_xmit_pkts() 549 header_len += tx_pkt->outer_l2_len + in enic_xmit_pkts() 554 if ((ol_flags & ol_flags_mask) && (header_len == 0)) { in enic_xmit_pkts() [all …]
|
| /dpdk/drivers/net/qede/base/ |
| H A D | eth_common.h | 432 u8 header_len /* Packet L2+L3+L4 header length */; member
|
| /dpdk/drivers/net/ena/ |
| H A D | ena_ethdev.c | 175 uint16_t *header_len); 2863 uint16_t *header_len) in ena_tx_map_mbuf() argument 2882 *header_len = push_len; in ena_tx_map_mbuf() 2900 *header_len = 0; in ena_tx_map_mbuf() 2937 uint16_t header_len; in ena_xmit_mbuf() local 2959 ena_tx_map_mbuf(tx_ring, tx_info, mbuf, &push_header, &header_len); in ena_xmit_mbuf() 2965 ena_tx_ctx.header_len = header_len; in ena_xmit_mbuf()
|
| /dpdk/drivers/net/qede/ |
| H A D | qede_rxtx.c | 1829 cqe_start_tpa->header_len, in qede_recv_pkts()
|