| /dpdk/examples/ipsec-secgw/ |
| H A D | esp.c | 149 uint8_t *nexthdr, *pad_len; in esp_inbound_post() local 182 pad_len = nexthdr - 1; in esp_inbound_post() 184 padding = pad_len - *pad_len; in esp_inbound_post() 185 for (i = 0; i < *pad_len; i++) { in esp_inbound_post() 361 padding[pad_len - 2] = pad_len - 2; in esp_outbound() 362 padding[pad_len - 1] = nlp; in esp_outbound() 379 for (i = 0; i < pad_len - 2; i++) in esp_outbound() 381 padding[pad_len - 2] = pad_len - 2; in esp_outbound() 382 padding[pad_len - 1] = nlp; in esp_outbound() 422 padding[pad_len - 2] = pad_len - 2; in esp_outbound() [all …]
|
| /dpdk/app/test/ |
| H A D | test_cryptodev_blockcipher.c | 99 uint32_t pad_len; in test_blockcipher_one_case() local 206 pad_len = RTE_ALIGN(buf_len, 16) - buf_len; in test_blockcipher_one_case() 208 buf_len += pad_len; in test_blockcipher_one_case() 243 buf_p = rte_pktmbuf_append(ibuf, pad_len); in test_blockcipher_one_case() 255 rte_memcpy(buf_p, temp_p, pad_len); in test_blockcipher_one_case() 257 memset(buf_p, 0xa5, pad_len); in test_blockcipher_one_case() 272 buf_p = rte_pktmbuf_append(obuf, buf_len + pad_len); in test_blockcipher_one_case() 476 sym_op->cipher.data.length += pad_len; in test_blockcipher_one_case() 716 digest_len + pad_len; in test_blockcipher_one_case()
|
| H A D | test_ipsec.c | 600 .pad_len = padlen, in setup_test_string_tunneled()
|
| /dpdk/lib/net/ |
| H A D | rte_esp.h | 32 uint8_t pad_len; /**< number of pad bytes (0-255) */ member
|
| /dpdk/drivers/net/txgbe/ |
| H A D | txgbe_ipsec.h | 81 uint8_t pad_len; member
|
| H A D | txgbe_ipsec.c | 471 mdata->pad_len = txgbe_crypto_compute_pad_len(m); in txgbe_crypto_update_mb()
|
| H A D | txgbe_rxtx.c | 423 type_tucmd_mlhl |= TXGBE_TXD_IPSEC_ESPLEN(md->pad_len); in txgbe_set_xmit_ctx() 798 tx_offload.sec_pad_len = ipsec_mdata->pad_len; in txgbe_xmit_pkts()
|
| /dpdk/drivers/net/ixgbe/ |
| H A D | ixgbe_ipsec.h | 97 uint8_t pad_len; member
|
| H A D | ixgbe_ipsec.c | 491 mdata->pad_len = ixgbe_crypto_compute_pad_len(m); in ixgbe_crypto_update_mb()
|
| H A D | ixgbe_rxtx.c | 465 (md->pad_len & IXGBE_ADVTXD_IPSEC_ESP_LEN_MASK); in ixgbe_set_xmit_ctx() 698 tx_offload.sec_pad_len = ipsec_mdata->pad_len; in ixgbe_xmit_pkts()
|
| /dpdk/lib/ipsec/ |
| H A D | esp_inb.c | 466 if (tofs[0] < espt.pad_len) { in trs_process_check() 470 tofs[0] -= espt.pad_len; in trs_process_check() 472 return check_pad_bytes(ml[0], tofs[0], espt.pad_len); in trs_process_check() 604 tl = tlen + espt[i].pad_len; in tun_process() 665 tl = tlen + espt[i].pad_len; in trs_process()
|
| H A D | esp_outb.c | 227 espt->pad_len = pdlen; in outb_tun_pkt_prepare() 424 espt->pad_len = pdlen; in outb_trs_pkt_prepare()
|
| /dpdk/examples/l2fwd-crypto/ |
| H A D | main.c | 401 uint32_t pad_len = 0; in l2fwd_simple_crypto_enqueue() local 436 pad_len = cparams->block_size - in l2fwd_simple_crypto_enqueue() 442 pad_len = cparams->cipher_dataunit_len - in l2fwd_simple_crypto_enqueue() 446 pad_len = 0; in l2fwd_simple_crypto_enqueue() 449 if (pad_len) { in l2fwd_simple_crypto_enqueue() 450 padding = rte_pktmbuf_append(m, pad_len); in l2fwd_simple_crypto_enqueue() 454 data_len += pad_len; in l2fwd_simple_crypto_enqueue() 455 memset(padding, 0, pad_len); in l2fwd_simple_crypto_enqueue() 628 uint32_t pad_len; in l2fwd_simple_forward() local 643 pad_len = m->pkt_len - rte_be_to_cpu_16(ip_hdr->total_length) - in l2fwd_simple_forward() [all …]
|
| /dpdk/drivers/common/cnxk/hw/ |
| H A D | cpt.h | 239 uint8_t pad_len : 3; member
|
| /dpdk/drivers/common/cnxk/ |
| H A D | roc_cpt_debug.c | 19 cpth->w0.pad_len, cpth->w0.num_frags, cpth->w0.pkt_out); in roc_cpt_parse_hdr_dump()
|
| /dpdk/drivers/crypto/qat/dev/ |
| H A D | qat_crypto_pmd_gens.h | 783 uint8_t pad_len = 0; in enqueue_one_aead_job_gen1() local 786 pad_len = ICP_QAT_HW_CCM_AAD_B0_LEN - in enqueue_one_aead_job_gen1() 793 memset(&aad_data[pad_idx], 0, pad_len); in enqueue_one_aead_job_gen1()
|
| /dpdk/drivers/net/iavf/ |
| H A D | iavf_ipsec_crypto.c | 1098 *esp_trailer_length += esp_trailer->pad_len; in iavf_ipsec_crypto_get_esp_trailer()
|