| /dpdk/drivers/crypto/dpaa_sec/ |
| H A D | dpaa_sec_raw_dp.c | 176 sg++; in build_dpaa_raw_dp_auth_fd() 206 sg++; in build_dpaa_raw_dp_auth_fd() 275 sg = &cf->sg[2]; in build_raw_cipher_auth_gcm_sg() 330 sg++; in build_raw_cipher_auth_gcm_sg() 348 sg++; in build_raw_cipher_auth_gcm_sg() 436 sg = &cf->sg[2]; in build_dpaa_raw_dp_chain_fd() 490 sg++; in build_dpaa_raw_dp_chain_fd() 583 sg = &cf->sg[2]; in build_dpaa_raw_dp_cipher_fd() 597 sg = &cf->sg[2]; in build_dpaa_raw_dp_cipher_fd() 691 sg = &cf->sg[2]; in build_dpaa_raw_proto_sg() [all …]
|
| H A D | dpaa_sec.c | 1045 sg = &cf->sg[0]; in build_auth_only() 1057 sg = &cf->sg[2]; in build_auth_only() 1159 sg = &cf->sg[2]; in build_cipher_only_sg() 1254 sg = &cf->sg[0]; in build_cipher_only() 1260 sg = &cf->sg[1]; in build_cipher_only() 1269 sg = &cf->sg[2]; in build_cipher_only() 1330 sg = &cf->sg[2]; in build_cipher_auth_gcm_sg() 1448 sg = &cf->sg[2]; in build_cipher_auth_gcm() 1576 sg = &cf->sg[2]; in build_cipher_auth_sg() 1685 sg = &cf->sg[2]; in build_cipher_auth() [all …]
|
| H A D | dpaa_sec.h | 237 struct qm_sg_entry sg[MAX_JOB_SG_ENTRIES]; member
|
| /dpdk/drivers/crypto/caam_jr/ |
| H A D | caam_jr.c | 754 sg = &ctx->sg[0]; in build_auth_only_sg() 830 sg = &ctx->sg[0]; in build_auth_only() 914 sg = &ctx->sg[0]; in build_cipher_only_sg() 939 sg++; in build_cipher_only_sg() 950 sg++; in build_cipher_only_sg() 1020 sg = &ctx->sg[0]; in build_cipher_only() 1029 sg = &ctx->sg[1]; in build_cipher_only() 1110 sg = &ctx->sg[0]; in build_cipher_auth_sg() 1141 sg++; in build_cipher_auth_sg() 1234 sg = &ctx->sg[0]; in build_cipher_auth() [all …]
|
| H A D | caam_jr_pvt.h | 183 struct sec4_sg_entry sg[MAX_SG_ENTRIES]; member
|
| /dpdk/drivers/net/octeontx_ep/ |
| H A D | otx_ep_rxtx.c | 383 rte_free(finfo->g.sg); in otx_ep_iqreq_delete() 577 finfo->g.sg = in otx_ep_xmit_pkts() 579 if (finfo->g.sg == NULL) { in otx_ep_xmit_pkts() 596 finfo->g.sg[(j >> 2)].ptr[(j & 3)] = in otx_ep_xmit_pkts() 598 set_sg_size(&finfo->g.sg[(j >> 2)], in otx_ep_xmit_pkts() 608 rte_free(finfo->g.sg); in otx_ep_xmit_pkts() 711 finfo->g.sg = in otx2_ep_xmit_pkts() 713 if (finfo->g.sg == NULL) { in otx2_ep_xmit_pkts() 730 finfo->g.sg[(j >> 2)].ptr[(j & 3)] = in otx2_ep_xmit_pkts() 732 set_sg_size(&finfo->g.sg[(j >> 2)], in otx2_ep_xmit_pkts() [all …]
|
| H A D | otx_ep_common.h | 488 struct otx_ep_sg_entry *sg; member
|
| /dpdk/drivers/net/cnxk/ |
| H A D | cn10k_tx.h | 876 sg_u = sg->u; in cn10k_nix_prepare_mseg() 916 sg->u = sg_u; in cn10k_nix_prepare_mseg() 917 sg->segs = 3; in cn10k_nix_prepare_mseg() 919 sg_u = sg->u; in cn10k_nix_prepare_mseg() 926 sg->u = sg_u; in cn10k_nix_prepare_mseg() 927 sg->segs = i; in cn10k_nix_prepare_mseg() 1276 sg_u = sg->u; in cn10k_nix_prepare_mseg_vec_list() 1317 sg->u = sg_u; in cn10k_nix_prepare_mseg_vec_list() 1318 sg->segs = 3; in cn10k_nix_prepare_mseg_vec_list() 1326 sg->u = sg_u; in cn10k_nix_prepare_mseg_vec_list() [all …]
|
| H A D | cn9k_tx.h | 425 sg_u = sg->u; in cn9k_nix_prepare_mseg() 470 sg->u = sg_u; in cn9k_nix_prepare_mseg() 471 sg->segs = 3; in cn9k_nix_prepare_mseg() 473 sg_u = sg->u; in cn9k_nix_prepare_mseg() 480 sg->u = sg_u; in cn9k_nix_prepare_mseg() 481 sg->segs = i; in cn9k_nix_prepare_mseg() 674 sg_u = sg->u; in cn9k_nix_prepare_mseg_vec_list() 715 sg->u = sg_u; in cn9k_nix_prepare_mseg_vec_list() 716 sg->segs = 3; in cn9k_nix_prepare_mseg_vec_list() 724 sg->u = sg_u; in cn9k_nix_prepare_mseg_vec_list() [all …]
|
| H A D | cn9k_rx.h | 125 uint64_t sg; in nix_cqe_xtract_mseg() local 127 sg = *(const uint64_t *)(rx + 1); in nix_cqe_xtract_mseg() 128 nb_segs = (sg >> 48) & 0x3; in nix_cqe_xtract_mseg() 137 mbuf->data_len = (sg & 0xFFFF) - (flags & NIX_RX_OFFLOAD_TSTAMP_F ? in nix_cqe_xtract_mseg() 140 sg = sg >> 16; in nix_cqe_xtract_mseg() 157 mbuf->data_len = sg & 0xFFFF; in nix_cqe_xtract_mseg() 158 sg = sg >> 16; in nix_cqe_xtract_mseg() 164 sg = *(const uint64_t *)(iova_list); in nix_cqe_xtract_mseg() 165 nb_segs = (sg >> 48) & 0x3; in nix_cqe_xtract_mseg()
|
| H A D | cn10k_rx.h | 688 uint64_t sg; in nix_cqe_xtract_mseg() local 690 sg = *(const uint64_t *)(rx + 1); in nix_cqe_xtract_mseg() 691 nb_segs = (sg >> 48) & 0x3; in nix_cqe_xtract_mseg() 700 mbuf->data_len = (sg & 0xFFFF) - (flags & NIX_RX_OFFLOAD_TSTAMP_F ? in nix_cqe_xtract_mseg() 703 sg = sg >> 16; in nix_cqe_xtract_mseg() 719 mbuf->data_len = sg & 0xFFFF; in nix_cqe_xtract_mseg() 720 sg = sg >> 16; in nix_cqe_xtract_mseg() 726 sg = *(const uint64_t *)(iova_list); in nix_cqe_xtract_mseg() 727 nb_segs = (sg >> 48) & 0x3; in nix_cqe_xtract_mseg()
|
| /dpdk/drivers/bus/fslmc/portal/ |
| H A D | dpaa2_hw_pvt.h | 348 #define DPAA2_SG_SET_FORMAT(sg, format) do { \ argument 349 (sg)->fin_bpid_offset &= 0xCFFFFFFF; \ 350 (sg)->fin_bpid_offset |= (uint32_t)format << 28; \ 353 #define DPAA2_SG_SET_FINAL(sg, fin) do { \ argument 354 (sg)->fin_bpid_offset &= 0x7FFFFFFF; \ 355 (sg)->fin_bpid_offset |= (uint32_t)fin << 31; \ 357 #define DPAA2_SG_IS_FINAL(sg) (!!((sg)->fin_bpid_offset >> 31)) argument
|
| /dpdk/drivers/net/liquidio/ |
| H A D | lio_rxtx.c | 1425 if (g->sg) in lio_delete_sglist() 1463 if (g->sg == NULL) { in lio_setup_sglists() 1472 if (((unsigned long)g->sg) & 7) { in lio_setup_sglists() 1473 g->adjust = 8 - (((unsigned long)g->sg) & 7); in lio_setup_sglists() 1474 g->sg = in lio_setup_sglists() 1721 memset(g->sg, 0, g->sg_size); in lio_dev_xmit_pkts() 1722 g->sg[0].ptr[0] = rte_mbuf_data_iova(m); in lio_dev_xmit_pkts() 1723 lio_add_sg_size(&g->sg[0], m->data_len, 0); in lio_dev_xmit_pkts() 1732 g->sg[(i >> 2)].ptr[(i & 3)] = in lio_dev_xmit_pkts() 1734 lio_add_sg_size(&g->sg[(i >> 2)], in lio_dev_xmit_pkts() [all …]
|
| H A D | lio_struct.h | 377 struct lio_sg_entry *sg; member
|
| /dpdk/drivers/net/netvsc/ |
| H A D | hn_nvs.h | 232 struct vmbus_gpa sg[], unsigned int sglen, in hn_nvs_send_sglist() argument 236 return rte_vmbus_chan_send_sglist(chan, sg, sglen, nvs_msg, nvs_msglen, in hn_nvs_send_sglist()
|
| H A D | hn_rxtx.c | 1402 static unsigned int hn_fill_sg(struct vmbus_gpa *sg, in hn_fill_sg() argument 1417 sg[segs].page = page; in hn_fill_sg() 1418 sg[segs].ofs = offset; in hn_fill_sg() 1419 sg[segs].len = bytes; in hn_fill_sg() 1437 struct vmbus_gpa sg[hn_get_slots(m)]; in hn_xmit_sg() local 1458 sg[0].page = addr / rte_mem_page_size(); in hn_xmit_sg() 1459 sg[0].ofs = addr & PAGE_MASK; in hn_xmit_sg() 1460 sg[0].len = RNDIS_PACKET_MSG_OFFSET_ABS(hn_rndis_pktlen(txd->rndis_pkt)); in hn_xmit_sg() 1465 segs += hn_fill_sg(sg + 1, m); in hn_xmit_sg() 1471 return hn_nvs_send_sglist(txq->chan, sg, segs, in hn_xmit_sg()
|
| H A D | hn_rndis.c | 259 struct vmbus_gpa sg; in hn_nvs_send_rndis_ctrl() local 274 sg.page = addr / rte_mem_page_size(); in hn_nvs_send_rndis_ctrl() 275 sg.ofs = addr & PAGE_MASK; in hn_nvs_send_rndis_ctrl() 276 sg.len = reqlen; in hn_nvs_send_rndis_ctrl() 278 if (sg.ofs + reqlen > rte_mem_page_size()) { in hn_nvs_send_rndis_ctrl() 285 return hn_nvs_send_sglist(chan, &sg, 1, in hn_nvs_send_rndis_ctrl()
|
| /dpdk/drivers/bus/dpaa/include/ |
| H A D | fsl_qman.h | 276 static inline u64 qm_sg_entry_get64(const struct qm_sg_entry *sg) in qm_sg_entry_get64() argument 278 return sg->addr; in qm_sg_entry_get64() 281 static inline dma_addr_t qm_sg_addr(const struct qm_sg_entry *sg) in qm_sg_addr() argument 283 return (dma_addr_t)sg->addr; in qm_sg_addr() 287 #define qm_sg_entry_set64(sg, v) \ argument 289 struct qm_sg_entry *__sg931 = (sg); \
|
| /dpdk/drivers/event/cnxk/ |
| H A D | cn9k_worker.h | 610 union nix_send_sg_s *sg; in cn9k_sso_hws_xmit_sec_one() local 619 sg = (union nix_send_sg_s *)&cmd[4]; in cn9k_sso_hws_xmit_sec_one() 621 sg = (union nix_send_sg_s *)&cmd[2]; in cn9k_sso_hws_xmit_sec_one() 629 dptr = *(uint64_t *)(sg + 1); in cn9k_sso_hws_xmit_sec_one() 641 sg->seg1_size = pkt_len + dlen_adj; in cn9k_sso_hws_xmit_sec_one()
|
| /dpdk/drivers/bus/vmbus/ |
| H A D | vmbus_channel.c | 143 struct vmbus_gpa sg[], uint32_t sglen, in rte_vmbus_chan_send_sglist() argument 168 iov[1].iov_base = sg; in rte_vmbus_chan_send_sglist()
|