| /dpdk/drivers/event/octeontx/ |
| H A D | ssovf_worker.c | 40 ssows_swtag_untag(ws); in ssows_fwd_swtag() 42 ws->swtag_req = 1; in ssows_fwd_swtag() 63 ssows_swtag_wait(ws); in ssows_fwd_group() 76 if (ws->cur_grp == grp) in ssows_forward_event() 91 ssows_swtag_untag(ws); in ssows_release_event() 154 struct ssows *ws = port; in ssows_enq() local 185 struct ssows *ws = port; in ssows_enq_new_burst() local 242 ws->lookup_mem); in ssows_flush_events() 265 ssows_desched(ws); in ssows_reset() 290 ssows_swtag_wait(ws); in __sso_event_tx_adapter_enqueue() [all …]
|
| H A D | ssovf_worker.h | 173 ssovf_load_pair(get_work0, get_work1, ws->getwork); in ssows_get_work() 176 ws->cur_tt = sched_type_queue & 0x3; in ssows_get_work() 177 ws->cur_grp = sched_type_queue >> 2; in ssows_get_work() 185 ws->lookup_mem); in ssows_get_work() 238 ssows_swtag_untag(struct ssows *ws) in ssows_swtag_untag() argument 241 ws->cur_tt = SSO_SYNC_UNTAGGED; in ssows_swtag_untag() 252 ssows_desched(struct ssows *ws) in ssows_desched() argument 254 ssovf_write64(0, ws->base + SSOW_VHWS_OP_DESCHED); in ssows_desched() 258 ssows_swtag_wait(struct ssows *ws) in ssows_swtag_wait() argument 261 while (ssovf_read64(ws->base + SSOW_VHWS_SWTP)) in ssows_swtag_wait() [all …]
|
| H A D | ssovf_evdev.c | 238 struct ssows *ws; in ssovf_port_setup() local 256 if (ws == NULL) { in ssovf_port_setup() 263 rte_free(ws); in ssovf_port_setup() 271 ws->getwork = ws->base + reg_off; in ssovf_port_setup() 272 ws->port = port_id; in ssovf_port_setup() 278 rte_free(ws); in ssovf_port_setup() 639 struct ssows *ws; in ssovf_start() local 646 ssows_reset(ws); in ssovf_start() 647 ws->swtag_req = 0; in ssovf_start() 677 struct ssows *ws; in ssovf_stop() local [all …]
|
| H A D | ssovf_evdev.h | 183 void ssows_flush_events(struct ssows *ws, uint8_t queue_id, 185 void ssows_reset(struct ssows *ws);
|
| /dpdk/drivers/event/cnxk/ |
| H A D | cn10k_worker.c | 12 struct cn10k_sso_hws *ws = port; in cn10k_sso_hws_enq() local 16 return cn10k_sso_hws_new_event(ws, ev); in cn10k_sso_hws_enq() 18 cn10k_sso_hws_forward_event(ws, ev); in cn10k_sso_hws_enq() 21 if (ws->swtag_req) { in cn10k_sso_hws_enq() 22 cnxk_sso_hws_desched(ev->u64, ws->base); in cn10k_sso_hws_enq() 23 ws->swtag_req = 0; in cn10k_sso_hws_enq() 26 cnxk_sso_hws_swtag_flush(ws->base); in cn10k_sso_hws_enq() 47 struct cn10k_sso_hws *ws = port; in cn10k_sso_hws_enq_new_burst() local 60 struct cn10k_sso_hws *ws = port; in cn10k_sso_hws_enq_fwd_burst() local 63 cn10k_sso_hws_forward_event(ws, ev); in cn10k_sso_hws_enq_fwd_burst() [all …]
|
| H A D | cn10k_worker.h | 30 if (ws->xaq_lmt <= *ws->fc_mem) in cn10k_sso_hws_new_event() 61 ws->swtag_req = 1; in cn10k_sso_hws_fwd_swtag() 248 ws->lookup_mem); in cn10k_sso_hws_post_process() 265 ws->tstamp, ws->lmt_base); in cn10k_sso_hws_post_process() 296 ws->gw_rdata = gw.u64[0]; in cn10k_sso_hws_get_work() 335 ws->gw_rdata = gw.u64[0]; in cn10k_sso_hws_get_work_empty() 565 ws->gw_rdata = roc_sso_hws_head_wait(ws->base); in cn10k_sso_tx_one() 630 lmt_addr = ws->lmt_base; in cn10k_sso_hws_event_tx() 653 rte_prefetch0(ws); in cn10k_sso_hws_event_tx() 681 ws, &ev[0], cmd, (const uint64_t *)ws->tx_adptr_data, \ [all …]
|
| H A D | cn10k_eventdev.c | 42 struct cn10k_sso_hws *ws; in cn10k_sso_init_hws_mem() local 48 if (ws == NULL) { in cn10k_sso_init_hws_mem() 54 ws = (struct cn10k_sso_hws *)((uint8_t *)ws + RTE_CACHE_LINE_SIZE); in cn10k_sso_init_hws_mem() 56 ws->hws_id = port_id; in cn10k_sso_init_hws_mem() 57 ws->swtag_req = 0; in cn10k_sso_init_hws_mem() 61 return ws; in cn10k_sso_init_hws_mem() 108 memset(ws, 0, sizeof(*ws)); in cn10k_sso_hws_release() 138 ws, &ev, in cn10k_sso_hws_flush_events() 186 ws->swtag_req) in cn10k_sso_hws_reset() 594 ws, &ev, in cn10k_sso_port_quiesce() [all …]
|
| H A D | cn9k_worker.c | 13 struct cn9k_sso_hws *ws = port; in cn9k_sso_hws_enq() local 17 return cn9k_sso_hws_new_event(ws, ev); in cn9k_sso_hws_enq() 19 cn9k_sso_hws_forward_event(ws, ev); in cn9k_sso_hws_enq() 22 if (ws->swtag_req) { in cn9k_sso_hws_enq() 23 cnxk_sso_hws_desched(ev->u64, ws->base); in cn9k_sso_hws_enq() 24 ws->swtag_req = 0; in cn9k_sso_hws_enq() 27 cnxk_sso_hws_swtag_flush(ws->base); in cn9k_sso_hws_enq() 48 struct cn9k_sso_hws *ws = port; in cn9k_sso_hws_enq_new_burst() local 61 struct cn9k_sso_hws *ws = port; in cn9k_sso_hws_enq_fwd_burst() local 64 cn9k_sso_hws_forward_event(ws, ev); in cn9k_sso_hws_enq_fwd_burst() [all …]
|
| H A D | cn9k_worker.h | 31 if (ws->xaq_lmt <= *ws->fc_mem) in cn9k_sso_hws_new_event() 35 ws->grp_base + (grp << 12)); in cn9k_sso_hws_new_event() 73 while (ws->xaq_lmt <= __atomic_load_n(ws->fc_mem, __ATOMIC_RELAXED)) in cn9k_sso_hws_new_event_wait() 88 ws->swtag_req = 1; in cn9k_sso_hws_forward_event() 96 roc_sso_hws_head_wait(ws->base); in cn9k_sso_hws_forward_event() 257 plt_write64(ws->gw_wdata, ws->base + SSOW_LF_GWS_OP_GET_WORK0); in cn9k_sso_hws_get_work() 289 ws->tstamp); in cn9k_sso_hws_get_work() 428 return cn9k_sso_hws_get_work(ws, ev, flags, ws->lookup_mem); \ 447 ret = cn9k_sso_hws_get_work(ws, ev, flags, ws->lookup_mem); \ 859 return cn9k_sso_hws_event_tx(ws->base[!ws->vws], &ev[0], cmd, \ [all …]
|
| H A D | cn9k_eventdev.c | 35 ws = port; in cn9k_sso_hws_link() 59 ws = port; in cn9k_sso_hws_unlink() 86 ws = hws; in cn9k_sso_hws_setup() 113 ws = hws; in cn9k_sso_hws_release() 116 memset(ws, 0, sizeof(*ws)); in cn9k_sso_hws_release() 154 ws = hws; in cn9k_sso_hws_flush_events() 202 ws = hws; in cn9k_sso_hws_reset() 239 ws->swtag_req = 0; in cn9k_sso_hws_reset() 723 ws = RTE_PTR_ADD(ws, sizeof(struct cnxk_sso_hws_cookie)); in cn9k_sso_init_hws_mem() 731 data = ws; in cn9k_sso_init_hws_mem() [all …]
|
| H A D | cnxk_eventdev.c | 186 void *ws; in cnxk_setup_event_ports() local 190 ws = event_dev->data->ports[i]; in cnxk_setup_event_ports() 192 ws = init_hws_fn(dev, i); in cnxk_setup_event_ports() 193 if (ws == NULL) in cnxk_setup_event_ports() 198 event_dev->data->ports[i] = ws; in cnxk_setup_event_ports() 471 void *ws; in cnxk_sso_cleanup() local 474 ws = event_dev->data->ports[i]; in cnxk_sso_cleanup() 475 reset_fn(dev, ws); in cnxk_sso_cleanup() 481 ws = event_dev->data->ports[0]; in cnxk_sso_cleanup() 548 void *ws; in cnxk_sso_close() local [all …]
|
| H A D | cnxk_eventdev.h | 73 typedef void (*cnxk_sso_hws_setup_t)(void *dev, void *ws, uintptr_t grp_base); 74 typedef void (*cnxk_sso_hws_release_t)(void *dev, void *ws); 75 typedef int (*cnxk_sso_link_t)(void *dev, void *ws, uint16_t *map, 77 typedef int (*cnxk_sso_unlink_t)(void *dev, void *ws, uint16_t *map, 80 typedef void (*cnxk_sso_hws_reset_t)(void *arg, void *ws); 81 typedef int (*cnxk_sso_hws_flush_t)(void *ws, uint8_t queue_id, uintptr_t base, 224 cnxk_sso_hws_get_cookie(void *ws) in cnxk_sso_hws_get_cookie() argument 226 return RTE_PTR_SUB(ws, sizeof(struct cnxk_sso_hws_cookie)); in cnxk_sso_hws_get_cookie()
|
| /dpdk/drivers/net/cnxk/ |
| H A D | cn10k_tx.h | 1019 if ((flags & NIX_TX_VWQE_F) && !(ws[1] & BIT_ULL(35))) in cn10k_nix_xmit_pkts() 1020 ws[1] = roc_sso_hws_head_wait(ws[0]); in cn10k_nix_xmit_pkts() 1072 cn10k_nix_xmit_pkts_mseg(void *tx_queue, uint64_t *ws, in cn10k_nix_xmit_pkts_mseg() argument 1163 if ((flags & NIX_TX_VWQE_F) && !(ws[1] & BIT_ULL(35))) in cn10k_nix_xmit_pkts_mseg() 1164 ws[1] = roc_sso_hws_head_wait(ws[0]); in cn10k_nix_xmit_pkts_mseg() 1586 cn10k_nix_xmit_pkts_vector(void *tx_queue, uint64_t *ws, in cn10k_nix_xmit_pkts_vector() argument 2712 if ((flags & NIX_TX_VWQE_F) && !(ws[1] & BIT_ULL(35))) in cn10k_nix_xmit_pkts_vector() 2713 ws[1] = roc_sso_hws_head_wait(ws[0]); in cn10k_nix_xmit_pkts_vector() 2781 pkts += cn10k_nix_xmit_pkts(tx_queue, ws, tx_pkts, in cn10k_nix_xmit_pkts_vector() 2790 cn10k_nix_xmit_pkts_vector(void *tx_queue, uint64_t *ws, in cn10k_nix_xmit_pkts_vector() argument [all …]
|
| /dpdk/drivers/net/mlx5/ |
| H A D | mlx5_tx.h | 816 struct mlx5_wqe_wseg *ws; in mlx5_tx_wseg_init() local 818 ws = RTE_PTR_ADD(wqe, MLX5_WSEG_SIZE); in mlx5_tx_wseg_init() 819 ws->operation = rte_cpu_to_be_32(MLX5_WAIT_COND_CYCLIC_BIGGER); in mlx5_tx_wseg_init() 820 ws->lkey = RTE_BE32(0); in mlx5_tx_wseg_init() 821 ws->va_high = RTE_BE32(0); in mlx5_tx_wseg_init() 822 ws->va_low = RTE_BE32(0); in mlx5_tx_wseg_init() 827 ws->value = rte_cpu_to_be_64(ts); in mlx5_tx_wseg_init() 828 ws->mask = txq->rt_timemask; in mlx5_tx_wseg_init()
|
| /dpdk/drivers/crypto/octeontx/ |
| H A D | otx_cryptodev_ops.c | 672 submit_request_to_sso(struct ssows *ws, uintptr_t req, in submit_request_to_sso() argument 681 ssows_head_wait(ws); in submit_request_to_sso() 684 ssovf_store_pair(add_work, req, ws->grps[rsp_info->queue_id]); in submit_request_to_sso()
|
| /dpdk/doc/guides/nics/ |
| H A D | ice.rst | 20 `https://www.intel.com/content/www/us/en/search.html?ws=text#q=e810&t=Downloads&layout=table`.
|