| /dpdk/drivers/event/sw/ |
| H A D | sw_evdev_xstats.c | 56 case rx: return sw->stats.rx_pkts; in get_dev_stat() 260 sw->qid_count * sw->port_count * in sw_xstats_init() 264 sw->xstats = rte_zmalloc_socket(NULL, sizeof(sw->xstats[0]) * count, 0, in sw_xstats_init() 265 sw->data->socket_id); in sw_xstats_init() 266 if (sw->xstats == NULL) in sw_xstats_init() 322 sw->xstats_count_mode_port = stat - sw->xstats_count_mode_dev; in sw_xstats_init() 376 (sw->xstats_count_mode_dev + sw->xstats_count_mode_port); in sw_xstats_init() 379 sw->xstats_count = stat; in sw_xstats_init() 387 rte_free(sw->xstats); in sw_xstats_uninit() 388 sw->xstats_count = 0; in sw_xstats_uninit() [all …]
|
| H A D | sw_evdev_scheduler.c | 90 sw->cq_ring_space[cq]--; in sw_schedule_atomic_to_cq() 160 sw->cq_ring_space[cq]--; in sw_schedule_parallel_to_cq() 172 sw->ports[cq].cq_buf[sw->ports[cq].cq_buf_count++] = *qe; in sw_schedule_parallel_to_cq() 219 sw->sched_cq_qid_called++; in sw_schedule_qid_to_cq() 315 iq_enqueue(sw, iq, qe); in sw_schedule_reorder() 343 RTE_SET_USED(sw); in sw_refill_pp_buf() 496 iq_enqueue(sw, iq, qe); in sw_schedule_pull_port_dir() 518 sw->sched_called++; in sw_event_schedule() 543 sw->qid_count); in sw_event_schedule() 588 sw->cq_ring_space[i] = in sw_event_schedule() [all …]
|
| H A D | sw_evdev.c | 160 p->sw = sw; in sw_port_setup() 536 if (!sw->chunks) in sw_dev_configure() 541 iq_free_chunk(sw, &sw->chunks[i]); in sw_dev_configure() 628 sw->port_count, sw->qid_count); in sw_dump() 631 sw->stats.rx_pkts, sw->stats.rx_dropped, sw->stats.tx_pkts); in sw_dump() 805 sw->qids_prioritized[qidx] = &sw->qids[i]; in sw_start() 817 sw->started = 1; in sw_start() 837 while (!(sw_qids_empty(sw) && sw_ports_empty(sw))) { in sw_stop() 845 sw->started = 0; in sw_stop() 866 memset(&sw->stats, 0, sizeof(sw->stats)); in sw_close() [all …]
|
| H A D | iq_chunk.h | 32 iq_alloc_chunk(struct sw_evdev *sw) in iq_alloc_chunk() argument 35 sw->chunk_list_head = chunk->next; in iq_alloc_chunk() 43 chunk->next = sw->chunk_list_head; in iq_free_chunk() 44 sw->chunk_list_head = chunk; in iq_free_chunk() 53 iq_free_chunk(sw, head); in iq_free_chunk_list() 61 iq->head = iq_alloc_chunk(sw); in iq_init() 94 iq_free_chunk(sw, iq->head); in iq_pop() 134 iq_free_chunk(sw, current); in iq_dequeue_burst() 142 iq_free_chunk(sw, current); in iq_dequeue_burst() 156 iq_put_back(struct sw_evdev *sw, in iq_put_back() argument [all …]
|
| H A D | sw_evdev_worker.c | 58 struct sw_evdev *sw = (void *)p->sw; in sw_event_enqueue_burst() local 59 uint32_t sw_inflights = rte_atomic32_read(&sw->inflights); in sw_event_enqueue_burst() 60 uint32_t credit_update_quanta = sw->credit_update_quanta; in sw_event_enqueue_burst() 74 if (sw_inflights + credit_update_quanta > sw->nb_events_limit) in sw_event_enqueue_burst() 77 rte_atomic32_add(&sw->inflights, credit_update_quanta); in sw_event_enqueue_burst() 89 const uint8_t invalid_qid = (ev[i].queue_id >= sw->qid_count); in sw_event_enqueue_burst() 127 rte_atomic32_sub(&sw->inflights, credit_update_quanta); in sw_event_enqueue_burst() 150 struct sw_evdev *sw = (void *)p->sw; in sw_event_dequeue_burst() local 151 uint32_t credit_update_quanta = sw->credit_update_quanta; in sw_event_dequeue_burst() 159 rte_atomic32_sub(&sw->inflights, credit_update_quanta); in sw_event_dequeue_burst()
|
| H A D | sw_evdev.h | 150 struct sw_evdev *sw; member
|
| /dpdk/lib/eventdev/ |
| H A D | rte_event_timer_adapter.c | 609 struct swtim *sw; in swtim_callback() local 657 sw->expired_timers[sw->n_expired_timers++] = tim; in swtim_callback() 770 rte_mempool_put_bulk(sw->tim_pool, (void **)sw->expired_timers, in swtim_service_func() 825 struct swtim *sw; in swtim_init() local 834 sw = rte_zmalloc_socket(swtim_name, sizeof(*sw), RTE_CACHE_LINE_SIZE, in swtim_init() 836 if (sw == NULL) { in swtim_init() 870 sw->in_use[i].v = 0; in swtim_init() 922 rte_free(sw); in swtim_init() 948 sw); in swtim_uninit() 957 rte_free(sw); in swtim_uninit() [all …]
|
| /dpdk/drivers/net/nfp/nfpcore/ |
| H A D | nfp_rtsym.c | 50 struct nfp_rtsym *sw, struct nfp_rtsym_entry *fw) in nfp_rtsym_sw_entry_init() argument 52 sw->type = fw->type; in nfp_rtsym_sw_entry_init() 54 sw->addr = ((uint64_t)fw->addr_hi << 32) | in nfp_rtsym_sw_entry_init() 56 sw->size = ((uint64_t)fw->size_hi << 32) | in nfp_rtsym_sw_entry_init() 62 sw->name, sw->addr, sw->size, sw->target); in nfp_rtsym_sw_entry_init() 66 sw->target = NFP_RTSYM_TARGET_LMEM; in nfp_rtsym_sw_entry_init() 69 sw->target = NFP_RTSYM_TARGET_EMU_CACHE; in nfp_rtsym_sw_entry_init() 72 sw->target = fw->target; in nfp_rtsym_sw_entry_init() 77 sw->domain = nfp_meid(fw->island, fw->menum); in nfp_rtsym_sw_entry_init() 79 sw->domain = fw->island; in nfp_rtsym_sw_entry_init() [all …]
|
| /dpdk/drivers/net/atlantic/hw_atl/ |
| H A D | hw_atl_utils.c | 424 struct aq_hw_atl_utils_fw_rpc_tid_s sw; in hw_atl_utils_fw_rpc_call() local 437 sw.tid = 0xFFFFU & (++self->rpc_tid); in hw_atl_utils_fw_rpc_call() 438 sw.len = (u16)rpc_size; in hw_atl_utils_fw_rpc_call() 439 aq_hw_write_reg(self, HW_ATL_RPC_CONTROL_ADR, sw.val); in hw_atl_utils_fw_rpc_call() 449 struct aq_hw_atl_utils_fw_rpc_tid_s sw; in hw_atl_utils_fw_rpc_wait() local 453 sw.val = aq_hw_read_reg(self, HW_ATL_RPC_CONTROL_ADR); in hw_atl_utils_fw_rpc_wait() 455 self->rpc_tid = sw.tid; in hw_atl_utils_fw_rpc_wait() 457 AQ_HW_WAIT_FOR(sw.tid == in hw_atl_utils_fw_rpc_wait() 465 err = hw_atl_utils_fw_rpc_call(self, sw.len); in hw_atl_utils_fw_rpc_wait() 469 } while (sw.tid != fw.tid || 0xFFFFU == fw.len); in hw_atl_utils_fw_rpc_wait()
|
| /dpdk/drivers/net/ice/base/ |
| H A D | ice_switch.c | 5150 struct ice_switch_info *sw) in ice_add_vlan_rule() argument 5304 struct ice_switch_info *sw) in ice_remove_eth_mac_rule() argument 6221 vlan_id, sw); in _ice_set_vlan_vsi_promisc() 6225 lport, sw); in _ice_set_vlan_vsi_promisc() 8829 struct ice_switch_info *sw; in ice_add_adv_rule() local 9009 sw = hw->switch_info; in ice_add_adv_rule() 9246 struct ice_switch_info *sw; in ice_rem_adv_rule_by_id() local 9248 sw = hw->switch_info; in ice_rem_adv_rule_by_id() 9282 struct ice_switch_info *sw; in ice_rem_adv_rule_for_vsi() local 9286 sw = hw->switch_info; in ice_rem_adv_rule_for_vsi() [all …]
|
| H A D | ice_common.c | 741 struct ice_switch_info *sw; in ice_init_fltr_mgmt_struct() local 747 sw = hw->switch_info; in ice_init_fltr_mgmt_struct() 749 if (!sw) in ice_init_fltr_mgmt_struct() 752 INIT_LIST_HEAD(&sw->vsi_list_map_head); in ice_init_fltr_mgmt_struct() 753 sw->prof_res_bm_init = 0; in ice_init_fltr_mgmt_struct() 776 if (!sw) in ice_cleanup_fltr_mgmt_single() 784 recps = sw->recp_list; in ice_cleanup_fltr_mgmt_single() 824 ice_rm_sw_replay_rule_info(hw, sw); in ice_cleanup_fltr_mgmt_single() 825 ice_free(hw, sw->recp_list); in ice_cleanup_fltr_mgmt_single() 826 ice_free(hw, sw); in ice_cleanup_fltr_mgmt_single() [all …]
|
| H A D | ice_common.h | 227 ice_replay_pre_init(struct ice_hw *hw, struct ice_switch_info *sw);
|
| H A D | ice_switch.h | 548 void ice_rm_sw_replay_rule_info(struct ice_hw *hw, struct ice_switch_info *sw);
|
| /dpdk/drivers/net/octeontx/base/ |
| H A D | octeontx_pki_var.h | 74 uint64_t sw : 1; member 120 uint64_t sw : 1; 228 uint64_t sw:9; member
|
| /dpdk/drivers/event/ |
| H A D | meson.build | 16 'sw',
|
| /dpdk/doc/guides/eventdevs/ |
| H A D | index.rst | 20 sw
|
| /dpdk/doc/guides/sample_app_ug/ |
| H A D | dma.rst | 48 ./<build_dir>/examples/dpdk-ioat [EAL options] -- [-p MASK] [-q NQ] [-s RS] [-c <sw|hw>] 58 * c CT: Performed packet copy type: software (sw) or hardware using 90 $ ./<build_dir>/examples/dpdk-dma -l 0-2 -n 2 -- -p 0x1 --mac-updating -c sw
|
| /dpdk/doc/guides/vdpadevs/ |
| H A D | ifc.rst | 30 The device argument "sw-live-migration=1" will configure the driver into SW
|
| /dpdk/doc/guides/nics/ |
| H A D | nfb.rst | 46 `Netcope common <https://github.com/CESNET/ndk-sw>`_.
|
| H A D | enic.rst | 118 <http://www.cisco.com/c/en/us/td/docs/unified_computing/ucs/sw/vm_fex/vmware/gui/config_guide/b_GUI…
|
| H A D | bnxt.rst | 350 testpmd> csum set (ip|tcp|udp|outer-ip|outer-udp) (sw|hw) (port_id)
|
| /dpdk/drivers/net/ice/ |
| H A D | ice_switch_filter.c | 1970 struct ice_switch_info *sw; in ice_switch_redirect() local 1979 sw = hw->switch_info; in ice_switch_redirect() 1980 if (!sw->recp_list[rdata->rid].recp_created) in ice_switch_redirect() 1988 list_head = &sw->recp_list[rdata->rid].filt_rules; in ice_switch_redirect()
|
| /dpdk/doc/guides/testpmd_app_ug/ |
| H A D | run_app.rst | 507 * ``--noisy-tx-sw-buffer-size`` 513 * ``--noisy-tx-sw-buffer-flushtime=N``
|
| /dpdk/doc/guides/contributing/ |
| H A D | coding_style.rst | 820 * The eventdev software poll mode driver in ``drivers/event/sw`` uses ``pmd.event.sw``
|
| /dpdk/ |
| H A D | MAINTAINERS | 1286 F: drivers/event/sw/ 1287 F: doc/guides/eventdevs/sw.rst
|