| /dpdk/drivers/net/ice/base/ |
| H A D | ice_flex_pipe.c | 2578 fv_ext = hw->blk[blk].es.t + (prof * hw->blk[blk].es.fvw); in ice_find_prot_off() 3082 &hw->blk[blk].xlt2.vsis[vsi]; in ice_vsig_add_mv_vsi() 3084 hw->blk[blk].xlt2.t[vsi] = vsig; in ice_vsig_add_mv_vsi() 3111 for (i = hw->blk[blk].masks.first; i < hw->blk[blk].masks.first + in ice_prof_has_mask_idx() 3445 i < hw->blk[blk].masks.first + hw->blk[blk].masks.count; i++) in ice_init_prof_masks() 3482 i < hw->blk[blk].masks.first + hw->blk[blk].masks.count; i++) in ice_alloc_prof_mask() 3542 mask_idx < hw->blk[blk].masks.first + hw->blk[blk].masks.count)) in ice_free_prof_mask() 3608 i < hw->blk[blk].masks.first + hw->blk[blk].masks.count; i++) { in ice_shutdown_prof_masks() 3696 ice_memset(&hw->blk[blk].es.t[off], 0, hw->blk[blk].es.fvw * in ice_write_es() 3700 ice_memcpy(&hw->blk[blk].es.t[off], fv, hw->blk[blk].es.fvw * in ice_write_es() [all …]
|
| H A D | ice_flow.c | 1327 u8 fv_words = hw->blk[params->blk].es.fvw; in ice_flow_xtract_pkt_flags() 1337 if (hw->blk[params->blk].es.reverse) in ice_flow_xtract_pkt_flags() 1367 u8 fv_words = hw->blk[params->blk].es.fvw; in ice_flow_xtract_fld() 1573 if (hw->blk[params->blk].es.reverse) in ice_flow_xtract_fld() 1616 fv_words = hw->blk[params->blk].es.fvw; in ice_flow_xtract_raws() 1651 if (hw->blk[params->blk].es.reverse) in ice_flow_xtract_raws() 1996 ice_acquire_lock(&hw->blk[blk].es.prof_map_lock); in ice_flow_get_hw_prof() 2002 ice_release_lock(&hw->blk[blk].es.prof_map_lock); in ice_flow_get_hw_prof() 2212 params->blk = blk; in ice_flow_add_prof_sync() 2551 u8 fv_words = hw->blk[blk].es.fvw; in ice_flow_set_hw_prof() [all …]
|
| H A D | ice_flex_pipe.h | 28 ice_find_prot_off(struct ice_hw *hw, enum ice_block blk, u8 prof, u16 fv_idx, 63 ice_vsig_find_vsi(struct ice_hw *hw, enum ice_block blk, u16 vsi, u16 *vsig); 66 ice_add_prof(struct ice_hw *hw, enum ice_block blk, u64 id, u8 ptypes[], 72 ice_search_prof_id(struct ice_hw *hw, enum ice_block blk, u64 id); 74 ice_add_vsi_flow(struct ice_hw *hw, enum ice_block blk, u16 vsi, u16 vsig); 76 ice_add_prof_id_flow(struct ice_hw *hw, enum ice_block blk, u16 vsi, u64 hdl); 78 ice_rem_prof_id_flow(struct ice_hw *hw, enum ice_block blk, u16 vsi, u64 hdl); 80 ice_flow_assoc_hw_prof(struct ice_hw *hw, enum ice_block blk, 91 ice_rem_prof(struct ice_hw *hw, enum ice_block blk, u64 id);
|
| H A D | ice_flow.h | 549 ice_flow_find_prof(struct ice_hw *hw, enum ice_block blk, enum ice_flow_dir dir, 552 ice_flow_add_prof(struct ice_hw *hw, enum ice_block blk, enum ice_flow_dir dir, 557 ice_flow_rem_prof(struct ice_hw *hw, enum ice_block blk, u64 prof_id); 559 ice_flow_assoc_prof(struct ice_hw *hw, enum ice_block blk, 562 ice_flow_assoc_vsig_vsi(struct ice_hw *hw, enum ice_block blk, u16 vsi_handle, 567 enum ice_block blk); 569 ice_flow_get_hw_prof(struct ice_hw *hw, enum ice_block blk, u64 prof_id, 572 u64 ice_flow_find_entry(struct ice_hw *hw, enum ice_block blk, u64 entry_id); 574 ice_flow_add_entry(struct ice_hw *hw, enum ice_block blk, u64 prof_id, 579 ice_flow_rem_entry(struct ice_hw *hw, enum ice_block blk, u64 entry_h); [all …]
|
| H A D | ice_parser.c | 489 int buf_len, enum ice_block blk, in ice_parser_profile_init() argument 499 if (blk == ICE_BLK_SW) { in ice_parser_profile_init() 502 } else if (blk == ICE_BLK_ACL) { in ice_parser_profile_init() 505 } else if (blk == ICE_BLK_FD) { in ice_parser_profile_init() 508 } else if (blk == ICE_BLK_RSS) { in ice_parser_profile_init()
|
| H A D | ice_parser.h | 108 int buf_len, enum ice_block blk,
|
| H A D | ice_type.h | 1244 struct ice_blk_info blk[ICE_BLK_COUNT]; member
|
| H A D | ice_switch.c | 7027 for (j = 0; j < hw->blk[ICE_BLK_SW].es.fvw; j++) in ice_fill_fv_word_index()
|
| /dpdk/drivers/net/bnxt/ |
| H A D | bnxt_rxtx_vec_common.h | 138 uint32_t ring_mask, ring_size, num, blk; in bnxt_tx_cmp_vec() local 151 blk = 0; in bnxt_tx_cmp_vec() 153 mbuf = txr->tx_buf_ring[cons + blk]; in bnxt_tx_cmp_vec() 157 blk++; in bnxt_tx_cmp_vec() 158 } while (blk < num); in bnxt_tx_cmp_vec() 160 if (blk) { in bnxt_tx_cmp_vec() 163 blk); in bnxt_tx_cmp_vec() 165 blk * sizeof(struct rte_mbuf *)); in bnxt_tx_cmp_vec() 166 raw_cons += blk; in bnxt_tx_cmp_vec() 167 nr_pkts -= blk; in bnxt_tx_cmp_vec()
|
| H A D | bnxt_txr.c | 365 unsigned int blk = 0; in bnxt_tx_cmp_fast() local 378 free[blk++] = *tx_buf; in bnxt_tx_cmp_fast() 385 if (blk) in bnxt_tx_cmp_fast() 386 rte_mempool_put_bulk(free[0]->pool, (void *)free, blk); in bnxt_tx_cmp_fast() 398 unsigned int blk = 0; in bnxt_tx_cmp() local 425 free[blk++] = mbuf; in bnxt_tx_cmp() 434 blk); in bnxt_tx_cmp() 439 blk = 1; in bnxt_tx_cmp() 443 if (blk) in bnxt_tx_cmp() 444 rte_mempool_put_bulk(pool, (void *)free, blk); in bnxt_tx_cmp()
|
| /dpdk/doc/guides/sample_app_ug/ |
| H A D | vhost_blk.rst | 8 which used as the backend of Qemu vhost-user-blk device. Users can extend 15 Development Kit, https://github.com/spdk/spdk) vhost-user-blk target, 54 -device vhost-user-blk-pci,packed=on,chardev=char0,num-queues=1 \ 58 You must check whether your Qemu can support "vhost-user-blk" or not, 65 vhost-blk reconnect and packed ring.
|
| /dpdk/examples/vhost_blk/ |
| H A D | Makefile | 5 APP = vhost-blk 8 SRCS-y := blk.c vhost_blk.c vhost_blk_compat.c
|
| H A D | meson.build | 22 'blk.c',
|
| /dpdk/drivers/common/cnxk/ |
| H A D | roc_cpt_priv.h | 24 int cpt_lfs_alloc(struct dev *dev, uint8_t eng_grpmsk, uint8_t blk,
|
| /dpdk/drivers/common/sfc_efx/base/ |
| H A D | ef10_nic.c | 899 uint32_t buf, blk; in ef10_nic_pio_alloc() local 923 for (blk = 0; blk < blk_per_buf; blk++) { in ef10_nic_pio_alloc() 924 if ((*map & (1u << blk)) == 0) { in ef10_nic_pio_alloc() 925 *map |= (1u << blk); in ef10_nic_pio_alloc() 936 *blknump = blk; in ef10_nic_pio_alloc() 938 *offsetp = blk * (*sizep); in ef10_nic_pio_alloc()
|
| /dpdk/doc/guides/nics/ |
| H A D | liquidio.rst | 99 -device virtio-blk-pci,scsi=off,drive=disk1,id=virtio-disk1,bootindex=1 \
|
| H A D | thunderx.rst | 134 -device virtio-blk-device,scsi=off,drive=disk1,id=virtio-disk1,bootindex=1 \
|
| /dpdk/drivers/net/ice/ |
| H A D | ice_fdir_filter.c | 1329 int ptg = hw->blk[ICE_BLK_FD].xlt1.t[id]; in ice_fdir_create_filter() 1488 int ptg = hw->blk[ICE_BLK_FD].xlt1.t[id]; in ice_fdir_destroy_filter() 1491 enum ice_block blk = ICE_BLK_FD; in ice_fdir_destroy_filter() local 1503 ice_rem_prof_id_flow(hw, blk, vsi_num, id); in ice_fdir_destroy_filter() 1506 ice_rem_prof_id_flow(hw, blk, vsi_num, id); in ice_fdir_destroy_filter()
|
| H A D | ice_acl_filter.c | 417 enum ice_block blk = ICE_BLK_ACL; in ice_acl_hw_set_conf() local 435 ret = ice_flow_add_entry(hw, blk, flow_type, in ice_acl_hw_set_conf()
|
| H A D | ice_hash.c | 1221 ptg = hw->blk[ICE_BLK_RSS].xlt1.t[id]; in ice_hash_add_raw_cfg() 1359 ptg = hw->blk[ICE_BLK_RSS].xlt1.t[id]; in ice_hash_rem_raw_cfg()
|
| /dpdk/drivers/net/qede/ |
| H A D | qede_debug.c | 1323 const struct dbg_block_chip *blk; in qed_update_blocks_reset_state() local 1327 blk = qed_get_dbg_block_per_chip(p_hwfn, (enum block_id)blk_id); in qed_update_blocks_reset_state() 1328 is_removed = GET_FIELD(blk->flags, DBG_BLOCK_CHIP_IS_REMOVED); in qed_update_blocks_reset_state() 1329 has_rst_reg = GET_FIELD(blk->flags, in qed_update_blocks_reset_state() 1334 !(reg_val[blk->reset_reg_id] & in qed_update_blocks_reset_state() 1335 OSAL_BIT(blk->reset_reg_bit_offset)); in qed_update_blocks_reset_state()
|
| /dpdk/drivers/net/bnx2x/ |
| H A D | bnx2x.c | 3114 __rte_unused const char *blk) in bnx2x_print_next_block() argument 3116 PMD_DRV_LOG(INFO, sc, "%s%s", idx ? ", " : "", blk); in bnx2x_print_next_block()
|