| /f-stack/freebsd/net80211/ |
| H A D | ieee80211_ageq.c | 56 memset(aq, 0, sizeof(*aq)); in ieee80211_ageq_init() 68 KASSERT(aq->aq_len == 0, ("%d frames on ageq", aq->aq_len)); in ieee80211_ageq_cleanup() 111 if (__predict_true(aq->aq_len < aq->aq_maxlen)) { in ieee80211_ageq_append() 113 aq->aq_head = m; in ieee80211_ageq_append() 121 aq->aq_tail = m; in ieee80211_ageq_append() 122 aq->aq_len++; in ieee80211_ageq_append() 129 aq->aq_drops++; in ieee80211_ageq_append() 175 KASSERT(aq->aq_len > 0, ("aq len %d", aq->aq_len)); in ieee80211_ageq_age() 176 aq->aq_len--; in ieee80211_ageq_age() 213 KASSERT(aq->aq_len > 0, ("aq len %d", aq->aq_len)); in ieee80211_ageq_remove() [all …]
|
| /f-stack/dpdk/drivers/net/i40e/base/ |
| H A D | i40e_adminq.c | 284 wr32(hw, hw->aq.asq.len, (hw->aq.num_asq_entries | in i40e_config_asq_regs() 287 wr32(hw, hw->aq.asq.len, (hw->aq.num_asq_entries | in i40e_config_asq_regs() 294 wr32(hw, hw->aq.asq.len, (hw->aq.num_asq_entries | in i40e_config_asq_regs() 297 wr32(hw, hw->aq.asq.len, (hw->aq.num_asq_entries | in i40e_config_asq_regs() 334 wr32(hw, hw->aq.arq.len, (hw->aq.num_arq_entries | in i40e_config_arq_regs() 344 wr32(hw, hw->aq.arq.len, (hw->aq.num_arq_entries | in i40e_config_arq_regs() 411 hw->aq.asq.count = hw->aq.num_asq_entries; in i40e_init_asq() 474 hw->aq.arq.count = hw->aq.num_arq_entries; in i40e_init_arq() 578 struct i40e_adminq_info *aq = &hw->aq; in i40e_set_hw_flags() local 649 struct i40e_adminq_info *aq = &hw->aq; in i40e_init_adminq() local [all …]
|
| H A D | i40e_nvm.c | 1044 hw->aq.asq_last_status); in i40e_nvmupd_state_init() 1055 hw->aq.asq_last_status); in i40e_nvmupd_state_init() 1069 hw->aq.asq_last_status); in i40e_nvmupd_state_init() 1086 hw->aq.asq_last_status); in i40e_nvmupd_state_init() 1103 hw->aq.asq_last_status); in i40e_nvmupd_state_init() 1119 hw->aq.asq_last_status); in i40e_nvmupd_state_init() 1238 *perrno = hw->aq.asq_last_status ? in i40e_nvmupd_state_writing() 1254 *perrno = hw->aq.asq_last_status ? in i40e_nvmupd_state_writing() 1312 hw->aq.asq_last_status); in i40e_nvmupd_state_writing() 1341 if (hw->aq.arq_last_status) { in i40e_nvmupd_clear_wait_state() [all …]
|
| H A D | i40e_dcb.c | 787 if (hw->aq.asq_last_status == I40E_AQ_RC_ENOENT) in i40e_get_ieee_dcb_config() 808 (((hw->aq.fw_maj_ver == 4) && (hw->aq.fw_min_ver < 33)) || in i40e_get_dcb_config() 809 (hw->aq.fw_maj_ver < 4))) in i40e_get_dcb_config() 814 ((hw->aq.fw_maj_ver == 4) && (hw->aq.fw_min_ver == 33))) { in i40e_get_dcb_config() 839 if (hw->aq.asq_last_status == I40E_AQ_RC_ENOENT) in i40e_get_dcb_config() 856 if (hw->aq.asq_last_status == I40E_AQ_RC_ENOENT) in i40e_get_dcb_config() 965 } else if (hw->aq.asq_last_status == I40E_AQ_RC_ENOENT) { in i40e_get_fw_lldp_status() 969 } else if (hw->aq.asq_last_status == I40E_AQ_RC_EPERM) { in i40e_get_fw_lldp_status()
|
| H A D | i40e_common.c | 378 if (hw->aq.asq.len) in i40e_check_asq_alive() 1742 switch (hw->aq.asq_last_status) { in i40e_aq_get_phy_capabilities() 2077 (hw->aq.fw_maj_ver < 4 || (hw->aq.fw_maj_ver == 4 && in i40e_aq_get_link_info() 2285 if (hw->aq.api_maj_ver > maj || in i40e_hw_ver_ge() 2286 (hw->aq.api_maj_ver == maj && hw->aq.api_min_ver >= min)) in i40e_hw_ver_ge() 6027 if (hw->aq.fw_maj_ver < 6) in i40e_aq_replace_cloud_filters() 7337 (hw->aq.api_min_ver < 5)) || in i40e_read_rx_ctl() 7398 (hw->aq.api_min_ver < 5)) || in i40e_write_rx_ctl() 8041 msglen = aq->datalen; in i40e_ddp_exec_aq_section() 8048 msg = &aq->data[0]; in i40e_ddp_exec_aq_section() [all …]
|
| H A D | i40e_type.h | 716 struct i40e_adminq_info aq; member
|
| /f-stack/dpdk/drivers/event/octeontx2/ |
| H A D | otx2_evdev_adptr.c | 41 aq->cq_mask.ena = ~(aq->cq_mask.ena); in sso_rxq_enable() 42 aq->cq_mask.caching = ~(aq->cq_mask.caching); in sso_rxq_enable() 92 aq->rq.bad_utag = aq->rq.good_utag; in sso_rxq_enable() 100 aq->rq_mask.sso_ena = ~(aq->rq_mask.sso_ena); in sso_rxq_enable() 101 aq->rq_mask.sso_tt = ~(aq->rq_mask.sso_tt); in sso_rxq_enable() 108 aq->rq_mask.ltag = ~(aq->rq_mask.ltag); in sso_rxq_enable() 111 aq->rq_mask.ena = ~(aq->rq_mask.ena); in sso_rxq_enable() 142 aq->cq_mask.ena = ~(aq->cq_mask.ena); in sso_rxq_disable() 143 aq->cq_mask.caching = ~(aq->cq_mask.caching); in sso_rxq_disable() 181 aq->rq_mask.ltag = ~(aq->rq_mask.ltag); in sso_rxq_disable() [all …]
|
| /f-stack/dpdk/drivers/common/iavf/ |
| H A D | iavf_adminq.c | 266 wr32(hw, hw->aq.asq.len, (hw->aq.num_asq_entries | in iavf_config_asq_regs() 268 wr32(hw, hw->aq.asq.bal, IAVF_LO_DWORD(hw->aq.asq.desc_buf.pa)); in iavf_config_asq_regs() 269 wr32(hw, hw->aq.asq.bah, IAVF_HI_DWORD(hw->aq.asq.desc_buf.pa)); in iavf_config_asq_regs() 295 wr32(hw, hw->aq.arq.len, (hw->aq.num_arq_entries | in iavf_config_arq_regs() 301 wr32(hw, hw->aq.arq.tail, hw->aq.num_arq_entries - 1); in iavf_config_arq_regs() 360 hw->aq.asq.count = hw->aq.num_asq_entries; in iavf_init_asq() 423 hw->aq.arq.count = hw->aq.num_arq_entries; in iavf_init_arq() 628 return rd32(hw, hw->aq.asq.head) == hw->aq.asq.next_to_use; in iavf_asq_done() 744 dma_buff = &(hw->aq.asq.r.asq_bi[hw->aq.asq.next_to_use]); in iavf_asq_send_command() 764 if (hw->aq.asq.next_to_use == hw->aq.asq.count) in iavf_asq_send_command() [all …]
|
| H A D | iavf_type.h | 341 struct iavf_adminq_info aq; member
|
| H A D | iavf_common.c | 332 if (hw->aq.asq.len) in iavf_check_asq_alive() 333 return !!(rd32(hw, hw->aq.asq.len) & in iavf_check_asq_alive()
|
| /f-stack/dpdk/drivers/net/octeontx2/ |
| H A D | otx2_stats.c | 326 otx2_mbox_memset(&aq->rq_mask, 0, sizeof(aq->rq_mask)); in nix_queue_stats_reset() 333 aq->rq_mask.octs = ~(aq->rq_mask.octs); in nix_queue_stats_reset() 334 aq->rq_mask.pkts = ~(aq->rq_mask.pkts); in nix_queue_stats_reset() 335 aq->rq_mask.drop_octs = ~(aq->rq_mask.drop_octs); in nix_queue_stats_reset() 336 aq->rq_mask.drop_pkts = ~(aq->rq_mask.drop_pkts); in nix_queue_stats_reset() 337 aq->rq_mask.re_pkts = ~(aq->rq_mask.re_pkts); in nix_queue_stats_reset() 360 otx2_mbox_memset(&aq->sq_mask, 0, sizeof(aq->sq_mask)); in nix_queue_stats_reset() 366 aq->sq_mask.octs = ~(aq->sq_mask.octs); in nix_queue_stats_reset() 367 aq->sq_mask.pkts = ~(aq->sq_mask.pkts); in nix_queue_stats_reset() 368 aq->sq_mask.drop_octs = ~(aq->sq_mask.drop_octs); in nix_queue_stats_reset() [all …]
|
| H A D | otx2_flow_ctrl.c | 87 struct nix_aq_enq_req *aq; in otx2_nix_cq_bp_cfg() local 95 if (!aq) { in otx2_nix_cq_bp_cfg() 105 if (!aq) in otx2_nix_cq_bp_cfg() 108 aq->qidx = rxq->rq; in otx2_nix_cq_bp_cfg() 109 aq->ctype = NIX_AQ_CTYPE_CQ; in otx2_nix_cq_bp_cfg() 113 aq->cq.bpid = fc->bpid[0]; in otx2_nix_cq_bp_cfg() 114 aq->cq_mask.bpid = ~(aq->cq_mask.bpid); in otx2_nix_cq_bp_cfg() 115 aq->cq.bp = rxq->cq_drop; in otx2_nix_cq_bp_cfg() 116 aq->cq_mask.bp = ~(aq->cq_mask.bp); in otx2_nix_cq_bp_cfg() 119 aq->cq.bp_ena = !!enb; in otx2_nix_cq_bp_cfg() [all …]
|
| H A D | otx2_ethdev.c | 308 aq->qidx = qid; in nix_cq_rq_init() 312 aq->cq.ena = 1; in nix_cq_rq_init() 354 aq->qidx = qid; in nix_cq_rq_init() 377 aq->rq.ena = 1; in nix_cq_rq_init() 399 if (!aq) { in nix_cq_rq_init() 411 if (!aq) { in nix_cq_rq_init() 444 aq->rq_mask.ena = ~(aq->rq_mask.ena); in nix_rq_enb_dis() 465 aq->cq_mask.ena = ~(aq->cq_mask.ena); in nix_cq_rq_uninit() 480 if (!aq) { in nix_cq_rq_uninit() 492 if (!aq) { in nix_cq_rq_uninit() [all …]
|
| H A D | otx2_ethdev_debug.c | 398 struct nix_aq_enq_req *aq; in otx2_nix_queues_ctx_dump() local 403 aq = otx2_mbox_alloc_msg_nix_aq_enq(mbox); in otx2_nix_queues_ctx_dump() 404 aq->qidx = q; in otx2_nix_queues_ctx_dump() 405 aq->ctype = NIX_AQ_CTYPE_CQ; in otx2_nix_queues_ctx_dump() 406 aq->op = NIX_AQ_INSTOP_READ; in otx2_nix_queues_ctx_dump() 420 aq->qidx = q; in otx2_nix_queues_ctx_dump() 421 aq->ctype = NIX_AQ_CTYPE_RQ; in otx2_nix_queues_ctx_dump() 422 aq->op = NIX_AQ_INSTOP_READ; in otx2_nix_queues_ctx_dump() 435 aq->qidx = q; in otx2_nix_queues_ctx_dump() 436 aq->ctype = NIX_AQ_CTYPE_SQ; in otx2_nix_queues_ctx_dump() [all …]
|
| H A D | otx2_ethdev_sec.c | 794 struct nix_aq_enq_req *aq; in otx2_eth_sec_update_tag_type() local 797 aq = otx2_mbox_alloc_msg_nix_aq_enq(mbox); in otx2_eth_sec_update_tag_type() 798 aq->qidx = 0; /* Read RQ:0 context */ in otx2_eth_sec_update_tag_type() 799 aq->ctype = NIX_AQ_CTYPE_RQ; in otx2_eth_sec_update_tag_type() 800 aq->op = NIX_AQ_INSTOP_READ; in otx2_eth_sec_update_tag_type()
|
| /f-stack/dpdk/drivers/mempool/octeontx2/ |
| H A D | otx2_mempool_debug.c | 91 struct npa_aq_enq_req *aq; in otx2_mempool_ctx_dump() local 101 aq = otx2_mbox_alloc_msg_npa_aq_enq(lf->mbox); in otx2_mempool_ctx_dump() 102 aq->aura_id = q; in otx2_mempool_ctx_dump() 103 aq->ctype = NPA_AQ_CTYPE_POOL; in otx2_mempool_ctx_dump() 104 aq->op = NPA_AQ_INSTOP_READ; in otx2_mempool_ctx_dump() 120 aq = otx2_mbox_alloc_msg_npa_aq_enq(lf->mbox); in otx2_mempool_ctx_dump() 121 aq->aura_id = q; in otx2_mempool_ctx_dump() 122 aq->ctype = NPA_AQ_CTYPE_AURA; in otx2_mempool_ctx_dump() 123 aq->op = NPA_AQ_INSTOP_READ; in otx2_mempool_ctx_dump()
|
| /f-stack/dpdk/drivers/net/iavf/ |
| H A D | iavf.h | 228 hw->aq.num_arq_entries = IAVF_AQ_LEN; in iavf_init_adminq_parameter() 229 hw->aq.num_asq_entries = IAVF_AQ_LEN; in iavf_init_adminq_parameter() 230 hw->aq.arq_buf_size = IAVF_AQ_BUF_SZ; in iavf_init_adminq_parameter() 231 hw->aq.asq_buf_size = IAVF_AQ_BUF_SZ; in iavf_init_adminq_parameter()
|
| /f-stack/freebsd/contrib/openzfs/module/os/linux/spl/ |
| H A D | spl-kmem.c | 77 va_list aq; in kmem_vasprintf() local 81 va_copy(aq, ap); in kmem_vasprintf() 82 ptr = kvasprintf(kmem_flags_convert(KM_SLEEP), fmt, aq); in kmem_vasprintf() 83 va_end(aq); in kmem_vasprintf()
|
| /f-stack/dpdk/drivers/net/mlx5/ |
| H A D | mlx5_txpp.c | 625 struct mlx5_txpp_wq *aq = &sh->txpp.rearm_queue; in mlx5_txpp_cq_arm() local 626 uint32_t arm_sn = aq->arm_sn << MLX5_CQ_SQN_OFFSET; in mlx5_txpp_cq_arm() 627 uint32_t db_hi = arm_sn | MLX5_CQ_DBR_CMD_ALL | aq->cq_ci; in mlx5_txpp_cq_arm() 628 uint64_t db_be = rte_cpu_to_be_64(((uint64_t)db_hi << 32) | aq->cq->id); in mlx5_txpp_cq_arm() 633 aq->cq_dbrec[MLX5_CQ_ARM_DB] = rte_cpu_to_be_32(db_hi); in mlx5_txpp_cq_arm() 642 aq->arm_sn++; in mlx5_txpp_cq_arm()
|
| /f-stack/dpdk/drivers/net/ice/ |
| H A D | ice_dcf.c | 594 hw->avf.aq.num_arq_entries = ICE_DCF_AQ_LEN; in ice_dcf_init_hw() 595 hw->avf.aq.num_asq_entries = ICE_DCF_AQ_LEN; in ice_dcf_init_hw() 596 hw->avf.aq.arq_buf_size = ICE_DCF_AQ_BUF_SZ; in ice_dcf_init_hw() 597 hw->avf.aq.asq_buf_size = ICE_DCF_AQ_BUF_SZ; in ice_dcf_init_hw()
|
| /f-stack/dpdk/drivers/net/i40e/ |
| H A D | i40e_ethdev.c | 1542 hw->aq.fw_maj_ver, hw->aq.fw_min_ver, in eth_i40e_dev_init() 1543 hw->aq.api_maj_ver, hw->aq.api_min_ver, in eth_i40e_dev_init() 5579 hw->aq.asq_last_status); in i40e_vsi_get_bw_config() 6729 hw->aq.asq_last_status); in i40e_dev_handle_aq_msg() 10113 if (hw->aq.fw_maj_ver < 6) in i40e_configure_registers() 10944 hw->aq.asq_last_status); in i40e_config_switch_comp_tc() 10954 hw->aq.asq_last_status); in i40e_config_switch_comp_tc() 10963 hw->aq.asq_last_status); in i40e_config_switch_comp_tc() 11030 hw->aq.asq_last_status); in i40e_vsi_config_tc() 11051 hw->aq.asq_last_status); in i40e_vsi_config_tc() [all …]
|
| H A D | i40e_ethdev.h | 1527 hw->aq.num_arq_entries = I40E_AQ_LEN; in i40e_init_adminq_parameter() 1528 hw->aq.num_asq_entries = I40E_AQ_LEN; in i40e_init_adminq_parameter() 1529 hw->aq.arq_buf_size = I40E_AQ_BUF_SZ; in i40e_init_adminq_parameter() 1530 hw->aq.asq_buf_size = I40E_AQ_BUF_SZ; in i40e_init_adminq_parameter()
|
| H A D | rte_pmd_i40e.c | 342 if (hw->aq.fw_maj_ver < 5 && hw->mac.type != I40E_MAC_X722) { in i40e_vsi_set_tx_loopback() 2585 hw->aq.asq_last_status); in i40e_vsi_update_queue_region_mapping() 2890 i40e_aq_str(hw, hw->aq.asq_last_status)); in i40e_queue_region_dcb_configure()
|
| H A D | i40e_pf.c | 240 hw->aq.asq_last_status); in i40e_pf_host_send_msg_to_vf()
|
| /f-stack/dpdk/drivers/baseband/acc100/ |
| H A D | rte_acc100_pmd.c | 185 uint16_t aq; in updateQtop() local 197 for (aq = 0; aq < ACC100_NUM_AQS; aq++) { in updateQtop() 199 0, qg, aq)); in updateQtop()
|