| /dpdk/drivers/net/tap/ |
| H A D | tap_bpf_api.c | 120 bzero(&attr, sizeof(attr)); in bpf_load() 121 attr.prog_type = type; in bpf_load() 126 attr.log_level = 0; in bpf_load() 127 attr.kern_version = 0; in bpf_load() 129 return sys_bpf(BPF_PROG_LOAD, &attr, sizeof(attr)); in bpf_load() 153 bzero(&attr, sizeof(attr)); in tap_flow_bpf_rss_map_create() 159 return sys_bpf(BPF_MAP_CREATE, &attr, sizeof(attr)); in tap_flow_bpf_rss_map_create() 181 bzero(&attr, sizeof(attr)); in tap_flow_bpf_update_rss_elem() 184 attr.map_fd = fd; in tap_flow_bpf_update_rss_elem() 187 attr.flags = BPF_ANY; in tap_flow_bpf_update_rss_elem() [all …]
|
| H A D | tap_netlink.c | 146 tap_nl_attr_next(const struct nlattr *attr) in tap_nl_attr_next() argument 149 + NLMSG_ALIGN(attr->nla_len)); in tap_nl_attr_next() 157 if (attr->nla_len < sizeof(struct nlattr)) in tap_nl_attr_ok() 159 if ((int)attr->nla_len > len) in tap_nl_attr_ok() 169 const struct nlattr *attr; in tap_nl_dump_ext_ack() local 180 for (attr = tap_nl_attr_first(nh, hlen); in tap_nl_dump_ext_ack() 181 tap_nl_attr_ok(attr, tail - (const char *)attr); in tap_nl_dump_ext_ack() 182 attr = tap_nl_attr_next(attr)) { in tap_nl_dump_ext_ack() 183 uint16_t type = attr->nla_type & NLA_TYPE_MASK; in tap_nl_dump_ext_ack() 186 const char *msg = (const char *)attr in tap_nl_dump_ext_ack() [all …]
|
| /dpdk/drivers/net/fm10k/base/ |
| H A D | fm10k_tlv.c | 41 u32 *attr; in fm10k_tlv_attr_put_null_string() local 100 attr++; in fm10k_tlv_attr_get_null_string() 122 u32 *attr; in fm10k_tlv_attr_put_mac_vlan() local 212 u32 *attr; in fm10k_tlv_attr_put_value() local 264 *(u64 *)value = ((u64)attr[2] << 32) | attr[1]; in fm10k_tlv_attr_get_value() 290 u32 *attr; in fm10k_tlv_attr_put_le_struct() local 341 attr++; in fm10k_tlv_attr_get_le_struct() 362 u32 *attr; in fm10k_tlv_attr_nest_start() local 443 attr++; in fm10k_tlv_attr_validate() 528 attr++; in fm10k_tlv_attr_parse() [all …]
|
| H A D | fm10k_tlv.h | 85 const struct fm10k_tlv_attr *attr; member 90 #define FM10K_MSG_HANDLER(id, attr, func) { id, attr, func } argument 114 #define fm10k_tlv_attr_get_u8(attr, ptr) \ argument 115 fm10k_tlv_attr_get_value(attr, ptr, sizeof(u8)) 116 #define fm10k_tlv_attr_get_u16(attr, ptr) \ argument 118 #define fm10k_tlv_attr_get_u32(attr, ptr) \ argument 120 #define fm10k_tlv_attr_get_u64(attr, ptr) \ argument 122 #define fm10k_tlv_attr_get_s8(attr, ptr) \ argument 124 #define fm10k_tlv_attr_get_s16(attr, ptr) \ argument 126 #define fm10k_tlv_attr_get_s32(attr, ptr) \ argument [all …]
|
| /dpdk/drivers/common/mlx5/ |
| H A D | mlx5_devx_cmds.c | 381 if (attr->crypto_en) { in mlx5_devx_cmd_mkey_create() 935 attr->reg_c_preserve = in mlx5_devx_cmd_query_hca_attr() 967 if (attr->crypto) in mlx5_devx_cmd_query_hca_attr() 987 if (attr->qos.sup) { in mlx5_devx_cmd_query_hca_attr() 1025 (ctx, &attr->flex); in mlx5_devx_cmd_query_hca_attr() 1029 if (attr->vdpa.valid) in mlx5_devx_cmd_query_hca_attr() 1118 if (attr->roce) { in mlx5_devx_cmd_query_hca_attr() 1129 if (attr->eth_virt && in mlx5_devx_cmd_query_hca_attr() 1133 attr->eth_virt = 0; in mlx5_devx_cmd_query_hca_attr() 2070 switch (attr->type) { in mlx5_devx_cmd_modify_virtq() [all …]
|
| H A D | mlx5_common_devx.c | 124 attr->q_umem_valid = 1; in mlx5_devx_cq_create() 126 attr->q_umem_offset = 0; in mlx5_devx_cq_create() 127 attr->db_umem_valid = 1; in mlx5_devx_cq_create() 128 attr->db_umem_id = attr->q_umem_id; in mlx5_devx_cq_create() 130 attr->eqn = eqn; in mlx5_devx_cq_create() 247 attr->wq_attr.dbr_umem_id = attr->wq_attr.wq_umem_id; in mlx5_devx_sq_create() 388 attr->wq_umem_offset = 0; in mlx5_devx_qp_create() 389 attr->dbr_umem_valid = 1; in mlx5_devx_qp_create() 390 attr->dbr_umem_id = attr->wq_umem_id; in mlx5_devx_qp_create() 647 attr->flush_in_error_en = 0; in mlx5_devx_rq_shared_create() [all …]
|
| /dpdk/drivers/vdpa/mlx5/ |
| H A D | mlx5_vdpa_virtq.c | 179 attr.hw_available_index, attr.hw_used_index); in mlx5_vdpa_virtq_query() 181 attr.hw_available_index, in mlx5_vdpa_virtq_query() 182 attr.hw_used_index); in mlx5_vdpa_virtq_query() 311 attr.umems[i].offset = 0; in mlx5_vdpa_virtq_setup() 321 attr.desc_addr = gpa; in mlx5_vdpa_virtq_setup() 328 attr.used_addr = gpa; in mlx5_vdpa_virtq_setup() 335 attr.available_addr = gpa; in mlx5_vdpa_virtq_setup() 350 attr.q_size = vq.size; in mlx5_vdpa_virtq_setup() 351 attr.mkey = priv->gpa_mkey_index; in mlx5_vdpa_virtq_setup() 353 attr.queue_index = index; in mlx5_vdpa_virtq_setup() [all …]
|
| H A D | mlx5_vdpa_lm.c | 14 struct mlx5_devx_virtq_attr attr = { in mlx5_vdpa_logging_enable() local 21 attr.queue_index = i; in mlx5_vdpa_logging_enable() 26 &attr)) { in mlx5_vdpa_logging_enable() 39 struct mlx5_devx_virtq_attr attr = { in mlx5_vdpa_dirty_bitmap_set() local 54 attr.dirty_bitmap_mkey = priv->lm_mr.lkey; in mlx5_vdpa_dirty_bitmap_set() 56 attr.queue_index = i; in mlx5_vdpa_dirty_bitmap_set() 60 &attr)) { in mlx5_vdpa_dirty_bitmap_set()
|
| H A D | mlx5_vdpa_steer.c | 65 struct mlx5_devx_rqt_attr *attr = rte_zmalloc(__func__, sizeof(*attr) in mlx5_vdpa_rqt_prepare() local 71 if (!attr) { in mlx5_vdpa_rqt_prepare() 79 attr->rq_list[k] = priv->virtqs[i].virtq->id; in mlx5_vdpa_rqt_prepare() 88 attr->rq_list[k] = attr->rq_list[j]; in mlx5_vdpa_rqt_prepare() 89 attr->rq_type = MLX5_INLINE_Q_TYPE_VIRTQ; in mlx5_vdpa_rqt_prepare() 90 attr->rqt_max_size = rqt_n; in mlx5_vdpa_rqt_prepare() 91 attr->rqt_actual_size = rqt_n; in mlx5_vdpa_rqt_prepare() 94 attr); in mlx5_vdpa_rqt_prepare() 100 ret = mlx5_devx_cmd_modify_rqt(priv->steer.rqt, attr); in mlx5_vdpa_rqt_prepare() 104 rte_free(attr); in mlx5_vdpa_rqt_prepare()
|
| H A D | mlx5_vdpa_event.c | 90 struct mlx5_devx_cq_attr attr = { in mlx5_vdpa_cq_create() local 98 &attr, SOCKET_ID_ANY); in mlx5_vdpa_cq_create() 470 pthread_attr_t attr; in mlx5_vdpa_cqe_event_setup() local 479 pthread_attr_init(&attr); in mlx5_vdpa_cqe_event_setup() 581 struct mlx5_devx_qp_attr attr = {0}; in mlx5_vdpa_event_qp_create() local 587 attr.pd = priv->cdev->pdn; in mlx5_vdpa_event_qp_create() 588 attr.ts_format = in mlx5_vdpa_event_qp_create() 596 attr.cqn = eqp->cq.cq_obj.cq->id; in mlx5_vdpa_event_qp_create() 600 attr.ts_format = in mlx5_vdpa_event_qp_create() 603 attr.num_of_receive_wqes * in mlx5_vdpa_event_qp_create() [all …]
|
| /dpdk/drivers/net/hinic/base/ |
| H A D | hinic_pmd_hwif.c | 164 struct hinic_func_attr *attr = &hwif->attr; in set_ppf() local 182 if (attr->ppf_idx == attr->func_global_idx) in set_ppf() 183 attr->func_type = TYPE_PPF; in set_ppf() 360 return hwif->attr.func_type; in hinic_func_type() 367 return hwif->attr.ppf_idx; in hinic_ppf_idx() 378 return hwif->attr.num_dma_attr; in hinic_dma_attr_entry_num() 456 hwif->attr.port_to_port_idx, hwif->attr.pci_intf_idx, in hinic_parse_hwif_attr() 457 hwif->attr.vf_in_pf, hwif->attr.ppf_idx, in hinic_parse_hwif_attr() 458 hwif->attr.global_vf_id_of_pf, hwif->attr.func_type); in hinic_parse_hwif_attr() 460 hwif->attr.num_aeqs, hwif->attr.num_ceqs, in hinic_parse_hwif_attr() [all …]
|
| H A D | hinic_pmd_hwif.h | 10 #define HINIC_HWIF_NUM_AEQS(hwif) ((hwif)->attr.num_aeqs) 11 #define HINIC_HWIF_NUM_CEQS(hwif) ((hwif)->attr.num_ceqs) 12 #define HINIC_HWIF_NUM_IRQS(hwif) ((hwif)->attr.num_irqs) 13 #define HINIC_HWIF_GLOBAL_IDX(hwif) ((hwif)->attr.func_global_idx) 14 #define HINIC_HWIF_GLOBAL_VF_OFFSET(hwif) ((hwif)->attr.global_vf_id_of_pf) 15 #define HINIC_HWIF_PPF_IDX(hwif) ((hwif)->attr.ppf_idx) 16 #define HINIC_PCI_INTF_IDX(hwif) ((hwif)->attr.pci_intf_idx) 18 #define HINIC_FUNC_TYPE(dev) ((dev)->hwif->attr.func_type) 78 struct hinic_func_attr attr; member
|
| /dpdk/drivers/crypto/dpaa2_sec/mc/ |
| H A D | dpseci.c | 289 struct dpseci_attr *attr) in dpseci_get_attributes() argument 307 attr->id = le32_to_cpu(rsp_params->id); in dpseci_get_attributes() 372 struct dpseci_rx_queue_attr *attr) in dpseci_get_rx_queue() argument 397 attr->dest_cfg.dest_type = in dpseci_get_rx_queue() 400 attr->order_preservation_en = in dpseci_get_rx_queue() 422 struct dpseci_tx_queue_attr *attr) in dpseci_get_tx_queue() argument 444 attr->priority = rsp_params->priority; in dpseci_get_tx_queue() 461 struct dpseci_sec_attr *attr) in dpseci_get_sec_attr() argument 480 attr->major_rev = rsp_params->major_rev; in dpseci_get_sec_attr() 482 attr->era = rsp_params->era; in dpseci_get_sec_attr() [all …]
|
| /dpdk/app/test-flow-perf/ |
| H A D | flow_gen.c | 20 fill_attributes(struct rte_flow_attr *attr, in fill_attributes() argument 28 attr->ingress = 1; in fill_attributes() 30 attr->egress = 1; in fill_attributes() 32 attr->transfer = 1; in fill_attributes() 34 attr->group = group; in fill_attributes() 35 attr->priority = rte_rand_max(max_priority); in fill_attributes() 56 struct rte_flow_attr attr; in generate_flow() local 63 memset(&attr, 0, sizeof(struct rte_flow_attr)); in generate_flow() 65 fill_attributes(&attr, flow_attrs, group, max_priority); in generate_flow() 74 flow = rte_flow_create(port_id, &attr, items, actions, error); in generate_flow()
|
| /dpdk/drivers/net/e1000/ |
| H A D | igb_flow.c | 122 if (!attr) { in cons_parse_ntuple_filter() 372 if (!attr->ingress) { in cons_parse_ntuple_filter() 381 if (attr->egress) { in cons_parse_ntuple_filter() 516 if (!attr) { in cons_parse_ethertype_filter() 635 if (attr->egress) { in cons_parse_ethertype_filter() 659 if (attr->group) { in cons_parse_ethertype_filter() 786 if (!attr) { in cons_parse_syn_filter() 942 if (attr->egress) { in cons_parse_syn_filter() 1072 if (!attr) { in cons_parse_flex_filter() 1239 if (attr->egress) { in cons_parse_flex_filter() [all …]
|
| /dpdk/drivers/net/txgbe/ |
| H A D | txgbe_flow.c | 174 if (!attr) { in cons_parse_ntuple_filter() 541 if (attr->egress) { in cons_parse_ntuple_filter() 670 if (!attr) { in cons_parse_ethertype_filter() 782 if (attr->egress) { in cons_parse_ethertype_filter() 806 if (attr->group) { in cons_parse_ethertype_filter() 915 if (!attr) { in cons_parse_syn_filter() 1069 if (attr->egress) { in cons_parse_syn_filter() 1171 if (!attr) { in cons_parse_l2_tn_filter() 1246 if (attr->egress) { in cons_parse_l2_tn_filter() 1574 if (!attr) { in txgbe_parse_fdir_filter_normal() [all …]
|
| /dpdk/app/test/ |
| H A D | test_flow_classify.c | 309 attr.ingress = 1; in test_valid_parameters() 310 attr.priority = 1; in test_valid_parameters() 358 attr.ingress = 1; in test_invalid_patterns() 359 attr.priority = 1; in test_invalid_patterns() 437 attr.ingress = 1; in test_invalid_actions() 438 attr.priority = 1; in test_invalid_actions() 649 attr.ingress = 1; in test_query_udp() 650 attr.priority = 1; in test_query_udp() 716 attr.ingress = 1; in test_query_tcp() 717 attr.priority = 1; in test_query_tcp() [all …]
|
| /dpdk/drivers/regex/mlx5/ |
| H A D | mlx5_regex.c | 74 struct mlx5_hca_attr *attr = &cdev->config.hca_attr; in mlx5_regex_dev_probe() local 78 if ((!attr->regexp_params && !attr->mmo_regex_sq_en && !attr->mmo_regex_qp_en) in mlx5_regex_dev_probe() 79 || attr->regexp_num_of_engines == 0) { in mlx5_regex_dev_probe() 92 priv->mmo_regex_qp_cap = attr->mmo_regex_qp_en; in mlx5_regex_dev_probe() 93 priv->mmo_regex_sq_cap = attr->mmo_regex_sq_en; in mlx5_regex_dev_probe() 96 if (attr->regexp_version == MLX5_RXP_BF2_IDENTIFIER) in mlx5_regex_dev_probe() 113 if (!attr->umr_indirect_mkey_disabled && in mlx5_regex_dev_probe() 114 !attr->umr_modify_entity_size_disabled) in mlx5_regex_dev_probe()
|
| /dpdk/drivers/net/ixgbe/ |
| H A D | ixgbe_flow.c | 208 if (!attr) { in cons_parse_ntuple_filter() 581 if (attr->egress) { in cons_parse_ntuple_filter() 708 if (!attr) { in cons_parse_ethertype_filter() 820 if (attr->egress) { in cons_parse_ethertype_filter() 844 if (attr->group) { in cons_parse_ethertype_filter() 956 if (!attr) { in cons_parse_syn_filter() 1110 if (attr->egress) { in cons_parse_syn_filter() 1215 if (!attr) { in cons_parse_l2_tn_filter() 1290 if (attr->egress) { in cons_parse_l2_tn_filter() 1633 if (!attr) { in ixgbe_parse_fdir_filter_normal() [all …]
|
| /dpdk/drivers/event/dpaa2/ |
| H A D | dpaa2_eventdev_selftest.c | 223 struct event_attr *attr; in update_event_and_validation_attr() local 227 attr->flow_id = flow_id; in update_event_and_validation_attr() 228 attr->event_type = event_type; in update_event_and_validation_attr() 230 attr->sched_type = sched_type; in update_event_and_validation_attr() 231 attr->queue = queue; in update_event_and_validation_attr() 232 attr->port = port; in update_event_and_validation_attr() 233 attr->seq = seq; in update_event_and_validation_attr() 317 struct event_attr *attr; in validate_event() local 322 attr->flow_id, ev->flow_id); in validate_event() 334 attr->queue, ev->queue_id); in validate_event() [all …]
|
| /dpdk/drivers/bus/fslmc/mc/ |
| H A D | dpdmai.c | 282 struct dpdmai_attr *attr) in dpdmai_get_attributes() argument 300 attr->id = le32_to_cpu(rsp_params->id); in dpdmai_get_attributes() 301 attr->num_of_priorities = rsp_params->num_of_priorities; in dpdmai_get_attributes() 302 attr->num_of_queues = rsp_params->num_of_queues; in dpdmai_get_attributes() 303 attr->options = le32_to_cpu(rsp_params->options); in dpdmai_get_attributes() 370 struct dpdmai_rx_queue_attr *attr) in dpdmai_get_rx_queue() argument 392 attr->user_ctx = le64_to_cpu(rsp_params->user_ctx); in dpdmai_get_rx_queue() 393 attr->fqid = le32_to_cpu(rsp_params->fqid); in dpdmai_get_rx_queue() 394 attr->dest_cfg.dest_id = le32_to_cpu(rsp_params->dest_id); in dpdmai_get_rx_queue() 420 struct dpdmai_tx_queue_attr *attr) in dpdmai_get_tx_queue() argument [all …]
|
| /dpdk/drivers/net/mlx5/ |
| H A D | mlx5_flow_hw.c | 563 if (attr->transfer) in flow_hw_actions_translate() 928 attr.ingress = 1; in flow_hw_actions_construct() 930 attr.egress = 1; in flow_hw_actions_construct() 931 attr.ingress = 0; in flow_hw_actions_construct() 933 attr.ingress = 1; in flow_hw_actions_construct() 1369 &attr, in flow_hw_q_flow_flush() 1651 at->attr = *attr; in flow_hw_actions_template_create() 1746 it->attr = *attr; in flow_hw_pattern_template_create() 1867 if (attr->group) { in flow_hw_grp_create_cb() 1956 attr->egress) || in flow_hw_grp_match_cb() [all …]
|
| H A D | mlx5_dr.c | 26 struct mlx5dr_context_attr *attr) in mlx5dr_context_open() argument 29 (void)attr; in mlx5dr_context_open() 56 struct mlx5dr_table_attr *attr) in mlx5dr_table_create() argument 59 (void)attr; in mlx5dr_table_create() 125 struct mlx5dr_matcher_attr *attr __rte_unused) in mlx5dr_matcher_create() 166 struct mlx5dr_rule_attr *attr __rte_unused, in mlx5dr_rule_create() 182 struct mlx5dr_rule_attr *attr __rte_unused) in mlx5dr_rule_destroy()
|
| H A D | mlx5_flow_flex.c | 380 if (!(attr->header_length_mode & in mlx5_flex_translate_length() 397 if (!(attr->header_length_mode & in mlx5_flex_translate_length() 422 if (!(attr->header_length_mode & in mlx5_flex_translate_length() 584 struct mlx5_hca_flex_attr *attr, in mlx5_flex_cover_sample() argument 594 if (!(attr->sample_offset_mode & in mlx5_flex_cover_sample() 615 if (!(attr->sample_offset_mode & in mlx5_flex_cover_sample() 628 if (!(attr->sample_offset_mode & in mlx5_flex_cover_sample() 828 struct mlx5_hca_flex_attr *attr, in mlx5_flex_alloc_sample() argument 935 attr, error); in mlx5_flex_translate_sample() 944 fl, attr, error); in mlx5_flex_translate_sample() [all …]
|
| /dpdk/drivers/net/igc/ |
| H A D | igc_flow.c | 71 if (!attr) in igc_parse_attribute() 74 if (attr->group) in igc_parse_attribute() 76 RTE_FLOW_ERROR_TYPE_ATTR_GROUP, attr, in igc_parse_attribute() 79 if (attr->egress) in igc_parse_attribute() 84 if (attr->transfer) in igc_parse_attribute() 89 if (!attr->ingress) in igc_parse_attribute() 94 if (attr->priority == 0) in igc_parse_attribute() 113 if (attr->priority > 1) { in igc_parse_attribute() 266 const struct rte_flow_attr *attr, in igc_parse_flow() argument 716 const struct rte_flow_attr *attr, in igc_flow_create() argument [all …]
|