| /linux-6.15/lib/ |
| H A D | test_xarray.c | 63 XA_BUG_ON(xa, xa_load(xa, index) != NULL); in xa_erase_index() 153 void *entry = xa_load(xa, j); in check_xa_load() 164 void *entry = xa_load(xa, j); in check_xa_load() 344 XA_BUG_ON(xa, xa_load(xa, 1) != NULL); in check_xa_shrink() 439 XA_BUG_ON(xa, xa_load(xa, i) != FIVE); in check_cmpxchg_order() 452 XA_BUG_ON(xa, xa_load(xa, i) != FIVE); in check_cmpxchg_order() 464 XA_BUG_ON(xa, xa_load(xa, i) != NULL); in check_cmpxchg_order() 489 XA_BUG_ON(xa, xa_load(xa, 12345678)); in check_reserve() 660 XA_BUG_ON(xa, xa_load(xa, 2) != NULL); in check_multi_store() 670 XA_BUG_ON(xa, xa_load(xa, 2) != NULL); in check_multi_store() [all …]
|
| /linux-6.15/drivers/gpu/drm/imagination/ |
| H A D | pvr_free_list.h | 143 free_list = pvr_free_list_get(xa_load(&pvr_file->free_list_handles, handle)); in pvr_free_list_lookup() 172 free_list = xa_load(&pvr_dev->free_list_ids, id); in pvr_free_list_lookup_id()
|
| H A D | pvr_context.h | 165 ctx = pvr_context_get(xa_load(&pvr_file->ctx_handles, handle)); in pvr_context_lookup() 195 ctx = xa_load(&pvr_dev->ctx_ids, id); in pvr_context_lookup_id()
|
| H A D | pvr_hwrt.h | 105 hwrt = xa_load(&pvr_file->hwrt_handles, handle); in pvr_hwrt_dataset_lookup()
|
| /linux-6.15/drivers/net/ethernet/intel/ice/ |
| H A D | ice_adapter.c | 75 adapter = xa_load(&ice_adapters, index); in ice_adapter_get() 108 adapter = xa_load(&ice_adapters, index); in ice_adapter_put()
|
| H A D | ice_eswitch.c | 184 repr = xa_load(&pf->eswitch.reprs, *repr_id); in ice_eswitch_update_repr() 578 struct ice_repr *repr = xa_load(&pf->eswitch.reprs, vf->repr_id); in ice_eswitch_detach_vf() 596 struct ice_repr *repr = xa_load(&pf->eswitch.reprs, sf->repr_id); in ice_eswitch_detach_sf() 620 repr = xa_load(&eswitch->reprs, le16_to_cpu(desc->src_vsi)); in ice_eswitch_get_target()
|
| H A D | ice_eswitch_br.c | 312 port = xa_load(&bridge->ports, vsi_idx); in ice_esw_br_port_vlan_lookup() 318 vlan = xa_load(&port->vlans, vid); in ice_esw_br_port_vlan_lookup() 742 port = xa_load(&bridge->ports, vsi_idx); in ice_eswitch_br_port_vlan_add() 753 vlan = xa_load(&port->vlans, vid); in ice_eswitch_br_port_vlan_add() 777 port = xa_load(&bridge->ports, vsi_idx); in ice_eswitch_br_port_vlan_del() 781 vlan = xa_load(&port->vlans, vid); in ice_eswitch_br_port_vlan_del()
|
| /linux-6.15/Documentation/translations/zh_CN/core-api/ |
| H A D | xarray.rst | 63 然后你可以用xa_store()来设置条目,用xa_load()来获取条目。xa_store将用新的条目覆盖任 153 * xa_load() 195 如果你想利用锁来保护你存储在XArray中的数据结构,你可以在调用xa_load()之前调用xa_lock(),然后在
|
| /linux-6.15/drivers/iommu/iommufd/ |
| H A D | device.c | 82 igroup = xa_load(&ictx->groups, id); in iommufd_get_group() 312 attach = xa_load(&igroup->pasid_attach, pasid); in iommufd_group_device_num() 359 return !xa_load(&igroup->pasid_attach, pasid); in iommufd_group_first_attach() 395 attach = xa_load(&idev->igroup->pasid_attach, pasid); in iommufd_device_is_attached() 396 return xa_load(&attach->device_array, idev->obj.id); in iommufd_device_is_attached() 420 attach = xa_load(&igroup->pasid_attach, IOMMU_NO_PASID); in iommufd_hwpt_pasid_compat() 646 attach = xa_load(&igroup->pasid_attach, pasid); in iommufd_hw_pagetable_detach() 691 attach = xa_load(&igroup->pasid_attach, IOMMU_NO_PASID); in iommufd_group_remove_reserved_iova() 708 attach = xa_load(&igroup->pasid_attach, IOMMU_NO_PASID); in iommufd_group_do_replace_reserved_iova() 744 attach = xa_load(&igroup->pasid_attach, pasid); in iommufd_device_do_replace()
|
| /linux-6.15/fs/xfs/libxfs/ |
| H A D | xfs_group.c | 40 xg = xa_load(&mp->m_groups[type].xa, index); in xfs_group_get() 81 xg = xa_load(&mp->m_groups[type].xa, index); in xfs_group_grab()
|
| /linux-6.15/drivers/gpu/drm/tegra/ |
| H A D | uapi.c | 170 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_close() 200 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_map() 281 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_unmap()
|
| /linux-6.15/drivers/firmware/arm_scmi/ |
| H A D | perf.c | 596 opp = xa_load(&dom->opps_by_lvl, min_perf); in scmi_perf_limits_set() 604 opp = xa_load(&dom->opps_by_lvl, max_perf); in scmi_perf_limits_set() 675 opp = xa_load(&dom->opps_by_idx, *min_perf); in scmi_perf_limits_get() 681 opp = xa_load(&dom->opps_by_idx, *max_perf); in scmi_perf_limits_get() 745 opp = xa_load(&dom->opps_by_lvl, level); in scmi_perf_level_set() 808 opp = xa_load(&dom->opps_by_idx, *level); in scmi_perf_level_get() 991 opp = xa_load(&dom->opps_by_idx, level); in scmi_dvfs_freq_get() 1137 opp = xa_load(&dom->opps_by_lvl, index); in scmi_perf_xlate_opp_to_freq() 1143 opp = xa_load(&dom->opps_by_idx, index); in scmi_perf_xlate_opp_to_freq()
|
| /linux-6.15/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/ |
| H A D | dr_domain.c | 81 recalc_cs_ft = xa_load(&dmn->csum_fts_xa, vport_num); in mlx5dr_domain_get_recalc_cs_ft_addr() 329 vport_caps = xa_load(&caps->vports.vports_caps_xa, vport); in mlx5dr_domain_get_vport_cap() 546 peer = xa_load(&dmn->peer_dmn_xa, peer_vhca_id); in mlx5dr_domain_set_peer() 552 peer = xa_load(&dmn->peer_dmn_xa, peer_vhca_id); in mlx5dr_domain_set_peer()
|
| /linux-6.15/include/linux/ |
| H A D | phy_link_topology.h | 55 pdn = xa_load(&topo->phys, phyindex); in phy_link_topo_get_phy()
|
| /linux-6.15/drivers/infiniband/hw/cxgb4/ |
| H A D | ev.c | 127 qhp = xa_load(&dev->qps, CQE_QPID(err_cqe)); in c4iw_ev_dispatch() 228 chp = xa_load(&dev->cqs, qid); in c4iw_ev_handler()
|
| /linux-6.15/fs/smb/server/mgmt/ |
| H A D | user_session.c | 148 entry = xa_load(&sess->rpc_handle_list, id); in ksmbd_session_rpc_method() 277 sess = xa_load(&conn->sessions, id); in is_ksmbd_session_in_connection() 293 sess = xa_load(&conn->sessions, id); in ksmbd_session_lookup()
|
| /linux-6.15/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | pagealloc.c | 106 root = xa_load(&dev->priv.page_root_xa, function); in page_root_per_function() 177 root = xa_load(&dev->priv.page_root_xa, function); in find_fw_page() 259 root = xa_load(&dev->priv.page_root_xa, fwp->function); in free_fwp() 443 root = xa_load(&dev->priv.page_root_xa, function); in release_all_pages() 501 root = xa_load(&dev->priv.page_root_xa, get_function(func_id, ec_function)); in reclaim_pages_cmd()
|
| H A D | eq.c | 814 irq = xa_load(&table->comp_irqs, vecidx); in comp_irq_release_pci() 862 irq = xa_load(&table->comp_irqs, vecidx); in comp_irq_release_sf() 1009 irq = xa_load(&table->comp_irqs, vecidx); in create_comp_eq() 1055 eq = xa_load(&table->comp_eqs, vecidx); in mlx5_comp_eqn_get() 1086 eq = xa_load(&table->comp_eqs, vector); in mlx5_comp_irqn_get() 1103 eq = xa_load(&table->comp_eqs, vector); in mlx5_comp_irq_get_affinity_mask()
|
| /linux-6.15/drivers/net/ethernet/mellanox/mlx5/core/en/ |
| H A D | mapping.c | 116 mi = xa_load(&ctx->xarray, index); in mapping_remove() 139 mi = xa_load(&ctx->xarray, index); in mapping_find()
|
| /linux-6.15/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| H A D | vport.c | 75 entry = xa_load(&ctx->vports.vport_gvmi_xa, vport); in mlx5hws_vport_get_gvmi()
|
| /linux-6.15/drivers/infiniband/sw/rxe/ |
| H A D | rxe_mr.c | 255 page = xa_load(&mr->page_list, index); in rxe_mr_copy_xarray() 452 page = xa_load(&mr->page_list, index); in rxe_flush_pmem_iova() 501 page = xa_load(&mr->page_list, index); in rxe_mr_do_atomic_op() 563 page = xa_load(&mr->page_list, index); in rxe_mr_do_atomic_write()
|
| /linux-6.15/drivers/cxl/core/ |
| H A D | cdat.c | 154 dent = xa_load(dsmas_xa, dslbis->handle); in cdat_dslbis_handler() 672 perf_ctx = xa_load(usp_xa, index); in cxl_endpoint_gather_bandwidth() 798 us_ctx = xa_load(res_xa, us_index); in DEFINE_FREE() 882 hb_ctx = xa_load(hb_xa, hb_index); in cxl_rp_gather_bandwidth() 931 mw_ctx = xa_load(mw_xa, mw_index); in cxl_hb_gather_bandwidth()
|
| /linux-6.15/drivers/base/ |
| H A D | auxiliary_sysfs.c | 102 struct auxiliary_irq_info *info __free(kfree) = xa_load(&auxdev->sysfs.irqs, irq); in auxiliary_device_sysfs_irq_remove()
|
| /linux-6.15/security/apparmor/ |
| H A D | secid.c | 47 return xa_load(&aa_secids, secid); in aa_secid_to_label()
|
| /linux-6.15/mm/ |
| H A D | readahead.c | 252 struct folio *folio = xa_load(&mapping->i_pages, index + i); in page_cache_ra_unbounded() 761 struct folio *folio = xa_load(&mapping->i_pages, index); in readahead_expand() 790 struct folio *folio = xa_load(&mapping->i_pages, index); in readahead_expand()
|