| /linux-6.15/net/ceph/ |
| H A D | debugfs.c | 69 for (n = rb_first(&map->pg_pools); n; n = rb_next(n)) { in osdmap_show() 92 for (n = rb_first(&map->pg_temp); n; n = rb_next(n)) { in osdmap_show() 103 for (n = rb_first(&map->primary_temp); n; n = rb_next(n)) { in osdmap_show() 110 for (n = rb_first(&map->pg_upmap); n; n = rb_next(n)) { in osdmap_show() 121 for (n = rb_first(&map->pg_upmap_items); n; n = rb_next(n)) { in osdmap_show() 239 for (n = rb_first(&osd->o_requests); n; n = rb_next(n)) { in dump_requests() 265 for (n = rb_first(&osd->o_linger_requests); n; n = rb_next(n)) { in dump_linger_requests() 326 for (n = rb_first(&osd->o_backoffs_by_id); n; n = rb_next(n)) { in dump_backoffs() 352 for (n = rb_first(&osdc->osds); n; n = rb_next(n)) { in osdc_show() 360 for (n = rb_first(&osdc->osds); n; n = rb_next(n)) { in osdc_show() [all …]
|
| H A D | osd_client.c | 1183 for (p = rb_first(&osd->o_requests); p; ) { in DEFINE_RB_FUNCS() 1351 for (n = rb_first(&osd->o_requests); n; ) { in close_osd() 1971 rb_entry(rb_first(&osd->o_backoff_mappings), in DEFINE_RB_FUNCS() 1976 rb_entry(rb_first(&spg->backoffs), in DEFINE_RB_FUNCS() 3447 for (p = rb_first(&osd->o_requests); p; ) { in handle_timeout() 3970 for (n = rb_first(&osd->o_requests); n; ) { in scan_requests() 4056 for (n = rb_first(&osdc->osds); n; ) { in handle_one_map() 4082 for (n = rb_first(need_resend); n; ) { in kick_requests() 4097 for (n = rb_first(need_resend); n; ) { in kick_requests() 4249 for (n = rb_first(&osd->o_requests); n; ) { in kick_osd_requests() [all …]
|
| H A D | osdmap.c | 223 rb_entry(rb_first(root), struct crush_name_node, cn_node); in clear_crush_names() 269 rb_entry(rb_first(&c->choose_args), in clear_choose_args() 762 for (rbp = rb_first(&map->pg_pools); rbp; rbp = rb_next(rbp)) { in ceph_pg_poolid_by_name() 1153 rb_entry(rb_first(&map->pg_temp), in ceph_osdmap_destroy() 1160 rb_entry(rb_first(&map->primary_temp), in ceph_osdmap_destroy() 1167 rb_entry(rb_first(&map->pg_upmap), in ceph_osdmap_destroy() 1174 rb_entry(rb_first(&map->pg_upmap_items), in ceph_osdmap_destroy() 1181 rb_entry(rb_first(&map->pg_pools), in ceph_osdmap_destroy() 2973 struct rb_node *n1 = rb_first(locs1); in ceph_compare_crush_locs() 2974 struct rb_node *n2 = rb_first(locs2); in ceph_compare_crush_locs() [all …]
|
| /linux-6.15/arch/powerpc/kernel/ |
| H A D | eeh_cache.c | 103 n = rb_first(&cache->rb_root); in eeh_addr_cache_print() 218 n = rb_first(&pci_io_addr_cache_root.rb_root); in __eeh_addr_cache_rmv_dev() 270 for (n = rb_first(&pci_io_addr_cache_root.rb_root); n; n = rb_next(n)) { in eeh_addr_cache_show()
|
| /linux-6.15/Documentation/translations/zh_CN/core-api/ |
| H A D | rbtree.rst | 165 struct rb_node *rb_first(struct rb_root *tree); 170 要开始迭代,需要使用一个指向树根的指针调用rb_first()或rb_last(),它将返回一个指向 181 for (node = rb_first(&mytree); node; node = rb_next(node)) 189 时间复杂度为O(logN)的rb_first()的调用,以简单地获取指针,避免了潜在的昂贵的树迭代。
|
| /linux-6.15/scripts/gdb/linux/ |
| H A D | rbtree.py | 25 def rb_first(root): function 122 result = rb_first(root)
|
| /linux-6.15/lib/ |
| H A D | rbtree_test.c | 198 for (rb = rb_first(&root.rb_root); rb; rb = rb_next(rb)) { in check() 224 for (rb = rb_first(&root.rb_root); rb; rb = rb_next(rb)) { in check_augmented() 291 for (node = rb_first(&root.rb_root); node; node = rb_next(node)) in basic_check() 305 node = rb_first(&root.rb_root); in basic_check()
|
| H A D | rbtree.c | 466 struct rb_node *rb_first(const struct rb_root *root) in rb_first() function 477 EXPORT_SYMBOL(rb_first);
|
| /linux-6.15/drivers/base/regmap/ |
| H A D | regcache-rbtree.c | 148 for (node = rb_first(&rbtree_ctx->root); node != NULL; in rbtree_show() 225 next = rb_first(&rbtree_ctx->root); in regcache_rbtree_exit() 476 for (node = rb_first(&rbtree_ctx->root); node; node = rb_next(node)) { in regcache_rbtree_sync() 518 for (node = rb_first(&rbtree_ctx->root); node; node = rb_next(node)) { in regcache_rbtree_drop()
|
| /linux-6.15/tools/perf/ui/gtk/ |
| H A D | hists.c | 101 bool has_single_node = (rb_first(root) == rb_last(root)); in perf_gtk__add_callchain_flat() 103 for (nd = rb_first(root); nd; nd = rb_next(nd)) { in perf_gtk__add_callchain_flat() 165 for (nd = rb_first(root); nd; nd = rb_next(nd)) { in perf_gtk__add_callchain_folded() 224 bool has_single_node = (rb_first(root) == rb_last(root)); in perf_gtk__add_callchain_graph() 226 for (nd = rb_first(root); nd; nd = rb_next(nd)) { in perf_gtk__add_callchain_graph()
|
| /linux-6.15/fs/ocfs2/ |
| H A D | reservations.c | 83 node = rb_first(&resmap->m_reservations); in ocfs2_dump_resv() 138 node = rb_first(&resmap->m_reservations); in ocfs2_check_resmap() 363 node = rb_first(&resmap->m_reservations); in ocfs2_find_resv_lhs() 522 next = rb_first(root); in __ocfs2_resv_find_window()
|
| /linux-6.15/kernel/trace/ |
| H A D | trace_stat.c | 185 node = rb_first(&session->stat_root); in stat_seq_start() 200 return rb_first(&session->stat_root); in stat_seq_next()
|
| /linux-6.15/tools/perf/util/ |
| H A D | callchain.c | 426 n = rb_first(&node->rb_root_in); in __sort_chain_flat() 457 n = rb_first(&node->rb_root_in); in __sort_chain_graph_abs() 488 n = rb_first(&node->rb_root_in); in __sort_chain_graph_rel() 553 n = rb_first(&new->rb_root_in); in create_child() 1052 n = rb_first(&src->rb_root_in); in merge_chain_branch() 1295 n = rb_first(&node->rb_root_in); in callchain_node_branch_counts_cumul() 1503 n = rb_first(&node->rb_root_in); in free_callchain_node() 1528 n = rb_first(&node->rb_root_in); in decay_callchain_node() 1755 struct rb_node *rb_node = rb_first(root); in count_callchain_hits()
|
| /linux-6.15/fs/proc/ |
| H A D | nommu.c | 86 for (p = rb_first(&nommu_region_tree); p; p = rb_next(p)) in nommu_region_list_start()
|
| /linux-6.15/drivers/android/ |
| H A D | binder_alloc.c | 439 for (n = rb_first(&alloc->allocated_buffers); n; n = rb_next(n)) { in debug_no_space_locked() 448 for (n = rb_first(&alloc->free_buffers); n; n = rb_next(n)) { in debug_no_space_locked() 488 for (n = rb_first(&alloc->allocated_buffers); n != NULL; in debug_low_async_space_locked() 974 while ((n = rb_first(&alloc->allocated_buffers))) { in binder_alloc_deferred_release() 1047 for (n = rb_first(&alloc->allocated_buffers); n; n = rb_next(n)) { in binder_alloc_print_allocated() 1106 for (n = rb_first(&alloc->allocated_buffers); n != NULL; n = rb_next(n)) in binder_alloc_get_allocated_count()
|
| /linux-6.15/net/netfilter/ |
| H A D | nf_conncount.c | 468 for (node = rb_first(root); node != NULL; node = rb_next(node)) { in tree_gc_worker() 483 node = rb_first(root); in tree_gc_worker() 568 while ((node = rb_first(r)) != NULL) { in destroy_tree()
|
| /linux-6.15/fs/btrfs/ |
| H A D | ref-verify.c | 242 while ((n = rb_first(&be->roots))) { in free_block_entry() 248 while((n = rb_first(&be->refs))) { in free_block_entry() 644 for (n = rb_first(&be->refs); n; n = rb_next(n)) { in dump_block_entry() 652 for (n = rb_first(&be->roots); n; n = rb_next(n)) { in dump_block_entry() 915 while ((n = rb_first(&fs_info->block_tree))) { in btrfs_free_ref_cache()
|
| /linux-6.15/tools/perf/ui/stdio/ |
| H A D | hist.c | 129 node = rb_first(root); in __callchain__fprintf_graph() 234 node = rb_first(root); in callchain__fprintf_graph() 315 struct rb_node *rb_node = rb_first(tree); in callchain__fprintf_flat() 366 struct rb_node *rb_node = rb_first(tree); in callchain__fprintf_folded()
|
| /linux-6.15/arch/x86/events/intel/ |
| H A D | uncore_discovery.c | 199 for (node = rb_first(root); node; node = rb_next(node)) { in uncore_find_unit() 401 node = rb_first(&type->units); in intel_uncore_clear_discovery_tables() 708 for (node = rb_first(&discovery_tables); node; node = rb_next(node)) { in intel_uncore_generic_init_uncores()
|
| /linux-6.15/drivers/infiniband/hw/mlx4/ |
| H A D | cm.c | 549 while (rb_first(sl_id_map)) { in mlx4_ib_cm_paravirt_clean() 551 rb_entry(rb_first(sl_id_map), in mlx4_ib_cm_paravirt_clean() 560 nd = rb_first(sl_id_map); in mlx4_ib_cm_paravirt_clean()
|
| /linux-6.15/tools/perf/ui/browsers/ |
| H A D | map.c | 120 for (nd = rb_first(mb.b.entries); nd; nd = rb_next(nd)) { in map__browse()
|
| H A D | hists.c | 182 for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) { in callchain_node__count_rows_rb_tree() 265 for (nd = rb_first(chain); nd; nd = rb_next(nd)) { in callchain__count_rows() 331 struct rb_node *nd = rb_first(&node->rb_root); in callchain_node__init_have_children_rb_tree() 333 for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) { in callchain_node__init_have_children_rb_tree() 370 struct rb_node *nd = rb_first(root); in callchain__init_have_children() 373 for (nd = rb_first(root); nd; nd = rb_next(nd)) { in callchain__init_have_children() 557 for (nd = rb_first(chain); nd; nd = rb_next(nd)) { in callchain__set_folding() 921 node = rb_first(root); in hist_browser__show_callchain_flat() 1025 node = rb_first(root); in hist_browser__show_callchain_folded() 1114 node = rb_first(root); in hist_browser__show_callchain_graph() [all …]
|
| /linux-6.15/tools/perf/tests/ |
| H A D | hists_cumulate.c | 216 for (node = rb_first(root), i = 0; in do_test() 238 cnode = rb_entry(rb_first(root), struct callchain_node, rb_node); in do_test()
|
| /linux-6.15/drivers/vfio/ |
| H A D | vfio_iommu_type1.c | 264 for (p = rb_first(&dma->pfn_list); p; p = rb_next(p)) { in vfio_dma_populate_bitmap() 276 for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) { in vfio_iommu_populate_bitmap_full() 287 for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) { in vfio_dma_bitmap_alloc_all() 312 for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) { in vfio_dma_bitmap_free_all() 1231 for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) { in vfio_iova_dirty_bitmap() 1707 n = rb_first(&iommu->dma_list); in vfio_iommu_replay() 1785 for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) { in vfio_iommu_replay() 2366 while ((node = rb_first(&iommu->dma_list))) in vfio_iommu_unmap_unpin_all() 2374 n = rb_first(&iommu->dma_list); in vfio_iommu_unmap_unpin_reaccount() 2381 p = rb_first(&dma->pfn_list); in vfio_iommu_unmap_unpin_reaccount()
|
| /linux-6.15/fs/xfs/scrub/ |
| H A D | bitmap.c | 63 for ((bn) = rb_entry_safe(rb_first(&(bitmap)->xb_root.rb_root), \ in INTERVAL_TREE_DEFINE() 338 for ((bn) = rb_entry_safe(rb_first(&(bitmap)->xb_root.rb_root), \ in INTERVAL_TREE_DEFINE()
|