| /linux-6.15/include/linux/ |
| H A D | rbtree.h | 110 bool leftmost) in rb_insert_color_cached() argument 112 if (leftmost) in rb_insert_color_cached() 121 struct rb_node *leftmost = NULL; in rb_erase_cached() local 124 leftmost = root->rb_leftmost = rb_next(node); in rb_erase_cached() 128 return leftmost; in rb_erase_cached() 170 bool leftmost = true; in rb_add_cached() local 178 leftmost = false; in rb_add_cached() 183 rb_insert_color_cached(node, tree, leftmost); in rb_add_cached() 185 return leftmost ? node : NULL; in rb_add_cached() 226 bool leftmost = true; in rb_find_add_cached() local [all …]
|
| H A D | interval_tree_generic.h | 44 bool leftmost = true; \ 55 leftmost = false; \ 62 leftmost, &ITPREFIX ## _augment); \ 118 ITSTRUCT *node, *leftmost; \ 140 leftmost = rb_entry(root->rb_leftmost, ITSTRUCT, ITRB); \ 141 if (ITSTART(leftmost) > last) \
|
| H A D | timerqueue.h | 25 struct rb_node *leftmost = rb_first_cached(&head->rb_root); in timerqueue_getnext() local 27 return rb_entry_safe(leftmost, struct timerqueue_node, node); in timerqueue_getnext()
|
| H A D | rbtree_augmented.h | 70 bool leftmost = true; in rb_add_augmented_cached() local 78 leftmost = false; in rb_add_augmented_cached() 84 rb_insert_augmented_cached(node, tree, leftmost, augment); in rb_add_augmented_cached() 86 return leftmost ? node : NULL; in rb_add_augmented_cached()
|
| /linux-6.15/tools/perf/util/ |
| H A D | rblist.c | 17 bool leftmost = true; in rblist__add_node() local 29 leftmost = false; in rblist__add_node() 40 rb_insert_color_cached(new_node, &rblist->entries, leftmost); in rblist__add_node() 59 bool leftmost = true; in __rblist__findnew() local 71 leftmost = false; in __rblist__findnew() 82 &rblist->entries, leftmost); in __rblist__findnew()
|
| H A D | hist.c | 628 bool leftmost = true; in hists__findnew_entry() local 677 leftmost = false; in hists__findnew_entry() 1567 bool leftmost = true; in hierarchy_insert_entry() local 1582 leftmost = false; in hierarchy_insert_entry() 1681 bool leftmost = true; in hists__collapse_insert_entry() local 1718 leftmost = false; in hists__collapse_insert_entry() 1873 leftmost = false; in hierarchy_insert_output_entry() 1973 leftmost = false; in __hists__insert_output_entry() 2294 leftmost = false; in resort_filtered_entry() 2497 leftmost = false; in hists__add_dummy_entry() [all …]
|
| H A D | srcline.c | 982 bool leftmost = true; in srcline__tree_insert() local 1000 leftmost = false; in srcline__tree_insert() 1004 rb_insert_color_cached(&node->rb_node, tree, leftmost); in srcline__tree_insert() 1075 bool leftmost = true; in inlines__tree_insert() local 1084 leftmost = false; in inlines__tree_insert() 1088 rb_insert_color_cached(&inlines->rb_node, tree, leftmost); in inlines__tree_insert()
|
| /linux-6.15/tools/include/linux/ |
| H A D | interval_tree_generic.h | 44 bool leftmost = true; \ 55 leftmost = false; \ 62 leftmost, &ITPREFIX ## _augment); \ 122 ITSTRUCT *node, *leftmost; \ 144 leftmost = rb_entry(root->rb_leftmost, ITSTRUCT, ITRB); \ 145 if (ITSTART(leftmost) > last) \
|
| H A D | rbtree.h | 131 bool leftmost) in rb_insert_color_cached() argument 133 if (leftmost) in rb_insert_color_cached() 183 bool leftmost = true; in rb_add_cached() local 191 leftmost = false; in rb_add_cached() 196 rb_insert_color_cached(node, tree, leftmost); in rb_add_cached()
|
| /linux-6.15/kernel/locking/ |
| H A D | rtmutex_common.h | 124 struct rb_node *leftmost = rb_first_cached(&lock->waiters); in rt_mutex_waiter_is_top_waiter() local 126 return rb_entry(leftmost, struct rt_mutex_waiter, tree.entry) == waiter; in rt_mutex_waiter_is_top_waiter() 131 struct rb_node *leftmost = rb_first_cached(&lock->waiters); in rt_mutex_top_waiter() local 136 if (leftmost) { in rt_mutex_top_waiter() 137 w = rb_entry(leftmost, struct rt_mutex_waiter, tree.entry); in rt_mutex_top_waiter()
|
| /linux-6.15/fs/f2fs/ |
| H A D | extent_cache.c | 223 bool *leftmost) in __lookup_extent_node_ret() argument 240 *leftmost = true; in __lookup_extent_node_ret() 250 *leftmost = false; in __lookup_extent_node_ret() 293 bool leftmost) in __attach_extent_node() argument 581 bool leftmost) in __insert_extent_tree() argument 594 leftmost = true; in __insert_extent_tree() 605 leftmost = false; in __insert_extent_tree() 662 bool leftmost = false; in __update_extent_tree_range() local 697 &leftmost); in __update_extent_tree_range() 825 bool leftmost = false; in f2fs_update_read_extent_tree_range_compressed() local [all …]
|
| /linux-6.15/lib/ |
| H A D | rbtree_test.c | 54 bool leftmost = true; in insert_cached() local 62 leftmost = false; in insert_cached() 67 rb_insert_color_cached(&node->rb, root, leftmost); in insert_cached() 117 bool leftmost = true; in insert_augmented_cached() local 128 leftmost = false; in insert_augmented_cached() 135 leftmost, &augment_callbacks); in insert_augmented_cached()
|
| /linux-6.15/kernel/bpf/ |
| H A D | range_tree.c | 91 bool leftmost = true; in __range_size_insert() local 99 leftmost = false; in __range_size_insert() 104 rb_insert_color_cached(&rn->rb_range_size, root, leftmost); in __range_size_insert()
|
| /linux-6.15/drivers/gpu/drm/ |
| H A D | drm_mm.c | 170 bool leftmost; in drm_mm_interval_tree_add_node() local 187 leftmost = false; in drm_mm_interval_tree_add_node() 191 leftmost = true; in drm_mm_interval_tree_add_node() 203 leftmost = false; in drm_mm_interval_tree_add_node() 208 rb_insert_augmented_cached(&node->rb, &mm->interval_tree, leftmost, in drm_mm_interval_tree_add_node()
|
| /linux-6.15/Documentation/userspace-api/media/v4l/ |
| H A D | pixfmt-intro.rst | 29 leftmost pixel of the topmost row. Following that is the pixel 34 leftmost pixel of the second row from the top, and so on. The last row
|
| H A D | pixfmt-v4l2-mplane.rst | 43 - Distance in bytes between the leftmost pixels in two adjacent
|
| /linux-6.15/net/sched/ |
| H A D | sch_etf.c | 168 bool leftmost = true; in etf_enqueue_timesortedlist() local 183 leftmost = false; in etf_enqueue_timesortedlist() 189 rb_insert_color_cached(&nskb->rbnode, &q->head, leftmost); in etf_enqueue_timesortedlist()
|
| /linux-6.15/Documentation/scheduler/ |
| H A D | sched-design-CFS.rst | 76 p->se.vruntime key. CFS picks the "leftmost" task from this tree and sticks to it. 79 to become the "leftmost task" and thus get on the CPU within a deterministic 86 becomes the "leftmost task" of the time-ordered rbtree it maintains (plus a 87 small amount of "granularity" distance relative to the leftmost task so that we 88 do not over-schedule tasks and trash the cache), then the new leftmost task is
|
| /linux-6.15/Documentation/translations/zh_CN/core-api/ |
| H A D | rbtree.rst | 281 * Iterate to find the leftmost such node N. 293 return node; /* node is leftmost match */
|
| /linux-6.15/kernel/sched/ |
| H A D | deadline.c | 621 struct rb_node *leftmost; in enqueue_pushable_dl_task() local 625 leftmost = rb_add_cached(&p->pushable_dl_tasks, in enqueue_pushable_dl_task() 628 if (leftmost) in enqueue_pushable_dl_task() 641 struct rb_node *leftmost; in dequeue_pushable_dl_task() local 646 leftmost = rb_erase_cached(&p->pushable_dl_tasks, root); in dequeue_pushable_dl_task() 647 if (leftmost) in dequeue_pushable_dl_task() 648 dl_rq->earliest_dl.next = __node_2_pdl(leftmost)->dl.deadline; in dequeue_pushable_dl_task() 1876 struct rb_node *leftmost = rb_first_cached(&dl_rq->root); in dec_dl_deadline() local 1877 struct sched_dl_entity *entry = __node_2_dle(leftmost); in dec_dl_deadline()
|
| /linux-6.15/Documentation/core-api/ |
| H A D | rbtree.rst | 199 Computing the leftmost (smallest) node is quite a common task for binary 212 leftmost node. This allows rb_root_cached to exist wherever rb_root does, 319 * Iterate to find the leftmost such node N. 331 return node; /* node is leftmost match */
|
| /linux-6.15/block/ |
| H A D | blk-throttle.c | 379 bool leftmost = true; in tg_service_queue_add() local 389 leftmost = false; in tg_service_queue_add() 395 leftmost); in tg_service_queue_add()
|
| /linux-6.15/Documentation/bpf/ |
| H A D | map_lpm_trie.rst | 105 from leftmost leaf first. This means that iteration will return more
|
| /linux-6.15/fs/ |
| H A D | eventpoll.c | 1478 bool leftmost = true; in ep_rbtree_insert() local 1486 leftmost = false; in ep_rbtree_insert() 1491 rb_insert_color_cached(&epi->rbn, &ep->rbr, leftmost); in ep_rbtree_insert()
|
| /linux-6.15/Documentation/RCU/Design/Memory-Ordering/ |
| H A D | Tree-RCU-Memory-Ordering.rst | 360 The diagram below shows the path of ordering if the leftmost 363 leftmost ``rcu_node`` structure offlines its last CPU and if the next 547 | The leftmost drive to root stopped before it reached the root | 624 running on a CPU corresponding to the leftmost leaf ``rcu_node``
|