| /linux-6.15/kernel/ |
| H A D | range.c | 29 int add_range_with_merge(struct range *range, int az, int nr_range, in add_range_with_merge() argument 53 memmove(&range[i], &range[i + 1], in add_range_with_merge() 76 if (start <= range[j].start && end >= range[j].end) { in subtract_range() 82 if (start <= range[j].start && end < range[j].end && in subtract_range() 89 if (start > range[j].start && end >= range[j].end && in subtract_range() 95 if (start > range[j].start && end < range[j].end) { in subtract_range() 102 range[i].end = range[j].end; in subtract_range() 126 int clean_sort_range(struct range *range, int az) in clean_sort_range() argument 141 range[i].start = range[k].start; in clean_sort_range() 142 range[i].end = range[k].end; in clean_sort_range() [all …]
|
| /linux-6.15/drivers/soc/ti/ |
| H A D | knav_qmss_acc.c | 90 range = _instdata; in knav_acc_int_handler() 92 kdev = range->kdev; in knav_acc_int_handler() 94 acc = range->acc; in knav_acc_int_handler() 205 acc = range->acc; in knav_range_setup_acc_irq() 233 range); in knav_range_setup_acc_irq() 311 acc = range->acc; in knav_acc_setup_cmd() 449 devm_kfree(range->kdev->dev, range->acc); in knav_acc_free_range() 530 range->name); in knav_init_acc_range() 536 range->name); in knav_init_acc_range() 547 range->acc = devm_kcalloc(kdev->dev, channels, sizeof(*range->acc), in knav_init_acc_range() [all …]
|
| H A D | knav_qmss_queue.c | 140 struct knav_range_info *range = inst->range; in knav_queue_free_irq() local 242 struct knav_range_info *range = inst->range; in __knav_queue_open() local 245 if (range->ops && range->ops->open_queue) in __knav_queue_open() 318 struct knav_range_info *range = inst->range; in knav_queue_set_notify() local 320 if (range->ops && range->ops->set_notify) in knav_queue_set_notify() 567 range->ops->close_queue(range, inst); in knav_queue_close() 1263 range->num_irqs = min(range->num_irqs, range->num_queues); in knav_setup_queue_range() 1296 range->name, range->queue_base, in knav_setup_queue_range() 1335 range->ops->free_range(range); in knav_free_queue_range() 1724 inst->range = range; in knav_queue_init_queue() [all …]
|
| /linux-6.15/arch/s390/include/asm/ |
| H A D | physmem_info.h | 139 for (range = &physmem_info.reserved[t], *p_start = range->start, *p_end = range->end; \ 140 range && range->end; range = range->chain ? __va(range->chain) : NULL, \ 141 *p_start = range ? range->start : 0, *p_end = range ? range->end : 0) 146 if (!range) { in __physmem_reserved_next() 148 if (range->end) in __physmem_reserved_next() 149 return range; in __physmem_reserved_next() 155 if (range->end) in __physmem_reserved_next() 156 return range; in __physmem_reserved_next() 163 *p_start = range ? range->start : 0, *p_end = range ? range->end : 0; \ 164 range; range = __physmem_reserved_next(&t, range), \ [all …]
|
| /linux-6.15/security/selinux/ss/ |
| H A D | context.h | 39 memset(&c->range, 0, sizeof(c->range)); in mls_context_init() 47 dst->range.level[0].sens = src->range.level[0].sens; in mls_context_cpy() 48 rc = ebitmap_cpy(&dst->range.level[0].cat, &src->range.level[0].cat); in mls_context_cpy() 52 dst->range.level[1].sens = src->range.level[1].sens; in mls_context_cpy() 68 dst->range.level[0].sens = src->range.level[0].sens; in mls_context_cpy_low() 73 dst->range.level[1].sens = src->range.level[0].sens; in mls_context_cpy_low() 89 dst->range.level[0].sens = src->range.level[1].sens; in mls_context_cpy_high() 94 dst->range.level[1].sens = src->range.level[1].sens; in mls_context_cpy_high() 107 const struct mls_range *r1 = &c1->range, *r2 = &c2->range; in mls_context_glblub() 138 return ((c1->range.level[0].sens == c2->range.level[0].sens) && in mls_context_equal() [all …]
|
| H A D | mls.c | 50 e = &context->range.level[l].cat; in mls_compute_context_len() 71 &context->range.level[1])) in mls_compute_context_len() 110 e = &context->range.level[l].cat; in mls_sid_to_context() 148 &context->range.level[1])) in mls_sid_to_context() 208 if (!mls_range_contains(usrdatum->range, c->range)) in mls_context_isvalid() 339 context->range.level[1].sens = context->range.level[0].sens; in mls_context_to_sid() 341 &context->range.level[0].cat); in mls_context_to_sid() 385 context->range.level[l].sens = range->level[l].sens; in mls_range_set() 387 &range->level[l].cat); in mls_range_set() 581 context->range.level[1].sens = context->range.level[0].sens; in mls_import_netlbl_lvl() [all …]
|
| /linux-6.15/mm/ |
| H A D | memremap.c | 66 static void pgmap_array_delete(struct range *range) in pgmap_array_delete() argument 75 struct range *range = &pgmap->ranges[range_id]; in pfn_first() local 88 struct range *range = &pgmap->ranges[i]; in pgmap_pfn_valid() local 100 const struct range *range = &pgmap->ranges[range_id]; in pfn_end() local 102 return (range->start + range_len(range)) >> PAGE_SHIFT; in pfn_end() 113 struct range *range = &pgmap->ranges[range_id]; in pageunmap_range() local 127 arch_remove_memory(range->start, range_len(range), in pageunmap_range() 174 struct range *range = &pgmap->ranges[range_id]; in pagemap_range() local 196 is_ram = region_intersects(range->start, range_len(range), in pagemap_range() 202 range->start, range->end); in pagemap_range() [all …]
|
| H A D | hmm.c | 90 struct hmm_range *range = hmm_vma_walk->range; in hmm_pte_need_fault() local 125 struct hmm_range *range = hmm_vma_walk->range; in hmm_range_need_fault() local 134 if (!((range->default_flags | range->pfn_flags_mask) & in hmm_range_need_fault() 151 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_walk_hole() local 192 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_handle_pmd() local 228 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_handle_pte() local 326 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_walk_pmd() local 415 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_walk_pud() local 474 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_walk_hugetlb_entry() local 524 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_walk_test() local [all …]
|
| H A D | execmem.c | 34 unsigned long start = range->start; in execmem_vmalloc() 35 unsigned long end = range->end; in execmem_vmalloc() 47 if (!p && range->fallback_start) { in execmem_vmalloc() 48 start = range->fallback_start; in execmem_vmalloc() 49 end = range->fallback_end; in execmem_vmalloc() 74 range->start, range->end, NUMA_NO_NODE, in execmem_vmap() 76 if (!area && range->fallback_start) in execmem_vmap() 78 range->fallback_start, range->fallback_end, in execmem_vmap() 194 if (addr >= range->start && addr + size < range->end) in within_range() 197 if (range->fallback_start && in within_range() [all …]
|
| /linux-6.15/lib/ |
| H A D | logic_pio.c | 59 if (start >= range->hw_start + range->size || in logic_pio_register_range() 61 mmio_end = range->io_start + range->size; in logic_pio_register_range() 132 found_range = range; in find_io_range_by_fwnode() 148 if (in_range(pio, range->io_start, range->size)) { in find_io_range() 149 found_range = range; in find_io_range() 175 if (range) in logic_pio_to_hwaddr() 176 return range->hw_start + pio - range->io_start; in logic_pio_to_hwaddr() 195 if (!range || range->flags == LOGIC_PIO_CPU_MMIO) { in logic_pio_trans_hwaddr() 204 return addr - range->hw_start + range->io_start; in logic_pio_trans_hwaddr() 215 if (in_range(addr, range->hw_start, range->size)) { in logic_pio_trans_cpuaddr() [all …]
|
| /linux-6.15/include/linux/ |
| H A D | range.h | 6 struct range { struct 11 static inline u64 range_len(const struct range *range) in range_len() argument 13 return range->end - range->start + 1; in range_len() 18 const struct range *r2) in range_contains() 25 const struct range *r2) in range_overlaps() 30 int add_range(struct range *range, int az, int nr_range, 34 int add_range_with_merge(struct range *range, int az, int nr_range, 37 void subtract_range(struct range *range, int az, u64 start, u64 end); 39 int clean_sort_range(struct range *range, int az); 41 void sort_range(struct range *range, int nr_range); [all …]
|
| H A D | mmu_notifier.h | 437 if (mm_has_notifiers(range->mm)) { in mmu_notifier_invalidate_range_start() 457 if (mm_has_notifiers(range->mm)) { in mmu_notifier_invalidate_range_start_nonblock() 471 if (mm_has_notifiers(range->mm)) in mmu_notifier_invalidate_range_end() 501 range->event = event; in mmu_notifier_range_init() 502 range->mm = mm; in mmu_notifier_range_init() 503 range->start = start; in mmu_notifier_range_init() 504 range->end = end; in mmu_notifier_range_init() 505 range->flags = flags; in mmu_notifier_range_init() 515 range->owner = owner; in mmu_notifier_range_init_owner() 577 range->start = start; in _mmu_notifier_range_init() [all …]
|
| /linux-6.15/arch/mips/loongson64/ |
| H A D | init.c | 159 range = kzalloc(sizeof(*range), GFP_ATOMIC); in add_legacy_isa_io() 160 if (!range) in add_legacy_isa_io() 170 kfree(range); in add_legacy_isa_io() 177 kfree(range); in add_legacy_isa_io() 208 range.cpu_addr, in reserve_pio_range() 209 range.cpu_addr + range.size - 1, in reserve_pio_range() 210 range.bus_addr); in reserve_pio_range() 211 if (add_legacy_isa_io(&np->fwnode, range.cpu_addr, range.size)) in reserve_pio_range() 216 range.cpu_addr, in reserve_pio_range() 217 range.cpu_addr + range.size - 1, in reserve_pio_range() [all …]
|
| /linux-6.15/drivers/dax/ |
| H A D | kmem.c | 34 struct range *range = &dax_range->range; in dax_kmem_range() local 41 r->end = range->end; in dax_kmem_range() 98 struct range range; in dev_dax_kmem_probe() local 103 i, range.start, range.end); in dev_dax_kmem_probe() 132 struct range range; in dev_dax_kmem_probe() local 139 res = request_mem_region(range.start, range_len(&range), data->res_name); in dev_dax_kmem_probe() 142 i, range.start, range.end); in dev_dax_kmem_probe() 175 i, range.start, range.end); in dev_dax_kmem_probe() 216 struct range range; in dev_dax_kmem_remove() local 223 rc = remove_memory(range.start, range_len(&range)); in dev_dax_kmem_remove() [all …]
|
| /linux-6.15/drivers/gpu/drm/ |
| H A D | drm_gpusvm.c | 803 range = kzalloc(sizeof(*range), GFP_KERNEL); in drm_gpusvm_range_alloc() 805 if (!range) in drm_gpusvm_range_alloc() 817 return range; in drm_gpusvm_range_alloc() 937 if (range) { in drm_gpusvm_range_chunk_size() 1052 if (range) in drm_gpusvm_range_find_or_insert() 1073 if (IS_ERR(range)) { in drm_gpusvm_range_find_or_insert() 1086 return range; in drm_gpusvm_range_find_or_insert() 1219 return range; in drm_gpusvm_range_get() 1240 kfree(range); in drm_gpusvm_range_destroy() 1275 return range->flags.has_devmem_pages || range->flags.has_dma_mapping; in drm_gpusvm_range_pages_valid() [all …]
|
| /linux-6.15/drivers/gpu/drm/xe/ |
| H A D | xe_svm.c | 34 return xe_svm_range_in_vram(range) && range->tile_present; in xe_svm_range_has_vram_binding() 87 struct xe_svm_range *range; in xe_svm_range_alloc() local 89 range = kzalloc(sizeof(*range), GFP_KERNEL); in xe_svm_range_alloc() 90 if (!range) in xe_svm_range_alloc() 96 return &range->base; in xe_svm_range_alloc() 102 kfree(range); in xe_svm_range_free() 146 if (range->base.flags.unmapped || !range->tile_present) in xe_svm_range_notifier_event_begin() 306 struct xe_svm_range *range; in xe_svm_garbage_collector() local 317 typeof(*range), in xe_svm_garbage_collector() 319 if (!range) in xe_svm_garbage_collector() [all …]
|
| /linux-6.15/drivers/of/ |
| H A D | address.c | 95 if (*addr != *range) in of_bus_default_flags_map() 231 range->size); in of_pci_range_to_resource() 801 if (!range) in of_pci_range_parser_one() 804 if (!parser->range || parser->range + np > parser->end) in of_pci_range_parser_one() 807 range->flags = parser->bus->get_flags(parser->range); in of_pci_range_parser_one() 840 if (bus_addr != range->bus_addr + range->size || in of_pci_range_parser_one() 841 cpu_addr != range->cpu_addr + range->size) in of_pci_range_parser_one() 848 return range; in of_pci_range_parser_one() 948 range.bus_addr, range.cpu_addr, range.size); in of_dma_get_range() 986 if (range.cpu_addr + range.size > cpu_end) in of_dma_get_max_cpu_address() [all …]
|
| /linux-6.15/drivers/pci/hotplug/ |
| H A D | ibmphp_res.c | 521 range = range->next; in fix_me() 938 range = range->next; in find_range() 1203 len_tmp = range->end - range->start; in ibmphp_check_resource() 1238 range = range->next; in ibmphp_check_resource() 1269 len_tmp = range->end - range->start; in ibmphp_check_resource() 1304 range = range->next; in ibmphp_check_resource() 1777 debug_pci("[%x - %x]\n", range->start, range->end); in ibmphp_print_test() 1778 range = range->next; in ibmphp_print_test() 1787 debug_pci("[%x - %x]\n", range->start, range->end); in ibmphp_print_test() 1788 range = range->next; in ibmphp_print_test() [all …]
|
| /linux-6.15/mm/damon/ |
| H A D | sysfs-common.c | 22 struct damon_sysfs_ul_range *range = kmalloc(sizeof(*range), in damon_sysfs_ul_range_alloc() local 25 if (!range) in damon_sysfs_ul_range_alloc() 27 range->kobj = (struct kobject){}; in damon_sysfs_ul_range_alloc() 28 range->min = min; in damon_sysfs_ul_range_alloc() 29 range->max = max; in damon_sysfs_ul_range_alloc() 31 return range; in damon_sysfs_ul_range_alloc() 37 struct damon_sysfs_ul_range *range = container_of(kobj, in min_show() local 40 return sysfs_emit(buf, "%lu\n", range->min); in min_show() 55 range->min = min; in min_store() 65 return sysfs_emit(buf, "%lu\n", range->max); in max_show() [all …]
|
| /linux-6.15/drivers/gpu/drm/sprd/ |
| H A D | megacores_pll.c | 236 range[L] = 50 * scale; in dphy_timing_config() 237 range[H] = INFINITY; in dphy_timing_config() 243 range[L] = 38 * scale; in dphy_timing_config() 245 tmp = AVERAGE(range[L], range[H]); in dphy_timing_config() 246 val[CLK] = DIV_ROUND_UP(AVERAGE(range[L], range[H]), t_half_byteck) - 1; in dphy_timing_config() 249 tmp |= AVERAGE(range[L], range[H]) << 16; in dphy_timing_config() 250 val[DATA] = DIV_ROUND_UP(AVERAGE(range[L], range[H]), t_half_byteck) - 1; in dphy_timing_config() 255 range[H] = INFINITY; in dphy_timing_config() 266 range[H] = INFINITY; in dphy_timing_config() 274 range[H] = INFINITY; in dphy_timing_config() [all …]
|
| /linux-6.15/include/drm/ |
| H A D | drm_gpusvm.h | 338 struct drm_gpusvm_range *range); 341 struct drm_gpusvm_range *range); 349 struct drm_gpusvm_range *range); 356 struct drm_gpusvm_range *range, 360 struct drm_gpusvm_range *range, 430 return range->itree.start; in drm_gpusvm_range_start() 442 return range->itree.last + 1; in drm_gpusvm_range_end() 454 return drm_gpusvm_range_end(range) - drm_gpusvm_range_start(range); in drm_gpusvm_range_size() 504 if (range && !list_is_last(&range->entry, in __drm_gpusvm_range_next() 505 &range->notifier->range_list)) in __drm_gpusvm_range_next() [all …]
|
| /linux-6.15/tools/testing/selftests/net/ |
| H A D | ip_local_port_range.c | 30 *lo = range & 0xffff; in unpack_port_range() 31 *hi = range >> 16; in unpack_port_range() 115 *range = val; in get_ip_local_port_range() 238 __u32 range; in TEST_F() local 281 __u32 range; in TEST_F() local 290 err = setsockopt(fd, SOL_IP, IP_LOCAL_PORT_RANGE, &range, sizeof(range)); in TEST_F() 308 __u32 range; in TEST_F() local 339 err = setsockopt(fd, SOL_IP, IP_LOCAL_PORT_RANGE, &range, sizeof(range)); in TEST_F() 365 __u32 range; in TEST_F() local 372 err = setsockopt(fd, SOL_IP, IP_LOCAL_PORT_RANGE, &range, sizeof(range)); in TEST_F() [all …]
|
| /linux-6.15/net/netfilter/ |
| H A D | nf_nat_core.c | 460 &range->min_proto, &range->max_proto); in nf_in_range() 531 if (nf_inet_addr_cmp(&range->min_addr, &range->max_addr)) { in find_best_ips_proto() 532 *var_ipp = range->min_addr; in find_best_ips_proto() 755 &range->min_proto, in get_unique_tuple() 756 &range->max_proto) && in get_unique_tuple() 757 (range->min_proto.all == range->max_proto.all || in get_unique_tuple() 1052 range->max_proto.all = range->min_proto.all; in nf_nat_l4proto_nlattr_to_range() 1094 range->min_addr.ip); in nf_nat_ipv4_nlattr_to_range() 1112 range->max_addr = range->min_addr; in nf_nat_ipv6_nlattr_to_range() 1124 memset(range, 0, sizeof(*range)); in nfnetlink_parse_nat() [all …]
|
| H A D | nf_nat_bpf.c | 37 struct nf_nat_range2 range; in bpf_ct_set_nat_info() local 42 memset(&range, 0, sizeof(struct nf_nat_range2)); in bpf_ct_set_nat_info() 43 range.flags = NF_NAT_RANGE_MAP_IPS; in bpf_ct_set_nat_info() 44 range.min_addr = *addr; in bpf_ct_set_nat_info() 45 range.max_addr = range.min_addr; in bpf_ct_set_nat_info() 47 range.flags |= NF_NAT_RANGE_PROTO_SPECIFIED; in bpf_ct_set_nat_info() 48 range.min_proto.all = cpu_to_be16(port); in bpf_ct_set_nat_info() 49 range.max_proto.all = range.min_proto.all; in bpf_ct_set_nat_info() 52 return nf_nat_setup_info(ct, &range, manip) == NF_DROP ? -ENOMEM : 0; in bpf_ct_set_nat_info()
|
| H A D | xt_nat.c | 55 struct nf_nat_range2 range; in xt_snat_target_v0() local 64 xt_nat_convert_range(&range, &mr->range[0]); in xt_snat_target_v0() 72 struct nf_nat_range2 range; in xt_dnat_target_v0() local 80 xt_nat_convert_range(&range, &mr->range[0]); in xt_dnat_target_v0() 88 struct nf_nat_range2 range; in xt_snat_target_v1() local 97 memcpy(&range, range_v1, sizeof(*range_v1)); in xt_snat_target_v1() 98 memset(&range.base_proto, 0, sizeof(range.base_proto)); in xt_snat_target_v1() 107 struct nf_nat_range2 range; in xt_dnat_target_v1() local 115 memcpy(&range, range_v1, sizeof(*range_v1)); in xt_dnat_target_v1() 116 memset(&range.base_proto, 0, sizeof(range.base_proto)); in xt_dnat_target_v1() [all …]
|