| /linux-6.15/drivers/gpu/drm/radeon/ |
| H A D | radeon_cs.c | 301 if (p->chunks == NULL) { in radeon_cs_parser_init() 316 p->chunk_relocs = &p->chunks[i]; in radeon_cs_parser_init() 319 p->chunk_ib = &p->chunks[i]; in radeon_cs_parser_init() 331 p->chunk_flags = &p->chunks[i]; in radeon_cs_parser_init() 337 size = p->chunks[i].length_dw; in radeon_cs_parser_init() 339 p->chunks[i].user_ptr = cdata; in radeon_cs_parser_init() 358 if (p->chunks[i].length_dw > 1) in radeon_cs_parser_init() 359 ring = p->chunks[i].kdata[1]; in radeon_cs_parser_init() 360 if (p->chunks[i].length_dw > 2) in radeon_cs_parser_init() 459 kvfree(parser->chunks[i].kdata); in radeon_cs_parser_fini() [all …]
|
| /linux-6.15/drivers/comedi/drivers/ni_routing/tools/ |
| H A D | convert_csv_to_c.py | 228 chunks = [ self.output_file_top, 244 chunks.append('\t&{},'.format(dev_table_name)) 273 chunks.append('\tNULL,') # terminate list 274 chunks.append('};') 275 return '\n'.join(chunks) 416 chunks = [ self.output_file_top, 432 chunks.append('\t&{},'.format(fam_table_name)) 462 chunks.append('\tNULL,') # terminate list 463 chunks.append('};') 464 return '\n'.join(chunks)
|
| /linux-6.15/drivers/net/ethernet/netronome/nfp/nfpcore/ |
| H A D | nfp_nsp.c | 505 } *chunks; in nfp_nsp_command_buf_dma_sg() local 517 chunks = kcalloc(nseg, sizeof(*chunks), GFP_KERNEL); in nfp_nsp_command_buf_dma_sg() 518 if (!chunks) in nfp_nsp_command_buf_dma_sg() 526 chunks[i].chunk = kmalloc(chunk_size, in nfp_nsp_command_buf_dma_sg() 528 if (!chunks[i].chunk) in nfp_nsp_command_buf_dma_sg() 540 off += chunks[i].len; in nfp_nsp_command_buf_dma_sg() 548 addr = dma_map_single(dev, chunks[i].chunk, chunks[i].len, in nfp_nsp_command_buf_dma_sg() 550 chunks[i].dma_addr = addr; in nfp_nsp_command_buf_dma_sg() 609 dma_unmap_single(dev, chunks[i].dma_addr, chunks[i].len, in nfp_nsp_command_buf_dma_sg() 614 kfree(chunks[i].chunk); in nfp_nsp_command_buf_dma_sg() [all …]
|
| /linux-6.15/tools/testing/selftests/bpf/progs/ |
| H A D | dynptr_success.c | 615 int i, chunks = 200; in test_dynptr_copy_xdp() local 619 bpf_ringbuf_reserve_dynptr(&ringbuf, len * chunks, 0, &ptr_buf); in test_dynptr_copy_xdp() 622 bpf_for(i, 0, chunks) { in test_dynptr_copy_xdp() 628 err = bpf_dynptr_copy(&ptr_xdp, 0, &ptr_buf, 0, len * chunks); in test_dynptr_copy_xdp() 632 bpf_for(i, 0, chunks) { in test_dynptr_copy_xdp() 643 bpf_for(i, 0, chunks) { in test_dynptr_copy_xdp() 649 err = bpf_dynptr_copy(&ptr_buf, 0, &ptr_xdp, 0, len * chunks); in test_dynptr_copy_xdp() 653 bpf_for(i, 0, chunks) { in test_dynptr_copy_xdp() 663 err = bpf_dynptr_copy(&ptr_xdp, 2, &ptr_xdp, len, len * (chunks - 1)); in test_dynptr_copy_xdp() 667 bpf_for(i, 0, chunks - 1) { in test_dynptr_copy_xdp() [all …]
|
| /linux-6.15/drivers/net/mctp/ |
| H A D | mctp-serial.c | 534 u8 chunks[MAX_CHUNKS]; member 567 .chunks = { 3, 1, 1, 0}, 572 .chunks = { 3, 1, 1, 0}, 577 .chunks = { 1, 2, 0}, 582 .chunks = { 1, 1, 1, 0}, 587 .chunks = { 1, 1, 1, 1, 0}, 597 .chunks = { 1, 0 }, 602 .chunks = { 1, 0 }, 607 .chunks = { 3, 0 }, 612 .chunks = { 7, 0 }, [all …]
|
| /linux-6.15/net/xdp/ |
| H A D | xdp_umem.c | 163 u64 chunks, npgs; in xdp_umem_reg() local 198 chunks = div_u64_rem(size, chunk_size, &chunks_rem); in xdp_umem_reg() 199 if (!chunks || chunks > U32_MAX) in xdp_umem_reg() 217 umem->chunks = chunks; in xdp_umem_reg()
|
| /linux-6.15/scripts/gdb/linux/ |
| H A D | timerlist.py | 164 chunks = [] 170 chunks.append(buf[start:end]) 172 chunks.append(',') 176 chunks[0] = chunks[0][0] # Cut off the first 0 178 return "".join(str(chunks))
|
| /linux-6.15/drivers/infiniband/hw/usnic/ |
| H A D | usnic_vnic.c | 44 struct usnic_vnic_res_chunk chunks[USNIC_VNIC_RES_TYPE_MAX]; member 117 for (i = 0; i < ARRAY_SIZE(vnic->chunks); i++) { in usnic_vnic_dump() 118 chunk = &vnic->chunks[i]; in usnic_vnic_dump() 222 return vnic->chunks[type].cnt; in usnic_vnic_res_cnt() 228 return vnic->chunks[type].free_cnt; in usnic_vnic_res_free_cnt() 254 src = &vnic->chunks[type]; in usnic_vnic_get_resources() 286 vnic->chunks[res->type].free_cnt++; in usnic_vnic_put_resources() 382 &vnic->chunks[res_type]); in usnic_vnic_discover_resources() 391 usnic_vnic_free_res_chunk(&vnic->chunks[res_type]); in usnic_vnic_discover_resources() 427 usnic_vnic_free_res_chunk(&vnic->chunks[res_type]); in usnic_vnic_release_resources()
|
| /linux-6.15/drivers/gpu/drm/panthor/ |
| H A D | panthor_heap.c | 54 struct list_head chunks; member 167 if (initial_chunk && !list_empty(&heap->chunks)) { in panthor_alloc_heap_chunk() 171 prev_chunk = list_first_entry(&heap->chunks, in panthor_alloc_heap_chunk() 183 list_add(&chunk->node, &heap->chunks); in panthor_alloc_heap_chunk() 205 list_for_each_entry_safe(chunk, tmp, &heap->chunks, node) in panthor_free_heap_chunks() 311 INIT_LIST_HEAD(&heap->chunks); in panthor_heap_create() 320 first_chunk = list_first_entry(&heap->chunks, in panthor_heap_create() 391 list_for_each_entry_safe(chunk, tmp, &heap->chunks, node) { in panthor_heap_return_chunk() 478 chunk = list_first_entry(&heap->chunks, in panthor_heap_grow()
|
| /linux-6.15/drivers/net/ethernet/intel/idpf/ |
| H A D | idpf_virtchnl.c | 1073 chunk = &chunks->chunks[num_chunks]; in idpf_vport_get_q_reg() 1172 chunks = &vc_aq->chunks; in idpf_queue_reg_init() 1175 chunks = &vport_params->chunks; in idpf_queue_reg_init() 1812 buf_sz = struct_size(eq, chunks.chunks, num_chunks); in idpf_send_ena_dis_queues_msg() 2058 chunks = &vport_config->req_qs_chunks->chunks; in idpf_send_delete_queues_msg() 2061 chunks = &vport_params->chunks; in idpf_send_delete_queues_msg() 2074 idpf_convert_reg_to_queue_chunks(eq->chunks.chunks, chunks->chunks, in idpf_send_delete_queues_msg() 2157 size = struct_size(vc_msg, chunks.chunks, in idpf_send_add_queues_msg() 3262 chunk = &chunks->chunks[num_chunks]; in idpf_vport_get_queue_ids() 3378 chunks = &vc_aq->chunks; in idpf_vport_queue_ids_init() [all …]
|
| /linux-6.15/lib/ |
| H A D | genalloc.c | 160 INIT_LIST_HEAD(&pool->chunks); in gen_pool_create() 203 list_add_rcu(&chunk->next_chunk, &pool->chunks); in gen_pool_add_owner() 223 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_virt_to_phys() 249 list_for_each_safe(_chunk, _next_chunk, &pool->chunks) { in gen_pool_destroy() 297 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_alloc_algo_owner() 503 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_free_owner() 538 list_for_each_entry_rcu(chunk, &(pool)->chunks, next_chunk) in gen_pool_for_each_chunk() 561 list_for_each_entry_rcu(chunk, &(pool)->chunks, next_chunk) { in gen_pool_has_addr() 586 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) in gen_pool_avail() 605 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) in gen_pool_size()
|
| /linux-6.15/tools/testing/selftests/drivers/net/mlxsw/spectrum/ |
| H A D | devlink_lib_spectrum.sh | 90 devlink_resource_size_set 32000 kvd linear chunks 99 devlink_resource_size_set 32000 kvd linear chunks 108 devlink_resource_size_set 49152 kvd linear chunks
|
| /linux-6.15/Documentation/networking/ |
| H A D | oa-tc6-framework.rst | 49 each chunk. Ethernet frames are transferred over one or more data chunks. 59 receive (RX) chunks. Chunks in both transmit and receive directions may 69 In parallel, receive data chunks are received on MISO. Each receive data 160 the MAC-PHY will be converted into multiple transmit data chunks. Each 177 transaction. For TX data chunks, this bit shall be ’1’. 254 host will be sent as multiple receive data chunks. Each receive data 286 data chunks of frame data that are available for 350 transmit data chunks (empty chunks) or data valid transmit data chunks in 356 once the receive data chunks become available again for reading, the 359 first data chunk will have the receive chunks available information. [all …]
|
| /linux-6.15/net/sctp/ |
| H A D | chunk.c | 43 INIT_LIST_HEAD(&msg->chunks); in sctp_datamsg_init() 65 list_for_each_entry(chunk, &msg->chunks, frag_list) in sctp_datamsg_free() 81 list_for_each_safe(pos, temp, &msg->chunks) { in sctp_datamsg_destroy() 280 list_add_tail(&chunk->frag_list, &msg->chunks); in sctp_datamsg_from_user() 289 list_for_each_safe(pos, temp, &msg->chunks) { in sctp_datamsg_from_user()
|
| H A D | auth.c | 186 struct sctp_chunks_param *chunks, in sctp_auth_make_key_vector() argument 197 if (chunks) in sctp_auth_make_key_vector() 198 chunks_len = ntohs(chunks->param_hdr.length); in sctp_auth_make_key_vector() 209 if (chunks) { in sctp_auth_make_key_vector() 210 memcpy(new->data + offset, chunks, chunks_len); in sctp_auth_make_key_vector() 656 switch (param->chunks[i]) { in __sctp_auth_cid() 664 if (param->chunks[i] == chunk) in __sctp_auth_cid() 772 p->chunks[nchunks] = chunk_id; in sctp_auth_ep_add_chunkid()
|
| /linux-6.15/kernel/ |
| H A D | audit_tree.c | 17 struct list_head chunks; member 101 INIT_LIST_HEAD(&tree->chunks); in alloc_tree() 435 list_add(&chunk->owners[0].list, &tree->chunks); in create_chunk() 507 list_add(&p->list, &tree->chunks); in tag_chunk() 572 while (!list_empty(&victim->chunks)) { in prune_tree_chunks() 577 p = list_first_entry(&victim->chunks, struct audit_node, list); in prune_tree_chunks() 618 for (p = tree->chunks.next; p != &tree->chunks; p = q) { in trim_marked() 623 list_add(p, &tree->chunks); in trim_marked() 705 list_for_each_entry(node, &tree->chunks, list) { in audit_trim_trees() 845 list_for_each_entry(node, &tree->chunks, list) in audit_add_tree_rule() [all …]
|
| /linux-6.15/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_cs.c | 204 if (!p->chunks) { in amdgpu_cs_pass1() 221 p->chunks[i].chunk_id = user_chunk.chunk_id; in amdgpu_cs_pass1() 224 size = p->chunks[i].length_dw; in amdgpu_cs_pass1() 229 if (p->chunks[i].kdata == NULL) { in amdgpu_cs_pass1() 242 switch (p->chunks[i].chunk_id) { in amdgpu_cs_pass1() 321 kvfree(p->chunks[i].kdata); in amdgpu_cs_pass1() 322 kvfree(p->chunks); in amdgpu_cs_pass1() 323 p->chunks = NULL; in amdgpu_cs_pass1() 601 chunk = &p->chunks[i]; in amdgpu_cs_pass2() 1405 kvfree(parser->chunks[i].kdata); in amdgpu_cs_parser_fini() [all …]
|
| /linux-6.15/drivers/infiniband/ulp/rtrs/ |
| H A D | README | 28 session. A session is associated with a set of memory chunks reserved on the 36 chunks reserved for him on the server side. Their number, size and addresses 45 which of the memory chunks has been accessed and at which offset the message 80 the server (number of memory chunks which are going to be allocated for that 122 1. When processing a write request client selects one of the memory chunks 139 1. When processing a write request client selects one of the memory chunks 144 using the IMM field, Server invalidate rkey associated to the memory chunks 162 1. When processing a read request client selects one of the memory chunks 181 1. When processing a read request client selects one of the memory chunks 186 Server invalidate rkey associated to the memory chunks first, when it finishes,
|
| /linux-6.15/drivers/infiniband/hw/efa/ |
| H A D | efa_verbs.c | 111 struct pbl_chunk *chunks; member 1346 sizeof(*chunk_list->chunks), in pbl_chunk_list_create() 1348 if (!chunk_list->chunks) in pbl_chunk_list_create() 1358 if (!chunk_list->chunks[i].buf) in pbl_chunk_list_create() 1370 cur_chunk_buf = chunk_list->chunks[0].buf; in pbl_chunk_list_create() 1386 chunk_list->chunks[i].buf, in pbl_chunk_list_create() 1387 chunk_list->chunks[i].length, in pbl_chunk_list_create() 1422 kfree(chunk_list->chunks[i].buf); in pbl_chunk_list_create() 1424 kfree(chunk_list->chunks); in pbl_chunk_list_create() 1436 kfree(chunk_list->chunks[i].buf); in pbl_chunk_list_destroy() [all …]
|
| /linux-6.15/arch/x86/kernel/cpu/resctrl/ |
| H A D | monitor.c | 309 u64 shift = 64 - width, chunks; in mbm_overflow_count() local 311 chunks = (cur_msr << shift) - (prev_msr << shift); in mbm_overflow_count() 312 return chunks >> shift; in mbm_overflow_count() 323 u64 msr_val, chunks; in resctrl_arch_rmid_read() local 336 am->chunks += mbm_overflow_count(am->prev_msr, msr_val, in resctrl_arch_rmid_read() 338 chunks = get_corrected_mbm_count(rmid, am->chunks); in resctrl_arch_rmid_read() 341 chunks = msr_val; in resctrl_arch_rmid_read() 344 *val = chunks * hw_res->mon_scale; in resctrl_arch_rmid_read()
|
| /linux-6.15/tools/testing/selftests/bpf/ |
| H A D | generate_udp_fragments.py | 46 chunks = [frag[i : i + 10] for i in range(0, len(frag), 10)] 47 chunks_fmted = [", ".join([str(hex(b)) for b in chunk]) for chunk in chunks]
|
| /linux-6.15/drivers/net/wireless/ti/wlcore/ |
| H A D | boot.c | 237 u32 chunks, addr, len; in wlcore_boot_upload_firmware() local 242 chunks = be32_to_cpup((__be32 *) fw); in wlcore_boot_upload_firmware() 245 wl1271_debug(DEBUG_BOOT, "firmware chunks to be uploaded: %u", chunks); in wlcore_boot_upload_firmware() 247 while (chunks--) { in wlcore_boot_upload_firmware() 258 chunks, addr, len); in wlcore_boot_upload_firmware()
|
| /linux-6.15/drivers/net/wireless/intel/iwlwifi/pcie/ |
| H A D | ctxt-info-gen3.c | 337 len0 = pnvm_data->chunks[0].len; in iwl_pcie_load_payloads_continuously() 338 len1 = pnvm_data->chunks[1].len; in iwl_pcie_load_payloads_continuously() 353 memcpy(dram->block, pnvm_data->chunks[0].data, len0); in iwl_pcie_load_payloads_continuously() 354 memcpy((u8 *)dram->block + len0, pnvm_data->chunks[1].data, len1); in iwl_pcie_load_payloads_continuously() 387 len = pnvm_data->chunks[i].len; in iwl_pcie_load_payloads_segments() 388 data = pnvm_data->chunks[i].data; in iwl_pcie_load_payloads_segments()
|
| /linux-6.15/drivers/md/ |
| H A D | md-bitmap.c | 163 unsigned long chunks; member 1018 bytes = DIV_ROUND_UP(chunks, 8); in md_bitmap_storage_alloc() 1323 unsigned long chunks = bitmap->counts.chunks; in md_bitmap_init_from_disk() local 1335 for (i = 0; i < chunks ; i++) { in md_bitmap_init_from_disk() 1405 for (i = 0; i < chunks; i++) { in md_bitmap_init_from_disk() 1431 bit_cnt, chunks); in md_bitmap_init_from_disk() 2396 unsigned long chunks; in __bitmap_resize() local 2432 bytes = DIV_ROUND_UP(chunks, 8); in __bitmap_resize() 2480 bitmap->counts.chunks = chunks; in __bitmap_resize() 2485 chunks << chunkshift); in __bitmap_resize() [all …]
|
| /linux-6.15/drivers/gpu/drm/sprd/ |
| H A D | sprd_dsi.c | 460 u32 chunks = 0; in sprd_dsi_dpi_video() local 539 chunks = vm->hactive / video_size; in sprd_dsi_dpi_video() 542 if (total_bytes >= (bytes_per_chunk * chunks)) { in sprd_dsi_dpi_video() 544 bytes_per_chunk * chunks; in sprd_dsi_dpi_video() 550 if (bytes_left > (pkt_header * chunks)) { in sprd_dsi_dpi_video() 552 pkt_header * chunks) / chunks; in sprd_dsi_dpi_video() 560 chunks = 1; in sprd_dsi_dpi_video() 571 dsi_reg_wr(ctx, VIDEO_PKT_CONFIG, VIDEO_LINE_CHUNK_NUM, 16, chunks); in sprd_dsi_dpi_video()
|