| /dpdk/drivers/event/octeontx/ |
| H A D | timvf_worker.h | 147 chunk = (struct tim_mem_entry *)(uintptr_t)(chunk + nb_chunk_slots)->w0; in timr_clr_bkt() 149 while (chunk) { in timr_clr_bkt() 153 chunk = pnext; in timr_clr_bkt() 198 (void **)&chunk))) { in timvf_refill_chunk_generic() 205 (uintptr_t) chunk; in timvf_refill_chunk_generic() 215 return chunk; in timvf_refill_chunk_generic() 232 (uintptr_t) chunk; in timvf_refill_chunk_fpa() 237 return chunk; in timvf_refill_chunk_fpa() 289 *chunk = *pent; in timvf_add_entry_sp() 342 *chunk = *pent; in timvf_add_entry_mp() [all …]
|
| /dpdk/drivers/event/cnxk/ |
| H A D | cnxk_tim_worker.h | 155 struct cnxk_tim_ent *chunk; in cnxk_tim_clr_bkt() local 160 chunk = (struct cnxk_tim_ent *)(uintptr_t)(chunk + in cnxk_tim_clr_bkt() 163 while (chunk) { in cnxk_tim_clr_bkt() 172 chunk = pnext; in cnxk_tim_clr_bkt() 186 struct cnxk_tim_ent *chunk; in cnxk_tim_refill_chunk() local 196 (uintptr_t)chunk; in cnxk_tim_refill_chunk() 206 return chunk; in cnxk_tim_refill_chunk() 227 return chunk; in cnxk_tim_insert_chunk() 302 *chunk = *pent; in cnxk_tim_add_entry_sp() 396 *chunk = *pent; in cnxk_tim_add_entry_mp() [all …]
|
| /dpdk/drivers/event/sw/ |
| H A D | iq_chunk.h | 34 struct sw_queue_chunk *chunk = sw->chunk_list_head; in iq_alloc_chunk() local 35 sw->chunk_list_head = chunk->next; in iq_alloc_chunk() 36 chunk->next = NULL; in iq_alloc_chunk() 37 return chunk; in iq_alloc_chunk() 41 iq_free_chunk(struct sw_evdev *sw, struct sw_queue_chunk *chunk) in iq_free_chunk() argument 43 chunk->next = sw->chunk_list_head; in iq_free_chunk() 44 sw->chunk_list_head = chunk; in iq_free_chunk() 79 struct sw_queue_chunk *chunk = iq_alloc_chunk(sw); in iq_enqueue() local 80 iq->tail->next = chunk; in iq_enqueue() 81 iq->tail = chunk; in iq_enqueue()
|
| /dpdk/drivers/common/cnxk/ |
| H A D | roc_npa.h | 345 unsigned int count, chunk, num_alloc; in roc_npa_aura_op_batch_alloc() local 353 chunk = (num > ROC_CN10K_NPA_BATCH_ALLOC_MAX_PTRS) ? in roc_npa_aura_op_batch_alloc() 358 chunk, dis_wait, drop)) in roc_npa_aura_op_batch_alloc() 362 chunk); in roc_npa_aura_op_batch_alloc() 368 if (num_alloc != chunk) in roc_npa_aura_op_batch_alloc() 473 unsigned int max_burst, chunk, bnum; in roc_npa_aura_op_batch_free() local 482 chunk = (bnum >= max_burst) ? max_burst : bnum; in roc_npa_aura_op_batch_free() 485 buf += chunk; in roc_npa_aura_op_batch_free() 486 bnum -= chunk; in roc_npa_aura_op_batch_free() 671 unsigned int chunk, count, num_alloc; in roc_npa_aura_op_bulk_alloc() local [all …]
|
| /dpdk/drivers/common/sfc_efx/base/ |
| H A D | siena_nvram.c | 93 size_t chunk; in siena_nvram_partn_read() local 97 chunk = MIN(size, SIENA_NVRAM_CHUNK); in siena_nvram_partn_read() 104 size -= chunk; in siena_nvram_partn_read() 105 data += chunk; in siena_nvram_partn_read() 106 offset += chunk; in siena_nvram_partn_read() 146 size_t chunk; in siena_nvram_partn_write() local 150 chunk = MIN(size, SIENA_NVRAM_CHUNK); in siena_nvram_partn_write() 153 data, chunk)) != 0) { in siena_nvram_partn_write() 157 size -= chunk; in siena_nvram_partn_write() 158 data += chunk; in siena_nvram_partn_write() [all …]
|
| H A D | siena_mcdi.c | 135 size_t chunk = MIN(remaining, sizeof (data)); in siena_mcdi_read_response() local 139 memcpy((uint8_t *)bufferp + pos, &data, chunk); in siena_mcdi_read_response() 140 pos += chunk; in siena_mcdi_read_response() 141 remaining -= chunk; in siena_mcdi_read_response()
|
| H A D | ef10_mcdi.c | 219 size_t chunk = MIN(remaining, sizeof (data)); in ef10_mcdi_read_response() local 222 memcpy((uint8_t *)bufferp + pos, &data, chunk); in ef10_mcdi_read_response() 223 pos += chunk; in ef10_mcdi_read_response() 224 remaining -= chunk; in ef10_mcdi_read_response()
|
| H A D | ef10_nvram.c | 2015 size_t chunk; in ef10_nvram_partn_read_mode() local 2022 data, chunk, mode)) != 0) { in ef10_nvram_partn_read_mode() 2026 size -= chunk; in ef10_nvram_partn_read_mode() 2027 data += chunk; in ef10_nvram_partn_read_mode() 2028 offset += chunk; in ef10_nvram_partn_read_mode() 2133 size_t chunk; in ef10_nvram_partn_write() local 2157 chunk = MIN(size, write_size); in ef10_nvram_partn_write() 2160 data, chunk)) != 0) { in ef10_nvram_partn_write() 2164 size -= chunk; in ef10_nvram_partn_write() 2165 data += chunk; in ef10_nvram_partn_write() [all …]
|
| /dpdk/lib/efd/ |
| H A D | rte_efd.c | 316 struct efd_online_chunk *chunk = &table->chunks[socket_id][chunk_id]; in efd_get_choice() local 323 chunk->bin_choice_list[bin_id / EFD_CHUNK_NUM_BIN_TO_GROUP_SETS]; in efd_get_choice() 802 struct efd_online_chunk *chunk = &table->chunks[socket_id][chunk_id]; in efd_apply_update() local 810 chunk->bin_choice_list[bin_index]; in efd_apply_update() 973 struct efd_offline_chunk_rules * const chunk = in efd_compute_update() local 981 &chunk->group_rules[current_group_id]; in efd_compute_update() 1096 chunk->group_rules[test_group_id].num_rules; in efd_compute_update() 1106 new_group = &chunk->group_rules[smallest_group_id]; in efd_compute_update() 1146 new_group = &chunk->group_rules[*group_id]; in efd_compute_update() 1192 struct efd_offline_chunk_rules * const chunk = in rte_efd_delete() local [all …]
|
| /dpdk/drivers/crypto/octeontx/ |
| H A D | otx_cryptodev_hw_access.c | 515 struct command_chunk *chunk = NULL; in otx_cpt_get_resource() local 596 chunk = &cptvf->cqueue.chead[i]; in otx_cpt_get_resource() 597 chunk->head = mem; in otx_cpt_get_resource() 598 chunk->dma_addr = dma_addr; in otx_cpt_get_resource() 608 *next_ptr = (uint64_t)chunk->dma_addr; in otx_cpt_get_resource() 610 chunk_prev = chunk; in otx_cpt_get_resource()
|
| /dpdk/lib/eal/linux/ |
| H A D | eal_vfio.c | 35 uint64_t chunk; /**< this mapping can be split in chunks of this size */ member 98 map->len == 0 && map->chunk == 0; in is_null_map() 132 if (umm_a->chunk < umm_b->chunk) in user_mem_map_cmp() 134 if (umm_a->chunk > umm_b->chunk) in user_mem_map_cmp() 161 dst_first->chunk = src_first->chunk; in process_maps() 172 last->chunk = src_last->chunk; in process_maps() 231 if (left->chunk != right->chunk) in merge_map() 258 bool addr_is_aligned = (map_va_off % map->chunk) == 0; in addr_is_chunk_aligned() 259 bool iova_is_aligned = (map_iova_off % map->chunk) == 0; in addr_is_chunk_aligned() 1910 new_map->chunk = has_partial_unmap ? 0 : len; in container_dma_map()
|
| /dpdk/doc/guides/eventdevs/ |
| H A D | cnxk.rst | 124 - ``TIM modify chunk slots`` 126 The ``tim_chnk_slots`` devargs can be used to modify number of chunk slots. 127 Chunks are used to store event timers, a chunk can be visualised as an array 128 where the last element points to the next chunk and rest of them are used to
|
| /dpdk/doc/guides/faq/ |
| H A D | faq.rst | 106 Therefore, each 64B chunk is interleaved across both memory domains. 108 The first 64B chunk is mapped to node 0, the second 64B chunk is mapped to node 1, the third to nod…
|
| /dpdk/doc/guides/prog_guide/ |
| H A D | compressdev.rst | 429 burst() multiple times to process related chunk of data because 448 file broken into multiple chunks then file is represented by a stream and each chunk of that file is 470 An example pseudocode to set up and process a stream having NUM_CHUNKS with each chunk size of CHUN… 532 /* set to final, if last chunk*/
|
| H A D | gpudev.rst | 12 it is possible to allocate a chunk of GPU memory and use it
|
| H A D | efd_lib.rst | 409 input key the hash result is split into two parts a chunk id and 8-bit 410 bin id. A chunk contains 64 different groups and 256 bins (i.e. for any
|
| /dpdk/doc/guides/nics/ |
| H A D | netvsc.rst | 133 set larger than the MTU, then all packets smaller than the chunk size
|
| H A D | mlx4.rst | 98 page in the extended chunk is freed, that doesn't become reusable until the
|
| /dpdk/drivers/net/bnxt/ |
| H A D | bnxt_ethdev.c | 4008 uint16_t chunk; in bnxt_get_module_eeprom() local 4014 chunk = RTE_MIN(length, 256 - offset); in bnxt_get_module_eeprom() 4019 chunk, data); in bnxt_get_module_eeprom() 4024 data += chunk; in bnxt_get_module_eeprom() 4025 length -= chunk; in bnxt_get_module_eeprom() 4027 page += 1 + (chunk > 128); in bnxt_get_module_eeprom()
|
| /dpdk/doc/guides/platform/ |
| H A D | mlx5.rst | 579 page in the extended chunk is freed, that doesn't become reusable until the
|
| /dpdk/doc/guides/rel_notes/ |
| H A D | release_18_05.rst | 49 ``RTE_DRIVER_MEMPOOL_BUCKET_SIZE_KB`` memory chunk which is a build time option.
|