Searched refs:dirty (Results 1 – 14 of 14) sorted by relevance
| /dpdk/drivers/net/axgbe/ |
| H A D | axgbe_rxtx.c | 67 rxq->dirty = 0; in axgbe_dev_rx_queue_setup() 209 uint64_t old_dirty = rxq->dirty; in axgbe_recv_pkts() 320 rxq->dirty++; in axgbe_recv_pkts() 323 if (rxq->dirty != old_dirty) { in axgbe_recv_pkts() 343 uint64_t old_dirty = rxq->dirty; in eth_axgbe_recv_scattered_pkts() 452 rxq->dirty++; in eth_axgbe_recv_scattered_pkts() 489 if (rxq->dirty != old_dirty) { in eth_axgbe_recv_scattered_pkts() 586 txq->dirty = 0; in axgbe_dev_tx_queue_setup() 758 while (txq->cur != txq->dirty) { in axgbe_xmit_cleanup() 769 txq->dirty++; in axgbe_xmit_cleanup() [all …]
|
| H A D | axgbe_rxtx_vec_sse.c | 42 idx = AXGBE_GET_DESC_IDX(txq, txq->dirty + txq->free_batch_cnt in axgbe_xmit_cleanup_vec() 54 txq->dirty += txq->free_batch_cnt; in axgbe_xmit_cleanup_vec()
|
| H A D | axgbe_rxtx.h | 77 uint64_t dirty; member 125 uint64_t dirty; member
|
| H A D | axgbe_dev.c | 921 txq->dirty = 0; in wrapper_tx_desc_init() 945 rxq->dirty = 0; in wrapper_rx_desc_init()
|
| /dpdk/lib/eal/linux/ |
| H A D | eal_memalloc.c | 283 bool *dirty) in get_seg_fd() argument 293 *dirty = false; in get_seg_fd() 324 dirty != NULL) in get_seg_fd() 325 *dirty = true; in get_seg_fd() 495 bool *dirty) in resize_hugefile() argument 505 *dirty = false; in resize_hugefile() 511 grow, dirty); in resize_hugefile() 528 bool dirty; in alloc_seg() local 558 dirty = false; in alloc_seg() 570 &dirty); in alloc_seg() [all …]
|
| /dpdk/lib/eal/common/ |
| H A D | malloc_elem.c | 126 struct malloc_elem *orig_elem, size_t orig_size, bool dirty) in malloc_elem_init() argument 134 elem->dirty = dirty; in malloc_elem_init() 298 elem->orig_elem, elem->orig_size, elem->dirty); in split_elem() 504 elem1->dirty |= elem2->dirty; in join_elem() 585 elem->dirty = false; in malloc_elem_free() 597 if (!elem->dirty) in malloc_elem_free()
|
| H A D | malloc_elem.h | 36 uint32_t dirty : 1; member 328 bool dirty);
|
| H A D | malloc_heap.c | 91 void *start, size_t len, bool dirty) in malloc_heap_add_memory() argument 95 malloc_elem_init(elem, heap, msl, len, elem, len, dirty); in malloc_heap_add_memory() 303 bool dirty = false; in alloc_pages_on_heap() local 374 dirty |= ms[i]->flags & RTE_MEMSEG_FLAG_DIRTY; in alloc_pages_on_heap() 377 elem = malloc_heap_add_memory(heap, msl, map_addr, alloc_sz, dirty); in alloc_pages_on_heap()
|
| H A D | rte_malloc.c | 116 if (elem->dirty) { in rte_zmalloc_socket()
|
| /dpdk/doc/guides/vdpadevs/ |
| H A D | ifc.rst | 10 directly by DMA. Besides, it supports dirty page logging and device state 32 thread when LM happens, this thread will help device to log dirty pages. Thus 33 this mode does not require HW to implement a dirty page logging function block,
|
| /dpdk/doc/guides/howto/ |
| H A D | lm_bond_virtio_sriov.rst | 269 dirty sync count: 2 290 dirty sync count: 2
|
| H A D | lm_virtio_vhost_user.rst | 215 dirty sync count: 2
|
| /dpdk/doc/guides/rel_notes/ |
| H A D | release_19_02.rst | 91 devices to track dirty pages caused by DMA. the IFC driver has enabled this
|
| /dpdk/doc/guides/prog_guide/ |
| H A D | env_abstraction_layer.rst | 290 with all the data left in the mapped hugepages ("dirty" memory). 292 Memory allocator detects dirty segments and handles them accordingly, 971 * dirty - this flag is only meaningful when ``state`` is ``FREE``.
|