| /dpdk/drivers/common/sfc_efx/base/ |
| H A D | mcdi_mon.c | 37 uint32_t page = 0; in mcdi_mon_decode_stats() local 71 page++; in mcdi_mon_decode_stats() 236 int page; in efx_mcdi_sensor_info_npages() local 241 page = 0; in efx_mcdi_sensor_info_npages() 261 *npagesp = page; in efx_mcdi_sensor_info_npages() 280 uint32_t page; in efx_mcdi_sensor_info() local 290 for (page = 0; page < npages; page++) { in efx_mcdi_sensor_info() 483 uint32_t page; in mcdi_mon_limits_update() local 489 page = 0; in mcdi_mon_limits_update() 490 page--; in mcdi_mon_limits_update() [all …]
|
| /dpdk/kernel/linux/kni/ |
| H A D | kni_dev.h | 101 struct page *page = NULL; in iova_to_phys() local 109 FOLL_TOUCH, &page, NULL, NULL); in iova_to_phys() 112 FOLL_TOUCH, &page, NULL, NULL); in iova_to_phys() 117 phys_addr = page_to_phys(page) | offset; in iova_to_phys() 118 put_page(page); in iova_to_phys()
|
| /dpdk/lib/vhost/ |
| H A D | vhost.h | 687 if (page) { in gpa_to_first_hpa() 689 page->guest_phys_addr + page->size) { in gpa_to_first_hpa() 691 page->host_iova; in gpa_to_first_hpa() 693 page->size) { in gpa_to_first_hpa() 695 page->size - gpa; in gpa_to_first_hpa() 697 page->host_iova; in gpa_to_first_hpa() 706 page->guest_phys_addr + page->size) { in gpa_to_first_hpa() 708 page->host_iova; in gpa_to_first_hpa() 710 page->size) { in gpa_to_first_hpa() 712 page->size - gpa; in gpa_to_first_hpa() [all …]
|
| H A D | vhost.c | 90 vhost_set_bit(page % 8, &log_base[page / 8]); in vhost_log_page() 96 uint64_t page; in __vhost_log_write() local 107 page = addr / VHOST_LOG_PAGE; in __vhost_log_write() 108 while (page * VHOST_LOG_PAGE < addr + len) { in __vhost_log_write() 110 page += 1; in __vhost_log_write() 173 uint64_t page) in vhost_log_cache_page() argument 216 uint64_t page; in __vhost_log_cache_write() local 224 page = addr / VHOST_LOG_PAGE; in __vhost_log_cache_write() 225 while (page * VHOST_LOG_PAGE < addr + len) { in __vhost_log_cache_write() 226 vhost_log_cache_page(dev, vq, page); in __vhost_log_cache_write() [all …]
|
| H A D | vhost_user.c | 121 struct guest_page *page; in async_dma_map() local 125 page = &dev->guest_pages[i]; in async_dma_map() 127 page->host_user_addr, in async_dma_map() 128 page->host_iova, in async_dma_map() 129 page->size); in async_dma_map() 153 page = &dev->guest_pages[i]; in async_dma_map() 155 page->host_user_addr, in async_dma_map() 156 page->host_iova, in async_dma_map() 157 page->size); in async_dma_map() 969 page->size = size; in add_one_guest_page() [all …]
|
| /dpdk/doc/guides/tools/ |
| H A D | hugepages.rst | 31 Print the current huge page configuration 35 Clear existing huge page reservation 39 Mount the huge page filesystem 43 Unmount the huge page filesystem 52 If not specified the default system huge page size is used. 72 To display current huge page settings::
|
| /dpdk/drivers/raw/ifpga/base/ |
| H A D | ifpga_sec_mgr.c | 403 static int n3000_reload_fpga(struct intel_max10_device *dev, int page) in n3000_reload_fpga() argument 409 if (!dev || ((page != 0) && (page != 1))) { in n3000_reload_fpga() 426 SFPGA_RP_LOAD | SFPGA_PAGE(page)); in n3000_reload_fpga() 443 FPGA_RP_LOAD | FPGA_PAGE(page)); in n3000_reload_fpga() 474 static int n3000_reload_bmc(struct intel_max10_device *dev, int page) in n3000_reload_bmc() argument 481 if (!dev || ((page != 0) && (page != 1))) { in n3000_reload_bmc() 490 CONFIG_SEL_S(page) | REBOOT_REQ); in n3000_reload_bmc() 492 val = (page == 0) ? 0x1 : 0x3; in n3000_reload_bmc() 520 static int n3000_reload(struct ifpga_sec_mgr *smgr, int type, int page) in n3000_reload() argument 529 psel = (page == IFPGA_BOOT_PAGE_FACTORY ? 0 : 1); in n3000_reload() [all …]
|
| H A D | ifpga_sec_mgr.h | 66 int (*reload)(struct ifpga_sec_mgr *smgr, int type, int page); 90 int fpga_reload(struct ifpga_fme_hw *fme, int type, int page);
|
| /dpdk/kernel/freebsd/contigmem/ |
| H A D | contigmem.c | 277 vm_page_t m_paddr, page; in contigmem_cdev_pager_fault() local 298 page = *mres; in contigmem_cdev_pager_fault() 300 vm_page_updatefake(page, paddr, memattr); in contigmem_cdev_pager_fault() 306 page = vm_page_getfake(paddr, memattr); in contigmem_cdev_pager_fault() 309 vm_page_replace(page, object, (*mres)->pindex, *mres); in contigmem_cdev_pager_fault() 311 vm_page_t mret = vm_page_replace(page, object, (*mres)->pindex); in contigmem_cdev_pager_fault() 318 *mres = page; in contigmem_cdev_pager_fault() 321 page->valid = VM_PAGE_BITS_ALL; in contigmem_cdev_pager_fault()
|
| /dpdk/drivers/net/e1000/base/ |
| H A D | e1000_phy.c | 3082 if ((page >= 768) || (page == 0 && reg == 25) || (reg == 31)) in e1000_get_phy_addr_for_bm_page() 3109 if (page == BM_WUC_PAGE) { in e1000_write_phy_reg_bm() 3169 if (page == BM_WUC_PAGE) { in e1000_read_phy_reg_bm() 3228 if (page == BM_WUC_PAGE) { in e1000_read_phy_reg_bm2() 3239 page); in e1000_read_phy_reg_bm2() 3284 page); in e1000_write_phy_reg_bm2() 3428 page); in e1000_access_phy_wakeup_reg_bm() 3540 if (page > 0 && page < HV_INTC_FC_PAGE_START) { in __e1000_read_phy_reg_hv() 3548 page = 0; in __e1000_read_phy_reg_hv() 3650 if (page > 0 && page < HV_INTC_FC_PAGE_START) { in __e1000_write_phy_reg_hv() [all …]
|
| H A D | e1000_ich8lan.h | 106 #define PHY_REG(page, reg) (((page) << PHY_PAGE_SHIFT) | \ argument
|
| H A D | e1000_phy.h | 45 s32 e1000_set_page_igp(struct e1000_hw *hw, u16 page); 135 #define BM_PHY_REG(page, reg) \ argument 137 (((page) & 0xFFFF) << PHY_PAGE_SHIFT) |\
|
| /dpdk/drivers/net/igc/base/ |
| H A D | igc_phy.c | 3125 if (page >= 768 || (page == 0 && reg == 25) || reg == 31) in igc_get_phy_addr_for_bm_page() 3152 if (page == BM_WUC_PAGE) { in igc_write_phy_reg_bm() 3212 if (page == BM_WUC_PAGE) { in igc_read_phy_reg_bm() 3271 if (page == BM_WUC_PAGE) { in igc_read_phy_reg_bm2() 3282 page); in igc_read_phy_reg_bm2() 3327 page); in igc_write_phy_reg_bm2() 3471 page); in igc_access_phy_wakeup_reg_bm() 3583 if (page > 0 && page < HV_INTC_FC_PAGE_START) { in __igc_read_phy_reg_hv() 3591 page = 0; in __igc_read_phy_reg_hv() 3693 if (page > 0 && page < HV_INTC_FC_PAGE_START) { in __igc_write_phy_reg_hv() [all …]
|
| H A D | igc_phy.h | 45 s32 igc_set_page_igp(struct igc_hw *hw, u16 page); 153 #define BM_PHY_REG(page, reg) ( \ argument 155 typeof(page) _page = (page); \
|
| H A D | igc_ich8lan.h | 105 #define PHY_REG(page, reg) (((page) << PHY_PAGE_SHIFT) | \ argument
|
| /dpdk/drivers/net/bnxt/hcapi/cfa/ |
| H A D | hcapi_cfa_p4.c | 101 uint64_t hcapi_get_table_page(struct hcapi_cfa_em_table *mem, uint32_t page) in hcapi_get_table_page() argument 114 addr = (uint64_t)mem->pg_tbl[level].pg_va_tbl[page]; in hcapi_get_table_page() 241 uint32_t page; in hcapi_cfa_p4_key_hw_op() local 246 page = key_obj->offset / key_tbl->page_size; in hcapi_cfa_p4_key_hw_op() 248 op->hw.base_addr = hcapi_get_table_page(em_tbl, page); in hcapi_cfa_p4_key_hw_op()
|
| /dpdk/lib/eal/windows/ |
| H A D | eal_memalloc.c | 44 PSAPI_WORKING_SET_EX_BLOCK *page; in alloc_seg() local 116 page = &info.VirtualAttributes; in alloc_seg() 117 if (!page->Valid || !page->LargePage) { in alloc_seg() 121 if (page->Node != numa_node) { in alloc_seg() 124 numa_node, socket_id, page->Node); in alloc_seg()
|
| /dpdk/drivers/net/i40e/base/ |
| H A D | i40e_hmc.c | 122 struct i40e_dma_mem *page = &mem; in i40e_add_pd_table_entry() local 145 page = rsrc_pg; in i40e_add_pd_table_entry() 148 ret_code = i40e_allocate_dma_mem(hw, page, i40e_mem_bp, in i40e_add_pd_table_entry() 156 i40e_memcpy(&pd_entry->bp.addr, page, in i40e_add_pd_table_entry() 161 page_desc = page->pa | 0x1; in i40e_add_pd_table_entry()
|
| /dpdk/lib/eal/linux/ |
| H A D | eal_memory.c | 90 uint64_t page, physaddr; in rte_mem_virt2phy() local 117 retval = read(fd, &page, PFN_MASK_SIZE); in rte_mem_virt2phy() 134 if ((page & 0x7fffffffffffffULL) == 0) in rte_mem_virt2phy() 137 physaddr = ((page & 0x7fffffffffffffULL) * page_size) in rte_mem_virt2phy() 574 int page, nrpages = 0; in unlink_hugepage_files() local 584 for (page = 0; page < nrpages; page++) { in unlink_hugepage_files() 585 struct hugepage_file *hp = &hugepg_tbl[page]; in unlink_hugepage_files() 605 int page, nrpages = 0; in unmap_unneeded_hugepages() local 619 for (page = 0; page < nrpages; page++) { in unmap_unneeded_hugepages() 620 struct hugepage_file *hp = &hugepg_tbl[page]; in unmap_unneeded_hugepages()
|
| /dpdk/doc/guides/nics/ |
| H A D | e1000em.rst | 81 `http://www.linux-kvm.org/page/HOWTO1 <http://www.linux-kvm.org/page/HOWTO1>`_. 95 — `http://www.linux-kvm.org/page/Networking <http://www.linux-kvm.org/page/Networking>`_
|
| H A D | af_xdp.rst | 85 per page. In the PMD we report the maximum MTU for zero copy to be equal 86 to the page size less the frame overhead introduced by AF_XDP (XDP HR = 256) 87 and DPDK (frame headroom = 320). With a 4K page size this works out at 3520. 92 which is less than the page size (4096B) may be 3072B. In this case, the maximum
|
| /dpdk/drivers/bus/vmbus/ |
| H A D | rte_vmbus_reg.h | 143 uint64_t page[0]; member 150 uint64_t page; member
|
| /dpdk/drivers/net/ngbe/base/ |
| H A D | ngbe_phy.h | 29 u16 page; member
|
| /dpdk/drivers/raw/ifpga/ |
| H A D | rte_pmd_ifpga.h | 257 rte_pmd_ifpga_reload(uint16_t dev_id, int type, int page);
|
| /dpdk/doc/guides/vdpadevs/ |
| H A D | ifc.rst | 10 directly by DMA. Besides, it supports dirty page logging and device state 33 this mode does not require HW to implement a dirty page logging function block,
|