| /dpdk/drivers/net/avp/ |
| H A D | rte_avp_fifo.h | 40 fifo->write = 0; in avp_fifo_init() 54 unsigned int fifo_write = fifo->write; in avp_fifo_put() 67 fifo->write = fifo_write; in avp_fifo_put() 79 unsigned int fifo_write = fifo->write; in avp_fifo_get() 102 return (fifo->len + fifo->write - fifo->read) & (fifo->len - 1); in avp_fifo_count() 111 return (fifo->read - fifo->write - 1) & (fifo->len - 1); in avp_fifo_free_count()
|
| /dpdk/drivers/raw/ioat/ |
| H A D | rte_ioat_rawdev_fns.h | 121 unsigned short write = ioat->next_write; in __ioat_burst_capacity() local 122 unsigned short space = size - (write - read); in __ioat_burst_capacity() 134 unsigned short write = ioat->next_write; in __ioat_write_desc() local 136 unsigned short space = mask + read - write; in __ioat_write_desc() 144 ioat->next_write = write + 1; in __ioat_write_desc() 145 write &= mask; in __ioat_write_desc() 147 desc = &ioat->desc_ring[write]; in __ioat_write_desc() 151 (!(write & 0xF) << IOAT_COMP_UPDATE_SHIFT)); in __ioat_write_desc() 191 unsigned short write = ioat->next_write; in __ioat_fence() local 195 write = (write - 1) & mask; in __ioat_fence() [all …]
|
| /dpdk/lib/kni/ |
| H A D | rte_kni_fifo.h | 47 fifo->write = 0; in kni_fifo_init() 60 unsigned fifo_write = fifo->write; in kni_fifo_put() 72 __KNI_STORE_RELEASE(&fifo->write, fifo_write); in kni_fifo_put() 84 unsigned fifo_write = __KNI_LOAD_ACQUIRE(&fifo->write); in kni_fifo_get() 103 unsigned fifo_write = __KNI_LOAD_ACQUIRE(&fifo->write); in kni_fifo_count() 114 uint32_t fifo_write = __KNI_LOAD_ACQUIRE(&fifo->write); in kni_fifo_free_count()
|
| H A D | rte_kni_common.h | 66 unsigned write; /**< Next position to be written*/ member 69 volatile unsigned write; /**< Next position to be written*/
|
| /dpdk/drivers/event/sw/ |
| H A D | event_ring.h | 75 uint32_t write = r->write_idx; in rob_ring_enqueue() local 76 const uint32_t space = read + size - write; in rob_ring_enqueue() 79 r->ring[write & mask] = re; in rob_ring_enqueue() 89 const uint32_t write = r->write_idx; in rob_ring_dequeue() local 90 const uint32_t items = write - read; in rob_ring_dequeue()
|
| /dpdk/kernel/linux/kni/ |
| H A D | kni_fifo.h | 26 uint32_t fifo_write = fifo->write; in kni_fifo_put() 38 smp_store_release(&fifo->write, fifo_write); in kni_fifo_put() 51 uint32_t fifo_write = smp_load_acquire(&fifo->write); in kni_fifo_get() 71 uint32_t fifo_write = smp_load_acquire(&fifo->write); in kni_fifo_count() 82 uint32_t fifo_write = smp_load_acquire(&fifo->write); in kni_fifo_free_count()
|
| /dpdk/doc/guides/rawdevs/ |
| H A D | ntb.rst | 13 allocation for the peer to access and read/write allocated memory from peer. 45 NTB PMD needs kernel PCI driver to support write combining (WC) to get 70 echo "base=0x39bfa0000000 size=0x20000000 type=write-combining" >> /proc/mtrr 71 echo "base=0x39bfc0000000 size=0x20000000 type=write-combining" >> /proc/mtrr 78 reg00: base=0x000000000 ( 0MB), size= 2048MB, count=1: write-back 80 reg02: base=0x39bfa0000000 (60553728MB), size= 512MB, count=1: write-combining 81 reg03: base=0x39bfc0000000 (60554240MB), size= 512MB, count=1: write-combining 93 Since read/write remote system's memory are through PCI bus, remote read 94 is much more expensive than remote write. Thus, the enqueue and dequeue 142 So in this way, only remote write happens and remote read can be avoid
|
| /dpdk/drivers/net/axgbe/ |
| H A D | axgbe_rxtx.c | 248 err = AXGMAC_GET_BITS_LE(desc->write.desc3, in axgbe_recv_pkts() 252 error_status = desc->write.desc3 & AXGBE_ERR_STATUS; in axgbe_recv_pkts() 278 mbuf->hash.rss = rte_le_to_cpu_32(desc->write.desc1); in axgbe_recv_pkts() 279 etlt = AXGMAC_GET_BITS_LE(desc->write.desc3, in axgbe_recv_pkts() 286 AXGMAC_GET_BITS_LE(desc->write.desc0, in axgbe_recv_pkts() 385 err = AXGMAC_GET_BITS_LE(desc->write.desc3, in eth_axgbe_recv_scattered_pkts() 389 error_status = desc->write.desc3 & AXGBE_ERR_STATUS; in eth_axgbe_recv_scattered_pkts() 399 if (!AXGMAC_GET_BITS_LE(desc->write.desc3, in eth_axgbe_recv_scattered_pkts() 406 pkt_len = AXGMAC_GET_BITS_LE(desc->write.desc3, in eth_axgbe_recv_scattered_pkts() 422 etlt = AXGMAC_GET_BITS_LE(desc->write.desc3, in eth_axgbe_recv_scattered_pkts() [all …]
|
| /dpdk/examples/ip_pipeline/ |
| H A D | conn.c | 189 status = write(fd_client, in conn_poll_for_conn() 197 status = write(fd_client, in conn_poll_for_conn() 242 status = write(fd_client, in data_event_handle() 254 status = write(fd_client, in data_event_handle() 265 status = write(fd_client, in data_event_handle()
|
| /dpdk/drivers/net/softnic/ |
| H A D | conn.c | 192 status = write(fd_client, in softnic_conn_poll_for_conn() 200 status = write(fd_client, in softnic_conn_poll_for_conn() 246 status = write(fd_client, in data_event_handle() 258 status = write(fd_client, in data_event_handle() 269 status = write(fd_client, in data_event_handle()
|
| /dpdk/examples/pipeline/ |
| H A D | conn.c | 191 status = write(fd_client, in conn_poll_for_conn() 199 status = write(fd_client, in conn_poll_for_conn() 245 status = write(fd_client, in data_event_handle() 257 status = write(fd_client, in data_event_handle() 268 status = write(fd_client, in data_event_handle()
|
| /dpdk/lib/eal/unix/ |
| H A D | eal_unix_thread.c | 21 n = write(m2w, &c, 1); in eal_thread_wake_worker() 59 n = write(w2m, &c, 1); in eal_thread_ack_command()
|
| /dpdk/drivers/net/e1000/base/ |
| H A D | e1000_mbx.c | 84 else if (mbx->ops.write) in e1000_write_mbx() 85 ret_val = mbx->ops.write(hw, msg, size, mbx_id); in e1000_write_mbx() 259 if (!mbx->ops.write || !mbx->timeout) in e1000_write_posted_mbx() 263 ret_val = mbx->ops.write(hw, msg, size, mbx_id); in e1000_write_posted_mbx() 283 mbx->ops.write = e1000_null_mbx_transact; in e1000_init_mbx_ops_generic() 524 mbx->ops.write = e1000_write_mbx_vf; in e1000_init_mbx_params_vf() 750 mbx->ops.write = e1000_write_mbx_pf; in e1000_init_mbx_params_pf()
|
| /dpdk/drivers/dma/ioat/ |
| H A D | ioat_dmadev.c | 253 unsigned short write = ioat->next_write; in __write_desc() local 254 const unsigned short space = mask + read - write; in __write_desc() 260 ioat->next_write = write + 1; in __write_desc() 261 write &= mask; in __write_desc() 263 desc = &ioat->desc_ring[write]; in __write_desc() 274 ioat->desc_ring[(write - 1) & mask].u.control.fence = 1; in __write_desc() 526 unsigned short write = ioat->next_write; in ioat_burst_capacity() local 527 unsigned short space = size - (write - read); in ioat_burst_capacity()
|
| /dpdk/drivers/net/ixgbe/base/ |
| H A D | ixgbe_mbx.c | 54 } else if (mbx->ops.write) in ixgbe_write_mbx() 55 ret_val = mbx->ops.write(hw, msg, size, mbx_id); in ixgbe_write_mbx() 232 if (!mbx->ops.write || !mbx->timeout) in ixgbe_write_posted_mbx() 236 ret_val = mbx->ops.write(hw, msg, size, mbx_id); in ixgbe_write_posted_mbx() 483 mbx->ops.write = ixgbe_write_mbx_vf; in ixgbe_init_mbx_params_vf() 728 mbx->ops.write = ixgbe_write_mbx_pf; in ixgbe_init_mbx_params_pf()
|
| /dpdk/drivers/net/nfp/nfpcore/ |
| H A D | nfp_cpp_pcie_ops.c | 397 int write; member 421 priv->width.write = PULL_WIDTH(pp); in nfp6000_area_init() 424 priv->width.write > 0 && priv->width.read != priv->width.write) in nfp6000_area_init() 430 priv->width.bar = priv->width.write; in nfp6000_area_init() 592 width = priv->width.write; in nfp6000_area_write()
|
| /dpdk/drivers/net/txgbe/base/ |
| H A D | txgbe_mbx.c | 51 } else if (mbx->write) { in txgbe_write_mbx() 52 ret_val = mbx->write(hw, msg, size, mbx_id); in txgbe_write_mbx() 214 if (!mbx->write || !mbx->timeout) in txgbe_write_posted_mbx() 218 ret_val = mbx->write(hw, msg, size, mbx_id); in txgbe_write_posted_mbx()
|
| /dpdk/usertools/ |
| H A D | dpdk-devbind.py | 338 f.write(dev_id) 380 f.write("%s" % driver) 397 f.write("%04x %04x" % (int(dev["Vendor"], 16), 416 f.write(dev_id) 442 f.write("\00")
|
| /dpdk/buildtools/chkincs/ |
| H A D | gen_c_file_for_header.py | 12 cf.write(contents)
|
| /dpdk/doc/guides/howto/ |
| H A D | openwrt.rst | 92 To cross compile with meson build, you need to write a customized cross file 140 You can use the ``dd`` tool to write the OpenWrt image to the drive you 141 want to write the image on.
|
| /dpdk/drivers/net/virtio/virtio_user/ |
| H A D | vhost_vdpa.c | 186 if (write(data->vhostfd, &msg, sizeof(msg)) != sizeof(msg)) { in vhost_vdpa_iotlb_batch_begin() 212 if (write(data->vhostfd, &msg, sizeof(msg)) != sizeof(msg)) { in vhost_vdpa_iotlb_batch_end() 243 if (write(data->vhostfd, &msg, sizeof(msg)) != sizeof(msg)) { in vhost_vdpa_dma_map() 272 if (write(data->vhostfd, &msg, sizeof(msg)) != sizeof(msg)) { in vhost_vdpa_dma_unmap()
|
| /dpdk/lib/eal/linux/ |
| H A D | eal_log.c | 39 .write = console_log_write,
|
| /dpdk/lib/telemetry/ |
| H A D | telemetry_legacy.c | 154 if (write(s, out_buf, used) < 0) { in send_error_response() 175 if (write(s, out_buf, used) < 0) in perform_command()
|
| /dpdk/drivers/net/ngbe/base/ |
| H A D | ngbe_mbx.c | 51 } else if (mbx->write) { in ngbe_write_mbx() 52 ret_val = mbx->write(hw, msg, size, mbx_id); in ngbe_write_mbx()
|
| /dpdk/doc/guides/sample_app_ug/ |
| H A D | qos_scheduler.rst | 29 The TX thread, if present, reads from the TX ring and write the packets to the TX port. 91 * B = I/O RX lcore write burst size to the output software rings, 96 * D = Worker lcore write burst size to the NIC TX (the default value is 64) 106 * C = RX write-back threshold (the default value is 4) 114 * C = TX write-back threshold (the default value is 0)
|