1b6e66be2SVincenzo Maffione /* 2b6e66be2SVincenzo Maffione * Copyright (C) 2016-2018 Vincenzo Maffione 3b6e66be2SVincenzo Maffione * Copyright (C) 2015 Stefano Garzarella 4b6e66be2SVincenzo Maffione * All rights reserved. 5b6e66be2SVincenzo Maffione * 6b6e66be2SVincenzo Maffione * Redistribution and use in source and binary forms, with or without 7b6e66be2SVincenzo Maffione * modification, are permitted provided that the following conditions 8b6e66be2SVincenzo Maffione * are met: 9b6e66be2SVincenzo Maffione * 1. Redistributions of source code must retain the above copyright 10b6e66be2SVincenzo Maffione * notice, this list of conditions and the following disclaimer. 11b6e66be2SVincenzo Maffione * 2. Redistributions in binary form must reproduce the above copyright 12b6e66be2SVincenzo Maffione * notice, this list of conditions and the following disclaimer in the 13b6e66be2SVincenzo Maffione * documentation and/or other materials provided with the distribution. 14b6e66be2SVincenzo Maffione * 15b6e66be2SVincenzo Maffione * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 16b6e66be2SVincenzo Maffione * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 17b6e66be2SVincenzo Maffione * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 18b6e66be2SVincenzo Maffione * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 19b6e66be2SVincenzo Maffione * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 20b6e66be2SVincenzo Maffione * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 21b6e66be2SVincenzo Maffione * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 22b6e66be2SVincenzo Maffione * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 23b6e66be2SVincenzo Maffione * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 24b6e66be2SVincenzo Maffione * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 25b6e66be2SVincenzo Maffione * SUCH DAMAGE. 26b6e66be2SVincenzo Maffione * 27b6e66be2SVincenzo Maffione * $FreeBSD$ 28b6e66be2SVincenzo Maffione */ 29b6e66be2SVincenzo Maffione 30b6e66be2SVincenzo Maffione /* 31b6e66be2SVincenzo Maffione * common headers 32b6e66be2SVincenzo Maffione */ 33b6e66be2SVincenzo Maffione #if defined(__FreeBSD__) 34b6e66be2SVincenzo Maffione #include <sys/cdefs.h> 35b6e66be2SVincenzo Maffione #include <sys/param.h> 36b6e66be2SVincenzo Maffione #include <sys/kernel.h> 37b6e66be2SVincenzo Maffione #include <sys/types.h> 38b6e66be2SVincenzo Maffione #include <sys/selinfo.h> 39b6e66be2SVincenzo Maffione #include <sys/socket.h> 40b6e66be2SVincenzo Maffione #include <net/if.h> 41b6e66be2SVincenzo Maffione #include <net/if_var.h> 42b6e66be2SVincenzo Maffione #include <machine/bus.h> 43b6e66be2SVincenzo Maffione 44b6e66be2SVincenzo Maffione #define usleep_range(_1, _2) \ 45b6e66be2SVincenzo Maffione pause_sbt("sync-kloop-sleep", SBT_1US * _1, SBT_1US * 1, C_ABSOLUTE) 46b6e66be2SVincenzo Maffione 47b6e66be2SVincenzo Maffione #elif defined(linux) 48b6e66be2SVincenzo Maffione #include <bsd_glue.h> 49b6e66be2SVincenzo Maffione #include <linux/file.h> 50b6e66be2SVincenzo Maffione #include <linux/eventfd.h> 51b6e66be2SVincenzo Maffione #endif 52b6e66be2SVincenzo Maffione 53b6e66be2SVincenzo Maffione #include <net/netmap.h> 54b6e66be2SVincenzo Maffione #include <dev/netmap/netmap_kern.h> 55b6e66be2SVincenzo Maffione #include <net/netmap_virt.h> 56b6e66be2SVincenzo Maffione #include <dev/netmap/netmap_mem2.h> 57b6e66be2SVincenzo Maffione 58b6e66be2SVincenzo Maffione /* Support for eventfd-based notifications. */ 59b6e66be2SVincenzo Maffione #if defined(linux) 60b6e66be2SVincenzo Maffione #define SYNC_KLOOP_POLL 61b6e66be2SVincenzo Maffione #endif 62b6e66be2SVincenzo Maffione 63b6e66be2SVincenzo Maffione /* Write kring pointers (hwcur, hwtail) to the CSB. 64b6e66be2SVincenzo Maffione * This routine is coupled with ptnetmap_guest_read_kring_csb(). */ 65b6e66be2SVincenzo Maffione static inline void 66b6e66be2SVincenzo Maffione sync_kloop_kernel_write(struct nm_csb_ktoa __user *ptr, uint32_t hwcur, 67b6e66be2SVincenzo Maffione uint32_t hwtail) 68b6e66be2SVincenzo Maffione { 69*f79ba6d7SVincenzo Maffione /* Issue a first store-store barrier to make sure writes to the 70*f79ba6d7SVincenzo Maffione * netmap ring do not overcome updates on ktoa->hwcur and ktoa->hwtail. */ 71*f79ba6d7SVincenzo Maffione nm_stst_barrier(); 72*f79ba6d7SVincenzo Maffione 73b6e66be2SVincenzo Maffione /* 74*f79ba6d7SVincenzo Maffione * The same scheme used in nm_sync_kloop_appl_write() applies here. 75b6e66be2SVincenzo Maffione * We allow the application to read a value of hwcur more recent than the value 76b6e66be2SVincenzo Maffione * of hwtail, since this would anyway result in a consistent view of the 77b6e66be2SVincenzo Maffione * ring state (and hwcur can never wraparound hwtail, since hwcur must be 78b6e66be2SVincenzo Maffione * behind head). 79b6e66be2SVincenzo Maffione * 80b6e66be2SVincenzo Maffione * The following memory barrier scheme is used to make this happen: 81b6e66be2SVincenzo Maffione * 82b6e66be2SVincenzo Maffione * Application Kernel 83b6e66be2SVincenzo Maffione * 84b6e66be2SVincenzo Maffione * STORE(hwcur) LOAD(hwtail) 85*f79ba6d7SVincenzo Maffione * wmb() <-------------> rmb() 86b6e66be2SVincenzo Maffione * STORE(hwtail) LOAD(hwcur) 87b6e66be2SVincenzo Maffione */ 88b6e66be2SVincenzo Maffione CSB_WRITE(ptr, hwcur, hwcur); 89b6e66be2SVincenzo Maffione nm_stst_barrier(); 90b6e66be2SVincenzo Maffione CSB_WRITE(ptr, hwtail, hwtail); 91b6e66be2SVincenzo Maffione } 92b6e66be2SVincenzo Maffione 93b6e66be2SVincenzo Maffione /* Read kring pointers (head, cur, sync_flags) from the CSB. 94b6e66be2SVincenzo Maffione * This routine is coupled with ptnetmap_guest_write_kring_csb(). */ 95b6e66be2SVincenzo Maffione static inline void 96b6e66be2SVincenzo Maffione sync_kloop_kernel_read(struct nm_csb_atok __user *ptr, 97b6e66be2SVincenzo Maffione struct netmap_ring *shadow_ring, 98b6e66be2SVincenzo Maffione uint32_t num_slots) 99b6e66be2SVincenzo Maffione { 100b6e66be2SVincenzo Maffione /* 101b6e66be2SVincenzo Maffione * We place a memory barrier to make sure that the update of head never 102b6e66be2SVincenzo Maffione * overtakes the update of cur. 103*f79ba6d7SVincenzo Maffione * (see explanation in sync_kloop_kernel_write). 104b6e66be2SVincenzo Maffione */ 105b6e66be2SVincenzo Maffione CSB_READ(ptr, head, shadow_ring->head); 106*f79ba6d7SVincenzo Maffione nm_ldld_barrier(); 107b6e66be2SVincenzo Maffione CSB_READ(ptr, cur, shadow_ring->cur); 108b6e66be2SVincenzo Maffione CSB_READ(ptr, sync_flags, shadow_ring->flags); 109*f79ba6d7SVincenzo Maffione 110*f79ba6d7SVincenzo Maffione /* Make sure that loads from atok->head and atok->cur are not delayed 111*f79ba6d7SVincenzo Maffione * after the loads from the netmap ring. */ 112*f79ba6d7SVincenzo Maffione nm_ldld_barrier(); 113b6e66be2SVincenzo Maffione } 114b6e66be2SVincenzo Maffione 115b6e66be2SVincenzo Maffione /* Enable or disable application --> kernel kicks. */ 116b6e66be2SVincenzo Maffione static inline void 117b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(struct nm_csb_ktoa __user *csb_ktoa, uint32_t val) 118b6e66be2SVincenzo Maffione { 119b6e66be2SVincenzo Maffione CSB_WRITE(csb_ktoa, kern_need_kick, val); 120b6e66be2SVincenzo Maffione } 121b6e66be2SVincenzo Maffione 122dde885deSVincenzo Maffione #ifdef SYNC_KLOOP_POLL 123b6e66be2SVincenzo Maffione /* Are application interrupt enabled or disabled? */ 124b6e66be2SVincenzo Maffione static inline uint32_t 125b6e66be2SVincenzo Maffione csb_atok_intr_enabled(struct nm_csb_atok __user *csb_atok) 126b6e66be2SVincenzo Maffione { 127b6e66be2SVincenzo Maffione uint32_t v; 128b6e66be2SVincenzo Maffione 129b6e66be2SVincenzo Maffione CSB_READ(csb_atok, appl_need_kick, v); 130b6e66be2SVincenzo Maffione 131b6e66be2SVincenzo Maffione return v; 132b6e66be2SVincenzo Maffione } 133dde885deSVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 134b6e66be2SVincenzo Maffione 135b6e66be2SVincenzo Maffione static inline void 136b6e66be2SVincenzo Maffione sync_kloop_kring_dump(const char *title, const struct netmap_kring *kring) 137b6e66be2SVincenzo Maffione { 138*f79ba6d7SVincenzo Maffione nm_prinf("%s, kring %s, hwcur %d, rhead %d, " 139*f79ba6d7SVincenzo Maffione "rcur %d, rtail %d, hwtail %d", 140*f79ba6d7SVincenzo Maffione title, kring->name, kring->nr_hwcur, kring->rhead, 141*f79ba6d7SVincenzo Maffione kring->rcur, kring->rtail, kring->nr_hwtail); 142b6e66be2SVincenzo Maffione } 143b6e66be2SVincenzo Maffione 144b6e66be2SVincenzo Maffione struct sync_kloop_ring_args { 145b6e66be2SVincenzo Maffione struct netmap_kring *kring; 146b6e66be2SVincenzo Maffione struct nm_csb_atok *csb_atok; 147b6e66be2SVincenzo Maffione struct nm_csb_ktoa *csb_ktoa; 148b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 149b6e66be2SVincenzo Maffione struct eventfd_ctx *irq_ctx; 150b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 151b6e66be2SVincenzo Maffione }; 152b6e66be2SVincenzo Maffione 153b6e66be2SVincenzo Maffione static void 154b6e66be2SVincenzo Maffione netmap_sync_kloop_tx_ring(const struct sync_kloop_ring_args *a) 155b6e66be2SVincenzo Maffione { 156b6e66be2SVincenzo Maffione struct netmap_kring *kring = a->kring; 157b6e66be2SVincenzo Maffione struct nm_csb_atok *csb_atok = a->csb_atok; 158b6e66be2SVincenzo Maffione struct nm_csb_ktoa *csb_ktoa = a->csb_ktoa; 159b6e66be2SVincenzo Maffione struct netmap_ring shadow_ring; /* shadow copy of the netmap_ring */ 160b6e66be2SVincenzo Maffione bool more_txspace = false; 161b6e66be2SVincenzo Maffione uint32_t num_slots; 162b6e66be2SVincenzo Maffione int batch; 163b6e66be2SVincenzo Maffione 164b6e66be2SVincenzo Maffione num_slots = kring->nkr_num_slots; 165b6e66be2SVincenzo Maffione 166b6e66be2SVincenzo Maffione /* Disable application --> kernel notifications. */ 167b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 0); 168b6e66be2SVincenzo Maffione /* Copy the application kring pointers from the CSB */ 169b6e66be2SVincenzo Maffione sync_kloop_kernel_read(csb_atok, &shadow_ring, num_slots); 170b6e66be2SVincenzo Maffione 171b6e66be2SVincenzo Maffione for (;;) { 172b6e66be2SVincenzo Maffione batch = shadow_ring.head - kring->nr_hwcur; 173b6e66be2SVincenzo Maffione if (batch < 0) 174b6e66be2SVincenzo Maffione batch += num_slots; 175b6e66be2SVincenzo Maffione 176b6e66be2SVincenzo Maffione #ifdef PTN_TX_BATCH_LIM 177b6e66be2SVincenzo Maffione if (batch > PTN_TX_BATCH_LIM(num_slots)) { 178b6e66be2SVincenzo Maffione /* If application moves ahead too fast, let's cut the move so 179b6e66be2SVincenzo Maffione * that we don't exceed our batch limit. */ 180b6e66be2SVincenzo Maffione uint32_t head_lim = kring->nr_hwcur + PTN_TX_BATCH_LIM(num_slots); 181b6e66be2SVincenzo Maffione 182b6e66be2SVincenzo Maffione if (head_lim >= num_slots) 183b6e66be2SVincenzo Maffione head_lim -= num_slots; 184b6e66be2SVincenzo Maffione nm_prdis(1, "batch: %d head: %d head_lim: %d", batch, shadow_ring.head, 185b6e66be2SVincenzo Maffione head_lim); 186b6e66be2SVincenzo Maffione shadow_ring.head = head_lim; 187b6e66be2SVincenzo Maffione batch = PTN_TX_BATCH_LIM(num_slots); 188b6e66be2SVincenzo Maffione } 189b6e66be2SVincenzo Maffione #endif /* PTN_TX_BATCH_LIM */ 190b6e66be2SVincenzo Maffione 191b6e66be2SVincenzo Maffione if (nm_kr_txspace(kring) <= (num_slots >> 1)) { 192b6e66be2SVincenzo Maffione shadow_ring.flags |= NAF_FORCE_RECLAIM; 193b6e66be2SVincenzo Maffione } 194b6e66be2SVincenzo Maffione 195b6e66be2SVincenzo Maffione /* Netmap prologue */ 196b6e66be2SVincenzo Maffione shadow_ring.tail = kring->rtail; 197b6e66be2SVincenzo Maffione if (unlikely(nm_txsync_prologue(kring, &shadow_ring) >= num_slots)) { 198b6e66be2SVincenzo Maffione /* Reinit ring and enable notifications. */ 199b6e66be2SVincenzo Maffione netmap_ring_reinit(kring); 200b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 1); 201b6e66be2SVincenzo Maffione break; 202b6e66be2SVincenzo Maffione } 203b6e66be2SVincenzo Maffione 204b6e66be2SVincenzo Maffione if (unlikely(netmap_debug & NM_DEBUG_TXSYNC)) { 205b6e66be2SVincenzo Maffione sync_kloop_kring_dump("pre txsync", kring); 206b6e66be2SVincenzo Maffione } 207b6e66be2SVincenzo Maffione 208b6e66be2SVincenzo Maffione if (unlikely(kring->nm_sync(kring, shadow_ring.flags))) { 209b6e66be2SVincenzo Maffione /* Reenable notifications. */ 210b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 1); 211b6e66be2SVincenzo Maffione nm_prerr("txsync() failed"); 212b6e66be2SVincenzo Maffione break; 213b6e66be2SVincenzo Maffione } 214b6e66be2SVincenzo Maffione 215b6e66be2SVincenzo Maffione /* 216b6e66be2SVincenzo Maffione * Finalize 217b6e66be2SVincenzo Maffione * Copy kernel hwcur and hwtail into the CSB for the application sync(), and 218b6e66be2SVincenzo Maffione * do the nm_sync_finalize. 219b6e66be2SVincenzo Maffione */ 220b6e66be2SVincenzo Maffione sync_kloop_kernel_write(csb_ktoa, kring->nr_hwcur, 221b6e66be2SVincenzo Maffione kring->nr_hwtail); 222b6e66be2SVincenzo Maffione if (kring->rtail != kring->nr_hwtail) { 223b6e66be2SVincenzo Maffione /* Some more room available in the parent adapter. */ 224b6e66be2SVincenzo Maffione kring->rtail = kring->nr_hwtail; 225b6e66be2SVincenzo Maffione more_txspace = true; 226b6e66be2SVincenzo Maffione } 227b6e66be2SVincenzo Maffione 228b6e66be2SVincenzo Maffione if (unlikely(netmap_debug & NM_DEBUG_TXSYNC)) { 229b6e66be2SVincenzo Maffione sync_kloop_kring_dump("post txsync", kring); 230b6e66be2SVincenzo Maffione } 231b6e66be2SVincenzo Maffione 232b6e66be2SVincenzo Maffione /* Interrupt the application if needed. */ 233b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 234b6e66be2SVincenzo Maffione if (a->irq_ctx && more_txspace && csb_atok_intr_enabled(csb_atok)) { 235b6e66be2SVincenzo Maffione /* Disable application kick to avoid sending unnecessary kicks */ 236b6e66be2SVincenzo Maffione eventfd_signal(a->irq_ctx, 1); 237b6e66be2SVincenzo Maffione more_txspace = false; 238b6e66be2SVincenzo Maffione } 239b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 240b6e66be2SVincenzo Maffione 241b6e66be2SVincenzo Maffione /* Read CSB to see if there is more work to do. */ 242b6e66be2SVincenzo Maffione sync_kloop_kernel_read(csb_atok, &shadow_ring, num_slots); 243b6e66be2SVincenzo Maffione if (shadow_ring.head == kring->rhead) { 244b6e66be2SVincenzo Maffione /* 245b6e66be2SVincenzo Maffione * No more packets to transmit. We enable notifications and 246b6e66be2SVincenzo Maffione * go to sleep, waiting for a kick from the application when new 247b6e66be2SVincenzo Maffione * new slots are ready for transmission. 248b6e66be2SVincenzo Maffione */ 249b6e66be2SVincenzo Maffione /* Reenable notifications. */ 250b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 1); 251*f79ba6d7SVincenzo Maffione /* Double check, with store-load memory barrier. */ 252*f79ba6d7SVincenzo Maffione nm_stld_barrier(); 253b6e66be2SVincenzo Maffione sync_kloop_kernel_read(csb_atok, &shadow_ring, num_slots); 254b6e66be2SVincenzo Maffione if (shadow_ring.head != kring->rhead) { 255b6e66be2SVincenzo Maffione /* We won the race condition, there are more packets to 256b6e66be2SVincenzo Maffione * transmit. Disable notifications and do another cycle */ 257b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 0); 258b6e66be2SVincenzo Maffione continue; 259b6e66be2SVincenzo Maffione } 260b6e66be2SVincenzo Maffione break; 261b6e66be2SVincenzo Maffione } 262b6e66be2SVincenzo Maffione 263b6e66be2SVincenzo Maffione if (nm_kr_txempty(kring)) { 264b6e66be2SVincenzo Maffione /* No more available TX slots. We stop waiting for a notification 265b6e66be2SVincenzo Maffione * from the backend (netmap_tx_irq). */ 266b6e66be2SVincenzo Maffione nm_prdis(1, "TX ring"); 267b6e66be2SVincenzo Maffione break; 268b6e66be2SVincenzo Maffione } 269b6e66be2SVincenzo Maffione } 270b6e66be2SVincenzo Maffione 271b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 272b6e66be2SVincenzo Maffione if (a->irq_ctx && more_txspace && csb_atok_intr_enabled(csb_atok)) { 273b6e66be2SVincenzo Maffione eventfd_signal(a->irq_ctx, 1); 274b6e66be2SVincenzo Maffione } 275b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 276b6e66be2SVincenzo Maffione } 277b6e66be2SVincenzo Maffione 278b6e66be2SVincenzo Maffione /* RX cycle without receive any packets */ 279b6e66be2SVincenzo Maffione #define SYNC_LOOP_RX_DRY_CYCLES_MAX 2 280b6e66be2SVincenzo Maffione 281b6e66be2SVincenzo Maffione static inline int 282b6e66be2SVincenzo Maffione sync_kloop_norxslots(struct netmap_kring *kring, uint32_t g_head) 283b6e66be2SVincenzo Maffione { 284b6e66be2SVincenzo Maffione return (NM_ACCESS_ONCE(kring->nr_hwtail) == nm_prev(g_head, 285b6e66be2SVincenzo Maffione kring->nkr_num_slots - 1)); 286b6e66be2SVincenzo Maffione } 287b6e66be2SVincenzo Maffione 288b6e66be2SVincenzo Maffione static void 289b6e66be2SVincenzo Maffione netmap_sync_kloop_rx_ring(const struct sync_kloop_ring_args *a) 290b6e66be2SVincenzo Maffione { 291b6e66be2SVincenzo Maffione 292b6e66be2SVincenzo Maffione struct netmap_kring *kring = a->kring; 293b6e66be2SVincenzo Maffione struct nm_csb_atok *csb_atok = a->csb_atok; 294b6e66be2SVincenzo Maffione struct nm_csb_ktoa *csb_ktoa = a->csb_ktoa; 295b6e66be2SVincenzo Maffione struct netmap_ring shadow_ring; /* shadow copy of the netmap_ring */ 296b6e66be2SVincenzo Maffione int dry_cycles = 0; 297b6e66be2SVincenzo Maffione bool some_recvd = false; 298b6e66be2SVincenzo Maffione uint32_t num_slots; 299b6e66be2SVincenzo Maffione 300b6e66be2SVincenzo Maffione num_slots = kring->nkr_num_slots; 301b6e66be2SVincenzo Maffione 302b6e66be2SVincenzo Maffione /* Get RX csb_atok and csb_ktoa pointers from the CSB. */ 303b6e66be2SVincenzo Maffione num_slots = kring->nkr_num_slots; 304b6e66be2SVincenzo Maffione 305b6e66be2SVincenzo Maffione /* Disable notifications. */ 306b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 0); 307b6e66be2SVincenzo Maffione /* Copy the application kring pointers from the CSB */ 308b6e66be2SVincenzo Maffione sync_kloop_kernel_read(csb_atok, &shadow_ring, num_slots); 309b6e66be2SVincenzo Maffione 310b6e66be2SVincenzo Maffione for (;;) { 311b6e66be2SVincenzo Maffione uint32_t hwtail; 312b6e66be2SVincenzo Maffione 313b6e66be2SVincenzo Maffione /* Netmap prologue */ 314b6e66be2SVincenzo Maffione shadow_ring.tail = kring->rtail; 315b6e66be2SVincenzo Maffione if (unlikely(nm_rxsync_prologue(kring, &shadow_ring) >= num_slots)) { 316b6e66be2SVincenzo Maffione /* Reinit ring and enable notifications. */ 317b6e66be2SVincenzo Maffione netmap_ring_reinit(kring); 318b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 1); 319b6e66be2SVincenzo Maffione break; 320b6e66be2SVincenzo Maffione } 321b6e66be2SVincenzo Maffione 322b6e66be2SVincenzo Maffione if (unlikely(netmap_debug & NM_DEBUG_RXSYNC)) { 323b6e66be2SVincenzo Maffione sync_kloop_kring_dump("pre rxsync", kring); 324b6e66be2SVincenzo Maffione } 325b6e66be2SVincenzo Maffione 326b6e66be2SVincenzo Maffione if (unlikely(kring->nm_sync(kring, shadow_ring.flags))) { 327b6e66be2SVincenzo Maffione /* Reenable notifications. */ 328b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 1); 329b6e66be2SVincenzo Maffione nm_prerr("rxsync() failed"); 330b6e66be2SVincenzo Maffione break; 331b6e66be2SVincenzo Maffione } 332b6e66be2SVincenzo Maffione 333b6e66be2SVincenzo Maffione /* 334b6e66be2SVincenzo Maffione * Finalize 335b6e66be2SVincenzo Maffione * Copy kernel hwcur and hwtail into the CSB for the application sync() 336b6e66be2SVincenzo Maffione */ 337b6e66be2SVincenzo Maffione hwtail = NM_ACCESS_ONCE(kring->nr_hwtail); 338b6e66be2SVincenzo Maffione sync_kloop_kernel_write(csb_ktoa, kring->nr_hwcur, hwtail); 339b6e66be2SVincenzo Maffione if (kring->rtail != hwtail) { 340b6e66be2SVincenzo Maffione kring->rtail = hwtail; 341b6e66be2SVincenzo Maffione some_recvd = true; 342b6e66be2SVincenzo Maffione dry_cycles = 0; 343b6e66be2SVincenzo Maffione } else { 344b6e66be2SVincenzo Maffione dry_cycles++; 345b6e66be2SVincenzo Maffione } 346b6e66be2SVincenzo Maffione 347b6e66be2SVincenzo Maffione if (unlikely(netmap_debug & NM_DEBUG_RXSYNC)) { 348b6e66be2SVincenzo Maffione sync_kloop_kring_dump("post rxsync", kring); 349b6e66be2SVincenzo Maffione } 350b6e66be2SVincenzo Maffione 351b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 352b6e66be2SVincenzo Maffione /* Interrupt the application if needed. */ 353b6e66be2SVincenzo Maffione if (a->irq_ctx && some_recvd && csb_atok_intr_enabled(csb_atok)) { 354b6e66be2SVincenzo Maffione /* Disable application kick to avoid sending unnecessary kicks */ 355b6e66be2SVincenzo Maffione eventfd_signal(a->irq_ctx, 1); 356b6e66be2SVincenzo Maffione some_recvd = false; 357b6e66be2SVincenzo Maffione } 358b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 359b6e66be2SVincenzo Maffione 360b6e66be2SVincenzo Maffione /* Read CSB to see if there is more work to do. */ 361b6e66be2SVincenzo Maffione sync_kloop_kernel_read(csb_atok, &shadow_ring, num_slots); 362b6e66be2SVincenzo Maffione if (sync_kloop_norxslots(kring, shadow_ring.head)) { 363b6e66be2SVincenzo Maffione /* 364b6e66be2SVincenzo Maffione * No more slots available for reception. We enable notification and 365b6e66be2SVincenzo Maffione * go to sleep, waiting for a kick from the application when new receive 366b6e66be2SVincenzo Maffione * slots are available. 367b6e66be2SVincenzo Maffione */ 368b6e66be2SVincenzo Maffione /* Reenable notifications. */ 369b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 1); 370*f79ba6d7SVincenzo Maffione /* Double check, with store-load memory barrier. */ 371*f79ba6d7SVincenzo Maffione nm_stld_barrier(); 372b6e66be2SVincenzo Maffione sync_kloop_kernel_read(csb_atok, &shadow_ring, num_slots); 373b6e66be2SVincenzo Maffione if (!sync_kloop_norxslots(kring, shadow_ring.head)) { 374b6e66be2SVincenzo Maffione /* We won the race condition, more slots are available. Disable 375b6e66be2SVincenzo Maffione * notifications and do another cycle. */ 376b6e66be2SVincenzo Maffione csb_ktoa_kick_enable(csb_ktoa, 0); 377b6e66be2SVincenzo Maffione continue; 378b6e66be2SVincenzo Maffione } 379b6e66be2SVincenzo Maffione break; 380b6e66be2SVincenzo Maffione } 381b6e66be2SVincenzo Maffione 382b6e66be2SVincenzo Maffione hwtail = NM_ACCESS_ONCE(kring->nr_hwtail); 383b6e66be2SVincenzo Maffione if (unlikely(hwtail == kring->rhead || 384b6e66be2SVincenzo Maffione dry_cycles >= SYNC_LOOP_RX_DRY_CYCLES_MAX)) { 385b6e66be2SVincenzo Maffione /* No more packets to be read from the backend. We stop and 386b6e66be2SVincenzo Maffione * wait for a notification from the backend (netmap_rx_irq). */ 387b6e66be2SVincenzo Maffione nm_prdis(1, "nr_hwtail: %d rhead: %d dry_cycles: %d", 388b6e66be2SVincenzo Maffione hwtail, kring->rhead, dry_cycles); 389b6e66be2SVincenzo Maffione break; 390b6e66be2SVincenzo Maffione } 391b6e66be2SVincenzo Maffione } 392b6e66be2SVincenzo Maffione 393b6e66be2SVincenzo Maffione nm_kr_put(kring); 394b6e66be2SVincenzo Maffione 395b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 396b6e66be2SVincenzo Maffione /* Interrupt the application if needed. */ 397b6e66be2SVincenzo Maffione if (a->irq_ctx && some_recvd && csb_atok_intr_enabled(csb_atok)) { 398b6e66be2SVincenzo Maffione eventfd_signal(a->irq_ctx, 1); 399b6e66be2SVincenzo Maffione } 400b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 401b6e66be2SVincenzo Maffione } 402b6e66be2SVincenzo Maffione 403b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 404b6e66be2SVincenzo Maffione struct sync_kloop_poll_entry { 405b6e66be2SVincenzo Maffione /* Support for receiving notifications from 406b6e66be2SVincenzo Maffione * a netmap ring or from the application. */ 407b6e66be2SVincenzo Maffione struct file *filp; 408b6e66be2SVincenzo Maffione wait_queue_t wait; 409b6e66be2SVincenzo Maffione wait_queue_head_t *wqh; 410b6e66be2SVincenzo Maffione 411b6e66be2SVincenzo Maffione /* Support for sending notifications to the application. */ 412b6e66be2SVincenzo Maffione struct eventfd_ctx *irq_ctx; 413b6e66be2SVincenzo Maffione struct file *irq_filp; 414b6e66be2SVincenzo Maffione }; 415b6e66be2SVincenzo Maffione 416b6e66be2SVincenzo Maffione struct sync_kloop_poll_ctx { 417b6e66be2SVincenzo Maffione poll_table wait_table; 418b6e66be2SVincenzo Maffione unsigned int next_entry; 419b6e66be2SVincenzo Maffione unsigned int num_entries; 420b6e66be2SVincenzo Maffione struct sync_kloop_poll_entry entries[0]; 421b6e66be2SVincenzo Maffione }; 422b6e66be2SVincenzo Maffione 423b6e66be2SVincenzo Maffione static void 424b6e66be2SVincenzo Maffione sync_kloop_poll_table_queue_proc(struct file *file, wait_queue_head_t *wqh, 425b6e66be2SVincenzo Maffione poll_table *pt) 426b6e66be2SVincenzo Maffione { 427b6e66be2SVincenzo Maffione struct sync_kloop_poll_ctx *poll_ctx = 428b6e66be2SVincenzo Maffione container_of(pt, struct sync_kloop_poll_ctx, wait_table); 429b6e66be2SVincenzo Maffione struct sync_kloop_poll_entry *entry = poll_ctx->entries + 430b6e66be2SVincenzo Maffione poll_ctx->next_entry; 431b6e66be2SVincenzo Maffione 432b6e66be2SVincenzo Maffione BUG_ON(poll_ctx->next_entry >= poll_ctx->num_entries); 433b6e66be2SVincenzo Maffione entry->wqh = wqh; 434b6e66be2SVincenzo Maffione entry->filp = file; 435b6e66be2SVincenzo Maffione /* Use the default wake up function. */ 436b6e66be2SVincenzo Maffione init_waitqueue_entry(&entry->wait, current); 437b6e66be2SVincenzo Maffione add_wait_queue(wqh, &entry->wait); 438b6e66be2SVincenzo Maffione poll_ctx->next_entry++; 439b6e66be2SVincenzo Maffione } 440b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 441b6e66be2SVincenzo Maffione 442b6e66be2SVincenzo Maffione int 443b6e66be2SVincenzo Maffione netmap_sync_kloop(struct netmap_priv_d *priv, struct nmreq_header *hdr) 444b6e66be2SVincenzo Maffione { 445b6e66be2SVincenzo Maffione struct nmreq_sync_kloop_start *req = 446b6e66be2SVincenzo Maffione (struct nmreq_sync_kloop_start *)(uintptr_t)hdr->nr_body; 447b6e66be2SVincenzo Maffione struct nmreq_opt_sync_kloop_eventfds *eventfds_opt = NULL; 448b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 449b6e66be2SVincenzo Maffione struct sync_kloop_poll_ctx *poll_ctx = NULL; 450b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 451b6e66be2SVincenzo Maffione int num_rx_rings, num_tx_rings, num_rings; 452*f79ba6d7SVincenzo Maffione struct sync_kloop_ring_args *args = NULL; 453b6e66be2SVincenzo Maffione uint32_t sleep_us = req->sleep_us; 454b6e66be2SVincenzo Maffione struct nm_csb_atok* csb_atok_base; 455b6e66be2SVincenzo Maffione struct nm_csb_ktoa* csb_ktoa_base; 456b6e66be2SVincenzo Maffione struct netmap_adapter *na; 457b6e66be2SVincenzo Maffione struct nmreq_option *opt; 458b6e66be2SVincenzo Maffione int err = 0; 459b6e66be2SVincenzo Maffione int i; 460b6e66be2SVincenzo Maffione 461b6e66be2SVincenzo Maffione if (sleep_us > 1000000) { 462b6e66be2SVincenzo Maffione /* We do not accept sleeping for more than a second. */ 463b6e66be2SVincenzo Maffione return EINVAL; 464b6e66be2SVincenzo Maffione } 465b6e66be2SVincenzo Maffione 466b6e66be2SVincenzo Maffione if (priv->np_nifp == NULL) { 467b6e66be2SVincenzo Maffione return ENXIO; 468b6e66be2SVincenzo Maffione } 469b6e66be2SVincenzo Maffione mb(); /* make sure following reads are not from cache */ 470b6e66be2SVincenzo Maffione 471b6e66be2SVincenzo Maffione na = priv->np_na; 472b6e66be2SVincenzo Maffione if (!nm_netmap_on(na)) { 473b6e66be2SVincenzo Maffione return ENXIO; 474b6e66be2SVincenzo Maffione } 475b6e66be2SVincenzo Maffione 476b6e66be2SVincenzo Maffione NMG_LOCK(); 477b6e66be2SVincenzo Maffione /* Make sure the application is working in CSB mode. */ 478b6e66be2SVincenzo Maffione if (!priv->np_csb_atok_base || !priv->np_csb_ktoa_base) { 479b6e66be2SVincenzo Maffione NMG_UNLOCK(); 480b6e66be2SVincenzo Maffione nm_prerr("sync-kloop on %s requires " 481b6e66be2SVincenzo Maffione "NETMAP_REQ_OPT_CSB option", na->name); 482b6e66be2SVincenzo Maffione return EINVAL; 483b6e66be2SVincenzo Maffione } 484b6e66be2SVincenzo Maffione 485b6e66be2SVincenzo Maffione csb_atok_base = priv->np_csb_atok_base; 486b6e66be2SVincenzo Maffione csb_ktoa_base = priv->np_csb_ktoa_base; 487b6e66be2SVincenzo Maffione 488b6e66be2SVincenzo Maffione /* Make sure that no kloop is currently running. */ 489b6e66be2SVincenzo Maffione if (priv->np_kloop_state & NM_SYNC_KLOOP_RUNNING) { 490b6e66be2SVincenzo Maffione err = EBUSY; 491b6e66be2SVincenzo Maffione } 492b6e66be2SVincenzo Maffione priv->np_kloop_state |= NM_SYNC_KLOOP_RUNNING; 493b6e66be2SVincenzo Maffione NMG_UNLOCK(); 494b6e66be2SVincenzo Maffione if (err) { 495b6e66be2SVincenzo Maffione return err; 496b6e66be2SVincenzo Maffione } 497b6e66be2SVincenzo Maffione 498b6e66be2SVincenzo Maffione num_rx_rings = priv->np_qlast[NR_RX] - priv->np_qfirst[NR_RX]; 499b6e66be2SVincenzo Maffione num_tx_rings = priv->np_qlast[NR_TX] - priv->np_qfirst[NR_TX]; 500b6e66be2SVincenzo Maffione num_rings = num_tx_rings + num_rx_rings; 501b6e66be2SVincenzo Maffione 502*f79ba6d7SVincenzo Maffione args = nm_os_malloc(num_rings * sizeof(args[0])); 503*f79ba6d7SVincenzo Maffione if (!args) { 504*f79ba6d7SVincenzo Maffione err = ENOMEM; 505*f79ba6d7SVincenzo Maffione goto out; 506*f79ba6d7SVincenzo Maffione } 507*f79ba6d7SVincenzo Maffione 508b6e66be2SVincenzo Maffione /* Validate notification options. */ 509b6e66be2SVincenzo Maffione opt = nmreq_findoption((struct nmreq_option *)(uintptr_t)hdr->nr_options, 510b6e66be2SVincenzo Maffione NETMAP_REQ_OPT_SYNC_KLOOP_EVENTFDS); 511b6e66be2SVincenzo Maffione if (opt != NULL) { 512b6e66be2SVincenzo Maffione err = nmreq_checkduplicate(opt); 513b6e66be2SVincenzo Maffione if (err) { 514b6e66be2SVincenzo Maffione opt->nro_status = err; 515b6e66be2SVincenzo Maffione goto out; 516b6e66be2SVincenzo Maffione } 517b6e66be2SVincenzo Maffione if (opt->nro_size != sizeof(*eventfds_opt) + 518b6e66be2SVincenzo Maffione sizeof(eventfds_opt->eventfds[0]) * num_rings) { 519b6e66be2SVincenzo Maffione /* Option size not consistent with the number of 520b6e66be2SVincenzo Maffione * entries. */ 521b6e66be2SVincenzo Maffione opt->nro_status = err = EINVAL; 522b6e66be2SVincenzo Maffione goto out; 523b6e66be2SVincenzo Maffione } 524b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 525b6e66be2SVincenzo Maffione eventfds_opt = (struct nmreq_opt_sync_kloop_eventfds *)opt; 526b6e66be2SVincenzo Maffione opt->nro_status = 0; 527b6e66be2SVincenzo Maffione /* We need 2 poll entries for TX and RX notifications coming 528b6e66be2SVincenzo Maffione * from the netmap adapter, plus one entries per ring for the 529b6e66be2SVincenzo Maffione * notifications coming from the application. */ 530b6e66be2SVincenzo Maffione poll_ctx = nm_os_malloc(sizeof(*poll_ctx) + 531b6e66be2SVincenzo Maffione (2 + num_rings) * sizeof(poll_ctx->entries[0])); 532b6e66be2SVincenzo Maffione init_poll_funcptr(&poll_ctx->wait_table, 533b6e66be2SVincenzo Maffione sync_kloop_poll_table_queue_proc); 534b6e66be2SVincenzo Maffione poll_ctx->num_entries = 2 + num_rings; 535b6e66be2SVincenzo Maffione poll_ctx->next_entry = 0; 536b6e66be2SVincenzo Maffione /* Poll for notifications coming from the applications through 537b6e66be2SVincenzo Maffione * eventfds . */ 538b6e66be2SVincenzo Maffione for (i = 0; i < num_rings; i++) { 539b6e66be2SVincenzo Maffione struct eventfd_ctx *irq; 540b6e66be2SVincenzo Maffione struct file *filp; 541b6e66be2SVincenzo Maffione unsigned long mask; 542b6e66be2SVincenzo Maffione 543b6e66be2SVincenzo Maffione filp = eventfd_fget(eventfds_opt->eventfds[i].ioeventfd); 544b6e66be2SVincenzo Maffione if (IS_ERR(filp)) { 545b6e66be2SVincenzo Maffione err = PTR_ERR(filp); 546b6e66be2SVincenzo Maffione goto out; 547b6e66be2SVincenzo Maffione } 548b6e66be2SVincenzo Maffione mask = filp->f_op->poll(filp, &poll_ctx->wait_table); 549b6e66be2SVincenzo Maffione if (mask & POLLERR) { 550b6e66be2SVincenzo Maffione err = EINVAL; 551b6e66be2SVincenzo Maffione goto out; 552b6e66be2SVincenzo Maffione } 553b6e66be2SVincenzo Maffione 554b6e66be2SVincenzo Maffione filp = eventfd_fget(eventfds_opt->eventfds[i].irqfd); 555b6e66be2SVincenzo Maffione if (IS_ERR(filp)) { 556b6e66be2SVincenzo Maffione err = PTR_ERR(filp); 557b6e66be2SVincenzo Maffione goto out; 558b6e66be2SVincenzo Maffione } 559b6e66be2SVincenzo Maffione poll_ctx->entries[i].irq_filp = filp; 560b6e66be2SVincenzo Maffione irq = eventfd_ctx_fileget(filp); 561b6e66be2SVincenzo Maffione if (IS_ERR(irq)) { 562b6e66be2SVincenzo Maffione err = PTR_ERR(irq); 563b6e66be2SVincenzo Maffione goto out; 564b6e66be2SVincenzo Maffione } 565b6e66be2SVincenzo Maffione poll_ctx->entries[i].irq_ctx = irq; 566b6e66be2SVincenzo Maffione } 567b6e66be2SVincenzo Maffione /* Poll for notifications coming from the netmap rings bound to 568b6e66be2SVincenzo Maffione * this file descriptor. */ 569b6e66be2SVincenzo Maffione { 570b6e66be2SVincenzo Maffione NM_SELINFO_T *si[NR_TXRX]; 571b6e66be2SVincenzo Maffione 572b6e66be2SVincenzo Maffione NMG_LOCK(); 573b6e66be2SVincenzo Maffione si[NR_RX] = nm_si_user(priv, NR_RX) ? &na->si[NR_RX] : 574b6e66be2SVincenzo Maffione &na->rx_rings[priv->np_qfirst[NR_RX]]->si; 575b6e66be2SVincenzo Maffione si[NR_TX] = nm_si_user(priv, NR_TX) ? &na->si[NR_TX] : 576b6e66be2SVincenzo Maffione &na->tx_rings[priv->np_qfirst[NR_TX]]->si; 577b6e66be2SVincenzo Maffione NMG_UNLOCK(); 578b6e66be2SVincenzo Maffione poll_wait(priv->np_filp, si[NR_TX], &poll_ctx->wait_table); 579*f79ba6d7SVincenzo Maffione poll_wait(priv->np_filp, si[NR_RX], &poll_ctx->wait_table); 580b6e66be2SVincenzo Maffione } 581b6e66be2SVincenzo Maffione #else /* SYNC_KLOOP_POLL */ 582b6e66be2SVincenzo Maffione opt->nro_status = EOPNOTSUPP; 583b6e66be2SVincenzo Maffione goto out; 584b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 585b6e66be2SVincenzo Maffione } 586b6e66be2SVincenzo Maffione 587*f79ba6d7SVincenzo Maffione /* Prepare the arguments for netmap_sync_kloop_tx_ring() 588*f79ba6d7SVincenzo Maffione * and netmap_sync_kloop_rx_ring(). */ 589*f79ba6d7SVincenzo Maffione for (i = 0; i < num_tx_rings; i++) { 590*f79ba6d7SVincenzo Maffione struct sync_kloop_ring_args *a = args + i; 591*f79ba6d7SVincenzo Maffione 592*f79ba6d7SVincenzo Maffione a->kring = NMR(na, NR_TX)[i + priv->np_qfirst[NR_TX]]; 593*f79ba6d7SVincenzo Maffione a->csb_atok = csb_atok_base + i; 594*f79ba6d7SVincenzo Maffione a->csb_ktoa = csb_ktoa_base + i; 595*f79ba6d7SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 596*f79ba6d7SVincenzo Maffione if (poll_ctx) 597*f79ba6d7SVincenzo Maffione a->irq_ctx = poll_ctx->entries[i].irq_ctx; 598*f79ba6d7SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 599*f79ba6d7SVincenzo Maffione } 600*f79ba6d7SVincenzo Maffione for (i = 0; i < num_rx_rings; i++) { 601*f79ba6d7SVincenzo Maffione struct sync_kloop_ring_args *a = args + num_tx_rings + i; 602*f79ba6d7SVincenzo Maffione 603*f79ba6d7SVincenzo Maffione a->kring = NMR(na, NR_RX)[i + priv->np_qfirst[NR_RX]]; 604*f79ba6d7SVincenzo Maffione a->csb_atok = csb_atok_base + num_tx_rings + i; 605*f79ba6d7SVincenzo Maffione a->csb_ktoa = csb_ktoa_base + num_tx_rings + i; 606*f79ba6d7SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 607*f79ba6d7SVincenzo Maffione if (poll_ctx) 608*f79ba6d7SVincenzo Maffione a->irq_ctx = poll_ctx->entries[num_tx_rings + i].irq_ctx; 609*f79ba6d7SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 610*f79ba6d7SVincenzo Maffione } 611*f79ba6d7SVincenzo Maffione 612b6e66be2SVincenzo Maffione /* Main loop. */ 613b6e66be2SVincenzo Maffione for (;;) { 614b6e66be2SVincenzo Maffione if (unlikely(NM_ACCESS_ONCE(priv->np_kloop_state) & NM_SYNC_KLOOP_STOPPING)) { 615b6e66be2SVincenzo Maffione break; 616b6e66be2SVincenzo Maffione } 617b6e66be2SVincenzo Maffione 618b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 619*f79ba6d7SVincenzo Maffione if (poll_ctx) { 620*f79ba6d7SVincenzo Maffione /* It is important to set the task state as 621*f79ba6d7SVincenzo Maffione * interruptible before processing any TX/RX ring, 622*f79ba6d7SVincenzo Maffione * so that if a notification on ring Y comes after 623*f79ba6d7SVincenzo Maffione * we have processed ring Y, but before we call 624*f79ba6d7SVincenzo Maffione * schedule(), we don't miss it. This is true because 625*f79ba6d7SVincenzo Maffione * the wake up function will change the the task state, 626*f79ba6d7SVincenzo Maffione * and therefore the schedule_timeout() call below 627*f79ba6d7SVincenzo Maffione * will observe the change). 628*f79ba6d7SVincenzo Maffione */ 629*f79ba6d7SVincenzo Maffione set_current_state(TASK_INTERRUPTIBLE); 630*f79ba6d7SVincenzo Maffione } 631b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 632b6e66be2SVincenzo Maffione 633b6e66be2SVincenzo Maffione /* Process all the TX rings bound to this file descriptor. */ 634b6e66be2SVincenzo Maffione for (i = 0; i < num_tx_rings; i++) { 635*f79ba6d7SVincenzo Maffione struct sync_kloop_ring_args *a = args + i; 636b6e66be2SVincenzo Maffione 637*f79ba6d7SVincenzo Maffione if (unlikely(nm_kr_tryget(a->kring, 1, NULL))) { 638b6e66be2SVincenzo Maffione continue; 639b6e66be2SVincenzo Maffione } 640*f79ba6d7SVincenzo Maffione netmap_sync_kloop_tx_ring(a); 641*f79ba6d7SVincenzo Maffione nm_kr_put(a->kring); 642b6e66be2SVincenzo Maffione } 643b6e66be2SVincenzo Maffione 644b6e66be2SVincenzo Maffione /* Process all the RX rings bound to this file descriptor. */ 645b6e66be2SVincenzo Maffione for (i = 0; i < num_rx_rings; i++) { 646*f79ba6d7SVincenzo Maffione struct sync_kloop_ring_args *a = args + num_tx_rings + i; 647b6e66be2SVincenzo Maffione 648*f79ba6d7SVincenzo Maffione if (unlikely(nm_kr_tryget(a->kring, 1, NULL))) { 649b6e66be2SVincenzo Maffione continue; 650b6e66be2SVincenzo Maffione } 651*f79ba6d7SVincenzo Maffione netmap_sync_kloop_rx_ring(a); 652*f79ba6d7SVincenzo Maffione nm_kr_put(a->kring); 653b6e66be2SVincenzo Maffione } 654b6e66be2SVincenzo Maffione 655b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 656b6e66be2SVincenzo Maffione if (poll_ctx) { 657b6e66be2SVincenzo Maffione /* If a poll context is present, yield to the scheduler 658b6e66be2SVincenzo Maffione * waiting for a notification to come either from 659b6e66be2SVincenzo Maffione * netmap or the application. */ 660*f79ba6d7SVincenzo Maffione schedule_timeout(msecs_to_jiffies(20000)); 661b6e66be2SVincenzo Maffione } else 662b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 663b6e66be2SVincenzo Maffione { 664b6e66be2SVincenzo Maffione /* Default synchronization method: sleep for a while. */ 665b6e66be2SVincenzo Maffione usleep_range(sleep_us, sleep_us); 666b6e66be2SVincenzo Maffione } 667b6e66be2SVincenzo Maffione } 668b6e66be2SVincenzo Maffione out: 669b6e66be2SVincenzo Maffione #ifdef SYNC_KLOOP_POLL 670b6e66be2SVincenzo Maffione if (poll_ctx) { 671b6e66be2SVincenzo Maffione /* Stop polling from netmap and the eventfds, and deallocate 672b6e66be2SVincenzo Maffione * the poll context. */ 673b6e66be2SVincenzo Maffione __set_current_state(TASK_RUNNING); 674b6e66be2SVincenzo Maffione for (i = 0; i < poll_ctx->next_entry; i++) { 675b6e66be2SVincenzo Maffione struct sync_kloop_poll_entry *entry = 676b6e66be2SVincenzo Maffione poll_ctx->entries + i; 677b6e66be2SVincenzo Maffione 678b6e66be2SVincenzo Maffione if (entry->wqh) 679b6e66be2SVincenzo Maffione remove_wait_queue(entry->wqh, &entry->wait); 680b6e66be2SVincenzo Maffione /* We did not get a reference to the eventfds, but 681b6e66be2SVincenzo Maffione * don't do that on netmap file descriptors (since 682b6e66be2SVincenzo Maffione * a reference was not taken. */ 683b6e66be2SVincenzo Maffione if (entry->filp && entry->filp != priv->np_filp) 684b6e66be2SVincenzo Maffione fput(entry->filp); 685b6e66be2SVincenzo Maffione if (entry->irq_ctx) 686b6e66be2SVincenzo Maffione eventfd_ctx_put(entry->irq_ctx); 687b6e66be2SVincenzo Maffione if (entry->irq_filp) 688b6e66be2SVincenzo Maffione fput(entry->irq_filp); 689b6e66be2SVincenzo Maffione } 690b6e66be2SVincenzo Maffione nm_os_free(poll_ctx); 691b6e66be2SVincenzo Maffione poll_ctx = NULL; 692b6e66be2SVincenzo Maffione } 693b6e66be2SVincenzo Maffione #endif /* SYNC_KLOOP_POLL */ 694b6e66be2SVincenzo Maffione 695*f79ba6d7SVincenzo Maffione if (args) { 696*f79ba6d7SVincenzo Maffione nm_os_free(args); 697*f79ba6d7SVincenzo Maffione args = NULL; 698*f79ba6d7SVincenzo Maffione } 699*f79ba6d7SVincenzo Maffione 700b6e66be2SVincenzo Maffione /* Reset the kloop state. */ 701b6e66be2SVincenzo Maffione NMG_LOCK(); 702b6e66be2SVincenzo Maffione priv->np_kloop_state = 0; 703b6e66be2SVincenzo Maffione NMG_UNLOCK(); 704b6e66be2SVincenzo Maffione 705b6e66be2SVincenzo Maffione return err; 706b6e66be2SVincenzo Maffione } 707b6e66be2SVincenzo Maffione 708b6e66be2SVincenzo Maffione int 709b6e66be2SVincenzo Maffione netmap_sync_kloop_stop(struct netmap_priv_d *priv) 710b6e66be2SVincenzo Maffione { 711b6e66be2SVincenzo Maffione bool running = true; 712b6e66be2SVincenzo Maffione int err = 0; 713b6e66be2SVincenzo Maffione 714b6e66be2SVincenzo Maffione NMG_LOCK(); 715b6e66be2SVincenzo Maffione priv->np_kloop_state |= NM_SYNC_KLOOP_STOPPING; 716b6e66be2SVincenzo Maffione NMG_UNLOCK(); 717b6e66be2SVincenzo Maffione while (running) { 718b6e66be2SVincenzo Maffione usleep_range(1000, 1500); 719b6e66be2SVincenzo Maffione NMG_LOCK(); 720b6e66be2SVincenzo Maffione running = (NM_ACCESS_ONCE(priv->np_kloop_state) 721b6e66be2SVincenzo Maffione & NM_SYNC_KLOOP_RUNNING); 722b6e66be2SVincenzo Maffione NMG_UNLOCK(); 723b6e66be2SVincenzo Maffione } 724b6e66be2SVincenzo Maffione 725b6e66be2SVincenzo Maffione return err; 726b6e66be2SVincenzo Maffione } 727b6e66be2SVincenzo Maffione 728b6e66be2SVincenzo Maffione #ifdef WITH_PTNETMAP 729b6e66be2SVincenzo Maffione /* 730b6e66be2SVincenzo Maffione * Guest ptnetmap txsync()/rxsync() routines, used in ptnet device drivers. 731b6e66be2SVincenzo Maffione * These routines are reused across the different operating systems supported 732b6e66be2SVincenzo Maffione * by netmap. 733b6e66be2SVincenzo Maffione */ 734b6e66be2SVincenzo Maffione 735b6e66be2SVincenzo Maffione /* 736b6e66be2SVincenzo Maffione * Reconcile host and guest views of the transmit ring. 737b6e66be2SVincenzo Maffione * 738b6e66be2SVincenzo Maffione * Guest user wants to transmit packets up to the one before ring->head, 739b6e66be2SVincenzo Maffione * and guest kernel knows tx_ring->hwcur is the first packet unsent 740b6e66be2SVincenzo Maffione * by the host kernel. 741b6e66be2SVincenzo Maffione * 742b6e66be2SVincenzo Maffione * We push out as many packets as possible, and possibly 743b6e66be2SVincenzo Maffione * reclaim buffers from previously completed transmission. 744b6e66be2SVincenzo Maffione * 745b6e66be2SVincenzo Maffione * Notifications from the host are enabled only if the user guest would 746b6e66be2SVincenzo Maffione * block (no space in the ring). 747b6e66be2SVincenzo Maffione */ 748b6e66be2SVincenzo Maffione bool 749b6e66be2SVincenzo Maffione netmap_pt_guest_txsync(struct nm_csb_atok *atok, struct nm_csb_ktoa *ktoa, 750b6e66be2SVincenzo Maffione struct netmap_kring *kring, int flags) 751b6e66be2SVincenzo Maffione { 752b6e66be2SVincenzo Maffione bool notify = false; 753b6e66be2SVincenzo Maffione 754b6e66be2SVincenzo Maffione /* Disable notifications */ 755b6e66be2SVincenzo Maffione atok->appl_need_kick = 0; 756b6e66be2SVincenzo Maffione 757b6e66be2SVincenzo Maffione /* 758b6e66be2SVincenzo Maffione * First part: tell the host (updating the CSB) to process the new 759b6e66be2SVincenzo Maffione * packets. 760b6e66be2SVincenzo Maffione */ 761b6e66be2SVincenzo Maffione kring->nr_hwcur = ktoa->hwcur; 762*f79ba6d7SVincenzo Maffione nm_sync_kloop_appl_write(atok, kring->rcur, kring->rhead); 763b6e66be2SVincenzo Maffione 764b6e66be2SVincenzo Maffione /* Ask for a kick from a guest to the host if needed. */ 765b6e66be2SVincenzo Maffione if (((kring->rhead != kring->nr_hwcur || nm_kr_txempty(kring)) 766b6e66be2SVincenzo Maffione && NM_ACCESS_ONCE(ktoa->kern_need_kick)) || 767b6e66be2SVincenzo Maffione (flags & NAF_FORCE_RECLAIM)) { 768b6e66be2SVincenzo Maffione atok->sync_flags = flags; 769b6e66be2SVincenzo Maffione notify = true; 770b6e66be2SVincenzo Maffione } 771b6e66be2SVincenzo Maffione 772b6e66be2SVincenzo Maffione /* 773b6e66be2SVincenzo Maffione * Second part: reclaim buffers for completed transmissions. 774b6e66be2SVincenzo Maffione */ 775b6e66be2SVincenzo Maffione if (nm_kr_txempty(kring) || (flags & NAF_FORCE_RECLAIM)) { 776*f79ba6d7SVincenzo Maffione nm_sync_kloop_appl_read(ktoa, &kring->nr_hwtail, 777*f79ba6d7SVincenzo Maffione &kring->nr_hwcur); 778b6e66be2SVincenzo Maffione } 779b6e66be2SVincenzo Maffione 780b6e66be2SVincenzo Maffione /* 781b6e66be2SVincenzo Maffione * No more room in the ring for new transmissions. The user thread will 782b6e66be2SVincenzo Maffione * go to sleep and we need to be notified by the host when more free 783b6e66be2SVincenzo Maffione * space is available. 784b6e66be2SVincenzo Maffione */ 785b6e66be2SVincenzo Maffione if (nm_kr_txempty(kring) && !(kring->nr_kflags & NKR_NOINTR)) { 786b6e66be2SVincenzo Maffione /* Reenable notifications. */ 787b6e66be2SVincenzo Maffione atok->appl_need_kick = 1; 788*f79ba6d7SVincenzo Maffione /* Double check, with store-load memory barrier. */ 789*f79ba6d7SVincenzo Maffione nm_stld_barrier(); 790*f79ba6d7SVincenzo Maffione nm_sync_kloop_appl_read(ktoa, &kring->nr_hwtail, 791*f79ba6d7SVincenzo Maffione &kring->nr_hwcur); 792b6e66be2SVincenzo Maffione /* If there is new free space, disable notifications */ 793b6e66be2SVincenzo Maffione if (unlikely(!nm_kr_txempty(kring))) { 794b6e66be2SVincenzo Maffione atok->appl_need_kick = 0; 795b6e66be2SVincenzo Maffione } 796b6e66be2SVincenzo Maffione } 797b6e66be2SVincenzo Maffione 798b6e66be2SVincenzo Maffione nm_prdis(1, "%s CSB(head:%u cur:%u hwtail:%u) KRING(head:%u cur:%u tail:%u)", 799b6e66be2SVincenzo Maffione kring->name, atok->head, atok->cur, ktoa->hwtail, 800b6e66be2SVincenzo Maffione kring->rhead, kring->rcur, kring->nr_hwtail); 801b6e66be2SVincenzo Maffione 802b6e66be2SVincenzo Maffione return notify; 803b6e66be2SVincenzo Maffione } 804b6e66be2SVincenzo Maffione 805b6e66be2SVincenzo Maffione /* 806b6e66be2SVincenzo Maffione * Reconcile host and guest view of the receive ring. 807b6e66be2SVincenzo Maffione * 808b6e66be2SVincenzo Maffione * Update hwcur/hwtail from host (reading from CSB). 809b6e66be2SVincenzo Maffione * 810b6e66be2SVincenzo Maffione * If guest user has released buffers up to the one before ring->head, we 811b6e66be2SVincenzo Maffione * also give them to the host. 812b6e66be2SVincenzo Maffione * 813b6e66be2SVincenzo Maffione * Notifications from the host are enabled only if the user guest would 814b6e66be2SVincenzo Maffione * block (no more completed slots in the ring). 815b6e66be2SVincenzo Maffione */ 816b6e66be2SVincenzo Maffione bool 817b6e66be2SVincenzo Maffione netmap_pt_guest_rxsync(struct nm_csb_atok *atok, struct nm_csb_ktoa *ktoa, 818b6e66be2SVincenzo Maffione struct netmap_kring *kring, int flags) 819b6e66be2SVincenzo Maffione { 820b6e66be2SVincenzo Maffione bool notify = false; 821b6e66be2SVincenzo Maffione 822b6e66be2SVincenzo Maffione /* Disable notifications */ 823b6e66be2SVincenzo Maffione atok->appl_need_kick = 0; 824b6e66be2SVincenzo Maffione 825b6e66be2SVincenzo Maffione /* 826b6e66be2SVincenzo Maffione * First part: import newly received packets, by updating the kring 827b6e66be2SVincenzo Maffione * hwtail to the hwtail known from the host (read from the CSB). 828b6e66be2SVincenzo Maffione * This also updates the kring hwcur. 829b6e66be2SVincenzo Maffione */ 830*f79ba6d7SVincenzo Maffione nm_sync_kloop_appl_read(ktoa, &kring->nr_hwtail, &kring->nr_hwcur); 831b6e66be2SVincenzo Maffione kring->nr_kflags &= ~NKR_PENDINTR; 832b6e66be2SVincenzo Maffione 833b6e66be2SVincenzo Maffione /* 834b6e66be2SVincenzo Maffione * Second part: tell the host about the slots that guest user has 835b6e66be2SVincenzo Maffione * released, by updating cur and head in the CSB. 836b6e66be2SVincenzo Maffione */ 837b6e66be2SVincenzo Maffione if (kring->rhead != kring->nr_hwcur) { 838*f79ba6d7SVincenzo Maffione nm_sync_kloop_appl_write(atok, kring->rcur, kring->rhead); 839b6e66be2SVincenzo Maffione /* Ask for a kick from the guest to the host if needed. */ 840b6e66be2SVincenzo Maffione if (NM_ACCESS_ONCE(ktoa->kern_need_kick)) { 841b6e66be2SVincenzo Maffione atok->sync_flags = flags; 842b6e66be2SVincenzo Maffione notify = true; 843b6e66be2SVincenzo Maffione } 844b6e66be2SVincenzo Maffione } 845b6e66be2SVincenzo Maffione 846b6e66be2SVincenzo Maffione /* 847b6e66be2SVincenzo Maffione * No more completed RX slots. The user thread will go to sleep and 848b6e66be2SVincenzo Maffione * we need to be notified by the host when more RX slots have been 849b6e66be2SVincenzo Maffione * completed. 850b6e66be2SVincenzo Maffione */ 851b6e66be2SVincenzo Maffione if (nm_kr_rxempty(kring) && !(kring->nr_kflags & NKR_NOINTR)) { 852b6e66be2SVincenzo Maffione /* Reenable notifications. */ 853b6e66be2SVincenzo Maffione atok->appl_need_kick = 1; 854*f79ba6d7SVincenzo Maffione /* Double check, with store-load memory barrier. */ 855*f79ba6d7SVincenzo Maffione nm_stld_barrier(); 856*f79ba6d7SVincenzo Maffione nm_sync_kloop_appl_read(ktoa, &kring->nr_hwtail, 857*f79ba6d7SVincenzo Maffione &kring->nr_hwcur); 858b6e66be2SVincenzo Maffione /* If there are new slots, disable notifications. */ 859b6e66be2SVincenzo Maffione if (!nm_kr_rxempty(kring)) { 860b6e66be2SVincenzo Maffione atok->appl_need_kick = 0; 861b6e66be2SVincenzo Maffione } 862b6e66be2SVincenzo Maffione } 863b6e66be2SVincenzo Maffione 864b6e66be2SVincenzo Maffione nm_prdis(1, "%s CSB(head:%u cur:%u hwtail:%u) KRING(head:%u cur:%u tail:%u)", 865b6e66be2SVincenzo Maffione kring->name, atok->head, atok->cur, ktoa->hwtail, 866b6e66be2SVincenzo Maffione kring->rhead, kring->rcur, kring->nr_hwtail); 867b6e66be2SVincenzo Maffione 868b6e66be2SVincenzo Maffione return notify; 869b6e66be2SVincenzo Maffione } 870b6e66be2SVincenzo Maffione 871b6e66be2SVincenzo Maffione /* 872b6e66be2SVincenzo Maffione * Callbacks for ptnet drivers: nm_krings_create, nm_krings_delete, nm_dtor. 873b6e66be2SVincenzo Maffione */ 874b6e66be2SVincenzo Maffione int 875b6e66be2SVincenzo Maffione ptnet_nm_krings_create(struct netmap_adapter *na) 876b6e66be2SVincenzo Maffione { 877b6e66be2SVincenzo Maffione struct netmap_pt_guest_adapter *ptna = 878b6e66be2SVincenzo Maffione (struct netmap_pt_guest_adapter *)na; /* Upcast. */ 879b6e66be2SVincenzo Maffione struct netmap_adapter *na_nm = &ptna->hwup.up; 880b6e66be2SVincenzo Maffione struct netmap_adapter *na_dr = &ptna->dr.up; 881b6e66be2SVincenzo Maffione int ret; 882b6e66be2SVincenzo Maffione 883b6e66be2SVincenzo Maffione if (ptna->backend_users) { 884b6e66be2SVincenzo Maffione return 0; 885b6e66be2SVincenzo Maffione } 886b6e66be2SVincenzo Maffione 887b6e66be2SVincenzo Maffione /* Create krings on the public netmap adapter. */ 888b6e66be2SVincenzo Maffione ret = netmap_hw_krings_create(na_nm); 889b6e66be2SVincenzo Maffione if (ret) { 890b6e66be2SVincenzo Maffione return ret; 891b6e66be2SVincenzo Maffione } 892b6e66be2SVincenzo Maffione 893b6e66be2SVincenzo Maffione /* Copy krings into the netmap adapter private to the driver. */ 894b6e66be2SVincenzo Maffione na_dr->tx_rings = na_nm->tx_rings; 895b6e66be2SVincenzo Maffione na_dr->rx_rings = na_nm->rx_rings; 896b6e66be2SVincenzo Maffione 897b6e66be2SVincenzo Maffione return 0; 898b6e66be2SVincenzo Maffione } 899b6e66be2SVincenzo Maffione 900b6e66be2SVincenzo Maffione void 901b6e66be2SVincenzo Maffione ptnet_nm_krings_delete(struct netmap_adapter *na) 902b6e66be2SVincenzo Maffione { 903b6e66be2SVincenzo Maffione struct netmap_pt_guest_adapter *ptna = 904b6e66be2SVincenzo Maffione (struct netmap_pt_guest_adapter *)na; /* Upcast. */ 905b6e66be2SVincenzo Maffione struct netmap_adapter *na_nm = &ptna->hwup.up; 906b6e66be2SVincenzo Maffione struct netmap_adapter *na_dr = &ptna->dr.up; 907b6e66be2SVincenzo Maffione 908b6e66be2SVincenzo Maffione if (ptna->backend_users) { 909b6e66be2SVincenzo Maffione return; 910b6e66be2SVincenzo Maffione } 911b6e66be2SVincenzo Maffione 912b6e66be2SVincenzo Maffione na_dr->tx_rings = NULL; 913b6e66be2SVincenzo Maffione na_dr->rx_rings = NULL; 914b6e66be2SVincenzo Maffione 915b6e66be2SVincenzo Maffione netmap_hw_krings_delete(na_nm); 916b6e66be2SVincenzo Maffione } 917b6e66be2SVincenzo Maffione 918b6e66be2SVincenzo Maffione void 919b6e66be2SVincenzo Maffione ptnet_nm_dtor(struct netmap_adapter *na) 920b6e66be2SVincenzo Maffione { 921b6e66be2SVincenzo Maffione struct netmap_pt_guest_adapter *ptna = 922b6e66be2SVincenzo Maffione (struct netmap_pt_guest_adapter *)na; 923b6e66be2SVincenzo Maffione 924b6e66be2SVincenzo Maffione netmap_mem_put(ptna->dr.up.nm_mem); 925b6e66be2SVincenzo Maffione memset(&ptna->dr, 0, sizeof(ptna->dr)); 926b6e66be2SVincenzo Maffione netmap_mem_pt_guest_ifp_del(na->nm_mem, na->ifp); 927b6e66be2SVincenzo Maffione } 928b6e66be2SVincenzo Maffione 929b6e66be2SVincenzo Maffione int 930b6e66be2SVincenzo Maffione netmap_pt_guest_attach(struct netmap_adapter *arg, 931b6e66be2SVincenzo Maffione unsigned int nifp_offset, unsigned int memid) 932b6e66be2SVincenzo Maffione { 933b6e66be2SVincenzo Maffione struct netmap_pt_guest_adapter *ptna; 934b6e66be2SVincenzo Maffione struct ifnet *ifp = arg ? arg->ifp : NULL; 935b6e66be2SVincenzo Maffione int error; 936b6e66be2SVincenzo Maffione 937b6e66be2SVincenzo Maffione /* get allocator */ 938b6e66be2SVincenzo Maffione arg->nm_mem = netmap_mem_pt_guest_new(ifp, nifp_offset, memid); 939b6e66be2SVincenzo Maffione if (arg->nm_mem == NULL) 940b6e66be2SVincenzo Maffione return ENOMEM; 941b6e66be2SVincenzo Maffione arg->na_flags |= NAF_MEM_OWNER; 942b6e66be2SVincenzo Maffione error = netmap_attach_ext(arg, sizeof(struct netmap_pt_guest_adapter), 1); 943b6e66be2SVincenzo Maffione if (error) 944b6e66be2SVincenzo Maffione return error; 945b6e66be2SVincenzo Maffione 946b6e66be2SVincenzo Maffione /* get the netmap_pt_guest_adapter */ 947b6e66be2SVincenzo Maffione ptna = (struct netmap_pt_guest_adapter *) NA(ifp); 948b6e66be2SVincenzo Maffione 949b6e66be2SVincenzo Maffione /* Initialize a separate pass-through netmap adapter that is going to 950b6e66be2SVincenzo Maffione * be used by the ptnet driver only, and so never exposed to netmap 951b6e66be2SVincenzo Maffione * applications. We only need a subset of the available fields. */ 952b6e66be2SVincenzo Maffione memset(&ptna->dr, 0, sizeof(ptna->dr)); 953b6e66be2SVincenzo Maffione ptna->dr.up.ifp = ifp; 954b6e66be2SVincenzo Maffione ptna->dr.up.nm_mem = netmap_mem_get(ptna->hwup.up.nm_mem); 955b6e66be2SVincenzo Maffione ptna->dr.up.nm_config = ptna->hwup.up.nm_config; 956b6e66be2SVincenzo Maffione 957b6e66be2SVincenzo Maffione ptna->backend_users = 0; 958b6e66be2SVincenzo Maffione 959b6e66be2SVincenzo Maffione return 0; 960b6e66be2SVincenzo Maffione } 961b6e66be2SVincenzo Maffione 962b6e66be2SVincenzo Maffione #endif /* WITH_PTNETMAP */ 963