1 /* $NetBSD: cache_r4k.h,v 1.10 2003/03/08 04:43:26 rafal Exp $ */ 2 3 /*- 4 * SPDX-License-Identifier: BSD-4-Clause 5 * 6 * Copyright 2001 Wasabi Systems, Inc. 7 * All rights reserved. 8 * 9 * Written by Jason R. Thorpe for Wasabi Systems, Inc. 10 * 11 * Redistribution and use in source and binary forms, with or without 12 * modification, are permitted provided that the following conditions 13 * are met: 14 * 1. Redistributions of source code must retain the above copyright 15 * notice, this list of conditions and the following disclaimer. 16 * 2. Redistributions in binary form must reproduce the above copyright 17 * notice, this list of conditions and the following disclaimer in the 18 * documentation and/or other materials provided with the distribution. 19 * 3. All advertising materials mentioning features or use of this software 20 * must display the following acknowledgement: 21 * This product includes software developed for the NetBSD Project by 22 * Wasabi Systems, Inc. 23 * 4. The name of Wasabi Systems, Inc. may not be used to endorse 24 * or promote products derived from this software without specific prior 25 * written permission. 26 * 27 * THIS SOFTWARE IS PROVIDED BY WASABI SYSTEMS, INC. ``AS IS'' AND 28 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED 29 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 30 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL WASABI SYSTEMS, INC 31 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 32 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 33 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 34 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 35 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 36 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 37 * POSSIBILITY OF SUCH DAMAGE. 38 * 39 * $FreeBSD$ 40 */ 41 42 /* 43 * Cache definitions/operations for R4000-style caches. 44 */ 45 46 #define CACHE_R4K_I 0 47 #define CACHE_R4K_D 1 48 #define CACHE_R4K_SI 2 49 #define CACHE_R4K_SD 3 50 51 #define CACHEOP_R4K_INDEX_INV (0 << 2) /* I, SI */ 52 #define CACHEOP_R4K_INDEX_WB_INV (0 << 2) /* D, SD */ 53 #define CACHEOP_R4K_INDEX_LOAD_TAG (1 << 2) /* all */ 54 #define CACHEOP_R4K_INDEX_STORE_TAG (2 << 2) /* all */ 55 #define CACHEOP_R4K_CREATE_DIRTY_EXCL (3 << 2) /* D, SD */ 56 #define CACHEOP_R4K_HIT_INV (4 << 2) /* all */ 57 #define CACHEOP_R4K_HIT_WB_INV (5 << 2) /* D, SD */ 58 #define CACHEOP_R4K_FILL (5 << 2) /* I */ 59 #define CACHEOP_R4K_HIT_WB (6 << 2) /* I, D, SD */ 60 #define CACHEOP_R4K_HIT_SET_VIRTUAL (7 << 2) /* SI, SD */ 61 62 #if !defined(LOCORE) 63 64 /* 65 * cache_r4k_op_line: 66 * 67 * Perform the specified cache operation on a single line. 68 */ 69 #define cache_op_r4k_line(va, op) \ 70 do { \ 71 __asm __volatile( \ 72 ".set noreorder \n\t" \ 73 "cache %1, 0(%0) \n\t" \ 74 ".set reorder" \ 75 : \ 76 : "r" (va), "i" (op) \ 77 : "memory"); \ 78 } while (/*CONSTCOND*/0) 79 80 /* 81 * cache_r4k_op_8lines_16: 82 * 83 * Perform the specified cache operation on 8 16-byte cache lines. 84 */ 85 #define cache_r4k_op_8lines_16(va, op) \ 86 do { \ 87 __asm __volatile( \ 88 ".set noreorder \n\t" \ 89 "cache %1, 0x00(%0); cache %1, 0x10(%0) \n\t" \ 90 "cache %1, 0x20(%0); cache %1, 0x30(%0) \n\t" \ 91 "cache %1, 0x40(%0); cache %1, 0x50(%0) \n\t" \ 92 "cache %1, 0x60(%0); cache %1, 0x70(%0) \n\t" \ 93 ".set reorder" \ 94 : \ 95 : "r" (va), "i" (op) \ 96 : "memory"); \ 97 } while (/*CONSTCOND*/0) 98 99 /* 100 * cache_r4k_op_8lines_32: 101 * 102 * Perform the specified cache operation on 8 32-byte cache lines. 103 */ 104 #define cache_r4k_op_8lines_32(va, op) \ 105 do { \ 106 __asm __volatile( \ 107 ".set noreorder \n\t" \ 108 "cache %1, 0x00(%0); cache %1, 0x20(%0) \n\t" \ 109 "cache %1, 0x40(%0); cache %1, 0x60(%0) \n\t" \ 110 "cache %1, 0x80(%0); cache %1, 0xa0(%0) \n\t" \ 111 "cache %1, 0xc0(%0); cache %1, 0xe0(%0) \n\t" \ 112 ".set reorder" \ 113 : \ 114 : "r" (va), "i" (op) \ 115 : "memory"); \ 116 } while (/*CONSTCOND*/0) 117 118 /* 119 * cache_r4k_op_8lines_64: 120 * 121 * Perform the specified cache operation on 8 64-byte cache lines. 122 */ 123 #define cache_r4k_op_8lines_64(va, op) \ 124 do { \ 125 __asm __volatile( \ 126 ".set noreorder \n\t" \ 127 "cache %1, 0x000(%0); cache %1, 0x040(%0) \n\t" \ 128 "cache %1, 0x080(%0); cache %1, 0x0c0(%0) \n\t" \ 129 "cache %1, 0x100(%0); cache %1, 0x140(%0) \n\t" \ 130 "cache %1, 0x180(%0); cache %1, 0x1c0(%0) \n\t" \ 131 ".set reorder" \ 132 : \ 133 : "r" (va), "i" (op) \ 134 : "memory"); \ 135 } while (/*CONSTCOND*/0) 136 137 /* 138 * cache_r4k_op_32lines_16: 139 * 140 * Perform the specified cache operation on 32 16-byte 141 * cache lines. 142 */ 143 #define cache_r4k_op_32lines_16(va, op) \ 144 do { \ 145 __asm __volatile( \ 146 ".set noreorder \n\t" \ 147 "cache %1, 0x000(%0); cache %1, 0x010(%0); \n\t" \ 148 "cache %1, 0x020(%0); cache %1, 0x030(%0); \n\t" \ 149 "cache %1, 0x040(%0); cache %1, 0x050(%0); \n\t" \ 150 "cache %1, 0x060(%0); cache %1, 0x070(%0); \n\t" \ 151 "cache %1, 0x080(%0); cache %1, 0x090(%0); \n\t" \ 152 "cache %1, 0x0a0(%0); cache %1, 0x0b0(%0); \n\t" \ 153 "cache %1, 0x0c0(%0); cache %1, 0x0d0(%0); \n\t" \ 154 "cache %1, 0x0e0(%0); cache %1, 0x0f0(%0); \n\t" \ 155 "cache %1, 0x100(%0); cache %1, 0x110(%0); \n\t" \ 156 "cache %1, 0x120(%0); cache %1, 0x130(%0); \n\t" \ 157 "cache %1, 0x140(%0); cache %1, 0x150(%0); \n\t" \ 158 "cache %1, 0x160(%0); cache %1, 0x170(%0); \n\t" \ 159 "cache %1, 0x180(%0); cache %1, 0x190(%0); \n\t" \ 160 "cache %1, 0x1a0(%0); cache %1, 0x1b0(%0); \n\t" \ 161 "cache %1, 0x1c0(%0); cache %1, 0x1d0(%0); \n\t" \ 162 "cache %1, 0x1e0(%0); cache %1, 0x1f0(%0); \n\t" \ 163 ".set reorder" \ 164 : \ 165 : "r" (va), "i" (op) \ 166 : "memory"); \ 167 } while (/*CONSTCOND*/0) 168 169 /* 170 * cache_r4k_op_32lines_32: 171 * 172 * Perform the specified cache operation on 32 32-byte 173 * cache lines. 174 */ 175 #define cache_r4k_op_32lines_32(va, op) \ 176 do { \ 177 __asm __volatile( \ 178 ".set noreorder \n\t" \ 179 "cache %1, 0x000(%0); cache %1, 0x020(%0); \n\t" \ 180 "cache %1, 0x040(%0); cache %1, 0x060(%0); \n\t" \ 181 "cache %1, 0x080(%0); cache %1, 0x0a0(%0); \n\t" \ 182 "cache %1, 0x0c0(%0); cache %1, 0x0e0(%0); \n\t" \ 183 "cache %1, 0x100(%0); cache %1, 0x120(%0); \n\t" \ 184 "cache %1, 0x140(%0); cache %1, 0x160(%0); \n\t" \ 185 "cache %1, 0x180(%0); cache %1, 0x1a0(%0); \n\t" \ 186 "cache %1, 0x1c0(%0); cache %1, 0x1e0(%0); \n\t" \ 187 "cache %1, 0x200(%0); cache %1, 0x220(%0); \n\t" \ 188 "cache %1, 0x240(%0); cache %1, 0x260(%0); \n\t" \ 189 "cache %1, 0x280(%0); cache %1, 0x2a0(%0); \n\t" \ 190 "cache %1, 0x2c0(%0); cache %1, 0x2e0(%0); \n\t" \ 191 "cache %1, 0x300(%0); cache %1, 0x320(%0); \n\t" \ 192 "cache %1, 0x340(%0); cache %1, 0x360(%0); \n\t" \ 193 "cache %1, 0x380(%0); cache %1, 0x3a0(%0); \n\t" \ 194 "cache %1, 0x3c0(%0); cache %1, 0x3e0(%0); \n\t" \ 195 ".set reorder" \ 196 : \ 197 : "r" (va), "i" (op) \ 198 : "memory"); \ 199 } while (/*CONSTCOND*/0) 200 201 /* 202 * cache_r4k_op_32lines_64: 203 * 204 * Perform the specified cache operation on 32 64-byte 205 * cache lines. 206 */ 207 #define cache_r4k_op_32lines_64(va, op) \ 208 do { \ 209 __asm __volatile( \ 210 ".set noreorder \n\t" \ 211 "cache %1, 0x000(%0); cache %1, 0x040(%0); \n\t" \ 212 "cache %1, 0x080(%0); cache %1, 0x0c0(%0); \n\t" \ 213 "cache %1, 0x100(%0); cache %1, 0x140(%0); \n\t" \ 214 "cache %1, 0x180(%0); cache %1, 0x1c0(%0); \n\t" \ 215 "cache %1, 0x200(%0); cache %1, 0x240(%0); \n\t" \ 216 "cache %1, 0x280(%0); cache %1, 0x2c0(%0); \n\t" \ 217 "cache %1, 0x300(%0); cache %1, 0x340(%0); \n\t" \ 218 "cache %1, 0x380(%0); cache %1, 0x3c0(%0); \n\t" \ 219 "cache %1, 0x400(%0); cache %1, 0x440(%0); \n\t" \ 220 "cache %1, 0x480(%0); cache %1, 0x4c0(%0); \n\t" \ 221 "cache %1, 0x500(%0); cache %1, 0x540(%0); \n\t" \ 222 "cache %1, 0x580(%0); cache %1, 0x5c0(%0); \n\t" \ 223 "cache %1, 0x600(%0); cache %1, 0x640(%0); \n\t" \ 224 "cache %1, 0x680(%0); cache %1, 0x6c0(%0); \n\t" \ 225 "cache %1, 0x700(%0); cache %1, 0x740(%0); \n\t" \ 226 "cache %1, 0x780(%0); cache %1, 0x7c0(%0); \n\t" \ 227 ".set reorder" \ 228 : \ 229 : "r" (va), "i" (op) \ 230 : "memory"); \ 231 } while (/*CONSTCOND*/0) 232 233 /* 234 * cache_r4k_op_32lines_128: 235 * 236 * Perform the specified cache operation on 32 128-byte 237 * cache lines. 238 */ 239 #define cache_r4k_op_32lines_128(va, op) \ 240 do { \ 241 __asm __volatile( \ 242 ".set noreorder \n\t" \ 243 "cache %1, 0x0000(%0); cache %1, 0x0080(%0); \n\t" \ 244 "cache %1, 0x0100(%0); cache %1, 0x0180(%0); \n\t" \ 245 "cache %1, 0x0200(%0); cache %1, 0x0280(%0); \n\t" \ 246 "cache %1, 0x0300(%0); cache %1, 0x0380(%0); \n\t" \ 247 "cache %1, 0x0400(%0); cache %1, 0x0480(%0); \n\t" \ 248 "cache %1, 0x0500(%0); cache %1, 0x0580(%0); \n\t" \ 249 "cache %1, 0x0600(%0); cache %1, 0x0680(%0); \n\t" \ 250 "cache %1, 0x0700(%0); cache %1, 0x0780(%0); \n\t" \ 251 "cache %1, 0x0800(%0); cache %1, 0x0880(%0); \n\t" \ 252 "cache %1, 0x0900(%0); cache %1, 0x0980(%0); \n\t" \ 253 "cache %1, 0x0a00(%0); cache %1, 0x0a80(%0); \n\t" \ 254 "cache %1, 0x0b00(%0); cache %1, 0x0b80(%0); \n\t" \ 255 "cache %1, 0x0c00(%0); cache %1, 0x0c80(%0); \n\t" \ 256 "cache %1, 0x0d00(%0); cache %1, 0x0d80(%0); \n\t" \ 257 "cache %1, 0x0e00(%0); cache %1, 0x0e80(%0); \n\t" \ 258 "cache %1, 0x0f00(%0); cache %1, 0x0f80(%0); \n\t" \ 259 ".set reorder" \ 260 : \ 261 : "r" (va), "i" (op) \ 262 : "memory"); \ 263 } while (/*CONSTCOND*/0) 264 265 /* 266 * cache_r4k_op_16lines_16_2way: 267 * 268 * Perform the specified cache operation on 16 16-byte 269 * cache lines, 2-ways. 270 */ 271 #define cache_r4k_op_16lines_16_2way(va1, va2, op) \ 272 do { \ 273 __asm __volatile( \ 274 ".set noreorder \n\t" \ 275 "cache %2, 0x000(%0); cache %2, 0x000(%1); \n\t" \ 276 "cache %2, 0x010(%0); cache %2, 0x010(%1); \n\t" \ 277 "cache %2, 0x020(%0); cache %2, 0x020(%1); \n\t" \ 278 "cache %2, 0x030(%0); cache %2, 0x030(%1); \n\t" \ 279 "cache %2, 0x040(%0); cache %2, 0x040(%1); \n\t" \ 280 "cache %2, 0x050(%0); cache %2, 0x050(%1); \n\t" \ 281 "cache %2, 0x060(%0); cache %2, 0x060(%1); \n\t" \ 282 "cache %2, 0x070(%0); cache %2, 0x070(%1); \n\t" \ 283 "cache %2, 0x080(%0); cache %2, 0x080(%1); \n\t" \ 284 "cache %2, 0x090(%0); cache %2, 0x090(%1); \n\t" \ 285 "cache %2, 0x0a0(%0); cache %2, 0x0a0(%1); \n\t" \ 286 "cache %2, 0x0b0(%0); cache %2, 0x0b0(%1); \n\t" \ 287 "cache %2, 0x0c0(%0); cache %2, 0x0c0(%1); \n\t" \ 288 "cache %2, 0x0d0(%0); cache %2, 0x0d0(%1); \n\t" \ 289 "cache %2, 0x0e0(%0); cache %2, 0x0e0(%1); \n\t" \ 290 "cache %2, 0x0f0(%0); cache %2, 0x0f0(%1); \n\t" \ 291 ".set reorder" \ 292 : \ 293 : "r" (va1), "r" (va2), "i" (op) \ 294 : "memory"); \ 295 } while (/*CONSTCOND*/0) 296 297 /* 298 * cache_r4k_op_16lines_32_2way: 299 * 300 * Perform the specified cache operation on 16 32-byte 301 * cache lines, 2-ways. 302 */ 303 #define cache_r4k_op_16lines_32_2way(va1, va2, op) \ 304 do { \ 305 __asm __volatile( \ 306 ".set noreorder \n\t" \ 307 "cache %2, 0x000(%0); cache %2, 0x000(%1); \n\t" \ 308 "cache %2, 0x020(%0); cache %2, 0x020(%1); \n\t" \ 309 "cache %2, 0x040(%0); cache %2, 0x040(%1); \n\t" \ 310 "cache %2, 0x060(%0); cache %2, 0x060(%1); \n\t" \ 311 "cache %2, 0x080(%0); cache %2, 0x080(%1); \n\t" \ 312 "cache %2, 0x0a0(%0); cache %2, 0x0a0(%1); \n\t" \ 313 "cache %2, 0x0c0(%0); cache %2, 0x0c0(%1); \n\t" \ 314 "cache %2, 0x0e0(%0); cache %2, 0x0e0(%1); \n\t" \ 315 "cache %2, 0x100(%0); cache %2, 0x100(%1); \n\t" \ 316 "cache %2, 0x120(%0); cache %2, 0x120(%1); \n\t" \ 317 "cache %2, 0x140(%0); cache %2, 0x140(%1); \n\t" \ 318 "cache %2, 0x160(%0); cache %2, 0x160(%1); \n\t" \ 319 "cache %2, 0x180(%0); cache %2, 0x180(%1); \n\t" \ 320 "cache %2, 0x1a0(%0); cache %2, 0x1a0(%1); \n\t" \ 321 "cache %2, 0x1c0(%0); cache %2, 0x1c0(%1); \n\t" \ 322 "cache %2, 0x1e0(%0); cache %2, 0x1e0(%1); \n\t" \ 323 ".set reorder" \ 324 : \ 325 : "r" (va1), "r" (va2), "i" (op) \ 326 : "memory"); \ 327 } while (/*CONSTCOND*/0) 328 329 /* 330 * cache_r4k_op_8lines_16_4way: 331 * 332 * Perform the specified cache operation on 8 16-byte 333 * cache lines, 4-ways. 334 */ 335 #define cache_r4k_op_8lines_16_4way(va1, va2, va3, va4, op) \ 336 do { \ 337 __asm __volatile( \ 338 ".set noreorder \n\t" \ 339 "cache %4, 0x000(%0); cache %4, 0x000(%1); \n\t" \ 340 "cache %4, 0x000(%2); cache %4, 0x000(%3); \n\t" \ 341 "cache %4, 0x010(%0); cache %4, 0x010(%1); \n\t" \ 342 "cache %4, 0x010(%2); cache %4, 0x010(%3); \n\t" \ 343 "cache %4, 0x020(%0); cache %4, 0x020(%1); \n\t" \ 344 "cache %4, 0x020(%2); cache %4, 0x020(%3); \n\t" \ 345 "cache %4, 0x030(%0); cache %4, 0x030(%1); \n\t" \ 346 "cache %4, 0x030(%2); cache %4, 0x030(%3); \n\t" \ 347 "cache %4, 0x040(%0); cache %4, 0x040(%1); \n\t" \ 348 "cache %4, 0x040(%2); cache %4, 0x040(%3); \n\t" \ 349 "cache %4, 0x050(%0); cache %4, 0x050(%1); \n\t" \ 350 "cache %4, 0x050(%2); cache %4, 0x050(%3); \n\t" \ 351 "cache %4, 0x060(%0); cache %4, 0x060(%1); \n\t" \ 352 "cache %4, 0x060(%2); cache %4, 0x060(%3); \n\t" \ 353 "cache %4, 0x070(%0); cache %4, 0x070(%1); \n\t" \ 354 "cache %4, 0x070(%2); cache %4, 0x070(%3); \n\t" \ 355 ".set reorder" \ 356 : \ 357 : "r" (va1), "r" (va2), "r" (va3), "r" (va4), "i" (op) \ 358 : "memory"); \ 359 } while (/*CONSTCOND*/0) 360 361 /* 362 * cache_r4k_op_8lines_32_4way: 363 * 364 * Perform the specified cache operation on 8 32-byte 365 * cache lines, 4-ways. 366 */ 367 #define cache_r4k_op_8lines_32_4way(va1, va2, va3, va4, op) \ 368 do { \ 369 __asm __volatile( \ 370 ".set noreorder \n\t" \ 371 "cache %4, 0x000(%0); cache %4, 0x000(%1); \n\t" \ 372 "cache %4, 0x000(%2); cache %4, 0x000(%3); \n\t" \ 373 "cache %4, 0x020(%0); cache %4, 0x020(%1); \n\t" \ 374 "cache %4, 0x020(%2); cache %4, 0x020(%3); \n\t" \ 375 "cache %4, 0x040(%0); cache %4, 0x040(%1); \n\t" \ 376 "cache %4, 0x040(%2); cache %4, 0x040(%3); \n\t" \ 377 "cache %4, 0x060(%0); cache %4, 0x060(%1); \n\t" \ 378 "cache %4, 0x060(%2); cache %4, 0x060(%3); \n\t" \ 379 "cache %4, 0x080(%0); cache %4, 0x080(%1); \n\t" \ 380 "cache %4, 0x080(%2); cache %4, 0x080(%3); \n\t" \ 381 "cache %4, 0x0a0(%0); cache %4, 0x0a0(%1); \n\t" \ 382 "cache %4, 0x0a0(%2); cache %4, 0x0a0(%3); \n\t" \ 383 "cache %4, 0x0c0(%0); cache %4, 0x0c0(%1); \n\t" \ 384 "cache %4, 0x0c0(%2); cache %4, 0x0c0(%3); \n\t" \ 385 "cache %4, 0x0e0(%0); cache %4, 0x0e0(%1); \n\t" \ 386 "cache %4, 0x0e0(%2); cache %4, 0x0e0(%3); \n\t" \ 387 ".set reorder" \ 388 : \ 389 : "r" (va1), "r" (va2), "r" (va3), "r" (va4), "i" (op) \ 390 : "memory"); \ 391 } while (/*CONSTCOND*/0) 392 393 void r4k_icache_sync_all_16(void); 394 void r4k_icache_sync_range_16(vm_paddr_t, vm_size_t); 395 void r4k_icache_sync_range_index_16(vm_paddr_t, vm_size_t); 396 397 void r4k_icache_sync_all_32(void); 398 void r4k_icache_sync_range_32(vm_paddr_t, vm_size_t); 399 void r4k_icache_sync_range_index_32(vm_paddr_t, vm_size_t); 400 401 void r4k_pdcache_wbinv_all_16(void); 402 void r4k_pdcache_wbinv_range_16(vm_paddr_t, vm_size_t); 403 void r4k_pdcache_wbinv_range_index_16(vm_paddr_t, vm_size_t); 404 405 void r4k_pdcache_inv_range_16(vm_paddr_t, vm_size_t); 406 void r4k_pdcache_wb_range_16(vm_paddr_t, vm_size_t); 407 408 void r4k_pdcache_wbinv_all_32(void); 409 void r4k_pdcache_wbinv_range_32(vm_paddr_t, vm_size_t); 410 void r4k_pdcache_wbinv_range_index_32(vm_paddr_t, vm_size_t); 411 412 void r4k_pdcache_inv_range_32(vm_paddr_t, vm_size_t); 413 void r4k_pdcache_wb_range_32(vm_paddr_t, vm_size_t); 414 415 void r4k_sdcache_wbinv_all_32(void); 416 void r4k_sdcache_wbinv_range_32(vm_paddr_t, vm_size_t); 417 void r4k_sdcache_wbinv_range_index_32(vm_paddr_t, vm_size_t); 418 419 void r4k_sdcache_inv_range_32(vm_paddr_t, vm_size_t); 420 void r4k_sdcache_wb_range_32(vm_paddr_t, vm_size_t); 421 422 void r4k_sdcache_wbinv_all_128(void); 423 void r4k_sdcache_wbinv_range_128(vm_paddr_t, vm_size_t); 424 void r4k_sdcache_wbinv_range_index_128(vm_paddr_t, vm_size_t); 425 426 void r4k_sdcache_inv_range_128(vm_paddr_t, vm_size_t); 427 void r4k_sdcache_wb_range_128(vm_paddr_t, vm_size_t); 428 429 void r4k_sdcache_wbinv_all_generic(void); 430 void r4k_sdcache_wbinv_range_generic(vm_paddr_t, vm_size_t); 431 void r4k_sdcache_wbinv_range_index_generic(vm_paddr_t, vm_size_t); 432 433 void r4k_sdcache_inv_range_generic(vm_paddr_t, vm_size_t); 434 void r4k_sdcache_wb_range_generic(vm_paddr_t, vm_size_t); 435 436 #endif /* !LOCORE */ 437