1 /*-
2 * Copyright (c) 2010 Isilon Systems, Inc.
3 * Copyright (c) 2010 iX Systems, Inc.
4 * Copyright (c) 2010 Panasas, Inc.
5 * Copyright (c) 2013-2017 Mellanox Technologies, Ltd.
6 * All rights reserved.
7 *
8 * Redistribution and use in source and binary forms, with or without
9 * modification, are permitted provided that the following conditions
10 * are met:
11 * 1. Redistributions of source code must retain the above copyright
12 * notice unmodified, this list of conditions, and the following
13 * disclaimer.
14 * 2. Redistributions in binary form must reproduce the above copyright
15 * notice, this list of conditions and the following disclaimer in the
16 * documentation and/or other materials provided with the distribution.
17 *
18 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
19 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
20 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
21 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
22 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
23 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
24 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
25 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
26 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
27 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
28 */
29
30 #include <sys/cdefs.h>
31 __FBSDID("$FreeBSD$");
32
33 #include <sys/param.h>
34 #include <sys/systm.h>
35 #include <sys/malloc.h>
36 #include <sys/kernel.h>
37 #include <sys/sysctl.h>
38 #include <sys/lock.h>
39 #include <sys/mutex.h>
40
41 #include <machine/stdarg.h>
42
43 #include <linux/bitmap.h>
44 #include <linux/kobject.h>
45 #include <linux/slab.h>
46 #include <linux/idr.h>
47 #include <linux/err.h>
48
49 #define MAX_IDR_LEVEL ((MAX_IDR_SHIFT + IDR_BITS - 1) / IDR_BITS)
50 #define MAX_IDR_FREE (MAX_IDR_LEVEL * 2)
51
52 struct linux_idr_cache {
53 spinlock_t lock;
54 struct idr_layer *head;
55 unsigned count;
56 };
57
58 DPCPU_DEFINE_STATIC(struct linux_idr_cache, linux_idr_cache);
59
60 /*
61 * IDR Implementation.
62 *
63 * This is quick and dirty and not as re-entrant as the linux version
64 * however it should be fairly fast. It is basically a radix tree with
65 * a builtin bitmap for allocation.
66 */
67 static MALLOC_DEFINE(M_IDR, "idr", "Linux IDR compat");
68
69 static struct idr_layer *
idr_preload_dequeue_locked(struct linux_idr_cache * lic)70 idr_preload_dequeue_locked(struct linux_idr_cache *lic)
71 {
72 struct idr_layer *retval;
73
74 /* check if wrong thread is trying to dequeue */
75 if (mtx_owned(&lic->lock.m) == 0)
76 return (NULL);
77
78 retval = lic->head;
79 if (likely(retval != NULL)) {
80 lic->head = retval->ary[0];
81 lic->count--;
82 retval->ary[0] = NULL;
83 }
84 return (retval);
85 }
86
87 static void
idr_preload_init(void * arg)88 idr_preload_init(void *arg)
89 {
90 int cpu;
91
92 CPU_FOREACH(cpu) {
93 struct linux_idr_cache *lic =
94 DPCPU_ID_PTR(cpu, linux_idr_cache);
95
96 spin_lock_init(&lic->lock);
97 }
98 }
99 SYSINIT(idr_preload_init, SI_SUB_CPU, SI_ORDER_ANY, idr_preload_init, NULL);
100
101 static void
idr_preload_uninit(void * arg)102 idr_preload_uninit(void *arg)
103 {
104 int cpu;
105
106 CPU_FOREACH(cpu) {
107 struct idr_layer *cacheval;
108 struct linux_idr_cache *lic =
109 DPCPU_ID_PTR(cpu, linux_idr_cache);
110
111 while (1) {
112 spin_lock(&lic->lock);
113 cacheval = idr_preload_dequeue_locked(lic);
114 spin_unlock(&lic->lock);
115
116 if (cacheval == NULL)
117 break;
118 free(cacheval, M_IDR);
119 }
120 spin_lock_destroy(&lic->lock);
121 }
122 }
123 SYSUNINIT(idr_preload_uninit, SI_SUB_LOCK, SI_ORDER_FIRST, idr_preload_uninit, NULL);
124
125 void
idr_preload(gfp_t gfp_mask)126 idr_preload(gfp_t gfp_mask)
127 {
128 struct linux_idr_cache *lic;
129 struct idr_layer *cacheval;
130
131 sched_pin();
132
133 lic = &DPCPU_GET(linux_idr_cache);
134
135 /* fill up cache */
136 spin_lock(&lic->lock);
137 while (lic->count < MAX_IDR_FREE) {
138 spin_unlock(&lic->lock);
139 cacheval = malloc(sizeof(*cacheval), M_IDR, M_ZERO | gfp_mask);
140 spin_lock(&lic->lock);
141 if (cacheval == NULL)
142 break;
143 cacheval->ary[0] = lic->head;
144 lic->head = cacheval;
145 lic->count++;
146 }
147 }
148
149 void
idr_preload_end(void)150 idr_preload_end(void)
151 {
152 struct linux_idr_cache *lic;
153
154 lic = &DPCPU_GET(linux_idr_cache);
155 spin_unlock(&lic->lock);
156 sched_unpin();
157 }
158
159 static inline int
idr_max(struct idr * idr)160 idr_max(struct idr *idr)
161 {
162 return (1 << (idr->layers * IDR_BITS)) - 1;
163 }
164
165 static inline int
idr_pos(int id,int layer)166 idr_pos(int id, int layer)
167 {
168 return (id >> (IDR_BITS * layer)) & IDR_MASK;
169 }
170
171 void
idr_init(struct idr * idr)172 idr_init(struct idr *idr)
173 {
174 bzero(idr, sizeof(*idr));
175 mtx_init(&idr->lock, "idr", NULL, MTX_DEF);
176 }
177
178 /* Only frees cached pages. */
179 void
idr_destroy(struct idr * idr)180 idr_destroy(struct idr *idr)
181 {
182 struct idr_layer *il, *iln;
183
184 idr_remove_all(idr);
185 mtx_lock(&idr->lock);
186 for (il = idr->free; il != NULL; il = iln) {
187 iln = il->ary[0];
188 free(il, M_IDR);
189 }
190 mtx_unlock(&idr->lock);
191 mtx_destroy(&idr->lock);
192 }
193
194 static void
idr_remove_layer(struct idr_layer * il,int layer)195 idr_remove_layer(struct idr_layer *il, int layer)
196 {
197 int i;
198
199 if (il == NULL)
200 return;
201 if (layer == 0) {
202 free(il, M_IDR);
203 return;
204 }
205 for (i = 0; i < IDR_SIZE; i++)
206 if (il->ary[i])
207 idr_remove_layer(il->ary[i], layer - 1);
208 }
209
210 void
idr_remove_all(struct idr * idr)211 idr_remove_all(struct idr *idr)
212 {
213
214 mtx_lock(&idr->lock);
215 idr_remove_layer(idr->top, idr->layers - 1);
216 idr->top = NULL;
217 idr->layers = 0;
218 mtx_unlock(&idr->lock);
219 }
220
221 static void *
idr_remove_locked(struct idr * idr,int id)222 idr_remove_locked(struct idr *idr, int id)
223 {
224 struct idr_layer *il;
225 void *res;
226 int layer;
227 int idx;
228
229 id &= MAX_ID_MASK;
230 il = idr->top;
231 layer = idr->layers - 1;
232 if (il == NULL || id > idr_max(idr))
233 return (NULL);
234 /*
235 * Walk down the tree to this item setting bitmaps along the way
236 * as we know at least one item will be free along this path.
237 */
238 while (layer && il) {
239 idx = idr_pos(id, layer);
240 il->bitmap |= 1 << idx;
241 il = il->ary[idx];
242 layer--;
243 }
244 idx = id & IDR_MASK;
245 /*
246 * At this point we've set free space bitmaps up the whole tree.
247 * We could make this non-fatal and unwind but linux dumps a stack
248 * and a warning so I don't think it's necessary.
249 */
250 if (il == NULL || (il->bitmap & (1 << idx)) != 0)
251 panic("idr_remove: Item %d not allocated (%p, %p)\n",
252 id, idr, il);
253 res = il->ary[idx];
254 il->ary[idx] = NULL;
255 il->bitmap |= 1 << idx;
256
257 return (res);
258 }
259
260 void *
idr_remove(struct idr * idr,int id)261 idr_remove(struct idr *idr, int id)
262 {
263 void *res;
264
265 mtx_lock(&idr->lock);
266 res = idr_remove_locked(idr, id);
267 mtx_unlock(&idr->lock);
268
269 return (res);
270 }
271
272
273 static inline struct idr_layer *
idr_find_layer_locked(struct idr * idr,int id)274 idr_find_layer_locked(struct idr *idr, int id)
275 {
276 struct idr_layer *il;
277 int layer;
278
279 id &= MAX_ID_MASK;
280 il = idr->top;
281 layer = idr->layers - 1;
282 if (il == NULL || id > idr_max(idr))
283 return (NULL);
284 while (layer && il) {
285 il = il->ary[idr_pos(id, layer)];
286 layer--;
287 }
288 return (il);
289 }
290
291 void *
idr_replace(struct idr * idr,void * ptr,int id)292 idr_replace(struct idr *idr, void *ptr, int id)
293 {
294 struct idr_layer *il;
295 void *res;
296 int idx;
297
298 mtx_lock(&idr->lock);
299 il = idr_find_layer_locked(idr, id);
300 idx = id & IDR_MASK;
301
302 /* Replace still returns an error if the item was not allocated. */
303 if (il == NULL || (il->bitmap & (1 << idx))) {
304 res = ERR_PTR(-ENOENT);
305 } else {
306 res = il->ary[idx];
307 il->ary[idx] = ptr;
308 }
309 mtx_unlock(&idr->lock);
310 return (res);
311 }
312
313 static inline void *
idr_find_locked(struct idr * idr,int id)314 idr_find_locked(struct idr *idr, int id)
315 {
316 struct idr_layer *il;
317 void *res;
318
319 mtx_assert(&idr->lock, MA_OWNED);
320 il = idr_find_layer_locked(idr, id);
321 if (il != NULL)
322 res = il->ary[id & IDR_MASK];
323 else
324 res = NULL;
325 return (res);
326 }
327
328 void *
idr_find(struct idr * idr,int id)329 idr_find(struct idr *idr, int id)
330 {
331 void *res;
332
333 mtx_lock(&idr->lock);
334 res = idr_find_locked(idr, id);
335 mtx_unlock(&idr->lock);
336 return (res);
337 }
338
339 void *
idr_get_next(struct idr * idr,int * nextidp)340 idr_get_next(struct idr *idr, int *nextidp)
341 {
342 void *res = NULL;
343 int id = *nextidp;
344
345 mtx_lock(&idr->lock);
346 for (; id <= idr_max(idr); id++) {
347 res = idr_find_locked(idr, id);
348 if (res == NULL)
349 continue;
350 *nextidp = id;
351 break;
352 }
353 mtx_unlock(&idr->lock);
354 return (res);
355 }
356
357 int
idr_pre_get(struct idr * idr,gfp_t gfp_mask)358 idr_pre_get(struct idr *idr, gfp_t gfp_mask)
359 {
360 struct idr_layer *il, *iln;
361 struct idr_layer *head;
362 int need;
363
364 mtx_lock(&idr->lock);
365 for (;;) {
366 need = idr->layers + 1;
367 for (il = idr->free; il != NULL; il = il->ary[0])
368 need--;
369 mtx_unlock(&idr->lock);
370 if (need <= 0)
371 break;
372 for (head = NULL; need; need--) {
373 iln = malloc(sizeof(*il), M_IDR, M_ZERO | gfp_mask);
374 if (iln == NULL)
375 break;
376 bitmap_fill(&iln->bitmap, IDR_SIZE);
377 if (head != NULL) {
378 il->ary[0] = iln;
379 il = iln;
380 } else
381 head = il = iln;
382 }
383 if (head == NULL)
384 return (0);
385 mtx_lock(&idr->lock);
386 il->ary[0] = idr->free;
387 idr->free = head;
388 }
389 return (1);
390 }
391
392 static struct idr_layer *
idr_free_list_get(struct idr * idp)393 idr_free_list_get(struct idr *idp)
394 {
395 struct idr_layer *il;
396
397 if ((il = idp->free) != NULL) {
398 idp->free = il->ary[0];
399 il->ary[0] = NULL;
400 }
401 return (il);
402 }
403
404 static inline struct idr_layer *
idr_get(struct idr * idp)405 idr_get(struct idr *idp)
406 {
407 struct idr_layer *il;
408
409 if ((il = idr_free_list_get(idp)) != NULL) {
410 MPASS(il->bitmap != 0);
411 } else if ((il = malloc(sizeof(*il), M_IDR, M_ZERO | M_NOWAIT)) != NULL) {
412 bitmap_fill(&il->bitmap, IDR_SIZE);
413 } else if ((il = idr_preload_dequeue_locked(&DPCPU_GET(linux_idr_cache))) != NULL) {
414 bitmap_fill(&il->bitmap, IDR_SIZE);
415 } else {
416 return (NULL);
417 }
418 return (il);
419 }
420
421 /*
422 * Could be implemented as get_new_above(idr, ptr, 0, idp) but written
423 * first for simplicity sake.
424 */
425 static int
idr_get_new_locked(struct idr * idr,void * ptr,int * idp)426 idr_get_new_locked(struct idr *idr, void *ptr, int *idp)
427 {
428 struct idr_layer *stack[MAX_LEVEL];
429 struct idr_layer *il;
430 int error;
431 int layer;
432 int idx;
433 int id;
434
435 mtx_assert(&idr->lock, MA_OWNED);
436
437 error = -EAGAIN;
438 /*
439 * Expand the tree until there is free space.
440 */
441 if (idr->top == NULL || idr->top->bitmap == 0) {
442 if (idr->layers == MAX_LEVEL + 1) {
443 error = -ENOSPC;
444 goto out;
445 }
446 il = idr_get(idr);
447 if (il == NULL)
448 goto out;
449 il->ary[0] = idr->top;
450 if (idr->top)
451 il->bitmap &= ~1;
452 idr->top = il;
453 idr->layers++;
454 }
455 il = idr->top;
456 id = 0;
457 /*
458 * Walk the tree following free bitmaps, record our path.
459 */
460 for (layer = idr->layers - 1;; layer--) {
461 stack[layer] = il;
462 idx = ffsl(il->bitmap);
463 if (idx == 0)
464 panic("idr_get_new: Invalid leaf state (%p, %p)\n",
465 idr, il);
466 idx--;
467 id |= idx << (layer * IDR_BITS);
468 if (layer == 0)
469 break;
470 if (il->ary[idx] == NULL) {
471 il->ary[idx] = idr_get(idr);
472 if (il->ary[idx] == NULL)
473 goto out;
474 }
475 il = il->ary[idx];
476 }
477 /*
478 * Allocate the leaf to the consumer.
479 */
480 il->bitmap &= ~(1 << idx);
481 il->ary[idx] = ptr;
482 *idp = id;
483 /*
484 * Clear bitmaps potentially up to the root.
485 */
486 while (il->bitmap == 0 && ++layer < idr->layers) {
487 il = stack[layer];
488 il->bitmap &= ~(1 << idr_pos(id, layer));
489 }
490 error = 0;
491 out:
492 #ifdef INVARIANTS
493 if (error == 0 && idr_find_locked(idr, id) != ptr) {
494 panic("idr_get_new: Failed for idr %p, id %d, ptr %p\n",
495 idr, id, ptr);
496 }
497 #endif
498 return (error);
499 }
500
501 int
idr_get_new(struct idr * idr,void * ptr,int * idp)502 idr_get_new(struct idr *idr, void *ptr, int *idp)
503 {
504 int retval;
505
506 mtx_lock(&idr->lock);
507 retval = idr_get_new_locked(idr, ptr, idp);
508 mtx_unlock(&idr->lock);
509 return (retval);
510 }
511
512 static int
idr_get_new_above_locked(struct idr * idr,void * ptr,int starting_id,int * idp)513 idr_get_new_above_locked(struct idr *idr, void *ptr, int starting_id, int *idp)
514 {
515 struct idr_layer *stack[MAX_LEVEL];
516 struct idr_layer *il;
517 int error;
518 int layer;
519 int idx, sidx;
520 int id;
521
522 mtx_assert(&idr->lock, MA_OWNED);
523
524 error = -EAGAIN;
525 /*
526 * Compute the layers required to support starting_id and the mask
527 * at the top layer.
528 */
529 restart:
530 idx = starting_id;
531 layer = 0;
532 while (idx & ~IDR_MASK) {
533 layer++;
534 idx >>= IDR_BITS;
535 }
536 if (layer == MAX_LEVEL + 1) {
537 error = -ENOSPC;
538 goto out;
539 }
540 /*
541 * Expand the tree until there is free space at or beyond starting_id.
542 */
543 while (idr->layers <= layer ||
544 idr->top->bitmap < (1 << idr_pos(starting_id, idr->layers - 1))) {
545 if (idr->layers == MAX_LEVEL + 1) {
546 error = -ENOSPC;
547 goto out;
548 }
549 il = idr_get(idr);
550 if (il == NULL)
551 goto out;
552 il->ary[0] = idr->top;
553 if (idr->top && idr->top->bitmap == 0)
554 il->bitmap &= ~1;
555 idr->top = il;
556 idr->layers++;
557 }
558 il = idr->top;
559 id = 0;
560 /*
561 * Walk the tree following free bitmaps, record our path.
562 */
563 for (layer = idr->layers - 1;; layer--) {
564 stack[layer] = il;
565 sidx = idr_pos(starting_id, layer);
566 /* Returns index numbered from 0 or size if none exists. */
567 idx = find_next_bit(&il->bitmap, IDR_SIZE, sidx);
568 if (idx == IDR_SIZE && sidx == 0)
569 panic("idr_get_new: Invalid leaf state (%p, %p)\n",
570 idr, il);
571 /*
572 * We may have walked a path where there was a free bit but
573 * it was lower than what we wanted. Restart the search with
574 * a larger starting id. id contains the progress we made so
575 * far. Search the leaf one above this level. This may
576 * restart as many as MAX_LEVEL times but that is expected
577 * to be rare.
578 */
579 if (idx == IDR_SIZE) {
580 starting_id = id + (1 << ((layer + 1) * IDR_BITS));
581 goto restart;
582 }
583 if (idx > sidx)
584 starting_id = 0; /* Search the whole subtree. */
585 id |= idx << (layer * IDR_BITS);
586 if (layer == 0)
587 break;
588 if (il->ary[idx] == NULL) {
589 il->ary[idx] = idr_get(idr);
590 if (il->ary[idx] == NULL)
591 goto out;
592 }
593 il = il->ary[idx];
594 }
595 /*
596 * Allocate the leaf to the consumer.
597 */
598 il->bitmap &= ~(1 << idx);
599 il->ary[idx] = ptr;
600 *idp = id;
601 /*
602 * Clear bitmaps potentially up to the root.
603 */
604 while (il->bitmap == 0 && ++layer < idr->layers) {
605 il = stack[layer];
606 il->bitmap &= ~(1 << idr_pos(id, layer));
607 }
608 error = 0;
609 out:
610 #ifdef INVARIANTS
611 if (error == 0 && idr_find_locked(idr, id) != ptr) {
612 panic("idr_get_new_above: Failed for idr %p, id %d, ptr %p\n",
613 idr, id, ptr);
614 }
615 #endif
616 return (error);
617 }
618
619 int
idr_get_new_above(struct idr * idr,void * ptr,int starting_id,int * idp)620 idr_get_new_above(struct idr *idr, void *ptr, int starting_id, int *idp)
621 {
622 int retval;
623
624 mtx_lock(&idr->lock);
625 retval = idr_get_new_above_locked(idr, ptr, starting_id, idp);
626 mtx_unlock(&idr->lock);
627 return (retval);
628 }
629
630 int
ida_get_new_above(struct ida * ida,int starting_id,int * p_id)631 ida_get_new_above(struct ida *ida, int starting_id, int *p_id)
632 {
633 return (idr_get_new_above(&ida->idr, NULL, starting_id, p_id));
634 }
635
636 static int
idr_alloc_locked(struct idr * idr,void * ptr,int start,int end)637 idr_alloc_locked(struct idr *idr, void *ptr, int start, int end)
638 {
639 int max = end > 0 ? end - 1 : INT_MAX;
640 int error;
641 int id;
642
643 mtx_assert(&idr->lock, MA_OWNED);
644
645 if (unlikely(start < 0))
646 return (-EINVAL);
647 if (unlikely(max < start))
648 return (-ENOSPC);
649
650 if (start == 0)
651 error = idr_get_new_locked(idr, ptr, &id);
652 else
653 error = idr_get_new_above_locked(idr, ptr, start, &id);
654
655 if (unlikely(error < 0))
656 return (error);
657 if (unlikely(id > max)) {
658 idr_remove_locked(idr, id);
659 return (-ENOSPC);
660 }
661 return (id);
662 }
663
664 int
idr_alloc(struct idr * idr,void * ptr,int start,int end,gfp_t gfp_mask)665 idr_alloc(struct idr *idr, void *ptr, int start, int end, gfp_t gfp_mask)
666 {
667 int retval;
668
669 mtx_lock(&idr->lock);
670 retval = idr_alloc_locked(idr, ptr, start, end);
671 mtx_unlock(&idr->lock);
672 return (retval);
673 }
674
675 int
idr_alloc_cyclic(struct idr * idr,void * ptr,int start,int end,gfp_t gfp_mask)676 idr_alloc_cyclic(struct idr *idr, void *ptr, int start, int end, gfp_t gfp_mask)
677 {
678 int retval;
679
680 mtx_lock(&idr->lock);
681 retval = idr_alloc_locked(idr, ptr, max(start, idr->next_cyclic_id), end);
682 if (unlikely(retval == -ENOSPC))
683 retval = idr_alloc_locked(idr, ptr, start, end);
684 if (likely(retval >= 0))
685 idr->next_cyclic_id = retval + 1;
686 mtx_unlock(&idr->lock);
687 return (retval);
688 }
689
690 static int
idr_for_each_layer(struct idr_layer * il,int offset,int layer,int (* f)(int id,void * p,void * data),void * data)691 idr_for_each_layer(struct idr_layer *il, int offset, int layer,
692 int (*f)(int id, void *p, void *data), void *data)
693 {
694 int i, err;
695
696 if (il == NULL)
697 return (0);
698 if (layer == 0) {
699 for (i = 0; i < IDR_SIZE; i++) {
700 if (il->ary[i] == NULL)
701 continue;
702 err = f(i + offset, il->ary[i], data);
703 if (err)
704 return (err);
705 }
706 return (0);
707 }
708 for (i = 0; i < IDR_SIZE; i++) {
709 if (il->ary[i] == NULL)
710 continue;
711 err = idr_for_each_layer(il->ary[i],
712 (i + offset) * IDR_SIZE, layer - 1, f, data);
713 if (err)
714 return (err);
715 }
716 return (0);
717 }
718
719 /* NOTE: It is not allowed to modify the IDR tree while it is being iterated */
720 int
idr_for_each(struct idr * idp,int (* f)(int id,void * p,void * data),void * data)721 idr_for_each(struct idr *idp, int (*f)(int id, void *p, void *data), void *data)
722 {
723 return (idr_for_each_layer(idp->top, 0, idp->layers - 1, f, data));
724 }
725
726 static int
idr_has_entry(int id,void * p,void * data)727 idr_has_entry(int id, void *p, void *data)
728 {
729
730 return (1);
731 }
732
733 bool
idr_is_empty(struct idr * idp)734 idr_is_empty(struct idr *idp)
735 {
736
737 return (idr_for_each(idp, idr_has_entry, NULL) == 0);
738 }
739
740 int
ida_pre_get(struct ida * ida,gfp_t flags)741 ida_pre_get(struct ida *ida, gfp_t flags)
742 {
743 if (idr_pre_get(&ida->idr, flags) == 0)
744 return (0);
745
746 if (ida->free_bitmap == NULL) {
747 ida->free_bitmap =
748 malloc(sizeof(struct ida_bitmap), M_IDR, flags);
749 }
750 return (ida->free_bitmap != NULL);
751 }
752
753 int
ida_simple_get(struct ida * ida,unsigned int start,unsigned int end,gfp_t flags)754 ida_simple_get(struct ida *ida, unsigned int start, unsigned int end,
755 gfp_t flags)
756 {
757 int ret, id;
758 unsigned int max;
759
760 MPASS((int)start >= 0);
761 MPASS((int)end >= 0);
762
763 if (end == 0)
764 max = 0x80000000;
765 else {
766 MPASS(end > start);
767 max = end - 1;
768 }
769 again:
770 if (!ida_pre_get(ida, flags))
771 return (-ENOMEM);
772
773 if ((ret = ida_get_new_above(ida, start, &id)) == 0) {
774 if (id > max) {
775 ida_remove(ida, id);
776 ret = -ENOSPC;
777 } else {
778 ret = id;
779 }
780 }
781 if (__predict_false(ret == -EAGAIN))
782 goto again;
783
784 return (ret);
785 }
786
787 void
ida_simple_remove(struct ida * ida,unsigned int id)788 ida_simple_remove(struct ida *ida, unsigned int id)
789 {
790 idr_remove(&ida->idr, id);
791 }
792
793 void
ida_remove(struct ida * ida,int id)794 ida_remove(struct ida *ida, int id)
795 {
796 idr_remove(&ida->idr, id);
797 }
798
799 void
ida_init(struct ida * ida)800 ida_init(struct ida *ida)
801 {
802 idr_init(&ida->idr);
803 }
804
805 void
ida_destroy(struct ida * ida)806 ida_destroy(struct ida *ida)
807 {
808 idr_destroy(&ida->idr);
809 free(ida->free_bitmap, M_IDR);
810 }
811