bitops.h (282741) | bitops.h (289621) |
---|---|
1/*- 2 * Copyright (c) 2010 Isilon Systems, Inc. 3 * Copyright (c) 2010 iX Systems, Inc. 4 * Copyright (c) 2010 Panasas, Inc. | 1/*- 2 * Copyright (c) 2010 Isilon Systems, Inc. 3 * Copyright (c) 2010 iX Systems, Inc. 4 * Copyright (c) 2010 Panasas, Inc. |
5 * Copyright (c) 2013, 2014 Mellanox Technologies, Ltd. | 5 * Copyright (c) 2013-2015 Mellanox Technologies, Ltd. |
6 * All rights reserved. 7 * 8 * Redistribution and use in source and binary forms, with or without 9 * modification, are permitted provided that the following conditions 10 * are met: 11 * 1. Redistributions of source code must retain the above copyright 12 * notice unmodified, this list of conditions, and the following 13 * disclaimer. --- 10 unchanged lines hidden (view full) --- 24 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 25 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 26 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 27 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 28 */ 29#ifndef _LINUX_BITOPS_H_ 30#define _LINUX_BITOPS_H_ 31 | 6 * All rights reserved. 7 * 8 * Redistribution and use in source and binary forms, with or without 9 * modification, are permitted provided that the following conditions 10 * are met: 11 * 1. Redistributions of source code must retain the above copyright 12 * notice unmodified, this list of conditions, and the following 13 * disclaimer. --- 10 unchanged lines hidden (view full) --- 24 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 25 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 26 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 27 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 28 */ 29#ifndef _LINUX_BITOPS_H_ 30#define _LINUX_BITOPS_H_ 31 |
32#include <sys/types.h> 33#include <sys/systm.h> 34 35#define BIT(nr) (1UL << (nr)) |
|
32#ifdef __LP64__ 33#define BITS_PER_LONG 64 34#else 35#define BITS_PER_LONG 32 36#endif | 36#ifdef __LP64__ 37#define BITS_PER_LONG 64 38#else 39#define BITS_PER_LONG 32 40#endif |
37#define BIT_MASK(n) (~0UL >> (BITS_PER_LONG - (n))) | 41#define BITMAP_FIRST_WORD_MASK(start) (~0UL << ((start) % BITS_PER_LONG)) 42#define BITMAP_LAST_WORD_MASK(n) (~0UL >> (BITS_PER_LONG - (n))) |
38#define BITS_TO_LONGS(n) howmany((n), BITS_PER_LONG) | 43#define BITS_TO_LONGS(n) howmany((n), BITS_PER_LONG) |
44#define BIT_MASK(nr) (1UL << ((nr) & (BITS_PER_LONG - 1))) |
|
39#define BIT_WORD(nr) ((nr) / BITS_PER_LONG) | 45#define BIT_WORD(nr) ((nr) / BITS_PER_LONG) |
40 | 46#define GENMASK(lo, hi) (((2UL << ((hi) - (lo))) - 1UL) << (lo)) |
41#define BITS_PER_BYTE 8 42 43static inline int 44__ffs(int mask) 45{ 46 return (ffs(mask) - 1); 47} 48 --- 36 unchanged lines hidden (view full) --- 85 86 for (bit = 0; size >= BITS_PER_LONG; 87 size -= BITS_PER_LONG, bit += BITS_PER_LONG, addr++) { 88 if (*addr == 0) 89 continue; 90 return (bit + __ffsl(*addr)); 91 } 92 if (size) { | 47#define BITS_PER_BYTE 8 48 49static inline int 50__ffs(int mask) 51{ 52 return (ffs(mask) - 1); 53} 54 --- 36 unchanged lines hidden (view full) --- 91 92 for (bit = 0; size >= BITS_PER_LONG; 93 size -= BITS_PER_LONG, bit += BITS_PER_LONG, addr++) { 94 if (*addr == 0) 95 continue; 96 return (bit + __ffsl(*addr)); 97 } 98 if (size) { |
93 mask = (*addr) & BIT_MASK(size); | 99 mask = (*addr) & BITMAP_LAST_WORD_MASK(size); |
94 if (mask) 95 bit += __ffsl(mask); 96 else 97 bit += size; 98 } 99 return (bit); 100} 101 --- 5 unchanged lines hidden (view full) --- 107 108 for (bit = 0; size >= BITS_PER_LONG; 109 size -= BITS_PER_LONG, bit += BITS_PER_LONG, addr++) { 110 if (~(*addr) == 0) 111 continue; 112 return (bit + __ffsl(~(*addr))); 113 } 114 if (size) { | 100 if (mask) 101 bit += __ffsl(mask); 102 else 103 bit += size; 104 } 105 return (bit); 106} 107 --- 5 unchanged lines hidden (view full) --- 113 114 for (bit = 0; size >= BITS_PER_LONG; 115 size -= BITS_PER_LONG, bit += BITS_PER_LONG, addr++) { 116 if (~(*addr) == 0) 117 continue; 118 return (bit + __ffsl(~(*addr))); 119 } 120 if (size) { |
115 mask = ~(*addr) & BIT_MASK(size); | 121 mask = ~(*addr) & BITMAP_LAST_WORD_MASK(size); |
116 if (mask) 117 bit += __ffsl(mask); 118 else 119 bit += size; 120 } 121 return (bit); 122} 123 --- 5 unchanged lines hidden (view full) --- 129 int bit; 130 int pos; 131 132 pos = size / BITS_PER_LONG; 133 offs = size % BITS_PER_LONG; 134 bit = BITS_PER_LONG * pos; 135 addr += pos; 136 if (offs) { | 122 if (mask) 123 bit += __ffsl(mask); 124 else 125 bit += size; 126 } 127 return (bit); 128} 129 --- 5 unchanged lines hidden (view full) --- 135 int bit; 136 int pos; 137 138 pos = size / BITS_PER_LONG; 139 offs = size % BITS_PER_LONG; 140 bit = BITS_PER_LONG * pos; 141 addr += pos; 142 if (offs) { |
137 mask = (*addr) & BIT_MASK(offs); | 143 mask = (*addr) & BITMAP_LAST_WORD_MASK(offs); |
138 if (mask) 139 return (bit + __flsl(mask)); 140 } 141 while (--pos) { 142 addr--; 143 bit -= BITS_PER_LONG; 144 if (*addr) 145 return (bit + __flsl(mask)); --- 11 unchanged lines hidden (view full) --- 157 158 if (offset >= size) 159 return (size); 160 pos = offset / BITS_PER_LONG; 161 offs = offset % BITS_PER_LONG; 162 bit = BITS_PER_LONG * pos; 163 addr += pos; 164 if (offs) { | 144 if (mask) 145 return (bit + __flsl(mask)); 146 } 147 while (--pos) { 148 addr--; 149 bit -= BITS_PER_LONG; 150 if (*addr) 151 return (bit + __flsl(mask)); --- 11 unchanged lines hidden (view full) --- 163 164 if (offset >= size) 165 return (size); 166 pos = offset / BITS_PER_LONG; 167 offs = offset % BITS_PER_LONG; 168 bit = BITS_PER_LONG * pos; 169 addr += pos; 170 if (offs) { |
165 mask = (*addr) & ~BIT_MASK(offs); | 171 mask = (*addr) & ~BITMAP_LAST_WORD_MASK(offs); |
166 if (mask) 167 return (bit + __ffsl(mask)); 168 if (size - bit <= BITS_PER_LONG) 169 return (size); 170 bit += BITS_PER_LONG; 171 addr++; 172 } 173 for (size -= bit; size >= BITS_PER_LONG; 174 size -= BITS_PER_LONG, bit += BITS_PER_LONG, addr++) { 175 if (*addr == 0) 176 continue; 177 return (bit + __ffsl(*addr)); 178 } 179 if (size) { | 172 if (mask) 173 return (bit + __ffsl(mask)); 174 if (size - bit <= BITS_PER_LONG) 175 return (size); 176 bit += BITS_PER_LONG; 177 addr++; 178 } 179 for (size -= bit; size >= BITS_PER_LONG; 180 size -= BITS_PER_LONG, bit += BITS_PER_LONG, addr++) { 181 if (*addr == 0) 182 continue; 183 return (bit + __ffsl(*addr)); 184 } 185 if (size) { |
180 mask = (*addr) & BIT_MASK(size); | 186 mask = (*addr) & BITMAP_LAST_WORD_MASK(size); |
181 if (mask) 182 bit += __ffsl(mask); 183 else 184 bit += size; 185 } 186 return (bit); 187} 188 --- 8 unchanged lines hidden (view full) --- 197 198 if (offset >= size) 199 return (size); 200 pos = offset / BITS_PER_LONG; 201 offs = offset % BITS_PER_LONG; 202 bit = BITS_PER_LONG * pos; 203 addr += pos; 204 if (offs) { | 187 if (mask) 188 bit += __ffsl(mask); 189 else 190 bit += size; 191 } 192 return (bit); 193} 194 --- 8 unchanged lines hidden (view full) --- 203 204 if (offset >= size) 205 return (size); 206 pos = offset / BITS_PER_LONG; 207 offs = offset % BITS_PER_LONG; 208 bit = BITS_PER_LONG * pos; 209 addr += pos; 210 if (offs) { |
205 mask = ~(*addr) & ~BIT_MASK(offs); | 211 mask = ~(*addr) & ~BITMAP_LAST_WORD_MASK(offs); |
206 if (mask) 207 return (bit + __ffsl(mask)); 208 if (size - bit <= BITS_PER_LONG) 209 return (size); 210 bit += BITS_PER_LONG; 211 addr++; 212 } 213 for (size -= bit; size >= BITS_PER_LONG; 214 size -= BITS_PER_LONG, bit += BITS_PER_LONG, addr++) { 215 if (~(*addr) == 0) 216 continue; 217 return (bit + __ffsl(~(*addr))); 218 } 219 if (size) { | 212 if (mask) 213 return (bit + __ffsl(mask)); 214 if (size - bit <= BITS_PER_LONG) 215 return (size); 216 bit += BITS_PER_LONG; 217 addr++; 218 } 219 for (size -= bit; size >= BITS_PER_LONG; 220 size -= BITS_PER_LONG, bit += BITS_PER_LONG, addr++) { 221 if (~(*addr) == 0) 222 continue; 223 return (bit + __ffsl(~(*addr))); 224 } 225 if (size) { |
220 mask = ~(*addr) & BIT_MASK(size); | 226 mask = ~(*addr) & BITMAP_LAST_WORD_MASK(size); |
221 if (mask) 222 bit += __ffsl(mask); 223 else 224 bit += size; 225 } 226 return (bit); 227} 228 --- 11 unchanged lines hidden (view full) --- 240{ 241 int tail; 242 int len; 243 244 len = (size / BITS_PER_LONG) * sizeof(long); 245 memset(addr, 0xff, len); 246 tail = size & (BITS_PER_LONG - 1); 247 if (tail) | 227 if (mask) 228 bit += __ffsl(mask); 229 else 230 bit += size; 231 } 232 return (bit); 233} 234 --- 11 unchanged lines hidden (view full) --- 246{ 247 int tail; 248 int len; 249 250 len = (size / BITS_PER_LONG) * sizeof(long); 251 memset(addr, 0xff, len); 252 tail = size & (BITS_PER_LONG - 1); 253 if (tail) |
248 addr[size / BITS_PER_LONG] = BIT_MASK(tail); | 254 addr[size / BITS_PER_LONG] = BITMAP_LAST_WORD_MASK(tail); |
249} 250 251static inline int 252bitmap_full(unsigned long *addr, int size) 253{ | 255} 256 257static inline int 258bitmap_full(unsigned long *addr, int size) 259{ |
254 long mask; | 260 unsigned long mask; |
255 int tail; 256 int len; 257 int i; 258 259 len = size / BITS_PER_LONG; 260 for (i = 0; i < len; i++) 261 if (addr[i] != ~0UL) 262 return (0); 263 tail = size & (BITS_PER_LONG - 1); 264 if (tail) { | 261 int tail; 262 int len; 263 int i; 264 265 len = size / BITS_PER_LONG; 266 for (i = 0; i < len; i++) 267 if (addr[i] != ~0UL) 268 return (0); 269 tail = size & (BITS_PER_LONG - 1); 270 if (tail) { |
265 mask = BIT_MASK(tail); | 271 mask = BITMAP_LAST_WORD_MASK(tail); |
266 if ((addr[i] & mask) != mask) 267 return (0); 268 } 269 return (1); 270} 271 272static inline int 273bitmap_empty(unsigned long *addr, int size) 274{ | 272 if ((addr[i] & mask) != mask) 273 return (0); 274 } 275 return (1); 276} 277 278static inline int 279bitmap_empty(unsigned long *addr, int size) 280{ |
275 long mask; | 281 unsigned long mask; |
276 int tail; 277 int len; 278 int i; 279 280 len = size / BITS_PER_LONG; 281 for (i = 0; i < len; i++) 282 if (addr[i] != 0) 283 return (0); 284 tail = size & (BITS_PER_LONG - 1); 285 if (tail) { | 282 int tail; 283 int len; 284 int i; 285 286 len = size / BITS_PER_LONG; 287 for (i = 0; i < len; i++) 288 if (addr[i] != 0) 289 return (0); 290 tail = size & (BITS_PER_LONG - 1); 291 if (tail) { |
286 mask = BIT_MASK(tail); | 292 mask = BITMAP_LAST_WORD_MASK(tail); |
287 if ((addr[i] & mask) != 0) 288 return (0); 289 } 290 return (1); 291} 292 | 293 if ((addr[i] & mask) != 0) 294 return (0); 295 } 296 return (1); 297} 298 |
293#define NBLONG (NBBY * sizeof(long)) 294 | |
295#define __set_bit(i, a) \ | 299#define __set_bit(i, a) \ |
296 atomic_set_long(&((volatile long *)(a))[(i)/NBLONG], 1UL << ((i) % NBLONG)) | 300 atomic_set_long(&((volatile long *)(a))[BIT_WORD(i)], BIT_MASK(i)) |
297 298#define set_bit(i, a) \ | 301 302#define set_bit(i, a) \ |
299 atomic_set_long(&((volatile long *)(a))[(i)/NBLONG], 1UL << ((i) % NBLONG)) | 303 atomic_set_long(&((volatile long *)(a))[BIT_WORD(i)], BIT_MASK(i)) |
300 301#define __clear_bit(i, a) \ | 304 305#define __clear_bit(i, a) \ |
302 atomic_clear_long(&((volatile long *)(a))[(i)/NBLONG], 1UL << ((i) % NBLONG)) | 306 atomic_clear_long(&((volatile long *)(a))[BIT_WORD(i)], BIT_MASK(i)) |
303 304#define clear_bit(i, a) \ | 307 308#define clear_bit(i, a) \ |
305 atomic_clear_long(&((volatile long *)(a))[(i)/NBLONG], 1UL << ((i) % NBLONG)) | 309 atomic_clear_long(&((volatile long *)(a))[BIT_WORD(i)], BIT_MASK(i)) |
306 307#define test_bit(i, a) \ | 310 311#define test_bit(i, a) \ |
308 !!(atomic_load_acq_long(&((volatile long *)(a))[(i)/NBLONG]) & \ 309 (1UL << ((i) % NBLONG))) | 312 !!(atomic_load_acq_long(&((volatile long *)(a))[BIT_WORD(i)]) & \ 313 BIT_MASK(i)) |
310 311static inline long 312test_and_clear_bit(long bit, long *var) 313{ 314 long val; 315 | 314 315static inline long 316test_and_clear_bit(long bit, long *var) 317{ 318 long val; 319 |
316 var += bit / (sizeof(long) * NBBY); 317 bit %= sizeof(long) * NBBY; | 320 var += BIT_WORD(bit); 321 bit %= BITS_PER_LONG; |
318 bit = (1UL << bit); 319 do { 320 val = *(volatile long *)var; 321 } while (atomic_cmpset_long(var, val, val & ~bit) == 0); 322 323 return !!(val & bit); 324} 325 326static inline long 327test_and_set_bit(long bit, long *var) 328{ 329 long val; 330 | 322 bit = (1UL << bit); 323 do { 324 val = *(volatile long *)var; 325 } while (atomic_cmpset_long(var, val, val & ~bit) == 0); 326 327 return !!(val & bit); 328} 329 330static inline long 331test_and_set_bit(long bit, long *var) 332{ 333 long val; 334 |
331 var += bit / (sizeof(long) * NBBY); 332 bit %= sizeof(long) * NBBY; | 335 var += BIT_WORD(bit); 336 bit %= BITS_PER_LONG; |
333 bit = (1UL << bit); 334 do { 335 val = *(volatile long *)var; 336 } while (atomic_cmpset_long(var, val, val | bit) == 0); 337 338 return !!(val & bit); 339} 340 | 337 bit = (1UL << bit); 338 do { 339 val = *(volatile long *)var; 340 } while (atomic_cmpset_long(var, val, val | bit) == 0); 341 342 return !!(val & bit); 343} 344 |
341 342#define BITMAP_FIRST_WORD_MASK(start) (~0UL << ((start) % BITS_PER_LONG)) 343#define BITMAP_LAST_WORD_MASK(nbits) \ 344( \ 345 ((nbits) % BITS_PER_LONG) ? \ 346 (1UL<<((nbits) % BITS_PER_LONG))-1 : ~0UL \ 347) 348 349 | |
350static inline void 351bitmap_set(unsigned long *map, int start, int nr) 352{ 353 unsigned long *p = map + BIT_WORD(start); 354 const int size = start + nr; 355 int bits_to_set = BITS_PER_LONG - (start % BITS_PER_LONG); 356 unsigned long mask_to_set = BITMAP_FIRST_WORD_MASK(start); 357 --- 27 unchanged lines hidden (view full) --- 385 } 386 if (nr) { 387 mask_to_clear &= BITMAP_LAST_WORD_MASK(size); 388 *p &= ~mask_to_clear; 389 } 390} 391 392enum { | 345static inline void 346bitmap_set(unsigned long *map, int start, int nr) 347{ 348 unsigned long *p = map + BIT_WORD(start); 349 const int size = start + nr; 350 int bits_to_set = BITS_PER_LONG - (start % BITS_PER_LONG); 351 unsigned long mask_to_set = BITMAP_FIRST_WORD_MASK(start); 352 --- 27 unchanged lines hidden (view full) --- 380 } 381 if (nr) { 382 mask_to_clear &= BITMAP_LAST_WORD_MASK(size); 383 *p &= ~mask_to_clear; 384 } 385} 386 387enum { |
393 REG_OP_ISFREE, /* true if region is all zero bits */ 394 REG_OP_ALLOC, /* set all bits in region */ 395 REG_OP_RELEASE, /* clear all bits in region */ | 388 REG_OP_ISFREE, 389 REG_OP_ALLOC, 390 REG_OP_RELEASE, |
396}; 397 398static int __reg_op(unsigned long *bitmap, int pos, int order, int reg_op) 399{ | 391}; 392 393static int __reg_op(unsigned long *bitmap, int pos, int order, int reg_op) 394{ |
400 int nbits_reg; /* number of bits in region */ 401 int index; /* index first long of region in bitmap */ 402 int offset; /* bit offset region in bitmap[index] */ 403 int nlongs_reg; /* num longs spanned by region in bitmap */ 404 int nbitsinlong; /* num bits of region in each spanned long */ 405 unsigned long mask; /* bitmask for one long of region */ 406 int i; /* scans bitmap by longs */ 407 int ret = 0; /* return value */ | 395 int nbits_reg; 396 int index; 397 int offset; 398 int nlongs_reg; 399 int nbitsinlong; 400 unsigned long mask; 401 int i; 402 int ret = 0; |
408 | 403 |
409 /* 410 * Either nlongs_reg == 1 (for small orders that fit in one long) 411 * or (offset == 0 && mask == ~0UL) (for larger multiword orders.) 412 */ | |
413 nbits_reg = 1 << order; 414 index = pos / BITS_PER_LONG; 415 offset = pos - (index * BITS_PER_LONG); 416 nlongs_reg = BITS_TO_LONGS(nbits_reg); 417 nbitsinlong = min(nbits_reg, BITS_PER_LONG); 418 | 404 nbits_reg = 1 << order; 405 index = pos / BITS_PER_LONG; 406 offset = pos - (index * BITS_PER_LONG); 407 nlongs_reg = BITS_TO_LONGS(nbits_reg); 408 nbitsinlong = min(nbits_reg, BITS_PER_LONG); 409 |
419 /* 420 * Can't do "mask = (1UL << nbitsinlong) - 1", as that 421 * overflows if nbitsinlong == BITS_PER_LONG. 422 */ | |
423 mask = (1UL << (nbitsinlong - 1)); 424 mask += mask - 1; 425 mask <<= offset; 426 427 switch (reg_op) { 428 case REG_OP_ISFREE: 429 for (i = 0; i < nlongs_reg; i++) { 430 if (bitmap[index + i] & mask) 431 goto done; 432 } | 410 mask = (1UL << (nbitsinlong - 1)); 411 mask += mask - 1; 412 mask <<= offset; 413 414 switch (reg_op) { 415 case REG_OP_ISFREE: 416 for (i = 0; i < nlongs_reg; i++) { 417 if (bitmap[index + i] & mask) 418 goto done; 419 } |
433 ret = 1; /* all bits in region free (zero) */ | 420 ret = 1; |
434 break; 435 436 case REG_OP_ALLOC: 437 for (i = 0; i < nlongs_reg; i++) 438 bitmap[index + i] |= mask; 439 break; 440 441 case REG_OP_RELEASE: 442 for (i = 0; i < nlongs_reg; i++) 443 bitmap[index + i] &= ~mask; 444 break; 445 } 446done: 447 return ret; 448} 449 | 421 break; 422 423 case REG_OP_ALLOC: 424 for (i = 0; i < nlongs_reg; i++) 425 bitmap[index + i] |= mask; 426 break; 427 428 case REG_OP_RELEASE: 429 for (i = 0; i < nlongs_reg; i++) 430 bitmap[index + i] &= ~mask; 431 break; 432 } 433done: 434 return ret; 435} 436 |
450/** 451 * bitmap_find_free_region - find a contiguous aligned mem region 452 * @bitmap: array of unsigned longs corresponding to the bitmap 453 * @bits: number of bits in the bitmap 454 * @order: region size (log base 2 of number of bits) to find 455 * 456 * Find a region of free (zero) bits in a @bitmap of @bits bits and 457 * allocate them (set them to one). Only consider regions of length 458 * a power (@order) of two, aligned to that power of two, which 459 * makes the search algorithm much faster. 460 * 461 * Return the bit offset in bitmap of the allocated region, 462 * or -errno on failure. 463 */ | |
464static inline int 465bitmap_find_free_region(unsigned long *bitmap, int bits, int order) 466{ | 437static inline int 438bitmap_find_free_region(unsigned long *bitmap, int bits, int order) 439{ |
467 int pos, end; /* scans bitmap by regions of size order */ | 440 int pos; 441 int end; |
468 469 for (pos = 0 ; (end = pos + (1 << order)) <= bits; pos = end) { 470 if (!__reg_op(bitmap, pos, order, REG_OP_ISFREE)) 471 continue; 472 __reg_op(bitmap, pos, order, REG_OP_ALLOC); 473 return pos; 474 } 475 return -ENOMEM; 476} 477 | 442 443 for (pos = 0 ; (end = pos + (1 << order)) <= bits; pos = end) { 444 if (!__reg_op(bitmap, pos, order, REG_OP_ISFREE)) 445 continue; 446 __reg_op(bitmap, pos, order, REG_OP_ALLOC); 447 return pos; 448 } 449 return -ENOMEM; 450} 451 |
478/** 479 * bitmap_allocate_region - allocate bitmap region 480 * @bitmap: array of unsigned longs corresponding to the bitmap 481 * @pos: beginning of bit region to allocate 482 * @order: region size (log base 2 of number of bits) to allocate 483 * 484 * Allocate (set bits in) a specified region of a bitmap. 485 * 486 * Return 0 on success, or %-EBUSY if specified region wasn't 487 * free (not all bits were zero). 488 */ 489 | |
490static inline int 491bitmap_allocate_region(unsigned long *bitmap, int pos, int order) 492{ 493 if (!__reg_op(bitmap, pos, order, REG_OP_ISFREE)) 494 return -EBUSY; 495 __reg_op(bitmap, pos, order, REG_OP_ALLOC); 496 return 0; 497} 498 | 452static inline int 453bitmap_allocate_region(unsigned long *bitmap, int pos, int order) 454{ 455 if (!__reg_op(bitmap, pos, order, REG_OP_ISFREE)) 456 return -EBUSY; 457 __reg_op(bitmap, pos, order, REG_OP_ALLOC); 458 return 0; 459} 460 |
499/** 500 * bitmap_release_region - release allocated bitmap region 501 * @bitmap: array of unsigned longs corresponding to the bitmap 502 * @pos: beginning of bit region to release 503 * @order: region size (log base 2 of number of bits) to release 504 * 505 * This is the complement to __bitmap_find_free_region() and releases 506 * the found region (by clearing it in the bitmap). 507 * 508 * No return value. 509 */ | |
510static inline void 511bitmap_release_region(unsigned long *bitmap, int pos, int order) 512{ 513 __reg_op(bitmap, pos, order, REG_OP_RELEASE); 514} 515 516 517#define for_each_set_bit(bit, addr, size) \ 518 for ((bit) = find_first_bit((addr), (size)); \ 519 (bit) < (size); \ 520 (bit) = find_next_bit((addr), (size), (bit) + 1)) 521 522#endif /* _LINUX_BITOPS_H_ */ | 461static inline void 462bitmap_release_region(unsigned long *bitmap, int pos, int order) 463{ 464 __reg_op(bitmap, pos, order, REG_OP_RELEASE); 465} 466 467 468#define for_each_set_bit(bit, addr, size) \ 469 for ((bit) = find_first_bit((addr), (size)); \ 470 (bit) < (size); \ 471 (bit) = find_next_bit((addr), (size), (bit) + 1)) 472 473#endif /* _LINUX_BITOPS_H_ */ |