/macosx-10.5.8/xnu-1228.15.4/bsd/sys/ |
H A D | quota.h | 217 #define dqhash1(id, shift, mask) \ 218 ((((id) * 2654435761UL) >> (shift)) & (mask)) 229 * Compute the hash shift value. 237 int shift; local 239 for (shift = 32; size > 1; size >>= 1, --shift) 241 return (shift); 262 int qf_shift; /* primary hash shift */
|
/macosx-10.5.8/xnu-1228.15.4/libkern/gen/ |
H A D | OSAtomicOperations.c | 138 int shift = (UInt32) *(((UInt8 *) &shiftValues) + alignment); local 143 mask <<= shift; local 146 oldValue = (oldValue & ~mask) | (oldValue8 << shift); 147 newValue = (oldValue & ~mask) | (newValue8 << shift); 242 UInt32 shift = (UInt32) *(((UInt8 *) &shiftValues) + alignment); local 247 mask <<= shift; local 250 oldValue = (oldValue & ~mask) | (oldValue16 << shift); 251 newValue = (oldValue & ~mask) | (newValue16 << shift);
|
/macosx-10.5.8/xnu-1228.15.4/osfmk/i386/ |
H A D | rtclock.h | 48 uint32_t shift; /* tsc -> nanosec shift/div */ member in struct:rtc_nanotime 49 /* shift is overloaded with 64 uint32_t shift,
|
H A D | rtclock.c | 99 * based on the scale factor and an implicit 32 bit shift. 169 * with this ratio expressed as a 32-bit scale and shift 173 * The tuple {tsc_base, ns_base, scale, shift} is exported in the commpage 182 commpage_set_nanotime(rntp->tsc_base, rntp->ns_base, rntp->scale, rntp->shift); 195 _rtc_nanotime_store(tsc, base, rntp->scale, rntp->shift, rntp); 267 _rtc_nanotime_store(tsc_base, base, rntp->scale, rntp->shift, rntp); 365 rntp->shift = cycles; 367 rntp->shift = 32;
|
H A D | genassym.c | 518 offsetof(rtc_nanotime_t *, shift));
|
/macosx-10.5.8/xnu-1228.15.4/osfmk/i386/commpage/ |
H A D | commpage.h | 97 extern void commpage_set_nanotime(uint64_t tsc_base, uint64_t ns_base, uint32_t scale, uint32_t shift);
|
H A D | commpage.c | 474 uint32_t shift ) 488 if ((shift != 32) && ((_cpu_capabilities & kSlow)==0) ) 507 p32->nt_shift = shift; 508 p64->nt_shift = shift;
|
/macosx-10.5.8/xnu-1228.15.4/osfmk/ddb/ |
H A D | makedis.c | 678 $[...] bitsplice. The mask covers the bitfield and the shift says how 682 int shift; member in struct:bits 742 bp->shift = i; 1117 it occurs in the bitsplice. `shift' indicates how much to shift left 1121 the bitsplicebits for y will have shift = 2 and value.arg pointing to y, 1122 and those for x will have shift = -2 and value.mask = binary 1100. 1141 int shift; member in struct:bitsplicebits 1155 int shift, nfrombits, ntobits; local 1167 shift [all...] |
/macosx-10.5.8/xnu-1228.15.4/iokit/Kernel/ |
H A D | IOLib.cpp | 771 register int shift; local 774 for (shift = 1; shift < intsize; shift++) { 776 return (IOAlignment)(intsize - shift);
|
/macosx-10.5.8/xnu-1228.15.4/osfmk/ppc/ |
H A D | AltiAssist.s | 80 vspltisw v1,8 ; Get half of the shift
|
H A D | savearea_asm.s | 1062 ori r0,r0,0x2200 ; Finish shift table 1169 ori r0,r0,0x2200 ; Finish shift table
|
H A D | lowmem_vectors.s | 916 addi r24,r22,16 ; Get shift to move cpu mask to syscall mask 1053 addi r24,r22,16 ; Get shift to move cpu mask to syscall mask 1460 addi r22,r22,10 ; Adjust code so we shift into CR5 1793 addi r22,r22,10 ; Adjust code so we shift into CR5
|
H A D | machine_routines_asm.s | 1682 rlwinm. r0,r0,pfSCOMFixUpb+1,31,31 ; Set shift if we need a fix me up 2254 rlwinm r0,r0,pfSCOMFixUpb+1,31,31 ; Set shift if we need a fix me up
|
H A D | hw_vm.s | 286 rlwinm r10,r30,18,14,17 ; Shift EA[32:35] down to correct spot in VSID (actually shift up 14) 903 hrmBlock32: lis r29,0xD000 ; Get shift to 32MB bsu 1114 // **** Need to adjust above shift based on the page size - large pages need to shift a bit more 1239 hrmBlock64: lis r29,0xD000 ; Get shift to 32MB bsu 4086 rlwinm r2,r2,0,27,31 ; Wrap shift so we do not shift cache entries 8-F out 4089 srw r0,r0,r2 ; Slide sub-tag mask to right slot (shift work for either half) 4090 srw r5,r30,r2 ; Slide sub-tag to right slot (shift work for either half) 4185 subi r4,r17,1 ; Adjust shift t [all...] |
H A D | cswtch.s | 1973 slwi r7,r8,2 ; shift groups-of-2 over by 2 1977 slwi r7,r8,4 ; shift groups-of-4 over by 4
|
H A D | Firmware.s | 1844 li r9,32 /* Get the initial shift calc */ 1851 sub r9,r9,r26 /* Get right shift justifier for pixel size */ 2440 rlwinm. r0,r0,pfSCOMFixUpb+1,31,31 ; Set shift if we need a fix me up
|
/macosx-10.5.8/xnu-1228.15.4/osfmk/console/ |
H A D | panic_dialog.c | 761 int shift = 4; local 765 runlen |= ((dataPtr[i++] & 0x7F) << shift); 766 shift+=7;
|
/macosx-10.5.8/xnu-1228.15.4/bsd/dev/ppc/ |
H A D | xsumas.s | 172 slwi r4,r4,8 ; shift last byte into proper lane 233 slwi r7,r7,8 ; shift last byte into proper lane 380 slwi r4,r4,8 ; shift last byte into proper lane
|
/macosx-10.5.8/xnu-1228.15.4/osfmk/console/panic_ui/ |
H A D | genimage.c | 1054 int shift = 4; local 1059 runlen |= ((dataPtr[i++] & 0x7F) << shift); 1060 shift+=7;
|
/macosx-10.5.8/xnu-1228.15.4/bsd/vfs/ |
H A D | vfs_subr.c | 3101 int shift; local 3104 * Work out how far we have to shift the block count down to make it fit. 3105 * Note that it's possible to have to shift so far that the resulting 3112 for (shift = 0; shift < 32; shift++) { 3113 if ((sp->f_blocks >> shift) <= LONG_MAX) 3115 if ((sp->f_bsize << (shift + 1)) > LONG_MAX) 3119 sfs.f_blocks = (long)__SHIFT_OR_CLIP(sp->f_blocks, shift); 3120 sfs.f_bfree = (long)__SHIFT_OR_CLIP(sp->f_bfree, shift); [all...] |
H A D | vfs_syscalls.c | 6902 int shift; local 6905 * Work out how far we have to shift the block count down to make it fit. 6906 * Note that it's possible to have to shift so far that the resulting 6913 for (shift = 0; shift < 32; shift++) { 6914 if ((sfsp->f_blocks >> shift) <= LONG_MAX) 6916 if ((sfsp->f_bsize << (shift + 1)) > LONG_MAX) 6920 sfs.f_blocks = (long)__SHIFT_OR_CLIP(sfsp->f_blocks, shift); 6921 sfs.f_bfree = (long)__SHIFT_OR_CLIP(sfsp->f_bfree, shift); [all...] |
/macosx-10.5.8/xnu-1228.15.4/bsd/dev/i386/ |
H A D | fasttrap_isa.c | 219 int shift = function_entry ? 1 : 0; local 245 stack = regs64->isf.rsp + sizeof(uint64_t) * (argno - 6 + shift); 252 value = dtrace_fuword32((user_addr_t)(unsigned long)&stack[argno + shift]);
|
/macosx-10.5.8/xnu-1228.15.4/osfmk/kern/ |
H A D | sched_prim.c | 278 uint32_t shift; local 313 for (shift = 0; abstime > BASEPRI_DEFAULT; ++shift) 315 sched_fixed_shift = shift;
|
/macosx-10.5.8/xnu-1228.15.4/osfmk/default_pager/ |
H A D | dp_backing_store.c | 107 * 0 means no shift to pages, so == 1 page/cluster. 1 would mean 466 * Set up default page shift, but only if not already 502 * Keep cluster size in bit shift because it's quicker 1055 ("device=0x%x,offset=0x%x,count=0x%x,record_size=0x%x,shift=%d,total_size=0x%x\n", 1312 unsigned int shift, 1336 ASSERT(ps->ps_clshift >= shift); 1405 (ps->ps_clshift >= shift)) { 1465 * Must pass cluster shift to find the most appropriate segment. 3797 ("device=0x%x,offset=0x%x,count=0x%x,record_size=0x%x,shift=%d,total_size=0x%x\n", 1311 ps_select_segment( unsigned int shift, int *psindex) argument
|
/macosx-10.5.8/xnu-1228.15.4/bsd/dev/dtrace/ |
H A D | dtrace.c | 4966 int shift = (sizeof (uintptr_t) * NBBY) - 4, i = 0; 5004 while (shift >= 0) { 5005 mask = (uintptr_t)0xf << shift; 5007 if (val >= ((uintptr_t)1 << shift)) 5008 c[i++] = "0123456789abcdef"[(val & mask) >> shift]; 5009 shift -= 4;
|