/macosx-10.10/ntfs-84/kext/ |
H A D | ntfs_runlist.c | 206 if ((dst->vcn + dst->length) != src->vcn) 333 d_rl[marker].vcn = d_rl[marker - 1].vcn + 383 if (src[0].vcn > 0) 392 if (src[0].vcn > d_rl[loc - 1].vcn + merged_length) 422 d_rl[marker].vcn = d_rl[marker - 1].vcn + d_rl[marker - 1].length; 425 d_rl[marker].length = d_rl[marker + 1].vcn 869 VCN vcn; /* Current vcn. */ local 1120 ntfs_rl_vcn_to_lcn(const ntfs_rl_element *rl, const VCN vcn, s64 *clusters) argument 1176 ntfs_rl_find_vcn_nolock(ntfs_rl_element *rl, const VCN vcn) argument 2247 VCN vcn; local 2383 VCN vcn; local [all...] |
H A D | ntfs_secure.c | 209 VCN vcn, old_vcn; local 328 /* Get the starting vcn of the index_block holding the child node. */ 329 vcn = sle64_to_cpup((sle64*)((u8*)ie + le16_to_cpu(ie->length) - 8)); 340 * Convert vcn to byte offset in the index allocation attribute and map 343 err = ntfs_page_map(ni, (vcn << ni->vcn_size_shift) & 352 ia = (INDEX_ALLOCATION*)(kaddr + ((vcn << ni->vcn_size_shift) & 362 ntfs_error(vol->mp, "Index record with vcn 0x%llx is " 364 (unsigned long long)vcn); 367 if (sle64_to_cpu(ia->index_block_vcn) != vcn) { 373 (unsigned long long)vcn); [all...] |
H A D | ntfs_runlist.h | 50 * ntfs_rl_element - in memory vcn to lcn mapping array element 51 * @vcn: starting vcn of the current array element 55 * The last vcn (in fact the last vcn + 1) is reached when length == 0. 57 * When lcn == -1 this means that the count vcns starting at vcn are not 60 typedef struct { /* In memory vcn to lcn mapping structure element. */ 61 VCN vcn; /* vcn = Starting virtual cluster number. */ member in struct:__anon11432 67 * ntfs_runlist - in memory vcn t [all...] |
H A D | ntfs_compress.c | 103 VCN start_vcn, vcn, end_vcn; local 116 vcn = start_vcn = ofs >> ni->vol->cluster_size_shift; 124 /* Seek to element containing target vcn. */ 125 while (rl->length && rl[1].vcn <= vcn) 127 lcn = ntfs_rl_vcn_to_lcn(rl, vcn, &clusters); 130 * starting vcn the entire compression block must be sparse. 136 if (vcn == start_vcn) 151 if (vcn + clusters >= end_vcn) { 155 vcn [all...] |
H A D | ntfs_attr.c | 88 VCN vcn, end_vcn; local 118 vcn = 0; 121 err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, vcn, 150 * vcn we looked up and that the highest_vcn of the attribute 151 * is above the current vcn. 153 if (sle64_to_cpu(a->lowest_vcn) != vcn || (vcn && 154 sle64_to_cpu(a->highest_vcn) < vcn)) 156 /* Determine the next vcn. */ 157 vcn 226 ntfs_map_runlist_nolock(ntfs_inode *ni, VCN vcn, ntfs_attr_search_ctx *ctx) argument 435 ntfs_attr_vcn_to_lcn_nolock(ntfs_inode *ni, const VCN vcn, const BOOL write_locked, s64 *clusters) argument 580 ntfs_attr_find_vcn_nolock(ntfs_inode *ni, const VCN vcn, ntfs_rl_element **run, ntfs_attr_search_ctx *ctx) argument 4073 VCN vcn, end_vcn; local 5947 VCN vcn, lowest_vcn, stop_vcn; local [all...] |
H A D | ntfs_attr.h | 61 __private_extern__ errno_t ntfs_map_runlist_nolock(ntfs_inode *ni, VCN vcn, 65 const VCN vcn, const BOOL write_locked, s64 *clusters); 68 const VCN vcn, ntfs_rl_element **run,
|
H A D | ntfs_lcnalloc.c | 69 * @start_vcn: vcn to use for the first allocated cluster 82 * @start_vcn specifies the vcn of the first allocated cluster. This makes 425 rl[rlpos].vcn = rl[rlpos - 1].vcn + 430 rl[rlpos].vcn = start_vcn; 768 rl[rlpos].vcn = rl[rlpos - 1].vcn + rl[rlpos - 1].length; 870 * @start_vcn: vcn in the runlist @rl at which to start freeing clusters 929 delta = start_vcn - rl->vcn; 960 "%lld at vcn 1200 VCN vcn; local [all...] |
H A D | ntfs_index.c | 501 * key plus the index data plus optionally the sub-node vcn pointer. 594 VCN vcn; local 628 /* Get the starting vcn of the child index block to descend into. */ 629 vcn = sle64_to_cpup((sle64*)((u8*)ictx->entry + 631 if (vcn < 0) { 637 ofs = (vcn << idx_ni->vcn_size_shift) & ~PAGE_MASK_64; 659 * If @vcn is in the same VM page as the existing page we reuse 684 ia = (INDEX_ALLOCATION*)(addr + ((unsigned)(vcn << 696 "chkdsk.", (unsigned long long)vcn, 700 if (sle64_to_cpu(ia->index_block_vcn) != vcn) { 1907 ntfs_index_block_lay_out(ntfs_inode *idx_ni, VCN vcn, INDEX_BLOCK *ia) argument 2312 VCN vcn; local 3593 VCN vcn = 0; local 4399 VCN vcn; member in union:__anon11313 5255 leVCN vcn; local [all...] |
H A D | ntfs_dir.c | 118 VCN vcn, old_vcn; local 355 /* Get the starting vcn of the index block holding the child node. */ 356 vcn = sle64_to_cpup((sle64*)((u8*)ie + le16_to_cpu(ie->length) - 8)); 367 * Convert vcn to byte offset in the index allocation attribute and map 370 err = ntfs_page_map(ia_ni, (vcn << ia_ni->vcn_size_shift) & 379 ia = (INDEX_ALLOCATION*)(kaddr + ((vcn << ia_ni->vcn_size_shift) & 392 (unsigned long long)vcn, 396 if (sle64_to_cpu(ia->index_block_vcn) != vcn) { 402 (unsigned long long)vcn, 413 (unsigned long long)vcn, [all...] |
H A D | ntfs_mft.c | 806 VCN vcn, lowest_vcn = 0; local 835 vcn = (allocated_size - 1) >> vol->cluster_size_shift; 836 err = ntfs_attr_find_vcn_nolock(mftbmp_ni, vcn, &rl, NULL); 838 rl[1].vcn != vcn + 1) { 888 rl[1].vcn++; 900 err = ntfs_cluster_alloc(vol, vcn + 1, 1, lcn, DATA_ZONE, 945 mftbmp_ni->name_len, vcn, NULL, 0, ctx); 1005 a->highest_vcn = cpu_to_sle64(vcn + 1); 1042 mftbmp_ni->name_len, vcn, NUL 1273 VCN vcn, lowest_vcn = 0; local [all...] |
H A D | ntfs_debug.c | 276 (unsigned long long)rl[u].vcn, 282 (unsigned long long)rl[u].vcn,
|
H A D | ntfs_index.h | 162 VCN vcn; /* The vcn of this index block. */ member in struct:_ntfs_index_context::__anon11316::__anon11317
|
H A D | ntfs_vnops.c | 10243 VCN vcn; local 10373 * All is ok, do the mapping. First, work out the vcn and vcn offset 10376 vcn = byte_offset >> vol->cluster_size_shift; 10379 * Convert the vcn to the corresponding lcn and obtain the number of 10380 * contiguous clusters starting at the vcn. 10383 lcn = ntfs_attr_vcn_to_lcn_nolock(ni, vcn, FALSE, 10466 /* The vcn was mapped successfully to a physical lcn, return it. */
|
/macosx-10.10/ntfs-84/util/ |
H A D | ntfs.util.c | 79 typedef struct { /* In memory vcn to lcn mapping structure element. */ 80 VCN vcn; /* vcn = Starting virtual cluster number. */ member in struct:__anon11447 410 VCN vcn; /* Current vcn. */ local 425 /* Start at vcn = lowest_vcn and lcn 0. */ 426 vcn = sle64_to_cpu(a->lowest_vcn); 434 if (!vcn && !*buf) 444 if (vcn) { 445 rl->vcn 615 ntfs_rl_vcn_to_lcn(const ntfs_rl_element *rl, const VCN vcn, s64 *clusters) argument 715 VCN vcn; local [all...] |
/macosx-10.10/lsof-53/lsof/lib/ |
H A D | rnch.c | 246 static int vcn = 0; local 254 for (i = 0; i < vcn; i++) { 297 if (vcn >= vca) { 310 vc[vcn++] = va;
|