/linux-master/drivers/nvdimm/ |
H A D | badrange.c | 150 u64 new_start = clr_end + 1; local 151 u64 new_len = bre_end - new_start + 1; 154 alloc_and_append_badrange_entry(badrange, new_start,
|
/linux-master/arch/arm/mm/ |
H A D | cache-l2x0.c | 1609 unsigned long new_start, new_end; local 1616 new_start = bcm_l2_phys_addr(start); 1621 l2c210_inv_range(new_start, new_end); 1628 l2c210_inv_range(new_start, 1636 unsigned long new_start, new_end; local 1643 new_start = bcm_l2_phys_addr(start); 1648 l2c210_clean_range(new_start, new_end); 1655 l2c210_clean_range(new_start, 1663 unsigned long new_start, new_end; local 1675 new_start [all...] |
/linux-master/drivers/media/pci/bt8xx/ |
H A D | bttv-vbi.c | 329 __s32 new_start; local 331 new_start = frt->fmt.vbi.start[i] + tvnorm->vbistart[i] 334 frt->fmt.vbi.start[i] = min(new_start, max_end - 1);
|
/linux-master/drivers/iommu/iommufd/ |
H A D | io_pagetable.c | 1173 unsigned long new_start = iova + 1; local 1188 if (new_start & (alignment - 1) || 1189 iopt_area_start_byte(area, new_start) & (alignment - 1)) 1224 (new_start - 1) - start_iova + 1, 1229 rc = iopt_insert_area(iopt, rhs, area->pages, new_start, 1230 iopt_area_start_byte(area, new_start), 1231 last_iova - new_start + 1, area->iommu_prot);
|
/linux-master/drivers/gpu/drm/radeon/ |
H A D | radeon_ttm.c | 138 uint64_t old_start, new_start; local 146 new_start = (u64)new_mem->start << PAGE_SHIFT; 161 new_start += rdev->mc.vram_start; 164 new_start += rdev->mc.gtt_start; 178 fence = radeon_copy(rdev, old_start, new_start, num_pages, bo->base.resv);
|
/linux-master/mm/ |
H A D | readahead.c | 756 * @new_start: The revised start 772 loff_t new_start, size_t new_len) 779 new_index = new_start / PAGE_SIZE; 805 new_len += new_start - readahead_pos(ractl); 771 readahead_expand(struct readahead_control *ractl, loff_t new_start, size_t new_len) argument
|
H A D | z3fold.c | 725 unsigned short new_start = TOTAL_CHUNKS - zhdr->last_chunks - local 727 mchunk_memmove(zhdr, new_start); 728 zhdr->start_middle = new_start;
|
H A D | mmap.c | 1933 unsigned long new_start; local 1948 new_start = (vma->vm_flags & VM_GROWSUP) ? vma->vm_start : 1950 if (is_hugepage_only_range(vma->vm_mm, new_start, size))
|
/linux-master/kernel/bpf/ |
H A D | log.c | 93 u64 new_end, new_start; local 98 new_start = new_end - log->len_total; 100 new_start = log->start_pos; 102 log->start_pos = new_start;
|
/linux-master/fs/bcachefs/ |
H A D | btree_update.c | 194 struct bpos new_start = bkey_start_pos(new.k); local 195 unsigned front_split = bkey_lt(bkey_start_pos(old.k), new_start); 216 bch2_cut_back(new_start, update); 232 bch2_cut_front(new_start, update);
|
/linux-master/drivers/md/dm-vdo/indexer/ |
H A D | delta-index.c | 158 u64 new_start; local 163 new_start = delta_zone->new_offsets[first]; 164 if (delta_list->start != new_start) { 169 delta_list->start = new_start; 184 new_start = delta_zone->new_offsets[middle]; 190 if (new_start > delta_list->start) {
|
/linux-master/fs/ |
H A D | exec.c | 690 unsigned long new_start = old_start - shift; local 692 VMA_ITERATOR(vmi, mm, new_start); 696 BUG_ON(new_start > new_end); 707 * cover the whole range: [new_start, old_end) 709 if (vma_expand(&vmi, vma, new_start, old_end, vma->vm_pgoff, NULL)) 717 vma, new_start, length, false, true)) 743 return vma_shrink(&vmi, vma, new_start, new_end, vma->vm_pgoff);
|
/linux-master/drivers/gpu/drm/ |
H A D | drm_buddy.c | 678 u64 new_start; local 708 new_start = drm_buddy_block_offset(block); 710 err = __alloc_range(mm, &dfs, new_start, new_size, blocks, NULL);
|
/linux-master/fs/orangefs/ |
H A D | inode.c | 248 loff_t new_start = readahead_pos(rac); local 261 readahead_expand(rac, new_start, new_len);
|
/linux-master/fs/ext4/ |
H A D | mballoc.c | 4217 * new_start The start of the range we want to compare 4222 ext4_mb_pa_rb_next_iter(ext4_lblk_t new_start, ext4_lblk_t cur_start, struct rb_node *node) argument 4224 if (new_start < cur_start) 4275 ext4_lblk_t new_start, tmp_pa_start, right_pa_start = -1; local 4278 new_start = *start; 4386 if (left_pa_end > new_start) 4387 new_start = left_pa_end; 4397 ext4_mb_pa_assert_overlap(ac, new_start, new_end); 4399 *start = new_start; 5122 ext4_lblk_t iter_start, new_start; local [all...] |
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_svm.c | 927 uint64_t new_start, uint64_t new_n, uint64_t *new_vram_pages) 938 d = (new_start - old_start) * size; 943 d = (new_start == old_start) ? new_n * size : 0; 1012 * range from new_start up to size new->npages, the remaining old range is from 1130 svm_range_split_head(struct svm_range *prange, uint64_t new_start, argument 1134 int r = svm_range_split(prange, new_start, prange->last, &head); 1138 if (!IS_ALIGNED(new_start, 1UL << prange->granularity)) 925 svm_range_split_array(void *ppnew, void *ppold, size_t size, uint64_t old_start, uint64_t old_n, uint64_t new_start, uint64_t new_n, uint64_t *new_vram_pages) argument
|
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc_submission.c | 1182 __extend_last_switch(struct intel_guc *guc, u64 *prev_start, u32 new_start) argument 1187 if (new_start == lower_32_bits(*prev_start)) 1196 * If neither gt_stamp nor new_start has rolled over, then the 1200 * The below conditions address the cases of new_start rollover and 1203 if (new_start < gt_stamp_last && 1204 (new_start - gt_stamp_last) <= POLL_TIME_CLKS) 1207 if (new_start > gt_stamp_last && 1208 (gt_stamp_last - new_start) <= POLL_TIME_CLKS && gt_stamp_hi) 1211 *prev_start = ((u64)gt_stamp_hi << 32) | new_start;
|
/linux-master/include/linux/ |
H A D | pagemap.h | 1279 loff_t new_start, size_t new_len);
|
/linux-master/drivers/net/ethernet/netronome/nfp/bpf/ |
H A D | jit.c | 4330 s16 new_start = range_start; local 4336 new_start = off; 4349 if (new_end - new_start <= 64) { 4351 range_start = new_start;
|
/linux-master/fs/ocfs2/ |
H A D | alloc.c | 5827 unsigned int new_start) 5840 return current_tail == new_start; 5826 ocfs2_truncate_log_can_coalesce(struct ocfs2_truncate_log *tl, unsigned int new_start) argument
|