/linux-master/drivers/block/ |
H A D | ublk_drv.c | 992 * FLUSH, DISCARD or WRITE_ZEROES usually won't return bytes returned, so end them 1304 unsigned long pfn, end, phys_off = vma->vm_pgoff << PAGE_SHIFT; local 1320 end = UBLKSRV_CMD_BUF_OFFSET + ub->dev_info.nr_hw_queues * max_sz; 1321 if (phys_off < UBLKSRV_CMD_BUF_OFFSET || phys_off >= end)
|
/linux-master/arch/xtensa/include/asm/ |
H A D | cacheflush.h | 100 void flush_icache_range(unsigned long start, unsigned long end); 123 #define flush_cache_vmap(start,end) flush_cache_all() 124 #define flush_cache_vmap_early(start,end) do { } while (0) 125 #define flush_cache_vunmap(start,end) flush_cache_all() 137 unsigned long start, unsigned long end); 150 #define local_flush_icache_range(start, end) \ 152 __flush_dcache_range(start, (end) - (start)); \ 153 __invalidate_icache_range(start,(end) - (start)); \
|
/linux-master/arch/s390/mm/ |
H A D | gmap.c | 777 * @end: end virtual address in the guest address space 780 unsigned long end) 785 nb->notifier_call(gmap, start, end); 2209 unsigned long start, end, bits, raddr; local 2220 end = start + ((sg->orig_asce & _ASCE_TABLE_LENGTH) + 1) * PAGE_SIZE; 2222 gaddr < end) { 2522 unsigned long end, struct mm_walk *walk) 2564 unsigned long end, struct mm_walk *walk) 2568 for (addr = start; addr != end; add 779 gmap_call_notifier(struct gmap *gmap, unsigned long start, unsigned long end) argument 2521 thp_split_walk_pmd_entry(pmd_t *pmd, unsigned long addr, unsigned long end, struct mm_walk *walk) argument 2563 __zap_zero_pages(pmd_t *pmd, unsigned long start, unsigned long end, struct mm_walk *walk) argument 2650 unsigned long start, end; local 2785 __s390_uv_destroy_range(struct mm_struct *mm, unsigned long start, unsigned long end, bool interruptible) argument [all...] |
/linux-master/arch/arm64/net/ |
H A D | bpf_jit_comp.c | 1515 * - offset[0] offset of the end of prologue, 1517 * - offset[1] - offset of the end of 1st instruction, 1520 * - offset[3] - offset of the end of 3rd instruction, 1542 * instruction (end of program) 1574 static inline void bpf_flush_icache(void *start, void *end) argument 1576 flush_icache_range((unsigned long)start, (unsigned long)end); 2281 * dummy_tramp is placed at the end: 2360 /* plt locates at the end of bpf prog */
|
/linux-master/kernel/ |
H A D | workqueue.c | 1152 * Linked worklist will always end before the end of the list, 1932 * $pwq still has inactive work items, rotate it to the end of the 3557 * Leave this pool. Notify regular workers; otherwise, we end up 3587 unsigned long end = jiffies + BH_WORKER_JIFFIES; local 3610 --nr_restarts && time_before(jiffies, end)); 4320 * smp_mb() at the end of set_work_pool_and_clear_pending() is paired 6437 * An unbound pool may end up with a cpumask which doesn't have any online
|
H A D | cpu.c | 312 ktime_t now, end, start = ktime_get(); local 315 end = start + 10ULL * NSEC_PER_SEC; 326 if (now > end) { 1077 goto end; 1105 end: 2304 enum cpuhp_state i, end; local 2310 end = CPUHP_AP_ONLINE_DYN_END; 2314 end = CPUHP_BP_PREPARE_DYN_END; 2320 for (i = state; i <= end; i++, step++) {
|
/linux-master/fs/nfsd/ |
H A D | nfs4xdr.c | 138 * buffer might end on a page boundary. 2603 char *str, *end, *next; local 2613 end = str = components; 2614 while (*end) { 2619 for (; *end && (*end != esc_exit); end++) 2620 /* find esc_exit or end of string */; 2621 next = end + 1; 2622 if (*end [all...] |
/linux-master/fs/nfs/ |
H A D | inode.c | 440 * This is our front-end to iget that looks up inodes by file handle 634 /* Optimization: if the end result is no change, don't RPC */ 1820 u64 start, u64 end) 1840 if (end == nfsi->ooo->gap[i].start) 1841 end = nfsi->ooo->gap[i].end; 1842 else if (start == nfsi->ooo->gap[i].end) 1851 if (start != end) { 1859 nfsi->ooo->gap[cnt].end = end; 1819 nfs_ooo_merge(struct nfs_inode *nfsi, u64 start, u64 end) argument [all...] |
/linux-master/fs/bcachefs/ |
H A D | inode.c | 38 static int inode_decode_field(const u8 *in, const u8 *end, argument 45 if (in >= end) 58 if (in + bytes > end) 77 u8 *end = (void *) &packed[1]; local 116 BUG_ON(out > end); 157 const u8 *end = bkey_val_end(inode); local 170 ret = inode_decode_field(in, end, field, &field_bits); \ 188 const u8 *in, const u8 *end, 197 ret = bch2_varint_decode_fast(in, end, &v[0]); \ 203 ret = bch2_varint_decode_fast(in, end, 187 bch2_inode_unpack_v2(struct bch_inode_unpacked *unpacked, const u8 *in, const u8 *end, unsigned nr_fields) argument 231 const u8 *end = bkey_val_end(inode); local 818 struct bpos end = POS(inum.inum, U64_MAX); local [all...] |
H A D | buckets.c | 512 * Backup superblock might be past the end of our normal usable space: 1040 unsigned new_ptrs_bytes = (void *) new_ptrs.end - (void *) new_ptrs.start; 1041 unsigned old_ptrs_bytes = (void *) old_ptrs.end - (void *) old_ptrs.start; 1131 * Backup superblock might be past the end of our normal usable space: 1175 u64 start, u64 end, 1182 min_t(u64, bucket_to_sector(ca, b + 1), end) - start; 1196 } while (start < end); 1173 bch2_trans_mark_metadata_sectors(struct btree_trans *trans, struct bch_dev *ca, u64 start, u64 end, enum bch_data_type type, u64 *bucket, unsigned *bucket_sectors) argument
|
H A D | btree_node_scan.c | 270 static void bubble_up(struct found_btree_node *n, struct found_btree_node *end) argument 272 while (n + 1 < end && 281 struct found_btree_node *end) 286 n < end && 300 bubble_up(n, end); 279 handle_overwrites(struct bch_fs *c, struct found_btree_node *start, struct found_btree_node *end) argument
|
/linux-master/drivers/scsi/ |
H A D | sd.c | 29 * - end of transfer (bh + scsi_lib): SCSI_LOG_HLCOMPLETE level 1 1238 scmd_printk(KERN_ERR, cmd, "access beyond end of device\n"); 1319 * This indicates that the command is ready from our end to be queued. 2200 * In case of bogus fw or device, we could end up having 3122 const struct scsi_io_group_descriptor *desc, *start, *end; local 3134 end = (void *)buffer + ALIGN_DOWN(data.header_length + data.length, 3135 sizeof(*end)); 3141 for (desc = start; desc < end; desc++) 3736 * on @disk reach beyond the end of the device. If the SCSI host 3777 char *end local [all...] |
/linux-master/kernel/sched/ |
H A D | fair.c | 394 * at the end of the enqueue. 688 * and end up with a larger lag than we started with. 1277 * Mark the end of the wait period if dequeueing a 2192 * end try selecting ourselves (current == env->p) as a swap candidate. 3213 unsigned long start, end; local 3379 end = ALIGN(start + (pages << PAGE_SHIFT), HPAGE_SIZE); 3380 end = min(end, vma->vm_end); 3381 nr_pte_updates = change_prot_numa(vma, start, end); 3392 pages -= (end [all...] |
/linux-master/drivers/irqchip/ |
H A D | irq-gic-v3-its.c | 2232 phys_addr_t start, end, addr_end; local 2245 for_each_reserved_mem_range(i, &start, &end) { 2246 if (addr >= start && addr_end <= end) 3917 * easily. So in the end, vpe->pending_last is only an 5488 const unsigned long end) 5494 const unsigned long end) 5564 const unsigned long end) 5575 res.end = its_entry->base_address + ACPI_GICV3_ITS_MEM_SIZE - 1; 5612 const unsigned long end) 5620 .end 5487 gic_acpi_match_srat_its(union acpi_subtable_headers *header, const unsigned long end) argument 5493 gic_acpi_parse_srat_its(union acpi_subtable_headers *header, const unsigned long end) argument 5563 gic_acpi_parse_madt_its(union acpi_subtable_headers *header, const unsigned long end) argument 5611 its_acpi_reset(union acpi_subtable_headers *header, const unsigned long end) argument [all...] |
/linux-master/drivers/i2c/ |
H A D | i2c-core-base.c | 1230 char *blank, end; local 1247 res = sscanf(++blank, "%hi%c", &info.addr, &end); 1252 if (res > 1 && end != '\n') { 1298 char end; local 1302 res = sscanf(buf, "%hi%c", &addr, &end); 1307 if (res > 1 && end != '\n') {
|
/linux-master/arch/riscv/mm/ |
H A D | init.c | 217 * Reserve from the start of the kernel to the end of the kernel 241 * if end of dram is equal to maximum addressable memory. For 64-bit 242 * kernel, this problem can't happen here as the end of the virtual 263 * in the device tree, otherwise the allocation could end up in a 1107 * kernel, whereas for 64-bit kernel, the end of the virtual address 1204 pr_warn("fixmap btmap start [%08lx] != end [%08lx]\n", 1220 phys_addr_t end, 1226 for (pa = start; pa < end; pa += map_size) { 1229 best_map_size(pa, va, end - pa); 1238 phys_addr_t start, end; local 1219 create_linear_mapping_range(phys_addr_t start, phys_addr_t end, uintptr_t fixed_map_size) argument 1414 vmemmap_populate(unsigned long start, unsigned long end, int node, struct vmem_altmap *altmap) argument 1434 preallocate_pgd_pages_range(unsigned long start, unsigned long end, const char *area) argument [all...] |
/linux-master/arch/riscv/errata/thead/ |
H A D | errata.c | 160 void thead_errata_patch_func(struct alt_entry *begin, struct alt_entry *end, argument 169 for (alt = begin; alt < end; alt++) {
|
/linux-master/mm/ |
H A D | hugetlb.c | 100 unsigned long start, unsigned long end); 881 /* New entry for end of split region */ 912 } else { /* Trim end of region */ 1465 * next node from which to allocate, handling wrap at end of node 2533 * if we would end up overcommiting the surpluses. Abuse 3444 unsigned long end, void *arg) 3448 for (nid = start; nid < end; nid++) 3526 static void __init hugetlb_pages_alloc_boot_node(unsigned long start, unsigned long end, void *arg) argument 3529 int i, num = end - start; 5216 unsigned long reserve, start, end; local 3443 gather_bootmem_prealloc_parallel(unsigned long start, unsigned long end, void *arg) argument 5650 __unmap_hugepage_range(struct mmu_gather *tlb, struct vm_area_struct *vma, unsigned long start, unsigned long end, struct page *ref_page, zap_flags_t zap_flags) argument 5806 __hugetlb_zap_begin(struct vm_area_struct *vma, unsigned long *start, unsigned long *end) argument 5845 unmap_hugepage_range(struct vm_area_struct *vma, unsigned long start, unsigned long end, struct page *ref_page, zap_flags_t zap_flags) argument 6947 hugetlb_change_protection(struct vm_area_struct *vma, unsigned long address, unsigned long end, pgprot_t newprot, unsigned long cp_flags) argument 7255 hugetlb_unreserve_pages(struct inode *inode, long start, long end, long freed) argument 7328 unsigned long end = start + PUD_SIZE; local 7351 adjust_range_if_pmd_sharing_possible(struct vm_area_struct *vma, unsigned long *start, unsigned long *end) argument 7472 adjust_range_if_pmd_sharing_possible(struct vm_area_struct *vma, unsigned long *start, unsigned long *end) argument 7689 hugetlb_unshare_pmds(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument [all...] |
/linux-master/include/linux/ |
H A D | mm.h | 349 * These VMAs will get a single end guard page. This helps userspace protect 587 unsigned long end, unsigned long newflags); 700 rcu_read_lock(); /* keeps vma alive till the end of up_read */ 1017 unsigned long start, unsigned long end, gfp_t gfp) 1019 __mas_set_range(&vmi->mas, start, end - 1); 1058 /* The MM code likes to work with exclusive end addresses */ 2388 unsigned long end, unsigned long tree_end, bool mm_wr_locked); 2393 unsigned long end, unsigned long floor, unsigned long ceiling); 2408 void truncate_pagecache_range(struct inode *inode, loff_t offset, loff_t end); 2573 unsigned long end, unsigne 1016 vma_iter_clear_gfp(struct vma_iterator *vmi, unsigned long start, unsigned long end, gfp_t gfp) argument 3299 vma_modify_flags(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long new_flags) argument 3312 vma_modify_flags_name(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long new_flags, struct anon_vma_name *new_name) argument 3326 vma_modify_policy(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, struct mempolicy *new_pol) argument 3338 vma_modify_flags_uffd(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long new_flags, struct vm_userfaultfd_ctx new_ctx) argument 3535 range_in_vma(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument 4188 range_contains_unaccepted_memory(phys_addr_t start, phys_addr_t end) argument 4194 accept_memory(phys_addr_t start, phys_addr_t end) argument [all...] |
/linux-master/fs/ntfs3/ |
H A D | super.c | 1667 u64 lbo, bytes, start, end; local 1685 /* Align down 'end' on discard_granularity. */ 1686 end = (lbo + bytes) & sbi->discard_granularity_mask_inv; 1689 if (start >= end) 1692 err = blkdev_issue_discard(sb->s_bdev, start >> 9, (end - start) >> 9,
|
H A D | inode.c | 785 loff_t end; local 802 end = vbo + ret; 804 end = vbo + iter_count; 810 if (end > valid && !S_ISBLK(inode->i_mode)) { 811 ni->i_valid = end; 814 } else if (vbo < valid && valid < end) { 816 iov_iter_revert(iter, end - valid); 817 iov_iter_zero(end - valid, iter);
|
H A D | file.c | 290 CLST end = bytes_to_cluster(sbi, to); local 293 for (; vcn < end; vcn += len) { 324 loff_t end = pos + count; local 328 if (end <= inode->i_size && !extend_init) 334 if (end > inode->i_size) { 335 err = ntfs_set_size(inode, end); 354 err = filemap_fdatawrite_range(mapping, pos, end - 1); 362 err = filemap_fdatawait_range(mapping, pos, end - 1); 439 loff_t end = vbo + len; local 474 new_size = max(end, i_siz [all...] |
H A D | dir.c | 182 const u8 *end; local 195 for (ret = 0, end = name + name_len; name < end; ret++, name += slen) { 199 slen = nls->char2uni(name, end - name, uname + ret); 351 u32 end = le32_to_cpu(hdr->used); local 355 if (off + sizeof(struct NTFS_DE) > end) 360 if (e_size < sizeof(struct NTFS_DE) || off + e_size > end) 517 u32 e_size, off, end; local 532 end = le32_to_cpu(hdr->used); 535 for (; off + sizeof(struct NTFS_DE) <= end; of [all...] |
/linux-master/fs/netfs/ |
H A D | buffered_write.c | 898 unsigned long long end) 909 _enter(",%lx,%llx-%llx,%u", folio->index, start, end, caching); 928 * end of the range. 948 max_len = min_t(unsigned long long, max_len, end - start + 1); 1003 unsigned long long end) 1012 _enter("%llx,%llx,", start, end); 1019 folio = xas_find_marked(xas, end / PAGE_SIZE, PAGECACHE_TAG_DIRTY); 1102 folio, start, end); 1117 unsigned long long end) 1125 _start, end); 892 netfs_write_back_from_locked_folio(struct address_space *mapping, struct writeback_control *wbc, struct netfs_group *group, struct xa_state *xas, struct folio *folio, unsigned long long start, unsigned long long end) argument 998 netfs_writepages_begin(struct address_space *mapping, struct writeback_control *wbc, struct netfs_group *group, struct xa_state *xas, unsigned long long *_start, unsigned long long end) argument 1113 netfs_writepages_region(struct address_space *mapping, struct writeback_control *wbc, struct netfs_group *group, unsigned long long *_start, unsigned long long end) argument 1140 loff_t start, end; local [all...] |
/linux-master/drivers/video/fbdev/core/ |
H A D | fb_defio.c | 122 int fb_deferred_io_fsync(struct file *file, loff_t start, loff_t end, int datasync) argument 126 int err = file_write_and_wait_range(file, start, end);
|