Searched refs:end (Results 76 - 100 of 3740) sorted by last modified time

1234567891011>>

/linux-master/drivers/block/
H A Dublk_drv.c992 * FLUSH, DISCARD or WRITE_ZEROES usually won't return bytes returned, so end them
1304 unsigned long pfn, end, phys_off = vma->vm_pgoff << PAGE_SHIFT; local
1320 end = UBLKSRV_CMD_BUF_OFFSET + ub->dev_info.nr_hw_queues * max_sz;
1321 if (phys_off < UBLKSRV_CMD_BUF_OFFSET || phys_off >= end)
/linux-master/arch/xtensa/include/asm/
H A Dcacheflush.h100 void flush_icache_range(unsigned long start, unsigned long end);
123 #define flush_cache_vmap(start,end) flush_cache_all()
124 #define flush_cache_vmap_early(start,end) do { } while (0)
125 #define flush_cache_vunmap(start,end) flush_cache_all()
137 unsigned long start, unsigned long end);
150 #define local_flush_icache_range(start, end) \
152 __flush_dcache_range(start, (end) - (start)); \
153 __invalidate_icache_range(start,(end) - (start)); \
/linux-master/arch/s390/mm/
H A Dgmap.c777 * @end: end virtual address in the guest address space
780 unsigned long end)
785 nb->notifier_call(gmap, start, end);
2209 unsigned long start, end, bits, raddr; local
2220 end = start + ((sg->orig_asce & _ASCE_TABLE_LENGTH) + 1) * PAGE_SIZE;
2222 gaddr < end) {
2522 unsigned long end, struct mm_walk *walk)
2564 unsigned long end, struct mm_walk *walk)
2568 for (addr = start; addr != end; add
779 gmap_call_notifier(struct gmap *gmap, unsigned long start, unsigned long end) argument
2521 thp_split_walk_pmd_entry(pmd_t *pmd, unsigned long addr, unsigned long end, struct mm_walk *walk) argument
2563 __zap_zero_pages(pmd_t *pmd, unsigned long start, unsigned long end, struct mm_walk *walk) argument
2650 unsigned long start, end; local
2785 __s390_uv_destroy_range(struct mm_struct *mm, unsigned long start, unsigned long end, bool interruptible) argument
[all...]
/linux-master/arch/arm64/net/
H A Dbpf_jit_comp.c1515 * - offset[0] offset of the end of prologue,
1517 * - offset[1] - offset of the end of 1st instruction,
1520 * - offset[3] - offset of the end of 3rd instruction,
1542 * instruction (end of program)
1574 static inline void bpf_flush_icache(void *start, void *end) argument
1576 flush_icache_range((unsigned long)start, (unsigned long)end);
2281 * dummy_tramp is placed at the end:
2360 /* plt locates at the end of bpf prog */
/linux-master/kernel/
H A Dworkqueue.c1152 * Linked worklist will always end before the end of the list,
1932 * $pwq still has inactive work items, rotate it to the end of the
3557 * Leave this pool. Notify regular workers; otherwise, we end up
3587 unsigned long end = jiffies + BH_WORKER_JIFFIES; local
3610 --nr_restarts && time_before(jiffies, end));
4320 * smp_mb() at the end of set_work_pool_and_clear_pending() is paired
6437 * An unbound pool may end up with a cpumask which doesn't have any online
H A Dcpu.c312 ktime_t now, end, start = ktime_get(); local
315 end = start + 10ULL * NSEC_PER_SEC;
326 if (now > end) {
1077 goto end;
1105 end:
2304 enum cpuhp_state i, end; local
2310 end = CPUHP_AP_ONLINE_DYN_END;
2314 end = CPUHP_BP_PREPARE_DYN_END;
2320 for (i = state; i <= end; i++, step++) {
/linux-master/fs/nfsd/
H A Dnfs4xdr.c138 * buffer might end on a page boundary.
2603 char *str, *end, *next; local
2613 end = str = components;
2614 while (*end) {
2619 for (; *end && (*end != esc_exit); end++)
2620 /* find esc_exit or end of string */;
2621 next = end + 1;
2622 if (*end
[all...]
/linux-master/fs/nfs/
H A Dinode.c440 * This is our front-end to iget that looks up inodes by file handle
634 /* Optimization: if the end result is no change, don't RPC */
1820 u64 start, u64 end)
1840 if (end == nfsi->ooo->gap[i].start)
1841 end = nfsi->ooo->gap[i].end;
1842 else if (start == nfsi->ooo->gap[i].end)
1851 if (start != end) {
1859 nfsi->ooo->gap[cnt].end = end;
1819 nfs_ooo_merge(struct nfs_inode *nfsi, u64 start, u64 end) argument
[all...]
/linux-master/fs/bcachefs/
H A Dinode.c38 static int inode_decode_field(const u8 *in, const u8 *end, argument
45 if (in >= end)
58 if (in + bytes > end)
77 u8 *end = (void *) &packed[1]; local
116 BUG_ON(out > end);
157 const u8 *end = bkey_val_end(inode); local
170 ret = inode_decode_field(in, end, field, &field_bits); \
188 const u8 *in, const u8 *end,
197 ret = bch2_varint_decode_fast(in, end, &v[0]); \
203 ret = bch2_varint_decode_fast(in, end,
187 bch2_inode_unpack_v2(struct bch_inode_unpacked *unpacked, const u8 *in, const u8 *end, unsigned nr_fields) argument
231 const u8 *end = bkey_val_end(inode); local
818 struct bpos end = POS(inum.inum, U64_MAX); local
[all...]
H A Dbuckets.c512 * Backup superblock might be past the end of our normal usable space:
1040 unsigned new_ptrs_bytes = (void *) new_ptrs.end - (void *) new_ptrs.start;
1041 unsigned old_ptrs_bytes = (void *) old_ptrs.end - (void *) old_ptrs.start;
1131 * Backup superblock might be past the end of our normal usable space:
1175 u64 start, u64 end,
1182 min_t(u64, bucket_to_sector(ca, b + 1), end) - start;
1196 } while (start < end);
1173 bch2_trans_mark_metadata_sectors(struct btree_trans *trans, struct bch_dev *ca, u64 start, u64 end, enum bch_data_type type, u64 *bucket, unsigned *bucket_sectors) argument
H A Dbtree_node_scan.c270 static void bubble_up(struct found_btree_node *n, struct found_btree_node *end) argument
272 while (n + 1 < end &&
281 struct found_btree_node *end)
286 n < end &&
300 bubble_up(n, end);
279 handle_overwrites(struct bch_fs *c, struct found_btree_node *start, struct found_btree_node *end) argument
/linux-master/drivers/scsi/
H A Dsd.c29 * - end of transfer (bh + scsi_lib): SCSI_LOG_HLCOMPLETE level 1
1238 scmd_printk(KERN_ERR, cmd, "access beyond end of device\n");
1319 * This indicates that the command is ready from our end to be queued.
2200 * In case of bogus fw or device, we could end up having
3122 const struct scsi_io_group_descriptor *desc, *start, *end; local
3134 end = (void *)buffer + ALIGN_DOWN(data.header_length + data.length,
3135 sizeof(*end));
3141 for (desc = start; desc < end; desc++)
3736 * on @disk reach beyond the end of the device. If the SCSI host
3777 char *end local
[all...]
/linux-master/kernel/sched/
H A Dfair.c394 * at the end of the enqueue.
688 * and end up with a larger lag than we started with.
1277 * Mark the end of the wait period if dequeueing a
2192 * end try selecting ourselves (current == env->p) as a swap candidate.
3213 unsigned long start, end; local
3379 end = ALIGN(start + (pages << PAGE_SHIFT), HPAGE_SIZE);
3380 end = min(end, vma->vm_end);
3381 nr_pte_updates = change_prot_numa(vma, start, end);
3392 pages -= (end
[all...]
/linux-master/drivers/irqchip/
H A Dirq-gic-v3-its.c2232 phys_addr_t start, end, addr_end; local
2245 for_each_reserved_mem_range(i, &start, &end) {
2246 if (addr >= start && addr_end <= end)
3917 * easily. So in the end, vpe->pending_last is only an
5488 const unsigned long end)
5494 const unsigned long end)
5564 const unsigned long end)
5575 res.end = its_entry->base_address + ACPI_GICV3_ITS_MEM_SIZE - 1;
5612 const unsigned long end)
5620 .end
5487 gic_acpi_match_srat_its(union acpi_subtable_headers *header, const unsigned long end) argument
5493 gic_acpi_parse_srat_its(union acpi_subtable_headers *header, const unsigned long end) argument
5563 gic_acpi_parse_madt_its(union acpi_subtable_headers *header, const unsigned long end) argument
5611 its_acpi_reset(union acpi_subtable_headers *header, const unsigned long end) argument
[all...]
/linux-master/drivers/i2c/
H A Di2c-core-base.c1230 char *blank, end; local
1247 res = sscanf(++blank, "%hi%c", &info.addr, &end);
1252 if (res > 1 && end != '\n') {
1298 char end; local
1302 res = sscanf(buf, "%hi%c", &addr, &end);
1307 if (res > 1 && end != '\n') {
/linux-master/arch/riscv/mm/
H A Dinit.c217 * Reserve from the start of the kernel to the end of the kernel
241 * if end of dram is equal to maximum addressable memory. For 64-bit
242 * kernel, this problem can't happen here as the end of the virtual
263 * in the device tree, otherwise the allocation could end up in a
1107 * kernel, whereas for 64-bit kernel, the end of the virtual address
1204 pr_warn("fixmap btmap start [%08lx] != end [%08lx]\n",
1220 phys_addr_t end,
1226 for (pa = start; pa < end; pa += map_size) {
1229 best_map_size(pa, va, end - pa);
1238 phys_addr_t start, end; local
1219 create_linear_mapping_range(phys_addr_t start, phys_addr_t end, uintptr_t fixed_map_size) argument
1414 vmemmap_populate(unsigned long start, unsigned long end, int node, struct vmem_altmap *altmap) argument
1434 preallocate_pgd_pages_range(unsigned long start, unsigned long end, const char *area) argument
[all...]
/linux-master/arch/riscv/errata/thead/
H A Derrata.c160 void thead_errata_patch_func(struct alt_entry *begin, struct alt_entry *end, argument
169 for (alt = begin; alt < end; alt++) {
/linux-master/mm/
H A Dhugetlb.c100 unsigned long start, unsigned long end);
881 /* New entry for end of split region */
912 } else { /* Trim end of region */
1465 * next node from which to allocate, handling wrap at end of node
2533 * if we would end up overcommiting the surpluses. Abuse
3444 unsigned long end, void *arg)
3448 for (nid = start; nid < end; nid++)
3526 static void __init hugetlb_pages_alloc_boot_node(unsigned long start, unsigned long end, void *arg) argument
3529 int i, num = end - start;
5216 unsigned long reserve, start, end; local
3443 gather_bootmem_prealloc_parallel(unsigned long start, unsigned long end, void *arg) argument
5650 __unmap_hugepage_range(struct mmu_gather *tlb, struct vm_area_struct *vma, unsigned long start, unsigned long end, struct page *ref_page, zap_flags_t zap_flags) argument
5806 __hugetlb_zap_begin(struct vm_area_struct *vma, unsigned long *start, unsigned long *end) argument
5845 unmap_hugepage_range(struct vm_area_struct *vma, unsigned long start, unsigned long end, struct page *ref_page, zap_flags_t zap_flags) argument
6947 hugetlb_change_protection(struct vm_area_struct *vma, unsigned long address, unsigned long end, pgprot_t newprot, unsigned long cp_flags) argument
7255 hugetlb_unreserve_pages(struct inode *inode, long start, long end, long freed) argument
7328 unsigned long end = start + PUD_SIZE; local
7351 adjust_range_if_pmd_sharing_possible(struct vm_area_struct *vma, unsigned long *start, unsigned long *end) argument
7472 adjust_range_if_pmd_sharing_possible(struct vm_area_struct *vma, unsigned long *start, unsigned long *end) argument
7689 hugetlb_unshare_pmds(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument
[all...]
/linux-master/include/linux/
H A Dmm.h349 * These VMAs will get a single end guard page. This helps userspace protect
587 unsigned long end, unsigned long newflags);
700 rcu_read_lock(); /* keeps vma alive till the end of up_read */
1017 unsigned long start, unsigned long end, gfp_t gfp)
1019 __mas_set_range(&vmi->mas, start, end - 1);
1058 /* The MM code likes to work with exclusive end addresses */
2388 unsigned long end, unsigned long tree_end, bool mm_wr_locked);
2393 unsigned long end, unsigned long floor, unsigned long ceiling);
2408 void truncate_pagecache_range(struct inode *inode, loff_t offset, loff_t end);
2573 unsigned long end, unsigne
1016 vma_iter_clear_gfp(struct vma_iterator *vmi, unsigned long start, unsigned long end, gfp_t gfp) argument
3299 vma_modify_flags(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long new_flags) argument
3312 vma_modify_flags_name(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long new_flags, struct anon_vma_name *new_name) argument
3326 vma_modify_policy(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, struct mempolicy *new_pol) argument
3338 vma_modify_flags_uffd(struct vma_iterator *vmi, struct vm_area_struct *prev, struct vm_area_struct *vma, unsigned long start, unsigned long end, unsigned long new_flags, struct vm_userfaultfd_ctx new_ctx) argument
3535 range_in_vma(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument
4188 range_contains_unaccepted_memory(phys_addr_t start, phys_addr_t end) argument
4194 accept_memory(phys_addr_t start, phys_addr_t end) argument
[all...]
/linux-master/fs/ntfs3/
H A Dsuper.c1667 u64 lbo, bytes, start, end; local
1685 /* Align down 'end' on discard_granularity. */
1686 end = (lbo + bytes) & sbi->discard_granularity_mask_inv;
1689 if (start >= end)
1692 err = blkdev_issue_discard(sb->s_bdev, start >> 9, (end - start) >> 9,
H A Dinode.c785 loff_t end; local
802 end = vbo + ret;
804 end = vbo + iter_count;
810 if (end > valid && !S_ISBLK(inode->i_mode)) {
811 ni->i_valid = end;
814 } else if (vbo < valid && valid < end) {
816 iov_iter_revert(iter, end - valid);
817 iov_iter_zero(end - valid, iter);
H A Dfile.c290 CLST end = bytes_to_cluster(sbi, to); local
293 for (; vcn < end; vcn += len) {
324 loff_t end = pos + count; local
328 if (end <= inode->i_size && !extend_init)
334 if (end > inode->i_size) {
335 err = ntfs_set_size(inode, end);
354 err = filemap_fdatawrite_range(mapping, pos, end - 1);
362 err = filemap_fdatawait_range(mapping, pos, end - 1);
439 loff_t end = vbo + len; local
474 new_size = max(end, i_siz
[all...]
H A Ddir.c182 const u8 *end; local
195 for (ret = 0, end = name + name_len; name < end; ret++, name += slen) {
199 slen = nls->char2uni(name, end - name, uname + ret);
351 u32 end = le32_to_cpu(hdr->used); local
355 if (off + sizeof(struct NTFS_DE) > end)
360 if (e_size < sizeof(struct NTFS_DE) || off + e_size > end)
517 u32 e_size, off, end; local
532 end = le32_to_cpu(hdr->used);
535 for (; off + sizeof(struct NTFS_DE) <= end; of
[all...]
/linux-master/fs/netfs/
H A Dbuffered_write.c898 unsigned long long end)
909 _enter(",%lx,%llx-%llx,%u", folio->index, start, end, caching);
928 * end of the range.
948 max_len = min_t(unsigned long long, max_len, end - start + 1);
1003 unsigned long long end)
1012 _enter("%llx,%llx,", start, end);
1019 folio = xas_find_marked(xas, end / PAGE_SIZE, PAGECACHE_TAG_DIRTY);
1102 folio, start, end);
1117 unsigned long long end)
1125 _start, end);
892 netfs_write_back_from_locked_folio(struct address_space *mapping, struct writeback_control *wbc, struct netfs_group *group, struct xa_state *xas, struct folio *folio, unsigned long long start, unsigned long long end) argument
998 netfs_writepages_begin(struct address_space *mapping, struct writeback_control *wbc, struct netfs_group *group, struct xa_state *xas, unsigned long long *_start, unsigned long long end) argument
1113 netfs_writepages_region(struct address_space *mapping, struct writeback_control *wbc, struct netfs_group *group, unsigned long long *_start, unsigned long long end) argument
1140 loff_t start, end; local
[all...]
/linux-master/drivers/video/fbdev/core/
H A Dfb_defio.c122 int fb_deferred_io_fsync(struct file *file, loff_t start, loff_t end, int datasync) argument
126 int err = file_write_and_wait_range(file, start, end);

Completed in 611 milliseconds

1234567891011>>