/linux-master/fs/btrfs/ |
H A D | extent-io-tree.c | 182 static struct extent_state *alloc_extent_state_atomic(struct extent_state *prealloc) argument 184 if (!prealloc) 185 prealloc = alloc_extent_state(GFP_ATOMIC); 187 return prealloc; 511 * struct 'prealloc' as the newly created second half. 'split' indicates an 517 * prealloc: [orig->start, split - 1] 524 struct extent_state *prealloc, u64 split) 533 prealloc->start = orig->start; 534 prealloc->end = split - 1; 535 prealloc 523 split_state(struct extent_io_tree *tree, struct extent_state *orig, struct extent_state *prealloc, u64 split) argument 631 struct extent_state *prealloc = NULL; local 1059 struct extent_state *prealloc = NULL; local 1312 struct extent_state *prealloc = NULL; local [all...] |
H A D | qgroup.c | 195 * Must be called with qgroup_lock held and @prealloc preallocated. 197 * The control on the lifespan of @prealloc would be transferred to this 198 * function, thus caller should no longer touch @prealloc. 201 struct btrfs_qgroup *prealloc, 208 /* Caller must have pre-allocated @prealloc. */ 209 ASSERT(prealloc); 220 kfree(prealloc); 225 qgroup = prealloc; 278 * Must be called with qgroup_lock held, the ownership of @prealloc is 285 static int __add_relation_rb(struct btrfs_qgroup_list *prealloc, argument 200 add_qgroup_rb(struct btrfs_fs_info *fs_info, struct btrfs_qgroup *prealloc, u64 qgroupid) argument 311 add_relation_rb(struct btrfs_fs_info *fs_info, struct btrfs_qgroup_list *prealloc, u64 memberid, u64 parentid) argument 470 struct btrfs_qgroup *prealloc; local 1006 struct btrfs_qgroup *prealloc = NULL; local 1569 struct btrfs_qgroup_list *prealloc = NULL; local 1699 struct btrfs_qgroup *prealloc = NULL; local 3215 struct btrfs_qgroup *prealloc; local [all...] |
H A D | extent_map.c | 303 bool prealloc = false; local 332 prealloc = true; 338 if (prealloc) {
|
H A D | extent_io.c | 881 struct btrfs_subpage *prealloc) 903 /* Already mapped, just free prealloc */ 905 btrfs_free_subpage(prealloc); 909 if (prealloc) 911 folio_attach_private(folio, prealloc); 2476 * delalloc in holes and prealloc extents. And this can lead to a deadlock 2582 * when we had a hole or prealloc extent with several delalloc ranges in 2594 * a hole or prealloc extent and we have now found a 2621 * processing a hole or prealloc extent for which we 2863 * Process a range which is a hole or a prealloc exten 879 attach_extent_buffer_folio(struct extent_buffer *eb, struct folio *folio, struct btrfs_subpage *prealloc) argument 3795 struct btrfs_subpage *prealloc = NULL; local [all...] |
/linux-master/lib/ |
H A D | stackdepot.c | 244 static bool depot_init_pool(void **prealloc) argument 256 if (!new_pool && *prealloc) { 258 WRITE_ONCE(new_pool, *prealloc); 259 *prealloc = NULL; 263 return false; /* new_pool and *prealloc are NULL */ 291 static void depot_keep_new_pool(void **prealloc) argument 302 WRITE_ONCE(new_pool, *prealloc); 303 *prealloc = NULL; 310 static struct stack_record *depot_pop_free_pool(void **prealloc, size_t size) argument 319 if (!depot_init_pool(prealloc)) 381 depot_alloc_stack(unsigned long *entries, unsigned int nr_entries, u32 hash, depot_flags_t flags, void **prealloc) argument 593 void *prealloc = NULL; local [all...] |
/linux-master/arch/arc/lib/ |
H A D | memset-archs.S | 10 * The memset implementation below is optimized to use prefetchw and prealloc 14 * we don't call any prefetchw/prealloc instruction for L1 cache lines which 21 prealloc [\reg, \off]
|
/linux-master/tools/testing/selftests/net/ |
H A D | ioam6.sh | 219 trace prealloc type 0x800000 ns 0 size 4 dev veth0 487 trace prealloc type 0x800000 ns 0 size 4 dev veth0 507 trace prealloc type 0xc00000 ns 123 size 4 dev veth0 533 trace prealloc type ${bit2type[$i]} ns 123 size ${bit2size[$i]} \ 572 trace prealloc type 0xfff002 ns 123 size 100 dev veth0 602 trace prealloc type 0x800000 ns 0 size 4 dev veth0 622 trace prealloc type 0xc00000 ns 123 size 4 dev veth0 648 trace prealloc type ${bit2type[$i]} ns 123 size ${bit2size[$i]} \ 677 trace prealloc type 0xc00000 ns 123 size 4 dev veth0 701 trace prealloc typ [all...] |
/linux-master/drivers/media/platform/renesas/vsp1/ |
H A D | vsp1_dl.h | 56 unsigned int prealloc);
|
H A D | vsp1_dl.c | 1091 unsigned int prealloc) 1112 * memory. An extra body is allocated on top of the prealloc to account 1121 dlm->pool = vsp1_dl_body_pool_create(vsp1, prealloc + 1, 1126 for (i = 0; i < prealloc; ++i) { 1145 VSP1_EXTCMD_AUTOFLD, prealloc); 1089 vsp1_dlm_create(struct vsp1_device *vsp1, unsigned int index, unsigned int prealloc) argument
|
/linux-master/arch/powerpc/platforms/ps3/ |
H A D | setup.c | 113 static void __init prealloc(struct ps3_prealloc *p) function 135 #define prealloc_ps3fb_videomemory() prealloc(&ps3fb_videomemory) 158 #define prealloc_ps3flash_bounce_buffer() prealloc(&ps3flash_bounce_buffer)
|
/linux-master/fs/ext4/ |
H A D | extents_status.c | 148 struct extent_status *prealloc); 151 struct extent_status *prealloc); 157 struct pending_reservation **prealloc); 786 struct extent_status *prealloc) 826 if (prealloc) 827 es = prealloc; 1359 * @prealloc - pre-allocated es to avoid memory allocation failures 1368 struct extent_status *prealloc) 1415 err = __es_insert_extent(inode, &newes, prealloc); 1939 * @prealloc 785 __es_insert_extent(struct inode *inode, struct extent_status *newes, struct extent_status *prealloc) argument 1366 __es_remove_extent(struct inode *inode, ext4_lblk_t lblk, ext4_lblk_t end, int *reserved, struct extent_status *prealloc) argument 1944 __insert_pending(struct inode *inode, ext4_lblk_t lblk, struct pending_reservation **prealloc) argument 2238 __revise_pending(struct inode *inode, ext4_lblk_t lblk, ext4_lblk_t len, struct pending_reservation **prealloc) argument [all...] |
/linux-master/net/sched/ |
H A D | sch_gred.c | 484 struct gred_sched_data **prealloc, 496 table->tab[dp] = q = *prealloc; 497 *prealloc = NULL; 651 struct gred_sched_data *prealloc; local 703 prealloc = kzalloc(sizeof(*prealloc), GFP_KERNEL); 706 err = gred_change_vq(sch, ctl->DP, ctl, prio, stab, max_P, &prealloc, 721 kfree(prealloc); 728 kfree(prealloc); 481 gred_change_vq(struct Qdisc *sch, int dp, struct tc_gred_qopt *ctl, int prio, u8 *stab, u32 max_P, struct gred_sched_data **prealloc, struct netlink_ext_ack *extack) argument
|
/linux-master/drivers/md/ |
H A D | dm-cache-target.c | 1398 struct dm_bio_prison_cell_v2 *prealloc; local 1400 prealloc = alloc_prison_cell(cache); 1410 prealloc, &mg->cell); 1412 free_prison_cell(cache, prealloc); 1417 if (mg->cell != prealloc) 1418 free_prison_cell(cache, prealloc); 1528 struct dm_bio_prison_cell_v2 *prealloc; local 1530 prealloc = alloc_prison_cell(cache); 1534 READ_WRITE_LOCK_LEVEL, prealloc, &mg->cell); 1536 free_prison_cell(cache, prealloc); [all...] |
/linux-master/drivers/gpu/drm/i915/display/ |
H A D | intel_fbdev.c | 164 bool prealloc = false; local 195 prealloc = true; 236 if (!i915_gem_object_is_shmem(obj) && !prealloc)
|
/linux-master/fs/sysfs/ |
H A D | file.c | 232 .prealloc = true, 237 .prealloc = true, 243 .prealloc = true,
|
/linux-master/fs/xfs/libxfs/ |
H A D | xfs_bmap.h | 216 xfs_fileoff_t off, xfs_filblks_t len, xfs_filblks_t prealloc,
|
H A D | xfs_bmap.c | 4060 xfs_filblks_t prealloc, 4073 * Cap the alloc length. Keep track of prealloc so we know whether to 4076 alen = XFS_FILBLKS_MIN(len + prealloc, XFS_MAX_BMBT_EXTLEN); 4079 if (prealloc && alen >= len) 4080 prealloc = alen - len; 4133 * prealloc == 0, so we must also check the aligned offset and length. 4135 if (whichfork == XFS_DATA_FORK && prealloc) 4137 if (whichfork == XFS_COW_FORK && (prealloc || aoff < off || alen > len)) 4055 xfs_bmapi_reserve_delalloc( struct xfs_inode *ip, int whichfork, xfs_fileoff_t off, xfs_filblks_t len, xfs_filblks_t prealloc, struct xfs_bmbt_irec *got, struct xfs_iext_cursor *icur, int eof) argument
|
/linux-master/fs/xfs/ |
H A D | xfs_dquot.c | 99 int prealloc = 0; local 106 prealloc = 1; 110 prealloc = 1; 121 if (prealloc) 632 /* initialize the dquot speculative prealloc thresholds */
|
/linux-master/kernel/bpf/ |
H A D | hashtab.c | 423 bool prealloc = !(attr->map_flags & BPF_F_NO_PREALLOC); local 441 if (lru && !prealloc) 478 bool prealloc = !(attr->map_flags & BPF_F_NO_PREALLOC); local 572 if (prealloc) { 1004 bool prealloc = htab_is_prealloc(htab); local 1008 if (prealloc) { 1045 if (prealloc) { 1061 if (!prealloc) 2221 bool prealloc = htab_is_prealloc(htab); local 2229 if (prealloc) { [all...] |
/linux-master/drivers/usb/gadget/function/ |
H A D | u_ether.c | 313 static int prealloc(struct list_head *list, struct usb_ep *ep, unsigned n) function 357 status = prealloc(&dev->tx_reqs, link->in_ep, n); 360 status = prealloc(&dev->rx_reqs, link->out_ep, n);
|
/linux-master/mm/ |
H A D | memory.c | 903 struct folio **prealloc, struct page *page) 908 new_folio = *prealloc; 913 * We have a prealloc page, all good! Take it 916 *prealloc = NULL; 966 int max_nr, int *rss, struct folio **prealloc) 985 if (unlikely(!*prealloc && folio_test_large(folio) && max_nr != 1)) { 1025 addr, rss, prealloc, page); 1077 struct folio *prealloc = NULL; local 1149 /* copy_present_ptes() will clear `*prealloc' if consumed */ 1152 ptent, addr, max_nr, rss, &prealloc); 901 copy_present_page(struct vm_area_struct *dst_vma, struct vm_area_struct *src_vma, pte_t *dst_pte, pte_t *src_pte, unsigned long addr, int *rss, struct folio **prealloc, struct page *page) argument 964 copy_present_ptes(struct vm_area_struct *dst_vma, struct vm_area_struct *src_vma, pte_t *dst_pte, pte_t *src_pte, pte_t pte, unsigned long addr, int max_nr, int *rss, struct folio **prealloc) argument [all...] |
/linux-master/include/linux/ |
H A D | kernfs.h | 308 * "prealloc" causes a buffer to be allocated at open for 311 * ->prealloc. Provide ->read and ->write with ->prealloc. 313 bool prealloc; member in struct:kernfs_ops
|
/linux-master/fs/ntfs3/ |
H A D | file.c | 397 &new_valid, ni->mi.sbi->options->prealloc, NULL); 1175 if (sbi->options->prealloc &&
|
/linux-master/fs/kernfs/ |
H A D | file.c | 665 * ->seq_show is incompatible with ->prealloc, 669 if (ops->prealloc && ops->seq_show) 671 if (ops->prealloc) {
|
/linux-master/fs/f2fs/ |
H A D | data.c | 1524 blkcnt_t prealloc; local 1563 prealloc = 0; 1587 prealloc++; 1677 err = f2fs_reserve_new_blocks(&dn, prealloc); 1682 if (prealloc && dn.ofs_in_node != last_ofs_in_node + 1) {
|