/linux-master/include/linux/ |
H A D | mm.h | 180 * per mm struct. Users can overwrite this number by sysctl but there is a 240 * The idea being to have a "virtual" mm in the same way 242 * mm details, and allowing different kinds of memory mappings 347 * support core mm. 453 /* This mask defines which mm->def_flags a process can inherit its parent */ 626 * in mm/mempolicy.c will do this automatically. 683 * vma->vm_lock_seq under vma->vm_lock protection and mm->mm_lock_seq 712 * mm->mm_lock_seq can't be concurrently modified. 778 struct vm_area_struct *lock_vma_under_rcu(struct mm_struct *mm, 792 static inline struct vm_area_struct *lock_vma_under_rcu(struct mm_struct *mm, argument 821 vma_init(struct vm_area_struct *vma, struct mm_struct *mm) argument 2433 fixup_user_fault(struct mm_struct *mm, unsigned long address, unsigned int fault_flags, bool *unlocked) argument 2472 get_user_page_vma_remote(struct mm_struct *mm, unsigned long addr, int gup_flags, struct vm_area_struct **vmap) argument 2590 get_mm_counter(struct mm_struct *mm, int member) argument 2597 add_mm_counter(struct mm_struct *mm, int member, long value) argument 2604 inc_mm_counter(struct mm_struct *mm, int member) argument 2611 dec_mm_counter(struct mm_struct *mm, int member) argument 2633 get_mm_rss(struct mm_struct *mm) argument 2640 get_mm_hiwater_rss(struct mm_struct *mm) argument 2645 get_mm_hiwater_vm(struct mm_struct *mm) argument 2650 update_hiwater_rss(struct mm_struct *mm) argument 2658 update_hiwater_vm(struct mm_struct *mm) argument 2664 reset_mm_hiwater_rss(struct mm_struct *mm) argument 2669 setmax_mm_hiwater_rss(unsigned long *maxrss, struct mm_struct *mm) argument 2699 get_locked_pte(struct mm_struct *mm, unsigned long addr, spinlock_t **ptl) argument 2708 __p4d_alloc(struct mm_struct *mm, pgd_t *pgd, unsigned long address) argument 2718 __pud_alloc(struct mm_struct *mm, p4d_t *p4d, unsigned long address) argument 2723 mm_inc_nr_puds(struct mm_struct *mm) argument 2724 mm_dec_nr_puds(struct mm_struct *mm) argument 2729 mm_inc_nr_puds(struct mm_struct *mm) argument 2736 mm_dec_nr_puds(struct mm_struct *mm) argument 2745 __pmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long address) argument 2751 mm_inc_nr_pmds(struct mm_struct *mm) argument 2752 mm_dec_nr_pmds(struct mm_struct *mm) argument 2757 mm_inc_nr_pmds(struct mm_struct *mm) argument 2764 mm_dec_nr_pmds(struct mm_struct *mm) argument 2773 mm_pgtables_bytes_init(struct mm_struct *mm) argument 2778 mm_pgtables_bytes(const struct mm_struct *mm) argument 2783 mm_inc_nr_ptes(struct mm_struct *mm) argument 2788 mm_dec_nr_ptes(struct mm_struct *mm) argument 2794 mm_pgtables_bytes_init(struct mm_struct *mm) argument 2795 mm_pgtables_bytes(const struct mm_struct *mm) argument 2800 mm_inc_nr_ptes(struct mm_struct *mm) argument 2801 mm_dec_nr_ptes(struct mm_struct *mm) argument 2809 p4d_alloc(struct mm_struct *mm, pgd_t *pgd, unsigned long address) argument 2816 pud_alloc(struct mm_struct *mm, p4d_t *p4d, unsigned long address) argument 2823 pmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long address) argument 2911 pte_lockptr(struct mm_struct *mm, pmd_t *pmd) argument 2936 pte_lockptr(struct mm_struct *mm, pmd_t *pmd) argument 2973 pte_offset_map_lock(struct mm_struct *mm, pmd_t *pmd, unsigned long addr, spinlock_t **ptlp) argument 3016 pmd_lockptr(struct mm_struct *mm, pmd_t *pmd) argument 3041 pmd_lockptr(struct mm_struct *mm, pmd_t *pmd) argument 3053 pmd_lock(struct mm_struct *mm, pmd_t *pmd) argument 3086 pud_lockptr(struct mm_struct *mm, pud_t *pud) argument 3091 pud_lock(struct mm_struct *mm, pud_t *pud) argument 3476 vma_lookup(struct mm_struct *mm, unsigned long addr) argument 3522 find_exact_vma(struct mm_struct *mm, unsigned long vm_start, unsigned long vm_end) argument 3825 get_gate_vma(struct mm_struct *mm) argument 3830 in_gate_area(struct mm_struct *mm, unsigned long addr) argument 4173 madvise_set_anon_name(struct mm_struct *mm, unsigned long start, unsigned long len_in, struct anon_vma_name *anon_name) argument [all...] |
H A D | mm_types.h | 652 /* VMA covers [vm_start; vm_end) addresses within mm */ 665 * Flags, see mm.h. 691 /* Flag to indicate areas detached from the mm->mm_mt tree */ 854 struct list_head mmlist; /* List of maybe swapped mm's. These 891 * protecting pages mapped by this mm to enforce a later COW, 920 * user/owner of this mm. All of the following must be true in 923 * current == mm->owner 924 * current->mm != mm 925 * new_owner->mm 1023 mm_init_cpumask(struct mm_struct *mm) argument 1032 mm_cpumask(struct mm_struct *mm) argument 1054 lru_gen_init_mm(struct mm_struct *mm) argument 1063 lru_gen_use_mm(struct mm_struct *mm) argument 1075 lru_gen_add_mm(struct mm_struct *mm) argument 1079 lru_gen_del_mm(struct mm_struct *mm) argument 1083 lru_gen_migrate_mm(struct mm_struct *mm) argument 1087 lru_gen_init_mm(struct mm_struct *mm) argument 1091 lru_gen_use_mm(struct mm_struct *mm) argument 1111 vma_iter_init(struct vma_iterator *vmi, struct mm_struct *mm, unsigned long addr) argument 1150 mm_cidmask(struct mm_struct *mm) argument 1160 mm_init_cid(struct mm_struct *mm) argument 1173 mm_alloc_cid(struct mm_struct *mm) argument 1182 mm_destroy_cid(struct mm_struct *mm) argument 1193 mm_init_cid(struct mm_struct *mm) argument 1194 mm_alloc_cid(struct mm_struct *mm) argument 1195 mm_destroy_cid(struct mm_struct *mm) argument [all...] |
/linux-master/arch/sparc/mm/ |
H A D | srmmu.c | 24 #include <linux/mm.h> 72 #define FLUSH_BEGIN(mm) 75 #define FLUSH_BEGIN(mm) if ((mm)->context != NO_CONTEXT) { 349 pgtable_t pte_alloc_one(struct mm_struct *mm) argument 354 if (!(ptep = pte_alloc_one_kernel(mm))) 357 spin_lock(&mm->page_table_lock); 363 spin_unlock(&mm->page_table_lock); 368 void pte_free(struct mm_struct *mm, pgtable_t ptep) argument 373 spin_lock(&mm 414 alloc_context(struct mm_struct *old_mm, struct mm_struct *mm) argument 472 switch_mm(struct mm_struct *old_mm, struct mm_struct *mm, struct task_struct *tsk) argument 999 init_new_context(struct task_struct *tsk, struct mm_struct *mm) argument 1005 destroy_context(struct mm_struct *mm) argument 1238 turbosparc_flush_cache_mm(struct mm_struct *mm) argument 1276 turbosparc_flush_sig_insns(struct mm_struct *mm, unsigned long insn_addr) argument 1290 turbosparc_flush_tlb_mm(struct mm_struct *mm) argument 1656 smp_flush_cache_mm(struct mm_struct *mm) argument 1668 smp_flush_tlb_mm(struct mm_struct *mm) argument 1688 struct mm_struct *mm = vma->vm_mm; local 1705 struct mm_struct *mm = vma->vm_mm; local 1720 struct mm_struct *mm = vma->vm_mm; local 1734 struct mm_struct *mm = vma->vm_mm; local 1760 smp_flush_sig_insns(struct mm_struct *mm, unsigned long insn_addr) argument [all...] |
H A D | fault_64.c | 3 * arch/sparc64/mm/fault.c: Page fault handlers for the 64-bit Sparc. 18 #include <linux/mm.h> 51 printk(KERN_ALERT "tsk->{mm,active_mm}->context = %016lx\n", 52 (tsk->mm ? 53 CTX_HWBITS(tsk->mm->context) : 55 printk(KERN_ALERT "tsk->{mm,active_mm}->pgd = %016lx\n", 56 (tsk->mm ? (unsigned long) tsk->mm->pgd : 81 pgd_t *pgdp = pgd_offset(current->mm, tpc); 270 struct mm_struct *mm local [all...] |
H A D | fault_32.c | 20 #include <linux/mm.h> 50 printk(KERN_ALERT "tsk->{mm,active_mm}->context = %08lx\n", 51 (tsk->mm ? tsk->mm->context : tsk->active_mm->context)); 52 printk(KERN_ALERT "tsk->{mm,active_mm}->pgd = %08lx\n", 53 (tsk->mm ? (unsigned long) tsk->mm->pgd : 117 struct mm_struct *mm = tsk->mm; local 143 if (pagefault_disabled() || !mm) 309 struct mm_struct *mm = tsk->mm; local [all...] |
/linux-master/mm/ |
H A D | mremap.c | 3 * mm/mremap.c 11 #include <linux/mm.h> 35 static pud_t *get_old_pud(struct mm_struct *mm, unsigned long addr) argument 41 pgd = pgd_offset(mm, addr); 56 static pmd_t *get_old_pmd(struct mm_struct *mm, unsigned long addr) argument 61 pud = get_old_pud(mm, addr); 72 static pud_t *alloc_new_pud(struct mm_struct *mm, struct vm_area_struct *vma, argument 78 pgd = pgd_offset(mm, addr); 79 p4d = p4d_alloc(mm, pgd, addr); 83 return pud_alloc(mm, p4 86 alloc_new_pmd(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr) argument 141 struct mm_struct *mm = vma->vm_mm; local 240 struct mm_struct *mm = vma->vm_mm; local 308 struct mm_struct *mm = vma->vm_mm; local 357 struct mm_struct *mm = vma->vm_mm; local 657 struct mm_struct *mm = vma->vm_mm; local 816 struct mm_struct *mm = current->mm; local 873 struct mm_struct *mm = current->mm; local 979 struct mm_struct *mm = current->mm; local [all...] |
H A D | nommu.c | 3 * linux/mm/nommu.c 5 * Replacement code for mm functions to support CPU's that don't 8 * See Documentation/admin-guide/mm/nommu-mmap.rst 20 #include <linux/mm.h> 21 #include <linux/sched/mm.h> 101 vma = find_vma(current->mm, (unsigned long)objp); 172 mmap_write_lock(current->mm); 173 vma = find_vma(current->mm, (unsigned long)ret); 176 mmap_write_unlock(current->mm); 381 struct mm_struct *mm local 545 setup_vma_to_mm(struct vm_area_struct *vma, struct mm_struct *mm) argument 599 delete_vma(struct mm_struct *mm, struct vm_area_struct *vma) argument 609 find_vma_intersection(struct mm_struct *mm, unsigned long start_addr, unsigned long end_addr) argument 624 find_vma(struct mm_struct *mm, unsigned long addr) argument 636 lock_mm_and_find_vma(struct mm_struct *mm, unsigned long addr, struct pt_regs *regs) argument 657 expand_stack(struct mm_struct *mm, unsigned long addr) argument 667 find_vma_exact(struct mm_struct *mm, unsigned long addr, unsigned long len) argument 1312 struct mm_struct *mm; local 1426 do_munmap(struct mm_struct *mm, unsigned long start, size_t len, struct list_head *uf) argument 1490 struct mm_struct *mm = current->mm; local 1508 exit_mmap(struct mm_struct *mm) argument 1647 __access_remote_vm(struct mm_struct *mm, unsigned long addr, void *buf, int len, unsigned int gup_flags) argument 1691 access_remote_vm(struct mm_struct *mm, unsigned long addr, void *buf, int len, unsigned int gup_flags) argument 1704 struct mm_struct *mm; local [all...] |
H A D | process_vm_access.c | 3 * linux/mm/process_vm_access.c 9 #include <linux/mm.h> 12 #include <linux/sched/mm.h> 68 * @mm: mm for task 77 struct mm_struct *mm, 102 * access remotely because task/mm might not 103 * current/current->mm 105 mmap_read_lock(mm); 106 pinned_pages = pin_user_pages_remote(mm, p 73 process_vm_rw_single_vec(unsigned long addr, unsigned long len, struct iov_iter *iter, struct page **process_pages, struct mm_struct *mm, struct task_struct *task, int vm_write) argument 159 struct mm_struct *mm; local [all...] |
H A D | pgtable-generic.c | 3 * mm/pgtable-generic.c 97 struct mm_struct *mm = (vma)->vm_mm; local 99 pte = ptep_get_and_clear(mm, address, ptep); 100 if (pte_accessible(mm, pte)) 165 void pgtable_trans_huge_deposit(struct mm_struct *mm, pmd_t *pmdp, argument 168 assert_spin_locked(pmd_lockptr(mm, pmdp)); 171 if (!pmd_huge_pte(mm, pmdp)) 174 list_add(&pgtable->lru, &pmd_huge_pte(mm, pmdp)->lru); 175 pmd_huge_pte(mm, pmdp) = pgtable; 181 pgtable_t pgtable_trans_huge_withdraw(struct mm_struct *mm, pmd_ argument 245 pte_free_defer(struct mm_struct *mm, pgtable_t pgtable) argument 306 pte_offset_map_nolock(struct mm_struct *mm, pmd_t *pmd, unsigned long addr, spinlock_t **ptlp) argument 362 __pte_offset_map_lock(struct mm_struct *mm, pmd_t *pmd, unsigned long addr, spinlock_t **ptlp) argument [all...] |
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_ttm_vram_mgr.c | 24 static inline bool xe_is_vram_mgr_blocks_contiguous(struct drm_buddy *mm, argument 36 size = drm_buddy_block_size(mm, block); 54 struct drm_buddy *mm = &mgr->mm; local 96 if (WARN_ON(min_page_size < mm->chunk_size)) { 140 err = drm_buddy_alloc_blocks(mm, (u64)place->fpfn << PAGE_SHIFT, 153 if (!drm_buddy_block_trim(mm, vres->base.size, &vres->blocks)) 166 u64 end = start + drm_buddy_block_size(mm, block); 178 xe_is_vram_mgr_blocks_contiguous(mm, &vres->blocks)) 199 drm_buddy_free_list(mm, 214 struct drm_buddy *mm = &mgr->mm; local 230 struct drm_buddy *mm = &mgr->mm; local 253 struct drm_buddy *mm = &mgr->mm; local 283 struct drm_buddy *mm = &mgr->mm; local [all...] |
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | pgtable.c | 23 #include <mm/mmu_decl.h> 105 void set_pmd_at(struct mm_struct *mm, unsigned long addr, argument 115 assert_spin_locked(pmd_lockptr(mm, pmdp)); 119 return set_pte_at(mm, addr, pmdp_ptep(pmdp), pmd_pte(pmd)); 122 void set_pud_at(struct mm_struct *mm, unsigned long addr, argument 132 assert_spin_locked(pud_lockptr(mm, pudp)); 136 return set_pte_at(mm, addr, pudp_ptep(pudp), pud_pte(pud)); 143 struct mm_struct *mm = arg; local 144 exit_lazy_flush_tlb(mm, false); 158 void serialize_against_pte_lookup(struct mm_struct *mm) argument 161 smp_call_function_many(mm_cpumask(mm), do_serialize, mm, 1); local 363 get_pmd_from_cache(struct mm_struct *mm) argument 385 __alloc_for_pmdcache(struct mm_struct *mm) argument 426 pmd_fragment_alloc(struct mm_struct *mm, unsigned long vmaddr) argument [all...] |
/linux-master/arch/powerpc/kernel/ |
H A D | vdso.c | 11 #include <linux/mm.h> 69 current->mm->context.vdso = (void __user *)new_vma->vm_start; 117 struct mm_struct *mm = task->mm; local 118 VMA_ITERATOR(vmi, mm, 0); 121 mmap_read_lock(mm); 126 mmap_read_unlock(mm); 168 * vDSO and insert it into the mm struct tree 175 struct mm_struct *mm = current->mm; local 234 struct mm_struct *mm = current->mm; local [all...] |
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_mqd_manager_v10.c | 45 static void update_cu_mask(struct mqd_manager *mm, void *mqd, argument 54 mqd_symmetrically_map_cu_mask(mm, 88 static void init_mqd(struct mqd_manager *mm, void **mqd, argument 129 if (mm->dev->kfd->cwsr_enabled) { 145 mm->update_mqd(mm, m, q, NULL); 148 static int load_mqd(struct mqd_manager *mm, void *mqd, argument 156 r = mm->dev->kfd2kgd->hqd_load(mm->dev->adev, mqd, pipe_id, queue_id, 162 static void update_mqd(struct mqd_manager *mm, voi argument 234 get_wave_state(struct mqd_manager *mm, void *mqd, struct queue_properties *q, void __user *ctl_stack, u32 *ctl_stack_used_size, u32 *save_area_used_size) argument 272 checkpoint_mqd(struct mqd_manager *mm, void *mqd, void *mqd_dst, void *ctl_stack_dst) argument 281 restore_mqd(struct mqd_manager *mm, void **mqd, struct kfd_mem_obj *mqd_mem_obj, uint64_t *gart_addr, struct queue_properties *qp, const void *mqd_src, const void *ctl_stack_src, const u32 ctl_stack_size) argument 308 init_mqd_hiq(struct mqd_manager *mm, void **mqd, struct kfd_mem_obj *mqd_mem_obj, uint64_t *gart_addr, struct queue_properties *q) argument 322 destroy_hiq_mqd(struct mqd_manager *mm, void *mqd, enum kfd_preempt_type type, unsigned int timeout, uint32_t pipe_id, uint32_t queue_id) argument 342 init_mqd_sdma(struct mqd_manager *mm, void **mqd, struct kfd_mem_obj *mqd_mem_obj, uint64_t *gart_addr, struct queue_properties *q) argument 361 update_mqd_sdma(struct mqd_manager *mm, void *mqd, struct queue_properties *q, struct mqd_update_info *minfo) argument 388 checkpoint_mqd_sdma(struct mqd_manager *mm, void *mqd, void *mqd_dst, void *ctl_stack_dst) argument 400 restore_mqd_sdma(struct mqd_manager *mm, void **mqd, struct kfd_mem_obj *mqd_mem_obj, uint64_t *gart_addr, struct queue_properties *qp, const void *mqd_src, const void *ctl_stack_src, const u32 ctl_stack_size) argument [all...] |
/linux-master/arch/powerpc/mm/ |
H A D | hugetlbpage.c | 11 #include <linux/mm.h> 36 pte_t *huge_pte_offset(struct mm_struct *mm, unsigned long addr, unsigned long sz) argument 42 return __find_linux_pte(mm->pgd, addr, NULL, NULL); 45 static int __hugepte_alloc(struct mm_struct *mm, hugepd_t *hpdp, argument 67 new = kmem_cache_alloc(cachep, pgtable_gfp_flags(mm, GFP_KERNEL)); 110 pte_t *huge_pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma, argument 123 pg = pgd_offset(mm, addr); 134 ptl = &mm->page_table_lock; 138 pu = pud_alloc(mm, p4, addr); 144 ptl = pud_lockptr(mm, p [all...] |
/linux-master/arch/arm/kernel/ |
H A D | smp_tlb.c | 33 struct mm_struct *mm = (struct mm_struct *)arg; local 35 local_flush_tlb_mm(mm); 164 static void broadcast_tlb_mm_a15_erratum(struct mm_struct *mm) argument 173 a15_erratum_get_cpumask(this_cpu, mm, &mask); 187 void flush_tlb_mm(struct mm_struct *mm) argument 190 on_each_cpu_mask(mm_cpumask(mm), ipi_flush_tlb_mm, mm, 1); local 192 __flush_tlb_mm(mm); 193 broadcast_tlb_mm_a15_erratum(mm);
|
/linux-master/arch/sh/mm/ |
H A D | mmap.c | 2 * arch/sh/mm/mmap.c 11 #include <linux/mm.h> 12 #include <linux/sched/mm.h> 57 struct mm_struct *mm = current->mm; local 85 vma = find_vma(mm, addr); 106 struct mm_struct *mm = current->mm; local 135 vma = find_vma(mm, addr); 144 info.high_limit = mm [all...] |
/linux-master/arch/nios2/mm/ |
H A D | fault.c | 5 * based on arch/mips/mm/fault.c which is: 24 #include <linux/mm.h> 48 struct mm_struct *mm = tsk->mm; local 81 if (faulthandler_disabled() || !mm) 90 vma = lock_mm_and_find_vma(mm, address, regs); 150 * No need to mmap_read_unlock(mm) as we would 152 * in mm/filemap.c. 158 mmap_read_unlock(mm); 166 mmap_read_unlock(mm); [all...] |
H A D | init.c | 8 * based on arch/m68k/mm/init.c 23 #include <linux/mm.h> 107 struct mm_struct *mm = current->mm; local 110 mmap_write_lock(mm); 113 ret = install_special_mapping(mm, KUSER_BASE, KUSER_SIZE, 117 mmap_write_unlock(mm);
|
/linux-master/arch/alpha/mm/ |
H A D | fault.c | 3 * linux/arch/alpha/mm/fault.c 10 #include <linux/mm.h> 89 struct mm_struct *mm = current->mm; local 111 if (!mm || faulthandler_disabled()) 122 vma = lock_mm_and_find_vma(mm, address, regs); 170 /* No need to mmap_read_unlock(mm) as we would 172 * in mm/filemap.c. 178 mmap_read_unlock(mm); 185 mmap_read_unlock(mm); [all...] |
/linux-master/arch/alpha/include/asm/ |
H A D | mmu_context.h | 105 * the "mm->context") are the ASN _version_ code. A version of 0 is 107 * need to do "p->mm->context = 0". 121 __get_new_mm_context(struct mm_struct *mm, long cpu) argument 160 a new mm->context (via flush_tlb_mm) without the ASN serial 170 optimize for switching between threads -- if the mm is 196 struct mm_struct * mm = current->active_mm; \ 198 if (!mm->context[cpu]) \ 199 __load_new_mm_context(mm); \ 234 init_new_context(struct task_struct *tsk, struct mm_struct *mm) argument 239 mm 248 enter_lazy_tlb(struct mm_struct *mm, struct task_struct *tsk) argument [all...] |
/linux-master/drivers/misc/cxl/ |
H A D | cxllib.c | 7 #include <linux/sched/mm.h> 184 struct mm_struct *mm = get_task_mm(task); local 185 if (mm == NULL) 191 attr->pid = mm->context.id; 192 mmput(mm); 202 static int get_vma_info(struct mm_struct *mm, u64 addr, argument 209 mmap_read_lock(mm); 211 vma = find_vma(mm, addr); 220 mmap_read_unlock(mm); 224 int cxllib_handle_fault(struct mm_struct *mm, u6 argument [all...] |
/linux-master/arch/arm/include/asm/ |
H A D | cacheflush.h | 10 #include <linux/mm.h> 29 * The arch/arm/mm/cache-*.S and arch/arm/mm/proc-*.S files 218 static inline void vivt_flush_cache_mm(struct mm_struct *mm) argument 220 if (cpumask_test_cpu(smp_processor_id(), mm_cpumask(mm))) 227 struct mm_struct *mm = vma->vm_mm; local 229 if (!mm || cpumask_test_cpu(smp_processor_id(), mm_cpumask(mm))) 237 struct mm_struct *mm = vma->vm_mm; local 239 if (!mm || cpumask_test_cp [all...] |
/linux-master/arch/parisc/mm/ |
H A D | fault.c | 13 #include <linux/mm.h> 272 struct mm_struct *mm; local 279 mm = tsk->mm; 280 if (!mm) { 294 mmap_read_lock(mm); 295 vma = find_vma_prev(mm, address, &prev_vma); 299 vma = expand_stack(mm, address); 349 * No need to mmap_read_unlock(mm) as we would 351 * in mm/filema 459 struct mm_struct *mm; local [all...] |
/linux-master/arch/um/kernel/ |
H A D | tlb.c | 6 #include <linux/mm.h> 41 struct mm_struct *mm; member in struct:host_vm_change 46 #define INIT_HVC(mm, force, userspace) \ 49 .mm = mm, \ 73 ret = map(&hvc->mm->context.id, op->u.mmap.addr, 84 ret = unmap(&hvc->mm->context.id, 96 ret = protect(&hvc->mm->context.id, 317 static void fix_range_common(struct mm_struct *mm, unsigned long start_addr, argument 325 hvc = INIT_HVC(mm, forc 353 struct mm_struct *mm; local 462 struct mm_struct *mm = vma->vm_mm; local 557 fix_range(struct mm_struct *mm, unsigned long start_addr, unsigned long end_addr, int force) argument 579 flush_tlb_mm_range(struct mm_struct *mm, unsigned long start, unsigned long end) argument 585 flush_tlb_mm(struct mm_struct *mm) argument 596 struct mm_struct *mm = current->mm; local [all...] |
/linux-master/arch/openrisc/mm/ |
H A D | tlb.c | 22 #include <linux/mm.h> 120 * Invalidate the selected mm context only. 123 * This should be changed to loop over over mm and call flush_tlb_range. 126 void local_flush_tlb_mm(struct mm_struct *mm) argument 129 /* Was seeing bugs with the mm struct passed to us. Scrapped most of 169 int init_new_context(struct task_struct *tsk, struct mm_struct *mm) argument 171 mm->context = NO_CONTEXT; 176 * destroying the mm itself. this is only called when the last user of the mm 180 void destroy_context(struct mm_struct *mm) argument [all...] |