Searched refs:mm (Results 276 - 300 of 2900) sorted by relevance

<<11121314151617181920>>

/linux-master/include/linux/
H A Dmm.h180 * per mm struct. Users can overwrite this number by sysctl but there is a
240 * The idea being to have a "virtual" mm in the same way
242 * mm details, and allowing different kinds of memory mappings
347 * support core mm.
453 /* This mask defines which mm->def_flags a process can inherit its parent */
626 * in mm/mempolicy.c will do this automatically.
683 * vma->vm_lock_seq under vma->vm_lock protection and mm->mm_lock_seq
712 * mm->mm_lock_seq can't be concurrently modified.
778 struct vm_area_struct *lock_vma_under_rcu(struct mm_struct *mm,
792 static inline struct vm_area_struct *lock_vma_under_rcu(struct mm_struct *mm, argument
821 vma_init(struct vm_area_struct *vma, struct mm_struct *mm) argument
2433 fixup_user_fault(struct mm_struct *mm, unsigned long address, unsigned int fault_flags, bool *unlocked) argument
2472 get_user_page_vma_remote(struct mm_struct *mm, unsigned long addr, int gup_flags, struct vm_area_struct **vmap) argument
2590 get_mm_counter(struct mm_struct *mm, int member) argument
2597 add_mm_counter(struct mm_struct *mm, int member, long value) argument
2604 inc_mm_counter(struct mm_struct *mm, int member) argument
2611 dec_mm_counter(struct mm_struct *mm, int member) argument
2633 get_mm_rss(struct mm_struct *mm) argument
2640 get_mm_hiwater_rss(struct mm_struct *mm) argument
2645 get_mm_hiwater_vm(struct mm_struct *mm) argument
2650 update_hiwater_rss(struct mm_struct *mm) argument
2658 update_hiwater_vm(struct mm_struct *mm) argument
2664 reset_mm_hiwater_rss(struct mm_struct *mm) argument
2669 setmax_mm_hiwater_rss(unsigned long *maxrss, struct mm_struct *mm) argument
2699 get_locked_pte(struct mm_struct *mm, unsigned long addr, spinlock_t **ptl) argument
2708 __p4d_alloc(struct mm_struct *mm, pgd_t *pgd, unsigned long address) argument
2718 __pud_alloc(struct mm_struct *mm, p4d_t *p4d, unsigned long address) argument
2723 mm_inc_nr_puds(struct mm_struct *mm) argument
2724 mm_dec_nr_puds(struct mm_struct *mm) argument
2729 mm_inc_nr_puds(struct mm_struct *mm) argument
2736 mm_dec_nr_puds(struct mm_struct *mm) argument
2745 __pmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long address) argument
2751 mm_inc_nr_pmds(struct mm_struct *mm) argument
2752 mm_dec_nr_pmds(struct mm_struct *mm) argument
2757 mm_inc_nr_pmds(struct mm_struct *mm) argument
2764 mm_dec_nr_pmds(struct mm_struct *mm) argument
2773 mm_pgtables_bytes_init(struct mm_struct *mm) argument
2778 mm_pgtables_bytes(const struct mm_struct *mm) argument
2783 mm_inc_nr_ptes(struct mm_struct *mm) argument
2788 mm_dec_nr_ptes(struct mm_struct *mm) argument
2794 mm_pgtables_bytes_init(struct mm_struct *mm) argument
2795 mm_pgtables_bytes(const struct mm_struct *mm) argument
2800 mm_inc_nr_ptes(struct mm_struct *mm) argument
2801 mm_dec_nr_ptes(struct mm_struct *mm) argument
2809 p4d_alloc(struct mm_struct *mm, pgd_t *pgd, unsigned long address) argument
2816 pud_alloc(struct mm_struct *mm, p4d_t *p4d, unsigned long address) argument
2823 pmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long address) argument
2911 pte_lockptr(struct mm_struct *mm, pmd_t *pmd) argument
2936 pte_lockptr(struct mm_struct *mm, pmd_t *pmd) argument
2973 pte_offset_map_lock(struct mm_struct *mm, pmd_t *pmd, unsigned long addr, spinlock_t **ptlp) argument
3016 pmd_lockptr(struct mm_struct *mm, pmd_t *pmd) argument
3041 pmd_lockptr(struct mm_struct *mm, pmd_t *pmd) argument
3053 pmd_lock(struct mm_struct *mm, pmd_t *pmd) argument
3086 pud_lockptr(struct mm_struct *mm, pud_t *pud) argument
3091 pud_lock(struct mm_struct *mm, pud_t *pud) argument
3476 vma_lookup(struct mm_struct *mm, unsigned long addr) argument
3522 find_exact_vma(struct mm_struct *mm, unsigned long vm_start, unsigned long vm_end) argument
3825 get_gate_vma(struct mm_struct *mm) argument
3830 in_gate_area(struct mm_struct *mm, unsigned long addr) argument
4173 madvise_set_anon_name(struct mm_struct *mm, unsigned long start, unsigned long len_in, struct anon_vma_name *anon_name) argument
[all...]
H A Dmm_types.h652 /* VMA covers [vm_start; vm_end) addresses within mm */
665 * Flags, see mm.h.
691 /* Flag to indicate areas detached from the mm->mm_mt tree */
854 struct list_head mmlist; /* List of maybe swapped mm's. These
891 * protecting pages mapped by this mm to enforce a later COW,
920 * user/owner of this mm. All of the following must be true in
923 * current == mm->owner
924 * current->mm != mm
925 * new_owner->mm
1023 mm_init_cpumask(struct mm_struct *mm) argument
1032 mm_cpumask(struct mm_struct *mm) argument
1054 lru_gen_init_mm(struct mm_struct *mm) argument
1063 lru_gen_use_mm(struct mm_struct *mm) argument
1075 lru_gen_add_mm(struct mm_struct *mm) argument
1079 lru_gen_del_mm(struct mm_struct *mm) argument
1083 lru_gen_migrate_mm(struct mm_struct *mm) argument
1087 lru_gen_init_mm(struct mm_struct *mm) argument
1091 lru_gen_use_mm(struct mm_struct *mm) argument
1111 vma_iter_init(struct vma_iterator *vmi, struct mm_struct *mm, unsigned long addr) argument
1150 mm_cidmask(struct mm_struct *mm) argument
1160 mm_init_cid(struct mm_struct *mm) argument
1173 mm_alloc_cid(struct mm_struct *mm) argument
1182 mm_destroy_cid(struct mm_struct *mm) argument
1193 mm_init_cid(struct mm_struct *mm) argument
1194 mm_alloc_cid(struct mm_struct *mm) argument
1195 mm_destroy_cid(struct mm_struct *mm) argument
[all...]
/linux-master/arch/sparc/mm/
H A Dsrmmu.c24 #include <linux/mm.h>
72 #define FLUSH_BEGIN(mm)
75 #define FLUSH_BEGIN(mm) if ((mm)->context != NO_CONTEXT) {
349 pgtable_t pte_alloc_one(struct mm_struct *mm) argument
354 if (!(ptep = pte_alloc_one_kernel(mm)))
357 spin_lock(&mm->page_table_lock);
363 spin_unlock(&mm->page_table_lock);
368 void pte_free(struct mm_struct *mm, pgtable_t ptep) argument
373 spin_lock(&mm
414 alloc_context(struct mm_struct *old_mm, struct mm_struct *mm) argument
472 switch_mm(struct mm_struct *old_mm, struct mm_struct *mm, struct task_struct *tsk) argument
999 init_new_context(struct task_struct *tsk, struct mm_struct *mm) argument
1005 destroy_context(struct mm_struct *mm) argument
1238 turbosparc_flush_cache_mm(struct mm_struct *mm) argument
1276 turbosparc_flush_sig_insns(struct mm_struct *mm, unsigned long insn_addr) argument
1290 turbosparc_flush_tlb_mm(struct mm_struct *mm) argument
1656 smp_flush_cache_mm(struct mm_struct *mm) argument
1668 smp_flush_tlb_mm(struct mm_struct *mm) argument
1688 struct mm_struct *mm = vma->vm_mm; local
1705 struct mm_struct *mm = vma->vm_mm; local
1720 struct mm_struct *mm = vma->vm_mm; local
1734 struct mm_struct *mm = vma->vm_mm; local
1760 smp_flush_sig_insns(struct mm_struct *mm, unsigned long insn_addr) argument
[all...]
H A Dfault_64.c3 * arch/sparc64/mm/fault.c: Page fault handlers for the 64-bit Sparc.
18 #include <linux/mm.h>
51 printk(KERN_ALERT "tsk->{mm,active_mm}->context = %016lx\n",
52 (tsk->mm ?
53 CTX_HWBITS(tsk->mm->context) :
55 printk(KERN_ALERT "tsk->{mm,active_mm}->pgd = %016lx\n",
56 (tsk->mm ? (unsigned long) tsk->mm->pgd :
81 pgd_t *pgdp = pgd_offset(current->mm, tpc);
270 struct mm_struct *mm local
[all...]
H A Dfault_32.c20 #include <linux/mm.h>
50 printk(KERN_ALERT "tsk->{mm,active_mm}->context = %08lx\n",
51 (tsk->mm ? tsk->mm->context : tsk->active_mm->context));
52 printk(KERN_ALERT "tsk->{mm,active_mm}->pgd = %08lx\n",
53 (tsk->mm ? (unsigned long) tsk->mm->pgd :
117 struct mm_struct *mm = tsk->mm; local
143 if (pagefault_disabled() || !mm)
309 struct mm_struct *mm = tsk->mm; local
[all...]
/linux-master/mm/
H A Dmremap.c3 * mm/mremap.c
11 #include <linux/mm.h>
35 static pud_t *get_old_pud(struct mm_struct *mm, unsigned long addr) argument
41 pgd = pgd_offset(mm, addr);
56 static pmd_t *get_old_pmd(struct mm_struct *mm, unsigned long addr) argument
61 pud = get_old_pud(mm, addr);
72 static pud_t *alloc_new_pud(struct mm_struct *mm, struct vm_area_struct *vma, argument
78 pgd = pgd_offset(mm, addr);
79 p4d = p4d_alloc(mm, pgd, addr);
83 return pud_alloc(mm, p4
86 alloc_new_pmd(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr) argument
141 struct mm_struct *mm = vma->vm_mm; local
240 struct mm_struct *mm = vma->vm_mm; local
308 struct mm_struct *mm = vma->vm_mm; local
357 struct mm_struct *mm = vma->vm_mm; local
657 struct mm_struct *mm = vma->vm_mm; local
816 struct mm_struct *mm = current->mm; local
873 struct mm_struct *mm = current->mm; local
979 struct mm_struct *mm = current->mm; local
[all...]
H A Dnommu.c3 * linux/mm/nommu.c
5 * Replacement code for mm functions to support CPU's that don't
8 * See Documentation/admin-guide/mm/nommu-mmap.rst
20 #include <linux/mm.h>
21 #include <linux/sched/mm.h>
101 vma = find_vma(current->mm, (unsigned long)objp);
172 mmap_write_lock(current->mm);
173 vma = find_vma(current->mm, (unsigned long)ret);
176 mmap_write_unlock(current->mm);
381 struct mm_struct *mm local
545 setup_vma_to_mm(struct vm_area_struct *vma, struct mm_struct *mm) argument
599 delete_vma(struct mm_struct *mm, struct vm_area_struct *vma) argument
609 find_vma_intersection(struct mm_struct *mm, unsigned long start_addr, unsigned long end_addr) argument
624 find_vma(struct mm_struct *mm, unsigned long addr) argument
636 lock_mm_and_find_vma(struct mm_struct *mm, unsigned long addr, struct pt_regs *regs) argument
657 expand_stack(struct mm_struct *mm, unsigned long addr) argument
667 find_vma_exact(struct mm_struct *mm, unsigned long addr, unsigned long len) argument
1312 struct mm_struct *mm; local
1426 do_munmap(struct mm_struct *mm, unsigned long start, size_t len, struct list_head *uf) argument
1490 struct mm_struct *mm = current->mm; local
1508 exit_mmap(struct mm_struct *mm) argument
1647 __access_remote_vm(struct mm_struct *mm, unsigned long addr, void *buf, int len, unsigned int gup_flags) argument
1691 access_remote_vm(struct mm_struct *mm, unsigned long addr, void *buf, int len, unsigned int gup_flags) argument
1704 struct mm_struct *mm; local
[all...]
H A Dprocess_vm_access.c3 * linux/mm/process_vm_access.c
9 #include <linux/mm.h>
12 #include <linux/sched/mm.h>
68 * @mm: mm for task
77 struct mm_struct *mm,
102 * access remotely because task/mm might not
103 * current/current->mm
105 mmap_read_lock(mm);
106 pinned_pages = pin_user_pages_remote(mm, p
73 process_vm_rw_single_vec(unsigned long addr, unsigned long len, struct iov_iter *iter, struct page **process_pages, struct mm_struct *mm, struct task_struct *task, int vm_write) argument
159 struct mm_struct *mm; local
[all...]
H A Dpgtable-generic.c3 * mm/pgtable-generic.c
97 struct mm_struct *mm = (vma)->vm_mm; local
99 pte = ptep_get_and_clear(mm, address, ptep);
100 if (pte_accessible(mm, pte))
165 void pgtable_trans_huge_deposit(struct mm_struct *mm, pmd_t *pmdp, argument
168 assert_spin_locked(pmd_lockptr(mm, pmdp));
171 if (!pmd_huge_pte(mm, pmdp))
174 list_add(&pgtable->lru, &pmd_huge_pte(mm, pmdp)->lru);
175 pmd_huge_pte(mm, pmdp) = pgtable;
181 pgtable_t pgtable_trans_huge_withdraw(struct mm_struct *mm, pmd_ argument
245 pte_free_defer(struct mm_struct *mm, pgtable_t pgtable) argument
306 pte_offset_map_nolock(struct mm_struct *mm, pmd_t *pmd, unsigned long addr, spinlock_t **ptlp) argument
362 __pte_offset_map_lock(struct mm_struct *mm, pmd_t *pmd, unsigned long addr, spinlock_t **ptlp) argument
[all...]
/linux-master/drivers/gpu/drm/xe/
H A Dxe_ttm_vram_mgr.c24 static inline bool xe_is_vram_mgr_blocks_contiguous(struct drm_buddy *mm, argument
36 size = drm_buddy_block_size(mm, block);
54 struct drm_buddy *mm = &mgr->mm; local
96 if (WARN_ON(min_page_size < mm->chunk_size)) {
140 err = drm_buddy_alloc_blocks(mm, (u64)place->fpfn << PAGE_SHIFT,
153 if (!drm_buddy_block_trim(mm, vres->base.size, &vres->blocks))
166 u64 end = start + drm_buddy_block_size(mm, block);
178 xe_is_vram_mgr_blocks_contiguous(mm, &vres->blocks))
199 drm_buddy_free_list(mm,
214 struct drm_buddy *mm = &mgr->mm; local
230 struct drm_buddy *mm = &mgr->mm; local
253 struct drm_buddy *mm = &mgr->mm; local
283 struct drm_buddy *mm = &mgr->mm; local
[all...]
/linux-master/arch/powerpc/mm/book3s64/
H A Dpgtable.c23 #include <mm/mmu_decl.h>
105 void set_pmd_at(struct mm_struct *mm, unsigned long addr, argument
115 assert_spin_locked(pmd_lockptr(mm, pmdp));
119 return set_pte_at(mm, addr, pmdp_ptep(pmdp), pmd_pte(pmd));
122 void set_pud_at(struct mm_struct *mm, unsigned long addr, argument
132 assert_spin_locked(pud_lockptr(mm, pudp));
136 return set_pte_at(mm, addr, pudp_ptep(pudp), pud_pte(pud));
143 struct mm_struct *mm = arg; local
144 exit_lazy_flush_tlb(mm, false);
158 void serialize_against_pte_lookup(struct mm_struct *mm) argument
161 smp_call_function_many(mm_cpumask(mm), do_serialize, mm, 1); local
363 get_pmd_from_cache(struct mm_struct *mm) argument
385 __alloc_for_pmdcache(struct mm_struct *mm) argument
426 pmd_fragment_alloc(struct mm_struct *mm, unsigned long vmaddr) argument
[all...]
/linux-master/arch/powerpc/kernel/
H A Dvdso.c11 #include <linux/mm.h>
69 current->mm->context.vdso = (void __user *)new_vma->vm_start;
117 struct mm_struct *mm = task->mm; local
118 VMA_ITERATOR(vmi, mm, 0);
121 mmap_read_lock(mm);
126 mmap_read_unlock(mm);
168 * vDSO and insert it into the mm struct tree
175 struct mm_struct *mm = current->mm; local
234 struct mm_struct *mm = current->mm; local
[all...]
/linux-master/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_mqd_manager_v10.c45 static void update_cu_mask(struct mqd_manager *mm, void *mqd, argument
54 mqd_symmetrically_map_cu_mask(mm,
88 static void init_mqd(struct mqd_manager *mm, void **mqd, argument
129 if (mm->dev->kfd->cwsr_enabled) {
145 mm->update_mqd(mm, m, q, NULL);
148 static int load_mqd(struct mqd_manager *mm, void *mqd, argument
156 r = mm->dev->kfd2kgd->hqd_load(mm->dev->adev, mqd, pipe_id, queue_id,
162 static void update_mqd(struct mqd_manager *mm, voi argument
234 get_wave_state(struct mqd_manager *mm, void *mqd, struct queue_properties *q, void __user *ctl_stack, u32 *ctl_stack_used_size, u32 *save_area_used_size) argument
272 checkpoint_mqd(struct mqd_manager *mm, void *mqd, void *mqd_dst, void *ctl_stack_dst) argument
281 restore_mqd(struct mqd_manager *mm, void **mqd, struct kfd_mem_obj *mqd_mem_obj, uint64_t *gart_addr, struct queue_properties *qp, const void *mqd_src, const void *ctl_stack_src, const u32 ctl_stack_size) argument
308 init_mqd_hiq(struct mqd_manager *mm, void **mqd, struct kfd_mem_obj *mqd_mem_obj, uint64_t *gart_addr, struct queue_properties *q) argument
322 destroy_hiq_mqd(struct mqd_manager *mm, void *mqd, enum kfd_preempt_type type, unsigned int timeout, uint32_t pipe_id, uint32_t queue_id) argument
342 init_mqd_sdma(struct mqd_manager *mm, void **mqd, struct kfd_mem_obj *mqd_mem_obj, uint64_t *gart_addr, struct queue_properties *q) argument
361 update_mqd_sdma(struct mqd_manager *mm, void *mqd, struct queue_properties *q, struct mqd_update_info *minfo) argument
388 checkpoint_mqd_sdma(struct mqd_manager *mm, void *mqd, void *mqd_dst, void *ctl_stack_dst) argument
400 restore_mqd_sdma(struct mqd_manager *mm, void **mqd, struct kfd_mem_obj *mqd_mem_obj, uint64_t *gart_addr, struct queue_properties *qp, const void *mqd_src, const void *ctl_stack_src, const u32 ctl_stack_size) argument
[all...]
/linux-master/arch/powerpc/mm/
H A Dhugetlbpage.c11 #include <linux/mm.h>
36 pte_t *huge_pte_offset(struct mm_struct *mm, unsigned long addr, unsigned long sz) argument
42 return __find_linux_pte(mm->pgd, addr, NULL, NULL);
45 static int __hugepte_alloc(struct mm_struct *mm, hugepd_t *hpdp, argument
67 new = kmem_cache_alloc(cachep, pgtable_gfp_flags(mm, GFP_KERNEL));
110 pte_t *huge_pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma, argument
123 pg = pgd_offset(mm, addr);
134 ptl = &mm->page_table_lock;
138 pu = pud_alloc(mm, p4, addr);
144 ptl = pud_lockptr(mm, p
[all...]
/linux-master/arch/arm/kernel/
H A Dsmp_tlb.c33 struct mm_struct *mm = (struct mm_struct *)arg; local
35 local_flush_tlb_mm(mm);
164 static void broadcast_tlb_mm_a15_erratum(struct mm_struct *mm) argument
173 a15_erratum_get_cpumask(this_cpu, mm, &mask);
187 void flush_tlb_mm(struct mm_struct *mm) argument
190 on_each_cpu_mask(mm_cpumask(mm), ipi_flush_tlb_mm, mm, 1); local
192 __flush_tlb_mm(mm);
193 broadcast_tlb_mm_a15_erratum(mm);
/linux-master/arch/sh/mm/
H A Dmmap.c2 * arch/sh/mm/mmap.c
11 #include <linux/mm.h>
12 #include <linux/sched/mm.h>
57 struct mm_struct *mm = current->mm; local
85 vma = find_vma(mm, addr);
106 struct mm_struct *mm = current->mm; local
135 vma = find_vma(mm, addr);
144 info.high_limit = mm
[all...]
/linux-master/arch/nios2/mm/
H A Dfault.c5 * based on arch/mips/mm/fault.c which is:
24 #include <linux/mm.h>
48 struct mm_struct *mm = tsk->mm; local
81 if (faulthandler_disabled() || !mm)
90 vma = lock_mm_and_find_vma(mm, address, regs);
150 * No need to mmap_read_unlock(mm) as we would
152 * in mm/filemap.c.
158 mmap_read_unlock(mm);
166 mmap_read_unlock(mm);
[all...]
H A Dinit.c8 * based on arch/m68k/mm/init.c
23 #include <linux/mm.h>
107 struct mm_struct *mm = current->mm; local
110 mmap_write_lock(mm);
113 ret = install_special_mapping(mm, KUSER_BASE, KUSER_SIZE,
117 mmap_write_unlock(mm);
/linux-master/arch/alpha/mm/
H A Dfault.c3 * linux/arch/alpha/mm/fault.c
10 #include <linux/mm.h>
89 struct mm_struct *mm = current->mm; local
111 if (!mm || faulthandler_disabled())
122 vma = lock_mm_and_find_vma(mm, address, regs);
170 /* No need to mmap_read_unlock(mm) as we would
172 * in mm/filemap.c.
178 mmap_read_unlock(mm);
185 mmap_read_unlock(mm);
[all...]
/linux-master/arch/alpha/include/asm/
H A Dmmu_context.h105 * the "mm->context") are the ASN _version_ code. A version of 0 is
107 * need to do "p->mm->context = 0".
121 __get_new_mm_context(struct mm_struct *mm, long cpu) argument
160 a new mm->context (via flush_tlb_mm) without the ASN serial
170 optimize for switching between threads -- if the mm is
196 struct mm_struct * mm = current->active_mm; \
198 if (!mm->context[cpu]) \
199 __load_new_mm_context(mm); \
234 init_new_context(struct task_struct *tsk, struct mm_struct *mm) argument
239 mm
248 enter_lazy_tlb(struct mm_struct *mm, struct task_struct *tsk) argument
[all...]
/linux-master/drivers/misc/cxl/
H A Dcxllib.c7 #include <linux/sched/mm.h>
184 struct mm_struct *mm = get_task_mm(task); local
185 if (mm == NULL)
191 attr->pid = mm->context.id;
192 mmput(mm);
202 static int get_vma_info(struct mm_struct *mm, u64 addr, argument
209 mmap_read_lock(mm);
211 vma = find_vma(mm, addr);
220 mmap_read_unlock(mm);
224 int cxllib_handle_fault(struct mm_struct *mm, u6 argument
[all...]
/linux-master/arch/arm/include/asm/
H A Dcacheflush.h10 #include <linux/mm.h>
29 * The arch/arm/mm/cache-*.S and arch/arm/mm/proc-*.S files
218 static inline void vivt_flush_cache_mm(struct mm_struct *mm) argument
220 if (cpumask_test_cpu(smp_processor_id(), mm_cpumask(mm)))
227 struct mm_struct *mm = vma->vm_mm; local
229 if (!mm || cpumask_test_cpu(smp_processor_id(), mm_cpumask(mm)))
237 struct mm_struct *mm = vma->vm_mm; local
239 if (!mm || cpumask_test_cp
[all...]
/linux-master/arch/parisc/mm/
H A Dfault.c13 #include <linux/mm.h>
272 struct mm_struct *mm; local
279 mm = tsk->mm;
280 if (!mm) {
294 mmap_read_lock(mm);
295 vma = find_vma_prev(mm, address, &prev_vma);
299 vma = expand_stack(mm, address);
349 * No need to mmap_read_unlock(mm) as we would
351 * in mm/filema
459 struct mm_struct *mm; local
[all...]
/linux-master/arch/um/kernel/
H A Dtlb.c6 #include <linux/mm.h>
41 struct mm_struct *mm; member in struct:host_vm_change
46 #define INIT_HVC(mm, force, userspace) \
49 .mm = mm, \
73 ret = map(&hvc->mm->context.id, op->u.mmap.addr,
84 ret = unmap(&hvc->mm->context.id,
96 ret = protect(&hvc->mm->context.id,
317 static void fix_range_common(struct mm_struct *mm, unsigned long start_addr, argument
325 hvc = INIT_HVC(mm, forc
353 struct mm_struct *mm; local
462 struct mm_struct *mm = vma->vm_mm; local
557 fix_range(struct mm_struct *mm, unsigned long start_addr, unsigned long end_addr, int force) argument
579 flush_tlb_mm_range(struct mm_struct *mm, unsigned long start, unsigned long end) argument
585 flush_tlb_mm(struct mm_struct *mm) argument
596 struct mm_struct *mm = current->mm; local
[all...]
/linux-master/arch/openrisc/mm/
H A Dtlb.c22 #include <linux/mm.h>
120 * Invalidate the selected mm context only.
123 * This should be changed to loop over over mm and call flush_tlb_range.
126 void local_flush_tlb_mm(struct mm_struct *mm) argument
129 /* Was seeing bugs with the mm struct passed to us. Scrapped most of
169 int init_new_context(struct task_struct *tsk, struct mm_struct *mm) argument
171 mm->context = NO_CONTEXT;
176 * destroying the mm itself. this is only called when the last user of the mm
180 void destroy_context(struct mm_struct *mm) argument
[all...]

Completed in 298 milliseconds

<<11121314151617181920>>