Searched refs:vma (Results 1 - 25 of 271) sorted by relevance

1234567891011

/asus-wl-520gu-7.0.1.45/src/linux/linux/drivers/char/drm-4.0/
H A Dvm.c60 unsigned long drm_vm_nopage(struct vm_area_struct *vma, argument
65 struct page *drm_vm_nopage(struct vm_area_struct *vma,
74 unsigned long drm_vm_shm_nopage(struct vm_area_struct *vma, argument
79 struct page *drm_vm_shm_nopage(struct vm_area_struct *vma,
85 drm_map_t *map = (drm_map_t *)vma->vm_private_data;
87 drm_map_t *map = (drm_map_t *)vma->vm_pte;
92 if (address > vma->vm_end) return NOPAGE_SIGBUS; /* Disallow mremap */
95 offset = address - vma->vm_start;
108 unsigned long drm_vm_shm_nopage_lock(struct vm_area_struct *vma, argument
113 struct page *drm_vm_shm_nopage_lock(struct vm_area_struct *vma,
141 drm_vm_dma_nopage(struct vm_area_struct *vma, unsigned long address, int write_access) argument
175 drm_vm_open(struct vm_area_struct *vma) argument
205 drm_vm_close(struct vm_area_struct *vma) argument
237 drm_mmap_dma(struct file *filp, struct vm_area_struct *vma) argument
270 drm_mmap(struct file *filp, struct vm_area_struct *vma) argument
[all...]
/asus-wl-520gu-7.0.1.45/src/linux/linux/mm/
H A Dmlock.c15 static inline int mlock_fixup_all(struct vm_area_struct * vma, int newflags) argument
17 spin_lock(&vma->vm_mm->page_table_lock);
18 vma->vm_flags = newflags;
19 spin_unlock(&vma->vm_mm->page_table_lock);
23 static inline int mlock_fixup_start(struct vm_area_struct * vma, argument
31 *n = *vma;
39 vma->vm_pgoff += (end - vma->vm_start) >> PAGE_SHIFT;
40 lock_vma_mappings(vma);
41 spin_lock(&vma
49 mlock_fixup_end(struct vm_area_struct * vma, unsigned long start, int newflags) argument
75 mlock_fixup_middle(struct vm_area_struct * vma, unsigned long start, unsigned long end, int newflags) argument
117 mlock_fixup(struct vm_area_struct * vma, unsigned long start, unsigned long end, unsigned int newflags) argument
151 struct vm_area_struct * vma, * next; local
242 struct vm_area_struct * vma; local
[all...]
H A Dmprotect.c94 static inline int mprotect_fixup_all(struct vm_area_struct * vma, struct vm_area_struct ** pprev, argument
98 struct mm_struct * mm = vma->vm_mm;
100 if (prev && prev->vm_end == vma->vm_start && can_vma_merge(prev, newflags) &&
101 !vma->vm_file && !(vma->vm_flags & VM_SHARED)) {
103 prev->vm_end = vma->vm_end;
104 __vma_unlink(mm, vma, prev);
107 kmem_cache_free(vm_area_cachep, vma);
114 vma->vm_flags = newflags;
115 vma
123 mprotect_fixup_start(struct vm_area_struct * vma, struct vm_area_struct ** pprev, unsigned long end, int newflags, pgprot_t prot) argument
163 mprotect_fixup_end(struct vm_area_struct * vma, struct vm_area_struct ** pprev, unsigned long start, int newflags, pgprot_t prot) argument
194 mprotect_fixup_middle(struct vm_area_struct * vma, struct vm_area_struct ** pprev, unsigned long start, unsigned long end, int newflags, pgprot_t prot) argument
239 mprotect_fixup(struct vm_area_struct * vma, struct vm_area_struct ** pprev, unsigned long start, unsigned long end, unsigned int newflags) argument
270 struct vm_area_struct * vma, * next, * prev; local
[all...]
H A Dmmap.c97 static inline void __remove_shared_vm_struct(struct vm_area_struct *vma) argument
99 struct file * file = vma->vm_file;
103 if (vma->vm_flags & VM_DENYWRITE)
105 if(vma->vm_next_share)
106 vma->vm_next_share->vm_pprev_share = vma->vm_pprev_share;
107 *vma->vm_pprev_share = vma->vm_next_share;
111 static inline void remove_shared_vm_struct(struct vm_area_struct *vma) argument
113 lock_vma_mappings(vma);
118 lock_vma_mappings(struct vm_area_struct *vma) argument
129 unlock_vma_mappings(struct vm_area_struct *vma) argument
250 struct vm_area_struct * vma; local
282 __vma_link_list(struct mm_struct * mm, struct vm_area_struct * vma, struct vm_area_struct * prev, rb_node_t * rb_parent) argument
297 __vma_link_rb(struct mm_struct * mm, struct vm_area_struct * vma, rb_node_t ** rb_link, rb_node_t * rb_parent) argument
304 __vma_link_file(struct vm_area_struct * vma) argument
329 __vma_link(struct mm_struct * mm, struct vm_area_struct * vma, struct vm_area_struct * prev, rb_node_t ** rb_link, rb_node_t * rb_parent) argument
337 vma_link(struct mm_struct * mm, struct vm_area_struct * vma, struct vm_area_struct * prev, rb_node_t ** rb_link, rb_node_t * rb_parent) argument
397 struct vm_area_struct * vma, * prev; local
615 struct vm_area_struct *vma; local
661 struct vm_area_struct *vma = NULL; local
699 struct vm_area_struct * vma; local
742 struct vm_area_struct * vma; local
1032 struct vm_area_struct * vma, * prev; local
1109 struct vm_area_struct * vma; local
1170 __insert_vm_struct(struct mm_struct * mm, struct vm_area_struct * vma) argument
1183 insert_vm_struct(struct mm_struct * mm, struct vm_area_struct * vma) argument
[all...]
H A Dmremap.c125 static inline unsigned long move_vma(struct vm_area_struct * vma, argument
129 struct mm_struct * mm = vma->vm_mm;
137 can_vma_merge(prev, vma->vm_flags) && !vma->vm_file && !(vma->vm_flags & VM_SHARED)) {
154 can_vma_merge(next, vma->vm_flags) && !vma->vm_file && !(vma->vm_flags & VM_SHARED)) {
163 can_vma_merge(prev, vma->vm_flags) && !vma
218 struct vm_area_struct *vma; local
[all...]
H A Dfilemap.c1867 static void nopage_sequential_readahead(struct vm_area_struct * vma, argument
1872 ra_window = get_max_readahead(vma->vm_file->f_dentry->d_inode);
1876 if (vma->vm_raend == 0)
1877 vma->vm_raend = vma->vm_pgoff + ra_window;
1884 if ((pgoff + (ra_window >> 1)) == vma->vm_raend) {
1885 unsigned long start = vma->vm_pgoff + vma->vm_raend;
1888 if (end > ((vma->vm_end >> PAGE_SHIFT) + vma
2071 filemap_sync_pte(pte_t * ptep, struct vm_area_struct *vma, unsigned long address, unsigned int flags) argument
2086 filemap_sync_pte_range(pmd_t * pmd, unsigned long address, unsigned long size, struct vm_area_struct *vma, unsigned long offset, unsigned int flags) argument
2116 filemap_sync_pmd_range(pgd_t * pgd, unsigned long address, unsigned long size, struct vm_area_struct *vma, unsigned int flags) argument
2146 filemap_sync(struct vm_area_struct * vma, unsigned long address, size_t size, unsigned int flags) argument
2180 generic_file_mmap(struct file * file, struct vm_area_struct * vma) argument
2209 msync_interval(struct vm_area_struct * vma, unsigned long start, unsigned long end, int flags) argument
2247 struct vm_area_struct * vma; local
2303 setup_read_behavior(struct vm_area_struct * vma, int behavior) argument
2320 madvise_fixup_start(struct vm_area_struct * vma, unsigned long end, int behavior) argument
2347 madvise_fixup_end(struct vm_area_struct * vma, unsigned long start, int behavior) argument
2374 madvise_fixup_middle(struct vm_area_struct * vma, unsigned long start, unsigned long end, int behavior) argument
2420 madvise_behavior(struct vm_area_struct * vma, unsigned long start, unsigned long end, int behavior) argument
2449 madvise_willneed(struct vm_area_struct * vma, unsigned long start, unsigned long end) argument
2520 madvise_dontneed(struct vm_area_struct * vma, unsigned long start, unsigned long end) argument
2530 madvise_vma(struct vm_area_struct * vma, unsigned long start, unsigned long end, int behavior) argument
2595 struct vm_area_struct * vma; local
2660 mincore_page(struct vm_area_struct * vma, unsigned long pgoff) argument
2676 mincore_vma(struct vm_area_struct * vma, unsigned long start, unsigned long end, unsigned char * vec) argument
2748 struct vm_area_struct * vma; local
[all...]
/asus-wl-520gu-7.0.1.45/src/linux/linux/drivers/char/drm/
H A Ddrm_vm.h58 struct page *DRM(vm_nopage)(struct vm_area_struct *vma, argument
63 drm_file_t *priv = vma->vm_file->private_data;
79 if (map->offset == VM_OFFSET(vma)) break;
83 unsigned long offset = address - vma->vm_start;
84 unsigned long baddr = VM_OFFSET(vma) + offset;
125 struct page *DRM(vm_shm_nopage)(struct vm_area_struct *vma, argument
129 drm_map_t *map = (drm_map_t *)vma->vm_private_data;
134 if (address > vma->vm_end) return NOPAGE_SIGBUS; /* Disallow mremap */
137 offset = address - vma->vm_start;
151 void DRM(vm_shm_close)(struct vm_area_struct *vma) argument
223 vm_dma_nopage(struct vm_area_struct *vma, unsigned long address, int write_access) argument
248 vm_sg_nopage(struct vm_area_struct *vma, unsigned long address, int write_access) argument
275 vm_open(struct vm_area_struct *vma) argument
296 vm_close(struct vm_area_struct *vma) argument
321 mmap_dma(struct file *filp, struct vm_area_struct *vma) argument
361 mmap(struct file *filp, struct vm_area_struct *vma) argument
[all...]
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/mips/kernel/
H A Dsysmips.c34 struct vm_area_struct * vma; local
36 vma = find_vma(current->mm, address);
37 if (!vma || vma->vm_start > address || !(vma->vm_flags & VM_READ))
39 address = vma->vm_end - address;
42 if (vma->vm_next && vma->vm_next->vm_start == vma->vm_end &&
43 (vma
[all...]
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/ia64/mm/
H A Dfault.c31 expand_backing_store (struct vm_area_struct *vma, unsigned long address) argument
36 if (address - vma->vm_start > current->rlim[RLIMIT_STACK].rlim_cur
37 || (((vma->vm_mm->total_vm + grow) << PAGE_SHIFT) > current->rlim[RLIMIT_AS].rlim_cur))
39 vma->vm_end += PAGE_SIZE;
40 vma->vm_mm->total_vm += grow;
41 if (vma->vm_flags & VM_LOCKED)
42 vma->vm_mm->locked_vm += grow;
50 struct vm_area_struct *vma, *prev_vma; local
65 * should be no vma for region 5 addr's anyway, so skip getting the
74 vma
[all...]
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/ia64/ia32/
H A Dbinfmt_elf32.c61 ia32_install_shared_page (struct vm_area_struct *vma, unsigned long address, int no_share) argument
63 struct page *pg = ia32_shared_page[(address - vma->vm_start)/PAGE_SIZE];
76 struct vm_area_struct *vma; local
83 vma = kmem_cache_alloc(vm_area_cachep, SLAB_KERNEL);
84 if (vma) {
85 vma->vm_mm = current->mm;
86 vma->vm_start = IA32_GDT_OFFSET;
87 vma->vm_end = vma->vm_start + max(PAGE_SIZE, 2*IA32_PAGE_SIZE);
88 vma
[all...]
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/sh/mm/
H A Dfault.c37 struct vm_area_struct * vma; local
43 vma = find_vma(current->mm, start);
44 if (!vma)
46 if (vma->vm_start > start)
50 if (!(vma->vm_flags & VM_WRITE))
58 if (handle_mm_fault(current->mm, vma, start, 1) <= 0)
64 if (start < vma->vm_end)
66 vma = vma->vm_next;
67 if (!vma || vm
94 struct vm_area_struct * vma; local
282 update_mmu_cache(struct vm_area_struct * vma, unsigned long address, pte_t pte) argument
353 flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument
[all...]
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/i386/mm/
H A Dfault.c35 struct vm_area_struct * vma; local
41 vma = find_vma(current->mm, start);
42 if (!vma)
44 if (vma->vm_start > start)
48 if (!(vma->vm_flags & VM_WRITE))
58 int fault = handle_mm_fault(current->mm, vma, start, 1);
68 if (start < vma->vm_end)
70 vma = vma->vm_next;
71 if (!vma || vm
144 struct vm_area_struct * vma; local
[all...]
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/ppc64/mm/
H A Dfault.c68 struct vm_area_struct * vma; local
104 vma = find_vma(mm, address);
105 PPCDBG(PPCDBG_MM, "\tdo_page_fault: vma = 0x%16.16lx\n", vma);
106 if (!vma) {
107 PPCDBG(PPCDBG_MM, "\tdo_page_fault: !vma\n");
110 PPCDBG(PPCDBG_MM, "\tdo_page_fault: vma->vm_start = 0x%16.16lx, vma->vm_flags = 0x%16.16lx\n", vma->vm_start, vma
[all...]
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/parisc/mm/
H A Dfault.c126 struct vm_area_struct *vma, *prev_vma; local
136 vma = find_vma_prev(mm, address, &prev_vma);
137 if (!vma || address < vma->vm_start)
148 if ((vma->vm_flags & acc_type) != acc_type)
157 switch (handle_mm_fault(mm, vma, address, (acc_type & VM_WRITE) != 0)) {
178 vma = prev_vma;
179 if (vma && (expand_stack(vma, address) == 0))
195 if (vma) {
[all...]
/asus-wl-520gu-7.0.1.45/src/linux/linux/drivers/sbus/char/
H A Dflash.c38 flash_mmap(struct file *file, struct vm_area_struct *vma) argument
48 if ((vma->vm_flags & VM_READ) &&
49 (vma->vm_flags & VM_WRITE)) {
53 if (vma->vm_flags & VM_READ) {
56 } else if (vma->vm_flags & VM_WRITE) {
66 if ((vma->vm_pgoff << PAGE_SHIFT) > size)
68 addr += (vma->vm_pgoff << PAGE_SHIFT);
70 if (vma->vm_end - (vma->vm_start + (vma
[all...]
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/mips64/mm/
H A Dc-sb1.c234 static void sb1_flush_icache_page(struct vm_area_struct *vma, argument
237 if (!(vma->vm_flags & VM_EXEC)) {
308 static void local_sb1_flush_cache_page(struct vm_area_struct *vma, argument
311 if (!(vma->vm_flags & VM_EXEC))
321 struct vm_area_struct *vma; member in struct:flush_cache_page_struct
329 local_sb1_flush_cache_page(args->vma, args->page);
332 static void sb1_flush_cache_page(struct vm_area_struct *vma, unsigned long addr) argument
335 args.vma = vma;
337 local_sb1_flush_cache_page(vma, add
[all...]
H A Dtlb-andes.c118 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument
120 if (cpu_context(smp_processor_id(), vma->vm_mm) != 0) {
125 printk("[tlbpage<%d,%08lx>]", vma->vm_mm->context, page);
127 newpid = (cpu_context(smp_processor_id(), vma->vm_mm) &
148 static void andes_update_mmu_cache(struct vm_area_struct * vma, argument
161 if (current->active_mm != vma->vm_mm)
166 if ((pid != cpu_asid(cpu, vma->vm_mm))
167 || (cpu_context(cpu, vma->vm_mm) == 0)) {
170 __FUNCTION__, (int) cpu_asid(cpu, vma->vm_mm), pid);
176 pgdp = pgd_offset(vma
[all...]
/asus-wl-520gu-7.0.1.45/src/router/iproute2/reference/asm-m68k/
H A Dpgalloc.h100 extern inline void flush_cache_page(struct vm_area_struct *vma, argument
103 if (vma->vm_mm == current->mm)
129 #define flush_icache_page(vma,pg) do { } while (0)
130 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0)
/asus-wl-520gu-7.0.1.45/src/linux/linux/include/asm-m68k/
H A Dpgalloc.h100 extern inline void flush_cache_page(struct vm_area_struct *vma, argument
103 if (vma->vm_mm == current->mm)
129 #define flush_icache_page(vma,pg) do { } while (0)
130 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0)
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/x86_64/kernel/
H A Dsys_x86_64.c70 struct vm_area_struct *vma; local
87 for (vma = find_vma(current->mm, addr); ; vma = vma->vm_next) {
88 /* At this point: (!vma || addr < vma->vm_end). */
91 if (!vma || addr + len <= vma->vm_start)
93 addr = vma->vm_end;
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/alpha/mm/
H A Dfault.c87 struct vm_area_struct * vma; local
120 vma = find_vma(mm, address);
121 if (!vma)
123 if (vma->vm_start <= address)
125 if (!(vma->vm_flags & VM_GROWSDOWN))
127 if (expand_stack(vma, address))
136 if (!(vma->vm_flags & VM_EXEC))
140 if (!(vma->vm_flags & (VM_READ | VM_WRITE)))
143 if (!(vma->vm_flags & VM_WRITE))
153 fault = handle_mm_fault(mm, vma, addres
[all...]
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/m68k/mm/
H A Dfault.c88 struct vm_area_struct * vma; local
106 vma = find_vma(mm, address);
107 if (!vma)
109 if (vma->vm_flags & VM_IO)
111 if (vma->vm_start <= address)
113 if (!(vma->vm_flags & VM_GROWSDOWN))
123 if (expand_stack(vma, address))
139 if (!(vma->vm_flags & VM_WRITE))
146 if (!(vma->vm_flags & (VM_READ | VM_EXEC)))
157 fault = handle_mm_fault(mm, vma, addres
[all...]
/asus-wl-520gu-7.0.1.45/src/router/iproute2/reference/asm-arm/proc-armo/
H A Dcache.h15 #define flush_cache_page(vma,vmaddr) do { } while (0)
26 #define flush_icache_page(vma,page) do { } while (0)
36 * - flush_tlb_page(vma, vmaddr) flushes one page
43 #define flush_tlb_page(vma, vmaddr) do { } while (0)
88 update_mmu_cache(struct vm_area_struct *vma, unsigned long addr, pte_t pte) argument
90 struct mm_struct *mm = vma->vm_mm;
/asus-wl-520gu-7.0.1.45/src/linux/linux/include/asm-arm/proc-armo/
H A Dcache.h15 #define flush_cache_page(vma,vmaddr) do { } while (0)
26 #define flush_icache_page(vma,page) do { } while (0)
36 * - flush_tlb_page(vma, vmaddr) flushes one page
43 #define flush_tlb_page(vma, vmaddr) do { } while (0)
88 update_mmu_cache(struct vm_area_struct *vma, unsigned long addr, pte_t pte) argument
90 struct mm_struct *mm = vma->vm_mm;
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/parisc/kernel/
H A Dsys_parisc.c38 struct vm_area_struct *vma; local
44 for (vma = find_vma(current->mm, addr); ; vma = vma->vm_next) {
45 /* At this point: (!vma || addr < vma->vm_end). */
48 if (!vma || addr + len <= vma->vm_start)
50 addr = vma->vm_end;
59 struct vm_area_struct *vma, *first_vm local
[all...]

Completed in 302 milliseconds

1234567891011