/asus-wl-520gu-7.0.1.45/src/linux/linux/drivers/char/drm-4.0/ |
H A D | vm.c | 60 unsigned long drm_vm_nopage(struct vm_area_struct *vma, argument 65 struct page *drm_vm_nopage(struct vm_area_struct *vma, 74 unsigned long drm_vm_shm_nopage(struct vm_area_struct *vma, argument 79 struct page *drm_vm_shm_nopage(struct vm_area_struct *vma, 85 drm_map_t *map = (drm_map_t *)vma->vm_private_data; 87 drm_map_t *map = (drm_map_t *)vma->vm_pte; 92 if (address > vma->vm_end) return NOPAGE_SIGBUS; /* Disallow mremap */ 95 offset = address - vma->vm_start; 108 unsigned long drm_vm_shm_nopage_lock(struct vm_area_struct *vma, argument 113 struct page *drm_vm_shm_nopage_lock(struct vm_area_struct *vma, 141 drm_vm_dma_nopage(struct vm_area_struct *vma, unsigned long address, int write_access) argument 175 drm_vm_open(struct vm_area_struct *vma) argument 205 drm_vm_close(struct vm_area_struct *vma) argument 237 drm_mmap_dma(struct file *filp, struct vm_area_struct *vma) argument 270 drm_mmap(struct file *filp, struct vm_area_struct *vma) argument [all...] |
/asus-wl-520gu-7.0.1.45/src/linux/linux/mm/ |
H A D | mlock.c | 15 static inline int mlock_fixup_all(struct vm_area_struct * vma, int newflags) argument 17 spin_lock(&vma->vm_mm->page_table_lock); 18 vma->vm_flags = newflags; 19 spin_unlock(&vma->vm_mm->page_table_lock); 23 static inline int mlock_fixup_start(struct vm_area_struct * vma, argument 31 *n = *vma; 39 vma->vm_pgoff += (end - vma->vm_start) >> PAGE_SHIFT; 40 lock_vma_mappings(vma); 41 spin_lock(&vma 49 mlock_fixup_end(struct vm_area_struct * vma, unsigned long start, int newflags) argument 75 mlock_fixup_middle(struct vm_area_struct * vma, unsigned long start, unsigned long end, int newflags) argument 117 mlock_fixup(struct vm_area_struct * vma, unsigned long start, unsigned long end, unsigned int newflags) argument 151 struct vm_area_struct * vma, * next; local 242 struct vm_area_struct * vma; local [all...] |
H A D | mprotect.c | 94 static inline int mprotect_fixup_all(struct vm_area_struct * vma, struct vm_area_struct ** pprev, argument 98 struct mm_struct * mm = vma->vm_mm; 100 if (prev && prev->vm_end == vma->vm_start && can_vma_merge(prev, newflags) && 101 !vma->vm_file && !(vma->vm_flags & VM_SHARED)) { 103 prev->vm_end = vma->vm_end; 104 __vma_unlink(mm, vma, prev); 107 kmem_cache_free(vm_area_cachep, vma); 114 vma->vm_flags = newflags; 115 vma 123 mprotect_fixup_start(struct vm_area_struct * vma, struct vm_area_struct ** pprev, unsigned long end, int newflags, pgprot_t prot) argument 163 mprotect_fixup_end(struct vm_area_struct * vma, struct vm_area_struct ** pprev, unsigned long start, int newflags, pgprot_t prot) argument 194 mprotect_fixup_middle(struct vm_area_struct * vma, struct vm_area_struct ** pprev, unsigned long start, unsigned long end, int newflags, pgprot_t prot) argument 239 mprotect_fixup(struct vm_area_struct * vma, struct vm_area_struct ** pprev, unsigned long start, unsigned long end, unsigned int newflags) argument 270 struct vm_area_struct * vma, * next, * prev; local [all...] |
H A D | mmap.c | 97 static inline void __remove_shared_vm_struct(struct vm_area_struct *vma) argument 99 struct file * file = vma->vm_file; 103 if (vma->vm_flags & VM_DENYWRITE) 105 if(vma->vm_next_share) 106 vma->vm_next_share->vm_pprev_share = vma->vm_pprev_share; 107 *vma->vm_pprev_share = vma->vm_next_share; 111 static inline void remove_shared_vm_struct(struct vm_area_struct *vma) argument 113 lock_vma_mappings(vma); 118 lock_vma_mappings(struct vm_area_struct *vma) argument 129 unlock_vma_mappings(struct vm_area_struct *vma) argument 250 struct vm_area_struct * vma; local 282 __vma_link_list(struct mm_struct * mm, struct vm_area_struct * vma, struct vm_area_struct * prev, rb_node_t * rb_parent) argument 297 __vma_link_rb(struct mm_struct * mm, struct vm_area_struct * vma, rb_node_t ** rb_link, rb_node_t * rb_parent) argument 304 __vma_link_file(struct vm_area_struct * vma) argument 329 __vma_link(struct mm_struct * mm, struct vm_area_struct * vma, struct vm_area_struct * prev, rb_node_t ** rb_link, rb_node_t * rb_parent) argument 337 vma_link(struct mm_struct * mm, struct vm_area_struct * vma, struct vm_area_struct * prev, rb_node_t ** rb_link, rb_node_t * rb_parent) argument 397 struct vm_area_struct * vma, * prev; local 615 struct vm_area_struct *vma; local 661 struct vm_area_struct *vma = NULL; local 699 struct vm_area_struct * vma; local 742 struct vm_area_struct * vma; local 1032 struct vm_area_struct * vma, * prev; local 1109 struct vm_area_struct * vma; local 1170 __insert_vm_struct(struct mm_struct * mm, struct vm_area_struct * vma) argument 1183 insert_vm_struct(struct mm_struct * mm, struct vm_area_struct * vma) argument [all...] |
H A D | mremap.c | 125 static inline unsigned long move_vma(struct vm_area_struct * vma, argument 129 struct mm_struct * mm = vma->vm_mm; 137 can_vma_merge(prev, vma->vm_flags) && !vma->vm_file && !(vma->vm_flags & VM_SHARED)) { 154 can_vma_merge(next, vma->vm_flags) && !vma->vm_file && !(vma->vm_flags & VM_SHARED)) { 163 can_vma_merge(prev, vma->vm_flags) && !vma 218 struct vm_area_struct *vma; local [all...] |
H A D | filemap.c | 1867 static void nopage_sequential_readahead(struct vm_area_struct * vma, argument 1872 ra_window = get_max_readahead(vma->vm_file->f_dentry->d_inode); 1876 if (vma->vm_raend == 0) 1877 vma->vm_raend = vma->vm_pgoff + ra_window; 1884 if ((pgoff + (ra_window >> 1)) == vma->vm_raend) { 1885 unsigned long start = vma->vm_pgoff + vma->vm_raend; 1888 if (end > ((vma->vm_end >> PAGE_SHIFT) + vma 2071 filemap_sync_pte(pte_t * ptep, struct vm_area_struct *vma, unsigned long address, unsigned int flags) argument 2086 filemap_sync_pte_range(pmd_t * pmd, unsigned long address, unsigned long size, struct vm_area_struct *vma, unsigned long offset, unsigned int flags) argument 2116 filemap_sync_pmd_range(pgd_t * pgd, unsigned long address, unsigned long size, struct vm_area_struct *vma, unsigned int flags) argument 2146 filemap_sync(struct vm_area_struct * vma, unsigned long address, size_t size, unsigned int flags) argument 2180 generic_file_mmap(struct file * file, struct vm_area_struct * vma) argument 2209 msync_interval(struct vm_area_struct * vma, unsigned long start, unsigned long end, int flags) argument 2247 struct vm_area_struct * vma; local 2303 setup_read_behavior(struct vm_area_struct * vma, int behavior) argument 2320 madvise_fixup_start(struct vm_area_struct * vma, unsigned long end, int behavior) argument 2347 madvise_fixup_end(struct vm_area_struct * vma, unsigned long start, int behavior) argument 2374 madvise_fixup_middle(struct vm_area_struct * vma, unsigned long start, unsigned long end, int behavior) argument 2420 madvise_behavior(struct vm_area_struct * vma, unsigned long start, unsigned long end, int behavior) argument 2449 madvise_willneed(struct vm_area_struct * vma, unsigned long start, unsigned long end) argument 2520 madvise_dontneed(struct vm_area_struct * vma, unsigned long start, unsigned long end) argument 2530 madvise_vma(struct vm_area_struct * vma, unsigned long start, unsigned long end, int behavior) argument 2595 struct vm_area_struct * vma; local 2660 mincore_page(struct vm_area_struct * vma, unsigned long pgoff) argument 2676 mincore_vma(struct vm_area_struct * vma, unsigned long start, unsigned long end, unsigned char * vec) argument 2748 struct vm_area_struct * vma; local [all...] |
/asus-wl-520gu-7.0.1.45/src/linux/linux/drivers/char/drm/ |
H A D | drm_vm.h | 58 struct page *DRM(vm_nopage)(struct vm_area_struct *vma, argument 63 drm_file_t *priv = vma->vm_file->private_data; 79 if (map->offset == VM_OFFSET(vma)) break; 83 unsigned long offset = address - vma->vm_start; 84 unsigned long baddr = VM_OFFSET(vma) + offset; 125 struct page *DRM(vm_shm_nopage)(struct vm_area_struct *vma, argument 129 drm_map_t *map = (drm_map_t *)vma->vm_private_data; 134 if (address > vma->vm_end) return NOPAGE_SIGBUS; /* Disallow mremap */ 137 offset = address - vma->vm_start; 151 void DRM(vm_shm_close)(struct vm_area_struct *vma) argument 223 vm_dma_nopage(struct vm_area_struct *vma, unsigned long address, int write_access) argument 248 vm_sg_nopage(struct vm_area_struct *vma, unsigned long address, int write_access) argument 275 vm_open(struct vm_area_struct *vma) argument 296 vm_close(struct vm_area_struct *vma) argument 321 mmap_dma(struct file *filp, struct vm_area_struct *vma) argument 361 mmap(struct file *filp, struct vm_area_struct *vma) argument [all...] |
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/mips/kernel/ |
H A D | sysmips.c | 34 struct vm_area_struct * vma; local 36 vma = find_vma(current->mm, address); 37 if (!vma || vma->vm_start > address || !(vma->vm_flags & VM_READ)) 39 address = vma->vm_end - address; 42 if (vma->vm_next && vma->vm_next->vm_start == vma->vm_end && 43 (vma [all...] |
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/ia64/mm/ |
H A D | fault.c | 31 expand_backing_store (struct vm_area_struct *vma, unsigned long address) argument 36 if (address - vma->vm_start > current->rlim[RLIMIT_STACK].rlim_cur 37 || (((vma->vm_mm->total_vm + grow) << PAGE_SHIFT) > current->rlim[RLIMIT_AS].rlim_cur)) 39 vma->vm_end += PAGE_SIZE; 40 vma->vm_mm->total_vm += grow; 41 if (vma->vm_flags & VM_LOCKED) 42 vma->vm_mm->locked_vm += grow; 50 struct vm_area_struct *vma, *prev_vma; local 65 * should be no vma for region 5 addr's anyway, so skip getting the 74 vma [all...] |
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/ia64/ia32/ |
H A D | binfmt_elf32.c | 61 ia32_install_shared_page (struct vm_area_struct *vma, unsigned long address, int no_share) argument 63 struct page *pg = ia32_shared_page[(address - vma->vm_start)/PAGE_SIZE]; 76 struct vm_area_struct *vma; local 83 vma = kmem_cache_alloc(vm_area_cachep, SLAB_KERNEL); 84 if (vma) { 85 vma->vm_mm = current->mm; 86 vma->vm_start = IA32_GDT_OFFSET; 87 vma->vm_end = vma->vm_start + max(PAGE_SIZE, 2*IA32_PAGE_SIZE); 88 vma [all...] |
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/sh/mm/ |
H A D | fault.c | 37 struct vm_area_struct * vma; local 43 vma = find_vma(current->mm, start); 44 if (!vma) 46 if (vma->vm_start > start) 50 if (!(vma->vm_flags & VM_WRITE)) 58 if (handle_mm_fault(current->mm, vma, start, 1) <= 0) 64 if (start < vma->vm_end) 66 vma = vma->vm_next; 67 if (!vma || vm 94 struct vm_area_struct * vma; local 282 update_mmu_cache(struct vm_area_struct * vma, unsigned long address, pte_t pte) argument 353 flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument [all...] |
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/i386/mm/ |
H A D | fault.c | 35 struct vm_area_struct * vma; local 41 vma = find_vma(current->mm, start); 42 if (!vma) 44 if (vma->vm_start > start) 48 if (!(vma->vm_flags & VM_WRITE)) 58 int fault = handle_mm_fault(current->mm, vma, start, 1); 68 if (start < vma->vm_end) 70 vma = vma->vm_next; 71 if (!vma || vm 144 struct vm_area_struct * vma; local [all...] |
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/ppc64/mm/ |
H A D | fault.c | 68 struct vm_area_struct * vma; local 104 vma = find_vma(mm, address); 105 PPCDBG(PPCDBG_MM, "\tdo_page_fault: vma = 0x%16.16lx\n", vma); 106 if (!vma) { 107 PPCDBG(PPCDBG_MM, "\tdo_page_fault: !vma\n"); 110 PPCDBG(PPCDBG_MM, "\tdo_page_fault: vma->vm_start = 0x%16.16lx, vma->vm_flags = 0x%16.16lx\n", vma->vm_start, vma [all...] |
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/parisc/mm/ |
H A D | fault.c | 126 struct vm_area_struct *vma, *prev_vma; local 136 vma = find_vma_prev(mm, address, &prev_vma); 137 if (!vma || address < vma->vm_start) 148 if ((vma->vm_flags & acc_type) != acc_type) 157 switch (handle_mm_fault(mm, vma, address, (acc_type & VM_WRITE) != 0)) { 178 vma = prev_vma; 179 if (vma && (expand_stack(vma, address) == 0)) 195 if (vma) { [all...] |
/asus-wl-520gu-7.0.1.45/src/linux/linux/drivers/sbus/char/ |
H A D | flash.c | 38 flash_mmap(struct file *file, struct vm_area_struct *vma) argument 48 if ((vma->vm_flags & VM_READ) && 49 (vma->vm_flags & VM_WRITE)) { 53 if (vma->vm_flags & VM_READ) { 56 } else if (vma->vm_flags & VM_WRITE) { 66 if ((vma->vm_pgoff << PAGE_SHIFT) > size) 68 addr += (vma->vm_pgoff << PAGE_SHIFT); 70 if (vma->vm_end - (vma->vm_start + (vma [all...] |
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/mips64/mm/ |
H A D | c-sb1.c | 234 static void sb1_flush_icache_page(struct vm_area_struct *vma, argument 237 if (!(vma->vm_flags & VM_EXEC)) { 308 static void local_sb1_flush_cache_page(struct vm_area_struct *vma, argument 311 if (!(vma->vm_flags & VM_EXEC)) 321 struct vm_area_struct *vma; member in struct:flush_cache_page_struct 329 local_sb1_flush_cache_page(args->vma, args->page); 332 static void sb1_flush_cache_page(struct vm_area_struct *vma, unsigned long addr) argument 335 args.vma = vma; 337 local_sb1_flush_cache_page(vma, add [all...] |
H A D | tlb-andes.c | 118 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page) argument 120 if (cpu_context(smp_processor_id(), vma->vm_mm) != 0) { 125 printk("[tlbpage<%d,%08lx>]", vma->vm_mm->context, page); 127 newpid = (cpu_context(smp_processor_id(), vma->vm_mm) & 148 static void andes_update_mmu_cache(struct vm_area_struct * vma, argument 161 if (current->active_mm != vma->vm_mm) 166 if ((pid != cpu_asid(cpu, vma->vm_mm)) 167 || (cpu_context(cpu, vma->vm_mm) == 0)) { 170 __FUNCTION__, (int) cpu_asid(cpu, vma->vm_mm), pid); 176 pgdp = pgd_offset(vma [all...] |
/asus-wl-520gu-7.0.1.45/src/router/iproute2/reference/asm-m68k/ |
H A D | pgalloc.h | 100 extern inline void flush_cache_page(struct vm_area_struct *vma, argument 103 if (vma->vm_mm == current->mm) 129 #define flush_icache_page(vma,pg) do { } while (0) 130 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0)
|
/asus-wl-520gu-7.0.1.45/src/linux/linux/include/asm-m68k/ |
H A D | pgalloc.h | 100 extern inline void flush_cache_page(struct vm_area_struct *vma, argument 103 if (vma->vm_mm == current->mm) 129 #define flush_icache_page(vma,pg) do { } while (0) 130 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0)
|
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/x86_64/kernel/ |
H A D | sys_x86_64.c | 70 struct vm_area_struct *vma; local 87 for (vma = find_vma(current->mm, addr); ; vma = vma->vm_next) { 88 /* At this point: (!vma || addr < vma->vm_end). */ 91 if (!vma || addr + len <= vma->vm_start) 93 addr = vma->vm_end;
|
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/alpha/mm/ |
H A D | fault.c | 87 struct vm_area_struct * vma; local 120 vma = find_vma(mm, address); 121 if (!vma) 123 if (vma->vm_start <= address) 125 if (!(vma->vm_flags & VM_GROWSDOWN)) 127 if (expand_stack(vma, address)) 136 if (!(vma->vm_flags & VM_EXEC)) 140 if (!(vma->vm_flags & (VM_READ | VM_WRITE))) 143 if (!(vma->vm_flags & VM_WRITE)) 153 fault = handle_mm_fault(mm, vma, addres [all...] |
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/m68k/mm/ |
H A D | fault.c | 88 struct vm_area_struct * vma; local 106 vma = find_vma(mm, address); 107 if (!vma) 109 if (vma->vm_flags & VM_IO) 111 if (vma->vm_start <= address) 113 if (!(vma->vm_flags & VM_GROWSDOWN)) 123 if (expand_stack(vma, address)) 139 if (!(vma->vm_flags & VM_WRITE)) 146 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) 157 fault = handle_mm_fault(mm, vma, addres [all...] |
/asus-wl-520gu-7.0.1.45/src/router/iproute2/reference/asm-arm/proc-armo/ |
H A D | cache.h | 15 #define flush_cache_page(vma,vmaddr) do { } while (0) 26 #define flush_icache_page(vma,page) do { } while (0) 36 * - flush_tlb_page(vma, vmaddr) flushes one page 43 #define flush_tlb_page(vma, vmaddr) do { } while (0) 88 update_mmu_cache(struct vm_area_struct *vma, unsigned long addr, pte_t pte) argument 90 struct mm_struct *mm = vma->vm_mm;
|
/asus-wl-520gu-7.0.1.45/src/linux/linux/include/asm-arm/proc-armo/ |
H A D | cache.h | 15 #define flush_cache_page(vma,vmaddr) do { } while (0) 26 #define flush_icache_page(vma,page) do { } while (0) 36 * - flush_tlb_page(vma, vmaddr) flushes one page 43 #define flush_tlb_page(vma, vmaddr) do { } while (0) 88 update_mmu_cache(struct vm_area_struct *vma, unsigned long addr, pte_t pte) argument 90 struct mm_struct *mm = vma->vm_mm;
|
/asus-wl-520gu-7.0.1.45/src/linux/linux/arch/parisc/kernel/ |
H A D | sys_parisc.c | 38 struct vm_area_struct *vma; local 44 for (vma = find_vma(current->mm, addr); ; vma = vma->vm_next) { 45 /* At this point: (!vma || addr < vma->vm_end). */ 48 if (!vma || addr + len <= vma->vm_start) 50 addr = vma->vm_end; 59 struct vm_area_struct *vma, *first_vm local [all...] |