/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-cris/ |
H A D | tlb.h | 8 * per-vma handling.. 10 #define tlb_start_vma(tlb, vma) do { } while (0) 11 #define tlb_end_vma(tlb, vma) do { } while (0)
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-frv/ |
H A D | cacheflush.h | 25 #define flush_cache_page(vma, vmaddr, pfn) do {} while(0) 76 extern void flush_icache_user_range(struct vm_area_struct *vma, struct page *page, 79 static inline void flush_icache_user_range(struct vm_area_struct *vma, struct page *page, argument 86 static inline void flush_icache_page(struct vm_area_struct *vma, struct page *page) argument 88 flush_icache_user_range(vma, page, page_to_phys(page), PAGE_SIZE); 95 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 98 flush_icache_user_range((vma), (page), (vaddr), (len)); \ 101 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
|
H A D | tlb.h | 13 * we don't need any special per-pte or per-vma handling... 15 #define tlb_start_vma(tlb, vma) do { } while (0) 16 #define tlb_end_vma(tlb, vma) do { } while (0)
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-xtensa/ |
H A D | tlb.h | 14 #define tlb_start_vma(tlb,vma) do { } while (0) 15 #define tlb_end_vma(tlb,vma) do { } while (0)
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-alpha/ |
H A D | tlb.h | 4 #define tlb_start_vma(tlb, vma) do { } while (0) 5 #define tlb_end_vma(tlb, vma) do { } while (0)
|
H A D | tlbflush.h | 37 struct vm_area_struct *vma, 41 if (vma->vm_flags & VM_EXEC) { 50 struct vm_area_struct *vma, 53 if (vma->vm_flags & VM_EXEC) 126 flush_tlb_page(struct vm_area_struct *vma, unsigned long addr) argument 128 struct mm_struct *mm = vma->vm_mm; 131 flush_tlb_current_page(mm, vma, addr); 139 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, argument 142 flush_tlb_mm(vma->vm_mm); 36 ev4_flush_tlb_current_page(struct mm_struct * mm, struct vm_area_struct *vma, unsigned long addr) argument 49 ev5_flush_tlb_current_page(struct mm_struct * mm, struct vm_area_struct *vma, unsigned long addr) argument
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-arm/ |
H A D | tlb.h | 69 * In the case of tlb vma handling, we can optimise these away in the 74 tlb_start_vma(struct mmu_gather *tlb, struct vm_area_struct *vma) argument 77 flush_cache_range(vma, vma->vm_start, vma->vm_end); 81 tlb_end_vma(struct mmu_gather *tlb, struct vm_area_struct *vma) argument 84 flush_tlb_range(vma, vma->vm_start, vma->vm_end);
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-avr32/ |
H A D | tlbflush.h | 19 * - flush_tlb_page(vma, vmaddr) flushes one page 20 * - flush_tlb_range(vma, start, end) flushes a range of pages 27 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, 29 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long page);
|
H A D | cacheflush.h | 91 #define flush_cache_range(vma, start, end) do { } while (0) 92 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) 101 * #define flush_icache_page(vma, page) do { } while (0) 103 extern void flush_icache_page(struct vm_area_struct *vma, struct page *page); 119 extern void flush_icache_user_range(struct vm_area_struct *vma, 123 #define copy_to_user_page(vma, page, vaddr, dst, src, len) do { \ 125 flush_icache_user_range(vma, page, vaddr, len); \ 127 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-sh/ |
H A D | cacheflush.h | 18 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 20 flush_cache_page(vma, vaddr, page_to_pfn(page));\ 22 flush_icache_user_range(vma, page, vaddr, len); \ 25 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ 27 flush_cache_page(vma, vaddr, page_to_pfn(page));\
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/mm/ |
H A D | fremap.c | 23 static int zap_pte(struct mm_struct *mm, struct vm_area_struct *vma, argument 30 flush_cache_page(vma, addr, pte_pfn(pte)); 31 pte = ptep_clear_flush(vma, addr, ptep); 32 page = vm_normal_page(vma, addr, pte); 36 page_remove_rmap(page, vma); 51 int install_page(struct mm_struct *mm, struct vm_area_struct *vma, argument 70 inode = vma->vm_file->f_mapping->host; 78 if (pte_none(*pte) || !zap_pte(mm, vma, addr, pte)) 81 flush_icache_page(vma, page); 85 update_mmu_cache(vma, add 99 install_file_pte(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, unsigned long pgoff, pgprot_t prot) argument 153 struct vm_area_struct *vma; local [all...] |
H A D | rmap.c | 57 int anon_vma_prepare(struct vm_area_struct *vma) argument 59 struct anon_vma *anon_vma = vma->anon_vma; 63 struct mm_struct *mm = vma->vm_mm; 66 anon_vma = find_mergeable_anon_vma(vma); 81 if (likely(!vma->anon_vma)) { 82 vma->anon_vma = anon_vma; 83 list_add_tail(&vma->anon_vma_node, &anon_vma->head); 96 void __anon_vma_merge(struct vm_area_struct *vma, struct vm_area_struct *next) argument 98 BUG_ON(vma->anon_vma != next->anon_vma); 102 void __anon_vma_link(struct vm_area_struct *vma) argument 110 anon_vma_link(struct vm_area_struct *vma) argument 121 anon_vma_unlink(struct vm_area_struct *vma) argument 189 vma_address(struct page *page, struct vm_area_struct *vma) argument 207 page_address_in_vma(struct page *page, struct vm_area_struct *vma) argument 269 page_referenced_one(struct page *page, struct vm_area_struct *vma, unsigned int *mapcount) argument 305 struct vm_area_struct *vma; local 339 struct vm_area_struct *vma; local 415 page_mkclean_one(struct page *page, struct vm_area_struct *vma) argument 451 struct vm_area_struct *vma; local 492 __page_set_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) argument 516 __page_check_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) argument 547 page_add_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) argument 568 page_add_new_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) argument 599 page_dup_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) argument 614 page_remove_rmap(struct page *page, struct vm_area_struct *vma) argument 653 try_to_unmap_one(struct page *page, struct vm_area_struct *vma, int migration) argument 765 try_to_unmap_cluster(unsigned long cursor, unsigned int *mapcount, struct vm_area_struct *vma) argument 835 struct vm_area_struct *vma; local 865 struct vm_area_struct *vma; local [all...] |
H A D | mremap.c | 68 static void move_ptes(struct vm_area_struct *vma, pmd_t *old_pmd, argument 74 struct mm_struct *mm = vma->vm_mm; 78 if (vma->vm_file) { 82 * since it might clean the dst vma before the src vma, 85 mapping = vma->vm_file->f_mapping; 88 new_vma->vm_truncate_count != vma->vm_truncate_count) 107 pte = ptep_clear_flush(vma, old_addr, old_pte); 123 static unsigned long move_page_tables(struct vm_area_struct *vma, argument 131 flush_cache_range(vma, old_add 157 move_vma(struct vm_area_struct *vma, unsigned long old_addr, unsigned long old_len, unsigned long new_len, unsigned long new_addr) argument 254 struct vm_area_struct *vma; local [all...] |
H A D | memory.c | 269 void free_pgtables(struct mmu_gather **tlb, struct vm_area_struct *vma, argument 272 while (vma) { 273 struct vm_area_struct *next = vma->vm_next; 274 unsigned long addr = vma->vm_start; 277 * Hide vma from rmap and vmtruncate before freeing pgtables 279 anon_vma_unlink(vma); 280 unlink_file_vma(vma); 282 if (is_vm_hugetlb_page(vma)) { 283 hugetlb_free_pgd_range(tlb, addr, vma->vm_end, 289 while (next && next->vm_start <= vma 353 print_bad_pte(struct vm_area_struct *vma, pte_t pte, unsigned long vaddr) argument 388 vm_normal_page(struct vm_area_struct *vma, unsigned long addr, pte_t pte) argument 428 copy_one_pte(struct mm_struct *dst_mm, struct mm_struct *src_mm, pte_t *dst_pte, pte_t *src_pte, struct vm_area_struct *vma, unsigned long addr, int *rss) argument 492 copy_pte_range(struct mm_struct *dst_mm, struct mm_struct *src_mm, pmd_t *dst_pmd, pmd_t *src_pmd, struct vm_area_struct *vma, unsigned long addr, unsigned long end) argument 542 copy_pmd_range(struct mm_struct *dst_mm, struct mm_struct *src_mm, pud_t *dst_pud, pud_t *src_pud, struct vm_area_struct *vma, unsigned long addr, unsigned long end) argument 564 copy_pud_range(struct mm_struct *dst_mm, struct mm_struct *src_mm, pgd_t *dst_pgd, pgd_t *src_pgd, struct vm_area_struct *vma, unsigned long addr, unsigned long end) argument 586 copy_page_range(struct mm_struct *dst_mm, struct mm_struct *src_mm, struct vm_area_struct *vma) argument 621 zap_pte_range(struct mmu_gather *tlb, struct vm_area_struct *vma, pmd_t *pmd, unsigned long addr, unsigned long end, long *zap_work, struct zap_details *details) argument 706 zap_pmd_range(struct mmu_gather *tlb, struct vm_area_struct *vma, pud_t *pud, unsigned long addr, unsigned long end, long *zap_work, struct zap_details *details) argument 728 zap_pud_range(struct mmu_gather *tlb, struct vm_area_struct *vma, pgd_t *pgd, unsigned long addr, unsigned long end, long *zap_work, struct zap_details *details) argument 750 unmap_page_range(struct mmu_gather *tlb, struct vm_area_struct *vma, unsigned long addr, unsigned long end, long *zap_work, struct zap_details *details) argument 811 unmap_vmas(struct mmu_gather **tlbp, struct vm_area_struct *vma, unsigned long start_addr, unsigned long end_addr, unsigned long *nr_accounted, struct zap_details *details) argument 883 zap_page_range(struct vm_area_struct *vma, unsigned long address, unsigned long size, struct zap_details *details) argument 903 follow_page(struct vm_area_struct *vma, unsigned long address, unsigned int flags) argument 995 struct vm_area_struct *vma; local 1176 zeromap_page_range(struct vm_area_struct *vma, unsigned long addr, unsigned long size, pgprot_t prot) argument 1269 vm_insert_page(struct vm_area_struct *vma, unsigned long addr, struct page *page) argument 1292 vm_insert_pfn(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn) argument 1401 remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) argument 1576 maybe_mkwrite(pte_t pte, struct vm_area_struct *vma) argument 1583 cow_user_page(struct page *dst, struct page *src, unsigned long va, struct vm_area_struct *vma) argument 1629 do_wp_page(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *page_table, pmd_t *pmd, spinlock_t *ptl, pte_t orig_pte) argument 1812 struct vm_area_struct *vma; local 1821 unmap_mapping_range_vma(struct vm_area_struct *vma, unsigned long start_addr, unsigned long end_addr, struct zap_details *details) argument 1865 struct vm_area_struct *vma; local 1897 struct vm_area_struct *vma; local 2069 swapin_readahead(swp_entry_t entry, unsigned long addr,struct vm_area_struct *vma) argument 2119 do_swap_page(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *page_table, pmd_t *pmd, int write_access, pte_t orig_pte) argument 2220 do_anonymous_page(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *page_table, pmd_t *pmd, int write_access) argument 2289 do_no_page(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *page_table, pmd_t *pmd, int write_access) argument 2439 do_no_pfn(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *page_table, pmd_t *pmd, int write_access) argument 2482 do_file_page(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *page_table, pmd_t *pmd, int write_access, pte_t orig_pte) argument 2524 handle_pte_fault(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *pte, pmd_t *pmd, int write_access) argument 2585 __handle_mm_fault(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, int write_access) argument 2668 struct vm_area_struct * vma; local 2776 struct vm_area_struct *vma; local [all...] |
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/fs/proc/ |
H A D | task_mmu.c | 77 struct vm_area_struct * vma; local 90 vma = mm->mmap; 91 while (vma) { 92 if ((vma->vm_flags & VM_EXECUTABLE) && vma->vm_file) 94 vma = vma->vm_next; 97 if (vma) { 98 *mnt = mntget(vma->vm_file->f_path.mnt); 99 *dentry = dget(vma 128 struct vm_area_struct *vma; member in struct:pmd_walker 138 struct vm_area_struct *vma = v; local 221 smaps_pte_range(struct vm_area_struct *vma, pmd_t *pmd, unsigned long addr, unsigned long end, void *private) argument 261 clear_refs_pte_range(struct vm_area_struct *vma, pmd_t *pmd, unsigned long addr, unsigned long end, void *private) argument 326 walk_page_range(struct vm_area_struct *vma, void (*action)(struct vm_area_struct *, pmd_t *, unsigned long, unsigned long, void *), void *private) argument 353 struct vm_area_struct *vma = v; local 364 struct vm_area_struct *vma; local 379 struct vm_area_struct *vma, *tail_vma = NULL; local 439 vma_stop(struct proc_maps_private *priv, struct vm_area_struct *vma) argument 451 struct vm_area_struct *vma = v; local 464 struct vm_area_struct *vma = v; local [all...] |
H A D | task_nommu.c | 22 if (!vml->vma) 27 atomic_read(&vml->vma->vm_usage) > 1 29 sbytes += kobjsize((void *) vml->vma->vm_start); 30 sbytes += kobjsize(vml->vma); 32 bytes += kobjsize((void *) vml->vma->vm_start); 33 bytes += kobjsize(vml->vma); 34 slack += kobjsize((void *) vml->vma->vm_start) - 35 (vml->vma->vm_end - vml->vma->vm_start); 78 if (tbp->vma) 110 struct vm_area_struct *vma; local [all...] |
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-powerpc/ |
H A D | cacheflush.h | 22 #define flush_cache_range(vma, start, end) do { } while (0) 23 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) 24 #define flush_icache_page(vma, page) do { } while (0) 39 extern void flush_icache_user_range(struct vm_area_struct *vma, 58 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 61 flush_icache_user_range(vma, page, vaddr, len); \ 63 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
|
H A D | tlbflush.h | 7 * - flush_tlb_page(vma, vmaddr) flushes one page 8 * - flush_tlb_page_nohash(vma, vmaddr) flushes one page if SW loaded TLB 9 * - flush_tlb_range(vma, start, end) flushes a range of pages 45 static inline void flush_tlb_page(struct vm_area_struct *vma, argument 51 static inline void flush_tlb_page_nohash(struct vm_area_struct *vma, argument 57 static inline void flush_tlb_range(struct vm_area_struct *vma, argument 77 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr); 78 extern void flush_tlb_page_nohash(struct vm_area_struct *vma, unsigned long addr); 79 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, 138 static inline void flush_tlb_page(struct vm_area_struct *vma, argument 143 flush_tlb_page_nohash(struct vm_area_struct *vma, unsigned long vmaddr) argument 148 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument [all...] |
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-ia64/ |
H A D | cacheflush.h | 22 #define flush_cache_range(vma, start, end) do { } while (0) 23 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) 24 #define flush_icache_page(vma,page) do { } while (0) 38 #define flush_icache_user_range(vma, page, user_addr, len) \ 44 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 46 flush_icache_user_range(vma, page, vaddr, len); \ 48 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/arch/arm/mm/ |
H A D | mmap.c | 29 struct vm_area_struct *vma; local 69 vma = find_vma(mm, addr); 71 (!vma || addr + len <= vma->vm_start)) 87 for (vma = find_vma(mm, addr); ; vma = vma->vm_next) { 88 /* At this point: (!vma || addr < vma->vm_end). */ 101 if (!vma || add [all...] |
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/arch/i386/mm/ |
H A D | hugetlbpage.c | 20 struct vm_area_struct *vma, 33 vma->vm_flags != svma->vm_flags || 40 static int vma_shareable(struct vm_area_struct *vma, unsigned long addr) argument 48 if (vma->vm_flags & VM_MAYSHARE && 49 vma->vm_start <= base && end <= vma->vm_end) 59 struct vm_area_struct *vma = find_vma(mm, addr); local 60 struct address_space *mapping = vma->vm_file->f_mapping; 61 pgoff_t idx = ((addr - vma->vm_start) >> PAGE_SHIFT) + 62 vma 19 page_table_shareable(struct vm_area_struct *svma, struct vm_area_struct *vma, unsigned long addr, pgoff_t idx) argument 191 struct vm_area_struct *vma; local 233 struct vm_area_struct *vma, *prev_vma; local 322 struct vm_area_struct *vma; local [all...] |
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/arch/arm/mach-netx/ |
H A D | fb.h | 22 int netx_clcd_mmap(struct clcd_fb *fb, struct vm_area_struct *vma);
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-v850/ |
H A D | tlbflush.h | 48 static inline void flush_tlb_page(struct vm_area_struct *vma, unsigned long addr) argument 53 static inline void flush_tlb_range(struct vm_area_struct *vma, argument
|
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/linux/ |
H A D | rmap.h | 19 * directly to a vma: instead it points to an anon_vma, on whose list 22 * After unlinking the last vma on the list, we must garbage collect 24 * pointing to this anon_vma once its vma list is empty. 27 spinlock_t lock; /* Serialize access to vma list */ 45 static inline void anon_vma_lock(struct vm_area_struct *vma) argument 47 struct anon_vma *anon_vma = vma->anon_vma; 52 static inline void anon_vma_unlock(struct vm_area_struct *vma) argument 54 struct anon_vma *anon_vma = vma->anon_vma; 78 void page_dup_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address); 80 static inline void page_dup_rmap(struct page *page, struct vm_area_struct *vma, unsigne argument [all...] |
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/sound/arm/ |
H A D | devdma.h | 3 int devdma_mmap(struct device *dev, struct snd_pcm_substream *substream, struct vm_area_struct *vma);
|