Searched refs:vma (Results 51 - 75 of 485) sorted by relevance

1234567891011>>

/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-cris/
H A Dtlb.h8 * per-vma handling..
10 #define tlb_start_vma(tlb, vma) do { } while (0)
11 #define tlb_end_vma(tlb, vma) do { } while (0)
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-frv/
H A Dcacheflush.h25 #define flush_cache_page(vma, vmaddr, pfn) do {} while(0)
76 extern void flush_icache_user_range(struct vm_area_struct *vma, struct page *page,
79 static inline void flush_icache_user_range(struct vm_area_struct *vma, struct page *page, argument
86 static inline void flush_icache_page(struct vm_area_struct *vma, struct page *page) argument
88 flush_icache_user_range(vma, page, page_to_phys(page), PAGE_SIZE);
95 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \
98 flush_icache_user_range((vma), (page), (vaddr), (len)); \
101 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
H A Dtlb.h13 * we don't need any special per-pte or per-vma handling...
15 #define tlb_start_vma(tlb, vma) do { } while (0)
16 #define tlb_end_vma(tlb, vma) do { } while (0)
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-xtensa/
H A Dtlb.h14 #define tlb_start_vma(tlb,vma) do { } while (0)
15 #define tlb_end_vma(tlb,vma) do { } while (0)
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-alpha/
H A Dtlb.h4 #define tlb_start_vma(tlb, vma) do { } while (0)
5 #define tlb_end_vma(tlb, vma) do { } while (0)
H A Dtlbflush.h37 struct vm_area_struct *vma,
41 if (vma->vm_flags & VM_EXEC) {
50 struct vm_area_struct *vma,
53 if (vma->vm_flags & VM_EXEC)
126 flush_tlb_page(struct vm_area_struct *vma, unsigned long addr) argument
128 struct mm_struct *mm = vma->vm_mm;
131 flush_tlb_current_page(mm, vma, addr);
139 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, argument
142 flush_tlb_mm(vma->vm_mm);
36 ev4_flush_tlb_current_page(struct mm_struct * mm, struct vm_area_struct *vma, unsigned long addr) argument
49 ev5_flush_tlb_current_page(struct mm_struct * mm, struct vm_area_struct *vma, unsigned long addr) argument
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-arm/
H A Dtlb.h69 * In the case of tlb vma handling, we can optimise these away in the
74 tlb_start_vma(struct mmu_gather *tlb, struct vm_area_struct *vma) argument
77 flush_cache_range(vma, vma->vm_start, vma->vm_end);
81 tlb_end_vma(struct mmu_gather *tlb, struct vm_area_struct *vma) argument
84 flush_tlb_range(vma, vma->vm_start, vma->vm_end);
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-avr32/
H A Dtlbflush.h19 * - flush_tlb_page(vma, vmaddr) flushes one page
20 * - flush_tlb_range(vma, start, end) flushes a range of pages
27 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
29 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long page);
H A Dcacheflush.h91 #define flush_cache_range(vma, start, end) do { } while (0)
92 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0)
101 * #define flush_icache_page(vma, page) do { } while (0)
103 extern void flush_icache_page(struct vm_area_struct *vma, struct page *page);
119 extern void flush_icache_user_range(struct vm_area_struct *vma,
123 #define copy_to_user_page(vma, page, vaddr, dst, src, len) do { \
125 flush_icache_user_range(vma, page, vaddr, len); \
127 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-sh/
H A Dcacheflush.h18 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \
20 flush_cache_page(vma, vaddr, page_to_pfn(page));\
22 flush_icache_user_range(vma, page, vaddr, len); \
25 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
27 flush_cache_page(vma, vaddr, page_to_pfn(page));\
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/mm/
H A Dfremap.c23 static int zap_pte(struct mm_struct *mm, struct vm_area_struct *vma, argument
30 flush_cache_page(vma, addr, pte_pfn(pte));
31 pte = ptep_clear_flush(vma, addr, ptep);
32 page = vm_normal_page(vma, addr, pte);
36 page_remove_rmap(page, vma);
51 int install_page(struct mm_struct *mm, struct vm_area_struct *vma, argument
70 inode = vma->vm_file->f_mapping->host;
78 if (pte_none(*pte) || !zap_pte(mm, vma, addr, pte))
81 flush_icache_page(vma, page);
85 update_mmu_cache(vma, add
99 install_file_pte(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long addr, unsigned long pgoff, pgprot_t prot) argument
153 struct vm_area_struct *vma; local
[all...]
H A Drmap.c57 int anon_vma_prepare(struct vm_area_struct *vma) argument
59 struct anon_vma *anon_vma = vma->anon_vma;
63 struct mm_struct *mm = vma->vm_mm;
66 anon_vma = find_mergeable_anon_vma(vma);
81 if (likely(!vma->anon_vma)) {
82 vma->anon_vma = anon_vma;
83 list_add_tail(&vma->anon_vma_node, &anon_vma->head);
96 void __anon_vma_merge(struct vm_area_struct *vma, struct vm_area_struct *next) argument
98 BUG_ON(vma->anon_vma != next->anon_vma);
102 void __anon_vma_link(struct vm_area_struct *vma) argument
110 anon_vma_link(struct vm_area_struct *vma) argument
121 anon_vma_unlink(struct vm_area_struct *vma) argument
189 vma_address(struct page *page, struct vm_area_struct *vma) argument
207 page_address_in_vma(struct page *page, struct vm_area_struct *vma) argument
269 page_referenced_one(struct page *page, struct vm_area_struct *vma, unsigned int *mapcount) argument
305 struct vm_area_struct *vma; local
339 struct vm_area_struct *vma; local
415 page_mkclean_one(struct page *page, struct vm_area_struct *vma) argument
451 struct vm_area_struct *vma; local
492 __page_set_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) argument
516 __page_check_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) argument
547 page_add_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) argument
568 page_add_new_anon_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) argument
599 page_dup_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address) argument
614 page_remove_rmap(struct page *page, struct vm_area_struct *vma) argument
653 try_to_unmap_one(struct page *page, struct vm_area_struct *vma, int migration) argument
765 try_to_unmap_cluster(unsigned long cursor, unsigned int *mapcount, struct vm_area_struct *vma) argument
835 struct vm_area_struct *vma; local
865 struct vm_area_struct *vma; local
[all...]
H A Dmremap.c68 static void move_ptes(struct vm_area_struct *vma, pmd_t *old_pmd, argument
74 struct mm_struct *mm = vma->vm_mm;
78 if (vma->vm_file) {
82 * since it might clean the dst vma before the src vma,
85 mapping = vma->vm_file->f_mapping;
88 new_vma->vm_truncate_count != vma->vm_truncate_count)
107 pte = ptep_clear_flush(vma, old_addr, old_pte);
123 static unsigned long move_page_tables(struct vm_area_struct *vma, argument
131 flush_cache_range(vma, old_add
157 move_vma(struct vm_area_struct *vma, unsigned long old_addr, unsigned long old_len, unsigned long new_len, unsigned long new_addr) argument
254 struct vm_area_struct *vma; local
[all...]
H A Dmemory.c269 void free_pgtables(struct mmu_gather **tlb, struct vm_area_struct *vma, argument
272 while (vma) {
273 struct vm_area_struct *next = vma->vm_next;
274 unsigned long addr = vma->vm_start;
277 * Hide vma from rmap and vmtruncate before freeing pgtables
279 anon_vma_unlink(vma);
280 unlink_file_vma(vma);
282 if (is_vm_hugetlb_page(vma)) {
283 hugetlb_free_pgd_range(tlb, addr, vma->vm_end,
289 while (next && next->vm_start <= vma
353 print_bad_pte(struct vm_area_struct *vma, pte_t pte, unsigned long vaddr) argument
388 vm_normal_page(struct vm_area_struct *vma, unsigned long addr, pte_t pte) argument
428 copy_one_pte(struct mm_struct *dst_mm, struct mm_struct *src_mm, pte_t *dst_pte, pte_t *src_pte, struct vm_area_struct *vma, unsigned long addr, int *rss) argument
492 copy_pte_range(struct mm_struct *dst_mm, struct mm_struct *src_mm, pmd_t *dst_pmd, pmd_t *src_pmd, struct vm_area_struct *vma, unsigned long addr, unsigned long end) argument
542 copy_pmd_range(struct mm_struct *dst_mm, struct mm_struct *src_mm, pud_t *dst_pud, pud_t *src_pud, struct vm_area_struct *vma, unsigned long addr, unsigned long end) argument
564 copy_pud_range(struct mm_struct *dst_mm, struct mm_struct *src_mm, pgd_t *dst_pgd, pgd_t *src_pgd, struct vm_area_struct *vma, unsigned long addr, unsigned long end) argument
586 copy_page_range(struct mm_struct *dst_mm, struct mm_struct *src_mm, struct vm_area_struct *vma) argument
621 zap_pte_range(struct mmu_gather *tlb, struct vm_area_struct *vma, pmd_t *pmd, unsigned long addr, unsigned long end, long *zap_work, struct zap_details *details) argument
706 zap_pmd_range(struct mmu_gather *tlb, struct vm_area_struct *vma, pud_t *pud, unsigned long addr, unsigned long end, long *zap_work, struct zap_details *details) argument
728 zap_pud_range(struct mmu_gather *tlb, struct vm_area_struct *vma, pgd_t *pgd, unsigned long addr, unsigned long end, long *zap_work, struct zap_details *details) argument
750 unmap_page_range(struct mmu_gather *tlb, struct vm_area_struct *vma, unsigned long addr, unsigned long end, long *zap_work, struct zap_details *details) argument
811 unmap_vmas(struct mmu_gather **tlbp, struct vm_area_struct *vma, unsigned long start_addr, unsigned long end_addr, unsigned long *nr_accounted, struct zap_details *details) argument
883 zap_page_range(struct vm_area_struct *vma, unsigned long address, unsigned long size, struct zap_details *details) argument
903 follow_page(struct vm_area_struct *vma, unsigned long address, unsigned int flags) argument
995 struct vm_area_struct *vma; local
1176 zeromap_page_range(struct vm_area_struct *vma, unsigned long addr, unsigned long size, pgprot_t prot) argument
1269 vm_insert_page(struct vm_area_struct *vma, unsigned long addr, struct page *page) argument
1292 vm_insert_pfn(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn) argument
1401 remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, unsigned long pfn, unsigned long size, pgprot_t prot) argument
1576 maybe_mkwrite(pte_t pte, struct vm_area_struct *vma) argument
1583 cow_user_page(struct page *dst, struct page *src, unsigned long va, struct vm_area_struct *vma) argument
1629 do_wp_page(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *page_table, pmd_t *pmd, spinlock_t *ptl, pte_t orig_pte) argument
1812 struct vm_area_struct *vma; local
1821 unmap_mapping_range_vma(struct vm_area_struct *vma, unsigned long start_addr, unsigned long end_addr, struct zap_details *details) argument
1865 struct vm_area_struct *vma; local
1897 struct vm_area_struct *vma; local
2069 swapin_readahead(swp_entry_t entry, unsigned long addr,struct vm_area_struct *vma) argument
2119 do_swap_page(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *page_table, pmd_t *pmd, int write_access, pte_t orig_pte) argument
2220 do_anonymous_page(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *page_table, pmd_t *pmd, int write_access) argument
2289 do_no_page(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *page_table, pmd_t *pmd, int write_access) argument
2439 do_no_pfn(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *page_table, pmd_t *pmd, int write_access) argument
2482 do_file_page(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *page_table, pmd_t *pmd, int write_access, pte_t orig_pte) argument
2524 handle_pte_fault(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, pte_t *pte, pmd_t *pmd, int write_access) argument
2585 __handle_mm_fault(struct mm_struct *mm, struct vm_area_struct *vma, unsigned long address, int write_access) argument
2668 struct vm_area_struct * vma; local
2776 struct vm_area_struct *vma; local
[all...]
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/fs/proc/
H A Dtask_mmu.c77 struct vm_area_struct * vma; local
90 vma = mm->mmap;
91 while (vma) {
92 if ((vma->vm_flags & VM_EXECUTABLE) && vma->vm_file)
94 vma = vma->vm_next;
97 if (vma) {
98 *mnt = mntget(vma->vm_file->f_path.mnt);
99 *dentry = dget(vma
128 struct vm_area_struct *vma; member in struct:pmd_walker
138 struct vm_area_struct *vma = v; local
221 smaps_pte_range(struct vm_area_struct *vma, pmd_t *pmd, unsigned long addr, unsigned long end, void *private) argument
261 clear_refs_pte_range(struct vm_area_struct *vma, pmd_t *pmd, unsigned long addr, unsigned long end, void *private) argument
326 walk_page_range(struct vm_area_struct *vma, void (*action)(struct vm_area_struct *, pmd_t *, unsigned long, unsigned long, void *), void *private) argument
353 struct vm_area_struct *vma = v; local
364 struct vm_area_struct *vma; local
379 struct vm_area_struct *vma, *tail_vma = NULL; local
439 vma_stop(struct proc_maps_private *priv, struct vm_area_struct *vma) argument
451 struct vm_area_struct *vma = v; local
464 struct vm_area_struct *vma = v; local
[all...]
H A Dtask_nommu.c22 if (!vml->vma)
27 atomic_read(&vml->vma->vm_usage) > 1
29 sbytes += kobjsize((void *) vml->vma->vm_start);
30 sbytes += kobjsize(vml->vma);
32 bytes += kobjsize((void *) vml->vma->vm_start);
33 bytes += kobjsize(vml->vma);
34 slack += kobjsize((void *) vml->vma->vm_start) -
35 (vml->vma->vm_end - vml->vma->vm_start);
78 if (tbp->vma)
110 struct vm_area_struct *vma; local
[all...]
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-powerpc/
H A Dcacheflush.h22 #define flush_cache_range(vma, start, end) do { } while (0)
23 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0)
24 #define flush_icache_page(vma, page) do { } while (0)
39 extern void flush_icache_user_range(struct vm_area_struct *vma,
58 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \
61 flush_icache_user_range(vma, page, vaddr, len); \
63 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
H A Dtlbflush.h7 * - flush_tlb_page(vma, vmaddr) flushes one page
8 * - flush_tlb_page_nohash(vma, vmaddr) flushes one page if SW loaded TLB
9 * - flush_tlb_range(vma, start, end) flushes a range of pages
45 static inline void flush_tlb_page(struct vm_area_struct *vma, argument
51 static inline void flush_tlb_page_nohash(struct vm_area_struct *vma, argument
57 static inline void flush_tlb_range(struct vm_area_struct *vma, argument
77 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr);
78 extern void flush_tlb_page_nohash(struct vm_area_struct *vma, unsigned long addr);
79 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
138 static inline void flush_tlb_page(struct vm_area_struct *vma, argument
143 flush_tlb_page_nohash(struct vm_area_struct *vma, unsigned long vmaddr) argument
148 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) argument
[all...]
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-ia64/
H A Dcacheflush.h22 #define flush_cache_range(vma, start, end) do { } while (0)
23 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0)
24 #define flush_icache_page(vma,page) do { } while (0)
38 #define flush_icache_user_range(vma, page, user_addr, len) \
44 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \
46 flush_icache_user_range(vma, page, vaddr, len); \
48 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/arch/arm/mm/
H A Dmmap.c29 struct vm_area_struct *vma; local
69 vma = find_vma(mm, addr);
71 (!vma || addr + len <= vma->vm_start))
87 for (vma = find_vma(mm, addr); ; vma = vma->vm_next) {
88 /* At this point: (!vma || addr < vma->vm_end). */
101 if (!vma || add
[all...]
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/arch/i386/mm/
H A Dhugetlbpage.c20 struct vm_area_struct *vma,
33 vma->vm_flags != svma->vm_flags ||
40 static int vma_shareable(struct vm_area_struct *vma, unsigned long addr) argument
48 if (vma->vm_flags & VM_MAYSHARE &&
49 vma->vm_start <= base && end <= vma->vm_end)
59 struct vm_area_struct *vma = find_vma(mm, addr); local
60 struct address_space *mapping = vma->vm_file->f_mapping;
61 pgoff_t idx = ((addr - vma->vm_start) >> PAGE_SHIFT) +
62 vma
19 page_table_shareable(struct vm_area_struct *svma, struct vm_area_struct *vma, unsigned long addr, pgoff_t idx) argument
191 struct vm_area_struct *vma; local
233 struct vm_area_struct *vma, *prev_vma; local
322 struct vm_area_struct *vma; local
[all...]
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/arch/arm/mach-netx/
H A Dfb.h22 int netx_clcd_mmap(struct clcd_fb *fb, struct vm_area_struct *vma);
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/asm-v850/
H A Dtlbflush.h48 static inline void flush_tlb_page(struct vm_area_struct *vma, unsigned long addr) argument
53 static inline void flush_tlb_range(struct vm_area_struct *vma, argument
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/include/linux/
H A Drmap.h19 * directly to a vma: instead it points to an anon_vma, on whose list
22 * After unlinking the last vma on the list, we must garbage collect
24 * pointing to this anon_vma once its vma list is empty.
27 spinlock_t lock; /* Serialize access to vma list */
45 static inline void anon_vma_lock(struct vm_area_struct *vma) argument
47 struct anon_vma *anon_vma = vma->anon_vma;
52 static inline void anon_vma_unlock(struct vm_area_struct *vma) argument
54 struct anon_vma *anon_vma = vma->anon_vma;
78 void page_dup_rmap(struct page *page, struct vm_area_struct *vma, unsigned long address);
80 static inline void page_dup_rmap(struct page *page, struct vm_area_struct *vma, unsigne argument
[all...]
/netgear-WNDR4500v2-V1.0.0.60_1.0.38/src/linux/linux-2.6/sound/arm/
H A Ddevdma.h3 int devdma_mmap(struct device *dev, struct snd_pcm_substream *substream, struct vm_area_struct *vma);

Completed in 282 milliseconds

1234567891011>>