1/*
2 * This file is subject to the terms and conditions of the GNU General Public
3 * License.  See the file "COPYING" in the main directory of this archive
4 * for more details.
5 *
6 * Copyright (C) 2001 - 2013 Tensilica Inc.
7 */
8
9#ifndef _XTENSA_TLBFLUSH_H
10#define _XTENSA_TLBFLUSH_H
11
12#include <linux/stringify.h>
13#include <asm/processor.h>
14
15#define DTLB_WAY_PGD	7
16
17#define ITLB_ARF_WAYS	4
18#define DTLB_ARF_WAYS	4
19
20#define ITLB_HIT_BIT	3
21#define DTLB_HIT_BIT	4
22
23#ifndef __ASSEMBLY__
24
25/* TLB flushing:
26 *
27 *  - flush_tlb_all() flushes all processes TLB entries
28 *  - flush_tlb_mm(mm) flushes the specified mm context TLB entries
29 *  - flush_tlb_page(vma, page) flushes a single page
30 *  - flush_tlb_range(vma, vmaddr, end) flushes a range of pages
31 */
32
33void local_flush_tlb_all(void);
34void local_flush_tlb_mm(struct mm_struct *mm);
35void local_flush_tlb_page(struct vm_area_struct *vma,
36		unsigned long page);
37void local_flush_tlb_range(struct vm_area_struct *vma,
38		unsigned long start, unsigned long end);
39void local_flush_tlb_kernel_range(unsigned long start, unsigned long end);
40
41#ifdef CONFIG_SMP
42
43void flush_tlb_all(void);
44void flush_tlb_mm(struct mm_struct *);
45void flush_tlb_page(struct vm_area_struct *, unsigned long);
46void flush_tlb_range(struct vm_area_struct *, unsigned long,
47		unsigned long);
48void flush_tlb_kernel_range(unsigned long start, unsigned long end);
49
50#else /* !CONFIG_SMP */
51
52#define flush_tlb_all()			   local_flush_tlb_all()
53#define flush_tlb_mm(mm)		   local_flush_tlb_mm(mm)
54#define flush_tlb_page(vma, page)	   local_flush_tlb_page(vma, page)
55#define flush_tlb_range(vma, vmaddr, end)  local_flush_tlb_range(vma, vmaddr, \
56								 end)
57#define flush_tlb_kernel_range(start, end) local_flush_tlb_kernel_range(start, \
58									end)
59
60#endif /* CONFIG_SMP */
61
62/* TLB operations. */
63
64static inline unsigned long itlb_probe(unsigned long addr)
65{
66	unsigned long tmp;
67	__asm__ __volatile__("pitlb  %0, %1\n\t" : "=a" (tmp) : "a" (addr));
68	return tmp;
69}
70
71static inline unsigned long dtlb_probe(unsigned long addr)
72{
73	unsigned long tmp;
74	__asm__ __volatile__("pdtlb  %0, %1\n\t" : "=a" (tmp) : "a" (addr));
75	return tmp;
76}
77
78static inline void invalidate_itlb_entry (unsigned long probe)
79{
80	__asm__ __volatile__("iitlb  %0; isync\n\t" : : "a" (probe));
81}
82
83static inline void invalidate_dtlb_entry (unsigned long probe)
84{
85	__asm__ __volatile__("idtlb  %0; dsync\n\t" : : "a" (probe));
86}
87
88/* Use the .._no_isync functions with caution.  Generally, these are
89 * handy for bulk invalidates followed by a single 'isync'.  The
90 * caller must follow up with an 'isync', which can be relatively
91 * expensive on some Xtensa implementations.
92 */
93static inline void invalidate_itlb_entry_no_isync (unsigned entry)
94{
95	/* Caller must follow up with 'isync'. */
96	__asm__ __volatile__ ("iitlb  %0\n" : : "a" (entry) );
97}
98
99static inline void invalidate_dtlb_entry_no_isync (unsigned entry)
100{
101	/* Caller must follow up with 'isync'. */
102	__asm__ __volatile__ ("idtlb  %0\n" : : "a" (entry) );
103}
104
105static inline void set_itlbcfg_register (unsigned long val)
106{
107	__asm__ __volatile__("wsr  %0, itlbcfg\n\t" "isync\n\t"
108			     : : "a" (val));
109}
110
111static inline void set_dtlbcfg_register (unsigned long val)
112{
113	__asm__ __volatile__("wsr  %0, dtlbcfg; dsync\n\t"
114	    		     : : "a" (val));
115}
116
117static inline void set_ptevaddr_register (unsigned long val)
118{
119	__asm__ __volatile__(" wsr  %0, ptevaddr; isync\n"
120			     : : "a" (val));
121}
122
123static inline unsigned long read_ptevaddr_register (void)
124{
125	unsigned long tmp;
126	__asm__ __volatile__("rsr  %0, ptevaddr\n\t" : "=a" (tmp));
127	return tmp;
128}
129
130static inline void write_dtlb_entry (pte_t entry, int way)
131{
132	__asm__ __volatile__("wdtlb  %1, %0; dsync\n\t"
133			     : : "r" (way), "r" (entry) );
134}
135
136static inline void write_itlb_entry (pte_t entry, int way)
137{
138	__asm__ __volatile__("witlb  %1, %0; isync\n\t"
139	                     : : "r" (way), "r" (entry) );
140}
141
142static inline void invalidate_page_directory (void)
143{
144	invalidate_dtlb_entry (DTLB_WAY_PGD);
145	invalidate_dtlb_entry (DTLB_WAY_PGD+1);
146	invalidate_dtlb_entry (DTLB_WAY_PGD+2);
147}
148
149static inline void invalidate_itlb_mapping (unsigned address)
150{
151	unsigned long tlb_entry;
152	if (((tlb_entry = itlb_probe(address)) & (1 << ITLB_HIT_BIT)) != 0)
153		invalidate_itlb_entry(tlb_entry);
154}
155
156static inline void invalidate_dtlb_mapping (unsigned address)
157{
158	unsigned long tlb_entry;
159	if (((tlb_entry = dtlb_probe(address)) & (1 << DTLB_HIT_BIT)) != 0)
160		invalidate_dtlb_entry(tlb_entry);
161}
162
163/*
164 * DO NOT USE THESE FUNCTIONS.  These instructions aren't part of the Xtensa
165 * ISA and exist only for test purposes..
166 * You may find it helpful for MMU debugging, however.
167 *
168 * 'at' is the unmodified input register
169 * 'as' is the output register, as follows (specific to the Linux config):
170 *
171 *      as[31..12] contain the virtual address
172 *      as[11..08] are meaningless
173 *      as[07..00] contain the asid
174 */
175
176static inline unsigned long read_dtlb_virtual (int way)
177{
178	unsigned long tmp;
179	__asm__ __volatile__("rdtlb0  %0, %1\n\t" : "=a" (tmp), "+a" (way));
180	return tmp;
181}
182
183static inline unsigned long read_dtlb_translation (int way)
184{
185	unsigned long tmp;
186	__asm__ __volatile__("rdtlb1  %0, %1\n\t" : "=a" (tmp), "+a" (way));
187	return tmp;
188}
189
190static inline unsigned long read_itlb_virtual (int way)
191{
192	unsigned long tmp;
193	__asm__ __volatile__("ritlb0  %0, %1\n\t" : "=a" (tmp), "+a" (way));
194	return tmp;
195}
196
197static inline unsigned long read_itlb_translation (int way)
198{
199	unsigned long tmp;
200	__asm__ __volatile__("ritlb1  %0, %1\n\t" : "=a" (tmp), "+a" (way));
201	return tmp;
202}
203
204#endif	/* __ASSEMBLY__ */
205#endif	/* _XTENSA_TLBFLUSH_H */
206