Merge branch 'for-linus' of git://neil.brown.name/md
[deliverable/linux.git] / arch / blackfin / kernel / cplb-mpu / cplbmgr.c
1 /*
2 * Blackfin CPLB exception handling.
3 * Copyright 2004-2007 Analog Devices Inc.
4 *
5 * This program is free software; you can redistribute it and/or modify
6 * it under the terms of the GNU General Public License as published by
7 * the Free Software Foundation; either version 2 of the License, or
8 * (at your option) any later version.
9 *
10 * This program is distributed in the hope that it will be useful,
11 * but WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
13 * GNU General Public License for more details.
14 *
15 * You should have received a copy of the GNU General Public License
16 * along with this program; if not, see the file COPYING, or write
17 * to the Free Software Foundation, Inc.,
18 * 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
19 */
20 #include <linux/module.h>
21 #include <linux/mm.h>
22
23 #include <asm/blackfin.h>
24 #include <asm/cacheflush.h>
25 #include <asm/cplbinit.h>
26 #include <asm/mmu_context.h>
27
28 /*
29 * WARNING
30 *
31 * This file is compiled with certain -ffixed-reg options. We have to
32 * make sure not to call any functions here that could clobber these
33 * registers.
34 */
35
36 int page_mask_nelts;
37 int page_mask_order;
38 unsigned long *current_rwx_mask[NR_CPUS];
39
40 int nr_dcplb_miss[NR_CPUS], nr_icplb_miss[NR_CPUS];
41 int nr_icplb_supv_miss[NR_CPUS], nr_dcplb_prot[NR_CPUS];
42 int nr_cplb_flush[NR_CPUS];
43
44 static inline void disable_dcplb(void)
45 {
46 unsigned long ctrl;
47 SSYNC();
48 ctrl = bfin_read_DMEM_CONTROL();
49 ctrl &= ~ENDCPLB;
50 bfin_write_DMEM_CONTROL(ctrl);
51 SSYNC();
52 }
53
54 static inline void enable_dcplb(void)
55 {
56 unsigned long ctrl;
57 SSYNC();
58 ctrl = bfin_read_DMEM_CONTROL();
59 ctrl |= ENDCPLB;
60 bfin_write_DMEM_CONTROL(ctrl);
61 SSYNC();
62 }
63
64 static inline void disable_icplb(void)
65 {
66 unsigned long ctrl;
67 SSYNC();
68 ctrl = bfin_read_IMEM_CONTROL();
69 ctrl &= ~ENICPLB;
70 bfin_write_IMEM_CONTROL(ctrl);
71 SSYNC();
72 }
73
74 static inline void enable_icplb(void)
75 {
76 unsigned long ctrl;
77 SSYNC();
78 ctrl = bfin_read_IMEM_CONTROL();
79 ctrl |= ENICPLB;
80 bfin_write_IMEM_CONTROL(ctrl);
81 SSYNC();
82 }
83
84 /*
85 * Given the contents of the status register, return the index of the
86 * CPLB that caused the fault.
87 */
88 static inline int faulting_cplb_index(int status)
89 {
90 int signbits = __builtin_bfin_norm_fr1x32(status & 0xFFFF);
91 return 30 - signbits;
92 }
93
94 /*
95 * Given the contents of the status register and the DCPLB_DATA contents,
96 * return true if a write access should be permitted.
97 */
98 static inline int write_permitted(int status, unsigned long data)
99 {
100 if (status & FAULT_USERSUPV)
101 return !!(data & CPLB_SUPV_WR);
102 else
103 return !!(data & CPLB_USER_WR);
104 }
105
106 /* Counters to implement round-robin replacement. */
107 static int icplb_rr_index[NR_CPUS], dcplb_rr_index[NR_CPUS];
108
109 /*
110 * Find an ICPLB entry to be evicted and return its index.
111 */
112 static int evict_one_icplb(unsigned int cpu)
113 {
114 int i;
115 for (i = first_switched_icplb; i < MAX_CPLBS; i++)
116 if ((icplb_tbl[cpu][i].data & CPLB_VALID) == 0)
117 return i;
118 i = first_switched_icplb + icplb_rr_index[cpu];
119 if (i >= MAX_CPLBS) {
120 i -= MAX_CPLBS - first_switched_icplb;
121 icplb_rr_index[cpu] -= MAX_CPLBS - first_switched_icplb;
122 }
123 icplb_rr_index[cpu]++;
124 return i;
125 }
126
127 static int evict_one_dcplb(unsigned int cpu)
128 {
129 int i;
130 for (i = first_switched_dcplb; i < MAX_CPLBS; i++)
131 if ((dcplb_tbl[cpu][i].data & CPLB_VALID) == 0)
132 return i;
133 i = first_switched_dcplb + dcplb_rr_index[cpu];
134 if (i >= MAX_CPLBS) {
135 i -= MAX_CPLBS - first_switched_dcplb;
136 dcplb_rr_index[cpu] -= MAX_CPLBS - first_switched_dcplb;
137 }
138 dcplb_rr_index[cpu]++;
139 return i;
140 }
141
142 static noinline int dcplb_miss(unsigned int cpu)
143 {
144 unsigned long addr = bfin_read_DCPLB_FAULT_ADDR();
145 int status = bfin_read_DCPLB_STATUS();
146 unsigned long *mask;
147 int idx;
148 unsigned long d_data;
149
150 nr_dcplb_miss[cpu]++;
151
152 d_data = CPLB_SUPV_WR | CPLB_VALID | CPLB_DIRTY | PAGE_SIZE_4KB;
153 #ifdef CONFIG_BFIN_DCACHE
154 if (bfin_addr_dcacheable(addr)) {
155 d_data |= CPLB_L1_CHBL | ANOMALY_05000158_WORKAROUND;
156 #ifdef CONFIG_BFIN_WT
157 d_data |= CPLB_L1_AOW | CPLB_WT;
158 #endif
159 }
160 #endif
161 if (addr >= physical_mem_end) {
162 if (addr >= ASYNC_BANK0_BASE && addr < ASYNC_BANK3_BASE + ASYNC_BANK3_SIZE
163 && (status & FAULT_USERSUPV)) {
164 addr &= ~0x3fffff;
165 d_data &= ~PAGE_SIZE_4KB;
166 d_data |= PAGE_SIZE_4MB;
167 } else if (addr >= BOOT_ROM_START && addr < BOOT_ROM_START + BOOT_ROM_LENGTH
168 && (status & (FAULT_RW | FAULT_USERSUPV)) == FAULT_USERSUPV) {
169 addr &= ~(1 * 1024 * 1024 - 1);
170 d_data &= ~PAGE_SIZE_4KB;
171 d_data |= PAGE_SIZE_1MB;
172 } else
173 return CPLB_PROT_VIOL;
174 } else if (addr >= _ramend) {
175 d_data |= CPLB_USER_RD | CPLB_USER_WR;
176 } else {
177 mask = current_rwx_mask[cpu];
178 if (mask) {
179 int page = addr >> PAGE_SHIFT;
180 int idx = page >> 5;
181 int bit = 1 << (page & 31);
182
183 if (mask[idx] & bit)
184 d_data |= CPLB_USER_RD;
185
186 mask += page_mask_nelts;
187 if (mask[idx] & bit)
188 d_data |= CPLB_USER_WR;
189 }
190 }
191 idx = evict_one_dcplb(cpu);
192
193 addr &= PAGE_MASK;
194 dcplb_tbl[cpu][idx].addr = addr;
195 dcplb_tbl[cpu][idx].data = d_data;
196
197 disable_dcplb();
198 bfin_write32(DCPLB_DATA0 + idx * 4, d_data);
199 bfin_write32(DCPLB_ADDR0 + idx * 4, addr);
200 enable_dcplb();
201
202 return 0;
203 }
204
205 static noinline int icplb_miss(unsigned int cpu)
206 {
207 unsigned long addr = bfin_read_ICPLB_FAULT_ADDR();
208 int status = bfin_read_ICPLB_STATUS();
209 int idx;
210 unsigned long i_data;
211
212 nr_icplb_miss[cpu]++;
213
214 /* If inside the uncached DMA region, fault. */
215 if (addr >= _ramend - DMA_UNCACHED_REGION && addr < _ramend)
216 return CPLB_PROT_VIOL;
217
218 if (status & FAULT_USERSUPV)
219 nr_icplb_supv_miss[cpu]++;
220
221 /*
222 * First, try to find a CPLB that matches this address. If we
223 * find one, then the fact that we're in the miss handler means
224 * that the instruction crosses a page boundary.
225 */
226 for (idx = first_switched_icplb; idx < MAX_CPLBS; idx++) {
227 if (icplb_tbl[cpu][idx].data & CPLB_VALID) {
228 unsigned long this_addr = icplb_tbl[cpu][idx].addr;
229 if (this_addr <= addr && this_addr + PAGE_SIZE > addr) {
230 addr += PAGE_SIZE;
231 break;
232 }
233 }
234 }
235
236 i_data = CPLB_VALID | CPLB_PORTPRIO | PAGE_SIZE_4KB;
237
238 #ifdef CONFIG_BFIN_ICACHE
239 /*
240 * Normal RAM, and possibly the reserved memory area, are
241 * cacheable.
242 */
243 if (addr < _ramend ||
244 (addr < physical_mem_end && reserved_mem_icache_on))
245 i_data |= CPLB_L1_CHBL | ANOMALY_05000158_WORKAROUND;
246 #endif
247
248 if (addr >= physical_mem_end) {
249 if (addr >= BOOT_ROM_START && addr < BOOT_ROM_START + BOOT_ROM_LENGTH
250 && (status & FAULT_USERSUPV)) {
251 addr &= ~(1 * 1024 * 1024 - 1);
252 i_data &= ~PAGE_SIZE_4KB;
253 i_data |= PAGE_SIZE_1MB;
254 } else
255 return CPLB_PROT_VIOL;
256 } else if (addr >= _ramend) {
257 i_data |= CPLB_USER_RD;
258 } else {
259 /*
260 * Two cases to distinguish - a supervisor access must
261 * necessarily be for a module page; we grant it
262 * unconditionally (could do better here in the future).
263 * Otherwise, check the x bitmap of the current process.
264 */
265 if (!(status & FAULT_USERSUPV)) {
266 unsigned long *mask = current_rwx_mask[cpu];
267
268 if (mask) {
269 int page = addr >> PAGE_SHIFT;
270 int idx = page >> 5;
271 int bit = 1 << (page & 31);
272
273 mask += 2 * page_mask_nelts;
274 if (mask[idx] & bit)
275 i_data |= CPLB_USER_RD;
276 }
277 }
278 }
279 idx = evict_one_icplb(cpu);
280 addr &= PAGE_MASK;
281 icplb_tbl[cpu][idx].addr = addr;
282 icplb_tbl[cpu][idx].data = i_data;
283
284 disable_icplb();
285 bfin_write32(ICPLB_DATA0 + idx * 4, i_data);
286 bfin_write32(ICPLB_ADDR0 + idx * 4, addr);
287 enable_icplb();
288
289 return 0;
290 }
291
292 static noinline int dcplb_protection_fault(unsigned int cpu)
293 {
294 int status = bfin_read_DCPLB_STATUS();
295
296 nr_dcplb_prot[cpu]++;
297
298 if (status & FAULT_RW) {
299 int idx = faulting_cplb_index(status);
300 unsigned long data = dcplb_tbl[cpu][idx].data;
301 if (!(data & CPLB_WT) && !(data & CPLB_DIRTY) &&
302 write_permitted(status, data)) {
303 data |= CPLB_DIRTY;
304 dcplb_tbl[cpu][idx].data = data;
305 bfin_write32(DCPLB_DATA0 + idx * 4, data);
306 return 0;
307 }
308 }
309 return CPLB_PROT_VIOL;
310 }
311
312 int cplb_hdr(int seqstat, struct pt_regs *regs)
313 {
314 int cause = seqstat & 0x3f;
315 unsigned int cpu = smp_processor_id();
316 switch (cause) {
317 case 0x23:
318 return dcplb_protection_fault(cpu);
319 case 0x2C:
320 return icplb_miss(cpu);
321 case 0x26:
322 return dcplb_miss(cpu);
323 default:
324 return 1;
325 }
326 }
327
328 void flush_switched_cplbs(unsigned int cpu)
329 {
330 int i;
331 unsigned long flags;
332
333 nr_cplb_flush[cpu]++;
334
335 local_irq_save_hw(flags);
336 disable_icplb();
337 for (i = first_switched_icplb; i < MAX_CPLBS; i++) {
338 icplb_tbl[cpu][i].data = 0;
339 bfin_write32(ICPLB_DATA0 + i * 4, 0);
340 }
341 enable_icplb();
342
343 disable_dcplb();
344 for (i = first_switched_dcplb; i < MAX_CPLBS; i++) {
345 dcplb_tbl[cpu][i].data = 0;
346 bfin_write32(DCPLB_DATA0 + i * 4, 0);
347 }
348 enable_dcplb();
349 local_irq_restore_hw(flags);
350
351 }
352
353 void set_mask_dcplbs(unsigned long *masks, unsigned int cpu)
354 {
355 int i;
356 unsigned long addr = (unsigned long)masks;
357 unsigned long d_data;
358 unsigned long flags;
359
360 if (!masks) {
361 current_rwx_mask[cpu] = masks;
362 return;
363 }
364
365 local_irq_save_hw(flags);
366 current_rwx_mask[cpu] = masks;
367
368 d_data = CPLB_SUPV_WR | CPLB_VALID | CPLB_DIRTY | PAGE_SIZE_4KB;
369 #ifdef CONFIG_BFIN_DCACHE
370 d_data |= CPLB_L1_CHBL;
371 #ifdef CONFIG_BFIN_WT
372 d_data |= CPLB_L1_AOW | CPLB_WT;
373 #endif
374 #endif
375
376 disable_dcplb();
377 for (i = first_mask_dcplb; i < first_switched_dcplb; i++) {
378 dcplb_tbl[cpu][i].addr = addr;
379 dcplb_tbl[cpu][i].data = d_data;
380 bfin_write32(DCPLB_DATA0 + i * 4, d_data);
381 bfin_write32(DCPLB_ADDR0 + i * 4, addr);
382 addr += PAGE_SIZE;
383 }
384 enable_dcplb();
385 local_irq_restore_hw(flags);
386 }
This page took 0.037903 seconds and 5 git commands to generate.