arm64: add Cortex-A53 cache errata workaround
[deliverable/linux.git] / arch / arm64 / mm / cache.S
CommitLineData
f1a0c4aa
CM
1/*
2 * Cache maintenance
3 *
4 * Copyright (C) 2001 Deep Blue Solutions Ltd.
5 * Copyright (C) 2012 ARM Ltd.
6 *
7 * This program is free software; you can redistribute it and/or modify
8 * it under the terms of the GNU General Public License version 2 as
9 * published by the Free Software Foundation.
10 *
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
15 *
16 * You should have received a copy of the GNU General Public License
17 * along with this program. If not, see <http://www.gnu.org/licenses/>.
18 */
19
20#include <linux/linkage.h>
21#include <linux/init.h>
22#include <asm/assembler.h>
301bcfac
AP
23#include <asm/cpufeature.h>
24#include <asm/alternative-asm.h>
f1a0c4aa
CM
25
26#include "proc-macros.S"
27
28/*
29 * __flush_dcache_all()
30 *
31 * Flush the whole D-cache.
32 *
33 * Corrupted registers: x0-x7, x9-x11
34 */
bff70595 35__flush_dcache_all:
dc60b777 36 dmb sy // ensure ordering with previous memory accesses
f1a0c4aa
CM
37 mrs x0, clidr_el1 // read clidr
38 and x3, x0, #0x7000000 // extract loc from clidr
39 lsr x3, x3, #23 // left align loc bit field
40 cbz x3, finished // if loc is 0, then no need to clean
41 mov x10, #0 // start clean at cache level 0
42loop1:
43 add x2, x10, x10, lsr #1 // work out 3x current cache level
44 lsr x1, x0, x2 // extract cache type bits from clidr
45 and x1, x1, #7 // mask of the bits for current cache only
46 cmp x1, #2 // see what cache we have at this level
47 b.lt skip // skip if no cache, or just i-cache
48 save_and_disable_irqs x9 // make CSSELR and CCSIDR access atomic
49 msr csselr_el1, x10 // select current cache level in csselr
50 isb // isb to sych the new cssr&csidr
51 mrs x1, ccsidr_el1 // read the new ccsidr
52 restore_irqs x9
53 and x2, x1, #7 // extract the length of the cache lines
54 add x2, x2, #4 // add 4 (line length offset)
55 mov x4, #0x3ff
56 and x4, x4, x1, lsr #3 // find maximum number on the way size
b4fed079 57 clz w5, w4 // find bit position of way size increment
f1a0c4aa
CM
58 mov x7, #0x7fff
59 and x7, x7, x1, lsr #13 // extract max number of the index size
60loop2:
61 mov x9, x4 // create working copy of max way size
62loop3:
63 lsl x6, x9, x5
64 orr x11, x10, x6 // factor way and cache number into x11
65 lsl x6, x7, x2
66 orr x11, x11, x6 // factor index number into x11
67 dc cisw, x11 // clean & invalidate by set/way
68 subs x9, x9, #1 // decrement the way
69 b.ge loop3
70 subs x7, x7, #1 // decrement the index
71 b.ge loop2
72skip:
73 add x10, x10, #2 // increment cache number
74 cmp x3, x10
75 b.gt loop1
76finished:
77 mov x10, #0 // swith back to cache level 0
78 msr csselr_el1, x10 // select current cache level in csselr
79 dsb sy
80 isb
81 ret
82ENDPROC(__flush_dcache_all)
83
84/*
85 * flush_cache_all()
86 *
87 * Flush the entire cache system. The data cache flush is now achieved
88 * using atomic clean / invalidates working outwards from L1 cache. This
89 * is done using Set/Way based cache maintainance instructions. The
90 * instruction cache can still be invalidated back to the point of
91 * unification in a single instruction.
92 */
93ENTRY(flush_cache_all)
94 mov x12, lr
95 bl __flush_dcache_all
96 mov x0, #0
97 ic ialluis // I+BTB cache invalidate
98 ret x12
99ENDPROC(flush_cache_all)
100
101/*
102 * flush_icache_range(start,end)
103 *
104 * Ensure that the I and D caches are coherent within specified region.
105 * This is typically used when code has been written to a memory region,
106 * and will be executed.
107 *
108 * - start - virtual start address of region
109 * - end - virtual end address of region
110 */
111ENTRY(flush_icache_range)
112 /* FALLTHROUGH */
113
114/*
115 * __flush_cache_user_range(start,end)
116 *
117 * Ensure that the I and D caches are coherent within specified region.
118 * This is typically used when code has been written to a memory region,
119 * and will be executed.
120 *
121 * - start - virtual start address of region
122 * - end - virtual end address of region
123 */
124ENTRY(__flush_cache_user_range)
125 dcache_line_size x2, x3
126 sub x3, x2, #1
127 bic x4, x0, x3
1281:
129USER(9f, dc cvau, x4 ) // clean D line to PoU
130 add x4, x4, x2
131 cmp x4, x1
132 b.lo 1b
dc60b777 133 dsb ish
f1a0c4aa
CM
134
135 icache_line_size x2, x3
136 sub x3, x2, #1
137 bic x4, x0, x3
1381:
139USER(9f, ic ivau, x4 ) // invalidate I line PoU
140 add x4, x4, x2
141 cmp x4, x1
142 b.lo 1b
1439: // ignore any faulting cache operation
dc60b777 144 dsb ish
f1a0c4aa
CM
145 isb
146 ret
147ENDPROC(flush_icache_range)
148ENDPROC(__flush_cache_user_range)
149
150/*
03324e6e 151 * __flush_dcache_area(kaddr, size)
f1a0c4aa
CM
152 *
153 * Ensure that the data held in the page kaddr is written back to the
154 * page in question.
155 *
156 * - kaddr - kernel address
157 * - size - size in question
158 */
159ENTRY(__flush_dcache_area)
160 dcache_line_size x2, x3
161 add x1, x0, x1
162 sub x3, x2, #1
163 bic x0, x0, x3
1641: dc civac, x0 // clean & invalidate D line / unified line
165 add x0, x0, x2
166 cmp x0, x1
167 b.lo 1b
168 dsb sy
169 ret
170ENDPROC(__flush_dcache_area)
7363590d 171
c218bca7
CM
172/*
173 * __inval_cache_range(start, end)
174 * - start - start address of region
175 * - end - end address of region
176 */
177ENTRY(__inval_cache_range)
178 /* FALLTHROUGH */
179
7363590d
CM
180/*
181 * __dma_inv_range(start, end)
182 * - start - virtual start address of region
183 * - end - virtual end address of region
184 */
185__dma_inv_range:
186 dcache_line_size x2, x3
187 sub x3, x2, #1
ebf81a93 188 tst x1, x3 // end cache line aligned?
7363590d 189 bic x1, x1, x3
ebf81a93
CM
190 b.eq 1f
191 dc civac, x1 // clean & invalidate D / U line
1921: tst x0, x3 // start cache line aligned?
193 bic x0, x0, x3
194 b.eq 2f
195 dc civac, x0 // clean & invalidate D / U line
196 b 3f
1972: dc ivac, x0 // invalidate D / U line
1983: add x0, x0, x2
7363590d 199 cmp x0, x1
ebf81a93 200 b.lo 2b
7363590d
CM
201 dsb sy
202 ret
c218bca7 203ENDPROC(__inval_cache_range)
7363590d
CM
204ENDPROC(__dma_inv_range)
205
206/*
207 * __dma_clean_range(start, end)
208 * - start - virtual start address of region
209 * - end - virtual end address of region
210 */
211__dma_clean_range:
212 dcache_line_size x2, x3
213 sub x3, x2, #1
214 bic x0, x0, x3
301bcfac 2151: alternative_insn "dc cvac, x0", "dc civac, x0", ARM64_WORKAROUND_CLEAN_CACHE
7363590d
CM
216 add x0, x0, x2
217 cmp x0, x1
218 b.lo 1b
219 dsb sy
220 ret
221ENDPROC(__dma_clean_range)
222
223/*
224 * __dma_flush_range(start, end)
225 * - start - virtual start address of region
226 * - end - virtual end address of region
227 */
228ENTRY(__dma_flush_range)
229 dcache_line_size x2, x3
230 sub x3, x2, #1
231 bic x0, x0, x3
2321: dc civac, x0 // clean & invalidate D / U line
233 add x0, x0, x2
234 cmp x0, x1
235 b.lo 1b
236 dsb sy
237 ret
238ENDPROC(__dma_flush_range)
239
240/*
241 * __dma_map_area(start, size, dir)
242 * - start - kernel virtual start address
243 * - size - size of region
244 * - dir - DMA direction
245 */
246ENTRY(__dma_map_area)
247 add x1, x1, x0
248 cmp w2, #DMA_FROM_DEVICE
249 b.eq __dma_inv_range
250 b __dma_clean_range
251ENDPROC(__dma_map_area)
252
253/*
254 * __dma_unmap_area(start, size, dir)
255 * - start - kernel virtual start address
256 * - size - size of region
257 * - dir - DMA direction
258 */
259ENTRY(__dma_unmap_area)
260 add x1, x1, x0
261 cmp w2, #DMA_TO_DEVICE
262 b.ne __dma_inv_range
263 ret
264ENDPROC(__dma_unmap_area)
This page took 0.122407 seconds and 5 git commands to generate.