Commit | Line | Data |
---|---|---|
1da177e4 LT |
1 | /* |
2 | * Simple MTD partitioning layer | |
3 | * | |
a1452a37 DW |
4 | * Copyright © 2000 Nicolas Pitre <nico@fluxnic.net> |
5 | * Copyright © 2002 Thomas Gleixner <gleixner@linutronix.de> | |
6 | * Copyright © 2000-2010 David Woodhouse <dwmw2@infradead.org> | |
1da177e4 | 7 | * |
a1452a37 DW |
8 | * This program is free software; you can redistribute it and/or modify |
9 | * it under the terms of the GNU General Public License as published by | |
10 | * the Free Software Foundation; either version 2 of the License, or | |
11 | * (at your option) any later version. | |
12 | * | |
13 | * This program is distributed in the hope that it will be useful, | |
14 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
15 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
16 | * GNU General Public License for more details. | |
17 | * | |
18 | * You should have received a copy of the GNU General Public License | |
19 | * along with this program; if not, write to the Free Software | |
20 | * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA | |
1da177e4 | 21 | * |
97894cda | 22 | */ |
1da177e4 LT |
23 | |
24 | #include <linux/module.h> | |
25 | #include <linux/types.h> | |
26 | #include <linux/kernel.h> | |
27 | #include <linux/slab.h> | |
28 | #include <linux/list.h> | |
1da177e4 LT |
29 | #include <linux/kmod.h> |
30 | #include <linux/mtd/mtd.h> | |
31 | #include <linux/mtd/partitions.h> | |
5daa7b21 | 32 | #include <linux/err.h> |
727dc612 | 33 | #include <linux/kconfig.h> |
1da177e4 | 34 | |
eea72d5f JI |
35 | #include "mtdcore.h" |
36 | ||
1da177e4 LT |
37 | /* Our partition linked list */ |
38 | static LIST_HEAD(mtd_partitions); | |
5daa7b21 | 39 | static DEFINE_MUTEX(mtd_partitions_mutex); |
1da177e4 LT |
40 | |
41 | /* Our partition node structure */ | |
42 | struct mtd_part { | |
43 | struct mtd_info mtd; | |
44 | struct mtd_info *master; | |
69423d99 | 45 | uint64_t offset; |
1da177e4 | 46 | struct list_head list; |
1da177e4 LT |
47 | }; |
48 | ||
49 | /* | |
50 | * Given a pointer to the MTD object in the mtd_part structure, we can retrieve | |
2524534d | 51 | * the pointer to that structure. |
1da177e4 | 52 | */ |
2524534d BN |
53 | static inline struct mtd_part *mtd_to_part(const struct mtd_info *mtd) |
54 | { | |
55 | return container_of(mtd, struct mtd_part, mtd); | |
56 | } | |
1da177e4 | 57 | |
97894cda TG |
58 | |
59 | /* | |
1da177e4 LT |
60 | * MTD methods which simply translate the effective address and pass through |
61 | * to the _real_ device. | |
62 | */ | |
63 | ||
b33a2887 AN |
64 | static int part_read(struct mtd_info *mtd, loff_t from, size_t len, |
65 | size_t *retlen, u_char *buf) | |
1da177e4 | 66 | { |
2524534d | 67 | struct mtd_part *part = mtd_to_part(mtd); |
d8877f19 | 68 | struct mtd_ecc_stats stats; |
f1a28c02 TG |
69 | int res; |
70 | ||
d8877f19 | 71 | stats = part->master->ecc_stats; |
994c8409 MD |
72 | res = part->master->_read(part->master, from + part->offset, len, |
73 | retlen, buf); | |
edbc4540 MD |
74 | if (unlikely(mtd_is_eccerr(res))) |
75 | mtd->ecc_stats.failed += | |
76 | part->master->ecc_stats.failed - stats.failed; | |
77 | else | |
78 | mtd->ecc_stats.corrected += | |
79 | part->master->ecc_stats.corrected - stats.corrected; | |
f1a28c02 | 80 | return res; |
1da177e4 LT |
81 | } |
82 | ||
b33a2887 AN |
83 | static int part_point(struct mtd_info *mtd, loff_t from, size_t len, |
84 | size_t *retlen, void **virt, resource_size_t *phys) | |
1da177e4 | 85 | { |
2524534d | 86 | struct mtd_part *part = mtd_to_part(mtd); |
5def4898 | 87 | |
994c8409 MD |
88 | return part->master->_point(part->master, from + part->offset, len, |
89 | retlen, virt, phys); | |
1da177e4 | 90 | } |
9223a456 | 91 | |
5e4e6e3f | 92 | static int part_unpoint(struct mtd_info *mtd, loff_t from, size_t len) |
1da177e4 | 93 | { |
2524534d | 94 | struct mtd_part *part = mtd_to_part(mtd); |
1da177e4 | 95 | |
994c8409 | 96 | return part->master->_unpoint(part->master, from + part->offset, len); |
1da177e4 LT |
97 | } |
98 | ||
402d3265 DH |
99 | static unsigned long part_get_unmapped_area(struct mtd_info *mtd, |
100 | unsigned long len, | |
101 | unsigned long offset, | |
102 | unsigned long flags) | |
103 | { | |
2524534d | 104 | struct mtd_part *part = mtd_to_part(mtd); |
402d3265 DH |
105 | |
106 | offset += part->offset; | |
994c8409 MD |
107 | return part->master->_get_unmapped_area(part->master, len, offset, |
108 | flags); | |
402d3265 DH |
109 | } |
110 | ||
8593fbc6 | 111 | static int part_read_oob(struct mtd_info *mtd, loff_t from, |
b33a2887 | 112 | struct mtd_oob_ops *ops) |
1da177e4 | 113 | { |
2524534d | 114 | struct mtd_part *part = mtd_to_part(mtd); |
f1a28c02 | 115 | int res; |
8593fbc6 | 116 | |
1da177e4 | 117 | if (from >= mtd->size) |
8593fbc6 | 118 | return -EINVAL; |
7014568b | 119 | if (ops->datbuf && from + ops->len > mtd->size) |
8593fbc6 | 120 | return -EINVAL; |
f1a28c02 | 121 | |
154bf89f AB |
122 | /* |
123 | * If OOB is also requested, make sure that we do not read past the end | |
124 | * of this partition. | |
125 | */ | |
126 | if (ops->oobbuf) { | |
127 | size_t len, pages; | |
128 | ||
29f1058a | 129 | len = mtd_oobavail(mtd, ops); |
154bf89f AB |
130 | pages = mtd_div_by_ws(mtd->size, mtd); |
131 | pages -= mtd_div_by_ws(from, mtd); | |
132 | if (ops->ooboffs + ops->ooblen > pages * len) | |
133 | return -EINVAL; | |
134 | } | |
135 | ||
994c8409 | 136 | res = part->master->_read_oob(part->master, from + part->offset, ops); |
f1a28c02 | 137 | if (unlikely(res)) { |
d57f4054 | 138 | if (mtd_is_bitflip(res)) |
f1a28c02 | 139 | mtd->ecc_stats.corrected++; |
d57f4054 | 140 | if (mtd_is_eccerr(res)) |
f1a28c02 TG |
141 | mtd->ecc_stats.failed++; |
142 | } | |
143 | return res; | |
1da177e4 LT |
144 | } |
145 | ||
b33a2887 AN |
146 | static int part_read_user_prot_reg(struct mtd_info *mtd, loff_t from, |
147 | size_t len, size_t *retlen, u_char *buf) | |
1da177e4 | 148 | { |
2524534d | 149 | struct mtd_part *part = mtd_to_part(mtd); |
994c8409 MD |
150 | return part->master->_read_user_prot_reg(part->master, from, len, |
151 | retlen, buf); | |
1da177e4 LT |
152 | } |
153 | ||
4b78fc42 CR |
154 | static int part_get_user_prot_info(struct mtd_info *mtd, size_t len, |
155 | size_t *retlen, struct otp_info *buf) | |
f77814dd | 156 | { |
2524534d | 157 | struct mtd_part *part = mtd_to_part(mtd); |
4b78fc42 CR |
158 | return part->master->_get_user_prot_info(part->master, len, retlen, |
159 | buf); | |
f77814dd NP |
160 | } |
161 | ||
b33a2887 AN |
162 | static int part_read_fact_prot_reg(struct mtd_info *mtd, loff_t from, |
163 | size_t len, size_t *retlen, u_char *buf) | |
1da177e4 | 164 | { |
2524534d | 165 | struct mtd_part *part = mtd_to_part(mtd); |
994c8409 MD |
166 | return part->master->_read_fact_prot_reg(part->master, from, len, |
167 | retlen, buf); | |
1da177e4 LT |
168 | } |
169 | ||
4b78fc42 CR |
170 | static int part_get_fact_prot_info(struct mtd_info *mtd, size_t len, |
171 | size_t *retlen, struct otp_info *buf) | |
f77814dd | 172 | { |
2524534d | 173 | struct mtd_part *part = mtd_to_part(mtd); |
4b78fc42 CR |
174 | return part->master->_get_fact_prot_info(part->master, len, retlen, |
175 | buf); | |
f77814dd NP |
176 | } |
177 | ||
b33a2887 AN |
178 | static int part_write(struct mtd_info *mtd, loff_t to, size_t len, |
179 | size_t *retlen, const u_char *buf) | |
1da177e4 | 180 | { |
2524534d | 181 | struct mtd_part *part = mtd_to_part(mtd); |
994c8409 MD |
182 | return part->master->_write(part->master, to + part->offset, len, |
183 | retlen, buf); | |
1da177e4 LT |
184 | } |
185 | ||
b33a2887 AN |
186 | static int part_panic_write(struct mtd_info *mtd, loff_t to, size_t len, |
187 | size_t *retlen, const u_char *buf) | |
388bbb09 | 188 | { |
2524534d | 189 | struct mtd_part *part = mtd_to_part(mtd); |
994c8409 MD |
190 | return part->master->_panic_write(part->master, to + part->offset, len, |
191 | retlen, buf); | |
388bbb09 RP |
192 | } |
193 | ||
8593fbc6 | 194 | static int part_write_oob(struct mtd_info *mtd, loff_t to, |
b33a2887 | 195 | struct mtd_oob_ops *ops) |
1da177e4 | 196 | { |
2524534d | 197 | struct mtd_part *part = mtd_to_part(mtd); |
8593fbc6 | 198 | |
1da177e4 | 199 | if (to >= mtd->size) |
8593fbc6 | 200 | return -EINVAL; |
7014568b | 201 | if (ops->datbuf && to + ops->len > mtd->size) |
8593fbc6 | 202 | return -EINVAL; |
994c8409 | 203 | return part->master->_write_oob(part->master, to + part->offset, ops); |
1da177e4 LT |
204 | } |
205 | ||
b33a2887 AN |
206 | static int part_write_user_prot_reg(struct mtd_info *mtd, loff_t from, |
207 | size_t len, size_t *retlen, u_char *buf) | |
1da177e4 | 208 | { |
2524534d | 209 | struct mtd_part *part = mtd_to_part(mtd); |
994c8409 MD |
210 | return part->master->_write_user_prot_reg(part->master, from, len, |
211 | retlen, buf); | |
1da177e4 LT |
212 | } |
213 | ||
b33a2887 AN |
214 | static int part_lock_user_prot_reg(struct mtd_info *mtd, loff_t from, |
215 | size_t len) | |
f77814dd | 216 | { |
2524534d | 217 | struct mtd_part *part = mtd_to_part(mtd); |
994c8409 | 218 | return part->master->_lock_user_prot_reg(part->master, from, len); |
f77814dd NP |
219 | } |
220 | ||
b33a2887 AN |
221 | static int part_writev(struct mtd_info *mtd, const struct kvec *vecs, |
222 | unsigned long count, loff_t to, size_t *retlen) | |
1da177e4 | 223 | { |
2524534d | 224 | struct mtd_part *part = mtd_to_part(mtd); |
994c8409 MD |
225 | return part->master->_writev(part->master, vecs, count, |
226 | to + part->offset, retlen); | |
1da177e4 LT |
227 | } |
228 | ||
b33a2887 | 229 | static int part_erase(struct mtd_info *mtd, struct erase_info *instr) |
1da177e4 | 230 | { |
2524534d | 231 | struct mtd_part *part = mtd_to_part(mtd); |
1da177e4 | 232 | int ret; |
664addc2 | 233 | |
1da177e4 | 234 | instr->addr += part->offset; |
994c8409 | 235 | ret = part->master->_erase(part->master, instr); |
74641d75 | 236 | if (ret) { |
bb0eb217 | 237 | if (instr->fail_addr != MTD_FAIL_ADDR_UNKNOWN) |
74641d75 AH |
238 | instr->fail_addr -= part->offset; |
239 | instr->addr -= part->offset; | |
240 | } | |
1da177e4 LT |
241 | return ret; |
242 | } | |
243 | ||
244 | void mtd_erase_callback(struct erase_info *instr) | |
245 | { | |
3c3c10bb | 246 | if (instr->mtd->_erase == part_erase) { |
2524534d | 247 | struct mtd_part *part = mtd_to_part(instr->mtd); |
1da177e4 | 248 | |
bb0eb217 | 249 | if (instr->fail_addr != MTD_FAIL_ADDR_UNKNOWN) |
1da177e4 LT |
250 | instr->fail_addr -= part->offset; |
251 | instr->addr -= part->offset; | |
252 | } | |
253 | if (instr->callback) | |
254 | instr->callback(instr); | |
255 | } | |
256 | EXPORT_SYMBOL_GPL(mtd_erase_callback); | |
257 | ||
69423d99 | 258 | static int part_lock(struct mtd_info *mtd, loff_t ofs, uint64_t len) |
1da177e4 | 259 | { |
2524534d | 260 | struct mtd_part *part = mtd_to_part(mtd); |
994c8409 | 261 | return part->master->_lock(part->master, ofs + part->offset, len); |
1da177e4 LT |
262 | } |
263 | ||
69423d99 | 264 | static int part_unlock(struct mtd_info *mtd, loff_t ofs, uint64_t len) |
1da177e4 | 265 | { |
2524534d | 266 | struct mtd_part *part = mtd_to_part(mtd); |
994c8409 | 267 | return part->master->_unlock(part->master, ofs + part->offset, len); |
1da177e4 LT |
268 | } |
269 | ||
9938424f RC |
270 | static int part_is_locked(struct mtd_info *mtd, loff_t ofs, uint64_t len) |
271 | { | |
2524534d | 272 | struct mtd_part *part = mtd_to_part(mtd); |
994c8409 | 273 | return part->master->_is_locked(part->master, ofs + part->offset, len); |
9938424f RC |
274 | } |
275 | ||
1da177e4 LT |
276 | static void part_sync(struct mtd_info *mtd) |
277 | { | |
2524534d | 278 | struct mtd_part *part = mtd_to_part(mtd); |
994c8409 | 279 | part->master->_sync(part->master); |
1da177e4 LT |
280 | } |
281 | ||
282 | static int part_suspend(struct mtd_info *mtd) | |
283 | { | |
2524534d | 284 | struct mtd_part *part = mtd_to_part(mtd); |
994c8409 | 285 | return part->master->_suspend(part->master); |
1da177e4 LT |
286 | } |
287 | ||
288 | static void part_resume(struct mtd_info *mtd) | |
289 | { | |
2524534d | 290 | struct mtd_part *part = mtd_to_part(mtd); |
994c8409 | 291 | part->master->_resume(part->master); |
1da177e4 LT |
292 | } |
293 | ||
8471bb73 EG |
294 | static int part_block_isreserved(struct mtd_info *mtd, loff_t ofs) |
295 | { | |
2524534d | 296 | struct mtd_part *part = mtd_to_part(mtd); |
8471bb73 EG |
297 | ofs += part->offset; |
298 | return part->master->_block_isreserved(part->master, ofs); | |
299 | } | |
300 | ||
b33a2887 | 301 | static int part_block_isbad(struct mtd_info *mtd, loff_t ofs) |
1da177e4 | 302 | { |
2524534d | 303 | struct mtd_part *part = mtd_to_part(mtd); |
1da177e4 | 304 | ofs += part->offset; |
994c8409 | 305 | return part->master->_block_isbad(part->master, ofs); |
1da177e4 LT |
306 | } |
307 | ||
b33a2887 | 308 | static int part_block_markbad(struct mtd_info *mtd, loff_t ofs) |
1da177e4 | 309 | { |
2524534d | 310 | struct mtd_part *part = mtd_to_part(mtd); |
f1a28c02 TG |
311 | int res; |
312 | ||
1da177e4 | 313 | ofs += part->offset; |
994c8409 | 314 | res = part->master->_block_markbad(part->master, ofs); |
f1a28c02 TG |
315 | if (!res) |
316 | mtd->ecc_stats.badblocks++; | |
317 | return res; | |
1da177e4 LT |
318 | } |
319 | ||
5daa7b21 RT |
320 | static inline void free_partition(struct mtd_part *p) |
321 | { | |
322 | kfree(p->mtd.name); | |
323 | kfree(p); | |
324 | } | |
325 | ||
97894cda TG |
326 | /* |
327 | * This function unregisters and destroy all slave MTD objects which are | |
1da177e4 LT |
328 | * attached to the given master MTD object. |
329 | */ | |
330 | ||
331 | int del_mtd_partitions(struct mtd_info *master) | |
332 | { | |
71a928c0 | 333 | struct mtd_part *slave, *next; |
5daa7b21 | 334 | int ret, err = 0; |
1da177e4 | 335 | |
5daa7b21 | 336 | mutex_lock(&mtd_partitions_mutex); |
71a928c0 | 337 | list_for_each_entry_safe(slave, next, &mtd_partitions, list) |
1da177e4 | 338 | if (slave->master == master) { |
5daa7b21 RT |
339 | ret = del_mtd_device(&slave->mtd); |
340 | if (ret < 0) { | |
341 | err = ret; | |
342 | continue; | |
343 | } | |
71a928c0 | 344 | list_del(&slave->list); |
5daa7b21 | 345 | free_partition(slave); |
1da177e4 | 346 | } |
5daa7b21 | 347 | mutex_unlock(&mtd_partitions_mutex); |
1da177e4 | 348 | |
5daa7b21 | 349 | return err; |
1da177e4 LT |
350 | } |
351 | ||
5daa7b21 RT |
352 | static struct mtd_part *allocate_partition(struct mtd_info *master, |
353 | const struct mtd_partition *part, int partno, | |
354 | uint64_t cur_offset) | |
7788ba71 AN |
355 | { |
356 | struct mtd_part *slave; | |
5daa7b21 | 357 | char *name; |
7788ba71 AN |
358 | |
359 | /* allocate the partition structure */ | |
b33a2887 | 360 | slave = kzalloc(sizeof(*slave), GFP_KERNEL); |
5daa7b21 RT |
361 | name = kstrdup(part->name, GFP_KERNEL); |
362 | if (!name || !slave) { | |
b33a2887 | 363 | printk(KERN_ERR"memory allocation error while creating partitions for \"%s\"\n", |
5daa7b21 RT |
364 | master->name); |
365 | kfree(name); | |
366 | kfree(slave); | |
367 | return ERR_PTR(-ENOMEM); | |
7788ba71 | 368 | } |
7788ba71 AN |
369 | |
370 | /* set up the MTD object for this partition */ | |
371 | slave->mtd.type = master->type; | |
372 | slave->mtd.flags = master->flags & ~part->mask_flags; | |
373 | slave->mtd.size = part->size; | |
374 | slave->mtd.writesize = master->writesize; | |
7fa33ac0 | 375 | slave->mtd.writebufsize = master->writebufsize; |
7788ba71 AN |
376 | slave->mtd.oobsize = master->oobsize; |
377 | slave->mtd.oobavail = master->oobavail; | |
378 | slave->mtd.subpage_sft = master->subpage_sft; | |
379 | ||
5daa7b21 | 380 | slave->mtd.name = name; |
7788ba71 AN |
381 | slave->mtd.owner = master->owner; |
382 | ||
727dc612 DE |
383 | /* NOTE: Historically, we didn't arrange MTDs as a tree out of |
384 | * concern for showing the same data in multiple partitions. | |
385 | * However, it is very useful to have the master node present, | |
386 | * so the MTD_PARTITIONED_MASTER option allows that. The master | |
387 | * will have device nodes etc only if this is set, so make the | |
388 | * parent conditional on that option. Note, this is a way to | |
389 | * distinguish between the master and the partition in sysfs. | |
1f24b5a8 | 390 | */ |
727dc612 DE |
391 | slave->mtd.dev.parent = IS_ENABLED(CONFIG_MTD_PARTITIONED_MASTER) ? |
392 | &master->dev : | |
393 | master->dev.parent; | |
1f24b5a8 | 394 | |
3c3c10bb AB |
395 | slave->mtd._read = part_read; |
396 | slave->mtd._write = part_write; | |
7788ba71 | 397 | |
3c3c10bb AB |
398 | if (master->_panic_write) |
399 | slave->mtd._panic_write = part_panic_write; | |
7788ba71 | 400 | |
3c3c10bb AB |
401 | if (master->_point && master->_unpoint) { |
402 | slave->mtd._point = part_point; | |
403 | slave->mtd._unpoint = part_unpoint; | |
7788ba71 AN |
404 | } |
405 | ||
3c3c10bb AB |
406 | if (master->_get_unmapped_area) |
407 | slave->mtd._get_unmapped_area = part_get_unmapped_area; | |
408 | if (master->_read_oob) | |
409 | slave->mtd._read_oob = part_read_oob; | |
410 | if (master->_write_oob) | |
411 | slave->mtd._write_oob = part_write_oob; | |
412 | if (master->_read_user_prot_reg) | |
413 | slave->mtd._read_user_prot_reg = part_read_user_prot_reg; | |
414 | if (master->_read_fact_prot_reg) | |
415 | slave->mtd._read_fact_prot_reg = part_read_fact_prot_reg; | |
416 | if (master->_write_user_prot_reg) | |
417 | slave->mtd._write_user_prot_reg = part_write_user_prot_reg; | |
418 | if (master->_lock_user_prot_reg) | |
419 | slave->mtd._lock_user_prot_reg = part_lock_user_prot_reg; | |
420 | if (master->_get_user_prot_info) | |
421 | slave->mtd._get_user_prot_info = part_get_user_prot_info; | |
422 | if (master->_get_fact_prot_info) | |
423 | slave->mtd._get_fact_prot_info = part_get_fact_prot_info; | |
424 | if (master->_sync) | |
425 | slave->mtd._sync = part_sync; | |
426 | if (!partno && !master->dev.class && master->_suspend && | |
427 | master->_resume) { | |
428 | slave->mtd._suspend = part_suspend; | |
429 | slave->mtd._resume = part_resume; | |
7788ba71 | 430 | } |
3c3c10bb AB |
431 | if (master->_writev) |
432 | slave->mtd._writev = part_writev; | |
433 | if (master->_lock) | |
434 | slave->mtd._lock = part_lock; | |
435 | if (master->_unlock) | |
436 | slave->mtd._unlock = part_unlock; | |
437 | if (master->_is_locked) | |
438 | slave->mtd._is_locked = part_is_locked; | |
8471bb73 EG |
439 | if (master->_block_isreserved) |
440 | slave->mtd._block_isreserved = part_block_isreserved; | |
3c3c10bb AB |
441 | if (master->_block_isbad) |
442 | slave->mtd._block_isbad = part_block_isbad; | |
443 | if (master->_block_markbad) | |
444 | slave->mtd._block_markbad = part_block_markbad; | |
445 | slave->mtd._erase = part_erase; | |
7788ba71 AN |
446 | slave->master = master; |
447 | slave->offset = part->offset; | |
7788ba71 AN |
448 | |
449 | if (slave->offset == MTDPART_OFS_APPEND) | |
450 | slave->offset = cur_offset; | |
451 | if (slave->offset == MTDPART_OFS_NXTBLK) { | |
452 | slave->offset = cur_offset; | |
69423d99 | 453 | if (mtd_mod_by_eb(cur_offset, master) != 0) { |
7788ba71 | 454 | /* Round up to next erasesize */ |
69423d99 | 455 | slave->offset = (mtd_div_by_eb(cur_offset, master) + 1) * master->erasesize; |
7788ba71 | 456 | printk(KERN_NOTICE "Moving partition %d: " |
69423d99 AH |
457 | "0x%012llx -> 0x%012llx\n", partno, |
458 | (unsigned long long)cur_offset, (unsigned long long)slave->offset); | |
7788ba71 AN |
459 | } |
460 | } | |
1a31368b DES |
461 | if (slave->offset == MTDPART_OFS_RETAIN) { |
462 | slave->offset = cur_offset; | |
463 | if (master->size - slave->offset >= slave->mtd.size) { | |
464 | slave->mtd.size = master->size - slave->offset | |
465 | - slave->mtd.size; | |
466 | } else { | |
467 | printk(KERN_ERR "mtd partition \"%s\" doesn't have enough space: %#llx < %#llx, disabled\n", | |
468 | part->name, master->size - slave->offset, | |
469 | slave->mtd.size); | |
470 | /* register to preserve ordering */ | |
471 | goto out_register; | |
472 | } | |
473 | } | |
7788ba71 AN |
474 | if (slave->mtd.size == MTDPART_SIZ_FULL) |
475 | slave->mtd.size = master->size - slave->offset; | |
476 | ||
69423d99 AH |
477 | printk(KERN_NOTICE "0x%012llx-0x%012llx : \"%s\"\n", (unsigned long long)slave->offset, |
478 | (unsigned long long)(slave->offset + slave->mtd.size), slave->mtd.name); | |
7788ba71 AN |
479 | |
480 | /* let's do some sanity checks */ | |
481 | if (slave->offset >= master->size) { | |
f636ffb4 | 482 | /* let's register it anyway to preserve ordering */ |
7788ba71 AN |
483 | slave->offset = 0; |
484 | slave->mtd.size = 0; | |
b33a2887 | 485 | printk(KERN_ERR"mtd: partition \"%s\" is out of reach -- disabled\n", |
7788ba71 | 486 | part->name); |
f636ffb4 | 487 | goto out_register; |
7788ba71 AN |
488 | } |
489 | if (slave->offset + slave->mtd.size > master->size) { | |
490 | slave->mtd.size = master->size - slave->offset; | |
69423d99 AH |
491 | printk(KERN_WARNING"mtd: partition \"%s\" extends beyond the end of device \"%s\" -- size truncated to %#llx\n", |
492 | part->name, master->name, (unsigned long long)slave->mtd.size); | |
7788ba71 | 493 | } |
b33a2887 | 494 | if (master->numeraseregions > 1) { |
7788ba71 | 495 | /* Deal with variable erase size stuff */ |
6910c136 | 496 | int i, max = master->numeraseregions; |
69423d99 | 497 | u64 end = slave->offset + slave->mtd.size; |
7788ba71 AN |
498 | struct mtd_erase_region_info *regions = master->eraseregions; |
499 | ||
6910c136 AN |
500 | /* Find the first erase regions which is part of this |
501 | * partition. */ | |
502 | for (i = 0; i < max && regions[i].offset <= slave->offset; i++) | |
7788ba71 | 503 | ; |
6910c136 | 504 | /* The loop searched for the region _behind_ the first one */ |
a57ca046 RK |
505 | if (i > 0) |
506 | i--; | |
7788ba71 | 507 | |
6910c136 AN |
508 | /* Pick biggest erasesize */ |
509 | for (; i < max && regions[i].offset < end; i++) { | |
7788ba71 AN |
510 | if (slave->mtd.erasesize < regions[i].erasesize) { |
511 | slave->mtd.erasesize = regions[i].erasesize; | |
512 | } | |
513 | } | |
6910c136 | 514 | BUG_ON(slave->mtd.erasesize == 0); |
7788ba71 AN |
515 | } else { |
516 | /* Single erase size */ | |
517 | slave->mtd.erasesize = master->erasesize; | |
518 | } | |
519 | ||
520 | if ((slave->mtd.flags & MTD_WRITEABLE) && | |
69423d99 | 521 | mtd_mod_by_eb(slave->offset, &slave->mtd)) { |
7788ba71 | 522 | /* Doesn't start on a boundary of major erase size */ |
b33a2887 AN |
523 | /* FIXME: Let it be writable if it is on a boundary of |
524 | * _minor_ erase size though */ | |
7788ba71 | 525 | slave->mtd.flags &= ~MTD_WRITEABLE; |
b33a2887 | 526 | printk(KERN_WARNING"mtd: partition \"%s\" doesn't start on an erase block boundary -- force read-only\n", |
7788ba71 AN |
527 | part->name); |
528 | } | |
529 | if ((slave->mtd.flags & MTD_WRITEABLE) && | |
69423d99 | 530 | mtd_mod_by_eb(slave->mtd.size, &slave->mtd)) { |
7788ba71 | 531 | slave->mtd.flags &= ~MTD_WRITEABLE; |
b33a2887 | 532 | printk(KERN_WARNING"mtd: partition \"%s\" doesn't end on an erase block -- force read-only\n", |
7788ba71 AN |
533 | part->name); |
534 | } | |
535 | ||
536 | slave->mtd.ecclayout = master->ecclayout; | |
bdf69c47 | 537 | slave->mtd.ecc_step_size = master->ecc_step_size; |
6a918bad | 538 | slave->mtd.ecc_strength = master->ecc_strength; |
d062d4ed MD |
539 | slave->mtd.bitflip_threshold = master->bitflip_threshold; |
540 | ||
3c3c10bb | 541 | if (master->_block_isbad) { |
69423d99 | 542 | uint64_t offs = 0; |
7788ba71 | 543 | |
b33a2887 | 544 | while (offs < slave->mtd.size) { |
fdf43a42 EG |
545 | if (mtd_block_isreserved(master, offs + slave->offset)) |
546 | slave->mtd.ecc_stats.bbtblocks++; | |
547 | else if (mtd_block_isbad(master, offs + slave->offset)) | |
7788ba71 AN |
548 | slave->mtd.ecc_stats.badblocks++; |
549 | offs += slave->mtd.erasesize; | |
550 | } | |
551 | } | |
552 | ||
f636ffb4 | 553 | out_register: |
7788ba71 AN |
554 | return slave; |
555 | } | |
556 | ||
a62c24d7 DE |
557 | static ssize_t mtd_partition_offset_show(struct device *dev, |
558 | struct device_attribute *attr, char *buf) | |
559 | { | |
560 | struct mtd_info *mtd = dev_get_drvdata(dev); | |
2524534d | 561 | struct mtd_part *part = mtd_to_part(mtd); |
a62c24d7 DE |
562 | return snprintf(buf, PAGE_SIZE, "%lld\n", part->offset); |
563 | } | |
564 | ||
565 | static DEVICE_ATTR(offset, S_IRUGO, mtd_partition_offset_show, NULL); | |
566 | ||
567 | static const struct attribute *mtd_partition_attrs[] = { | |
568 | &dev_attr_offset.attr, | |
569 | NULL | |
570 | }; | |
571 | ||
572 | static int mtd_add_partition_attrs(struct mtd_part *new) | |
573 | { | |
574 | int ret = sysfs_create_files(&new->mtd.dev.kobj, mtd_partition_attrs); | |
575 | if (ret) | |
576 | printk(KERN_WARNING | |
577 | "mtd: failed to create partition attrs, err=%d\n", ret); | |
578 | return ret; | |
579 | } | |
580 | ||
26a6d240 | 581 | int mtd_add_partition(struct mtd_info *master, const char *name, |
5daa7b21 RT |
582 | long long offset, long long length) |
583 | { | |
584 | struct mtd_partition part; | |
3a434f66 | 585 | struct mtd_part *new; |
5daa7b21 RT |
586 | int ret = 0; |
587 | ||
588 | /* the direct offset is expected */ | |
589 | if (offset == MTDPART_OFS_APPEND || | |
590 | offset == MTDPART_OFS_NXTBLK) | |
591 | return -EINVAL; | |
592 | ||
593 | if (length == MTDPART_SIZ_FULL) | |
594 | length = master->size - offset; | |
595 | ||
596 | if (length <= 0) | |
597 | return -EINVAL; | |
598 | ||
93867233 | 599 | memset(&part, 0, sizeof(part)); |
5daa7b21 RT |
600 | part.name = name; |
601 | part.size = length; | |
602 | part.offset = offset; | |
5daa7b21 RT |
603 | |
604 | new = allocate_partition(master, &part, -1, offset); | |
605 | if (IS_ERR(new)) | |
606 | return PTR_ERR(new); | |
607 | ||
5daa7b21 | 608 | mutex_lock(&mtd_partitions_mutex); |
5daa7b21 RT |
609 | list_add(&new->list, &mtd_partitions); |
610 | mutex_unlock(&mtd_partitions_mutex); | |
611 | ||
612 | add_mtd_device(&new->mtd); | |
613 | ||
a62c24d7 DE |
614 | mtd_add_partition_attrs(new); |
615 | ||
5daa7b21 | 616 | return ret; |
5daa7b21 RT |
617 | } |
618 | EXPORT_SYMBOL_GPL(mtd_add_partition); | |
619 | ||
620 | int mtd_del_partition(struct mtd_info *master, int partno) | |
621 | { | |
622 | struct mtd_part *slave, *next; | |
623 | int ret = -EINVAL; | |
624 | ||
625 | mutex_lock(&mtd_partitions_mutex); | |
626 | list_for_each_entry_safe(slave, next, &mtd_partitions, list) | |
627 | if ((slave->master == master) && | |
628 | (slave->mtd.index == partno)) { | |
a62c24d7 DE |
629 | sysfs_remove_files(&slave->mtd.dev.kobj, |
630 | mtd_partition_attrs); | |
5daa7b21 RT |
631 | ret = del_mtd_device(&slave->mtd); |
632 | if (ret < 0) | |
633 | break; | |
634 | ||
635 | list_del(&slave->list); | |
636 | free_partition(slave); | |
637 | break; | |
638 | } | |
639 | mutex_unlock(&mtd_partitions_mutex); | |
640 | ||
641 | return ret; | |
642 | } | |
643 | EXPORT_SYMBOL_GPL(mtd_del_partition); | |
644 | ||
1da177e4 LT |
645 | /* |
646 | * This function, given a master MTD object and a partition table, creates | |
647 | * and registers slave MTD objects which are bound to the master according to | |
648 | * the partition definitions. | |
1f24b5a8 | 649 | * |
727dc612 DE |
650 | * For historical reasons, this function's caller only registers the master |
651 | * if the MTD_PARTITIONED_MASTER config option is set. | |
1da177e4 LT |
652 | */ |
653 | ||
97894cda | 654 | int add_mtd_partitions(struct mtd_info *master, |
1da177e4 LT |
655 | const struct mtd_partition *parts, |
656 | int nbparts) | |
657 | { | |
658 | struct mtd_part *slave; | |
69423d99 | 659 | uint64_t cur_offset = 0; |
1da177e4 LT |
660 | int i; |
661 | ||
b33a2887 | 662 | printk(KERN_NOTICE "Creating %d MTD partitions on \"%s\":\n", nbparts, master->name); |
1da177e4 LT |
663 | |
664 | for (i = 0; i < nbparts; i++) { | |
5daa7b21 | 665 | slave = allocate_partition(master, parts + i, i, cur_offset); |
e5bae867 BB |
666 | if (IS_ERR(slave)) { |
667 | del_mtd_partitions(master); | |
5daa7b21 | 668 | return PTR_ERR(slave); |
e5bae867 | 669 | } |
5daa7b21 RT |
670 | |
671 | mutex_lock(&mtd_partitions_mutex); | |
672 | list_add(&slave->list, &mtd_partitions); | |
673 | mutex_unlock(&mtd_partitions_mutex); | |
674 | ||
675 | add_mtd_device(&slave->mtd); | |
a62c24d7 | 676 | mtd_add_partition_attrs(slave); |
5daa7b21 | 677 | |
1da177e4 | 678 | cur_offset = slave->offset + slave->mtd.size; |
1da177e4 LT |
679 | } |
680 | ||
681 | return 0; | |
682 | } | |
1da177e4 LT |
683 | |
684 | static DEFINE_SPINLOCK(part_parser_lock); | |
685 | static LIST_HEAD(part_parsers); | |
686 | ||
5531ae48 | 687 | static struct mtd_part_parser *mtd_part_parser_get(const char *name) |
1da177e4 | 688 | { |
71a928c0 | 689 | struct mtd_part_parser *p, *ret = NULL; |
1da177e4 | 690 | |
71a928c0 | 691 | spin_lock(&part_parser_lock); |
1da177e4 | 692 | |
71a928c0 | 693 | list_for_each_entry(p, &part_parsers, list) |
1da177e4 LT |
694 | if (!strcmp(p->name, name) && try_module_get(p->owner)) { |
695 | ret = p; | |
696 | break; | |
697 | } | |
71a928c0 | 698 | |
1da177e4 LT |
699 | spin_unlock(&part_parser_lock); |
700 | ||
701 | return ret; | |
702 | } | |
703 | ||
5531ae48 BN |
704 | static inline void mtd_part_parser_put(const struct mtd_part_parser *p) |
705 | { | |
706 | module_put(p->owner); | |
707 | } | |
953b3bd1 | 708 | |
adc83bf8 BN |
709 | /* |
710 | * Many partition parsers just expected the core to kfree() all their data in | |
711 | * one chunk. Do that by default. | |
712 | */ | |
713 | static void mtd_part_parser_cleanup_default(const struct mtd_partition *pparts, | |
714 | int nr_parts) | |
715 | { | |
716 | kfree(pparts); | |
717 | } | |
718 | ||
b9eab011 | 719 | int __register_mtd_parser(struct mtd_part_parser *p, struct module *owner) |
1da177e4 | 720 | { |
b9eab011 BN |
721 | p->owner = owner; |
722 | ||
adc83bf8 BN |
723 | if (!p->cleanup) |
724 | p->cleanup = &mtd_part_parser_cleanup_default; | |
725 | ||
1da177e4 LT |
726 | spin_lock(&part_parser_lock); |
727 | list_add(&p->list, &part_parsers); | |
728 | spin_unlock(&part_parser_lock); | |
b9eab011 BN |
729 | |
730 | return 0; | |
1da177e4 | 731 | } |
b9eab011 | 732 | EXPORT_SYMBOL_GPL(__register_mtd_parser); |
1da177e4 | 733 | |
cf3b2b1e | 734 | void deregister_mtd_parser(struct mtd_part_parser *p) |
1da177e4 LT |
735 | { |
736 | spin_lock(&part_parser_lock); | |
737 | list_del(&p->list); | |
738 | spin_unlock(&part_parser_lock); | |
1da177e4 | 739 | } |
b33a2887 | 740 | EXPORT_SYMBOL_GPL(deregister_mtd_parser); |
1da177e4 | 741 | |
ad274cec AB |
742 | /* |
743 | * Do not forget to update 'parse_mtd_partitions()' kerneldoc comment if you | |
744 | * are changing this array! | |
745 | */ | |
ccef4dcc | 746 | static const char * const default_mtd_part_types[] = { |
d26c87d6 DES |
747 | "cmdlinepart", |
748 | "ofpart", | |
749 | NULL | |
750 | }; | |
5c4eefbd | 751 | |
ad274cec AB |
752 | /** |
753 | * parse_mtd_partitions - parse MTD partitions | |
754 | * @master: the master partition (describes whole MTD device) | |
755 | * @types: names of partition parsers to try or %NULL | |
07fd2f87 | 756 | * @pparts: info about partitions found is returned here |
c7975330 | 757 | * @data: MTD partition parser-specific data |
ad274cec AB |
758 | * |
759 | * This function tries to find partition on MTD device @master. It uses MTD | |
760 | * partition parsers, specified in @types. However, if @types is %NULL, then | |
761 | * the default list of parsers is used. The default list contains only the | |
d26c87d6 | 762 | * "cmdlinepart" and "ofpart" parsers ATM. |
c51803dd HS |
763 | * Note: If there are more then one parser in @types, the kernel only takes the |
764 | * partitions parsed out by the first parser. | |
ad274cec AB |
765 | * |
766 | * This function may return: | |
767 | * o a negative error code in case of failure | |
07fd2f87 | 768 | * o zero otherwise, and @pparts will describe the partitions, number of |
adc83bf8 BN |
769 | * partitions, and the parser which parsed them. Caller must release |
770 | * resources with mtd_part_parser_cleanup() when finished with the returned | |
771 | * data. | |
ad274cec | 772 | */ |
26a47346 | 773 | int parse_mtd_partitions(struct mtd_info *master, const char *const *types, |
07fd2f87 | 774 | struct mtd_partitions *pparts, |
c7975330 | 775 | struct mtd_part_parser_data *data) |
1da177e4 LT |
776 | { |
777 | struct mtd_part_parser *parser; | |
5a2415b0 | 778 | int ret, err = 0; |
97894cda | 779 | |
5c4eefbd DES |
780 | if (!types) |
781 | types = default_mtd_part_types; | |
782 | ||
5a2415b0 | 783 | for ( ; *types; types++) { |
8e2c992b | 784 | pr_debug("%s: parsing partitions %s\n", master->name, *types); |
5531ae48 | 785 | parser = mtd_part_parser_get(*types); |
1da177e4 | 786 | if (!parser && !request_module("%s", *types)) |
5531ae48 | 787 | parser = mtd_part_parser_get(*types); |
8e2c992b MS |
788 | pr_debug("%s: got parser %s\n", master->name, |
789 | parser ? parser->name : NULL); | |
7c802fbd | 790 | if (!parser) |
1da177e4 | 791 | continue; |
07fd2f87 | 792 | ret = (*parser->parse_fn)(master, &pparts->parts, data); |
8e2c992b MS |
793 | pr_debug("%s: parser %s: %i\n", |
794 | master->name, parser->name, ret); | |
1da177e4 | 795 | if (ret > 0) { |
97894cda | 796 | printk(KERN_NOTICE "%d %s partitions found on MTD device %s\n", |
1da177e4 | 797 | ret, parser->name, master->name); |
07fd2f87 BN |
798 | pparts->nr_parts = ret; |
799 | pparts->parser = parser; | |
800 | return 0; | |
1da177e4 | 801 | } |
adc83bf8 | 802 | mtd_part_parser_put(parser); |
5a2415b0 BN |
803 | /* |
804 | * Stash the first error we see; only report it if no parser | |
805 | * succeeds | |
806 | */ | |
807 | if (ret < 0 && !err) | |
808 | err = ret; | |
1da177e4 | 809 | } |
5a2415b0 | 810 | return err; |
1da177e4 | 811 | } |
5daa7b21 | 812 | |
adc83bf8 BN |
813 | void mtd_part_parser_cleanup(struct mtd_partitions *parts) |
814 | { | |
815 | const struct mtd_part_parser *parser; | |
816 | ||
817 | if (!parts) | |
818 | return; | |
819 | ||
820 | parser = parts->parser; | |
821 | if (parser) { | |
822 | if (parser->cleanup) | |
823 | parser->cleanup(parts->parts, parts->nr_parts); | |
824 | ||
825 | mtd_part_parser_put(parser); | |
826 | } | |
827 | } | |
828 | ||
5dee4674 | 829 | int mtd_is_partition(const struct mtd_info *mtd) |
5daa7b21 RT |
830 | { |
831 | struct mtd_part *part; | |
a7e93dcd | 832 | int ispart = 0; |
5daa7b21 RT |
833 | |
834 | mutex_lock(&mtd_partitions_mutex); | |
835 | list_for_each_entry(part, &mtd_partitions, list) | |
836 | if (&part->mtd == mtd) { | |
a7e93dcd | 837 | ispart = 1; |
5daa7b21 RT |
838 | break; |
839 | } | |
840 | mutex_unlock(&mtd_partitions_mutex); | |
841 | ||
a7e93dcd | 842 | return ispart; |
5daa7b21 | 843 | } |
a7e93dcd | 844 | EXPORT_SYMBOL_GPL(mtd_is_partition); |
62082e56 RG |
845 | |
846 | /* Returns the size of the entire flash chip */ | |
847 | uint64_t mtd_get_device_size(const struct mtd_info *mtd) | |
848 | { | |
849 | if (!mtd_is_partition(mtd)) | |
850 | return mtd->size; | |
851 | ||
2524534d | 852 | return mtd_to_part(mtd)->master->size; |
62082e56 RG |
853 | } |
854 | EXPORT_SYMBOL_GPL(mtd_get_device_size); |