arm64: remove broken cachepolicy code
[deliverable/linux.git] / arch / arm64 / mm / mmu.c
index 6032f3e3056a6c69899b06ee4c86cc22aa985893..e57c170a91f35f2c77dcffd19ff3d88d97a80202 100644 (file)
 struct page *empty_zero_page;
 EXPORT_SYMBOL(empty_zero_page);
 
-struct cachepolicy {
-       const char      policy[16];
-       u64             mair;
-       u64             tcr;
-};
-
-static struct cachepolicy cache_policies[] __initdata = {
-       {
-               .policy         = "uncached",
-               .mair           = 0x44,                 /* inner, outer non-cacheable */
-               .tcr            = TCR_IRGN_NC | TCR_ORGN_NC,
-       }, {
-               .policy         = "writethrough",
-               .mair           = 0xaa,                 /* inner, outer write-through, read-allocate */
-               .tcr            = TCR_IRGN_WT | TCR_ORGN_WT,
-       }, {
-               .policy         = "writeback",
-               .mair           = 0xee,                 /* inner, outer write-back, read-allocate */
-               .tcr            = TCR_IRGN_WBnWA | TCR_ORGN_WBnWA,
-       }
-};
-
-/*
- * These are useful for identifying cache coherency problems by allowing the
- * cache or the cache and writebuffer to be turned off. It changes the Normal
- * memory caching attributes in the MAIR_EL1 register.
- */
-static int __init early_cachepolicy(char *p)
-{
-       int i;
-       u64 tmp;
-
-       for (i = 0; i < ARRAY_SIZE(cache_policies); i++) {
-               int len = strlen(cache_policies[i].policy);
-
-               if (memcmp(p, cache_policies[i].policy, len) == 0)
-                       break;
-       }
-       if (i == ARRAY_SIZE(cache_policies)) {
-               pr_err("ERROR: unknown or unsupported cache policy: %s\n", p);
-               return 0;
-       }
-
-       flush_cache_all();
-
-       /*
-        * Modify MT_NORMAL attributes in MAIR_EL1.
-        */
-       asm volatile(
-       "       mrs     %0, mair_el1\n"
-       "       bfi     %0, %1, %2, #8\n"
-       "       msr     mair_el1, %0\n"
-       "       isb\n"
-       : "=&r" (tmp)
-       : "r" (cache_policies[i].mair), "i" (MT_NORMAL * 8));
-
-       /*
-        * Modify TCR PTW cacheability attributes.
-        */
-       asm volatile(
-       "       mrs     %0, tcr_el1\n"
-       "       bic     %0, %0, %2\n"
-       "       orr     %0, %0, %1\n"
-       "       msr     tcr_el1, %0\n"
-       "       isb\n"
-       : "=&r" (tmp)
-       : "r" (cache_policies[i].tcr), "r" (TCR_IRGN_MASK | TCR_ORGN_MASK));
-
-       flush_cache_all();
-
-       return 0;
-}
-early_param("cachepolicy", early_cachepolicy);
-
 pgprot_t phys_mem_access_prot(struct file *file, unsigned long pfn,
                              unsigned long size, pgprot_t vma_prot)
 {
@@ -156,29 +82,19 @@ static void __init alloc_init_pte(pmd_t *pmd, unsigned long addr,
        } while (pte++, addr += PAGE_SIZE, addr != end);
 }
 
-static void __init alloc_init_pmd(pud_t *pud, unsigned long addr,
-                                 unsigned long end, phys_addr_t phys,
-                                 int map_io)
+static void __init alloc_init_pmd(struct mm_struct *mm, pud_t *pud,
+                                 unsigned long addr, unsigned long end,
+                                 phys_addr_t phys, pgprot_t prot)
 {
        pmd_t *pmd;
        unsigned long next;
-       pmdval_t prot_sect;
-       pgprot_t prot_pte;
-
-       if (map_io) {
-               prot_sect = PROT_SECT_DEVICE_nGnRE;
-               prot_pte = __pgprot(PROT_DEVICE_nGnRE);
-       } else {
-               prot_sect = PROT_SECT_NORMAL_EXEC;
-               prot_pte = PAGE_KERNEL_EXEC;
-       }
 
        /*
         * Check for initial section mappings in the pgd/pud and remove them.
         */
        if (pud_none(*pud) || pud_bad(*pud)) {
                pmd = early_alloc(PTRS_PER_PMD * sizeof(pmd_t));
-               pud_populate(&init_mm, pud, pmd);
+               pud_populate(mm, pud, pmd);
        }
 
        pmd = pmd_offset(pud, addr);
@@ -187,7 +103,8 @@ static void __init alloc_init_pmd(pud_t *pud, unsigned long addr,
                /* try section mapping first */
                if (((addr | next | phys) & ~SECTION_MASK) == 0) {
                        pmd_t old_pmd =*pmd;
-                       set_pmd(pmd, __pmd(phys | prot_sect));
+                       set_pmd(pmd, __pmd(phys |
+                                          pgprot_val(mk_sect_prot(prot))));
                        /*
                         * Check for previous table entries created during
                         * boot (__create_page_tables) and flush them.
@@ -196,22 +113,22 @@ static void __init alloc_init_pmd(pud_t *pud, unsigned long addr,
                                flush_tlb_all();
                } else {
                        alloc_init_pte(pmd, addr, next, __phys_to_pfn(phys),
-                                      prot_pte);
+                                      prot);
                }
                phys += next - addr;
        } while (pmd++, addr = next, addr != end);
 }
 
-static void __init alloc_init_pud(pgd_t *pgd, unsigned long addr,
-                                 unsigned long end, phys_addr_t phys,
-                                 int map_io)
+static void __init alloc_init_pud(struct mm_struct *mm, pgd_t *pgd,
+                                 unsigned long addr, unsigned long end,
+                                 phys_addr_t phys, pgprot_t prot)
 {
        pud_t *pud;
        unsigned long next;
 
        if (pgd_none(*pgd)) {
                pud = early_alloc(PTRS_PER_PUD * sizeof(pud_t));
-               pgd_populate(&init_mm, pgd, pud);
+               pgd_populate(mm, pgd, pud);
        }
        BUG_ON(pgd_bad(*pgd));
 
@@ -222,10 +139,11 @@ static void __init alloc_init_pud(pgd_t *pgd, unsigned long addr,
                /*
                 * For 4K granule only, attempt to put down a 1GB block
                 */
-               if (!map_io && (PAGE_SHIFT == 12) &&
+               if ((PAGE_SHIFT == 12) &&
                    ((addr | next | phys) & ~PUD_MASK) == 0) {
                        pud_t old_pud = *pud;
-                       set_pud(pud, __pud(phys | PROT_SECT_NORMAL_EXEC));
+                       set_pud(pud, __pud(phys |
+                                          pgprot_val(mk_sect_prot(prot))));
 
                        /*
                         * If we have an old value for a pud, it will
@@ -240,7 +158,7 @@ static void __init alloc_init_pud(pgd_t *pgd, unsigned long addr,
                                flush_tlb_all();
                        }
                } else {
-                       alloc_init_pmd(pud, addr, next, phys, map_io);
+                       alloc_init_pmd(mm, pud, addr, next, phys, prot);
                }
                phys += next - addr;
        } while (pud++, addr = next, addr != end);
@@ -250,9 +168,9 @@ static void __init alloc_init_pud(pgd_t *pgd, unsigned long addr,
  * Create the page directory entries and any necessary page tables for the
  * mapping specified by 'md'.
  */
-static void __init __create_mapping(pgd_t *pgd, phys_addr_t phys,
-                                   unsigned long virt, phys_addr_t size,
-                                   int map_io)
+static void __init __create_mapping(struct mm_struct *mm, pgd_t *pgd,
+                                   phys_addr_t phys, unsigned long virt,
+                                   phys_addr_t size, pgprot_t prot)
 {
        unsigned long addr, length, end, next;
 
@@ -262,7 +180,7 @@ static void __init __create_mapping(pgd_t *pgd, phys_addr_t phys,
        end = addr + length;
        do {
                next = pgd_addr_end(addr, end);
-               alloc_init_pud(pgd, addr, next, phys, map_io);
+               alloc_init_pud(mm, pgd, addr, next, phys, prot);
                phys += next - addr;
        } while (pgd++, addr = next, addr != end);
 }
@@ -275,17 +193,15 @@ static void __init create_mapping(phys_addr_t phys, unsigned long virt,
                        &phys, virt);
                return;
        }
-       __create_mapping(pgd_offset_k(virt & PAGE_MASK), phys, virt, size, 0);
+       __create_mapping(&init_mm, pgd_offset_k(virt & PAGE_MASK), phys, virt,
+                        size, PAGE_KERNEL_EXEC);
 }
 
-void __init create_id_mapping(phys_addr_t addr, phys_addr_t size, int map_io)
+void __init create_pgd_mapping(struct mm_struct *mm, phys_addr_t phys,
+                              unsigned long virt, phys_addr_t size,
+                              pgprot_t prot)
 {
-       if ((addr >> PGDIR_SHIFT) >= ARRAY_SIZE(idmap_pg_dir)) {
-               pr_warn("BUG: not creating id mapping for %pa\n", &addr);
-               return;
-       }
-       __create_mapping(&idmap_pg_dir[pgd_index(addr)],
-                        addr, addr, size, map_io);
+       __create_mapping(mm, pgd_offset(mm, virt), phys, virt, size, prot);
 }
 
 static void __init map_mem(void)
This page took 0.033661 seconds and 5 git commands to generate.