Merge git://git.kernel.org/pub/scm/linux/kernel/git/davem/sparc
[deliverable/linux.git] / arch / tile / kernel / tlb.c
CommitLineData
867e359b
CM
1/*
2 * Copyright 2010 Tilera Corporation. All Rights Reserved.
3 *
4 * This program is free software; you can redistribute it and/or
5 * modify it under the terms of the GNU General Public License
6 * as published by the Free Software Foundation, version 2.
7 *
8 * This program is distributed in the hope that it will be useful, but
9 * WITHOUT ANY WARRANTY; without even the implied warranty of
10 * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
11 * NON INFRINGEMENT. See the GNU General Public License for
12 * more details.
13 *
14 */
15
16#include <linux/cpumask.h>
17#include <linux/module.h>
621b1955 18#include <linux/hugetlb.h>
867e359b
CM
19#include <asm/tlbflush.h>
20#include <asm/homecache.h>
21#include <hv/hypervisor.h>
22
23/* From tlbflush.h */
24DEFINE_PER_CPU(int, current_asid);
25int min_asid, max_asid;
26
27/*
28 * Note that we flush the L1I (for VM_EXEC pages) as well as the TLB
29 * so that when we are unmapping an executable page, we also flush it.
30 * Combined with flushing the L1I at context switch time, this means
31 * we don't have to do any other icache flushes.
32 */
33
34void flush_tlb_mm(struct mm_struct *mm)
35{
36 HV_Remote_ASID asids[NR_CPUS];
37 int i = 0, cpu;
dc0b124d 38 for_each_cpu(cpu, mm_cpumask(mm)) {
867e359b
CM
39 HV_Remote_ASID *asid = &asids[i++];
40 asid->y = cpu / smp_topology.width;
41 asid->x = cpu % smp_topology.width;
42 asid->asid = per_cpu(current_asid, cpu);
43 }
dc0b124d 44 flush_remote(0, HV_FLUSH_EVICT_L1I, mm_cpumask(mm),
867e359b
CM
45 0, 0, 0, NULL, asids, i);
46}
47
48void flush_tlb_current_task(void)
49{
50 flush_tlb_mm(current->mm);
51}
52
621b1955 53void flush_tlb_page_mm(struct vm_area_struct *vma, struct mm_struct *mm,
867e359b
CM
54 unsigned long va)
55{
621b1955 56 unsigned long size = vma_kernel_pagesize(vma);
867e359b 57 int cache = (vma->vm_flags & VM_EXEC) ? HV_FLUSH_EVICT_L1I : 0;
dc0b124d
KM
58 flush_remote(0, cache, mm_cpumask(mm),
59 va, size, size, mm_cpumask(mm), NULL, 0);
867e359b
CM
60}
61
621b1955 62void flush_tlb_page(struct vm_area_struct *vma, unsigned long va)
867e359b
CM
63{
64 flush_tlb_page_mm(vma, vma->vm_mm, va);
65}
66EXPORT_SYMBOL(flush_tlb_page);
67
621b1955 68void flush_tlb_range(struct vm_area_struct *vma,
867e359b
CM
69 unsigned long start, unsigned long end)
70{
621b1955 71 unsigned long size = vma_kernel_pagesize(vma);
867e359b
CM
72 struct mm_struct *mm = vma->vm_mm;
73 int cache = (vma->vm_flags & VM_EXEC) ? HV_FLUSH_EVICT_L1I : 0;
dc0b124d
KM
74 flush_remote(0, cache, mm_cpumask(mm), start, end - start, size,
75 mm_cpumask(mm), NULL, 0);
867e359b
CM
76}
77
78void flush_tlb_all(void)
79{
80 int i;
81 for (i = 0; ; ++i) {
82 HV_VirtAddrRange r = hv_inquire_virtual(i);
83 if (r.size == 0)
84 break;
85 flush_remote(0, HV_FLUSH_EVICT_L1I, cpu_online_mask,
86 r.start, r.size, PAGE_SIZE, cpu_online_mask,
87 NULL, 0);
88 flush_remote(0, 0, NULL,
89 r.start, r.size, HPAGE_SIZE, cpu_online_mask,
90 NULL, 0);
91 }
92}
93
abe3265a
CM
94/*
95 * Callers need to flush the L1I themselves if necessary, e.g. for
96 * kernel module unload. Otherwise we assume callers are not using
97 * executable pgprot_t's. Using EVICT_L1I means that dataplane cpus
98 * will get an unnecessary interrupt otherwise.
99 */
867e359b
CM
100void flush_tlb_kernel_range(unsigned long start, unsigned long end)
101{
abe3265a 102 flush_remote(0, 0, NULL,
867e359b
CM
103 start, end - start, PAGE_SIZE, cpu_online_mask, NULL, 0);
104}
This page took 0.36719 seconds and 5 git commands to generate.