/* * Copyright (c) 2006 Jakub Jermar * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * * - Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * - Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * - The name of the author may not be used to endorse or promote products * derived from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ /** @addtogroup ia64mm * @{ */ /** @file */ /* * TLB management. */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include /** Invalidate all TLB entries. */ void tlb_invalidate_all(void) { ipl_t ipl; uintptr_t adr; uint32_t count1, count2, stride1, stride2; int i, j; adr = PAL_PTCE_INFO_BASE(); count1 = PAL_PTCE_INFO_COUNT1(); count2 = PAL_PTCE_INFO_COUNT2(); stride1 = PAL_PTCE_INFO_STRIDE1(); stride2 = PAL_PTCE_INFO_STRIDE2(); ipl = interrupts_disable(); for(i = 0; i < count1; i++) { for(j = 0; j < count2; j++) { asm volatile ( "ptc.e %0 ;;" : : "r" (adr) ); adr += stride2; } adr += stride1; } interrupts_restore(ipl); srlz_d(); srlz_i(); #ifdef CONFIG_VHPT vhpt_invalidate_all(); #endif } /** Invalidate entries belonging to an address space. * * @param asid Address space identifier. */ void tlb_invalidate_asid(asid_t asid) { tlb_invalidate_all(); } void tlb_invalidate_pages(asid_t asid, uintptr_t page, count_t cnt) { region_register rr; bool restore_rr = false; int b = 0; int c = cnt; uintptr_t va; va = page; rr.word = rr_read(VA2VRN(va)); if ((restore_rr = (rr.map.rid != ASID2RID(asid, VA2VRN(va))))) { /* * The selected region register does not contain required RID. * Save the old content of the register and replace the RID. */ region_register rr0; rr0 = rr; rr0.map.rid = ASID2RID(asid, VA2VRN(va)); rr_write(VA2VRN(va), rr0.word); srlz_d(); srlz_i(); } while(c >>= 1) b++; b >>= 1; uint64_t ps; switch (b) { case 0: /*cnt 1-3*/ ps = PAGE_WIDTH; break; case 1: /*cnt 4-15*/ /*cnt=((cnt-1)/4)+1;*/ ps = PAGE_WIDTH+2; va &= ~((1<> PPN_SHIFT; entry.ps = PAGE_WIDTH; if (dtr) dtr_mapping_insert(page, ASID_KERNEL, entry, tr); else dtc_mapping_insert(page, ASID_KERNEL, entry); } /** Purge kernel entries from DTR. * * Purge DTR entries used by the kernel. * * @param page Virtual page address including VRN bits. * @param width Width of the purge in bits. */ void dtr_purge(uintptr_t page, count_t width) { asm volatile ("ptr.d %0, %1\n" : : "r" (page), "r" (width<<2)); } /** Copy content of PTE into data translation cache. * * @param t PTE. */ void dtc_pte_copy(pte_t *t) { tlb_entry_t entry; entry.word[0] = 0; entry.word[1] = 0; entry.p = t->p; entry.ma = t->c ? MA_WRITEBACK : MA_UNCACHEABLE; entry.a = t->a; entry.d = t->d; entry.pl = t->k ? PL_KERNEL : PL_USER; entry.ar = t->w ? AR_WRITE : AR_READ; entry.ppn = t->frame >> PPN_SHIFT; entry.ps = PAGE_WIDTH; dtc_mapping_insert(t->page, t->as->asid, entry); #ifdef CONFIG_VHPT vhpt_mapping_insert(t->page, t->as->asid, entry); #endif } /** Copy content of PTE into instruction translation cache. * * @param t PTE. */ void itc_pte_copy(pte_t *t) { tlb_entry_t entry; entry.word[0] = 0; entry.word[1] = 0; ASSERT(t->x); entry.p = t->p; entry.ma = t->c ? MA_WRITEBACK : MA_UNCACHEABLE; entry.a = t->a; entry.pl = t->k ? PL_KERNEL : PL_USER; entry.ar = t->x ? (AR_EXECUTE | AR_READ) : AR_READ; entry.ppn = t->frame >> PPN_SHIFT; entry.ps = PAGE_WIDTH; itc_mapping_insert(t->page, t->as->asid, entry); #ifdef CONFIG_VHPT vhpt_mapping_insert(t->page, t->as->asid, entry); #endif } /** Instruction TLB fault handler for faults with VHPT turned off. * * @param vector Interruption vector. * @param istate Structure with saved interruption state. */ void alternate_instruction_tlb_fault(uint64_t vector, istate_t *istate) { region_register rr; rid_t rid; uintptr_t va; pte_t *t; va = istate->cr_ifa; /* faulting address */ rr.word = rr_read(VA2VRN(va)); rid = rr.map.rid; page_table_lock(AS, true); t = page_mapping_find(AS, va); if (t) { /* * The mapping was found in software page hash table. * Insert it into data translation cache. */ itc_pte_copy(t); page_table_unlock(AS, true); } else { /* * Forward the page fault to address space page fault handler. */ page_table_unlock(AS, true); if (as_page_fault(va, PF_ACCESS_EXEC, istate) == AS_PF_FAULT) { fault_if_from_uspace(istate,"Page fault at %p",va); panic("%s: va=%p, rid=%d, iip=%p\n", __func__, va, rid, istate->cr_iip); } } } /** Data TLB fault handler for faults with VHPT turned off. * * @param vector Interruption vector. * @param istate Structure with saved interruption state. */ void alternate_data_tlb_fault(uint64_t vector, istate_t *istate) { region_register rr; rid_t rid; uintptr_t va; pte_t *t; va = istate->cr_ifa; /* faulting address */ rr.word = rr_read(VA2VRN(va)); rid = rr.map.rid; if (RID2ASID(rid) == ASID_KERNEL) { if (VA2VRN(va) == VRN_KERNEL) { /* * Provide KA2PA(identity) mapping for faulting piece of * kernel address space. */ dtlb_kernel_mapping_insert(va, KA2PA(va), false, 0); return; } } page_table_lock(AS, true); t = page_mapping_find(AS, va); if (t) { /* * The mapping was found in the software page hash table. * Insert it into data translation cache. */ dtc_pte_copy(t); page_table_unlock(AS, true); } else { /* * Forward the page fault to the address space page fault handler. */ page_table_unlock(AS, true); if (as_page_fault(va, PF_ACCESS_READ, istate) == AS_PF_FAULT) { fault_if_from_uspace(istate,"Page fault at %p",va); panic("%s: va=%p, rid=%d, iip=%p\n", __func__, va, rid, istate->cr_iip); } } } /** Data nested TLB fault handler. * * This fault should not occur. * * @param vector Interruption vector. * @param istate Structure with saved interruption state. */ void data_nested_tlb_fault(uint64_t vector, istate_t *istate) { panic("%s\n", __func__); } /** Data Dirty bit fault handler. * * @param vector Interruption vector. * @param istate Structure with saved interruption state. */ void data_dirty_bit_fault(uint64_t vector, istate_t *istate) { region_register rr; rid_t rid; uintptr_t va; pte_t *t; va = istate->cr_ifa; /* faulting address */ rr.word = rr_read(VA2VRN(va)); rid = rr.map.rid; page_table_lock(AS, true); t = page_mapping_find(AS, va); ASSERT(t && t->p); if (t && t->p && t->w) { /* * Update the Dirty bit in page tables and reinsert * the mapping into DTC. */ t->d = true; dtc_pte_copy(t); } else { if (as_page_fault(va, PF_ACCESS_WRITE, istate) == AS_PF_FAULT) { fault_if_from_uspace(istate,"Page fault at %p",va); panic("%s: va=%p, rid=%d, iip=%p\n", __func__, va, rid, istate->cr_iip); t->d = true; dtc_pte_copy(t); } } page_table_unlock(AS, true); } /** Instruction access bit fault handler. * * @param vector Interruption vector. * @param istate Structure with saved interruption state. */ void instruction_access_bit_fault(uint64_t vector, istate_t *istate) { region_register rr; rid_t rid; uintptr_t va; pte_t *t; va = istate->cr_ifa; /* faulting address */ rr.word = rr_read(VA2VRN(va)); rid = rr.map.rid; page_table_lock(AS, true); t = page_mapping_find(AS, va); ASSERT(t && t->p); if (t && t->p && t->x) { /* * Update the Accessed bit in page tables and reinsert * the mapping into ITC. */ t->a = true; itc_pte_copy(t); } else { if (as_page_fault(va, PF_ACCESS_EXEC, istate) == AS_PF_FAULT) { fault_if_from_uspace(istate,"Page fault at %p",va); panic("%s: va=%p, rid=%d, iip=%p\n", __func__, va, rid, istate->cr_iip); t->a = true; itc_pte_copy(t); } } page_table_unlock(AS, true); } /** Data access bit fault handler. * * @param vector Interruption vector. * @param istate Structure with saved interruption state. */ void data_access_bit_fault(uint64_t vector, istate_t *istate) { region_register rr; rid_t rid; uintptr_t va; pte_t *t; va = istate->cr_ifa; /* faulting address */ rr.word = rr_read(VA2VRN(va)); rid = rr.map.rid; page_table_lock(AS, true); t = page_mapping_find(AS, va); ASSERT(t && t->p); if (t && t->p) { /* * Update the Accessed bit in page tables and reinsert * the mapping into DTC. */ t->a = true; dtc_pte_copy(t); } else { if (as_page_fault(va, PF_ACCESS_READ, istate) == AS_PF_FAULT) { fault_if_from_uspace(istate,"Page fault at %p",va); panic("%s: va=%p, rid=%d, iip=%p\n", __func__, va, rid, istate->cr_iip); t->a = true; itc_pte_copy(t); } } page_table_unlock(AS, true); } /** Page not present fault handler. * * @param vector Interruption vector. * @param istate Structure with saved interruption state. */ void page_not_present(uint64_t vector, istate_t *istate) { region_register rr; rid_t rid; uintptr_t va; pte_t *t; va = istate->cr_ifa; /* faulting address */ rr.word = rr_read(VA2VRN(va)); rid = rr.map.rid; page_table_lock(AS, true); t = page_mapping_find(AS, va); ASSERT(t); if (t->p) { /* * If the Present bit is set in page hash table, just copy it * and update ITC/DTC. */ if (t->x) itc_pte_copy(t); else dtc_pte_copy(t); page_table_unlock(AS, true); } else { page_table_unlock(AS, true); if (as_page_fault(va, PF_ACCESS_READ, istate) == AS_PF_FAULT) { fault_if_from_uspace(istate,"Page fault at %p",va); panic("%s: va=%p, rid=%d\n", __func__, va, rid); } } } /** @} */