diff --git a/sys/alpha/alpha/pmap.c b/sys/alpha/alpha/pmap.c index 21dfef1c14e6..5c8156d14c8f 100644 --- a/sys/alpha/alpha/pmap.c +++ b/sys/alpha/alpha/pmap.c @@ -197,21 +197,6 @@ #define PMAP_INLINE #endif -#if 0 - -static void -pmap_break(void) -{ -} - -/* #define PMAP_DEBUG_VA(va) if ((va) == 0x120058000) pmap_break(); else */ - -#endif - -#ifndef PMAP_DEBUG_VA -#define PMAP_DEBUG_VA(va) do {} while(0) -#endif - /* * Some macros for manipulating virtual addresses */ @@ -824,7 +809,6 @@ pmap_qenter(vm_offset_t va, vm_page_t *m, int count) pt_entry_t opte; pte = vtopte(tva); opte = *pte; - PMAP_DEBUG_VA(va); *pte = npte; if (opte) pmap_invalidate_page(kernel_pmap, tva); @@ -845,7 +829,6 @@ pmap_qremove(va, count) for (i = 0; i < count; i++) { pte = vtopte(va); - PMAP_DEBUG_VA(va); *pte = 0; pmap_invalidate_page(kernel_pmap, va); va += PAGE_SIZE; @@ -866,7 +849,6 @@ pmap_kenter(vm_offset_t va, vm_offset_t pa) npte = pmap_phys_to_pte(pa) | PG_ASM | PG_KRE | PG_KWE | PG_V; pte = vtopte(va); opte = *pte; - PMAP_DEBUG_VA(va); *pte = npte; if (opte) pmap_invalidate_page(kernel_pmap, va); @@ -881,7 +863,6 @@ pmap_kremove(vm_offset_t va) register pt_entry_t *pte; pte = vtopte(va); - PMAP_DEBUG_VA(va); *pte = 0; pmap_invalidate_page(kernel_pmap, va); } @@ -1942,7 +1923,6 @@ pmap_remove_pte(pmap_t pmap, pt_entry_t* ptq, vm_offset_t va) vm_page_t m; oldpte = *ptq; - PMAP_DEBUG_VA(va); *ptq = 0; if (oldpte & PG_W) pmap->pm_stats.wired_count -= 1; @@ -2075,7 +2055,6 @@ pmap_remove_all(vm_page_t m) tpte = *pte; - PMAP_DEBUG_VA(pv->pv_va); *pte = 0; if (tpte & PG_W) pv->pv_pmap->pm_stats.wired_count--; @@ -2328,7 +2307,6 @@ pmap_enter(pmap_t pmap, vm_offset_t va, vm_page_t m, vm_prot_t prot, * to update the pte. */ if (origpte != newpte) { - PMAP_DEBUG_VA(va); *pte = newpte; if (origpte) pmap_invalidate_page(pmap, va); @@ -2417,7 +2395,6 @@ pmap_enter_quick(pmap_t pmap, vm_offset_t va, vm_page_t m, vm_page_t mpte) * raise IPL while manipulating pv_table since pmap_enter can be * called at interrupt time. */ - PMAP_DEBUG_VA(va); pmap_insert_entry(pmap, va, mpte, m); /* @@ -2847,7 +2824,6 @@ pmap_remove_pages(pmap, sva, eva) npv = TAILQ_NEXT(pv, pv_plist); continue; } - PMAP_DEBUG_VA(pv->pv_va); *pte = 0; m = PHYS_TO_VM_PAGE(pmap_pte_pa(&tpte)); diff --git a/sys/ia64/ia64/pmap.c b/sys/ia64/ia64/pmap.c index dd78edcfba28..41b5e15f20ea 100644 --- a/sys/ia64/ia64/pmap.c +++ b/sys/ia64/ia64/pmap.c @@ -145,21 +145,6 @@ MALLOC_DEFINE(M_PMAP, "PMAP", "PMAP Structures"); #define PMAP_INLINE #endif -#if 0 - -static void -pmap_break(void) -{ -} - -/* #define PMAP_DEBUG_VA(va) if ((va) == 0x120058000) pmap_break(); else */ - -#endif - -#ifndef PMAP_DEBUG_VA -#define PMAP_DEBUG_VA(va) do {} while(0) -#endif - /* * Get PDEs and PTEs for user/kernel address space */ @@ -1376,7 +1361,6 @@ pmap_qenter(vm_offset_t va, vm_page_t *m, int count) for (i = 0; i < count; i++) { vm_offset_t tva = va + i * PAGE_SIZE; int wasvalid; - PMAP_DEBUG_VA(tva); pte = pmap_find_kpte(tva); wasvalid = pte->pte_p; pmap_set_pte(pte, tva, VM_PAGE_TO_PHYS(m[i]), @@ -1397,7 +1381,6 @@ pmap_qremove(vm_offset_t va, int count) struct ia64_lpte *pte; for (i = 0; i < count; i++) { - PMAP_DEBUG_VA(va); pte = pmap_find_kpte(va); pmap_clear_pte(pte, va); va += PAGE_SIZE; @@ -1781,10 +1764,8 @@ pmap_enter(pmap_t pmap, vm_offset_t va, vm_page_t m, vm_prot_t prot, * if the mapping or permission bits are different, we need * to invalidate the page. */ - if (!pmap_equal_pte(&origpte, pte)) { - PMAP_DEBUG_VA(va); + if (!pmap_equal_pte(&origpte, pte)) pmap_invalidate_page(pmap, va); - } pmap_install(oldpmap); } @@ -1814,8 +1795,6 @@ pmap_enter_quick(pmap_t pmap, vm_offset_t va, vm_page_t m) if (pte->pte_p) return; - PMAP_DEBUG_VA(va); - /* * Enter on the PV list since its part of our managed memory. */