|
@@ -30,8 +30,30 @@
|
|
|
#include "exec/ram_addr.h"
|
|
|
#include "tcg/tcg.h"
|
|
|
|
|
|
-//#define DEBUG_TLB
|
|
|
-//#define DEBUG_TLB_CHECK
|
|
|
+/* DEBUG defines, enable DEBUG_TLB_LOG to log to the CPU_LOG_MMU target */
|
|
|
+/* #define DEBUG_TLB */
|
|
|
+/* #define DEBUG_TLB_LOG */
|
|
|
+
|
|
|
+#ifdef DEBUG_TLB
|
|
|
+# define DEBUG_TLB_GATE 1
|
|
|
+# ifdef DEBUG_TLB_LOG
|
|
|
+# define DEBUG_TLB_LOG_GATE 1
|
|
|
+# else
|
|
|
+# define DEBUG_TLB_LOG_GATE 0
|
|
|
+# endif
|
|
|
+#else
|
|
|
+# define DEBUG_TLB_GATE 0
|
|
|
+# define DEBUG_TLB_LOG_GATE 0
|
|
|
+#endif
|
|
|
+
|
|
|
+#define tlb_debug(fmt, ...) do { \
|
|
|
+ if (DEBUG_TLB_LOG_GATE) { \
|
|
|
+ qemu_log_mask(CPU_LOG_MMU, "%s: " fmt, __func__, \
|
|
|
+ ## __VA_ARGS__); \
|
|
|
+ } else if (DEBUG_TLB_GATE) { \
|
|
|
+ fprintf(stderr, "%s: " fmt, __func__, ## __VA_ARGS__); \
|
|
|
+ } \
|
|
|
+} while (0)
|
|
|
|
|
|
/* statistics */
|
|
|
int tlb_flush_count;
|
|
@@ -52,9 +74,8 @@ void tlb_flush(CPUState *cpu, int flush_global)
|
|
|
{
|
|
|
CPUArchState *env = cpu->env_ptr;
|
|
|
|
|
|
-#if defined(DEBUG_TLB)
|
|
|
- printf("tlb_flush:\n");
|
|
|
-#endif
|
|
|
+ tlb_debug("(%d)\n", flush_global);
|
|
|
+
|
|
|
/* must reset current TB so that interrupts cannot modify the
|
|
|
links while we are modifying them */
|
|
|
cpu->current_tb = NULL;
|
|
@@ -73,9 +94,7 @@ static inline void v_tlb_flush_by_mmuidx(CPUState *cpu, va_list argp)
|
|
|
{
|
|
|
CPUArchState *env = cpu->env_ptr;
|
|
|
|
|
|
-#if defined(DEBUG_TLB)
|
|
|
- printf("tlb_flush_by_mmuidx:");
|
|
|
-#endif
|
|
|
+ tlb_debug("start\n");
|
|
|
/* must reset current TB so that interrupts cannot modify the
|
|
|
links while we are modifying them */
|
|
|
cpu->current_tb = NULL;
|
|
@@ -87,18 +106,12 @@ static inline void v_tlb_flush_by_mmuidx(CPUState *cpu, va_list argp)
|
|
|
break;
|
|
|
}
|
|
|
|
|
|
-#if defined(DEBUG_TLB)
|
|
|
- printf(" %d", mmu_idx);
|
|
|
-#endif
|
|
|
+ tlb_debug("%d\n", mmu_idx);
|
|
|
|
|
|
memset(env->tlb_table[mmu_idx], -1, sizeof(env->tlb_table[0]));
|
|
|
memset(env->tlb_v_table[mmu_idx], -1, sizeof(env->tlb_v_table[0]));
|
|
|
}
|
|
|
|
|
|
-#if defined(DEBUG_TLB)
|
|
|
- printf("\n");
|
|
|
-#endif
|
|
|
-
|
|
|
memset(cpu->tb_jmp_cache, 0, sizeof(cpu->tb_jmp_cache));
|
|
|
}
|
|
|
|
|
@@ -128,16 +141,14 @@ void tlb_flush_page(CPUState *cpu, target_ulong addr)
|
|
|
int i;
|
|
|
int mmu_idx;
|
|
|
|
|
|
-#if defined(DEBUG_TLB)
|
|
|
- printf("tlb_flush_page: " TARGET_FMT_lx "\n", addr);
|
|
|
-#endif
|
|
|
+ tlb_debug("page :" TARGET_FMT_lx "\n", addr);
|
|
|
+
|
|
|
/* Check if we need to flush due to large pages. */
|
|
|
if ((addr & env->tlb_flush_mask) == env->tlb_flush_addr) {
|
|
|
-#if defined(DEBUG_TLB)
|
|
|
- printf("tlb_flush_page: forced full flush ("
|
|
|
- TARGET_FMT_lx "/" TARGET_FMT_lx ")\n",
|
|
|
- env->tlb_flush_addr, env->tlb_flush_mask);
|
|
|
-#endif
|
|
|
+ tlb_debug("forcing full flush ("
|
|
|
+ TARGET_FMT_lx "/" TARGET_FMT_lx ")\n",
|
|
|
+ env->tlb_flush_addr, env->tlb_flush_mask);
|
|
|
+
|
|
|
tlb_flush(cpu, 1);
|
|
|
return;
|
|
|
}
|
|
@@ -170,16 +181,14 @@ void tlb_flush_page_by_mmuidx(CPUState *cpu, target_ulong addr, ...)
|
|
|
|
|
|
va_start(argp, addr);
|
|
|
|
|
|
-#if defined(DEBUG_TLB)
|
|
|
- printf("tlb_flush_page_by_mmu_idx: " TARGET_FMT_lx, addr);
|
|
|
-#endif
|
|
|
+ tlb_debug("addr "TARGET_FMT_lx"\n", addr);
|
|
|
+
|
|
|
/* Check if we need to flush due to large pages. */
|
|
|
if ((addr & env->tlb_flush_mask) == env->tlb_flush_addr) {
|
|
|
-#if defined(DEBUG_TLB)
|
|
|
- printf(" forced full flush ("
|
|
|
- TARGET_FMT_lx "/" TARGET_FMT_lx ")\n",
|
|
|
- env->tlb_flush_addr, env->tlb_flush_mask);
|
|
|
-#endif
|
|
|
+ tlb_debug("forced full flush ("
|
|
|
+ TARGET_FMT_lx "/" TARGET_FMT_lx ")\n",
|
|
|
+ env->tlb_flush_addr, env->tlb_flush_mask);
|
|
|
+
|
|
|
v_tlb_flush_by_mmuidx(cpu, argp);
|
|
|
va_end(argp);
|
|
|
return;
|
|
@@ -198,9 +207,7 @@ void tlb_flush_page_by_mmuidx(CPUState *cpu, target_ulong addr, ...)
|
|
|
break;
|
|
|
}
|
|
|
|
|
|
-#if defined(DEBUG_TLB)
|
|
|
- printf(" %d", mmu_idx);
|
|
|
-#endif
|
|
|
+ tlb_debug("idx %d\n", mmu_idx);
|
|
|
|
|
|
tlb_flush_entry(&env->tlb_table[mmu_idx][i], addr);
|
|
|
|
|
@@ -211,10 +218,6 @@ void tlb_flush_page_by_mmuidx(CPUState *cpu, target_ulong addr, ...)
|
|
|
}
|
|
|
va_end(argp);
|
|
|
|
|
|
-#if defined(DEBUG_TLB)
|
|
|
- printf("\n");
|
|
|
-#endif
|
|
|
-
|
|
|
tb_flush_jmp_cache(cpu, addr);
|
|
|
}
|
|
|
|
|
@@ -367,12 +370,9 @@ void tlb_set_page_with_attrs(CPUState *cpu, target_ulong vaddr,
|
|
|
section = address_space_translate_for_iotlb(cpu, asidx, paddr, &xlat, &sz);
|
|
|
assert(sz >= TARGET_PAGE_SIZE);
|
|
|
|
|
|
-#if defined(DEBUG_TLB)
|
|
|
- qemu_log_mask(CPU_LOG_MMU,
|
|
|
- "tlb_set_page: vaddr=" TARGET_FMT_lx " paddr=0x" TARGET_FMT_plx
|
|
|
- " prot=%x idx=%d\n",
|
|
|
- vaddr, paddr, prot, mmu_idx);
|
|
|
-#endif
|
|
|
+ tlb_debug("vaddr=" TARGET_FMT_lx " paddr=0x" TARGET_FMT_plx
|
|
|
+ " prot=%x idx=%d\n",
|
|
|
+ vaddr, paddr, prot, mmu_idx);
|
|
|
|
|
|
address = vaddr;
|
|
|
if (!memory_region_is_ram(section->mr) && !memory_region_is_romd(section->mr)) {
|