|
@@ -161,18 +161,18 @@ static void xen_ram_init(ram_addr_t ram_size)
|
|
|
ram_addr_t block_len;
|
|
|
|
|
|
block_len = ram_size;
|
|
|
- if (ram_size >= HVM_BELOW_4G_RAM_END) {
|
|
|
+ if (ram_size >= QEMU_BELOW_4G_RAM_END) {
|
|
|
/* Xen does not allocate the memory continuously, and keep a hole at
|
|
|
- * HVM_BELOW_4G_MMIO_START of HVM_BELOW_4G_MMIO_LENGTH
|
|
|
+ * QEMU_BELOW_4G_RAM_END of QEMU_BELOW_4G_MMIO_LENGTH
|
|
|
*/
|
|
|
- block_len += HVM_BELOW_4G_MMIO_LENGTH;
|
|
|
+ block_len += QEMU_BELOW_4G_MMIO_LENGTH;
|
|
|
}
|
|
|
memory_region_init_ram(&ram_memory, "xen.ram", block_len);
|
|
|
vmstate_register_ram_global(&ram_memory);
|
|
|
|
|
|
- if (ram_size >= HVM_BELOW_4G_RAM_END) {
|
|
|
- above_4g_mem_size = ram_size - HVM_BELOW_4G_RAM_END;
|
|
|
- below_4g_mem_size = HVM_BELOW_4G_RAM_END;
|
|
|
+ if (ram_size >= QEMU_BELOW_4G_RAM_END) {
|
|
|
+ above_4g_mem_size = ram_size - QEMU_BELOW_4G_RAM_END;
|
|
|
+ below_4g_mem_size = QEMU_BELOW_4G_RAM_END;
|
|
|
} else {
|
|
|
below_4g_mem_size = ram_size;
|
|
|
}
|
|
@@ -574,29 +574,6 @@ void qmp_xen_set_global_dirty_log(bool enable, Error **errp)
|
|
|
}
|
|
|
}
|
|
|
|
|
|
-/* VCPU Operations, MMIO, IO ring ... */
|
|
|
-
|
|
|
-static void xen_reset_vcpu(void *opaque)
|
|
|
-{
|
|
|
- CPUState *cpu = opaque;
|
|
|
-
|
|
|
- cpu->halted = 1;
|
|
|
-}
|
|
|
-
|
|
|
-void xen_vcpu_init(void)
|
|
|
-{
|
|
|
- if (first_cpu != NULL) {
|
|
|
- CPUState *cpu = ENV_GET_CPU(first_cpu);
|
|
|
-
|
|
|
- qemu_register_reset(xen_reset_vcpu, cpu);
|
|
|
- xen_reset_vcpu(cpu);
|
|
|
- }
|
|
|
- /* if rtc_clock is left to default (host_clock), disable it */
|
|
|
- if (rtc_clock == host_clock) {
|
|
|
- qemu_clock_enable(rtc_clock, false);
|
|
|
- }
|
|
|
-}
|
|
|
-
|
|
|
/* get the ioreq packets from share mem */
|
|
|
static ioreq_t *cpu_get_ioreq_from_shared_memory(XenIOState *state, int vcpu)
|
|
|
{
|