2
0

memory.rst 16 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372
  1. ==============
  2. The memory API
  3. ==============
  4. The memory API models the memory and I/O buses and controllers of a QEMU
  5. machine. It attempts to allow modelling of:
  6. - ordinary RAM
  7. - memory-mapped I/O (MMIO)
  8. - memory controllers that can dynamically reroute physical memory regions
  9. to different destinations
  10. The memory model provides support for
  11. - tracking RAM changes by the guest
  12. - setting up coalesced memory for kvm
  13. - setting up ioeventfd regions for kvm
  14. Memory is modelled as an acyclic graph of MemoryRegion objects. Sinks
  15. (leaves) are RAM and MMIO regions, while other nodes represent
  16. buses, memory controllers, and memory regions that have been rerouted.
  17. In addition to MemoryRegion objects, the memory API provides AddressSpace
  18. objects for every root and possibly for intermediate MemoryRegions too.
  19. These represent memory as seen from the CPU or a device's viewpoint.
  20. Types of regions
  21. ----------------
  22. There are multiple types of memory regions (all represented by a single C type
  23. MemoryRegion):
  24. - RAM: a RAM region is simply a range of host memory that can be made available
  25. to the guest.
  26. You typically initialize these with memory_region_init_ram(). Some special
  27. purposes require the variants memory_region_init_resizeable_ram(),
  28. memory_region_init_ram_from_file(), or memory_region_init_ram_ptr().
  29. - MMIO: a range of guest memory that is implemented by host callbacks;
  30. each read or write causes a callback to be called on the host.
  31. You initialize these with memory_region_init_io(), passing it a
  32. MemoryRegionOps structure describing the callbacks.
  33. - ROM: a ROM memory region works like RAM for reads (directly accessing
  34. a region of host memory), and forbids writes. You initialize these with
  35. memory_region_init_rom().
  36. - ROM device: a ROM device memory region works like RAM for reads
  37. (directly accessing a region of host memory), but like MMIO for
  38. writes (invoking a callback). You initialize these with
  39. memory_region_init_rom_device().
  40. - IOMMU region: an IOMMU region translates addresses of accesses made to it
  41. and forwards them to some other target memory region. As the name suggests,
  42. these are only needed for modelling an IOMMU, not for simple devices.
  43. You initialize these with memory_region_init_iommu().
  44. - container: a container simply includes other memory regions, each at
  45. a different offset. Containers are useful for grouping several regions
  46. into one unit. For example, a PCI BAR may be composed of a RAM region
  47. and an MMIO region.
  48. A container's subregions are usually non-overlapping. In some cases it is
  49. useful to have overlapping regions; for example a memory controller that
  50. can overlay a subregion of RAM with MMIO or ROM, or a PCI controller
  51. that does not prevent card from claiming overlapping BARs.
  52. You initialize a pure container with memory_region_init().
  53. - alias: a subsection of another region. Aliases allow a region to be
  54. split apart into discontiguous regions. Examples of uses are memory
  55. banks used when the guest address space is smaller than the amount
  56. of RAM addressed, or a memory controller that splits main memory to
  57. expose a "PCI hole". You can also create aliases to avoid trying to
  58. add the original region to multiple parents via
  59. `memory_region_add_subregion`.
  60. Aliases may point to any type of region, including other aliases,
  61. but an alias may not point back to itself, directly or indirectly.
  62. You initialize these with memory_region_init_alias().
  63. - reservation region: a reservation region is primarily for debugging.
  64. It claims I/O space that is not supposed to be handled by QEMU itself.
  65. The typical use is to track parts of the address space which will be
  66. handled by the host kernel when KVM is enabled. You initialize these
  67. by passing a NULL callback parameter to memory_region_init_io().
  68. It is valid to add subregions to a region which is not a pure container
  69. (that is, to an MMIO, RAM or ROM region). This means that the region
  70. will act like a container, except that any addresses within the container's
  71. region which are not claimed by any subregion are handled by the
  72. container itself (ie by its MMIO callbacks or RAM backing). However
  73. it is generally possible to achieve the same effect with a pure container
  74. one of whose subregions is a low priority "background" region covering
  75. the whole address range; this is often clearer and is preferred.
  76. Subregions cannot be added to an alias region.
  77. Migration
  78. ---------
  79. Where the memory region is backed by host memory (RAM, ROM and
  80. ROM device memory region types), this host memory needs to be
  81. copied to the destination on migration. These APIs which allocate
  82. the host memory for you will also register the memory so it is
  83. migrated:
  84. - memory_region_init_ram()
  85. - memory_region_init_rom()
  86. - memory_region_init_rom_device()
  87. For most devices and boards this is the correct thing. If you
  88. have a special case where you need to manage the migration of
  89. the backing memory yourself, you can call the functions:
  90. - memory_region_init_ram_nomigrate()
  91. - memory_region_init_rom_nomigrate()
  92. - memory_region_init_rom_device_nomigrate()
  93. which only initialize the MemoryRegion and leave handling
  94. migration to the caller.
  95. The functions:
  96. - memory_region_init_resizeable_ram()
  97. - memory_region_init_ram_from_file()
  98. - memory_region_init_ram_from_fd()
  99. - memory_region_init_ram_ptr()
  100. - memory_region_init_ram_device_ptr()
  101. are for special cases only, and so they do not automatically
  102. register the backing memory for migration; the caller must
  103. manage migration if necessary.
  104. Region names
  105. ------------
  106. Regions are assigned names by the constructor. For most regions these are
  107. only used for debugging purposes, but RAM regions also use the name to identify
  108. live migration sections. This means that RAM region names need to have ABI
  109. stability.
  110. Region lifecycle
  111. ----------------
  112. A region is created by one of the memory_region_init*() functions and
  113. attached to an object, which acts as its owner or parent. QEMU ensures
  114. that the owner object remains alive as long as the region is visible to
  115. the guest, or as long as the region is in use by a virtual CPU or another
  116. device. For example, the owner object will not die between an
  117. address_space_map operation and the corresponding address_space_unmap.
  118. After creation, a region can be added to an address space or a
  119. container with memory_region_add_subregion(), and removed using
  120. memory_region_del_subregion().
  121. Various region attributes (read-only, dirty logging, coalesced mmio,
  122. ioeventfd) can be changed during the region lifecycle. They take effect
  123. as soon as the region is made visible. This can be immediately, later,
  124. or never.
  125. Destruction of a memory region happens automatically when the owner
  126. object dies.
  127. If however the memory region is part of a dynamically allocated data
  128. structure, you should call object_unparent() to destroy the memory region
  129. before the data structure is freed. For an example see VFIOMSIXInfo
  130. and VFIOQuirk in hw/vfio/pci.c.
  131. You must not destroy a memory region as long as it may be in use by a
  132. device or CPU. In order to do this, as a general rule do not create or
  133. destroy memory regions dynamically during a device's lifetime, and only
  134. call object_unparent() in the memory region owner's instance_finalize
  135. callback. The dynamically allocated data structure that contains the
  136. memory region then should obviously be freed in the instance_finalize
  137. callback as well.
  138. If you break this rule, the following situation can happen:
  139. - the memory region's owner had a reference taken via memory_region_ref
  140. (for example by address_space_map)
  141. - the region is unparented, and has no owner anymore
  142. - when address_space_unmap is called, the reference to the memory region's
  143. owner is leaked.
  144. There is an exception to the above rule: it is okay to call
  145. object_unparent at any time for an alias or a container region. It is
  146. therefore also okay to create or destroy alias and container regions
  147. dynamically during a device's lifetime.
  148. This exceptional usage is valid because aliases and containers only help
  149. QEMU building the guest's memory map; they are never accessed directly.
  150. memory_region_ref and memory_region_unref are never called on aliases
  151. or containers, and the above situation then cannot happen. Exploiting
  152. this exception is rarely necessary, and therefore it is discouraged,
  153. but nevertheless it is used in a few places.
  154. For regions that "have no owner" (NULL is passed at creation time), the
  155. machine object is actually used as the owner. Since instance_finalize is
  156. never called for the machine object, you must never call object_unparent
  157. on regions that have no owner, unless they are aliases or containers.
  158. Overlapping regions and priority
  159. --------------------------------
  160. Usually, regions may not overlap each other; a memory address decodes into
  161. exactly one target. In some cases it is useful to allow regions to overlap,
  162. and sometimes to control which of an overlapping regions is visible to the
  163. guest. This is done with memory_region_add_subregion_overlap(), which
  164. allows the region to overlap any other region in the same container, and
  165. specifies a priority that allows the core to decide which of two regions at
  166. the same address are visible (highest wins).
  167. Priority values are signed, and the default value is zero. This means that
  168. you can use memory_region_add_subregion_overlap() both to specify a region
  169. that must sit 'above' any others (with a positive priority) and also a
  170. background region that sits 'below' others (with a negative priority).
  171. If the higher priority region in an overlap is a container or alias, then
  172. the lower priority region will appear in any "holes" that the higher priority
  173. region has left by not mapping subregions to that area of its address range.
  174. (This applies recursively -- if the subregions are themselves containers or
  175. aliases that leave holes then the lower priority region will appear in these
  176. holes too.)
  177. For example, suppose we have a container A of size 0x8000 with two subregions
  178. B and C. B is a container mapped at 0x2000, size 0x4000, priority 2; C is
  179. an MMIO region mapped at 0x0, size 0x6000, priority 1. B currently has two
  180. of its own subregions: D of size 0x1000 at offset 0 and E of size 0x1000 at
  181. offset 0x2000. As a diagram::
  182. 0 1000 2000 3000 4000 5000 6000 7000 8000
  183. |------|------|------|------|------|------|------|------|
  184. A: [ ]
  185. C: [CCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCCC]
  186. B: [ ]
  187. D: [DDDDD]
  188. E: [EEEEE]
  189. The regions that will be seen within this address range then are::
  190. [CCCCCCCCCCCC][DDDDD][CCCCC][EEEEE][CCCCC]
  191. Since B has higher priority than C, its subregions appear in the flat map
  192. even where they overlap with C. In ranges where B has not mapped anything
  193. C's region appears.
  194. If B had provided its own MMIO operations (ie it was not a pure container)
  195. then these would be used for any addresses in its range not handled by
  196. D or E, and the result would be::
  197. [CCCCCCCCCCCC][DDDDD][BBBBB][EEEEE][BBBBB]
  198. Priority values are local to a container, because the priorities of two
  199. regions are only compared when they are both children of the same container.
  200. This means that the device in charge of the container (typically modelling
  201. a bus or a memory controller) can use them to manage the interaction of
  202. its child regions without any side effects on other parts of the system.
  203. In the example above, the priorities of D and E are unimportant because
  204. they do not overlap each other. It is the relative priority of B and C
  205. that causes D and E to appear on top of C: D and E's priorities are never
  206. compared against the priority of C.
  207. Visibility
  208. ----------
  209. The memory core uses the following rules to select a memory region when the
  210. guest accesses an address:
  211. - all direct subregions of the root region are matched against the address, in
  212. descending priority order
  213. - if the address lies outside the region offset/size, the subregion is
  214. discarded
  215. - if the subregion is a leaf (RAM or MMIO), the search terminates, returning
  216. this leaf region
  217. - if the subregion is a container, the same algorithm is used within the
  218. subregion (after the address is adjusted by the subregion offset)
  219. - if the subregion is an alias, the search is continued at the alias target
  220. (after the address is adjusted by the subregion offset and alias offset)
  221. - if a recursive search within a container or alias subregion does not
  222. find a match (because of a "hole" in the container's coverage of its
  223. address range), then if this is a container with its own MMIO or RAM
  224. backing the search terminates, returning the container itself. Otherwise
  225. we continue with the next subregion in priority order
  226. - if none of the subregions match the address then the search terminates
  227. with no match found
  228. Example memory map
  229. ------------------
  230. ::
  231. system_memory: container@0-2^48-1
  232. |
  233. +---- lomem: alias@0-0xdfffffff ---> #ram (0-0xdfffffff)
  234. |
  235. +---- himem: alias@0x100000000-0x11fffffff ---> #ram (0xe0000000-0xffffffff)
  236. |
  237. +---- vga-window: alias@0xa0000-0xbffff ---> #pci (0xa0000-0xbffff)
  238. | (prio 1)
  239. |
  240. +---- pci-hole: alias@0xe0000000-0xffffffff ---> #pci (0xe0000000-0xffffffff)
  241. pci (0-2^32-1)
  242. |
  243. +--- vga-area: container@0xa0000-0xbffff
  244. | |
  245. | +--- alias@0x00000-0x7fff ---> #vram (0x010000-0x017fff)
  246. | |
  247. | +--- alias@0x08000-0xffff ---> #vram (0x020000-0x027fff)
  248. |
  249. +---- vram: ram@0xe1000000-0xe1ffffff
  250. |
  251. +---- vga-mmio: mmio@0xe2000000-0xe200ffff
  252. ram: ram@0x00000000-0xffffffff
  253. This is a (simplified) PC memory map. The 4GB RAM block is mapped into the
  254. system address space via two aliases: "lomem" is a 1:1 mapping of the first
  255. 3.5GB; "himem" maps the last 0.5GB at address 4GB. This leaves 0.5GB for the
  256. so-called PCI hole, that allows a 32-bit PCI bus to exist in a system with
  257. 4GB of memory.
  258. The memory controller diverts addresses in the range 640K-768K to the PCI
  259. address space. This is modelled using the "vga-window" alias, mapped at a
  260. higher priority so it obscures the RAM at the same addresses. The vga window
  261. can be removed by programming the memory controller; this is modelled by
  262. removing the alias and exposing the RAM underneath.
  263. The pci address space is not a direct child of the system address space, since
  264. we only want parts of it to be visible (we accomplish this using aliases).
  265. It has two subregions: vga-area models the legacy vga window and is occupied
  266. by two 32K memory banks pointing at two sections of the framebuffer.
  267. In addition the vram is mapped as a BAR at address e1000000, and an additional
  268. BAR containing MMIO registers is mapped after it.
  269. Note that if the guest maps a BAR outside the PCI hole, it would not be
  270. visible as the pci-hole alias clips it to a 0.5GB range.
  271. MMIO Operations
  272. ---------------
  273. MMIO regions are provided with ->read() and ->write() callbacks,
  274. which are sufficient for most devices. Some devices change behaviour
  275. based on the attributes used for the memory transaction, or need
  276. to be able to respond that the access should provoke a bus error
  277. rather than completing successfully; those devices can use the
  278. ->read_with_attrs() and ->write_with_attrs() callbacks instead.
  279. In addition various constraints can be supplied to control how these
  280. callbacks are called:
  281. - .valid.min_access_size, .valid.max_access_size define the access sizes
  282. (in bytes) which the device accepts; accesses outside this range will
  283. have device and bus specific behaviour (ignored, or machine check)
  284. - .valid.unaligned specifies that the *device being modelled* supports
  285. unaligned accesses; if false, unaligned accesses will invoke the
  286. appropriate bus or CPU specific behaviour.
  287. - .impl.min_access_size, .impl.max_access_size define the access sizes
  288. (in bytes) supported by the *implementation*; other access sizes will be
  289. emulated using the ones available. For example a 4-byte write will be
  290. emulated using four 1-byte writes, if .impl.max_access_size = 1.
  291. - .impl.unaligned specifies that the *implementation* supports unaligned
  292. accesses; if false, unaligned accesses will be emulated by two aligned
  293. accesses.
  294. API Reference
  295. -------------
  296. .. kernel-doc:: include/exec/memory.h