helpers.c 20 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730
  1. /*
  2. * low level and IOMMU backend agnostic helpers used by VFIO devices,
  3. * related to regions, interrupts, capabilities
  4. *
  5. * Copyright Red Hat, Inc. 2012
  6. *
  7. * Authors:
  8. * Alex Williamson <alex.williamson@redhat.com>
  9. *
  10. * This work is licensed under the terms of the GNU GPL, version 2. See
  11. * the COPYING file in the top-level directory.
  12. *
  13. * Based on qemu-kvm device-assignment:
  14. * Adapted for KVM by Qumranet.
  15. * Copyright (c) 2007, Neocleus, Alex Novik (alex@neocleus.com)
  16. * Copyright (c) 2007, Neocleus, Guy Zana (guy@neocleus.com)
  17. * Copyright (C) 2008, Qumranet, Amit Shah (amit.shah@qumranet.com)
  18. * Copyright (C) 2008, Red Hat, Amit Shah (amit.shah@redhat.com)
  19. * Copyright (C) 2008, IBM, Muli Ben-Yehuda (muli@il.ibm.com)
  20. */
  21. #include "qemu/osdep.h"
  22. #include <sys/ioctl.h>
  23. #include "hw/vfio/vfio-common.h"
  24. #include "hw/hw.h"
  25. #include "trace.h"
  26. #include "qapi/error.h"
  27. #include "qemu/error-report.h"
  28. #include "qemu/units.h"
  29. #include "monitor/monitor.h"
  30. /*
  31. * Common VFIO interrupt disable
  32. */
  33. void vfio_disable_irqindex(VFIODevice *vbasedev, int index)
  34. {
  35. struct vfio_irq_set irq_set = {
  36. .argsz = sizeof(irq_set),
  37. .flags = VFIO_IRQ_SET_DATA_NONE | VFIO_IRQ_SET_ACTION_TRIGGER,
  38. .index = index,
  39. .start = 0,
  40. .count = 0,
  41. };
  42. ioctl(vbasedev->fd, VFIO_DEVICE_SET_IRQS, &irq_set);
  43. }
  44. void vfio_unmask_single_irqindex(VFIODevice *vbasedev, int index)
  45. {
  46. struct vfio_irq_set irq_set = {
  47. .argsz = sizeof(irq_set),
  48. .flags = VFIO_IRQ_SET_DATA_NONE | VFIO_IRQ_SET_ACTION_UNMASK,
  49. .index = index,
  50. .start = 0,
  51. .count = 1,
  52. };
  53. ioctl(vbasedev->fd, VFIO_DEVICE_SET_IRQS, &irq_set);
  54. }
  55. void vfio_mask_single_irqindex(VFIODevice *vbasedev, int index)
  56. {
  57. struct vfio_irq_set irq_set = {
  58. .argsz = sizeof(irq_set),
  59. .flags = VFIO_IRQ_SET_DATA_NONE | VFIO_IRQ_SET_ACTION_MASK,
  60. .index = index,
  61. .start = 0,
  62. .count = 1,
  63. };
  64. ioctl(vbasedev->fd, VFIO_DEVICE_SET_IRQS, &irq_set);
  65. }
  66. static inline const char *action_to_str(int action)
  67. {
  68. switch (action) {
  69. case VFIO_IRQ_SET_ACTION_MASK:
  70. return "MASK";
  71. case VFIO_IRQ_SET_ACTION_UNMASK:
  72. return "UNMASK";
  73. case VFIO_IRQ_SET_ACTION_TRIGGER:
  74. return "TRIGGER";
  75. default:
  76. return "UNKNOWN ACTION";
  77. }
  78. }
  79. static const char *index_to_str(VFIODevice *vbasedev, int index)
  80. {
  81. if (vbasedev->type != VFIO_DEVICE_TYPE_PCI) {
  82. return NULL;
  83. }
  84. switch (index) {
  85. case VFIO_PCI_INTX_IRQ_INDEX:
  86. return "INTX";
  87. case VFIO_PCI_MSI_IRQ_INDEX:
  88. return "MSI";
  89. case VFIO_PCI_MSIX_IRQ_INDEX:
  90. return "MSIX";
  91. case VFIO_PCI_ERR_IRQ_INDEX:
  92. return "ERR";
  93. case VFIO_PCI_REQ_IRQ_INDEX:
  94. return "REQ";
  95. default:
  96. return NULL;
  97. }
  98. }
  99. bool vfio_set_irq_signaling(VFIODevice *vbasedev, int index, int subindex,
  100. int action, int fd, Error **errp)
  101. {
  102. ERRP_GUARD();
  103. g_autofree struct vfio_irq_set *irq_set = NULL;
  104. int argsz;
  105. const char *name;
  106. int32_t *pfd;
  107. argsz = sizeof(*irq_set) + sizeof(*pfd);
  108. irq_set = g_malloc0(argsz);
  109. irq_set->argsz = argsz;
  110. irq_set->flags = VFIO_IRQ_SET_DATA_EVENTFD | action;
  111. irq_set->index = index;
  112. irq_set->start = subindex;
  113. irq_set->count = 1;
  114. pfd = (int32_t *)&irq_set->data;
  115. *pfd = fd;
  116. if (!ioctl(vbasedev->fd, VFIO_DEVICE_SET_IRQS, irq_set)) {
  117. return true;
  118. }
  119. error_setg_errno(errp, errno, "VFIO_DEVICE_SET_IRQS failure");
  120. name = index_to_str(vbasedev, index);
  121. if (name) {
  122. error_prepend(errp, "%s-%d: ", name, subindex);
  123. } else {
  124. error_prepend(errp, "index %d-%d: ", index, subindex);
  125. }
  126. error_prepend(errp,
  127. "Failed to %s %s eventfd signaling for interrupt ",
  128. fd < 0 ? "tear down" : "set up", action_to_str(action));
  129. return false;
  130. }
  131. /*
  132. * IO Port/MMIO - Beware of the endians, VFIO is always little endian
  133. */
  134. void vfio_region_write(void *opaque, hwaddr addr,
  135. uint64_t data, unsigned size)
  136. {
  137. VFIORegion *region = opaque;
  138. VFIODevice *vbasedev = region->vbasedev;
  139. union {
  140. uint8_t byte;
  141. uint16_t word;
  142. uint32_t dword;
  143. uint64_t qword;
  144. } buf;
  145. switch (size) {
  146. case 1:
  147. buf.byte = data;
  148. break;
  149. case 2:
  150. buf.word = cpu_to_le16(data);
  151. break;
  152. case 4:
  153. buf.dword = cpu_to_le32(data);
  154. break;
  155. case 8:
  156. buf.qword = cpu_to_le64(data);
  157. break;
  158. default:
  159. hw_error("vfio: unsupported write size, %u bytes", size);
  160. break;
  161. }
  162. if (pwrite(vbasedev->fd, &buf, size, region->fd_offset + addr) != size) {
  163. error_report("%s(%s:region%d+0x%"HWADDR_PRIx", 0x%"PRIx64
  164. ",%d) failed: %m",
  165. __func__, vbasedev->name, region->nr,
  166. addr, data, size);
  167. }
  168. trace_vfio_region_write(vbasedev->name, region->nr, addr, data, size);
  169. /*
  170. * A read or write to a BAR always signals an INTx EOI. This will
  171. * do nothing if not pending (including not in INTx mode). We assume
  172. * that a BAR access is in response to an interrupt and that BAR
  173. * accesses will service the interrupt. Unfortunately, we don't know
  174. * which access will service the interrupt, so we're potentially
  175. * getting quite a few host interrupts per guest interrupt.
  176. */
  177. vbasedev->ops->vfio_eoi(vbasedev);
  178. }
  179. uint64_t vfio_region_read(void *opaque,
  180. hwaddr addr, unsigned size)
  181. {
  182. VFIORegion *region = opaque;
  183. VFIODevice *vbasedev = region->vbasedev;
  184. union {
  185. uint8_t byte;
  186. uint16_t word;
  187. uint32_t dword;
  188. uint64_t qword;
  189. } buf;
  190. uint64_t data = 0;
  191. if (pread(vbasedev->fd, &buf, size, region->fd_offset + addr) != size) {
  192. error_report("%s(%s:region%d+0x%"HWADDR_PRIx", %d) failed: %m",
  193. __func__, vbasedev->name, region->nr,
  194. addr, size);
  195. return (uint64_t)-1;
  196. }
  197. switch (size) {
  198. case 1:
  199. data = buf.byte;
  200. break;
  201. case 2:
  202. data = le16_to_cpu(buf.word);
  203. break;
  204. case 4:
  205. data = le32_to_cpu(buf.dword);
  206. break;
  207. case 8:
  208. data = le64_to_cpu(buf.qword);
  209. break;
  210. default:
  211. hw_error("vfio: unsupported read size, %u bytes", size);
  212. break;
  213. }
  214. trace_vfio_region_read(vbasedev->name, region->nr, addr, size, data);
  215. /* Same as write above */
  216. vbasedev->ops->vfio_eoi(vbasedev);
  217. return data;
  218. }
  219. const MemoryRegionOps vfio_region_ops = {
  220. .read = vfio_region_read,
  221. .write = vfio_region_write,
  222. .endianness = DEVICE_LITTLE_ENDIAN,
  223. .valid = {
  224. .min_access_size = 1,
  225. .max_access_size = 8,
  226. },
  227. .impl = {
  228. .min_access_size = 1,
  229. .max_access_size = 8,
  230. },
  231. };
  232. int vfio_bitmap_alloc(VFIOBitmap *vbmap, hwaddr size)
  233. {
  234. vbmap->pages = REAL_HOST_PAGE_ALIGN(size) / qemu_real_host_page_size();
  235. vbmap->size = ROUND_UP(vbmap->pages, sizeof(__u64) * BITS_PER_BYTE) /
  236. BITS_PER_BYTE;
  237. vbmap->bitmap = g_try_malloc0(vbmap->size);
  238. if (!vbmap->bitmap) {
  239. return -ENOMEM;
  240. }
  241. return 0;
  242. }
  243. struct vfio_info_cap_header *
  244. vfio_get_cap(void *ptr, uint32_t cap_offset, uint16_t id)
  245. {
  246. struct vfio_info_cap_header *hdr;
  247. for (hdr = ptr + cap_offset; hdr != ptr; hdr = ptr + hdr->next) {
  248. if (hdr->id == id) {
  249. return hdr;
  250. }
  251. }
  252. return NULL;
  253. }
  254. struct vfio_info_cap_header *
  255. vfio_get_region_info_cap(struct vfio_region_info *info, uint16_t id)
  256. {
  257. if (!(info->flags & VFIO_REGION_INFO_FLAG_CAPS)) {
  258. return NULL;
  259. }
  260. return vfio_get_cap((void *)info, info->cap_offset, id);
  261. }
  262. struct vfio_info_cap_header *
  263. vfio_get_device_info_cap(struct vfio_device_info *info, uint16_t id)
  264. {
  265. if (!(info->flags & VFIO_DEVICE_FLAGS_CAPS)) {
  266. return NULL;
  267. }
  268. return vfio_get_cap((void *)info, info->cap_offset, id);
  269. }
  270. static int vfio_setup_region_sparse_mmaps(VFIORegion *region,
  271. struct vfio_region_info *info)
  272. {
  273. struct vfio_info_cap_header *hdr;
  274. struct vfio_region_info_cap_sparse_mmap *sparse;
  275. int i, j;
  276. hdr = vfio_get_region_info_cap(info, VFIO_REGION_INFO_CAP_SPARSE_MMAP);
  277. if (!hdr) {
  278. return -ENODEV;
  279. }
  280. sparse = container_of(hdr, struct vfio_region_info_cap_sparse_mmap, header);
  281. trace_vfio_region_sparse_mmap_header(region->vbasedev->name,
  282. region->nr, sparse->nr_areas);
  283. region->mmaps = g_new0(VFIOMmap, sparse->nr_areas);
  284. for (i = 0, j = 0; i < sparse->nr_areas; i++) {
  285. if (sparse->areas[i].size) {
  286. trace_vfio_region_sparse_mmap_entry(i, sparse->areas[i].offset,
  287. sparse->areas[i].offset +
  288. sparse->areas[i].size - 1);
  289. region->mmaps[j].offset = sparse->areas[i].offset;
  290. region->mmaps[j].size = sparse->areas[i].size;
  291. j++;
  292. }
  293. }
  294. region->nr_mmaps = j;
  295. region->mmaps = g_realloc(region->mmaps, j * sizeof(VFIOMmap));
  296. return 0;
  297. }
  298. int vfio_region_setup(Object *obj, VFIODevice *vbasedev, VFIORegion *region,
  299. int index, const char *name)
  300. {
  301. g_autofree struct vfio_region_info *info = NULL;
  302. int ret;
  303. ret = vfio_get_region_info(vbasedev, index, &info);
  304. if (ret) {
  305. return ret;
  306. }
  307. region->vbasedev = vbasedev;
  308. region->flags = info->flags;
  309. region->size = info->size;
  310. region->fd_offset = info->offset;
  311. region->nr = index;
  312. if (region->size) {
  313. region->mem = g_new0(MemoryRegion, 1);
  314. memory_region_init_io(region->mem, obj, &vfio_region_ops,
  315. region, name, region->size);
  316. if (!vbasedev->no_mmap &&
  317. region->flags & VFIO_REGION_INFO_FLAG_MMAP) {
  318. ret = vfio_setup_region_sparse_mmaps(region, info);
  319. if (ret) {
  320. region->nr_mmaps = 1;
  321. region->mmaps = g_new0(VFIOMmap, region->nr_mmaps);
  322. region->mmaps[0].offset = 0;
  323. region->mmaps[0].size = region->size;
  324. }
  325. }
  326. }
  327. trace_vfio_region_setup(vbasedev->name, index, name,
  328. region->flags, region->fd_offset, region->size);
  329. return 0;
  330. }
  331. static void vfio_subregion_unmap(VFIORegion *region, int index)
  332. {
  333. trace_vfio_region_unmap(memory_region_name(&region->mmaps[index].mem),
  334. region->mmaps[index].offset,
  335. region->mmaps[index].offset +
  336. region->mmaps[index].size - 1);
  337. memory_region_del_subregion(region->mem, &region->mmaps[index].mem);
  338. munmap(region->mmaps[index].mmap, region->mmaps[index].size);
  339. object_unparent(OBJECT(&region->mmaps[index].mem));
  340. region->mmaps[index].mmap = NULL;
  341. }
  342. int vfio_region_mmap(VFIORegion *region)
  343. {
  344. int i, ret, prot = 0;
  345. char *name;
  346. if (!region->mem) {
  347. return 0;
  348. }
  349. prot |= region->flags & VFIO_REGION_INFO_FLAG_READ ? PROT_READ : 0;
  350. prot |= region->flags & VFIO_REGION_INFO_FLAG_WRITE ? PROT_WRITE : 0;
  351. for (i = 0; i < region->nr_mmaps; i++) {
  352. size_t align = MIN(1ULL << ctz64(region->mmaps[i].size), 1 * GiB);
  353. void *map_base, *map_align;
  354. /*
  355. * Align the mmap for more efficient mapping in the kernel. Ideally
  356. * we'd know the PMD and PUD mapping sizes to use as discrete alignment
  357. * intervals, but we don't. As of Linux v6.12, the largest PUD size
  358. * supporting huge pfnmap is 1GiB (ARCH_SUPPORTS_PUD_PFNMAP is only set
  359. * on x86_64). Align by power-of-two size, capped at 1GiB.
  360. *
  361. * NB. qemu_memalign() and friends actually allocate memory, whereas
  362. * the region size here can exceed host memory, therefore we manually
  363. * create an oversized anonymous mapping and clean it up for alignment.
  364. */
  365. map_base = mmap(0, region->mmaps[i].size + align, PROT_NONE,
  366. MAP_PRIVATE | MAP_ANONYMOUS, -1, 0);
  367. if (map_base == MAP_FAILED) {
  368. ret = -errno;
  369. goto no_mmap;
  370. }
  371. map_align = (void *)ROUND_UP((uintptr_t)map_base, (uintptr_t)align);
  372. munmap(map_base, map_align - map_base);
  373. munmap(map_align + region->mmaps[i].size,
  374. align - (map_align - map_base));
  375. region->mmaps[i].mmap = mmap(map_align, region->mmaps[i].size, prot,
  376. MAP_SHARED | MAP_FIXED,
  377. region->vbasedev->fd,
  378. region->fd_offset +
  379. region->mmaps[i].offset);
  380. if (region->mmaps[i].mmap == MAP_FAILED) {
  381. ret = -errno;
  382. goto no_mmap;
  383. }
  384. name = g_strdup_printf("%s mmaps[%d]",
  385. memory_region_name(region->mem), i);
  386. memory_region_init_ram_device_ptr(&region->mmaps[i].mem,
  387. memory_region_owner(region->mem),
  388. name, region->mmaps[i].size,
  389. region->mmaps[i].mmap);
  390. g_free(name);
  391. memory_region_add_subregion(region->mem, region->mmaps[i].offset,
  392. &region->mmaps[i].mem);
  393. trace_vfio_region_mmap(memory_region_name(&region->mmaps[i].mem),
  394. region->mmaps[i].offset,
  395. region->mmaps[i].offset +
  396. region->mmaps[i].size - 1);
  397. }
  398. return 0;
  399. no_mmap:
  400. trace_vfio_region_mmap_fault(memory_region_name(region->mem), i,
  401. region->fd_offset + region->mmaps[i].offset,
  402. region->fd_offset + region->mmaps[i].offset +
  403. region->mmaps[i].size - 1, ret);
  404. region->mmaps[i].mmap = NULL;
  405. for (i--; i >= 0; i--) {
  406. vfio_subregion_unmap(region, i);
  407. }
  408. return ret;
  409. }
  410. void vfio_region_unmap(VFIORegion *region)
  411. {
  412. int i;
  413. if (!region->mem) {
  414. return;
  415. }
  416. for (i = 0; i < region->nr_mmaps; i++) {
  417. if (region->mmaps[i].mmap) {
  418. vfio_subregion_unmap(region, i);
  419. }
  420. }
  421. }
  422. void vfio_region_exit(VFIORegion *region)
  423. {
  424. int i;
  425. if (!region->mem) {
  426. return;
  427. }
  428. for (i = 0; i < region->nr_mmaps; i++) {
  429. if (region->mmaps[i].mmap) {
  430. memory_region_del_subregion(region->mem, &region->mmaps[i].mem);
  431. }
  432. }
  433. trace_vfio_region_exit(region->vbasedev->name, region->nr);
  434. }
  435. void vfio_region_finalize(VFIORegion *region)
  436. {
  437. int i;
  438. if (!region->mem) {
  439. return;
  440. }
  441. for (i = 0; i < region->nr_mmaps; i++) {
  442. if (region->mmaps[i].mmap) {
  443. munmap(region->mmaps[i].mmap, region->mmaps[i].size);
  444. object_unparent(OBJECT(&region->mmaps[i].mem));
  445. }
  446. }
  447. object_unparent(OBJECT(region->mem));
  448. g_free(region->mem);
  449. g_free(region->mmaps);
  450. trace_vfio_region_finalize(region->vbasedev->name, region->nr);
  451. region->mem = NULL;
  452. region->mmaps = NULL;
  453. region->nr_mmaps = 0;
  454. region->size = 0;
  455. region->flags = 0;
  456. region->nr = 0;
  457. }
  458. void vfio_region_mmaps_set_enabled(VFIORegion *region, bool enabled)
  459. {
  460. int i;
  461. if (!region->mem) {
  462. return;
  463. }
  464. for (i = 0; i < region->nr_mmaps; i++) {
  465. if (region->mmaps[i].mmap) {
  466. memory_region_set_enabled(&region->mmaps[i].mem, enabled);
  467. }
  468. }
  469. trace_vfio_region_mmaps_set_enabled(memory_region_name(region->mem),
  470. enabled);
  471. }
  472. int vfio_get_region_info(VFIODevice *vbasedev, int index,
  473. struct vfio_region_info **info)
  474. {
  475. size_t argsz = sizeof(struct vfio_region_info);
  476. *info = g_malloc0(argsz);
  477. (*info)->index = index;
  478. retry:
  479. (*info)->argsz = argsz;
  480. if (ioctl(vbasedev->fd, VFIO_DEVICE_GET_REGION_INFO, *info)) {
  481. g_free(*info);
  482. *info = NULL;
  483. return -errno;
  484. }
  485. if ((*info)->argsz > argsz) {
  486. argsz = (*info)->argsz;
  487. *info = g_realloc(*info, argsz);
  488. goto retry;
  489. }
  490. return 0;
  491. }
  492. int vfio_get_dev_region_info(VFIODevice *vbasedev, uint32_t type,
  493. uint32_t subtype, struct vfio_region_info **info)
  494. {
  495. int i;
  496. for (i = 0; i < vbasedev->num_regions; i++) {
  497. struct vfio_info_cap_header *hdr;
  498. struct vfio_region_info_cap_type *cap_type;
  499. if (vfio_get_region_info(vbasedev, i, info)) {
  500. continue;
  501. }
  502. hdr = vfio_get_region_info_cap(*info, VFIO_REGION_INFO_CAP_TYPE);
  503. if (!hdr) {
  504. g_free(*info);
  505. continue;
  506. }
  507. cap_type = container_of(hdr, struct vfio_region_info_cap_type, header);
  508. trace_vfio_get_dev_region(vbasedev->name, i,
  509. cap_type->type, cap_type->subtype);
  510. if (cap_type->type == type && cap_type->subtype == subtype) {
  511. return 0;
  512. }
  513. g_free(*info);
  514. }
  515. *info = NULL;
  516. return -ENODEV;
  517. }
  518. bool vfio_has_region_cap(VFIODevice *vbasedev, int region, uint16_t cap_type)
  519. {
  520. g_autofree struct vfio_region_info *info = NULL;
  521. bool ret = false;
  522. if (!vfio_get_region_info(vbasedev, region, &info)) {
  523. if (vfio_get_region_info_cap(info, cap_type)) {
  524. ret = true;
  525. }
  526. }
  527. return ret;
  528. }
  529. bool vfio_device_get_name(VFIODevice *vbasedev, Error **errp)
  530. {
  531. ERRP_GUARD();
  532. struct stat st;
  533. if (vbasedev->fd < 0) {
  534. if (stat(vbasedev->sysfsdev, &st) < 0) {
  535. error_setg_errno(errp, errno, "no such host device");
  536. error_prepend(errp, VFIO_MSG_PREFIX, vbasedev->sysfsdev);
  537. return false;
  538. }
  539. /* User may specify a name, e.g: VFIO platform device */
  540. if (!vbasedev->name) {
  541. vbasedev->name = g_path_get_basename(vbasedev->sysfsdev);
  542. }
  543. } else {
  544. if (!vbasedev->iommufd) {
  545. error_setg(errp, "Use FD passing only with iommufd backend");
  546. return false;
  547. }
  548. /*
  549. * Give a name with fd so any function printing out vbasedev->name
  550. * will not break.
  551. */
  552. if (!vbasedev->name) {
  553. vbasedev->name = g_strdup_printf("VFIO_FD%d", vbasedev->fd);
  554. }
  555. }
  556. return true;
  557. }
  558. void vfio_device_set_fd(VFIODevice *vbasedev, const char *str, Error **errp)
  559. {
  560. ERRP_GUARD();
  561. int fd = monitor_fd_param(monitor_cur(), str, errp);
  562. if (fd < 0) {
  563. error_prepend(errp, "Could not parse remote object fd %s:", str);
  564. return;
  565. }
  566. vbasedev->fd = fd;
  567. }
  568. void vfio_device_init(VFIODevice *vbasedev, int type, VFIODeviceOps *ops,
  569. DeviceState *dev, bool ram_discard)
  570. {
  571. vbasedev->type = type;
  572. vbasedev->ops = ops;
  573. vbasedev->dev = dev;
  574. vbasedev->fd = -1;
  575. vbasedev->ram_block_discard_allowed = ram_discard;
  576. }
  577. int vfio_device_get_aw_bits(VFIODevice *vdev)
  578. {
  579. /*
  580. * iova_ranges is a sorted list. For old kernels that support
  581. * VFIO but not support query of iova ranges, iova_ranges is NULL,
  582. * in this case HOST_IOMMU_DEVICE_CAP_AW_BITS_MAX(64) is returned.
  583. */
  584. GList *l = g_list_last(vdev->bcontainer->iova_ranges);
  585. if (l) {
  586. Range *range = l->data;
  587. return range_get_last_bit(range) + 1;
  588. }
  589. return HOST_IOMMU_DEVICE_CAP_AW_BITS_MAX;
  590. }
  591. bool vfio_device_is_mdev(VFIODevice *vbasedev)
  592. {
  593. g_autofree char *subsys = NULL;
  594. g_autofree char *tmp = NULL;
  595. if (!vbasedev->sysfsdev) {
  596. return false;
  597. }
  598. tmp = g_strdup_printf("%s/subsystem", vbasedev->sysfsdev);
  599. subsys = realpath(tmp, NULL);
  600. return subsys && (strcmp(subsys, "/sys/bus/mdev") == 0);
  601. }
  602. bool vfio_device_hiod_realize(VFIODevice *vbasedev, Error **errp)
  603. {
  604. HostIOMMUDevice *hiod = vbasedev->hiod;
  605. if (!hiod) {
  606. return true;
  607. }
  608. return HOST_IOMMU_DEVICE_GET_CLASS(hiod)->realize(hiod, vbasedev, errp);
  609. }