vmxnet3.c 74 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625162616271628162916301631163216331634163516361637163816391640164116421643164416451646164716481649165016511652165316541655165616571658165916601661166216631664166516661667166816691670167116721673167416751676167716781679168016811682168316841685168616871688168916901691169216931694169516961697169816991700170117021703170417051706170717081709171017111712171317141715171617171718171917201721172217231724172517261727172817291730173117321733173417351736173717381739174017411742174317441745174617471748174917501751175217531754175517561757175817591760176117621763176417651766176717681769177017711772177317741775177617771778177917801781178217831784178517861787178817891790179117921793179417951796179717981799180018011802180318041805180618071808180918101811181218131814181518161817181818191820182118221823182418251826182718281829183018311832183318341835183618371838183918401841184218431844184518461847184818491850185118521853185418551856185718581859186018611862186318641865186618671868186918701871187218731874187518761877187818791880188118821883188418851886188718881889189018911892189318941895189618971898189919001901190219031904190519061907190819091910191119121913191419151916191719181919192019211922192319241925192619271928192919301931193219331934193519361937193819391940194119421943194419451946194719481949195019511952195319541955195619571958195919601961196219631964196519661967196819691970197119721973197419751976197719781979198019811982198319841985198619871988198919901991199219931994199519961997199819992000200120022003200420052006200720082009201020112012201320142015201620172018201920202021202220232024202520262027202820292030203120322033203420352036203720382039204020412042204320442045204620472048204920502051205220532054205520562057205820592060206120622063206420652066206720682069207020712072207320742075207620772078207920802081208220832084208520862087208820892090209120922093209420952096209720982099210021012102210321042105210621072108210921102111211221132114211521162117211821192120212121222123212421252126212721282129213021312132213321342135213621372138213921402141214221432144214521462147214821492150215121522153215421552156215721582159216021612162216321642165216621672168216921702171217221732174217521762177217821792180218121822183218421852186218721882189219021912192219321942195219621972198219922002201220222032204220522062207220822092210221122122213221422152216221722182219222022212222222322242225222622272228222922302231223222332234223522362237223822392240224122422243224422452246224722482249225022512252225322542255225622572258225922602261226222632264226522662267226822692270227122722273227422752276227722782279228022812282228322842285228622872288228922902291229222932294229522962297229822992300230123022303230423052306230723082309231023112312231323142315231623172318231923202321232223232324232523262327232823292330233123322333233423352336233723382339234023412342234323442345234623472348234923502351235223532354235523562357235823592360236123622363236423652366236723682369237023712372237323742375237623772378237923802381238223832384238523862387238823892390239123922393239423952396239723982399240024012402240324042405240624072408240924102411241224132414241524162417241824192420242124222423242424252426242724282429243024312432243324342435243624372438243924402441244224432444244524462447244824492450245124522453245424552456245724582459246024612462246324642465246624672468246924702471247224732474247524762477247824792480248124822483248424852486248724882489249024912492249324942495249624972498249925002501250225032504250525062507250825092510251125122513251425152516251725182519252025212522252325242525252625272528252925302531
  1. /*
  2. * QEMU VMWARE VMXNET3 paravirtual NIC
  3. *
  4. * Copyright (c) 2012 Ravello Systems LTD (http://ravellosystems.com)
  5. *
  6. * Developed by Daynix Computing LTD (http://www.daynix.com)
  7. *
  8. * Authors:
  9. * Dmitry Fleytman <dmitry@daynix.com>
  10. * Tamir Shomer <tamirs@daynix.com>
  11. * Yan Vugenfirer <yan@daynix.com>
  12. *
  13. * This work is licensed under the terms of the GNU GPL, version 2.
  14. * See the COPYING file in the top-level directory.
  15. *
  16. */
  17. #include "hw/hw.h"
  18. #include "hw/pci/pci.h"
  19. #include "net/net.h"
  20. #include "net/tap.h"
  21. #include "net/checksum.h"
  22. #include "sysemu/sysemu.h"
  23. #include "qemu-common.h"
  24. #include "qemu/bswap.h"
  25. #include "hw/pci/msix.h"
  26. #include "hw/pci/msi.h"
  27. #include "vmxnet3.h"
  28. #include "vmxnet_debug.h"
  29. #include "vmware_utils.h"
  30. #include "vmxnet_tx_pkt.h"
  31. #include "vmxnet_rx_pkt.h"
  32. #define PCI_DEVICE_ID_VMWARE_VMXNET3_REVISION 0x1
  33. #define VMXNET3_MSIX_BAR_SIZE 0x2000
  34. #define VMXNET3_BAR0_IDX (0)
  35. #define VMXNET3_BAR1_IDX (1)
  36. #define VMXNET3_MSIX_BAR_IDX (2)
  37. #define VMXNET3_OFF_MSIX_TABLE (0x000)
  38. #define VMXNET3_OFF_MSIX_PBA (0x800)
  39. /* Link speed in Mbps should be shifted by 16 */
  40. #define VMXNET3_LINK_SPEED (1000 << 16)
  41. /* Link status: 1 - up, 0 - down. */
  42. #define VMXNET3_LINK_STATUS_UP 0x1
  43. /* Least significant bit should be set for revision and version */
  44. #define VMXNET3_DEVICE_VERSION 0x1
  45. #define VMXNET3_DEVICE_REVISION 0x1
  46. /* Number of interrupt vectors for non-MSIx modes */
  47. #define VMXNET3_MAX_NMSIX_INTRS (1)
  48. /* Macros for rings descriptors access */
  49. #define VMXNET3_READ_TX_QUEUE_DESCR8(dpa, field) \
  50. (vmw_shmem_ld8(dpa + offsetof(struct Vmxnet3_TxQueueDesc, field)))
  51. #define VMXNET3_WRITE_TX_QUEUE_DESCR8(dpa, field, value) \
  52. (vmw_shmem_st8(dpa + offsetof(struct Vmxnet3_TxQueueDesc, field, value)))
  53. #define VMXNET3_READ_TX_QUEUE_DESCR32(dpa, field) \
  54. (vmw_shmem_ld32(dpa + offsetof(struct Vmxnet3_TxQueueDesc, field)))
  55. #define VMXNET3_WRITE_TX_QUEUE_DESCR32(dpa, field, value) \
  56. (vmw_shmem_st32(dpa + offsetof(struct Vmxnet3_TxQueueDesc, field), value))
  57. #define VMXNET3_READ_TX_QUEUE_DESCR64(dpa, field) \
  58. (vmw_shmem_ld64(dpa + offsetof(struct Vmxnet3_TxQueueDesc, field)))
  59. #define VMXNET3_WRITE_TX_QUEUE_DESCR64(dpa, field, value) \
  60. (vmw_shmem_st64(dpa + offsetof(struct Vmxnet3_TxQueueDesc, field), value))
  61. #define VMXNET3_READ_RX_QUEUE_DESCR64(dpa, field) \
  62. (vmw_shmem_ld64(dpa + offsetof(struct Vmxnet3_RxQueueDesc, field)))
  63. #define VMXNET3_READ_RX_QUEUE_DESCR32(dpa, field) \
  64. (vmw_shmem_ld32(dpa + offsetof(struct Vmxnet3_RxQueueDesc, field)))
  65. #define VMXNET3_WRITE_RX_QUEUE_DESCR64(dpa, field, value) \
  66. (vmw_shmem_st64(dpa + offsetof(struct Vmxnet3_RxQueueDesc, field), value))
  67. #define VMXNET3_WRITE_RX_QUEUE_DESCR8(dpa, field, value) \
  68. (vmw_shmem_st8(dpa + offsetof(struct Vmxnet3_RxQueueDesc, field), value))
  69. /* Macros for guest driver shared area access */
  70. #define VMXNET3_READ_DRV_SHARED64(shpa, field) \
  71. (vmw_shmem_ld64(shpa + offsetof(struct Vmxnet3_DriverShared, field)))
  72. #define VMXNET3_READ_DRV_SHARED32(shpa, field) \
  73. (vmw_shmem_ld32(shpa + offsetof(struct Vmxnet3_DriverShared, field)))
  74. #define VMXNET3_WRITE_DRV_SHARED32(shpa, field, val) \
  75. (vmw_shmem_st32(shpa + offsetof(struct Vmxnet3_DriverShared, field), val))
  76. #define VMXNET3_READ_DRV_SHARED16(shpa, field) \
  77. (vmw_shmem_ld16(shpa + offsetof(struct Vmxnet3_DriverShared, field)))
  78. #define VMXNET3_READ_DRV_SHARED8(shpa, field) \
  79. (vmw_shmem_ld8(shpa + offsetof(struct Vmxnet3_DriverShared, field)))
  80. #define VMXNET3_READ_DRV_SHARED(shpa, field, b, l) \
  81. (vmw_shmem_read(shpa + offsetof(struct Vmxnet3_DriverShared, field), b, l))
  82. #define VMXNET_FLAG_IS_SET(field, flag) (((field) & (flag)) == (flag))
  83. #define TYPE_VMXNET3 "vmxnet3"
  84. #define VMXNET3(obj) OBJECT_CHECK(VMXNET3State, (obj), TYPE_VMXNET3)
  85. /* Cyclic ring abstraction */
  86. typedef struct {
  87. hwaddr pa;
  88. size_t size;
  89. size_t cell_size;
  90. size_t next;
  91. uint8_t gen;
  92. } Vmxnet3Ring;
  93. static inline void vmxnet3_ring_init(Vmxnet3Ring *ring,
  94. hwaddr pa,
  95. size_t size,
  96. size_t cell_size,
  97. bool zero_region)
  98. {
  99. ring->pa = pa;
  100. ring->size = size;
  101. ring->cell_size = cell_size;
  102. ring->gen = VMXNET3_INIT_GEN;
  103. ring->next = 0;
  104. if (zero_region) {
  105. vmw_shmem_set(pa, 0, size * cell_size);
  106. }
  107. }
  108. #define VMXNET3_RING_DUMP(macro, ring_name, ridx, r) \
  109. macro("%s#%d: base %" PRIx64 " size %lu cell_size %lu gen %d next %lu", \
  110. (ring_name), (ridx), \
  111. (r)->pa, (r)->size, (r)->cell_size, (r)->gen, (r)->next)
  112. static inline void vmxnet3_ring_inc(Vmxnet3Ring *ring)
  113. {
  114. if (++ring->next >= ring->size) {
  115. ring->next = 0;
  116. ring->gen ^= 1;
  117. }
  118. }
  119. static inline void vmxnet3_ring_dec(Vmxnet3Ring *ring)
  120. {
  121. if (ring->next-- == 0) {
  122. ring->next = ring->size - 1;
  123. ring->gen ^= 1;
  124. }
  125. }
  126. static inline hwaddr vmxnet3_ring_curr_cell_pa(Vmxnet3Ring *ring)
  127. {
  128. return ring->pa + ring->next * ring->cell_size;
  129. }
  130. static inline void vmxnet3_ring_read_curr_cell(Vmxnet3Ring *ring, void *buff)
  131. {
  132. vmw_shmem_read(vmxnet3_ring_curr_cell_pa(ring), buff, ring->cell_size);
  133. }
  134. static inline void vmxnet3_ring_write_curr_cell(Vmxnet3Ring *ring, void *buff)
  135. {
  136. vmw_shmem_write(vmxnet3_ring_curr_cell_pa(ring), buff, ring->cell_size);
  137. }
  138. static inline size_t vmxnet3_ring_curr_cell_idx(Vmxnet3Ring *ring)
  139. {
  140. return ring->next;
  141. }
  142. static inline uint8_t vmxnet3_ring_curr_gen(Vmxnet3Ring *ring)
  143. {
  144. return ring->gen;
  145. }
  146. /* Debug trace-related functions */
  147. static inline void
  148. vmxnet3_dump_tx_descr(struct Vmxnet3_TxDesc *descr)
  149. {
  150. VMW_PKPRN("TX DESCR: "
  151. "addr %" PRIx64 ", len: %d, gen: %d, rsvd: %d, "
  152. "dtype: %d, ext1: %d, msscof: %d, hlen: %d, om: %d, "
  153. "eop: %d, cq: %d, ext2: %d, ti: %d, tci: %d",
  154. le64_to_cpu(descr->addr), descr->len, descr->gen, descr->rsvd,
  155. descr->dtype, descr->ext1, descr->msscof, descr->hlen, descr->om,
  156. descr->eop, descr->cq, descr->ext2, descr->ti, descr->tci);
  157. }
  158. static inline void
  159. vmxnet3_dump_virt_hdr(struct virtio_net_hdr *vhdr)
  160. {
  161. VMW_PKPRN("VHDR: flags 0x%x, gso_type: 0x%x, hdr_len: %d, gso_size: %d, "
  162. "csum_start: %d, csum_offset: %d",
  163. vhdr->flags, vhdr->gso_type, vhdr->hdr_len, vhdr->gso_size,
  164. vhdr->csum_start, vhdr->csum_offset);
  165. }
  166. static inline void
  167. vmxnet3_dump_rx_descr(struct Vmxnet3_RxDesc *descr)
  168. {
  169. VMW_PKPRN("RX DESCR: addr %" PRIx64 ", len: %d, gen: %d, rsvd: %d, "
  170. "dtype: %d, ext1: %d, btype: %d",
  171. le64_to_cpu(descr->addr), descr->len, descr->gen,
  172. descr->rsvd, descr->dtype, descr->ext1, descr->btype);
  173. }
  174. /* Device state and helper functions */
  175. #define VMXNET3_RX_RINGS_PER_QUEUE (2)
  176. typedef struct {
  177. Vmxnet3Ring tx_ring;
  178. Vmxnet3Ring comp_ring;
  179. uint8_t intr_idx;
  180. hwaddr tx_stats_pa;
  181. struct UPT1_TxStats txq_stats;
  182. } Vmxnet3TxqDescr;
  183. typedef struct {
  184. Vmxnet3Ring rx_ring[VMXNET3_RX_RINGS_PER_QUEUE];
  185. Vmxnet3Ring comp_ring;
  186. uint8_t intr_idx;
  187. hwaddr rx_stats_pa;
  188. struct UPT1_RxStats rxq_stats;
  189. } Vmxnet3RxqDescr;
  190. typedef struct {
  191. bool is_masked;
  192. bool is_pending;
  193. bool is_asserted;
  194. } Vmxnet3IntState;
  195. typedef struct {
  196. PCIDevice parent_obj;
  197. NICState *nic;
  198. NICConf conf;
  199. MemoryRegion bar0;
  200. MemoryRegion bar1;
  201. MemoryRegion msix_bar;
  202. Vmxnet3RxqDescr rxq_descr[VMXNET3_DEVICE_MAX_RX_QUEUES];
  203. Vmxnet3TxqDescr txq_descr[VMXNET3_DEVICE_MAX_TX_QUEUES];
  204. /* Whether MSI-X support was installed successfully */
  205. bool msix_used;
  206. /* Whether MSI support was installed successfully */
  207. bool msi_used;
  208. hwaddr drv_shmem;
  209. hwaddr temp_shared_guest_driver_memory;
  210. uint8_t txq_num;
  211. /* This boolean tells whether RX packet being indicated has to */
  212. /* be split into head and body chunks from different RX rings */
  213. bool rx_packets_compound;
  214. bool rx_vlan_stripping;
  215. bool lro_supported;
  216. uint8_t rxq_num;
  217. /* Network MTU */
  218. uint32_t mtu;
  219. /* Maximum number of fragments for indicated TX packets */
  220. uint32_t max_tx_frags;
  221. /* Maximum number of fragments for indicated RX packets */
  222. uint16_t max_rx_frags;
  223. /* Index for events interrupt */
  224. uint8_t event_int_idx;
  225. /* Whether automatic interrupts masking enabled */
  226. bool auto_int_masking;
  227. bool peer_has_vhdr;
  228. /* TX packets to QEMU interface */
  229. struct VmxnetTxPkt *tx_pkt;
  230. uint32_t offload_mode;
  231. uint32_t cso_or_gso_size;
  232. uint16_t tci;
  233. bool needs_vlan;
  234. struct VmxnetRxPkt *rx_pkt;
  235. bool tx_sop;
  236. bool skip_current_tx_pkt;
  237. uint32_t device_active;
  238. uint32_t last_command;
  239. uint32_t link_status_and_speed;
  240. Vmxnet3IntState interrupt_states[VMXNET3_MAX_INTRS];
  241. uint32_t temp_mac; /* To store the low part first */
  242. MACAddr perm_mac;
  243. uint32_t vlan_table[VMXNET3_VFT_SIZE];
  244. uint32_t rx_mode;
  245. MACAddr *mcast_list;
  246. uint32_t mcast_list_len;
  247. uint32_t mcast_list_buff_size; /* needed for live migration. */
  248. } VMXNET3State;
  249. /* Interrupt management */
  250. /*
  251. *This function returns sign whether interrupt line is in asserted state
  252. * This depends on the type of interrupt used. For INTX interrupt line will
  253. * be asserted until explicit deassertion, for MSI(X) interrupt line will
  254. * be deasserted automatically due to notification semantics of the MSI(X)
  255. * interrupts
  256. */
  257. static bool _vmxnet3_assert_interrupt_line(VMXNET3State *s, uint32_t int_idx)
  258. {
  259. PCIDevice *d = PCI_DEVICE(s);
  260. if (s->msix_used && msix_enabled(d)) {
  261. VMW_IRPRN("Sending MSI-X notification for vector %u", int_idx);
  262. msix_notify(d, int_idx);
  263. return false;
  264. }
  265. if (s->msi_used && msi_enabled(d)) {
  266. VMW_IRPRN("Sending MSI notification for vector %u", int_idx);
  267. msi_notify(d, int_idx);
  268. return false;
  269. }
  270. VMW_IRPRN("Asserting line for interrupt %u", int_idx);
  271. pci_irq_assert(d);
  272. return true;
  273. }
  274. static void _vmxnet3_deassert_interrupt_line(VMXNET3State *s, int lidx)
  275. {
  276. PCIDevice *d = PCI_DEVICE(s);
  277. /*
  278. * This function should never be called for MSI(X) interrupts
  279. * because deassertion never required for message interrupts
  280. */
  281. assert(!s->msix_used || !msix_enabled(d));
  282. /*
  283. * This function should never be called for MSI(X) interrupts
  284. * because deassertion never required for message interrupts
  285. */
  286. assert(!s->msi_used || !msi_enabled(d));
  287. VMW_IRPRN("Deasserting line for interrupt %u", lidx);
  288. pci_irq_deassert(d);
  289. }
  290. static void vmxnet3_update_interrupt_line_state(VMXNET3State *s, int lidx)
  291. {
  292. if (!s->interrupt_states[lidx].is_pending &&
  293. s->interrupt_states[lidx].is_asserted) {
  294. VMW_IRPRN("New interrupt line state for index %d is DOWN", lidx);
  295. _vmxnet3_deassert_interrupt_line(s, lidx);
  296. s->interrupt_states[lidx].is_asserted = false;
  297. return;
  298. }
  299. if (s->interrupt_states[lidx].is_pending &&
  300. !s->interrupt_states[lidx].is_masked &&
  301. !s->interrupt_states[lidx].is_asserted) {
  302. VMW_IRPRN("New interrupt line state for index %d is UP", lidx);
  303. s->interrupt_states[lidx].is_asserted =
  304. _vmxnet3_assert_interrupt_line(s, lidx);
  305. s->interrupt_states[lidx].is_pending = false;
  306. return;
  307. }
  308. }
  309. static void vmxnet3_trigger_interrupt(VMXNET3State *s, int lidx)
  310. {
  311. PCIDevice *d = PCI_DEVICE(s);
  312. s->interrupt_states[lidx].is_pending = true;
  313. vmxnet3_update_interrupt_line_state(s, lidx);
  314. if (s->msix_used && msix_enabled(d) && s->auto_int_masking) {
  315. goto do_automask;
  316. }
  317. if (s->msi_used && msi_enabled(d) && s->auto_int_masking) {
  318. goto do_automask;
  319. }
  320. return;
  321. do_automask:
  322. s->interrupt_states[lidx].is_masked = true;
  323. vmxnet3_update_interrupt_line_state(s, lidx);
  324. }
  325. static bool vmxnet3_interrupt_asserted(VMXNET3State *s, int lidx)
  326. {
  327. return s->interrupt_states[lidx].is_asserted;
  328. }
  329. static void vmxnet3_clear_interrupt(VMXNET3State *s, int int_idx)
  330. {
  331. s->interrupt_states[int_idx].is_pending = false;
  332. if (s->auto_int_masking) {
  333. s->interrupt_states[int_idx].is_masked = true;
  334. }
  335. vmxnet3_update_interrupt_line_state(s, int_idx);
  336. }
  337. static void
  338. vmxnet3_on_interrupt_mask_changed(VMXNET3State *s, int lidx, bool is_masked)
  339. {
  340. s->interrupt_states[lidx].is_masked = is_masked;
  341. vmxnet3_update_interrupt_line_state(s, lidx);
  342. }
  343. static bool vmxnet3_verify_driver_magic(hwaddr dshmem)
  344. {
  345. return (VMXNET3_READ_DRV_SHARED32(dshmem, magic) == VMXNET3_REV1_MAGIC);
  346. }
  347. #define VMXNET3_GET_BYTE(x, byte_num) (((x) >> (byte_num)*8) & 0xFF)
  348. #define VMXNET3_MAKE_BYTE(byte_num, val) \
  349. (((uint32_t)((val) & 0xFF)) << (byte_num)*8)
  350. static void vmxnet3_set_variable_mac(VMXNET3State *s, uint32_t h, uint32_t l)
  351. {
  352. s->conf.macaddr.a[0] = VMXNET3_GET_BYTE(l, 0);
  353. s->conf.macaddr.a[1] = VMXNET3_GET_BYTE(l, 1);
  354. s->conf.macaddr.a[2] = VMXNET3_GET_BYTE(l, 2);
  355. s->conf.macaddr.a[3] = VMXNET3_GET_BYTE(l, 3);
  356. s->conf.macaddr.a[4] = VMXNET3_GET_BYTE(h, 0);
  357. s->conf.macaddr.a[5] = VMXNET3_GET_BYTE(h, 1);
  358. VMW_CFPRN("Variable MAC: " VMXNET_MF, VMXNET_MA(s->conf.macaddr.a));
  359. qemu_format_nic_info_str(qemu_get_queue(s->nic), s->conf.macaddr.a);
  360. }
  361. static uint64_t vmxnet3_get_mac_low(MACAddr *addr)
  362. {
  363. return VMXNET3_MAKE_BYTE(0, addr->a[0]) |
  364. VMXNET3_MAKE_BYTE(1, addr->a[1]) |
  365. VMXNET3_MAKE_BYTE(2, addr->a[2]) |
  366. VMXNET3_MAKE_BYTE(3, addr->a[3]);
  367. }
  368. static uint64_t vmxnet3_get_mac_high(MACAddr *addr)
  369. {
  370. return VMXNET3_MAKE_BYTE(0, addr->a[4]) |
  371. VMXNET3_MAKE_BYTE(1, addr->a[5]);
  372. }
  373. static void
  374. vmxnet3_inc_tx_consumption_counter(VMXNET3State *s, int qidx)
  375. {
  376. vmxnet3_ring_inc(&s->txq_descr[qidx].tx_ring);
  377. }
  378. static inline void
  379. vmxnet3_inc_rx_consumption_counter(VMXNET3State *s, int qidx, int ridx)
  380. {
  381. vmxnet3_ring_inc(&s->rxq_descr[qidx].rx_ring[ridx]);
  382. }
  383. static inline void
  384. vmxnet3_inc_tx_completion_counter(VMXNET3State *s, int qidx)
  385. {
  386. vmxnet3_ring_inc(&s->txq_descr[qidx].comp_ring);
  387. }
  388. static void
  389. vmxnet3_inc_rx_completion_counter(VMXNET3State *s, int qidx)
  390. {
  391. vmxnet3_ring_inc(&s->rxq_descr[qidx].comp_ring);
  392. }
  393. static void
  394. vmxnet3_dec_rx_completion_counter(VMXNET3State *s, int qidx)
  395. {
  396. vmxnet3_ring_dec(&s->rxq_descr[qidx].comp_ring);
  397. }
  398. static void vmxnet3_complete_packet(VMXNET3State *s, int qidx, uint32 tx_ridx)
  399. {
  400. struct Vmxnet3_TxCompDesc txcq_descr;
  401. VMXNET3_RING_DUMP(VMW_RIPRN, "TXC", qidx, &s->txq_descr[qidx].comp_ring);
  402. txcq_descr.txdIdx = tx_ridx;
  403. txcq_descr.gen = vmxnet3_ring_curr_gen(&s->txq_descr[qidx].comp_ring);
  404. vmxnet3_ring_write_curr_cell(&s->txq_descr[qidx].comp_ring, &txcq_descr);
  405. /* Flush changes in TX descriptor before changing the counter value */
  406. smp_wmb();
  407. vmxnet3_inc_tx_completion_counter(s, qidx);
  408. vmxnet3_trigger_interrupt(s, s->txq_descr[qidx].intr_idx);
  409. }
  410. static bool
  411. vmxnet3_setup_tx_offloads(VMXNET3State *s)
  412. {
  413. switch (s->offload_mode) {
  414. case VMXNET3_OM_NONE:
  415. vmxnet_tx_pkt_build_vheader(s->tx_pkt, false, false, 0);
  416. break;
  417. case VMXNET3_OM_CSUM:
  418. vmxnet_tx_pkt_build_vheader(s->tx_pkt, false, true, 0);
  419. VMW_PKPRN("L4 CSO requested\n");
  420. break;
  421. case VMXNET3_OM_TSO:
  422. vmxnet_tx_pkt_build_vheader(s->tx_pkt, true, true,
  423. s->cso_or_gso_size);
  424. vmxnet_tx_pkt_update_ip_checksums(s->tx_pkt);
  425. VMW_PKPRN("GSO offload requested.");
  426. break;
  427. default:
  428. g_assert_not_reached();
  429. return false;
  430. }
  431. return true;
  432. }
  433. static void
  434. vmxnet3_tx_retrieve_metadata(VMXNET3State *s,
  435. const struct Vmxnet3_TxDesc *txd)
  436. {
  437. s->offload_mode = txd->om;
  438. s->cso_or_gso_size = txd->msscof;
  439. s->tci = txd->tci;
  440. s->needs_vlan = txd->ti;
  441. }
  442. typedef enum {
  443. VMXNET3_PKT_STATUS_OK,
  444. VMXNET3_PKT_STATUS_ERROR,
  445. VMXNET3_PKT_STATUS_DISCARD,/* only for tx */
  446. VMXNET3_PKT_STATUS_OUT_OF_BUF /* only for rx */
  447. } Vmxnet3PktStatus;
  448. static void
  449. vmxnet3_on_tx_done_update_stats(VMXNET3State *s, int qidx,
  450. Vmxnet3PktStatus status)
  451. {
  452. size_t tot_len = vmxnet_tx_pkt_get_total_len(s->tx_pkt);
  453. struct UPT1_TxStats *stats = &s->txq_descr[qidx].txq_stats;
  454. switch (status) {
  455. case VMXNET3_PKT_STATUS_OK:
  456. switch (vmxnet_tx_pkt_get_packet_type(s->tx_pkt)) {
  457. case ETH_PKT_BCAST:
  458. stats->bcastPktsTxOK++;
  459. stats->bcastBytesTxOK += tot_len;
  460. break;
  461. case ETH_PKT_MCAST:
  462. stats->mcastPktsTxOK++;
  463. stats->mcastBytesTxOK += tot_len;
  464. break;
  465. case ETH_PKT_UCAST:
  466. stats->ucastPktsTxOK++;
  467. stats->ucastBytesTxOK += tot_len;
  468. break;
  469. default:
  470. g_assert_not_reached();
  471. }
  472. if (s->offload_mode == VMXNET3_OM_TSO) {
  473. /*
  474. * According to VMWARE headers this statistic is a number
  475. * of packets after segmentation but since we don't have
  476. * this information in QEMU model, the best we can do is to
  477. * provide number of non-segmented packets
  478. */
  479. stats->TSOPktsTxOK++;
  480. stats->TSOBytesTxOK += tot_len;
  481. }
  482. break;
  483. case VMXNET3_PKT_STATUS_DISCARD:
  484. stats->pktsTxDiscard++;
  485. break;
  486. case VMXNET3_PKT_STATUS_ERROR:
  487. stats->pktsTxError++;
  488. break;
  489. default:
  490. g_assert_not_reached();
  491. }
  492. }
  493. static void
  494. vmxnet3_on_rx_done_update_stats(VMXNET3State *s,
  495. int qidx,
  496. Vmxnet3PktStatus status)
  497. {
  498. struct UPT1_RxStats *stats = &s->rxq_descr[qidx].rxq_stats;
  499. size_t tot_len = vmxnet_rx_pkt_get_total_len(s->rx_pkt);
  500. switch (status) {
  501. case VMXNET3_PKT_STATUS_OUT_OF_BUF:
  502. stats->pktsRxOutOfBuf++;
  503. break;
  504. case VMXNET3_PKT_STATUS_ERROR:
  505. stats->pktsRxError++;
  506. break;
  507. case VMXNET3_PKT_STATUS_OK:
  508. switch (vmxnet_rx_pkt_get_packet_type(s->rx_pkt)) {
  509. case ETH_PKT_BCAST:
  510. stats->bcastPktsRxOK++;
  511. stats->bcastBytesRxOK += tot_len;
  512. break;
  513. case ETH_PKT_MCAST:
  514. stats->mcastPktsRxOK++;
  515. stats->mcastBytesRxOK += tot_len;
  516. break;
  517. case ETH_PKT_UCAST:
  518. stats->ucastPktsRxOK++;
  519. stats->ucastBytesRxOK += tot_len;
  520. break;
  521. default:
  522. g_assert_not_reached();
  523. }
  524. if (tot_len > s->mtu) {
  525. stats->LROPktsRxOK++;
  526. stats->LROBytesRxOK += tot_len;
  527. }
  528. break;
  529. default:
  530. g_assert_not_reached();
  531. }
  532. }
  533. static inline bool
  534. vmxnet3_pop_next_tx_descr(VMXNET3State *s,
  535. int qidx,
  536. struct Vmxnet3_TxDesc *txd,
  537. uint32_t *descr_idx)
  538. {
  539. Vmxnet3Ring *ring = &s->txq_descr[qidx].tx_ring;
  540. vmxnet3_ring_read_curr_cell(ring, txd);
  541. if (txd->gen == vmxnet3_ring_curr_gen(ring)) {
  542. /* Only read after generation field verification */
  543. smp_rmb();
  544. /* Re-read to be sure we got the latest version */
  545. vmxnet3_ring_read_curr_cell(ring, txd);
  546. VMXNET3_RING_DUMP(VMW_RIPRN, "TX", qidx, ring);
  547. *descr_idx = vmxnet3_ring_curr_cell_idx(ring);
  548. vmxnet3_inc_tx_consumption_counter(s, qidx);
  549. return true;
  550. }
  551. return false;
  552. }
  553. static bool
  554. vmxnet3_send_packet(VMXNET3State *s, uint32_t qidx)
  555. {
  556. Vmxnet3PktStatus status = VMXNET3_PKT_STATUS_OK;
  557. if (!vmxnet3_setup_tx_offloads(s)) {
  558. status = VMXNET3_PKT_STATUS_ERROR;
  559. goto func_exit;
  560. }
  561. /* debug prints */
  562. vmxnet3_dump_virt_hdr(vmxnet_tx_pkt_get_vhdr(s->tx_pkt));
  563. vmxnet_tx_pkt_dump(s->tx_pkt);
  564. if (!vmxnet_tx_pkt_send(s->tx_pkt, qemu_get_queue(s->nic))) {
  565. status = VMXNET3_PKT_STATUS_DISCARD;
  566. goto func_exit;
  567. }
  568. func_exit:
  569. vmxnet3_on_tx_done_update_stats(s, qidx, status);
  570. return (status == VMXNET3_PKT_STATUS_OK);
  571. }
  572. static void vmxnet3_process_tx_queue(VMXNET3State *s, int qidx)
  573. {
  574. struct Vmxnet3_TxDesc txd;
  575. uint32_t txd_idx;
  576. uint32_t data_len;
  577. hwaddr data_pa;
  578. for (;;) {
  579. if (!vmxnet3_pop_next_tx_descr(s, qidx, &txd, &txd_idx)) {
  580. break;
  581. }
  582. vmxnet3_dump_tx_descr(&txd);
  583. if (!s->skip_current_tx_pkt) {
  584. data_len = (txd.len > 0) ? txd.len : VMXNET3_MAX_TX_BUF_SIZE;
  585. data_pa = le64_to_cpu(txd.addr);
  586. if (!vmxnet_tx_pkt_add_raw_fragment(s->tx_pkt,
  587. data_pa,
  588. data_len)) {
  589. s->skip_current_tx_pkt = true;
  590. }
  591. }
  592. if (s->tx_sop) {
  593. vmxnet3_tx_retrieve_metadata(s, &txd);
  594. s->tx_sop = false;
  595. }
  596. if (txd.eop) {
  597. if (!s->skip_current_tx_pkt) {
  598. vmxnet_tx_pkt_parse(s->tx_pkt);
  599. if (s->needs_vlan) {
  600. vmxnet_tx_pkt_setup_vlan_header(s->tx_pkt, s->tci);
  601. }
  602. vmxnet3_send_packet(s, qidx);
  603. } else {
  604. vmxnet3_on_tx_done_update_stats(s, qidx,
  605. VMXNET3_PKT_STATUS_ERROR);
  606. }
  607. vmxnet3_complete_packet(s, qidx, txd_idx);
  608. s->tx_sop = true;
  609. s->skip_current_tx_pkt = false;
  610. vmxnet_tx_pkt_reset(s->tx_pkt);
  611. }
  612. }
  613. }
  614. static inline void
  615. vmxnet3_read_next_rx_descr(VMXNET3State *s, int qidx, int ridx,
  616. struct Vmxnet3_RxDesc *dbuf, uint32_t *didx)
  617. {
  618. Vmxnet3Ring *ring = &s->rxq_descr[qidx].rx_ring[ridx];
  619. *didx = vmxnet3_ring_curr_cell_idx(ring);
  620. vmxnet3_ring_read_curr_cell(ring, dbuf);
  621. }
  622. static inline uint8_t
  623. vmxnet3_get_rx_ring_gen(VMXNET3State *s, int qidx, int ridx)
  624. {
  625. return s->rxq_descr[qidx].rx_ring[ridx].gen;
  626. }
  627. static inline hwaddr
  628. vmxnet3_pop_rxc_descr(VMXNET3State *s, int qidx, uint32_t *descr_gen)
  629. {
  630. uint8_t ring_gen;
  631. struct Vmxnet3_RxCompDesc rxcd;
  632. hwaddr daddr =
  633. vmxnet3_ring_curr_cell_pa(&s->rxq_descr[qidx].comp_ring);
  634. cpu_physical_memory_read(daddr, &rxcd, sizeof(struct Vmxnet3_RxCompDesc));
  635. ring_gen = vmxnet3_ring_curr_gen(&s->rxq_descr[qidx].comp_ring);
  636. if (rxcd.gen != ring_gen) {
  637. *descr_gen = ring_gen;
  638. vmxnet3_inc_rx_completion_counter(s, qidx);
  639. return daddr;
  640. }
  641. return 0;
  642. }
  643. static inline void
  644. vmxnet3_revert_rxc_descr(VMXNET3State *s, int qidx)
  645. {
  646. vmxnet3_dec_rx_completion_counter(s, qidx);
  647. }
  648. #define RXQ_IDX (0)
  649. #define RX_HEAD_BODY_RING (0)
  650. #define RX_BODY_ONLY_RING (1)
  651. static bool
  652. vmxnet3_get_next_head_rx_descr(VMXNET3State *s,
  653. struct Vmxnet3_RxDesc *descr_buf,
  654. uint32_t *descr_idx,
  655. uint32_t *ridx)
  656. {
  657. for (;;) {
  658. uint32_t ring_gen;
  659. vmxnet3_read_next_rx_descr(s, RXQ_IDX, RX_HEAD_BODY_RING,
  660. descr_buf, descr_idx);
  661. /* If no more free descriptors - return */
  662. ring_gen = vmxnet3_get_rx_ring_gen(s, RXQ_IDX, RX_HEAD_BODY_RING);
  663. if (descr_buf->gen != ring_gen) {
  664. return false;
  665. }
  666. /* Only read after generation field verification */
  667. smp_rmb();
  668. /* Re-read to be sure we got the latest version */
  669. vmxnet3_read_next_rx_descr(s, RXQ_IDX, RX_HEAD_BODY_RING,
  670. descr_buf, descr_idx);
  671. /* Mark current descriptor as used/skipped */
  672. vmxnet3_inc_rx_consumption_counter(s, RXQ_IDX, RX_HEAD_BODY_RING);
  673. /* If this is what we are looking for - return */
  674. if (descr_buf->btype == VMXNET3_RXD_BTYPE_HEAD) {
  675. *ridx = RX_HEAD_BODY_RING;
  676. return true;
  677. }
  678. }
  679. }
  680. static bool
  681. vmxnet3_get_next_body_rx_descr(VMXNET3State *s,
  682. struct Vmxnet3_RxDesc *d,
  683. uint32_t *didx,
  684. uint32_t *ridx)
  685. {
  686. vmxnet3_read_next_rx_descr(s, RXQ_IDX, RX_HEAD_BODY_RING, d, didx);
  687. /* Try to find corresponding descriptor in head/body ring */
  688. if (d->gen == vmxnet3_get_rx_ring_gen(s, RXQ_IDX, RX_HEAD_BODY_RING)) {
  689. /* Only read after generation field verification */
  690. smp_rmb();
  691. /* Re-read to be sure we got the latest version */
  692. vmxnet3_read_next_rx_descr(s, RXQ_IDX, RX_HEAD_BODY_RING, d, didx);
  693. if (d->btype == VMXNET3_RXD_BTYPE_BODY) {
  694. vmxnet3_inc_rx_consumption_counter(s, RXQ_IDX, RX_HEAD_BODY_RING);
  695. *ridx = RX_HEAD_BODY_RING;
  696. return true;
  697. }
  698. }
  699. /*
  700. * If there is no free descriptors on head/body ring or next free
  701. * descriptor is a head descriptor switch to body only ring
  702. */
  703. vmxnet3_read_next_rx_descr(s, RXQ_IDX, RX_BODY_ONLY_RING, d, didx);
  704. /* If no more free descriptors - return */
  705. if (d->gen == vmxnet3_get_rx_ring_gen(s, RXQ_IDX, RX_BODY_ONLY_RING)) {
  706. /* Only read after generation field verification */
  707. smp_rmb();
  708. /* Re-read to be sure we got the latest version */
  709. vmxnet3_read_next_rx_descr(s, RXQ_IDX, RX_BODY_ONLY_RING, d, didx);
  710. assert(d->btype == VMXNET3_RXD_BTYPE_BODY);
  711. *ridx = RX_BODY_ONLY_RING;
  712. vmxnet3_inc_rx_consumption_counter(s, RXQ_IDX, RX_BODY_ONLY_RING);
  713. return true;
  714. }
  715. return false;
  716. }
  717. static inline bool
  718. vmxnet3_get_next_rx_descr(VMXNET3State *s, bool is_head,
  719. struct Vmxnet3_RxDesc *descr_buf,
  720. uint32_t *descr_idx,
  721. uint32_t *ridx)
  722. {
  723. if (is_head || !s->rx_packets_compound) {
  724. return vmxnet3_get_next_head_rx_descr(s, descr_buf, descr_idx, ridx);
  725. } else {
  726. return vmxnet3_get_next_body_rx_descr(s, descr_buf, descr_idx, ridx);
  727. }
  728. }
  729. static void vmxnet3_rx_update_descr(struct VmxnetRxPkt *pkt,
  730. struct Vmxnet3_RxCompDesc *rxcd)
  731. {
  732. int csum_ok, is_gso;
  733. bool isip4, isip6, istcp, isudp;
  734. struct virtio_net_hdr *vhdr;
  735. uint8_t offload_type;
  736. if (vmxnet_rx_pkt_is_vlan_stripped(pkt)) {
  737. rxcd->ts = 1;
  738. rxcd->tci = vmxnet_rx_pkt_get_vlan_tag(pkt);
  739. }
  740. if (!vmxnet_rx_pkt_has_virt_hdr(pkt)) {
  741. goto nocsum;
  742. }
  743. vhdr = vmxnet_rx_pkt_get_vhdr(pkt);
  744. /*
  745. * Checksum is valid when lower level tell so or when lower level
  746. * requires checksum offload telling that packet produced/bridged
  747. * locally and did travel over network after last checksum calculation
  748. * or production
  749. */
  750. csum_ok = VMXNET_FLAG_IS_SET(vhdr->flags, VIRTIO_NET_HDR_F_DATA_VALID) ||
  751. VMXNET_FLAG_IS_SET(vhdr->flags, VIRTIO_NET_HDR_F_NEEDS_CSUM);
  752. offload_type = vhdr->gso_type & ~VIRTIO_NET_HDR_GSO_ECN;
  753. is_gso = (offload_type != VIRTIO_NET_HDR_GSO_NONE) ? 1 : 0;
  754. if (!csum_ok && !is_gso) {
  755. goto nocsum;
  756. }
  757. vmxnet_rx_pkt_get_protocols(pkt, &isip4, &isip6, &isudp, &istcp);
  758. if ((!istcp && !isudp) || (!isip4 && !isip6)) {
  759. goto nocsum;
  760. }
  761. rxcd->cnc = 0;
  762. rxcd->v4 = isip4 ? 1 : 0;
  763. rxcd->v6 = isip6 ? 1 : 0;
  764. rxcd->tcp = istcp ? 1 : 0;
  765. rxcd->udp = isudp ? 1 : 0;
  766. rxcd->fcs = rxcd->tuc = rxcd->ipc = 1;
  767. return;
  768. nocsum:
  769. rxcd->cnc = 1;
  770. return;
  771. }
  772. static void
  773. vmxnet3_physical_memory_writev(const struct iovec *iov,
  774. size_t start_iov_off,
  775. hwaddr target_addr,
  776. size_t bytes_to_copy)
  777. {
  778. size_t curr_off = 0;
  779. size_t copied = 0;
  780. while (bytes_to_copy) {
  781. if (start_iov_off < (curr_off + iov->iov_len)) {
  782. size_t chunk_len =
  783. MIN((curr_off + iov->iov_len) - start_iov_off, bytes_to_copy);
  784. cpu_physical_memory_write(target_addr + copied,
  785. iov->iov_base + start_iov_off - curr_off,
  786. chunk_len);
  787. copied += chunk_len;
  788. start_iov_off += chunk_len;
  789. curr_off = start_iov_off;
  790. bytes_to_copy -= chunk_len;
  791. } else {
  792. curr_off += iov->iov_len;
  793. }
  794. iov++;
  795. }
  796. }
  797. static bool
  798. vmxnet3_indicate_packet(VMXNET3State *s)
  799. {
  800. struct Vmxnet3_RxDesc rxd;
  801. bool is_head = true;
  802. uint32_t rxd_idx;
  803. uint32_t rx_ridx = 0;
  804. struct Vmxnet3_RxCompDesc rxcd;
  805. uint32_t new_rxcd_gen = VMXNET3_INIT_GEN;
  806. hwaddr new_rxcd_pa = 0;
  807. hwaddr ready_rxcd_pa = 0;
  808. struct iovec *data = vmxnet_rx_pkt_get_iovec(s->rx_pkt);
  809. size_t bytes_copied = 0;
  810. size_t bytes_left = vmxnet_rx_pkt_get_total_len(s->rx_pkt);
  811. uint16_t num_frags = 0;
  812. size_t chunk_size;
  813. vmxnet_rx_pkt_dump(s->rx_pkt);
  814. while (bytes_left > 0) {
  815. /* cannot add more frags to packet */
  816. if (num_frags == s->max_rx_frags) {
  817. break;
  818. }
  819. new_rxcd_pa = vmxnet3_pop_rxc_descr(s, RXQ_IDX, &new_rxcd_gen);
  820. if (!new_rxcd_pa) {
  821. break;
  822. }
  823. if (!vmxnet3_get_next_rx_descr(s, is_head, &rxd, &rxd_idx, &rx_ridx)) {
  824. break;
  825. }
  826. chunk_size = MIN(bytes_left, rxd.len);
  827. vmxnet3_physical_memory_writev(data, bytes_copied,
  828. le64_to_cpu(rxd.addr), chunk_size);
  829. bytes_copied += chunk_size;
  830. bytes_left -= chunk_size;
  831. vmxnet3_dump_rx_descr(&rxd);
  832. if (0 != ready_rxcd_pa) {
  833. cpu_physical_memory_write(ready_rxcd_pa, &rxcd, sizeof(rxcd));
  834. }
  835. memset(&rxcd, 0, sizeof(struct Vmxnet3_RxCompDesc));
  836. rxcd.rxdIdx = rxd_idx;
  837. rxcd.len = chunk_size;
  838. rxcd.sop = is_head;
  839. rxcd.gen = new_rxcd_gen;
  840. rxcd.rqID = RXQ_IDX + rx_ridx * s->rxq_num;
  841. if (0 == bytes_left) {
  842. vmxnet3_rx_update_descr(s->rx_pkt, &rxcd);
  843. }
  844. VMW_RIPRN("RX Completion descriptor: rxRing: %lu rxIdx %lu len %lu "
  845. "sop %d csum_correct %lu",
  846. (unsigned long) rx_ridx,
  847. (unsigned long) rxcd.rxdIdx,
  848. (unsigned long) rxcd.len,
  849. (int) rxcd.sop,
  850. (unsigned long) rxcd.tuc);
  851. is_head = false;
  852. ready_rxcd_pa = new_rxcd_pa;
  853. new_rxcd_pa = 0;
  854. num_frags++;
  855. }
  856. if (0 != ready_rxcd_pa) {
  857. rxcd.eop = 1;
  858. rxcd.err = (0 != bytes_left);
  859. cpu_physical_memory_write(ready_rxcd_pa, &rxcd, sizeof(rxcd));
  860. /* Flush RX descriptor changes */
  861. smp_wmb();
  862. }
  863. if (0 != new_rxcd_pa) {
  864. vmxnet3_revert_rxc_descr(s, RXQ_IDX);
  865. }
  866. vmxnet3_trigger_interrupt(s, s->rxq_descr[RXQ_IDX].intr_idx);
  867. if (bytes_left == 0) {
  868. vmxnet3_on_rx_done_update_stats(s, RXQ_IDX, VMXNET3_PKT_STATUS_OK);
  869. return true;
  870. } else if (num_frags == s->max_rx_frags) {
  871. vmxnet3_on_rx_done_update_stats(s, RXQ_IDX, VMXNET3_PKT_STATUS_ERROR);
  872. return false;
  873. } else {
  874. vmxnet3_on_rx_done_update_stats(s, RXQ_IDX,
  875. VMXNET3_PKT_STATUS_OUT_OF_BUF);
  876. return false;
  877. }
  878. }
  879. static void
  880. vmxnet3_io_bar0_write(void *opaque, hwaddr addr,
  881. uint64_t val, unsigned size)
  882. {
  883. VMXNET3State *s = opaque;
  884. if (VMW_IS_MULTIREG_ADDR(addr, VMXNET3_REG_TXPROD,
  885. VMXNET3_DEVICE_MAX_TX_QUEUES, VMXNET3_REG_ALIGN)) {
  886. int tx_queue_idx =
  887. VMW_MULTIREG_IDX_BY_ADDR(addr, VMXNET3_REG_TXPROD,
  888. VMXNET3_REG_ALIGN);
  889. assert(tx_queue_idx <= s->txq_num);
  890. vmxnet3_process_tx_queue(s, tx_queue_idx);
  891. return;
  892. }
  893. if (VMW_IS_MULTIREG_ADDR(addr, VMXNET3_REG_IMR,
  894. VMXNET3_MAX_INTRS, VMXNET3_REG_ALIGN)) {
  895. int l = VMW_MULTIREG_IDX_BY_ADDR(addr, VMXNET3_REG_IMR,
  896. VMXNET3_REG_ALIGN);
  897. VMW_CBPRN("Interrupt mask for line %d written: 0x%" PRIx64, l, val);
  898. vmxnet3_on_interrupt_mask_changed(s, l, val);
  899. return;
  900. }
  901. if (VMW_IS_MULTIREG_ADDR(addr, VMXNET3_REG_RXPROD,
  902. VMXNET3_DEVICE_MAX_RX_QUEUES, VMXNET3_REG_ALIGN) ||
  903. VMW_IS_MULTIREG_ADDR(addr, VMXNET3_REG_RXPROD2,
  904. VMXNET3_DEVICE_MAX_RX_QUEUES, VMXNET3_REG_ALIGN)) {
  905. return;
  906. }
  907. VMW_WRPRN("BAR0 unknown write [%" PRIx64 "] = %" PRIx64 ", size %d",
  908. (uint64_t) addr, val, size);
  909. }
  910. static uint64_t
  911. vmxnet3_io_bar0_read(void *opaque, hwaddr addr, unsigned size)
  912. {
  913. if (VMW_IS_MULTIREG_ADDR(addr, VMXNET3_REG_IMR,
  914. VMXNET3_MAX_INTRS, VMXNET3_REG_ALIGN)) {
  915. g_assert_not_reached();
  916. }
  917. VMW_CBPRN("BAR0 unknown read [%" PRIx64 "], size %d", addr, size);
  918. return 0;
  919. }
  920. static void vmxnet3_reset_interrupt_states(VMXNET3State *s)
  921. {
  922. int i;
  923. for (i = 0; i < ARRAY_SIZE(s->interrupt_states); i++) {
  924. s->interrupt_states[i].is_asserted = false;
  925. s->interrupt_states[i].is_pending = false;
  926. s->interrupt_states[i].is_masked = true;
  927. }
  928. }
  929. static void vmxnet3_reset_mac(VMXNET3State *s)
  930. {
  931. memcpy(&s->conf.macaddr.a, &s->perm_mac.a, sizeof(s->perm_mac.a));
  932. VMW_CFPRN("MAC address set to: " VMXNET_MF, VMXNET_MA(s->conf.macaddr.a));
  933. }
  934. static void vmxnet3_deactivate_device(VMXNET3State *s)
  935. {
  936. VMW_CBPRN("Deactivating vmxnet3...");
  937. s->device_active = false;
  938. }
  939. static void vmxnet3_reset(VMXNET3State *s)
  940. {
  941. VMW_CBPRN("Resetting vmxnet3...");
  942. vmxnet3_deactivate_device(s);
  943. vmxnet3_reset_interrupt_states(s);
  944. vmxnet_tx_pkt_reset(s->tx_pkt);
  945. s->drv_shmem = 0;
  946. s->tx_sop = true;
  947. s->skip_current_tx_pkt = false;
  948. }
  949. static void vmxnet3_update_rx_mode(VMXNET3State *s)
  950. {
  951. s->rx_mode = VMXNET3_READ_DRV_SHARED32(s->drv_shmem,
  952. devRead.rxFilterConf.rxMode);
  953. VMW_CFPRN("RX mode: 0x%08X", s->rx_mode);
  954. }
  955. static void vmxnet3_update_vlan_filters(VMXNET3State *s)
  956. {
  957. int i;
  958. /* Copy configuration from shared memory */
  959. VMXNET3_READ_DRV_SHARED(s->drv_shmem,
  960. devRead.rxFilterConf.vfTable,
  961. s->vlan_table,
  962. sizeof(s->vlan_table));
  963. /* Invert byte order when needed */
  964. for (i = 0; i < ARRAY_SIZE(s->vlan_table); i++) {
  965. s->vlan_table[i] = le32_to_cpu(s->vlan_table[i]);
  966. }
  967. /* Dump configuration for debugging purposes */
  968. VMW_CFPRN("Configured VLANs:");
  969. for (i = 0; i < sizeof(s->vlan_table) * 8; i++) {
  970. if (VMXNET3_VFTABLE_ENTRY_IS_SET(s->vlan_table, i)) {
  971. VMW_CFPRN("\tVLAN %d is present", i);
  972. }
  973. }
  974. }
  975. static void vmxnet3_update_mcast_filters(VMXNET3State *s)
  976. {
  977. uint16_t list_bytes =
  978. VMXNET3_READ_DRV_SHARED16(s->drv_shmem,
  979. devRead.rxFilterConf.mfTableLen);
  980. s->mcast_list_len = list_bytes / sizeof(s->mcast_list[0]);
  981. s->mcast_list = g_realloc(s->mcast_list, list_bytes);
  982. if (NULL == s->mcast_list) {
  983. if (0 == s->mcast_list_len) {
  984. VMW_CFPRN("Current multicast list is empty");
  985. } else {
  986. VMW_ERPRN("Failed to allocate multicast list of %d elements",
  987. s->mcast_list_len);
  988. }
  989. s->mcast_list_len = 0;
  990. } else {
  991. int i;
  992. hwaddr mcast_list_pa =
  993. VMXNET3_READ_DRV_SHARED64(s->drv_shmem,
  994. devRead.rxFilterConf.mfTablePA);
  995. cpu_physical_memory_read(mcast_list_pa, s->mcast_list, list_bytes);
  996. VMW_CFPRN("Current multicast list len is %d:", s->mcast_list_len);
  997. for (i = 0; i < s->mcast_list_len; i++) {
  998. VMW_CFPRN("\t" VMXNET_MF, VMXNET_MA(s->mcast_list[i].a));
  999. }
  1000. }
  1001. }
  1002. static void vmxnet3_setup_rx_filtering(VMXNET3State *s)
  1003. {
  1004. vmxnet3_update_rx_mode(s);
  1005. vmxnet3_update_vlan_filters(s);
  1006. vmxnet3_update_mcast_filters(s);
  1007. }
  1008. static uint32_t vmxnet3_get_interrupt_config(VMXNET3State *s)
  1009. {
  1010. uint32_t interrupt_mode = VMXNET3_IT_AUTO | (VMXNET3_IMM_AUTO << 2);
  1011. VMW_CFPRN("Interrupt config is 0x%X", interrupt_mode);
  1012. return interrupt_mode;
  1013. }
  1014. static void vmxnet3_fill_stats(VMXNET3State *s)
  1015. {
  1016. int i;
  1017. for (i = 0; i < s->txq_num; i++) {
  1018. cpu_physical_memory_write(s->txq_descr[i].tx_stats_pa,
  1019. &s->txq_descr[i].txq_stats,
  1020. sizeof(s->txq_descr[i].txq_stats));
  1021. }
  1022. for (i = 0; i < s->rxq_num; i++) {
  1023. cpu_physical_memory_write(s->rxq_descr[i].rx_stats_pa,
  1024. &s->rxq_descr[i].rxq_stats,
  1025. sizeof(s->rxq_descr[i].rxq_stats));
  1026. }
  1027. }
  1028. static void vmxnet3_adjust_by_guest_type(VMXNET3State *s)
  1029. {
  1030. struct Vmxnet3_GOSInfo gos;
  1031. VMXNET3_READ_DRV_SHARED(s->drv_shmem, devRead.misc.driverInfo.gos,
  1032. &gos, sizeof(gos));
  1033. s->rx_packets_compound =
  1034. (gos.gosType == VMXNET3_GOS_TYPE_WIN) ? false : true;
  1035. VMW_CFPRN("Guest type specifics: RXCOMPOUND: %d", s->rx_packets_compound);
  1036. }
  1037. static void
  1038. vmxnet3_dump_conf_descr(const char *name,
  1039. struct Vmxnet3_VariableLenConfDesc *pm_descr)
  1040. {
  1041. VMW_CFPRN("%s descriptor dump: Version %u, Length %u",
  1042. name, pm_descr->confVer, pm_descr->confLen);
  1043. };
  1044. static void vmxnet3_update_pm_state(VMXNET3State *s)
  1045. {
  1046. struct Vmxnet3_VariableLenConfDesc pm_descr;
  1047. pm_descr.confLen =
  1048. VMXNET3_READ_DRV_SHARED32(s->drv_shmem, devRead.pmConfDesc.confLen);
  1049. pm_descr.confVer =
  1050. VMXNET3_READ_DRV_SHARED32(s->drv_shmem, devRead.pmConfDesc.confVer);
  1051. pm_descr.confPA =
  1052. VMXNET3_READ_DRV_SHARED64(s->drv_shmem, devRead.pmConfDesc.confPA);
  1053. vmxnet3_dump_conf_descr("PM State", &pm_descr);
  1054. }
  1055. static void vmxnet3_update_features(VMXNET3State *s)
  1056. {
  1057. uint32_t guest_features;
  1058. int rxcso_supported;
  1059. guest_features = VMXNET3_READ_DRV_SHARED32(s->drv_shmem,
  1060. devRead.misc.uptFeatures);
  1061. rxcso_supported = VMXNET_FLAG_IS_SET(guest_features, UPT1_F_RXCSUM);
  1062. s->rx_vlan_stripping = VMXNET_FLAG_IS_SET(guest_features, UPT1_F_RXVLAN);
  1063. s->lro_supported = VMXNET_FLAG_IS_SET(guest_features, UPT1_F_LRO);
  1064. VMW_CFPRN("Features configuration: LRO: %d, RXCSUM: %d, VLANSTRIP: %d",
  1065. s->lro_supported, rxcso_supported,
  1066. s->rx_vlan_stripping);
  1067. if (s->peer_has_vhdr) {
  1068. qemu_set_offload(qemu_get_queue(s->nic)->peer,
  1069. rxcso_supported,
  1070. s->lro_supported,
  1071. s->lro_supported,
  1072. 0,
  1073. 0);
  1074. }
  1075. }
  1076. static bool vmxnet3_verify_intx(VMXNET3State *s, int intx)
  1077. {
  1078. return s->msix_used || s->msi_used || (intx ==
  1079. (pci_get_byte(s->parent_obj.config + PCI_INTERRUPT_PIN) - 1));
  1080. }
  1081. static void vmxnet3_validate_interrupt_idx(bool is_msix, int idx)
  1082. {
  1083. int max_ints = is_msix ? VMXNET3_MAX_INTRS : VMXNET3_MAX_NMSIX_INTRS;
  1084. if (idx >= max_ints) {
  1085. hw_error("Bad interrupt index: %d\n", idx);
  1086. }
  1087. }
  1088. static void vmxnet3_validate_interrupts(VMXNET3State *s)
  1089. {
  1090. int i;
  1091. VMW_CFPRN("Verifying event interrupt index (%d)", s->event_int_idx);
  1092. vmxnet3_validate_interrupt_idx(s->msix_used, s->event_int_idx);
  1093. for (i = 0; i < s->txq_num; i++) {
  1094. int idx = s->txq_descr[i].intr_idx;
  1095. VMW_CFPRN("Verifying TX queue %d interrupt index (%d)", i, idx);
  1096. vmxnet3_validate_interrupt_idx(s->msix_used, idx);
  1097. }
  1098. for (i = 0; i < s->rxq_num; i++) {
  1099. int idx = s->rxq_descr[i].intr_idx;
  1100. VMW_CFPRN("Verifying RX queue %d interrupt index (%d)", i, idx);
  1101. vmxnet3_validate_interrupt_idx(s->msix_used, idx);
  1102. }
  1103. }
  1104. static void vmxnet3_validate_queues(VMXNET3State *s)
  1105. {
  1106. /*
  1107. * txq_num and rxq_num are total number of queues
  1108. * configured by guest. These numbers must not
  1109. * exceed corresponding maximal values.
  1110. */
  1111. if (s->txq_num > VMXNET3_DEVICE_MAX_TX_QUEUES) {
  1112. hw_error("Bad TX queues number: %d\n", s->txq_num);
  1113. }
  1114. if (s->rxq_num > VMXNET3_DEVICE_MAX_RX_QUEUES) {
  1115. hw_error("Bad RX queues number: %d\n", s->rxq_num);
  1116. }
  1117. }
  1118. static void vmxnet3_activate_device(VMXNET3State *s)
  1119. {
  1120. int i;
  1121. static const uint32_t VMXNET3_DEF_TX_THRESHOLD = 1;
  1122. hwaddr qdescr_table_pa;
  1123. uint64_t pa;
  1124. uint32_t size;
  1125. /* Verify configuration consistency */
  1126. if (!vmxnet3_verify_driver_magic(s->drv_shmem)) {
  1127. VMW_ERPRN("Device configuration received from driver is invalid");
  1128. return;
  1129. }
  1130. vmxnet3_adjust_by_guest_type(s);
  1131. vmxnet3_update_features(s);
  1132. vmxnet3_update_pm_state(s);
  1133. vmxnet3_setup_rx_filtering(s);
  1134. /* Cache fields from shared memory */
  1135. s->mtu = VMXNET3_READ_DRV_SHARED32(s->drv_shmem, devRead.misc.mtu);
  1136. VMW_CFPRN("MTU is %u", s->mtu);
  1137. s->max_rx_frags =
  1138. VMXNET3_READ_DRV_SHARED16(s->drv_shmem, devRead.misc.maxNumRxSG);
  1139. if (s->max_rx_frags == 0) {
  1140. s->max_rx_frags = 1;
  1141. }
  1142. VMW_CFPRN("Max RX fragments is %u", s->max_rx_frags);
  1143. s->event_int_idx =
  1144. VMXNET3_READ_DRV_SHARED8(s->drv_shmem, devRead.intrConf.eventIntrIdx);
  1145. assert(vmxnet3_verify_intx(s, s->event_int_idx));
  1146. VMW_CFPRN("Events interrupt line is %u", s->event_int_idx);
  1147. s->auto_int_masking =
  1148. VMXNET3_READ_DRV_SHARED8(s->drv_shmem, devRead.intrConf.autoMask);
  1149. VMW_CFPRN("Automatic interrupt masking is %d", (int)s->auto_int_masking);
  1150. s->txq_num =
  1151. VMXNET3_READ_DRV_SHARED8(s->drv_shmem, devRead.misc.numTxQueues);
  1152. s->rxq_num =
  1153. VMXNET3_READ_DRV_SHARED8(s->drv_shmem, devRead.misc.numRxQueues);
  1154. VMW_CFPRN("Number of TX/RX queues %u/%u", s->txq_num, s->rxq_num);
  1155. vmxnet3_validate_queues(s);
  1156. qdescr_table_pa =
  1157. VMXNET3_READ_DRV_SHARED64(s->drv_shmem, devRead.misc.queueDescPA);
  1158. VMW_CFPRN("TX queues descriptors table is at 0x%" PRIx64, qdescr_table_pa);
  1159. /*
  1160. * Worst-case scenario is a packet that holds all TX rings space so
  1161. * we calculate total size of all TX rings for max TX fragments number
  1162. */
  1163. s->max_tx_frags = 0;
  1164. /* TX queues */
  1165. for (i = 0; i < s->txq_num; i++) {
  1166. hwaddr qdescr_pa =
  1167. qdescr_table_pa + i * sizeof(struct Vmxnet3_TxQueueDesc);
  1168. /* Read interrupt number for this TX queue */
  1169. s->txq_descr[i].intr_idx =
  1170. VMXNET3_READ_TX_QUEUE_DESCR8(qdescr_pa, conf.intrIdx);
  1171. assert(vmxnet3_verify_intx(s, s->txq_descr[i].intr_idx));
  1172. VMW_CFPRN("TX Queue %d interrupt: %d", i, s->txq_descr[i].intr_idx);
  1173. /* Read rings memory locations for TX queues */
  1174. pa = VMXNET3_READ_TX_QUEUE_DESCR64(qdescr_pa, conf.txRingBasePA);
  1175. size = VMXNET3_READ_TX_QUEUE_DESCR32(qdescr_pa, conf.txRingSize);
  1176. vmxnet3_ring_init(&s->txq_descr[i].tx_ring, pa, size,
  1177. sizeof(struct Vmxnet3_TxDesc), false);
  1178. VMXNET3_RING_DUMP(VMW_CFPRN, "TX", i, &s->txq_descr[i].tx_ring);
  1179. s->max_tx_frags += size;
  1180. /* TXC ring */
  1181. pa = VMXNET3_READ_TX_QUEUE_DESCR64(qdescr_pa, conf.compRingBasePA);
  1182. size = VMXNET3_READ_TX_QUEUE_DESCR32(qdescr_pa, conf.compRingSize);
  1183. vmxnet3_ring_init(&s->txq_descr[i].comp_ring, pa, size,
  1184. sizeof(struct Vmxnet3_TxCompDesc), true);
  1185. VMXNET3_RING_DUMP(VMW_CFPRN, "TXC", i, &s->txq_descr[i].comp_ring);
  1186. s->txq_descr[i].tx_stats_pa =
  1187. qdescr_pa + offsetof(struct Vmxnet3_TxQueueDesc, stats);
  1188. memset(&s->txq_descr[i].txq_stats, 0,
  1189. sizeof(s->txq_descr[i].txq_stats));
  1190. /* Fill device-managed parameters for queues */
  1191. VMXNET3_WRITE_TX_QUEUE_DESCR32(qdescr_pa,
  1192. ctrl.txThreshold,
  1193. VMXNET3_DEF_TX_THRESHOLD);
  1194. }
  1195. /* Preallocate TX packet wrapper */
  1196. VMW_CFPRN("Max TX fragments is %u", s->max_tx_frags);
  1197. vmxnet_tx_pkt_init(&s->tx_pkt, s->max_tx_frags, s->peer_has_vhdr);
  1198. vmxnet_rx_pkt_init(&s->rx_pkt, s->peer_has_vhdr);
  1199. /* Read rings memory locations for RX queues */
  1200. for (i = 0; i < s->rxq_num; i++) {
  1201. int j;
  1202. hwaddr qd_pa =
  1203. qdescr_table_pa + s->txq_num * sizeof(struct Vmxnet3_TxQueueDesc) +
  1204. i * sizeof(struct Vmxnet3_RxQueueDesc);
  1205. /* Read interrupt number for this RX queue */
  1206. s->rxq_descr[i].intr_idx =
  1207. VMXNET3_READ_TX_QUEUE_DESCR8(qd_pa, conf.intrIdx);
  1208. assert(vmxnet3_verify_intx(s, s->rxq_descr[i].intr_idx));
  1209. VMW_CFPRN("RX Queue %d interrupt: %d", i, s->rxq_descr[i].intr_idx);
  1210. /* Read rings memory locations */
  1211. for (j = 0; j < VMXNET3_RX_RINGS_PER_QUEUE; j++) {
  1212. /* RX rings */
  1213. pa = VMXNET3_READ_RX_QUEUE_DESCR64(qd_pa, conf.rxRingBasePA[j]);
  1214. size = VMXNET3_READ_RX_QUEUE_DESCR32(qd_pa, conf.rxRingSize[j]);
  1215. vmxnet3_ring_init(&s->rxq_descr[i].rx_ring[j], pa, size,
  1216. sizeof(struct Vmxnet3_RxDesc), false);
  1217. VMW_CFPRN("RX queue %d:%d: Base: %" PRIx64 ", Size: %d",
  1218. i, j, pa, size);
  1219. }
  1220. /* RXC ring */
  1221. pa = VMXNET3_READ_RX_QUEUE_DESCR64(qd_pa, conf.compRingBasePA);
  1222. size = VMXNET3_READ_RX_QUEUE_DESCR32(qd_pa, conf.compRingSize);
  1223. vmxnet3_ring_init(&s->rxq_descr[i].comp_ring, pa, size,
  1224. sizeof(struct Vmxnet3_RxCompDesc), true);
  1225. VMW_CFPRN("RXC queue %d: Base: %" PRIx64 ", Size: %d", i, pa, size);
  1226. s->rxq_descr[i].rx_stats_pa =
  1227. qd_pa + offsetof(struct Vmxnet3_RxQueueDesc, stats);
  1228. memset(&s->rxq_descr[i].rxq_stats, 0,
  1229. sizeof(s->rxq_descr[i].rxq_stats));
  1230. }
  1231. vmxnet3_validate_interrupts(s);
  1232. /* Make sure everything is in place before device activation */
  1233. smp_wmb();
  1234. vmxnet3_reset_mac(s);
  1235. s->device_active = true;
  1236. }
  1237. static void vmxnet3_handle_command(VMXNET3State *s, uint64_t cmd)
  1238. {
  1239. s->last_command = cmd;
  1240. switch (cmd) {
  1241. case VMXNET3_CMD_GET_PERM_MAC_HI:
  1242. VMW_CBPRN("Set: Get upper part of permanent MAC");
  1243. break;
  1244. case VMXNET3_CMD_GET_PERM_MAC_LO:
  1245. VMW_CBPRN("Set: Get lower part of permanent MAC");
  1246. break;
  1247. case VMXNET3_CMD_GET_STATS:
  1248. VMW_CBPRN("Set: Get device statistics");
  1249. vmxnet3_fill_stats(s);
  1250. break;
  1251. case VMXNET3_CMD_ACTIVATE_DEV:
  1252. VMW_CBPRN("Set: Activating vmxnet3 device");
  1253. vmxnet3_activate_device(s);
  1254. break;
  1255. case VMXNET3_CMD_UPDATE_RX_MODE:
  1256. VMW_CBPRN("Set: Update rx mode");
  1257. vmxnet3_update_rx_mode(s);
  1258. break;
  1259. case VMXNET3_CMD_UPDATE_VLAN_FILTERS:
  1260. VMW_CBPRN("Set: Update VLAN filters");
  1261. vmxnet3_update_vlan_filters(s);
  1262. break;
  1263. case VMXNET3_CMD_UPDATE_MAC_FILTERS:
  1264. VMW_CBPRN("Set: Update MAC filters");
  1265. vmxnet3_update_mcast_filters(s);
  1266. break;
  1267. case VMXNET3_CMD_UPDATE_FEATURE:
  1268. VMW_CBPRN("Set: Update features");
  1269. vmxnet3_update_features(s);
  1270. break;
  1271. case VMXNET3_CMD_UPDATE_PMCFG:
  1272. VMW_CBPRN("Set: Update power management config");
  1273. vmxnet3_update_pm_state(s);
  1274. break;
  1275. case VMXNET3_CMD_GET_LINK:
  1276. VMW_CBPRN("Set: Get link");
  1277. break;
  1278. case VMXNET3_CMD_RESET_DEV:
  1279. VMW_CBPRN("Set: Reset device");
  1280. vmxnet3_reset(s);
  1281. break;
  1282. case VMXNET3_CMD_QUIESCE_DEV:
  1283. VMW_CBPRN("Set: VMXNET3_CMD_QUIESCE_DEV - pause the device");
  1284. vmxnet3_deactivate_device(s);
  1285. break;
  1286. case VMXNET3_CMD_GET_CONF_INTR:
  1287. VMW_CBPRN("Set: VMXNET3_CMD_GET_CONF_INTR - interrupt configuration");
  1288. break;
  1289. default:
  1290. VMW_CBPRN("Received unknown command: %" PRIx64, cmd);
  1291. break;
  1292. }
  1293. }
  1294. static uint64_t vmxnet3_get_command_status(VMXNET3State *s)
  1295. {
  1296. uint64_t ret;
  1297. switch (s->last_command) {
  1298. case VMXNET3_CMD_ACTIVATE_DEV:
  1299. ret = (s->device_active) ? 0 : -1;
  1300. VMW_CFPRN("Device active: %" PRIx64, ret);
  1301. break;
  1302. case VMXNET3_CMD_RESET_DEV:
  1303. case VMXNET3_CMD_QUIESCE_DEV:
  1304. case VMXNET3_CMD_GET_QUEUE_STATUS:
  1305. ret = 0;
  1306. break;
  1307. case VMXNET3_CMD_GET_LINK:
  1308. ret = s->link_status_and_speed;
  1309. VMW_CFPRN("Link and speed: %" PRIx64, ret);
  1310. break;
  1311. case VMXNET3_CMD_GET_PERM_MAC_LO:
  1312. ret = vmxnet3_get_mac_low(&s->perm_mac);
  1313. break;
  1314. case VMXNET3_CMD_GET_PERM_MAC_HI:
  1315. ret = vmxnet3_get_mac_high(&s->perm_mac);
  1316. break;
  1317. case VMXNET3_CMD_GET_CONF_INTR:
  1318. ret = vmxnet3_get_interrupt_config(s);
  1319. break;
  1320. default:
  1321. VMW_WRPRN("Received request for unknown command: %x", s->last_command);
  1322. ret = -1;
  1323. break;
  1324. }
  1325. return ret;
  1326. }
  1327. static void vmxnet3_set_events(VMXNET3State *s, uint32_t val)
  1328. {
  1329. uint32_t events;
  1330. VMW_CBPRN("Setting events: 0x%x", val);
  1331. events = VMXNET3_READ_DRV_SHARED32(s->drv_shmem, ecr) | val;
  1332. VMXNET3_WRITE_DRV_SHARED32(s->drv_shmem, ecr, events);
  1333. }
  1334. static void vmxnet3_ack_events(VMXNET3State *s, uint32_t val)
  1335. {
  1336. uint32_t events;
  1337. VMW_CBPRN("Clearing events: 0x%x", val);
  1338. events = VMXNET3_READ_DRV_SHARED32(s->drv_shmem, ecr) & ~val;
  1339. VMXNET3_WRITE_DRV_SHARED32(s->drv_shmem, ecr, events);
  1340. }
  1341. static void
  1342. vmxnet3_io_bar1_write(void *opaque,
  1343. hwaddr addr,
  1344. uint64_t val,
  1345. unsigned size)
  1346. {
  1347. VMXNET3State *s = opaque;
  1348. switch (addr) {
  1349. /* Vmxnet3 Revision Report Selection */
  1350. case VMXNET3_REG_VRRS:
  1351. VMW_CBPRN("Write BAR1 [VMXNET3_REG_VRRS] = %" PRIx64 ", size %d",
  1352. val, size);
  1353. break;
  1354. /* UPT Version Report Selection */
  1355. case VMXNET3_REG_UVRS:
  1356. VMW_CBPRN("Write BAR1 [VMXNET3_REG_UVRS] = %" PRIx64 ", size %d",
  1357. val, size);
  1358. break;
  1359. /* Driver Shared Address Low */
  1360. case VMXNET3_REG_DSAL:
  1361. VMW_CBPRN("Write BAR1 [VMXNET3_REG_DSAL] = %" PRIx64 ", size %d",
  1362. val, size);
  1363. /*
  1364. * Guest driver will first write the low part of the shared
  1365. * memory address. We save it to temp variable and set the
  1366. * shared address only after we get the high part
  1367. */
  1368. if (0 == val) {
  1369. s->device_active = false;
  1370. }
  1371. s->temp_shared_guest_driver_memory = val;
  1372. s->drv_shmem = 0;
  1373. break;
  1374. /* Driver Shared Address High */
  1375. case VMXNET3_REG_DSAH:
  1376. VMW_CBPRN("Write BAR1 [VMXNET3_REG_DSAH] = %" PRIx64 ", size %d",
  1377. val, size);
  1378. /*
  1379. * Set the shared memory between guest driver and device.
  1380. * We already should have low address part.
  1381. */
  1382. s->drv_shmem = s->temp_shared_guest_driver_memory | (val << 32);
  1383. break;
  1384. /* Command */
  1385. case VMXNET3_REG_CMD:
  1386. VMW_CBPRN("Write BAR1 [VMXNET3_REG_CMD] = %" PRIx64 ", size %d",
  1387. val, size);
  1388. vmxnet3_handle_command(s, val);
  1389. break;
  1390. /* MAC Address Low */
  1391. case VMXNET3_REG_MACL:
  1392. VMW_CBPRN("Write BAR1 [VMXNET3_REG_MACL] = %" PRIx64 ", size %d",
  1393. val, size);
  1394. s->temp_mac = val;
  1395. break;
  1396. /* MAC Address High */
  1397. case VMXNET3_REG_MACH:
  1398. VMW_CBPRN("Write BAR1 [VMXNET3_REG_MACH] = %" PRIx64 ", size %d",
  1399. val, size);
  1400. vmxnet3_set_variable_mac(s, val, s->temp_mac);
  1401. break;
  1402. /* Interrupt Cause Register */
  1403. case VMXNET3_REG_ICR:
  1404. VMW_CBPRN("Write BAR1 [VMXNET3_REG_ICR] = %" PRIx64 ", size %d",
  1405. val, size);
  1406. g_assert_not_reached();
  1407. break;
  1408. /* Event Cause Register */
  1409. case VMXNET3_REG_ECR:
  1410. VMW_CBPRN("Write BAR1 [VMXNET3_REG_ECR] = %" PRIx64 ", size %d",
  1411. val, size);
  1412. vmxnet3_ack_events(s, val);
  1413. break;
  1414. default:
  1415. VMW_CBPRN("Unknown Write to BAR1 [%" PRIx64 "] = %" PRIx64 ", size %d",
  1416. addr, val, size);
  1417. break;
  1418. }
  1419. }
  1420. static uint64_t
  1421. vmxnet3_io_bar1_read(void *opaque, hwaddr addr, unsigned size)
  1422. {
  1423. VMXNET3State *s = opaque;
  1424. uint64_t ret = 0;
  1425. switch (addr) {
  1426. /* Vmxnet3 Revision Report Selection */
  1427. case VMXNET3_REG_VRRS:
  1428. VMW_CBPRN("Read BAR1 [VMXNET3_REG_VRRS], size %d", size);
  1429. ret = VMXNET3_DEVICE_REVISION;
  1430. break;
  1431. /* UPT Version Report Selection */
  1432. case VMXNET3_REG_UVRS:
  1433. VMW_CBPRN("Read BAR1 [VMXNET3_REG_UVRS], size %d", size);
  1434. ret = VMXNET3_DEVICE_VERSION;
  1435. break;
  1436. /* Command */
  1437. case VMXNET3_REG_CMD:
  1438. VMW_CBPRN("Read BAR1 [VMXNET3_REG_CMD], size %d", size);
  1439. ret = vmxnet3_get_command_status(s);
  1440. break;
  1441. /* MAC Address Low */
  1442. case VMXNET3_REG_MACL:
  1443. VMW_CBPRN("Read BAR1 [VMXNET3_REG_MACL], size %d", size);
  1444. ret = vmxnet3_get_mac_low(&s->conf.macaddr);
  1445. break;
  1446. /* MAC Address High */
  1447. case VMXNET3_REG_MACH:
  1448. VMW_CBPRN("Read BAR1 [VMXNET3_REG_MACH], size %d", size);
  1449. ret = vmxnet3_get_mac_high(&s->conf.macaddr);
  1450. break;
  1451. /*
  1452. * Interrupt Cause Register
  1453. * Used for legacy interrupts only so interrupt index always 0
  1454. */
  1455. case VMXNET3_REG_ICR:
  1456. VMW_CBPRN("Read BAR1 [VMXNET3_REG_ICR], size %d", size);
  1457. if (vmxnet3_interrupt_asserted(s, 0)) {
  1458. vmxnet3_clear_interrupt(s, 0);
  1459. ret = true;
  1460. } else {
  1461. ret = false;
  1462. }
  1463. break;
  1464. default:
  1465. VMW_CBPRN("Unknow read BAR1[%" PRIx64 "], %d bytes", addr, size);
  1466. break;
  1467. }
  1468. return ret;
  1469. }
  1470. static int
  1471. vmxnet3_can_receive(NetClientState *nc)
  1472. {
  1473. VMXNET3State *s = qemu_get_nic_opaque(nc);
  1474. return s->device_active &&
  1475. VMXNET_FLAG_IS_SET(s->link_status_and_speed, VMXNET3_LINK_STATUS_UP);
  1476. }
  1477. static inline bool
  1478. vmxnet3_is_registered_vlan(VMXNET3State *s, const void *data)
  1479. {
  1480. uint16_t vlan_tag = eth_get_pkt_tci(data) & VLAN_VID_MASK;
  1481. if (IS_SPECIAL_VLAN_ID(vlan_tag)) {
  1482. return true;
  1483. }
  1484. return VMXNET3_VFTABLE_ENTRY_IS_SET(s->vlan_table, vlan_tag);
  1485. }
  1486. static bool
  1487. vmxnet3_is_allowed_mcast_group(VMXNET3State *s, const uint8_t *group_mac)
  1488. {
  1489. int i;
  1490. for (i = 0; i < s->mcast_list_len; i++) {
  1491. if (!memcmp(group_mac, s->mcast_list[i].a, sizeof(s->mcast_list[i]))) {
  1492. return true;
  1493. }
  1494. }
  1495. return false;
  1496. }
  1497. static bool
  1498. vmxnet3_rx_filter_may_indicate(VMXNET3State *s, const void *data,
  1499. size_t size)
  1500. {
  1501. struct eth_header *ehdr = PKT_GET_ETH_HDR(data);
  1502. if (VMXNET_FLAG_IS_SET(s->rx_mode, VMXNET3_RXM_PROMISC)) {
  1503. return true;
  1504. }
  1505. if (!vmxnet3_is_registered_vlan(s, data)) {
  1506. return false;
  1507. }
  1508. switch (vmxnet_rx_pkt_get_packet_type(s->rx_pkt)) {
  1509. case ETH_PKT_UCAST:
  1510. if (!VMXNET_FLAG_IS_SET(s->rx_mode, VMXNET3_RXM_UCAST)) {
  1511. return false;
  1512. }
  1513. if (memcmp(s->conf.macaddr.a, ehdr->h_dest, ETH_ALEN)) {
  1514. return false;
  1515. }
  1516. break;
  1517. case ETH_PKT_BCAST:
  1518. if (!VMXNET_FLAG_IS_SET(s->rx_mode, VMXNET3_RXM_BCAST)) {
  1519. return false;
  1520. }
  1521. break;
  1522. case ETH_PKT_MCAST:
  1523. if (VMXNET_FLAG_IS_SET(s->rx_mode, VMXNET3_RXM_ALL_MULTI)) {
  1524. return true;
  1525. }
  1526. if (!VMXNET_FLAG_IS_SET(s->rx_mode, VMXNET3_RXM_MCAST)) {
  1527. return false;
  1528. }
  1529. if (!vmxnet3_is_allowed_mcast_group(s, ehdr->h_dest)) {
  1530. return false;
  1531. }
  1532. break;
  1533. default:
  1534. g_assert_not_reached();
  1535. }
  1536. return true;
  1537. }
  1538. static ssize_t
  1539. vmxnet3_receive(NetClientState *nc, const uint8_t *buf, size_t size)
  1540. {
  1541. VMXNET3State *s = qemu_get_nic_opaque(nc);
  1542. size_t bytes_indicated;
  1543. if (!vmxnet3_can_receive(nc)) {
  1544. VMW_PKPRN("Cannot receive now");
  1545. return -1;
  1546. }
  1547. if (s->peer_has_vhdr) {
  1548. vmxnet_rx_pkt_set_vhdr(s->rx_pkt, (struct virtio_net_hdr *)buf);
  1549. buf += sizeof(struct virtio_net_hdr);
  1550. size -= sizeof(struct virtio_net_hdr);
  1551. }
  1552. vmxnet_rx_pkt_set_packet_type(s->rx_pkt,
  1553. get_eth_packet_type(PKT_GET_ETH_HDR(buf)));
  1554. if (vmxnet3_rx_filter_may_indicate(s, buf, size)) {
  1555. vmxnet_rx_pkt_attach_data(s->rx_pkt, buf, size, s->rx_vlan_stripping);
  1556. bytes_indicated = vmxnet3_indicate_packet(s) ? size : -1;
  1557. if (bytes_indicated < size) {
  1558. VMW_PKPRN("RX: %lu of %lu bytes indicated", bytes_indicated, size);
  1559. }
  1560. } else {
  1561. VMW_PKPRN("Packet dropped by RX filter");
  1562. bytes_indicated = size;
  1563. }
  1564. assert(size > 0);
  1565. assert(bytes_indicated != 0);
  1566. return bytes_indicated;
  1567. }
  1568. static void vmxnet3_cleanup(NetClientState *nc)
  1569. {
  1570. VMXNET3State *s = qemu_get_nic_opaque(nc);
  1571. s->nic = NULL;
  1572. }
  1573. static void vmxnet3_set_link_status(NetClientState *nc)
  1574. {
  1575. VMXNET3State *s = qemu_get_nic_opaque(nc);
  1576. if (nc->link_down) {
  1577. s->link_status_and_speed &= ~VMXNET3_LINK_STATUS_UP;
  1578. } else {
  1579. s->link_status_and_speed |= VMXNET3_LINK_STATUS_UP;
  1580. }
  1581. vmxnet3_set_events(s, VMXNET3_ECR_LINK);
  1582. vmxnet3_trigger_interrupt(s, s->event_int_idx);
  1583. }
  1584. static NetClientInfo net_vmxnet3_info = {
  1585. .type = NET_CLIENT_OPTIONS_KIND_NIC,
  1586. .size = sizeof(NICState),
  1587. .can_receive = vmxnet3_can_receive,
  1588. .receive = vmxnet3_receive,
  1589. .cleanup = vmxnet3_cleanup,
  1590. .link_status_changed = vmxnet3_set_link_status,
  1591. };
  1592. static bool vmxnet3_peer_has_vnet_hdr(VMXNET3State *s)
  1593. {
  1594. NetClientState *nc = qemu_get_queue(s->nic);
  1595. if (qemu_has_vnet_hdr(nc->peer)) {
  1596. return true;
  1597. }
  1598. VMW_WRPRN("Peer has no virtio extension. Task offloads will be emulated.");
  1599. return false;
  1600. }
  1601. static void vmxnet3_net_uninit(VMXNET3State *s)
  1602. {
  1603. g_free(s->mcast_list);
  1604. vmxnet_tx_pkt_reset(s->tx_pkt);
  1605. vmxnet_tx_pkt_uninit(s->tx_pkt);
  1606. vmxnet_rx_pkt_uninit(s->rx_pkt);
  1607. qemu_del_nic(s->nic);
  1608. }
  1609. static void vmxnet3_net_init(VMXNET3State *s)
  1610. {
  1611. DeviceState *d = DEVICE(s);
  1612. VMW_CBPRN("vmxnet3_net_init called...");
  1613. qemu_macaddr_default_if_unset(&s->conf.macaddr);
  1614. /* Windows guest will query the address that was set on init */
  1615. memcpy(&s->perm_mac.a, &s->conf.macaddr.a, sizeof(s->perm_mac.a));
  1616. s->mcast_list = NULL;
  1617. s->mcast_list_len = 0;
  1618. s->link_status_and_speed = VMXNET3_LINK_SPEED | VMXNET3_LINK_STATUS_UP;
  1619. VMW_CFPRN("Permanent MAC: " MAC_FMT, MAC_ARG(s->perm_mac.a));
  1620. s->nic = qemu_new_nic(&net_vmxnet3_info, &s->conf,
  1621. object_get_typename(OBJECT(s)),
  1622. d->id, s);
  1623. s->peer_has_vhdr = vmxnet3_peer_has_vnet_hdr(s);
  1624. s->tx_sop = true;
  1625. s->skip_current_tx_pkt = false;
  1626. s->tx_pkt = NULL;
  1627. s->rx_pkt = NULL;
  1628. s->rx_vlan_stripping = false;
  1629. s->lro_supported = false;
  1630. if (s->peer_has_vhdr) {
  1631. qemu_set_vnet_hdr_len(qemu_get_queue(s->nic)->peer,
  1632. sizeof(struct virtio_net_hdr));
  1633. qemu_using_vnet_hdr(qemu_get_queue(s->nic)->peer, 1);
  1634. }
  1635. qemu_format_nic_info_str(qemu_get_queue(s->nic), s->conf.macaddr.a);
  1636. }
  1637. static void
  1638. vmxnet3_unuse_msix_vectors(VMXNET3State *s, int num_vectors)
  1639. {
  1640. PCIDevice *d = PCI_DEVICE(s);
  1641. int i;
  1642. for (i = 0; i < num_vectors; i++) {
  1643. msix_vector_unuse(d, i);
  1644. }
  1645. }
  1646. static bool
  1647. vmxnet3_use_msix_vectors(VMXNET3State *s, int num_vectors)
  1648. {
  1649. PCIDevice *d = PCI_DEVICE(s);
  1650. int i;
  1651. for (i = 0; i < num_vectors; i++) {
  1652. int res = msix_vector_use(d, i);
  1653. if (0 > res) {
  1654. VMW_WRPRN("Failed to use MSI-X vector %d, error %d", i, res);
  1655. vmxnet3_unuse_msix_vectors(s, i);
  1656. return false;
  1657. }
  1658. }
  1659. return true;
  1660. }
  1661. static bool
  1662. vmxnet3_init_msix(VMXNET3State *s)
  1663. {
  1664. PCIDevice *d = PCI_DEVICE(s);
  1665. int res = msix_init(d, VMXNET3_MAX_INTRS,
  1666. &s->msix_bar,
  1667. VMXNET3_MSIX_BAR_IDX, VMXNET3_OFF_MSIX_TABLE,
  1668. &s->msix_bar,
  1669. VMXNET3_MSIX_BAR_IDX, VMXNET3_OFF_MSIX_PBA,
  1670. 0);
  1671. if (0 > res) {
  1672. VMW_WRPRN("Failed to initialize MSI-X, error %d", res);
  1673. s->msix_used = false;
  1674. } else {
  1675. if (!vmxnet3_use_msix_vectors(s, VMXNET3_MAX_INTRS)) {
  1676. VMW_WRPRN("Failed to use MSI-X vectors, error %d", res);
  1677. msix_uninit(d, &s->msix_bar, &s->msix_bar);
  1678. s->msix_used = false;
  1679. } else {
  1680. s->msix_used = true;
  1681. }
  1682. }
  1683. return s->msix_used;
  1684. }
  1685. static void
  1686. vmxnet3_cleanup_msix(VMXNET3State *s)
  1687. {
  1688. PCIDevice *d = PCI_DEVICE(s);
  1689. if (s->msix_used) {
  1690. msix_vector_unuse(d, VMXNET3_MAX_INTRS);
  1691. msix_uninit(d, &s->msix_bar, &s->msix_bar);
  1692. }
  1693. }
  1694. #define VMXNET3_MSI_OFFSET (0x50)
  1695. #define VMXNET3_USE_64BIT (true)
  1696. #define VMXNET3_PER_VECTOR_MASK (false)
  1697. static bool
  1698. vmxnet3_init_msi(VMXNET3State *s)
  1699. {
  1700. PCIDevice *d = PCI_DEVICE(s);
  1701. int res;
  1702. res = msi_init(d, VMXNET3_MSI_OFFSET, VMXNET3_MAX_NMSIX_INTRS,
  1703. VMXNET3_USE_64BIT, VMXNET3_PER_VECTOR_MASK);
  1704. if (0 > res) {
  1705. VMW_WRPRN("Failed to initialize MSI, error %d", res);
  1706. s->msi_used = false;
  1707. } else {
  1708. s->msi_used = true;
  1709. }
  1710. return s->msi_used;
  1711. }
  1712. static void
  1713. vmxnet3_cleanup_msi(VMXNET3State *s)
  1714. {
  1715. PCIDevice *d = PCI_DEVICE(s);
  1716. if (s->msi_used) {
  1717. msi_uninit(d);
  1718. }
  1719. }
  1720. static void
  1721. vmxnet3_msix_save(QEMUFile *f, void *opaque)
  1722. {
  1723. PCIDevice *d = PCI_DEVICE(opaque);
  1724. msix_save(d, f);
  1725. }
  1726. static int
  1727. vmxnet3_msix_load(QEMUFile *f, void *opaque, int version_id)
  1728. {
  1729. PCIDevice *d = PCI_DEVICE(opaque);
  1730. msix_load(d, f);
  1731. return 0;
  1732. }
  1733. static const MemoryRegionOps b0_ops = {
  1734. .read = vmxnet3_io_bar0_read,
  1735. .write = vmxnet3_io_bar0_write,
  1736. .endianness = DEVICE_LITTLE_ENDIAN,
  1737. .impl = {
  1738. .min_access_size = 4,
  1739. .max_access_size = 4,
  1740. },
  1741. };
  1742. static const MemoryRegionOps b1_ops = {
  1743. .read = vmxnet3_io_bar1_read,
  1744. .write = vmxnet3_io_bar1_write,
  1745. .endianness = DEVICE_LITTLE_ENDIAN,
  1746. .impl = {
  1747. .min_access_size = 4,
  1748. .max_access_size = 4,
  1749. },
  1750. };
  1751. static int vmxnet3_pci_init(PCIDevice *pci_dev)
  1752. {
  1753. DeviceState *dev = DEVICE(pci_dev);
  1754. VMXNET3State *s = VMXNET3(pci_dev);
  1755. VMW_CBPRN("Starting init...");
  1756. memory_region_init_io(&s->bar0, OBJECT(s), &b0_ops, s,
  1757. "vmxnet3-b0", VMXNET3_PT_REG_SIZE);
  1758. pci_register_bar(pci_dev, VMXNET3_BAR0_IDX,
  1759. PCI_BASE_ADDRESS_SPACE_MEMORY, &s->bar0);
  1760. memory_region_init_io(&s->bar1, OBJECT(s), &b1_ops, s,
  1761. "vmxnet3-b1", VMXNET3_VD_REG_SIZE);
  1762. pci_register_bar(pci_dev, VMXNET3_BAR1_IDX,
  1763. PCI_BASE_ADDRESS_SPACE_MEMORY, &s->bar1);
  1764. memory_region_init(&s->msix_bar, OBJECT(s), "vmxnet3-msix-bar",
  1765. VMXNET3_MSIX_BAR_SIZE);
  1766. pci_register_bar(pci_dev, VMXNET3_MSIX_BAR_IDX,
  1767. PCI_BASE_ADDRESS_SPACE_MEMORY, &s->msix_bar);
  1768. vmxnet3_reset_interrupt_states(s);
  1769. /* Interrupt pin A */
  1770. pci_dev->config[PCI_INTERRUPT_PIN] = 0x01;
  1771. if (!vmxnet3_init_msix(s)) {
  1772. VMW_WRPRN("Failed to initialize MSI-X, configuration is inconsistent.");
  1773. }
  1774. if (!vmxnet3_init_msi(s)) {
  1775. VMW_WRPRN("Failed to initialize MSI, configuration is inconsistent.");
  1776. }
  1777. vmxnet3_net_init(s);
  1778. register_savevm(dev, "vmxnet3-msix", -1, 1,
  1779. vmxnet3_msix_save, vmxnet3_msix_load, s);
  1780. add_boot_device_path(s->conf.bootindex, dev, "/ethernet-phy@0");
  1781. return 0;
  1782. }
  1783. static void vmxnet3_pci_uninit(PCIDevice *pci_dev)
  1784. {
  1785. DeviceState *dev = DEVICE(pci_dev);
  1786. VMXNET3State *s = VMXNET3(pci_dev);
  1787. VMW_CBPRN("Starting uninit...");
  1788. unregister_savevm(dev, "vmxnet3-msix", s);
  1789. vmxnet3_net_uninit(s);
  1790. vmxnet3_cleanup_msix(s);
  1791. vmxnet3_cleanup_msi(s);
  1792. memory_region_destroy(&s->bar0);
  1793. memory_region_destroy(&s->bar1);
  1794. memory_region_destroy(&s->msix_bar);
  1795. }
  1796. static void vmxnet3_qdev_reset(DeviceState *dev)
  1797. {
  1798. PCIDevice *d = PCI_DEVICE(dev);
  1799. VMXNET3State *s = VMXNET3(d);
  1800. VMW_CBPRN("Starting QDEV reset...");
  1801. vmxnet3_reset(s);
  1802. }
  1803. static bool vmxnet3_mc_list_needed(void *opaque)
  1804. {
  1805. return true;
  1806. }
  1807. static int vmxnet3_mcast_list_pre_load(void *opaque)
  1808. {
  1809. VMXNET3State *s = opaque;
  1810. s->mcast_list = g_malloc(s->mcast_list_buff_size);
  1811. return 0;
  1812. }
  1813. static void vmxnet3_pre_save(void *opaque)
  1814. {
  1815. VMXNET3State *s = opaque;
  1816. s->mcast_list_buff_size = s->mcast_list_len * sizeof(MACAddr);
  1817. }
  1818. static const VMStateDescription vmxstate_vmxnet3_mcast_list = {
  1819. .name = "vmxnet3/mcast_list",
  1820. .version_id = 1,
  1821. .minimum_version_id = 1,
  1822. .minimum_version_id_old = 1,
  1823. .pre_load = vmxnet3_mcast_list_pre_load,
  1824. .fields = (VMStateField[]) {
  1825. VMSTATE_VBUFFER_UINT32(mcast_list, VMXNET3State, 0, NULL, 0,
  1826. mcast_list_buff_size),
  1827. VMSTATE_END_OF_LIST()
  1828. }
  1829. };
  1830. static void vmxnet3_get_ring_from_file(QEMUFile *f, Vmxnet3Ring *r)
  1831. {
  1832. r->pa = qemu_get_be64(f);
  1833. r->size = qemu_get_be32(f);
  1834. r->cell_size = qemu_get_be32(f);
  1835. r->next = qemu_get_be32(f);
  1836. r->gen = qemu_get_byte(f);
  1837. }
  1838. static void vmxnet3_put_ring_to_file(QEMUFile *f, Vmxnet3Ring *r)
  1839. {
  1840. qemu_put_be64(f, r->pa);
  1841. qemu_put_be32(f, r->size);
  1842. qemu_put_be32(f, r->cell_size);
  1843. qemu_put_be32(f, r->next);
  1844. qemu_put_byte(f, r->gen);
  1845. }
  1846. static void vmxnet3_get_tx_stats_from_file(QEMUFile *f,
  1847. struct UPT1_TxStats *tx_stat)
  1848. {
  1849. tx_stat->TSOPktsTxOK = qemu_get_be64(f);
  1850. tx_stat->TSOBytesTxOK = qemu_get_be64(f);
  1851. tx_stat->ucastPktsTxOK = qemu_get_be64(f);
  1852. tx_stat->ucastBytesTxOK = qemu_get_be64(f);
  1853. tx_stat->mcastPktsTxOK = qemu_get_be64(f);
  1854. tx_stat->mcastBytesTxOK = qemu_get_be64(f);
  1855. tx_stat->bcastPktsTxOK = qemu_get_be64(f);
  1856. tx_stat->bcastBytesTxOK = qemu_get_be64(f);
  1857. tx_stat->pktsTxError = qemu_get_be64(f);
  1858. tx_stat->pktsTxDiscard = qemu_get_be64(f);
  1859. }
  1860. static void vmxnet3_put_tx_stats_to_file(QEMUFile *f,
  1861. struct UPT1_TxStats *tx_stat)
  1862. {
  1863. qemu_put_be64(f, tx_stat->TSOPktsTxOK);
  1864. qemu_put_be64(f, tx_stat->TSOBytesTxOK);
  1865. qemu_put_be64(f, tx_stat->ucastPktsTxOK);
  1866. qemu_put_be64(f, tx_stat->ucastBytesTxOK);
  1867. qemu_put_be64(f, tx_stat->mcastPktsTxOK);
  1868. qemu_put_be64(f, tx_stat->mcastBytesTxOK);
  1869. qemu_put_be64(f, tx_stat->bcastPktsTxOK);
  1870. qemu_put_be64(f, tx_stat->bcastBytesTxOK);
  1871. qemu_put_be64(f, tx_stat->pktsTxError);
  1872. qemu_put_be64(f, tx_stat->pktsTxDiscard);
  1873. }
  1874. static int vmxnet3_get_txq_descr(QEMUFile *f, void *pv, size_t size)
  1875. {
  1876. Vmxnet3TxqDescr *r = pv;
  1877. vmxnet3_get_ring_from_file(f, &r->tx_ring);
  1878. vmxnet3_get_ring_from_file(f, &r->comp_ring);
  1879. r->intr_idx = qemu_get_byte(f);
  1880. r->tx_stats_pa = qemu_get_be64(f);
  1881. vmxnet3_get_tx_stats_from_file(f, &r->txq_stats);
  1882. return 0;
  1883. }
  1884. static void vmxnet3_put_txq_descr(QEMUFile *f, void *pv, size_t size)
  1885. {
  1886. Vmxnet3TxqDescr *r = pv;
  1887. vmxnet3_put_ring_to_file(f, &r->tx_ring);
  1888. vmxnet3_put_ring_to_file(f, &r->comp_ring);
  1889. qemu_put_byte(f, r->intr_idx);
  1890. qemu_put_be64(f, r->tx_stats_pa);
  1891. vmxnet3_put_tx_stats_to_file(f, &r->txq_stats);
  1892. }
  1893. const VMStateInfo txq_descr_info = {
  1894. .name = "txq_descr",
  1895. .get = vmxnet3_get_txq_descr,
  1896. .put = vmxnet3_put_txq_descr
  1897. };
  1898. static void vmxnet3_get_rx_stats_from_file(QEMUFile *f,
  1899. struct UPT1_RxStats *rx_stat)
  1900. {
  1901. rx_stat->LROPktsRxOK = qemu_get_be64(f);
  1902. rx_stat->LROBytesRxOK = qemu_get_be64(f);
  1903. rx_stat->ucastPktsRxOK = qemu_get_be64(f);
  1904. rx_stat->ucastBytesRxOK = qemu_get_be64(f);
  1905. rx_stat->mcastPktsRxOK = qemu_get_be64(f);
  1906. rx_stat->mcastBytesRxOK = qemu_get_be64(f);
  1907. rx_stat->bcastPktsRxOK = qemu_get_be64(f);
  1908. rx_stat->bcastBytesRxOK = qemu_get_be64(f);
  1909. rx_stat->pktsRxOutOfBuf = qemu_get_be64(f);
  1910. rx_stat->pktsRxError = qemu_get_be64(f);
  1911. }
  1912. static void vmxnet3_put_rx_stats_to_file(QEMUFile *f,
  1913. struct UPT1_RxStats *rx_stat)
  1914. {
  1915. qemu_put_be64(f, rx_stat->LROPktsRxOK);
  1916. qemu_put_be64(f, rx_stat->LROBytesRxOK);
  1917. qemu_put_be64(f, rx_stat->ucastPktsRxOK);
  1918. qemu_put_be64(f, rx_stat->ucastBytesRxOK);
  1919. qemu_put_be64(f, rx_stat->mcastPktsRxOK);
  1920. qemu_put_be64(f, rx_stat->mcastBytesRxOK);
  1921. qemu_put_be64(f, rx_stat->bcastPktsRxOK);
  1922. qemu_put_be64(f, rx_stat->bcastBytesRxOK);
  1923. qemu_put_be64(f, rx_stat->pktsRxOutOfBuf);
  1924. qemu_put_be64(f, rx_stat->pktsRxError);
  1925. }
  1926. static int vmxnet3_get_rxq_descr(QEMUFile *f, void *pv, size_t size)
  1927. {
  1928. Vmxnet3RxqDescr *r = pv;
  1929. int i;
  1930. for (i = 0; i < VMXNET3_RX_RINGS_PER_QUEUE; i++) {
  1931. vmxnet3_get_ring_from_file(f, &r->rx_ring[i]);
  1932. }
  1933. vmxnet3_get_ring_from_file(f, &r->comp_ring);
  1934. r->intr_idx = qemu_get_byte(f);
  1935. r->rx_stats_pa = qemu_get_be64(f);
  1936. vmxnet3_get_rx_stats_from_file(f, &r->rxq_stats);
  1937. return 0;
  1938. }
  1939. static void vmxnet3_put_rxq_descr(QEMUFile *f, void *pv, size_t size)
  1940. {
  1941. Vmxnet3RxqDescr *r = pv;
  1942. int i;
  1943. for (i = 0; i < VMXNET3_RX_RINGS_PER_QUEUE; i++) {
  1944. vmxnet3_put_ring_to_file(f, &r->rx_ring[i]);
  1945. }
  1946. vmxnet3_put_ring_to_file(f, &r->comp_ring);
  1947. qemu_put_byte(f, r->intr_idx);
  1948. qemu_put_be64(f, r->rx_stats_pa);
  1949. vmxnet3_put_rx_stats_to_file(f, &r->rxq_stats);
  1950. }
  1951. static int vmxnet3_post_load(void *opaque, int version_id)
  1952. {
  1953. VMXNET3State *s = opaque;
  1954. PCIDevice *d = PCI_DEVICE(s);
  1955. vmxnet_tx_pkt_init(&s->tx_pkt, s->max_tx_frags, s->peer_has_vhdr);
  1956. vmxnet_rx_pkt_init(&s->rx_pkt, s->peer_has_vhdr);
  1957. if (s->msix_used) {
  1958. if (!vmxnet3_use_msix_vectors(s, VMXNET3_MAX_INTRS)) {
  1959. VMW_WRPRN("Failed to re-use MSI-X vectors");
  1960. msix_uninit(d, &s->msix_bar, &s->msix_bar);
  1961. s->msix_used = false;
  1962. return -1;
  1963. }
  1964. }
  1965. vmxnet3_validate_queues(s);
  1966. vmxnet3_validate_interrupts(s);
  1967. return 0;
  1968. }
  1969. const VMStateInfo rxq_descr_info = {
  1970. .name = "rxq_descr",
  1971. .get = vmxnet3_get_rxq_descr,
  1972. .put = vmxnet3_put_rxq_descr
  1973. };
  1974. static int vmxnet3_get_int_state(QEMUFile *f, void *pv, size_t size)
  1975. {
  1976. Vmxnet3IntState *r = pv;
  1977. r->is_masked = qemu_get_byte(f);
  1978. r->is_pending = qemu_get_byte(f);
  1979. r->is_asserted = qemu_get_byte(f);
  1980. return 0;
  1981. }
  1982. static void vmxnet3_put_int_state(QEMUFile *f, void *pv, size_t size)
  1983. {
  1984. Vmxnet3IntState *r = pv;
  1985. qemu_put_byte(f, r->is_masked);
  1986. qemu_put_byte(f, r->is_pending);
  1987. qemu_put_byte(f, r->is_asserted);
  1988. }
  1989. const VMStateInfo int_state_info = {
  1990. .name = "int_state",
  1991. .get = vmxnet3_get_int_state,
  1992. .put = vmxnet3_put_int_state
  1993. };
  1994. static const VMStateDescription vmstate_vmxnet3 = {
  1995. .name = "vmxnet3",
  1996. .version_id = 1,
  1997. .minimum_version_id = 1,
  1998. .minimum_version_id_old = 1,
  1999. .pre_save = vmxnet3_pre_save,
  2000. .post_load = vmxnet3_post_load,
  2001. .fields = (VMStateField[]) {
  2002. VMSTATE_PCI_DEVICE(parent_obj, VMXNET3State),
  2003. VMSTATE_BOOL(rx_packets_compound, VMXNET3State),
  2004. VMSTATE_BOOL(rx_vlan_stripping, VMXNET3State),
  2005. VMSTATE_BOOL(lro_supported, VMXNET3State),
  2006. VMSTATE_UINT32(rx_mode, VMXNET3State),
  2007. VMSTATE_UINT32(mcast_list_len, VMXNET3State),
  2008. VMSTATE_UINT32(mcast_list_buff_size, VMXNET3State),
  2009. VMSTATE_UINT32_ARRAY(vlan_table, VMXNET3State, VMXNET3_VFT_SIZE),
  2010. VMSTATE_UINT32(mtu, VMXNET3State),
  2011. VMSTATE_UINT16(max_rx_frags, VMXNET3State),
  2012. VMSTATE_UINT32(max_tx_frags, VMXNET3State),
  2013. VMSTATE_UINT8(event_int_idx, VMXNET3State),
  2014. VMSTATE_BOOL(auto_int_masking, VMXNET3State),
  2015. VMSTATE_UINT8(txq_num, VMXNET3State),
  2016. VMSTATE_UINT8(rxq_num, VMXNET3State),
  2017. VMSTATE_UINT32(device_active, VMXNET3State),
  2018. VMSTATE_UINT32(last_command, VMXNET3State),
  2019. VMSTATE_UINT32(link_status_and_speed, VMXNET3State),
  2020. VMSTATE_UINT32(temp_mac, VMXNET3State),
  2021. VMSTATE_UINT64(drv_shmem, VMXNET3State),
  2022. VMSTATE_UINT64(temp_shared_guest_driver_memory, VMXNET3State),
  2023. VMSTATE_ARRAY(txq_descr, VMXNET3State,
  2024. VMXNET3_DEVICE_MAX_TX_QUEUES, 0, txq_descr_info,
  2025. Vmxnet3TxqDescr),
  2026. VMSTATE_ARRAY(rxq_descr, VMXNET3State,
  2027. VMXNET3_DEVICE_MAX_RX_QUEUES, 0, rxq_descr_info,
  2028. Vmxnet3RxqDescr),
  2029. VMSTATE_ARRAY(interrupt_states, VMXNET3State, VMXNET3_MAX_INTRS,
  2030. 0, int_state_info, Vmxnet3IntState),
  2031. VMSTATE_END_OF_LIST()
  2032. },
  2033. .subsections = (VMStateSubsection[]) {
  2034. {
  2035. .vmsd = &vmxstate_vmxnet3_mcast_list,
  2036. .needed = vmxnet3_mc_list_needed
  2037. },
  2038. {
  2039. /* empty element. */
  2040. }
  2041. }
  2042. };
  2043. static void
  2044. vmxnet3_write_config(PCIDevice *pci_dev, uint32_t addr, uint32_t val, int len)
  2045. {
  2046. pci_default_write_config(pci_dev, addr, val, len);
  2047. msix_write_config(pci_dev, addr, val, len);
  2048. msi_write_config(pci_dev, addr, val, len);
  2049. }
  2050. static Property vmxnet3_properties[] = {
  2051. DEFINE_NIC_PROPERTIES(VMXNET3State, conf),
  2052. DEFINE_PROP_END_OF_LIST(),
  2053. };
  2054. static void vmxnet3_class_init(ObjectClass *class, void *data)
  2055. {
  2056. DeviceClass *dc = DEVICE_CLASS(class);
  2057. PCIDeviceClass *c = PCI_DEVICE_CLASS(class);
  2058. c->init = vmxnet3_pci_init;
  2059. c->exit = vmxnet3_pci_uninit;
  2060. c->vendor_id = PCI_VENDOR_ID_VMWARE;
  2061. c->device_id = PCI_DEVICE_ID_VMWARE_VMXNET3;
  2062. c->revision = PCI_DEVICE_ID_VMWARE_VMXNET3_REVISION;
  2063. c->class_id = PCI_CLASS_NETWORK_ETHERNET;
  2064. c->subsystem_vendor_id = PCI_VENDOR_ID_VMWARE;
  2065. c->subsystem_id = PCI_DEVICE_ID_VMWARE_VMXNET3;
  2066. c->config_write = vmxnet3_write_config,
  2067. dc->desc = "VMWare Paravirtualized Ethernet v3";
  2068. dc->reset = vmxnet3_qdev_reset;
  2069. dc->vmsd = &vmstate_vmxnet3;
  2070. dc->props = vmxnet3_properties;
  2071. set_bit(DEVICE_CATEGORY_NETWORK, dc->categories);
  2072. }
  2073. static const TypeInfo vmxnet3_info = {
  2074. .name = TYPE_VMXNET3,
  2075. .parent = TYPE_PCI_DEVICE,
  2076. .instance_size = sizeof(VMXNET3State),
  2077. .class_init = vmxnet3_class_init,
  2078. };
  2079. static void vmxnet3_register_types(void)
  2080. {
  2081. VMW_CBPRN("vmxnet3_register_types called...");
  2082. type_register_static(&vmxnet3_info);
  2083. }
  2084. type_init(vmxnet3_register_types)