xics.c 15 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588
  1. /*
  2. * QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator
  3. *
  4. * PAPR Virtualized Interrupt System, aka ICS/ICP aka xics
  5. *
  6. * Copyright (c) 2010,2011 David Gibson, IBM Corporation.
  7. *
  8. * Permission is hereby granted, free of charge, to any person obtaining a copy
  9. * of this software and associated documentation files (the "Software"), to deal
  10. * in the Software without restriction, including without limitation the rights
  11. * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
  12. * copies of the Software, and to permit persons to whom the Software is
  13. * furnished to do so, subject to the following conditions:
  14. *
  15. * The above copyright notice and this permission notice shall be included in
  16. * all copies or substantial portions of the Software.
  17. *
  18. * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
  19. * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
  20. * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
  21. * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
  22. * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
  23. * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
  24. * THE SOFTWARE.
  25. *
  26. */
  27. #include "hw.h"
  28. #include "trace.h"
  29. #include "hw/spapr.h"
  30. #include "hw/xics.h"
  31. /*
  32. * ICP: Presentation layer
  33. */
  34. struct icp_server_state {
  35. uint32_t xirr;
  36. uint8_t pending_priority;
  37. uint8_t mfrr;
  38. qemu_irq output;
  39. };
  40. #define XISR_MASK 0x00ffffff
  41. #define CPPR_MASK 0xff000000
  42. #define XISR(ss) (((ss)->xirr) & XISR_MASK)
  43. #define CPPR(ss) (((ss)->xirr) >> 24)
  44. struct ics_state;
  45. struct icp_state {
  46. long nr_servers;
  47. struct icp_server_state *ss;
  48. struct ics_state *ics;
  49. };
  50. static void ics_reject(struct ics_state *ics, int nr);
  51. static void ics_resend(struct ics_state *ics);
  52. static void ics_eoi(struct ics_state *ics, int nr);
  53. static void icp_check_ipi(struct icp_state *icp, int server)
  54. {
  55. struct icp_server_state *ss = icp->ss + server;
  56. if (XISR(ss) && (ss->pending_priority <= ss->mfrr)) {
  57. return;
  58. }
  59. trace_xics_icp_check_ipi(server, ss->mfrr);
  60. if (XISR(ss)) {
  61. ics_reject(icp->ics, XISR(ss));
  62. }
  63. ss->xirr = (ss->xirr & ~XISR_MASK) | XICS_IPI;
  64. ss->pending_priority = ss->mfrr;
  65. qemu_irq_raise(ss->output);
  66. }
  67. static void icp_resend(struct icp_state *icp, int server)
  68. {
  69. struct icp_server_state *ss = icp->ss + server;
  70. if (ss->mfrr < CPPR(ss)) {
  71. icp_check_ipi(icp, server);
  72. }
  73. ics_resend(icp->ics);
  74. }
  75. static void icp_set_cppr(struct icp_state *icp, int server, uint8_t cppr)
  76. {
  77. struct icp_server_state *ss = icp->ss + server;
  78. uint8_t old_cppr;
  79. uint32_t old_xisr;
  80. old_cppr = CPPR(ss);
  81. ss->xirr = (ss->xirr & ~CPPR_MASK) | (cppr << 24);
  82. if (cppr < old_cppr) {
  83. if (XISR(ss) && (cppr <= ss->pending_priority)) {
  84. old_xisr = XISR(ss);
  85. ss->xirr &= ~XISR_MASK; /* Clear XISR */
  86. qemu_irq_lower(ss->output);
  87. ics_reject(icp->ics, old_xisr);
  88. }
  89. } else {
  90. if (!XISR(ss)) {
  91. icp_resend(icp, server);
  92. }
  93. }
  94. }
  95. static void icp_set_mfrr(struct icp_state *icp, int server, uint8_t mfrr)
  96. {
  97. struct icp_server_state *ss = icp->ss + server;
  98. ss->mfrr = mfrr;
  99. if (mfrr < CPPR(ss)) {
  100. icp_check_ipi(icp, server);
  101. }
  102. }
  103. static uint32_t icp_accept(struct icp_server_state *ss)
  104. {
  105. uint32_t xirr = ss->xirr;
  106. qemu_irq_lower(ss->output);
  107. ss->xirr = ss->pending_priority << 24;
  108. trace_xics_icp_accept(xirr, ss->xirr);
  109. return xirr;
  110. }
  111. static void icp_eoi(struct icp_state *icp, int server, uint32_t xirr)
  112. {
  113. struct icp_server_state *ss = icp->ss + server;
  114. /* Send EOI -> ICS */
  115. ss->xirr = (ss->xirr & ~CPPR_MASK) | (xirr & CPPR_MASK);
  116. trace_xics_icp_eoi(server, xirr, ss->xirr);
  117. ics_eoi(icp->ics, xirr & XISR_MASK);
  118. if (!XISR(ss)) {
  119. icp_resend(icp, server);
  120. }
  121. }
  122. static void icp_irq(struct icp_state *icp, int server, int nr, uint8_t priority)
  123. {
  124. struct icp_server_state *ss = icp->ss + server;
  125. trace_xics_icp_irq(server, nr, priority);
  126. if ((priority >= CPPR(ss))
  127. || (XISR(ss) && (ss->pending_priority <= priority))) {
  128. ics_reject(icp->ics, nr);
  129. } else {
  130. if (XISR(ss)) {
  131. ics_reject(icp->ics, XISR(ss));
  132. }
  133. ss->xirr = (ss->xirr & ~XISR_MASK) | (nr & XISR_MASK);
  134. ss->pending_priority = priority;
  135. trace_xics_icp_raise(ss->xirr, ss->pending_priority);
  136. qemu_irq_raise(ss->output);
  137. }
  138. }
  139. /*
  140. * ICS: Source layer
  141. */
  142. struct ics_irq_state {
  143. int server;
  144. uint8_t priority;
  145. uint8_t saved_priority;
  146. #define XICS_STATUS_ASSERTED 0x1
  147. #define XICS_STATUS_SENT 0x2
  148. #define XICS_STATUS_REJECTED 0x4
  149. #define XICS_STATUS_MASKED_PENDING 0x8
  150. uint8_t status;
  151. };
  152. struct ics_state {
  153. int nr_irqs;
  154. int offset;
  155. qemu_irq *qirqs;
  156. bool *islsi;
  157. struct ics_irq_state *irqs;
  158. struct icp_state *icp;
  159. };
  160. static int ics_valid_irq(struct ics_state *ics, uint32_t nr)
  161. {
  162. return (nr >= ics->offset)
  163. && (nr < (ics->offset + ics->nr_irqs));
  164. }
  165. static void resend_msi(struct ics_state *ics, int srcno)
  166. {
  167. struct ics_irq_state *irq = ics->irqs + srcno;
  168. /* FIXME: filter by server#? */
  169. if (irq->status & XICS_STATUS_REJECTED) {
  170. irq->status &= ~XICS_STATUS_REJECTED;
  171. if (irq->priority != 0xff) {
  172. icp_irq(ics->icp, irq->server, srcno + ics->offset,
  173. irq->priority);
  174. }
  175. }
  176. }
  177. static void resend_lsi(struct ics_state *ics, int srcno)
  178. {
  179. struct ics_irq_state *irq = ics->irqs + srcno;
  180. if ((irq->priority != 0xff)
  181. && (irq->status & XICS_STATUS_ASSERTED)
  182. && !(irq->status & XICS_STATUS_SENT)) {
  183. irq->status |= XICS_STATUS_SENT;
  184. icp_irq(ics->icp, irq->server, srcno + ics->offset, irq->priority);
  185. }
  186. }
  187. static void set_irq_msi(struct ics_state *ics, int srcno, int val)
  188. {
  189. struct ics_irq_state *irq = ics->irqs + srcno;
  190. trace_xics_set_irq_msi(srcno, srcno + ics->offset);
  191. if (val) {
  192. if (irq->priority == 0xff) {
  193. irq->status |= XICS_STATUS_MASKED_PENDING;
  194. trace_xics_masked_pending();
  195. } else {
  196. icp_irq(ics->icp, irq->server, srcno + ics->offset, irq->priority);
  197. }
  198. }
  199. }
  200. static void set_irq_lsi(struct ics_state *ics, int srcno, int val)
  201. {
  202. struct ics_irq_state *irq = ics->irqs + srcno;
  203. trace_xics_set_irq_lsi(srcno, srcno + ics->offset);
  204. if (val) {
  205. irq->status |= XICS_STATUS_ASSERTED;
  206. } else {
  207. irq->status &= ~XICS_STATUS_ASSERTED;
  208. }
  209. resend_lsi(ics, srcno);
  210. }
  211. static void ics_set_irq(void *opaque, int srcno, int val)
  212. {
  213. struct ics_state *ics = (struct ics_state *)opaque;
  214. if (ics->islsi[srcno]) {
  215. set_irq_lsi(ics, srcno, val);
  216. } else {
  217. set_irq_msi(ics, srcno, val);
  218. }
  219. }
  220. static void write_xive_msi(struct ics_state *ics, int srcno)
  221. {
  222. struct ics_irq_state *irq = ics->irqs + srcno;
  223. if (!(irq->status & XICS_STATUS_MASKED_PENDING)
  224. || (irq->priority == 0xff)) {
  225. return;
  226. }
  227. irq->status &= ~XICS_STATUS_MASKED_PENDING;
  228. icp_irq(ics->icp, irq->server, srcno + ics->offset, irq->priority);
  229. }
  230. static void write_xive_lsi(struct ics_state *ics, int srcno)
  231. {
  232. resend_lsi(ics, srcno);
  233. }
  234. static void ics_write_xive(struct ics_state *ics, int nr, int server,
  235. uint8_t priority, uint8_t saved_priority)
  236. {
  237. int srcno = nr - ics->offset;
  238. struct ics_irq_state *irq = ics->irqs + srcno;
  239. irq->server = server;
  240. irq->priority = priority;
  241. irq->saved_priority = saved_priority;
  242. trace_xics_ics_write_xive(nr, srcno, server, priority);
  243. if (ics->islsi[srcno]) {
  244. write_xive_lsi(ics, srcno);
  245. } else {
  246. write_xive_msi(ics, srcno);
  247. }
  248. }
  249. static void ics_reject(struct ics_state *ics, int nr)
  250. {
  251. struct ics_irq_state *irq = ics->irqs + nr - ics->offset;
  252. trace_xics_ics_reject(nr, nr - ics->offset);
  253. irq->status |= XICS_STATUS_REJECTED; /* Irrelevant but harmless for LSI */
  254. irq->status &= ~XICS_STATUS_SENT; /* Irrelevant but harmless for MSI */
  255. }
  256. static void ics_resend(struct ics_state *ics)
  257. {
  258. int i;
  259. for (i = 0; i < ics->nr_irqs; i++) {
  260. /* FIXME: filter by server#? */
  261. if (ics->islsi[i]) {
  262. resend_lsi(ics, i);
  263. } else {
  264. resend_msi(ics, i);
  265. }
  266. }
  267. }
  268. static void ics_eoi(struct ics_state *ics, int nr)
  269. {
  270. int srcno = nr - ics->offset;
  271. struct ics_irq_state *irq = ics->irqs + srcno;
  272. trace_xics_ics_eoi(nr);
  273. if (ics->islsi[srcno]) {
  274. irq->status &= ~XICS_STATUS_SENT;
  275. }
  276. }
  277. /*
  278. * Exported functions
  279. */
  280. qemu_irq xics_get_qirq(struct icp_state *icp, int irq)
  281. {
  282. if (!ics_valid_irq(icp->ics, irq)) {
  283. return NULL;
  284. }
  285. return icp->ics->qirqs[irq - icp->ics->offset];
  286. }
  287. void xics_set_irq_type(struct icp_state *icp, int irq, bool lsi)
  288. {
  289. assert(ics_valid_irq(icp->ics, irq));
  290. icp->ics->islsi[irq - icp->ics->offset] = lsi;
  291. }
  292. static target_ulong h_cppr(PowerPCCPU *cpu, sPAPREnvironment *spapr,
  293. target_ulong opcode, target_ulong *args)
  294. {
  295. CPUState *cs = CPU(cpu);
  296. target_ulong cppr = args[0];
  297. icp_set_cppr(spapr->icp, cs->cpu_index, cppr);
  298. return H_SUCCESS;
  299. }
  300. static target_ulong h_ipi(PowerPCCPU *cpu, sPAPREnvironment *spapr,
  301. target_ulong opcode, target_ulong *args)
  302. {
  303. target_ulong server = args[0];
  304. target_ulong mfrr = args[1];
  305. if (server >= spapr->icp->nr_servers) {
  306. return H_PARAMETER;
  307. }
  308. icp_set_mfrr(spapr->icp, server, mfrr);
  309. return H_SUCCESS;
  310. }
  311. static target_ulong h_xirr(PowerPCCPU *cpu, sPAPREnvironment *spapr,
  312. target_ulong opcode, target_ulong *args)
  313. {
  314. CPUState *cs = CPU(cpu);
  315. uint32_t xirr = icp_accept(spapr->icp->ss + cs->cpu_index);
  316. args[0] = xirr;
  317. return H_SUCCESS;
  318. }
  319. static target_ulong h_eoi(PowerPCCPU *cpu, sPAPREnvironment *spapr,
  320. target_ulong opcode, target_ulong *args)
  321. {
  322. CPUState *cs = CPU(cpu);
  323. target_ulong xirr = args[0];
  324. icp_eoi(spapr->icp, cs->cpu_index, xirr);
  325. return H_SUCCESS;
  326. }
  327. static void rtas_set_xive(sPAPREnvironment *spapr, uint32_t token,
  328. uint32_t nargs, target_ulong args,
  329. uint32_t nret, target_ulong rets)
  330. {
  331. struct ics_state *ics = spapr->icp->ics;
  332. uint32_t nr, server, priority;
  333. if ((nargs != 3) || (nret != 1)) {
  334. rtas_st(rets, 0, -3);
  335. return;
  336. }
  337. nr = rtas_ld(args, 0);
  338. server = rtas_ld(args, 1);
  339. priority = rtas_ld(args, 2);
  340. if (!ics_valid_irq(ics, nr) || (server >= ics->icp->nr_servers)
  341. || (priority > 0xff)) {
  342. rtas_st(rets, 0, -3);
  343. return;
  344. }
  345. ics_write_xive(ics, nr, server, priority, priority);
  346. rtas_st(rets, 0, 0); /* Success */
  347. }
  348. static void rtas_get_xive(sPAPREnvironment *spapr, uint32_t token,
  349. uint32_t nargs, target_ulong args,
  350. uint32_t nret, target_ulong rets)
  351. {
  352. struct ics_state *ics = spapr->icp->ics;
  353. uint32_t nr;
  354. if ((nargs != 1) || (nret != 3)) {
  355. rtas_st(rets, 0, -3);
  356. return;
  357. }
  358. nr = rtas_ld(args, 0);
  359. if (!ics_valid_irq(ics, nr)) {
  360. rtas_st(rets, 0, -3);
  361. return;
  362. }
  363. rtas_st(rets, 0, 0); /* Success */
  364. rtas_st(rets, 1, ics->irqs[nr - ics->offset].server);
  365. rtas_st(rets, 2, ics->irqs[nr - ics->offset].priority);
  366. }
  367. static void rtas_int_off(sPAPREnvironment *spapr, uint32_t token,
  368. uint32_t nargs, target_ulong args,
  369. uint32_t nret, target_ulong rets)
  370. {
  371. struct ics_state *ics = spapr->icp->ics;
  372. uint32_t nr;
  373. if ((nargs != 1) || (nret != 1)) {
  374. rtas_st(rets, 0, -3);
  375. return;
  376. }
  377. nr = rtas_ld(args, 0);
  378. if (!ics_valid_irq(ics, nr)) {
  379. rtas_st(rets, 0, -3);
  380. return;
  381. }
  382. ics_write_xive(ics, nr, ics->irqs[nr - ics->offset].server, 0xff,
  383. ics->irqs[nr - ics->offset].priority);
  384. rtas_st(rets, 0, 0); /* Success */
  385. }
  386. static void rtas_int_on(sPAPREnvironment *spapr, uint32_t token,
  387. uint32_t nargs, target_ulong args,
  388. uint32_t nret, target_ulong rets)
  389. {
  390. struct ics_state *ics = spapr->icp->ics;
  391. uint32_t nr;
  392. if ((nargs != 1) || (nret != 1)) {
  393. rtas_st(rets, 0, -3);
  394. return;
  395. }
  396. nr = rtas_ld(args, 0);
  397. if (!ics_valid_irq(ics, nr)) {
  398. rtas_st(rets, 0, -3);
  399. return;
  400. }
  401. ics_write_xive(ics, nr, ics->irqs[nr - ics->offset].server,
  402. ics->irqs[nr - ics->offset].saved_priority,
  403. ics->irqs[nr - ics->offset].saved_priority);
  404. rtas_st(rets, 0, 0); /* Success */
  405. }
  406. static void xics_reset(void *opaque)
  407. {
  408. struct icp_state *icp = (struct icp_state *)opaque;
  409. struct ics_state *ics = icp->ics;
  410. int i;
  411. for (i = 0; i < icp->nr_servers; i++) {
  412. icp->ss[i].xirr = 0;
  413. icp->ss[i].pending_priority = 0xff;
  414. icp->ss[i].mfrr = 0xff;
  415. /* Make all outputs are deasserted */
  416. qemu_set_irq(icp->ss[i].output, 0);
  417. }
  418. memset(ics->irqs, 0, sizeof(struct ics_irq_state) * ics->nr_irqs);
  419. for (i = 0; i < ics->nr_irqs; i++) {
  420. ics->irqs[i].priority = 0xff;
  421. ics->irqs[i].saved_priority = 0xff;
  422. }
  423. }
  424. struct icp_state *xics_system_init(int nr_irqs)
  425. {
  426. CPUPPCState *env;
  427. CPUState *cpu;
  428. int max_server_num;
  429. struct icp_state *icp;
  430. struct ics_state *ics;
  431. max_server_num = -1;
  432. for (env = first_cpu; env != NULL; env = env->next_cpu) {
  433. cpu = CPU(ppc_env_get_cpu(env));
  434. if (cpu->cpu_index > max_server_num) {
  435. max_server_num = cpu->cpu_index;
  436. }
  437. }
  438. icp = g_malloc0(sizeof(*icp));
  439. icp->nr_servers = max_server_num + 1;
  440. icp->ss = g_malloc0(icp->nr_servers*sizeof(struct icp_server_state));
  441. for (env = first_cpu; env != NULL; env = env->next_cpu) {
  442. cpu = CPU(ppc_env_get_cpu(env));
  443. struct icp_server_state *ss = &icp->ss[cpu->cpu_index];
  444. switch (PPC_INPUT(env)) {
  445. case PPC_FLAGS_INPUT_POWER7:
  446. ss->output = env->irq_inputs[POWER7_INPUT_INT];
  447. break;
  448. case PPC_FLAGS_INPUT_970:
  449. ss->output = env->irq_inputs[PPC970_INPUT_INT];
  450. break;
  451. default:
  452. hw_error("XICS interrupt model does not support this CPU bus "
  453. "model\n");
  454. exit(1);
  455. }
  456. }
  457. ics = g_malloc0(sizeof(*ics));
  458. ics->nr_irqs = nr_irqs;
  459. ics->offset = XICS_IRQ_BASE;
  460. ics->irqs = g_malloc0(nr_irqs * sizeof(struct ics_irq_state));
  461. ics->islsi = g_malloc0(nr_irqs * sizeof(bool));
  462. icp->ics = ics;
  463. ics->icp = icp;
  464. ics->qirqs = qemu_allocate_irqs(ics_set_irq, ics, nr_irqs);
  465. spapr_register_hypercall(H_CPPR, h_cppr);
  466. spapr_register_hypercall(H_IPI, h_ipi);
  467. spapr_register_hypercall(H_XIRR, h_xirr);
  468. spapr_register_hypercall(H_EOI, h_eoi);
  469. spapr_rtas_register("ibm,set-xive", rtas_set_xive);
  470. spapr_rtas_register("ibm,get-xive", rtas_get_xive);
  471. spapr_rtas_register("ibm,int-off", rtas_int_off);
  472. spapr_rtas_register("ibm,int-on", rtas_int_on);
  473. qemu_register_reset(xics_reset, icp);
  474. return icp;
  475. }