|
@@ -274,7 +274,11 @@ void qemu_event_destroy(QemuEvent *ev)
|
|
|
|
|
|
void qemu_event_set(QemuEvent *ev)
|
|
void qemu_event_set(QemuEvent *ev)
|
|
{
|
|
{
|
|
- if (atomic_mb_read(&ev->value) != EV_SET) {
|
|
|
|
|
|
+ /* qemu_event_set has release semantics, but because it *loads*
|
|
|
|
+ * ev->value we need a full memory barrier here.
|
|
|
|
+ */
|
|
|
|
+ smp_mb();
|
|
|
|
+ if (atomic_read(&ev->value) != EV_SET) {
|
|
if (atomic_xchg(&ev->value, EV_SET) == EV_BUSY) {
|
|
if (atomic_xchg(&ev->value, EV_SET) == EV_BUSY) {
|
|
/* There were waiters, wake them up. */
|
|
/* There were waiters, wake them up. */
|
|
SetEvent(ev->event);
|
|
SetEvent(ev->event);
|
|
@@ -284,7 +288,11 @@ void qemu_event_set(QemuEvent *ev)
|
|
|
|
|
|
void qemu_event_reset(QemuEvent *ev)
|
|
void qemu_event_reset(QemuEvent *ev)
|
|
{
|
|
{
|
|
- if (atomic_mb_read(&ev->value) == EV_SET) {
|
|
|
|
|
|
+ unsigned value;
|
|
|
|
+
|
|
|
|
+ value = atomic_read(&ev->value);
|
|
|
|
+ smp_mb_acquire();
|
|
|
|
+ if (value == EV_SET) {
|
|
/* If there was a concurrent reset (or even reset+wait),
|
|
/* If there was a concurrent reset (or even reset+wait),
|
|
* do nothing. Otherwise change EV_SET->EV_FREE.
|
|
* do nothing. Otherwise change EV_SET->EV_FREE.
|
|
*/
|
|
*/
|
|
@@ -296,7 +304,8 @@ void qemu_event_wait(QemuEvent *ev)
|
|
{
|
|
{
|
|
unsigned value;
|
|
unsigned value;
|
|
|
|
|
|
- value = atomic_mb_read(&ev->value);
|
|
|
|
|
|
+ value = atomic_read(&ev->value);
|
|
|
|
+ smp_mb_acquire();
|
|
if (value != EV_SET) {
|
|
if (value != EV_SET) {
|
|
if (value == EV_FREE) {
|
|
if (value == EV_FREE) {
|
|
/* qemu_event_set is not yet going to call SetEvent, but we are
|
|
/* qemu_event_set is not yet going to call SetEvent, but we are
|