|
@@ -0,0 +1,47 @@
|
|
|
+/*
|
|
|
+ * SPDX-License-Identifier: GPL-2.0-or-later
|
|
|
+ * Atomic store insert into 128-bit, AArch64 version.
|
|
|
+ *
|
|
|
+ * Copyright (C) 2023 Linaro, Ltd.
|
|
|
+ */
|
|
|
+
|
|
|
+#ifndef AARCH64_STORE_INSERT_AL16_H
|
|
|
+#define AARCH64_STORE_INSERT_AL16_H
|
|
|
+
|
|
|
+/**
|
|
|
+ * store_atom_insert_al16:
|
|
|
+ * @p: host address
|
|
|
+ * @val: shifted value to store
|
|
|
+ * @msk: mask for value to store
|
|
|
+ *
|
|
|
+ * Atomically store @val to @p masked by @msk.
|
|
|
+ */
|
|
|
+static inline void ATTRIBUTE_ATOMIC128_OPT
|
|
|
+store_atom_insert_al16(Int128 *ps, Int128 val, Int128 msk)
|
|
|
+{
|
|
|
+ /*
|
|
|
+ * GCC only implements __sync* primitives for int128 on aarch64.
|
|
|
+ * We can do better without the barriers, and integrating the
|
|
|
+ * arithmetic into the load-exclusive/store-conditional pair.
|
|
|
+ */
|
|
|
+ uint64_t tl, th, vl, vh, ml, mh;
|
|
|
+ uint32_t fail;
|
|
|
+
|
|
|
+ qemu_build_assert(!HOST_BIG_ENDIAN);
|
|
|
+ vl = int128_getlo(val);
|
|
|
+ vh = int128_gethi(val);
|
|
|
+ ml = int128_getlo(msk);
|
|
|
+ mh = int128_gethi(msk);
|
|
|
+
|
|
|
+ asm("0: ldxp %[l], %[h], %[mem]\n\t"
|
|
|
+ "bic %[l], %[l], %[ml]\n\t"
|
|
|
+ "bic %[h], %[h], %[mh]\n\t"
|
|
|
+ "orr %[l], %[l], %[vl]\n\t"
|
|
|
+ "orr %[h], %[h], %[vh]\n\t"
|
|
|
+ "stxp %w[f], %[l], %[h], %[mem]\n\t"
|
|
|
+ "cbnz %w[f], 0b\n"
|
|
|
+ : [mem] "+Q"(*ps), [f] "=&r"(fail), [l] "=&r"(tl), [h] "=&r"(th)
|
|
|
+ : [vl] "r"(vl), [vh] "r"(vh), [ml] "r"(ml), [mh] "r"(mh));
|
|
|
+}
|
|
|
+
|
|
|
+#endif /* AARCH64_STORE_INSERT_AL16_H */
|