summary refs log tree commit diff stats
path: root/host/include/aarch64/host/store-insert-al16.h.inc
blob: 1943155bc6032311fe822f5c3c91d071dc37c64f (plain) (blame)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
/*
 * SPDX-License-Identifier: GPL-2.0-or-later
 * Atomic store insert into 128-bit, AArch64 version.
 *
 * Copyright (C) 2023 Linaro, Ltd.
 */

#ifndef AARCH64_STORE_INSERT_AL16_H
#define AARCH64_STORE_INSERT_AL16_H

/**
 * store_atom_insert_al16:
 * @p: host address
 * @val: shifted value to store
 * @msk: mask for value to store
 *
 * Atomically store @val to @p masked by @msk.
 */
static inline void ATTRIBUTE_ATOMIC128_OPT
store_atom_insert_al16(Int128 *ps, Int128 val, Int128 msk)
{
    /*
     * GCC only implements __sync* primitives for int128 on aarch64.
     * We can do better without the barriers, and integrating the
     * arithmetic into the load-exclusive/store-conditional pair.
     */
    uint64_t tl, th, vl, vh, ml, mh;
    uint32_t fail;

    qemu_build_assert(!HOST_BIG_ENDIAN);
    vl = int128_getlo(val);
    vh = int128_gethi(val);
    ml = int128_getlo(msk);
    mh = int128_gethi(msk);

    asm("0: ldxp %[l], %[h], %[mem]\n\t"
        "bic %[l], %[l], %[ml]\n\t"
        "bic %[h], %[h], %[mh]\n\t"
        "orr %[l], %[l], %[vl]\n\t"
        "orr %[h], %[h], %[vh]\n\t"
        "stxp %w[f], %[l], %[h], %[mem]\n\t"
        "cbnz %w[f], 0b\n"
        : [mem] "+Q"(*ps), [f] "=&r"(fail), [l] "=&r"(tl), [h] "=&r"(th)
        : [vl] "r"(vl), [vh] "r"(vh), [ml] "r"(ml), [mh] "r"(mh));
}

#endif /* AARCH64_STORE_INSERT_AL16_H */