mirror of
https://git.FreeBSD.org/ports.git
synced 2025-01-01 05:45:45 +00:00
lang/ghc: Update to 9.6.5
This commit is contained in:
parent
4ec2bfdbda
commit
920765491c
@ -1,6 +1,6 @@
|
||||
PORTNAME= ghc
|
||||
PORTVERSION= ${GHC_VERSION}
|
||||
PORTREVISION?= 1
|
||||
PORTREVISION?= 0
|
||||
CATEGORIES= lang haskell
|
||||
MASTER_SITES= https://www.haskell.org/ghc/dist/${PORTVERSION}/:source \
|
||||
LOCAL/arrowd/:boot
|
||||
@ -71,19 +71,19 @@ GMP_VARS_OFF= hadrian_gmp_arg=--bignum=native
|
||||
PROFILE_VARS= enable_profile=YES hadrian_setting_profile=True
|
||||
PROFILE_VARS_OFF= enable_profile=NO hadrian_setting_profile=False
|
||||
|
||||
GHC_VERSION?= 9.6.4
|
||||
GHC_VERSION?= 9.6.5
|
||||
LLVM_VERSION?= 15
|
||||
BOOT_GHC_VERSION?= 9.2.8
|
||||
# LLVM version that bootstrap compiler uses
|
||||
BOOT_LLVM_VERSION?= 12
|
||||
|
||||
BASE_PACKAGES?= Cabal-3.10.1.0 array-0.5.6.0 base-4.18.2.0 binary-0.8.9.1 \
|
||||
BASE_PACKAGES?= Cabal-3.10.3.0 array-0.5.6.0 base-4.18.2.1 binary-0.8.9.1 \
|
||||
bytestring-0.11.5.3 containers-0.6.7 deepseq-1.4.8.1 \
|
||||
directory-1.3.8.1 exceptions-0.10.7 filepath-1.4.200.1 \
|
||||
directory-1.3.8.4 exceptions-0.10.7 filepath-1.4.300.1 \
|
||||
ghc-${GHC_VERSION} ghc-bignum-1.3 ghc-compact-0.1.0.0 \
|
||||
ghc-prim-0.10.0 haskeline-0.8.2.1 hpc-0.6.2.0 \
|
||||
integer-gmp-1.1 mtl-2.3.1 parsec-3.1.16.1 pretty-1.1.3.6 \
|
||||
process-1.6.17.0 stm-2.5.1.0 template-haskell-2.20.0.0 \
|
||||
process-1.6.19.0 stm-2.5.1.0 template-haskell-2.20.0.0 \
|
||||
terminfo-0.4.1.6 text-2.0.2 time-1.12.2 \
|
||||
transformers-0.6.1.0 unix-2.8.4.0 xhtml-3000.2.2.1
|
||||
|
||||
@ -338,7 +338,7 @@ check-hadrian-bootstrap:
|
||||
${FALSE}
|
||||
.endif
|
||||
# Install bootstrap GHC
|
||||
${MAKE_CMD} -C ${.CURDIR} pre-configure
|
||||
${MAKE} -C ${.CURDIR} pre-configure
|
||||
# Compile Hadrian
|
||||
cd ${WRKSRC}/hadrian/bootstrap && \
|
||||
./bootstrap.py -w ${BOOT_GHC} -s /tmp/hadrian-${GHC_VERSION}-boot.tar.gz
|
||||
|
@ -1,14 +1,14 @@
|
||||
TIMESTAMP = 1708789087
|
||||
SHA256 (ghc-9.6.4-src.tar.xz) = 10bf25b8b07174fdd9868b5c0c56c17c0ef1edcb6247b4b864be933651bfd4c0
|
||||
SIZE (ghc-9.6.4-src.tar.xz) = 29451856
|
||||
TIMESTAMP = 1719575256
|
||||
SHA256 (ghc-9.6.5-src.tar.xz) = 87b389924f98c1a26c205122757338c8dab33ad1fcf670faa22622742432b93c
|
||||
SIZE (ghc-9.6.5-src.tar.xz) = 29463284
|
||||
SHA256 (ghc-9.4.8-src.tar.xz) = 0bf407eb67fe3e3c24b0f4c8dea8cb63e07f63ca0f76cf2058565143507ab85e
|
||||
SIZE (ghc-9.4.8-src.tar.xz) = 28655220
|
||||
SHA256 (ghc-9.2.8-src.tar.xz) = 5f13d1786bf4fd12f4b45faa37abedb5bb3f36d5e58f7da5307e8bfe88a567a1
|
||||
SIZE (ghc-9.2.8-src.tar.xz) = 24661260
|
||||
SHA256 (ghc-8.10.7-src.tar.xz) = e3eef6229ce9908dfe1ea41436befb0455fefb1932559e860ad4c606b0d03c9d
|
||||
SIZE (ghc-8.10.7-src.tar.xz) = 19932668
|
||||
SHA256 (hadrian-9.6.4-boot.tar.gz) = 93152bb56ece601b6ea9924aa2ded2a66768c345607f790f3fe015055c1d6cdf
|
||||
SIZE (hadrian-9.6.4-boot.tar.gz) = 1495579
|
||||
SHA256 (hadrian-9.6.5-boot.tar.gz) = a5a7e6721d693c0e004fecff5745e47b2ec0c703b8378ccc5144e20a76049f95
|
||||
SIZE (hadrian-9.6.5-boot.tar.gz) = 1495606
|
||||
SHA256 (hadrian-9.4.8-boot.tar.gz) = 5eb060655686921b086645dcbdd6268f09bc66baf8cc76fce74dea9b8b55ab5d
|
||||
SIZE (hadrian-9.4.8-boot.tar.gz) = 1495586
|
||||
SHA256 (hadrian-9.2.8-boot.tar.gz) = 730412e1c17309670075be8f76dbca9ebadedfa10b758e605f9e67804e9e8ac5
|
||||
|
@ -1,822 +0,0 @@
|
||||
https://gitlab.haskell.org/ghc/ghc/-/merge_requests/12227
|
||||
|
||||
https://gitlab.haskell.org/ghc/ghc/-/issues/24580
|
||||
|
||||
--- rts/ARMOutlineAtomicsSymbols.h.orig 2024-01-08 13:24:59 UTC
|
||||
+++ rts/ARMOutlineAtomicsSymbols.h
|
||||
@@ -10,583 +10,583 @@
|
||||
#include <stdint.h>
|
||||
#include <stdatomic.h>
|
||||
|
||||
-uint8_t ghc___aarch64_cas1_relax(uint8_t old, uint8_t new, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_cas1_relax(uint8_t old, uint8_t new, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_cas1_relax(uint8_t old, uint8_t new, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_cas1_relax(uint8_t old, uint8_t new, _Atomic uint8_t* p) {
|
||||
atomic_compare_exchange_strong_explicit(p, &old, new, memory_order_relaxed, memory_order_relaxed); return old;
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_cas1_acq(uint8_t old, uint8_t new, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_cas1_acq(uint8_t old, uint8_t new, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_cas1_acq(uint8_t old, uint8_t new, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_cas1_acq(uint8_t old, uint8_t new, _Atomic uint8_t* p) {
|
||||
atomic_compare_exchange_strong_explicit(p, &old, new, memory_order_acquire, memory_order_acquire); return old;
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_cas1_acq_rel(uint8_t old, uint8_t new, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_cas1_acq_rel(uint8_t old, uint8_t new, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_cas1_acq_rel(uint8_t old, uint8_t new, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_cas1_acq_rel(uint8_t old, uint8_t new, _Atomic uint8_t* p) {
|
||||
atomic_compare_exchange_strong_explicit(p, &old, new, memory_order_acq_rel, memory_order_acquire); return old;
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_cas1_sync(uint8_t old, uint8_t new, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_cas1_sync(uint8_t old, uint8_t new, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_cas1_sync(uint8_t old, uint8_t new, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_cas1_sync(uint8_t old, uint8_t new, _Atomic uint8_t* p) {
|
||||
atomic_compare_exchange_strong_explicit(p, &old, new, memory_order_seq_cst, memory_order_seq_cst); return old;
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_cas2_relax(uint16_t old, uint16_t new, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_cas2_relax(uint16_t old, uint16_t new, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_cas2_relax(uint16_t old, uint16_t new, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_cas2_relax(uint16_t old, uint16_t new, _Atomic uint16_t* p) {
|
||||
atomic_compare_exchange_strong_explicit(p, &old, new, memory_order_relaxed, memory_order_relaxed); return old;
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_cas2_acq(uint16_t old, uint16_t new, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_cas2_acq(uint16_t old, uint16_t new, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_cas2_acq(uint16_t old, uint16_t new, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_cas2_acq(uint16_t old, uint16_t new, _Atomic uint16_t* p) {
|
||||
atomic_compare_exchange_strong_explicit(p, &old, new, memory_order_acquire, memory_order_acquire); return old;
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_cas2_acq_rel(uint16_t old, uint16_t new, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_cas2_acq_rel(uint16_t old, uint16_t new, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_cas2_acq_rel(uint16_t old, uint16_t new, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_cas2_acq_rel(uint16_t old, uint16_t new, _Atomic uint16_t* p) {
|
||||
atomic_compare_exchange_strong_explicit(p, &old, new, memory_order_acq_rel, memory_order_acquire); return old;
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_cas2_sync(uint16_t old, uint16_t new, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_cas2_sync(uint16_t old, uint16_t new, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_cas2_sync(uint16_t old, uint16_t new, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_cas2_sync(uint16_t old, uint16_t new, _Atomic uint16_t* p) {
|
||||
atomic_compare_exchange_strong_explicit(p, &old, new, memory_order_seq_cst, memory_order_seq_cst); return old;
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_cas4_relax(uint32_t old, uint32_t new, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_cas4_relax(uint32_t old, uint32_t new, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_cas4_relax(uint32_t old, uint32_t new, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_cas4_relax(uint32_t old, uint32_t new, _Atomic uint32_t* p) {
|
||||
atomic_compare_exchange_strong_explicit(p, &old, new, memory_order_relaxed, memory_order_relaxed); return old;
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_cas4_acq(uint32_t old, uint32_t new, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_cas4_acq(uint32_t old, uint32_t new, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_cas4_acq(uint32_t old, uint32_t new, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_cas4_acq(uint32_t old, uint32_t new, _Atomic uint32_t* p) {
|
||||
atomic_compare_exchange_strong_explicit(p, &old, new, memory_order_acquire, memory_order_acquire); return old;
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_cas4_acq_rel(uint32_t old, uint32_t new, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_cas4_acq_rel(uint32_t old, uint32_t new, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_cas4_acq_rel(uint32_t old, uint32_t new, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_cas4_acq_rel(uint32_t old, uint32_t new, _Atomic uint32_t* p) {
|
||||
atomic_compare_exchange_strong_explicit(p, &old, new, memory_order_acq_rel, memory_order_acquire); return old;
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_cas4_sync(uint32_t old, uint32_t new, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_cas4_sync(uint32_t old, uint32_t new, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_cas4_sync(uint32_t old, uint32_t new, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_cas4_sync(uint32_t old, uint32_t new, _Atomic uint32_t* p) {
|
||||
atomic_compare_exchange_strong_explicit(p, &old, new, memory_order_seq_cst, memory_order_seq_cst); return old;
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_cas8_relax(uint64_t old, uint64_t new, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_cas8_relax(uint64_t old, uint64_t new, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_cas8_relax(uint64_t old, uint64_t new, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_cas8_relax(uint64_t old, uint64_t new, _Atomic uint64_t* p) {
|
||||
atomic_compare_exchange_strong_explicit(p, &old, new, memory_order_relaxed, memory_order_relaxed); return old;
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_cas8_acq(uint64_t old, uint64_t new, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_cas8_acq(uint64_t old, uint64_t new, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_cas8_acq(uint64_t old, uint64_t new, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_cas8_acq(uint64_t old, uint64_t new, _Atomic uint64_t* p) {
|
||||
atomic_compare_exchange_strong_explicit(p, &old, new, memory_order_acquire, memory_order_acquire); return old;
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_cas8_acq_rel(uint64_t old, uint64_t new, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_cas8_acq_rel(uint64_t old, uint64_t new, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_cas8_acq_rel(uint64_t old, uint64_t new, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_cas8_acq_rel(uint64_t old, uint64_t new, _Atomic uint64_t* p) {
|
||||
atomic_compare_exchange_strong_explicit(p, &old, new, memory_order_acq_rel, memory_order_acquire); return old;
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_cas8_sync(uint64_t old, uint64_t new, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_cas8_sync(uint64_t old, uint64_t new, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_cas8_sync(uint64_t old, uint64_t new, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_cas8_sync(uint64_t old, uint64_t new, _Atomic uint64_t* p) {
|
||||
atomic_compare_exchange_strong_explicit(p, &old, new, memory_order_seq_cst, memory_order_seq_cst); return old;
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_swp1_relax(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_swp1_relax(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_swp1_relax(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_swp1_relax(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_swp1_acq(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_swp1_acq(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_swp1_acq(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_swp1_acq(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_swp1_rel(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_swp1_rel(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_swp1_rel(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_swp1_rel(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_swp1_acq_rel(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_swp1_acq_rel(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_swp1_acq_rel(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_swp1_acq_rel(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_swp1_sync(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_swp1_sync(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_swp1_sync(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_swp1_sync(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_swp2_relax(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_swp2_relax(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_swp2_relax(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_swp2_relax(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_swp2_acq(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_swp2_acq(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_swp2_acq(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_swp2_acq(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_swp2_rel(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_swp2_rel(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_swp2_rel(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_swp2_rel(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_swp2_acq_rel(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_swp2_acq_rel(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_swp2_acq_rel(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_swp2_acq_rel(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_swp2_sync(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_swp2_sync(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_swp2_sync(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_swp2_sync(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_swp4_relax(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_swp4_relax(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_swp4_relax(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_swp4_relax(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_swp4_acq(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_swp4_acq(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_swp4_acq(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_swp4_acq(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_swp4_rel(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_swp4_rel(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_swp4_rel(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_swp4_rel(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_swp4_acq_rel(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_swp4_acq_rel(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_swp4_acq_rel(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_swp4_acq_rel(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_swp4_sync(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_swp4_sync(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_swp4_sync(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_swp4_sync(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_swp8_relax(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_swp8_relax(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_swp8_relax(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_swp8_relax(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_swp8_acq(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_swp8_acq(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_swp8_acq(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_swp8_acq(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_swp8_rel(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_swp8_rel(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_swp8_rel(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_swp8_rel(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_swp8_acq_rel(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_swp8_acq_rel(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_swp8_acq_rel(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_swp8_acq_rel(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_swp8_sync(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_swp8_sync(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_swp8_sync(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_swp8_sync(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_exchange_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldadd1_relax(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldadd1_relax(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldadd1_relax(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldadd1_relax(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldadd1_acq(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldadd1_acq(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldadd1_acq(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldadd1_acq(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldadd1_rel(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldadd1_rel(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldadd1_rel(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldadd1_rel(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldadd1_acq_rel(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldadd1_acq_rel(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldadd1_acq_rel(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldadd1_acq_rel(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldadd1_sync(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldadd1_sync(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldadd1_sync(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldadd1_sync(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldadd2_relax(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldadd2_relax(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldadd2_relax(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldadd2_relax(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldadd2_acq(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldadd2_acq(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldadd2_acq(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldadd2_acq(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldadd2_rel(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldadd2_rel(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldadd2_rel(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldadd2_rel(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldadd2_acq_rel(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldadd2_acq_rel(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldadd2_acq_rel(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldadd2_acq_rel(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldadd2_sync(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldadd2_sync(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldadd2_sync(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldadd2_sync(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldadd4_relax(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldadd4_relax(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldadd4_relax(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldadd4_relax(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldadd4_acq(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldadd4_acq(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldadd4_acq(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldadd4_acq(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldadd4_rel(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldadd4_rel(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldadd4_rel(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldadd4_rel(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldadd4_acq_rel(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldadd4_acq_rel(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldadd4_acq_rel(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldadd4_acq_rel(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldadd4_sync(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldadd4_sync(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldadd4_sync(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldadd4_sync(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldadd8_relax(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldadd8_relax(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldadd8_relax(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldadd8_relax(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldadd8_acq(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldadd8_acq(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldadd8_acq(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldadd8_acq(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldadd8_rel(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldadd8_rel(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldadd8_rel(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldadd8_rel(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldadd8_acq_rel(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldadd8_acq_rel(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldadd8_acq_rel(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldadd8_acq_rel(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldadd8_sync(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldadd8_sync(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldadd8_sync(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldadd8_sync(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_add_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldclr1_relax(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldclr1_relax(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldclr1_relax(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldclr1_relax(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldclr1_acq(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldclr1_acq(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldclr1_acq(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldclr1_acq(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldclr1_rel(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldclr1_rel(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldclr1_rel(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldclr1_rel(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldclr1_acq_rel(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldclr1_acq_rel(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldclr1_acq_rel(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldclr1_acq_rel(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldclr1_sync(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldclr1_sync(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldclr1_sync(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldclr1_sync(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldclr2_relax(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldclr2_relax(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldclr2_relax(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldclr2_relax(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldclr2_acq(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldclr2_acq(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldclr2_acq(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldclr2_acq(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldclr2_rel(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldclr2_rel(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldclr2_rel(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldclr2_rel(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldclr2_acq_rel(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldclr2_acq_rel(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldclr2_acq_rel(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldclr2_acq_rel(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldclr2_sync(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldclr2_sync(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldclr2_sync(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldclr2_sync(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldclr4_relax(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldclr4_relax(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldclr4_relax(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldclr4_relax(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldclr4_acq(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldclr4_acq(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldclr4_acq(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldclr4_acq(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldclr4_rel(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldclr4_rel(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldclr4_rel(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldclr4_rel(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldclr4_acq_rel(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldclr4_acq_rel(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldclr4_acq_rel(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldclr4_acq_rel(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldclr4_sync(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldclr4_sync(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldclr4_sync(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldclr4_sync(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldclr8_relax(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldclr8_relax(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldclr8_relax(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldclr8_relax(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldclr8_acq(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldclr8_acq(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldclr8_acq(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldclr8_acq(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldclr8_rel(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldclr8_rel(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldclr8_rel(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldclr8_rel(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldclr8_acq_rel(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldclr8_acq_rel(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldclr8_acq_rel(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldclr8_acq_rel(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldclr8_sync(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldclr8_sync(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldclr8_sync(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldclr8_sync(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_and_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldeor1_relax(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldeor1_relax(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldeor1_relax(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldeor1_relax(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldeor1_acq(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldeor1_acq(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldeor1_acq(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldeor1_acq(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldeor1_rel(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldeor1_rel(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldeor1_rel(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldeor1_rel(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldeor1_acq_rel(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldeor1_acq_rel(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldeor1_acq_rel(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldeor1_acq_rel(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldeor1_sync(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldeor1_sync(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldeor1_sync(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldeor1_sync(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldeor2_relax(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldeor2_relax(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldeor2_relax(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldeor2_relax(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldeor2_acq(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldeor2_acq(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldeor2_acq(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldeor2_acq(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldeor2_rel(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldeor2_rel(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldeor2_rel(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldeor2_rel(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldeor2_acq_rel(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldeor2_acq_rel(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldeor2_acq_rel(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldeor2_acq_rel(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldeor2_sync(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldeor2_sync(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldeor2_sync(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldeor2_sync(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldeor4_relax(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldeor4_relax(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldeor4_relax(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldeor4_relax(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldeor4_acq(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldeor4_acq(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldeor4_acq(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldeor4_acq(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldeor4_rel(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldeor4_rel(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldeor4_rel(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldeor4_rel(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldeor4_acq_rel(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldeor4_acq_rel(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldeor4_acq_rel(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldeor4_acq_rel(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldeor4_sync(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldeor4_sync(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldeor4_sync(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldeor4_sync(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldeor8_relax(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldeor8_relax(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldeor8_relax(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldeor8_relax(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldeor8_acq(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldeor8_acq(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldeor8_acq(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldeor8_acq(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldeor8_rel(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldeor8_rel(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldeor8_rel(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldeor8_rel(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldeor8_acq_rel(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldeor8_acq_rel(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldeor8_acq_rel(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldeor8_acq_rel(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldeor8_sync(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldeor8_sync(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldeor8_sync(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldeor8_sync(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_xor_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldset1_relax(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldset1_relax(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldset1_relax(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldset1_relax(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldset1_acq(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldset1_acq(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldset1_acq(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldset1_acq(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldset1_rel(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldset1_rel(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldset1_rel(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldset1_rel(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldset1_acq_rel(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldset1_acq_rel(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldset1_acq_rel(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldset1_acq_rel(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint8_t ghc___aarch64_ldset1_sync(uint8_t v, uint8_t* p);
|
||||
-uint8_t ghc___aarch64_ldset1_sync(uint8_t v, uint8_t* p) {
|
||||
+uint8_t ghc___aarch64_ldset1_sync(uint8_t v, _Atomic uint8_t* p);
|
||||
+uint8_t ghc___aarch64_ldset1_sync(uint8_t v, _Atomic uint8_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldset2_relax(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldset2_relax(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldset2_relax(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldset2_relax(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldset2_acq(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldset2_acq(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldset2_acq(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldset2_acq(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldset2_rel(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldset2_rel(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldset2_rel(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldset2_rel(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldset2_acq_rel(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldset2_acq_rel(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldset2_acq_rel(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldset2_acq_rel(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint16_t ghc___aarch64_ldset2_sync(uint16_t v, uint16_t* p);
|
||||
-uint16_t ghc___aarch64_ldset2_sync(uint16_t v, uint16_t* p) {
|
||||
+uint16_t ghc___aarch64_ldset2_sync(uint16_t v, _Atomic uint16_t* p);
|
||||
+uint16_t ghc___aarch64_ldset2_sync(uint16_t v, _Atomic uint16_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldset4_relax(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldset4_relax(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldset4_relax(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldset4_relax(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldset4_acq(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldset4_acq(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldset4_acq(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldset4_acq(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldset4_rel(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldset4_rel(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldset4_rel(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldset4_rel(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldset4_acq_rel(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldset4_acq_rel(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldset4_acq_rel(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldset4_acq_rel(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint32_t ghc___aarch64_ldset4_sync(uint32_t v, uint32_t* p);
|
||||
-uint32_t ghc___aarch64_ldset4_sync(uint32_t v, uint32_t* p) {
|
||||
+uint32_t ghc___aarch64_ldset4_sync(uint32_t v, _Atomic uint32_t* p);
|
||||
+uint32_t ghc___aarch64_ldset4_sync(uint32_t v, _Atomic uint32_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldset8_relax(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldset8_relax(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldset8_relax(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldset8_relax(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_relaxed);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldset8_acq(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldset8_acq(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldset8_acq(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldset8_acq(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_acquire);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldset8_rel(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldset8_rel(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldset8_rel(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldset8_rel(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_release);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldset8_acq_rel(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldset8_acq_rel(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldset8_acq_rel(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldset8_acq_rel(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_acq_rel);
|
||||
}
|
||||
|
||||
-uint64_t ghc___aarch64_ldset8_sync(uint64_t v, uint64_t* p);
|
||||
-uint64_t ghc___aarch64_ldset8_sync(uint64_t v, uint64_t* p) {
|
||||
+uint64_t ghc___aarch64_ldset8_sync(uint64_t v, _Atomic uint64_t* p);
|
||||
+uint64_t ghc___aarch64_ldset8_sync(uint64_t v, _Atomic uint64_t* p) {
|
||||
return atomic_fetch_or_explicit(p, v, memory_order_seq_cst);
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user