|  | ; RUN: llc < %s -mtriple=powerpc-apple-darwin -verify-machineinstrs | FileCheck %s --check-prefix=CHECK --check-prefix=PPC32 | 
|  | ; FIXME: -verify-machineinstrs currently fail on ppc64 (mismatched register/instruction). | 
|  | ; This is already checked for in Atomics-64.ll | 
|  | ; RUN: llc < %s -mtriple=powerpc64-apple-darwin | FileCheck %s --check-prefix=CHECK --check-prefix=PPC64 | 
|  |  | 
|  | ; FIXME: we don't currently check for the operations themselves with CHECK-NEXT, | 
|  | ;   because they are implemented in a very messy way with lwarx/stwcx. | 
|  | ;   It should be fixed soon in another patch. | 
|  |  | 
|  | ; We first check loads, for all sizes from i8 to i64. | 
|  | ; We also vary orderings to check for barriers. | 
|  | define i8 @load_i8_unordered(i8* %mem) { | 
|  | ; CHECK-LABEL: load_i8_unordered | 
|  | ; CHECK: lbz | 
|  | ; CHECK-NOT: sync | 
|  | %val = load atomic i8, i8* %mem unordered, align 1 | 
|  | ret i8 %val | 
|  | } | 
|  | define i16 @load_i16_monotonic(i16* %mem) { | 
|  | ; CHECK-LABEL: load_i16_monotonic | 
|  | ; CHECK: lhz | 
|  | ; CHECK-NOT: sync | 
|  | %val = load atomic i16, i16* %mem monotonic, align 2 | 
|  | ret i16 %val | 
|  | } | 
|  | define i32 @load_i32_acquire(i32* %mem) { | 
|  | ; CHECK-LABEL: load_i32_acquire | 
|  | ; CHECK: lwz [[VAL:r[0-9]+]] | 
|  | %val = load atomic i32, i32* %mem acquire, align 4 | 
|  | ; CHECK-PPC32: lwsync | 
|  | ; CHECK-PPC64: cmpw [[CR:cr[0-9]+]], [[VAL]], [[VAL]] | 
|  | ; CHECK-PPC64: bne- [[CR]], .+4 | 
|  | ; CHECK-PPC64: isync | 
|  | ret i32 %val | 
|  | } | 
|  | define i64 @load_i64_seq_cst(i64* %mem) { | 
|  | ; CHECK-LABEL: load_i64_seq_cst | 
|  | ; CHECK: sync | 
|  | ; PPC32: __sync_ | 
|  | ; PPC64-NOT: __sync_ | 
|  | ; PPC64: ld [[VAL:r[0-9]+]] | 
|  | %val = load atomic i64, i64* %mem seq_cst, align 8 | 
|  | ; CHECK-PPC32: lwsync | 
|  | ; CHECK-PPC64: cmpw [[CR:cr[0-9]+]], [[VAL]], [[VAL]] | 
|  | ; CHECK-PPC64: bne- [[CR]], .+4 | 
|  | ; CHECK-PPC64: isync | 
|  | ret i64 %val | 
|  | } | 
|  |  | 
|  | ; Stores | 
|  | define void @store_i8_unordered(i8* %mem) { | 
|  | ; CHECK-LABEL: store_i8_unordered | 
|  | ; CHECK-NOT: sync | 
|  | ; CHECK: stb | 
|  | store atomic i8 42, i8* %mem unordered, align 1 | 
|  | ret void | 
|  | } | 
|  | define void @store_i16_monotonic(i16* %mem) { | 
|  | ; CHECK-LABEL: store_i16_monotonic | 
|  | ; CHECK-NOT: sync | 
|  | ; CHECK: sth | 
|  | store atomic i16 42, i16* %mem monotonic, align 2 | 
|  | ret void | 
|  | } | 
|  | define void @store_i32_release(i32* %mem) { | 
|  | ; CHECK-LABEL: store_i32_release | 
|  | ; CHECK: lwsync | 
|  | ; CHECK: stw | 
|  | store atomic i32 42, i32* %mem release, align 4 | 
|  | ret void | 
|  | } | 
|  | define void @store_i64_seq_cst(i64* %mem) { | 
|  | ; CHECK-LABEL: store_i64_seq_cst | 
|  | ; CHECK: sync | 
|  | ; PPC32: __sync_ | 
|  | ; PPC64-NOT: __sync_ | 
|  | ; PPC64: std | 
|  | store atomic i64 42, i64* %mem seq_cst, align 8 | 
|  | ret void | 
|  | } | 
|  |  | 
|  | ; Atomic CmpXchg | 
|  | define i8 @cas_strong_i8_sc_sc(i8* %mem) { | 
|  | ; CHECK-LABEL: cas_strong_i8_sc_sc | 
|  | ; CHECK: sync | 
|  | %val = cmpxchg i8* %mem, i8 0, i8 1 seq_cst seq_cst | 
|  | ; CHECK: lwsync | 
|  | %loaded = extractvalue { i8, i1} %val, 0 | 
|  | ret i8 %loaded | 
|  | } | 
|  | define i16 @cas_weak_i16_acquire_acquire(i16* %mem) { | 
|  | ; CHECK-LABEL: cas_weak_i16_acquire_acquire | 
|  | ;CHECK-NOT: sync | 
|  | %val = cmpxchg weak i16* %mem, i16 0, i16 1 acquire acquire | 
|  | ; CHECK: lwsync | 
|  | %loaded = extractvalue { i16, i1} %val, 0 | 
|  | ret i16 %loaded | 
|  | } | 
|  | define i32 @cas_strong_i32_acqrel_acquire(i32* %mem) { | 
|  | ; CHECK-LABEL: cas_strong_i32_acqrel_acquire | 
|  | ; CHECK: lwsync | 
|  | %val = cmpxchg i32* %mem, i32 0, i32 1 acq_rel acquire | 
|  | ; CHECK: lwsync | 
|  | %loaded = extractvalue { i32, i1} %val, 0 | 
|  | ret i32 %loaded | 
|  | } | 
|  | define i64 @cas_weak_i64_release_monotonic(i64* %mem) { | 
|  | ; CHECK-LABEL: cas_weak_i64_release_monotonic | 
|  | ; CHECK: lwsync | 
|  | %val = cmpxchg weak i64* %mem, i64 0, i64 1 release monotonic | 
|  | ; CHECK-NOT: [sync ] | 
|  | %loaded = extractvalue { i64, i1} %val, 0 | 
|  | ret i64 %loaded | 
|  | } | 
|  |  | 
|  | ; AtomicRMW | 
|  | define i8 @add_i8_monotonic(i8* %mem, i8 %operand) { | 
|  | ; CHECK-LABEL: add_i8_monotonic | 
|  | ; CHECK-NOT: sync | 
|  | %val = atomicrmw add i8* %mem, i8 %operand monotonic | 
|  | ret i8 %val | 
|  | } | 
|  | define i16 @xor_i16_seq_cst(i16* %mem, i16 %operand) { | 
|  | ; CHECK-LABEL: xor_i16_seq_cst | 
|  | ; CHECK: sync | 
|  | %val = atomicrmw xor i16* %mem, i16 %operand seq_cst | 
|  | ; CHECK: lwsync | 
|  | ret i16 %val | 
|  | } | 
|  | define i32 @xchg_i32_acq_rel(i32* %mem, i32 %operand) { | 
|  | ; CHECK-LABEL: xchg_i32_acq_rel | 
|  | ; CHECK: lwsync | 
|  | %val = atomicrmw xchg i32* %mem, i32 %operand acq_rel | 
|  | ; CHECK: lwsync | 
|  | ret i32 %val | 
|  | } | 
|  | define i64 @and_i64_release(i64* %mem, i64 %operand) { | 
|  | ; CHECK-LABEL: and_i64_release | 
|  | ; CHECK: lwsync | 
|  | %val = atomicrmw and i64* %mem, i64 %operand release | 
|  | ; CHECK-NOT: [sync ] | 
|  | ret i64 %val | 
|  | } |