|  | ; RUN: llc < %s -mtriple=x86_64-apple-darwin -mcpu=corei7     | FileCheck --check-prefix=SSE %s | 
|  | ; RUN: llc < %s -mtriple=x86_64-apple-darwin -mcpu=corei7-avx | FileCheck --check-prefix=AVX %s | 
|  |  | 
|  | ; Every GPR should be saved - except r11 | 
|  | define preserve_mostcc void @preserve_mostcc1() nounwind { | 
|  | entry: | 
|  | ;SSE-LABEL: preserve_mostcc1 | 
|  | ;SSE:       pushq %r10 | 
|  | ;SSE-NEXT:  pushq %r9 | 
|  | ;SSE-NEXT:  pushq %r8 | 
|  | ;SSE-NEXT:  pushq %rdi | 
|  | ;SSE-NEXT:  pushq %rsi | 
|  | ;SSE-NEXT:  pushq %rdx | 
|  | ;SSE-NEXT:  pushq %rcx | 
|  | ;SSE-NEXT:  pushq %rax | 
|  | ;SSE-NEXT:  pushq %rbp | 
|  | ;SSE-NEXT:  pushq %r15 | 
|  | ;SSE-NEXT:  pushq %r14 | 
|  | ;SSE-NEXT:  pushq %r13 | 
|  | ;SSE-NEXT:  pushq %r12 | 
|  | ;SSE-NEXT:  pushq %rbx | 
|  | ;AVX-LABEL: preserve_mostcc1 | 
|  | ;AVX:       pushq %r10 | 
|  | ;AVX-NEXT:  pushq %r9 | 
|  | ;AVX-NEXT:  pushq %r8 | 
|  | ;AVX-NEXT:  pushq %rdi | 
|  | ;AVX-NEXT:  pushq %rsi | 
|  | ;AVX-NEXT:  pushq %rdx | 
|  | ;AVX-NEXT:  pushq %rcx | 
|  | ;AVX-NEXT:  pushq %rax | 
|  | ;AVX-NEXT:  pushq %rbp | 
|  | ;AVX-NEXT:  pushq %r15 | 
|  | ;AVX-NEXT:  pushq %r14 | 
|  | ;AVX-NEXT:  pushq %r13 | 
|  | ;AVX-NEXT:  pushq %r12 | 
|  | ;AVX-NEXT:  pushq %rbx | 
|  | call void asm sideeffect "", "~{rax},~{rbx},~{rcx},~{rdx},~{rsi},~{rdi},~{r8},~{r9},~{r10},~{r11},~{r12},~{r13},~{r14},~{r15},~{rbp},~{xmm0},~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15}"() | 
|  | ret void | 
|  | } | 
|  |  | 
|  | ; Make sure R11 and XMMs are saved before the call | 
|  | declare preserve_mostcc void @foo(i64, i64, double, double) | 
|  | define void @preserve_mostcc2() nounwind { | 
|  | entry: | 
|  | ;SSE-LABEL: preserve_mostcc2 | 
|  | ;SSE:       movq %r11, [[REG:%[a-z0-9]+]] | 
|  | ;SSE:       movaps %xmm2 | 
|  | ;SSE:       movaps %xmm3 | 
|  | ;SSE:       movaps %xmm4 | 
|  | ;SSE:       movaps %xmm5 | 
|  | ;SSE:       movaps %xmm6 | 
|  | ;SSE:       movaps %xmm7 | 
|  | ;SSE:       movaps %xmm8 | 
|  | ;SSE:       movaps %xmm9 | 
|  | ;SSE:       movaps %xmm10 | 
|  | ;SSE:       movaps %xmm11 | 
|  | ;SSE:       movaps %xmm12 | 
|  | ;SSE:       movaps %xmm13 | 
|  | ;SSE:       movaps %xmm14 | 
|  | ;SSE:       movaps %xmm15 | 
|  | ;SSE:       movq [[REG]], %r11 | 
|  | %a0 = call i64 asm sideeffect "", "={rax}"() nounwind | 
|  | %a1 = call i64 asm sideeffect "", "={rcx}"() nounwind | 
|  | %a2 = call i64 asm sideeffect "", "={rdx}"() nounwind | 
|  | %a3 = call i64 asm sideeffect "", "={r8}"() nounwind | 
|  | %a4 = call i64 asm sideeffect "", "={r9}"() nounwind | 
|  | %a5 = call i64 asm sideeffect "", "={r10}"() nounwind | 
|  | %a6 = call i64 asm sideeffect "", "={r11}"() nounwind | 
|  | %a10 = call <2 x double> asm sideeffect "", "={xmm2}"() nounwind | 
|  | %a11 = call <2 x double> asm sideeffect "", "={xmm3}"() nounwind | 
|  | %a12 = call <2 x double> asm sideeffect "", "={xmm4}"() nounwind | 
|  | %a13 = call <2 x double> asm sideeffect "", "={xmm5}"() nounwind | 
|  | %a14 = call <2 x double> asm sideeffect "", "={xmm6}"() nounwind | 
|  | %a15 = call <2 x double> asm sideeffect "", "={xmm7}"() nounwind | 
|  | %a16 = call <2 x double> asm sideeffect "", "={xmm8}"() nounwind | 
|  | %a17 = call <2 x double> asm sideeffect "", "={xmm9}"() nounwind | 
|  | %a18 = call <2 x double> asm sideeffect "", "={xmm10}"() nounwind | 
|  | %a19 = call <2 x double> asm sideeffect "", "={xmm11}"() nounwind | 
|  | %a20 = call <2 x double> asm sideeffect "", "={xmm12}"() nounwind | 
|  | %a21 = call <2 x double> asm sideeffect "", "={xmm13}"() nounwind | 
|  | %a22 = call <2 x double> asm sideeffect "", "={xmm14}"() nounwind | 
|  | %a23 = call <2 x double> asm sideeffect "", "={xmm15}"() nounwind | 
|  | call preserve_mostcc void @foo(i64 1, i64 2, double 3.0, double 4.0) | 
|  | call void asm sideeffect "", "{rax},{rcx},{rdx},{r8},{r9},{r10},{r11},{xmm2},{xmm3},{xmm4},{xmm5},{xmm6},{xmm7},{xmm8},{xmm9},{xmm10},{xmm11},{xmm12},{xmm13},{xmm14},{xmm15}"(i64 %a0, i64 %a1, i64 %a2, i64 %a3, i64 %a4, i64 %a5, i64 %a6, <2 x double> %a10, <2 x double> %a11, <2 x double> %a12, <2 x double> %a13, <2 x double> %a14, <2 x double> %a15, <2 x double> %a16, <2 x double> %a17, <2 x double> %a18, <2 x double> %a19, <2 x double> %a20, <2 x double> %a21, <2 x double> %a22, <2 x double> %a23) | 
|  | ret void | 
|  | } |