1 ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
2 ; RUN: llc < %s -mtriple=i686-apple-darwin8 -mcpu=yonah | FileCheck %s
4 define void @test(<1 x i64> %c64, <1 x i64> %mask1, i8* %P) {
6 ; CHECK: ## %bb.0: ## %entry
7 ; CHECK-NEXT: pushl %edi
8 ; CHECK-NEXT: .cfi_def_cfa_offset 8
9 ; CHECK-NEXT: subl $16, %esp
10 ; CHECK-NEXT: .cfi_def_cfa_offset 24
11 ; CHECK-NEXT: .cfi_offset %edi, -8
12 ; CHECK-NEXT: movl {{[0-9]+}}(%esp), %eax
13 ; CHECK-NEXT: movl {{[0-9]+}}(%esp), %ecx
14 ; CHECK-NEXT: movl %ecx, {{[0-9]+}}(%esp)
15 ; CHECK-NEXT: movl %eax, (%esp)
16 ; CHECK-NEXT: movl {{[0-9]+}}(%esp), %eax
17 ; CHECK-NEXT: movl {{[0-9]+}}(%esp), %ecx
18 ; CHECK-NEXT: movl %ecx, {{[0-9]+}}(%esp)
19 ; CHECK-NEXT: movl %eax, {{[0-9]+}}(%esp)
20 ; CHECK-NEXT: movl {{[0-9]+}}(%esp), %edi
21 ; CHECK-NEXT: movq (%esp), %mm0
22 ; CHECK-NEXT: movq {{[0-9]+}}(%esp), %mm1
23 ; CHECK-NEXT: maskmovq %mm0, %mm1
24 ; CHECK-NEXT: addl $16, %esp
25 ; CHECK-NEXT: popl %edi
28 %tmp4 = bitcast <1 x i64> %mask1 to x86_mmx ; <x86_mmx> [#uses=1]
29 %tmp6 = bitcast <1 x i64> %c64 to x86_mmx ; <x86_mmx> [#uses=1]
30 tail call void @llvm.x86.mmx.maskmovq( x86_mmx %tmp4, x86_mmx %tmp6, i8* %P )
34 declare void @llvm.x86.mmx.maskmovq(x86_mmx, x86_mmx, i8*)