| ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 4 |
| ; RUN: llc -verify-machineinstrs -force-streaming < %s | FileCheck %s |
| |
| target triple = "aarch64-linux" |
| |
| define void @test_svzero_za64_vg1x2(i32 %slice) #0 { |
| ; CHECK-LABEL: test_svzero_za64_vg1x2: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: mov w8, w0 |
| ; CHECK-NEXT: zero za.d[w8, 0, vgx2] |
| ; CHECK-NEXT: ret |
| entry: |
| tail call void @llvm.aarch64.sme.zero.za64.vg1x2(i32 %slice) |
| ret void |
| } |
| |
| define void @test_svzero_za64_vg1x2_offset(i32 %slice) #0 { |
| ; CHECK-LABEL: test_svzero_za64_vg1x2_offset: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: mov w8, w0 |
| ; CHECK-NEXT: zero za.d[w8, 7, vgx2] |
| ; CHECK-NEXT: ret |
| entry: |
| %slice.max = add i32 %slice, 7 |
| tail call void @llvm.aarch64.sme.zero.za64.vg1x2(i32 %slice.max) |
| ret void |
| } |
| |
| define void @test_svzero_za64_vg1x4(i32 %slice) #0 { |
| ; CHECK-LABEL: test_svzero_za64_vg1x4: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: mov w8, w0 |
| ; CHECK-NEXT: zero za.d[w8, 0, vgx4] |
| ; CHECK-NEXT: ret |
| entry: |
| tail call void @llvm.aarch64.sme.zero.za64.vg1x4(i32 %slice) |
| ret void |
| } |
| |
| define void @test_svzero_za64_vg1x4_offset(i32 %slice) #0 { |
| ; CHECK-LABEL: test_svzero_za64_vg1x4_offset: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: mov w8, w0 |
| ; CHECK-NEXT: zero za.d[w8, 1, vgx4] |
| ; CHECK-NEXT: ret |
| entry: |
| %slice.min = add i32 %slice, 1 |
| tail call void @llvm.aarch64.sme.zero.za64.vg1x4(i32 %slice.min) |
| ret void |
| } |
| |
| define void @test_svzero_za64_vg2x1(i32 %slice) #0 { |
| ; CHECK-LABEL: test_svzero_za64_vg2x1: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: mov w8, w0 |
| ; CHECK-NEXT: zero za.d[w8, 0:1] |
| ; CHECK-NEXT: ret |
| entry: |
| tail call void @llvm.aarch64.sme.zero.za64.vg2x1(i32 %slice) |
| ret void |
| } |
| |
| define void @test_svzero_za64_vg2x1_offset(i32 %slice) #0 { |
| ; CHECK-LABEL: test_svzero_za64_vg2x1_offset: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: mov w8, w0 |
| ; CHECK-NEXT: zero za.d[w8, 6:7] |
| ; CHECK-NEXT: ret |
| entry: |
| %slice.max = add i32 %slice, 6 |
| tail call void @llvm.aarch64.sme.zero.za64.vg2x1(i32 %slice.max) |
| ret void |
| } |
| |
| define void @test_svzero_za64_vg2x2(i32 %slice) #0 { |
| ; CHECK-LABEL: test_svzero_za64_vg2x2: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: mov w8, w0 |
| ; CHECK-NEXT: zero za.d[w8, 0:1, vgx2] |
| ; CHECK-NEXT: ret |
| entry: |
| tail call void @llvm.aarch64.sme.zero.za64.vg2x2(i32 %slice) |
| ret void |
| } |
| |
| define void @test_svzero_za64_vg2x2_offset(i32 %slice) #0 { |
| ; CHECK-LABEL: test_svzero_za64_vg2x2_offset: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: mov w8, w0 |
| ; CHECK-NEXT: zero za.d[w8, 2:3, vgx2] |
| ; CHECK-NEXT: ret |
| entry: |
| %slice.max = add i32 %slice, 2 |
| tail call void @llvm.aarch64.sme.zero.za64.vg2x2(i32 %slice.max) |
| ret void |
| } |
| |
| define void @test_svzero_za64_vg2x4(i32 %slice) #0 { |
| ; CHECK-LABEL: test_svzero_za64_vg2x4: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: mov w8, w0 |
| ; CHECK-NEXT: zero za.d[w8, 0:1, vgx4] |
| ; CHECK-NEXT: ret |
| entry: |
| tail call void @llvm.aarch64.sme.zero.za64.vg2x4(i32 %slice) |
| ret void |
| } |
| |
| define void @test_svzero_za64_vg2x4_offset(i32 %slice) #0 { |
| ; CHECK-LABEL: test_svzero_za64_vg2x4_offset: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: add w8, w0, #1 |
| ; CHECK-NEXT: zero za.d[w8, 0:1, vgx4] |
| ; CHECK-NEXT: ret |
| entry: |
| %slice.min = add i32 %slice, 1 |
| tail call void @llvm.aarch64.sme.zero.za64.vg2x4(i32 %slice.min) |
| ret void |
| } |
| |
| define void @test_svzero_za64_vg4x1(i32 %slice) #0 { |
| ; CHECK-LABEL: test_svzero_za64_vg4x1: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: mov w8, w0 |
| ; CHECK-NEXT: zero za.d[w8, 0:3] |
| ; CHECK-NEXT: ret |
| entry: |
| tail call void @llvm.aarch64.sme.zero.za64.vg4x1(i32 %slice) |
| ret void |
| } |
| |
| define void @test_svzero_za64_vg4x1_offset(i32 %slice) #0 { |
| ; CHECK-LABEL: test_svzero_za64_vg4x1_offset: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: mov w8, w0 |
| ; CHECK-NEXT: zero za.d[w8, 4:7] |
| ; CHECK-NEXT: ret |
| entry: |
| %slice.max = add i32 %slice, 4 |
| tail call void @llvm.aarch64.sme.zero.za64.vg4x1(i32 %slice.max) |
| ret void |
| } |
| |
| define void @test_svzero_za64_vg4x2(i32 %slice) #0 { |
| ; CHECK-LABEL: test_svzero_za64_vg4x2: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: mov w8, w0 |
| ; CHECK-NEXT: zero za.d[w8, 0:3, vgx2] |
| ; CHECK-NEXT: ret |
| entry: |
| tail call void @llvm.aarch64.sme.zero.za64.vg4x2(i32 %slice) |
| ret void |
| } |
| |
| define void @test_svzero_za64_vg4x2_offset(i32 %slice) #0 { |
| ; CHECK-LABEL: test_svzero_za64_vg4x2_offset: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: mov w8, w0 |
| ; CHECK-NEXT: zero za.d[w8, 0:3, vgx2] |
| ; CHECK-NEXT: ret |
| entry: |
| %slice.max = add i32 %slice, 0 |
| tail call void @llvm.aarch64.sme.zero.za64.vg4x2(i32 %slice.max) |
| ret void |
| } |
| |
| define void @test_svzero_za64_vg4x4(i32 %slice) #0 { |
| ; CHECK-LABEL: test_svzero_za64_vg4x4: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: mov w8, w0 |
| ; CHECK-NEXT: zero za.d[w8, 0:3, vgx4] |
| ; CHECK-NEXT: ret |
| entry: |
| tail call void @llvm.aarch64.sme.zero.za64.vg4x4(i32 %slice) |
| ret void |
| } |
| |
| define void @test_svzero_za64_vg4x4_offset(i32 %slice) #0 { |
| ; CHECK-LABEL: test_svzero_za64_vg4x4_offset: |
| ; CHECK: // %bb.0: // %entry |
| ; CHECK-NEXT: add w8, w0, #1 |
| ; CHECK-NEXT: zero za.d[w8, 0:3, vgx4] |
| ; CHECK-NEXT: ret |
| entry: |
| %slice.min = add i32 %slice, 1 |
| tail call void @llvm.aarch64.sme.zero.za64.vg4x4(i32 %slice.min) |
| ret void |
| } |
| |
| attributes #0 = { nounwind "target-features" = "+sme2p1"} |