|
1 | 1 | ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
2 |
| -; RUN: llc -mtriple=riscv64 -mattr=+experimental-v -verify-machineinstrs < %s | FileCheck %s |
| 2 | +; RUN: llc -mtriple=riscv32 -mattr=+experimental-v -verify-machineinstrs < %s | FileCheck %s --check-prefixes=CHECK,RV32 |
| 3 | +; RUN: llc -mtriple=riscv64 -mattr=+experimental-v -verify-machineinstrs < %s | FileCheck %s --check-prefixes=CHECK,RV64 |
3 | 4 |
|
4 | 5 | define <vscale x 1 x i8> @vadd_vx_nxv1i8(<vscale x 1 x i8> %va, i8 signext %b) {
|
5 | 6 | ; CHECK-LABEL: vadd_vx_nxv1i8:
|
@@ -665,11 +666,24 @@ define <vscale x 16 x i32> @vadd_vx_nxv16i32_1(<vscale x 16 x i32> %va) {
|
665 | 666 | }
|
666 | 667 |
|
667 | 668 | define <vscale x 1 x i64> @vadd_vx_nxv1i64(<vscale x 1 x i64> %va, i64 %b) {
|
668 |
| -; CHECK-LABEL: vadd_vx_nxv1i64: |
669 |
| -; CHECK: # %bb.0: |
670 |
| -; CHECK-NEXT: vsetvli a1, zero, e64, m1, ta, mu |
671 |
| -; CHECK-NEXT: vadd.vx v8, v8, a0 |
672 |
| -; CHECK-NEXT: ret |
| 669 | +; RV32-LABEL: vadd_vx_nxv1i64: |
| 670 | +; RV32: # %bb.0: |
| 671 | +; RV32-NEXT: addi sp, sp, -16 |
| 672 | +; RV32-NEXT: .cfi_def_cfa_offset 16 |
| 673 | +; RV32-NEXT: sw a1, 12(sp) |
| 674 | +; RV32-NEXT: sw a0, 8(sp) |
| 675 | +; RV32-NEXT: vsetvli a0, zero, e64, m1, ta, mu |
| 676 | +; RV32-NEXT: addi a0, sp, 8 |
| 677 | +; RV32-NEXT: vlse64.v v9, (a0), zero |
| 678 | +; RV32-NEXT: vadd.vv v8, v8, v9 |
| 679 | +; RV32-NEXT: addi sp, sp, 16 |
| 680 | +; RV32-NEXT: ret |
| 681 | +; |
| 682 | +; RV64-LABEL: vadd_vx_nxv1i64: |
| 683 | +; RV64: # %bb.0: |
| 684 | +; RV64-NEXT: vsetvli a1, zero, e64, m1, ta, mu |
| 685 | +; RV64-NEXT: vadd.vx v8, v8, a0 |
| 686 | +; RV64-NEXT: ret |
673 | 687 | %head = insertelement <vscale x 1 x i64> undef, i64 %b, i32 0
|
674 | 688 | %splat = shufflevector <vscale x 1 x i64> %head, <vscale x 1 x i64> undef, <vscale x 1 x i32> zeroinitializer
|
675 | 689 | %vc = add <vscale x 1 x i64> %va, %splat
|
@@ -701,11 +715,24 @@ define <vscale x 1 x i64> @vadd_vx_nxv1i64_1(<vscale x 1 x i64> %va) {
|
701 | 715 | }
|
702 | 716 |
|
703 | 717 | define <vscale x 2 x i64> @vadd_vx_nxv2i64(<vscale x 2 x i64> %va, i64 %b) {
|
704 |
| -; CHECK-LABEL: vadd_vx_nxv2i64: |
705 |
| -; CHECK: # %bb.0: |
706 |
| -; CHECK-NEXT: vsetvli a1, zero, e64, m2, ta, mu |
707 |
| -; CHECK-NEXT: vadd.vx v8, v8, a0 |
708 |
| -; CHECK-NEXT: ret |
| 718 | +; RV32-LABEL: vadd_vx_nxv2i64: |
| 719 | +; RV32: # %bb.0: |
| 720 | +; RV32-NEXT: addi sp, sp, -16 |
| 721 | +; RV32-NEXT: .cfi_def_cfa_offset 16 |
| 722 | +; RV32-NEXT: sw a1, 12(sp) |
| 723 | +; RV32-NEXT: sw a0, 8(sp) |
| 724 | +; RV32-NEXT: vsetvli a0, zero, e64, m2, ta, mu |
| 725 | +; RV32-NEXT: addi a0, sp, 8 |
| 726 | +; RV32-NEXT: vlse64.v v10, (a0), zero |
| 727 | +; RV32-NEXT: vadd.vv v8, v8, v10 |
| 728 | +; RV32-NEXT: addi sp, sp, 16 |
| 729 | +; RV32-NEXT: ret |
| 730 | +; |
| 731 | +; RV64-LABEL: vadd_vx_nxv2i64: |
| 732 | +; RV64: # %bb.0: |
| 733 | +; RV64-NEXT: vsetvli a1, zero, e64, m2, ta, mu |
| 734 | +; RV64-NEXT: vadd.vx v8, v8, a0 |
| 735 | +; RV64-NEXT: ret |
709 | 736 | %head = insertelement <vscale x 2 x i64> undef, i64 %b, i32 0
|
710 | 737 | %splat = shufflevector <vscale x 2 x i64> %head, <vscale x 2 x i64> undef, <vscale x 2 x i32> zeroinitializer
|
711 | 738 | %vc = add <vscale x 2 x i64> %va, %splat
|
@@ -737,11 +764,24 @@ define <vscale x 2 x i64> @vadd_vx_nxv2i64_1(<vscale x 2 x i64> %va) {
|
737 | 764 | }
|
738 | 765 |
|
739 | 766 | define <vscale x 4 x i64> @vadd_vx_nxv4i64(<vscale x 4 x i64> %va, i64 %b) {
|
740 |
| -; CHECK-LABEL: vadd_vx_nxv4i64: |
741 |
| -; CHECK: # %bb.0: |
742 |
| -; CHECK-NEXT: vsetvli a1, zero, e64, m4, ta, mu |
743 |
| -; CHECK-NEXT: vadd.vx v8, v8, a0 |
744 |
| -; CHECK-NEXT: ret |
| 767 | +; RV32-LABEL: vadd_vx_nxv4i64: |
| 768 | +; RV32: # %bb.0: |
| 769 | +; RV32-NEXT: addi sp, sp, -16 |
| 770 | +; RV32-NEXT: .cfi_def_cfa_offset 16 |
| 771 | +; RV32-NEXT: sw a1, 12(sp) |
| 772 | +; RV32-NEXT: sw a0, 8(sp) |
| 773 | +; RV32-NEXT: vsetvli a0, zero, e64, m4, ta, mu |
| 774 | +; RV32-NEXT: addi a0, sp, 8 |
| 775 | +; RV32-NEXT: vlse64.v v12, (a0), zero |
| 776 | +; RV32-NEXT: vadd.vv v8, v8, v12 |
| 777 | +; RV32-NEXT: addi sp, sp, 16 |
| 778 | +; RV32-NEXT: ret |
| 779 | +; |
| 780 | +; RV64-LABEL: vadd_vx_nxv4i64: |
| 781 | +; RV64: # %bb.0: |
| 782 | +; RV64-NEXT: vsetvli a1, zero, e64, m4, ta, mu |
| 783 | +; RV64-NEXT: vadd.vx v8, v8, a0 |
| 784 | +; RV64-NEXT: ret |
745 | 785 | %head = insertelement <vscale x 4 x i64> undef, i64 %b, i32 0
|
746 | 786 | %splat = shufflevector <vscale x 4 x i64> %head, <vscale x 4 x i64> undef, <vscale x 4 x i32> zeroinitializer
|
747 | 787 | %vc = add <vscale x 4 x i64> %va, %splat
|
@@ -773,11 +813,24 @@ define <vscale x 4 x i64> @vadd_vx_nxv4i64_1(<vscale x 4 x i64> %va) {
|
773 | 813 | }
|
774 | 814 |
|
775 | 815 | define <vscale x 8 x i64> @vadd_vx_nxv8i64(<vscale x 8 x i64> %va, i64 %b) {
|
776 |
| -; CHECK-LABEL: vadd_vx_nxv8i64: |
777 |
| -; CHECK: # %bb.0: |
778 |
| -; CHECK-NEXT: vsetvli a1, zero, e64, m8, ta, mu |
779 |
| -; CHECK-NEXT: vadd.vx v8, v8, a0 |
780 |
| -; CHECK-NEXT: ret |
| 816 | +; RV32-LABEL: vadd_vx_nxv8i64: |
| 817 | +; RV32: # %bb.0: |
| 818 | +; RV32-NEXT: addi sp, sp, -16 |
| 819 | +; RV32-NEXT: .cfi_def_cfa_offset 16 |
| 820 | +; RV32-NEXT: sw a1, 12(sp) |
| 821 | +; RV32-NEXT: sw a0, 8(sp) |
| 822 | +; RV32-NEXT: vsetvli a0, zero, e64, m8, ta, mu |
| 823 | +; RV32-NEXT: addi a0, sp, 8 |
| 824 | +; RV32-NEXT: vlse64.v v16, (a0), zero |
| 825 | +; RV32-NEXT: vadd.vv v8, v8, v16 |
| 826 | +; RV32-NEXT: addi sp, sp, 16 |
| 827 | +; RV32-NEXT: ret |
| 828 | +; |
| 829 | +; RV64-LABEL: vadd_vx_nxv8i64: |
| 830 | +; RV64: # %bb.0: |
| 831 | +; RV64-NEXT: vsetvli a1, zero, e64, m8, ta, mu |
| 832 | +; RV64-NEXT: vadd.vx v8, v8, a0 |
| 833 | +; RV64-NEXT: ret |
781 | 834 | %head = insertelement <vscale x 8 x i64> undef, i64 %b, i32 0
|
782 | 835 | %splat = shufflevector <vscale x 8 x i64> %head, <vscale x 8 x i64> undef, <vscale x 8 x i32> zeroinitializer
|
783 | 836 | %vc = add <vscale x 8 x i64> %va, %splat
|
|
0 commit comments