@@ -1504,9 +1504,11 @@ entry:
1504
1504
define <vscale x 1 x i1 > @intrinsic_vmslt_mask_vi_nxv1i8_i8 (<vscale x 1 x i1 > %0 , <vscale x 1 x i8 > %1 , <vscale x 1 x i1 > %2 , i32 %3 ) nounwind {
1505
1505
; CHECK-LABEL: intrinsic_vmslt_mask_vi_nxv1i8_i8:
1506
1506
; CHECK: # %bb.0: # %entry
1507
+ ; CHECK-NEXT: vmv1r.v v25, v0
1507
1508
; CHECK-NEXT: vsetvli a0, a0, e8,mf8,tu,mu
1508
- ; CHECK-NEXT: vmsle.vi v9, v8, -15, v0.t
1509
1509
; CHECK-NEXT: vmv1r.v v0, v9
1510
+ ; CHECK-NEXT: vmsle.vi v25, v8, -15, v0.t
1511
+ ; CHECK-NEXT: vmv1r.v v0, v25
1510
1512
; CHECK-NEXT: jalr zero, 0(ra)
1511
1513
entry:
1512
1514
%a = call <vscale x 1 x i1 > @llvm.riscv.vmslt.mask.nxv1i8.i8 (
@@ -1537,9 +1539,11 @@ entry:
1537
1539
define <vscale x 2 x i1 > @intrinsic_vmslt_mask_vi_nxv2i8_i8 (<vscale x 2 x i1 > %0 , <vscale x 2 x i8 > %1 , <vscale x 2 x i1 > %2 , i32 %3 ) nounwind {
1538
1540
; CHECK-LABEL: intrinsic_vmslt_mask_vi_nxv2i8_i8:
1539
1541
; CHECK: # %bb.0: # %entry
1542
+ ; CHECK-NEXT: vmv1r.v v25, v0
1540
1543
; CHECK-NEXT: vsetvli a0, a0, e8,mf4,tu,mu
1541
- ; CHECK-NEXT: vmsle.vi v9, v8, -13, v0.t
1542
1544
; CHECK-NEXT: vmv1r.v v0, v9
1545
+ ; CHECK-NEXT: vmsle.vi v25, v8, -13, v0.t
1546
+ ; CHECK-NEXT: vmv1r.v v0, v25
1543
1547
; CHECK-NEXT: jalr zero, 0(ra)
1544
1548
entry:
1545
1549
%a = call <vscale x 2 x i1 > @llvm.riscv.vmslt.mask.nxv2i8.i8 (
@@ -1570,9 +1574,11 @@ entry:
1570
1574
define <vscale x 4 x i1 > @intrinsic_vmslt_mask_vi_nxv4i8_i8 (<vscale x 4 x i1 > %0 , <vscale x 4 x i8 > %1 , <vscale x 4 x i1 > %2 , i32 %3 ) nounwind {
1571
1575
; CHECK-LABEL: intrinsic_vmslt_mask_vi_nxv4i8_i8:
1572
1576
; CHECK: # %bb.0: # %entry
1577
+ ; CHECK-NEXT: vmv1r.v v25, v0
1573
1578
; CHECK-NEXT: vsetvli a0, a0, e8,mf2,tu,mu
1574
- ; CHECK-NEXT: vmsle.vi v9, v8, -11, v0.t
1575
1579
; CHECK-NEXT: vmv1r.v v0, v9
1580
+ ; CHECK-NEXT: vmsle.vi v25, v8, -11, v0.t
1581
+ ; CHECK-NEXT: vmv1r.v v0, v25
1576
1582
; CHECK-NEXT: jalr zero, 0(ra)
1577
1583
entry:
1578
1584
%a = call <vscale x 4 x i1 > @llvm.riscv.vmslt.mask.nxv4i8.i8 (
@@ -1603,9 +1609,11 @@ entry:
1603
1609
define <vscale x 8 x i1 > @intrinsic_vmslt_mask_vi_nxv8i8_i8 (<vscale x 8 x i1 > %0 , <vscale x 8 x i8 > %1 , <vscale x 8 x i1 > %2 , i32 %3 ) nounwind {
1604
1610
; CHECK-LABEL: intrinsic_vmslt_mask_vi_nxv8i8_i8:
1605
1611
; CHECK: # %bb.0: # %entry
1612
+ ; CHECK-NEXT: vmv1r.v v25, v0
1606
1613
; CHECK-NEXT: vsetvli a0, a0, e8,m1,tu,mu
1607
- ; CHECK-NEXT: vmsle.vi v9, v8, -9, v0.t
1608
1614
; CHECK-NEXT: vmv1r.v v0, v9
1615
+ ; CHECK-NEXT: vmsle.vi v25, v8, -9, v0.t
1616
+ ; CHECK-NEXT: vmv1r.v v0, v25
1609
1617
; CHECK-NEXT: jalr zero, 0(ra)
1610
1618
entry:
1611
1619
%a = call <vscale x 8 x i1 > @llvm.riscv.vmslt.mask.nxv8i8.i8 (
@@ -1636,9 +1644,11 @@ entry:
1636
1644
define <vscale x 16 x i1 > @intrinsic_vmslt_mask_vi_nxv16i8_i8 (<vscale x 16 x i1 > %0 , <vscale x 16 x i8 > %1 , <vscale x 16 x i1 > %2 , i32 %3 ) nounwind {
1637
1645
; CHECK-LABEL: intrinsic_vmslt_mask_vi_nxv16i8_i8:
1638
1646
; CHECK: # %bb.0: # %entry
1647
+ ; CHECK-NEXT: vmv1r.v v25, v0
1639
1648
; CHECK-NEXT: vsetvli a0, a0, e8,m2,tu,mu
1640
- ; CHECK-NEXT: vmsle.vi v10, v8, -7, v0.t
1641
1649
; CHECK-NEXT: vmv1r.v v0, v10
1650
+ ; CHECK-NEXT: vmsle.vi v25, v8, -7, v0.t
1651
+ ; CHECK-NEXT: vmv1r.v v0, v25
1642
1652
; CHECK-NEXT: jalr zero, 0(ra)
1643
1653
entry:
1644
1654
%a = call <vscale x 16 x i1 > @llvm.riscv.vmslt.mask.nxv16i8.i8 (
@@ -1669,9 +1679,11 @@ entry:
1669
1679
define <vscale x 32 x i1 > @intrinsic_vmslt_mask_vi_nxv32i8_i8 (<vscale x 32 x i1 > %0 , <vscale x 32 x i8 > %1 , <vscale x 32 x i1 > %2 , i32 %3 ) nounwind {
1670
1680
; CHECK-LABEL: intrinsic_vmslt_mask_vi_nxv32i8_i8:
1671
1681
; CHECK: # %bb.0: # %entry
1682
+ ; CHECK-NEXT: vmv1r.v v25, v0
1672
1683
; CHECK-NEXT: vsetvli a0, a0, e8,m4,tu,mu
1673
- ; CHECK-NEXT: vmsle.vi v12, v8, -5, v0.t
1674
1684
; CHECK-NEXT: vmv1r.v v0, v12
1685
+ ; CHECK-NEXT: vmsle.vi v25, v8, -5, v0.t
1686
+ ; CHECK-NEXT: vmv1r.v v0, v25
1675
1687
; CHECK-NEXT: jalr zero, 0(ra)
1676
1688
entry:
1677
1689
%a = call <vscale x 32 x i1 > @llvm.riscv.vmslt.mask.nxv32i8.i8 (
@@ -1702,9 +1714,11 @@ entry:
1702
1714
define <vscale x 1 x i1 > @intrinsic_vmslt_mask_vi_nxv1i16_i16 (<vscale x 1 x i1 > %0 , <vscale x 1 x i16 > %1 , <vscale x 1 x i1 > %2 , i32 %3 ) nounwind {
1703
1715
; CHECK-LABEL: intrinsic_vmslt_mask_vi_nxv1i16_i16:
1704
1716
; CHECK: # %bb.0: # %entry
1717
+ ; CHECK-NEXT: vmv1r.v v25, v0
1705
1718
; CHECK-NEXT: vsetvli a0, a0, e16,mf4,tu,mu
1706
- ; CHECK-NEXT: vmsle.vi v9, v8, -3, v0.t
1707
1719
; CHECK-NEXT: vmv1r.v v0, v9
1720
+ ; CHECK-NEXT: vmsle.vi v25, v8, -3, v0.t
1721
+ ; CHECK-NEXT: vmv1r.v v0, v25
1708
1722
; CHECK-NEXT: jalr zero, 0(ra)
1709
1723
entry:
1710
1724
%a = call <vscale x 1 x i1 > @llvm.riscv.vmslt.mask.nxv1i16.i16 (
@@ -1735,9 +1749,11 @@ entry:
1735
1749
define <vscale x 2 x i1 > @intrinsic_vmslt_mask_vi_nxv2i16_i16 (<vscale x 2 x i1 > %0 , <vscale x 2 x i16 > %1 , <vscale x 2 x i1 > %2 , i32 %3 ) nounwind {
1736
1750
; CHECK-LABEL: intrinsic_vmslt_mask_vi_nxv2i16_i16:
1737
1751
; CHECK: # %bb.0: # %entry
1752
+ ; CHECK-NEXT: vmv1r.v v25, v0
1738
1753
; CHECK-NEXT: vsetvli a0, a0, e16,mf2,tu,mu
1739
- ; CHECK-NEXT: vmsle.vi v9, v8, -1, v0.t
1740
1754
; CHECK-NEXT: vmv1r.v v0, v9
1755
+ ; CHECK-NEXT: vmsle.vi v25, v8, -1, v0.t
1756
+ ; CHECK-NEXT: vmv1r.v v0, v25
1741
1757
; CHECK-NEXT: jalr zero, 0(ra)
1742
1758
entry:
1743
1759
%a = call <vscale x 2 x i1 > @llvm.riscv.vmslt.mask.nxv2i16.i16 (
@@ -1768,9 +1784,11 @@ entry:
1768
1784
define <vscale x 4 x i1 > @intrinsic_vmslt_mask_vi_nxv4i16_i16 (<vscale x 4 x i1 > %0 , <vscale x 4 x i16 > %1 , <vscale x 4 x i1 > %2 , i32 %3 ) nounwind {
1769
1785
; CHECK-LABEL: intrinsic_vmslt_mask_vi_nxv4i16_i16:
1770
1786
; CHECK: # %bb.0: # %entry
1787
+ ; CHECK-NEXT: vmv1r.v v25, v0
1771
1788
; CHECK-NEXT: vsetvli a0, a0, e16,m1,tu,mu
1772
- ; CHECK-NEXT: vmsle.vi v9, v8, 0, v0.t
1773
1789
; CHECK-NEXT: vmv1r.v v0, v9
1790
+ ; CHECK-NEXT: vmsle.vi v25, v8, 0, v0.t
1791
+ ; CHECK-NEXT: vmv1r.v v0, v25
1774
1792
; CHECK-NEXT: jalr zero, 0(ra)
1775
1793
entry:
1776
1794
%a = call <vscale x 4 x i1 > @llvm.riscv.vmslt.mask.nxv4i16.i16 (
@@ -1801,9 +1819,11 @@ entry:
1801
1819
define <vscale x 8 x i1 > @intrinsic_vmslt_mask_vi_nxv8i16_i16 (<vscale x 8 x i1 > %0 , <vscale x 8 x i16 > %1 , <vscale x 8 x i1 > %2 , i32 %3 ) nounwind {
1802
1820
; CHECK-LABEL: intrinsic_vmslt_mask_vi_nxv8i16_i16:
1803
1821
; CHECK: # %bb.0: # %entry
1822
+ ; CHECK-NEXT: vmv1r.v v25, v0
1804
1823
; CHECK-NEXT: vsetvli a0, a0, e16,m2,tu,mu
1805
- ; CHECK-NEXT: vmsle.vi v10, v8, 2, v0.t
1806
1824
; CHECK-NEXT: vmv1r.v v0, v10
1825
+ ; CHECK-NEXT: vmsle.vi v25, v8, 2, v0.t
1826
+ ; CHECK-NEXT: vmv1r.v v0, v25
1807
1827
; CHECK-NEXT: jalr zero, 0(ra)
1808
1828
entry:
1809
1829
%a = call <vscale x 8 x i1 > @llvm.riscv.vmslt.mask.nxv8i16.i16 (
@@ -1834,9 +1854,11 @@ entry:
1834
1854
define <vscale x 16 x i1 > @intrinsic_vmslt_mask_vi_nxv16i16_i16 (<vscale x 16 x i1 > %0 , <vscale x 16 x i16 > %1 , <vscale x 16 x i1 > %2 , i32 %3 ) nounwind {
1835
1855
; CHECK-LABEL: intrinsic_vmslt_mask_vi_nxv16i16_i16:
1836
1856
; CHECK: # %bb.0: # %entry
1857
+ ; CHECK-NEXT: vmv1r.v v25, v0
1837
1858
; CHECK-NEXT: vsetvli a0, a0, e16,m4,tu,mu
1838
- ; CHECK-NEXT: vmsle.vi v12, v8, 4, v0.t
1839
1859
; CHECK-NEXT: vmv1r.v v0, v12
1860
+ ; CHECK-NEXT: vmsle.vi v25, v8, 4, v0.t
1861
+ ; CHECK-NEXT: vmv1r.v v0, v25
1840
1862
; CHECK-NEXT: jalr zero, 0(ra)
1841
1863
entry:
1842
1864
%a = call <vscale x 16 x i1 > @llvm.riscv.vmslt.mask.nxv16i16.i16 (
@@ -1867,9 +1889,11 @@ entry:
1867
1889
define <vscale x 1 x i1 > @intrinsic_vmslt_mask_vi_nxv1i32_i32 (<vscale x 1 x i1 > %0 , <vscale x 1 x i32 > %1 , <vscale x 1 x i1 > %2 , i32 %3 ) nounwind {
1868
1890
; CHECK-LABEL: intrinsic_vmslt_mask_vi_nxv1i32_i32:
1869
1891
; CHECK: # %bb.0: # %entry
1892
+ ; CHECK-NEXT: vmv1r.v v25, v0
1870
1893
; CHECK-NEXT: vsetvli a0, a0, e32,mf2,tu,mu
1871
- ; CHECK-NEXT: vmsle.vi v9, v8, 6, v0.t
1872
1894
; CHECK-NEXT: vmv1r.v v0, v9
1895
+ ; CHECK-NEXT: vmsle.vi v25, v8, 6, v0.t
1896
+ ; CHECK-NEXT: vmv1r.v v0, v25
1873
1897
; CHECK-NEXT: jalr zero, 0(ra)
1874
1898
entry:
1875
1899
%a = call <vscale x 1 x i1 > @llvm.riscv.vmslt.mask.nxv1i32.i32 (
@@ -1900,9 +1924,11 @@ entry:
1900
1924
define <vscale x 2 x i1 > @intrinsic_vmslt_mask_vi_nxv2i32_i32 (<vscale x 2 x i1 > %0 , <vscale x 2 x i32 > %1 , <vscale x 2 x i1 > %2 , i32 %3 ) nounwind {
1901
1925
; CHECK-LABEL: intrinsic_vmslt_mask_vi_nxv2i32_i32:
1902
1926
; CHECK: # %bb.0: # %entry
1927
+ ; CHECK-NEXT: vmv1r.v v25, v0
1903
1928
; CHECK-NEXT: vsetvli a0, a0, e32,m1,tu,mu
1904
- ; CHECK-NEXT: vmsle.vi v9, v8, 8, v0.t
1905
1929
; CHECK-NEXT: vmv1r.v v0, v9
1930
+ ; CHECK-NEXT: vmsle.vi v25, v8, 8, v0.t
1931
+ ; CHECK-NEXT: vmv1r.v v0, v25
1906
1932
; CHECK-NEXT: jalr zero, 0(ra)
1907
1933
entry:
1908
1934
%a = call <vscale x 2 x i1 > @llvm.riscv.vmslt.mask.nxv2i32.i32 (
@@ -1933,9 +1959,11 @@ entry:
1933
1959
define <vscale x 4 x i1 > @intrinsic_vmslt_mask_vi_nxv4i32_i32 (<vscale x 4 x i1 > %0 , <vscale x 4 x i32 > %1 , <vscale x 4 x i1 > %2 , i32 %3 ) nounwind {
1934
1960
; CHECK-LABEL: intrinsic_vmslt_mask_vi_nxv4i32_i32:
1935
1961
; CHECK: # %bb.0: # %entry
1962
+ ; CHECK-NEXT: vmv1r.v v25, v0
1936
1963
; CHECK-NEXT: vsetvli a0, a0, e32,m2,tu,mu
1937
- ; CHECK-NEXT: vmsle.vi v10, v8, 10, v0.t
1938
1964
; CHECK-NEXT: vmv1r.v v0, v10
1965
+ ; CHECK-NEXT: vmsle.vi v25, v8, 10, v0.t
1966
+ ; CHECK-NEXT: vmv1r.v v0, v25
1939
1967
; CHECK-NEXT: jalr zero, 0(ra)
1940
1968
entry:
1941
1969
%a = call <vscale x 4 x i1 > @llvm.riscv.vmslt.mask.nxv4i32.i32 (
@@ -1966,9 +1994,11 @@ entry:
1966
1994
define <vscale x 8 x i1 > @intrinsic_vmslt_mask_vi_nxv8i32_i32 (<vscale x 8 x i1 > %0 , <vscale x 8 x i32 > %1 , <vscale x 8 x i1 > %2 , i32 %3 ) nounwind {
1967
1995
; CHECK-LABEL: intrinsic_vmslt_mask_vi_nxv8i32_i32:
1968
1996
; CHECK: # %bb.0: # %entry
1997
+ ; CHECK-NEXT: vmv1r.v v25, v0
1969
1998
; CHECK-NEXT: vsetvli a0, a0, e32,m4,tu,mu
1970
- ; CHECK-NEXT: vmsle.vi v12, v8, 12, v0.t
1971
1999
; CHECK-NEXT: vmv1r.v v0, v12
2000
+ ; CHECK-NEXT: vmsle.vi v25, v8, 12, v0.t
2001
+ ; CHECK-NEXT: vmv1r.v v0, v25
1972
2002
; CHECK-NEXT: jalr zero, 0(ra)
1973
2003
entry:
1974
2004
%a = call <vscale x 8 x i1 > @llvm.riscv.vmslt.mask.nxv8i32.i32 (
0 commit comments