summaryrefslogtreecommitdiffstats
path: root/test
diff options
context:
space:
mode:
authorBill Wendling <isanbard@gmail.com>2013-12-08 00:07:30 +0000
committerBill Wendling <isanbard@gmail.com>2013-12-08 00:07:30 +0000
commit488aab6df3723d49e256042b99e5ef2f5a9cf46b (patch)
treede686c91b36487eb9e525470e751763ac9c99a38 /test
parent4a5b649babdb800f9a578d4ee91427a98a795874 (diff)
downloadexternal_llvm-488aab6df3723d49e256042b99e5ef2f5a9cf46b.zip
external_llvm-488aab6df3723d49e256042b99e5ef2f5a9cf46b.tar.gz
external_llvm-488aab6df3723d49e256042b99e5ef2f5a9cf46b.tar.bz2
Merging r196362:
------------------------------------------------------------------------ r196362 | kevinqin | 2013-12-04 00:02:34 -0800 (Wed, 04 Dec 2013) | 1 line [AArch64 Neon] Add ACLE intrinsic vceqz_f64. ------------------------------------------------------------------------ git-svn-id: https://llvm.org/svn/llvm-project/llvm/branches/release_34@196699 91177308-0d34-0410-b5e6-96231b3b80d8
Diffstat (limited to 'test')
-rw-r--r--test/CodeGen/AArch64/neon-scalar-fp-compare.ll26
1 files changed, 19 insertions, 7 deletions
diff --git a/test/CodeGen/AArch64/neon-scalar-fp-compare.ll b/test/CodeGen/AArch64/neon-scalar-fp-compare.ll
index aa6348d..a6e5859 100644
--- a/test/CodeGen/AArch64/neon-scalar-fp-compare.ll
+++ b/test/CodeGen/AArch64/neon-scalar-fp-compare.ll
@@ -24,6 +24,15 @@ entry:
ret i64 %0
}
+define <1 x i64> @test_vceqz_f64(<1 x double> %a) #0 {
+; CHECK: test_vceqz_f64
+; CHECK: fcmeq {{d[0-9]+}}, {{d[0-9]+}}, #0.0
+entry:
+ %0 = fcmp oeq <1 x double> %a, zeroinitializer
+ %vceqz.i = zext <1 x i1> %0 to <1 x i64>
+ ret <1 x i64> %vceqz.i
+}
+
define i32 @test_vceqzs_f32(float %a) {
; CHECK: test_vceqzs_f32
; CHECK: fcmeq {{s[0-9]}}, {{s[0-9]}}, #0.0
@@ -39,7 +48,7 @@ define i64 @test_vceqzd_f64(double %a) {
; CHECK: fcmeq {{d[0-9]}}, {{d[0-9]}}, #0.0
entry:
%vceq.i = insertelement <1 x double> undef, double %a, i32 0
- %vceq1.i = call <1 x i64> @llvm.aarch64.neon.vceq.v1i64.v1f64.v1f64(<1 x double> %vceq.i, <1 x double> zeroinitializer)
+ %vceq1.i = tail call <1 x i64> @llvm.aarch64.neon.vceq.v1i64.v1f64.v1f32(<1 x double> %vceq.i, <1 x float> zeroinitializer) #5
%0 = extractelement <1 x i64> %vceq1.i, i32 0
ret i64 %0
}
@@ -81,7 +90,7 @@ define i64 @test_vcgezd_f64(double %a) {
; CHECK: fcmge {{d[0-9]}}, {{d[0-9]}}, #0.0
entry:
%vcge.i = insertelement <1 x double> undef, double %a, i32 0
- %vcge1.i = call <1 x i64> @llvm.aarch64.neon.vcge.v1i64.v1f64.v1f64(<1 x double> %vcge.i, <1 x double> zeroinitializer)
+ %vcge1.i = tail call <1 x i64> @llvm.aarch64.neon.vcge.v1i64.v1f64.v1f32(<1 x double> %vcge.i, <1 x float> zeroinitializer) #5
%0 = extractelement <1 x i64> %vcge1.i, i32 0
ret i64 %0
}
@@ -123,7 +132,7 @@ define i64 @test_vcgtzd_f64(double %a) {
; CHECK: fcmgt {{d[0-9]}}, {{d[0-9]}}, #0.0
entry:
%vcgt.i = insertelement <1 x double> undef, double %a, i32 0
- %vcgt1.i = call <1 x i64> @llvm.aarch64.neon.vcgt.v1i64.v1f64.v1f64(<1 x double> %vcgt.i, <1 x double> zeroinitializer)
+ %vcgt1.i = tail call <1 x i64> @llvm.aarch64.neon.vcgt.v1i64.v1f64.v1f32(<1 x double> %vcgt.i, <1 x float> zeroinitializer) #5
%0 = extractelement <1 x i64> %vcgt1.i, i32 0
ret i64 %0
}
@@ -165,7 +174,7 @@ define i64 @test_vclezd_f64(double %a) {
; CHECK: fcmle {{d[0-9]}}, {{d[0-9]}}, #0.0
entry:
%vcle.i = insertelement <1 x double> undef, double %a, i32 0
- %vcle1.i = call <1 x i64> @llvm.aarch64.neon.vclez.v1i64.v1f64.v1f64(<1 x double> %vcle.i, <1 x double> zeroinitializer)
+ %vcle1.i = tail call <1 x i64> @llvm.aarch64.neon.vclez.v1i64.v1f64.v1f32(<1 x double> %vcle.i, <1 x float> zeroinitializer) #5
%0 = extractelement <1 x i64> %vcle1.i, i32 0
ret i64 %0
}
@@ -207,7 +216,7 @@ define i64 @test_vcltzd_f64(double %a) {
; CHECK: fcmlt {{d[0-9]}}, {{d[0-9]}}, #0.0
entry:
%vclt.i = insertelement <1 x double> undef, double %a, i32 0
- %vclt1.i = call <1 x i64> @llvm.aarch64.neon.vcltz.v1i64.v1f64.v1f64(<1 x double> %vclt.i, <1 x double> zeroinitializer)
+ %vclt1.i = tail call <1 x i64> @llvm.aarch64.neon.vcltz.v1i64.v1f64.v1f32(<1 x double> %vclt.i, <1 x float> zeroinitializer) #5
%0 = extractelement <1 x i64> %vclt1.i, i32 0
ret i64 %0
}
@@ -301,15 +310,18 @@ entry:
}
declare <1 x i32> @llvm.aarch64.neon.vceq.v1i32.v1f32.v1f32(<1 x float>, <1 x float>)
+declare <1 x i64> @llvm.aarch64.neon.vceq.v1i64.v1f64.v1f32(<1 x double>, <1 x float>)
declare <1 x i64> @llvm.aarch64.neon.vceq.v1i64.v1f64.v1f64(<1 x double>, <1 x double>)
declare <1 x i32> @llvm.aarch64.neon.vcge.v1i32.v1f32.v1f32(<1 x float>, <1 x float>)
+declare <1 x i64> @llvm.aarch64.neon.vcge.v1i64.v1f64.v1f32(<1 x double>, <1 x float>)
declare <1 x i64> @llvm.aarch64.neon.vcge.v1i64.v1f64.v1f64(<1 x double>, <1 x double>)
declare <1 x i32> @llvm.aarch64.neon.vclez.v1i32.v1f32.v1f32(<1 x float>, <1 x float>)
-declare <1 x i64> @llvm.aarch64.neon.vclez.v1i64.v1f64.v1f64(<1 x double>, <1 x double>)
+declare <1 x i64> @llvm.aarch64.neon.vclez.v1i64.v1f64.v1f32(<1 x double>, <1 x float>)
declare <1 x i32> @llvm.aarch64.neon.vcgt.v1i32.v1f32.v1f32(<1 x float>, <1 x float>)
+declare <1 x i64> @llvm.aarch64.neon.vcgt.v1i64.v1f64.v1f32(<1 x double>, <1 x float>)
declare <1 x i64> @llvm.aarch64.neon.vcgt.v1i64.v1f64.v1f64(<1 x double>, <1 x double>)
declare <1 x i32> @llvm.aarch64.neon.vcltz.v1i32.v1f32.v1f32(<1 x float>, <1 x float>)
-declare <1 x i64> @llvm.aarch64.neon.vcltz.v1i64.v1f64.v1f64(<1 x double>, <1 x double>)
+declare <1 x i64> @llvm.aarch64.neon.vcltz.v1i64.v1f64.v1f32(<1 x double>, <1 x float>)
declare <1 x i32> @llvm.aarch64.neon.vcage.v1i32.v1f32.v1f32(<1 x float>, <1 x float>)
declare <1 x i64> @llvm.aarch64.neon.vcage.v1i64.v1f64.v1f64(<1 x double>, <1 x double>)
declare <1 x i32> @llvm.aarch64.neon.vcagt.v1i32.v1f32.v1f32(<1 x float>, <1 x float>)