From 7db9e31367807161121df907de9a2cb9fa2ca8f3 Mon Sep 17 00:00:00 2001 From: Abdul Rafae Noor <arnoor2@tyler.cs.illinois.edu> Date: Wed, 27 Jan 2021 14:50:47 -0600 Subject: [PATCH] Initializing tests for DFG2LLVM_WrapperAPI pass --- .../DFG2LLVM_WrapperAPI/wrapConv.ll | 136 ++++++++++++++++ .../DFG2LLVM_WrapperAPI/wrapMatMul.ll | 153 ++++++++++++++++++ 2 files changed, 289 insertions(+) create mode 100644 hpvm/test/regressionTests/DFG2LLVM_WrapperAPI/wrapConv.ll create mode 100644 hpvm/test/regressionTests/DFG2LLVM_WrapperAPI/wrapMatMul.ll diff --git a/hpvm/test/regressionTests/DFG2LLVM_WrapperAPI/wrapConv.ll b/hpvm/test/regressionTests/DFG2LLVM_WrapperAPI/wrapConv.ll new file mode 100644 index 0000000000..b61979c89d --- /dev/null +++ b/hpvm/test/regressionTests/DFG2LLVM_WrapperAPI/wrapConv.ll @@ -0,0 +1,136 @@ +; RUN: opt -load LLVMBuildDFG.so -load LLVMInPlaceDFGAnalysis.so -load LLVMDFG2LLVM_WrapperAPI.so -S -inplace -dfg2llvm-wrapperapi -quantization-levels-filename=q_file.txt --configuration-inputs-filename=conf_file.txt < %s | FileCheck %s + +; ModuleID = 'fuseConv.hpvm_.ll' +source_filename = "fuseConv.cpp" +target datalayout = "e-m:e-i64:64-f80:128-n8:16:32:64-S128" +target triple = "x86_64-unknown-linux-gnu" + +%struct.out._Z4rootPvmS_mS_m = type <{ i8*, i64 }> +%struct.out._Z13pool_max_nodePvm = type <{ i8*, i64 }> + +; CHECK-LABEL: i32 @main( +; CHECK: call void @llvm_hpvm_initApproxhpvmRt(i32 0) +; CHECK-NEXT: call void @llvm_hpvm_initializeRuntimeController( +; CHECK-NEXT: call void @llvm.hpvm.init() +; CHECK: call void @hpvm_request_tensor( +; CHECK: call void @llvm_hpvm_cleanupApproxhpvmRt() +; CHECK: call void @llvm_hpvm_clearRuntimeController() +; CHECK-NEXT: call void @llvm.hpvm.cleanup() + + + +; CHECK-LABEL: @_Z9conv_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned__Z13pool_max_nodePvm_cloned_wrapper_api( +; CHECK: call void @hpvm_request_tensor( +; CHECK: call void @hpvm_request_tensor( +; CHECK: call void @hpvm_request_tensor( +; CHECK: call i8* @wrapper_ConvLayer2( +; CHECK: ret + + + +; Function Attrs: norecurse nounwind uwtable +define dso_local i32 @main() local_unnamed_addr #0 { +entry: + call void @llvm.hpvm.init() + %call = tail call noalias i8* @malloc(i64 64) #3 + %graphID = call i8* @llvm.hpvm.launch(i8* bitcast (%struct.out._Z4rootPvmS_mS_m (i8*, i64, i8*, i64, i8*, i64)* @_Z4rootPvmS_mS_m_cloned to i8*), i8* %call, i1 false) + call void @llvm.hpvm.wait(i8* %graphID) + %input = bitcast i8* %call to i8** + %0 = load i8*, i8** %input, align 1, !tbaa !4 + tail call void @hpvm_request_tensor(i8* %0, i32 1) #3 + call void @llvm.hpvm.cleanup() + ret i32 0 +} + +; Function Attrs: nofree nounwind +declare dso_local noalias i8* @malloc(i64) local_unnamed_addr #1 + +declare dso_local void @hpvm_request_tensor(i8*, i32) local_unnamed_addr #2 + +; Function Attrs: nounwind +declare i8* @llvm.hpvm.tensor.convolution(i8*, i8*, i32, i32, i32, i32) #3 + +; Function Attrs: nounwind +declare i8* @llvm.hpvm.tensor.add(i8*, i8*) #3 + +; Function Attrs: nounwind +declare i8* @llvm.hpvm.tensor.tanh(i8*) #3 + +; Function Attrs: nounwind +declare i8* @llvm.hpvm.tensor.pool.max(i8*, i32, i32, i32, i32, i32, i32) #3 + +; Function Attrs: nounwind +declare i8* @llvm.hpvm.createNode(i8*) #3 + +; Function Attrs: nounwind +declare void @llvm.hpvm.bind.input(i8*, i32, i32, i1) #3 + +; Function Attrs: nounwind +declare i8* @llvm.hpvm.createEdge(i8*, i8*, i1, i32, i32, i1) #3 + +; Function Attrs: nounwind +declare void @llvm.hpvm.bind.output(i8*, i32, i32, i1) #3 + +; Function Attrs: nounwind uwtable +define dso_local %struct.out._Z4rootPvmS_mS_m @_Z4rootPvmS_mS_m_cloned(i8* in %input, i64 %input_bytes, i8* in %conv2d_1_w, i64 %conv2d_1_w_bytes, i8* in %conv2d_1_b, i64 %conv2d_1_b_bytes) #4 { +entry: + %_Z9conv_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned__Z13pool_max_nodePvm_cloned.node = call i8* @llvm.hpvm.createNode(i8* bitcast (%struct.out._Z13pool_max_nodePvm (i8*, i64, i8*, i64, i8*, i64)* @_Z9conv_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned__Z13pool_max_nodePvm_cloned to i8*)) + call void @llvm.hpvm.bind.input(i8* %_Z9conv_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned__Z13pool_max_nodePvm_cloned.node, i32 0, i32 0, i1 false) + call void @llvm.hpvm.bind.input(i8* %_Z9conv_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned__Z13pool_max_nodePvm_cloned.node, i32 1, i32 1, i1 false) + call void @llvm.hpvm.bind.input(i8* %_Z9conv_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned__Z13pool_max_nodePvm_cloned.node, i32 2, i32 2, i1 false) + call void @llvm.hpvm.bind.input(i8* %_Z9conv_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned__Z13pool_max_nodePvm_cloned.node, i32 3, i32 3, i1 false) + call void @llvm.hpvm.bind.input(i8* %_Z9conv_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned__Z13pool_max_nodePvm_cloned.node, i32 4, i32 4, i1 false) + call void @llvm.hpvm.bind.input(i8* %_Z9conv_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned__Z13pool_max_nodePvm_cloned.node, i32 5, i32 5, i1 false) + call void @llvm.hpvm.bind.output(i8* %_Z9conv_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned__Z13pool_max_nodePvm_cloned.node, i32 0, i32 0, i1 false) + call void @llvm.hpvm.bind.output(i8* %_Z9conv_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned__Z13pool_max_nodePvm_cloned.node, i32 1, i32 1, i1 false) + ret %struct.out._Z4rootPvmS_mS_m undef +} + +; Function Attrs: nounwind +declare void @llvm.hpvm.init() #3 + +; Function Attrs: nounwind +declare i8* @llvm.hpvm.launch(i8*, i8*, i1) #3 + +; Function Attrs: nounwind +declare void @llvm.hpvm.wait(i8*) #3 + +; Function Attrs: nounwind +declare void @llvm.hpvm.cleanup() #3 + +define %struct.out._Z13pool_max_nodePvm @_Z9conv_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned__Z13pool_max_nodePvm_cloned(i8* in %s_s_s_t1, i64 %s_s_s_bytes_t1, i8* in %s_s_s_t2, i64 %s_s_s_bytes_t2, i8* in %s_s_d_t2, i64 %s_s_d_bytes_t2) { +entry: + %s_s_s_call1 = call i8* @llvm.hpvm.tensor.convolution(i8* %s_s_s_t1, i8* %s_s_s_t2, i32 2, i32 2, i32 1, i32 1) + %s_s_call1 = call i8* @llvm.hpvm.tensor.add(i8* %s_s_s_call1, i8* %s_s_d_t2) + %s_call1 = call i8* @llvm.hpvm.tensor.tanh(i8* %s_s_call1) + %call1 = call i8* @llvm.hpvm.tensor.pool.max(i8* %s_call1, i32 2, i32 2, i32 0, i32 0, i32 2, i32 2) + %returnStruct = insertvalue %struct.out._Z13pool_max_nodePvm undef, i8* %call1, 0 + %returnStruct2 = insertvalue %struct.out._Z13pool_max_nodePvm %returnStruct, i64 0, 1 + ret %struct.out._Z13pool_max_nodePvm %returnStruct2 +} + +attributes #0 = { norecurse nounwind uwtable "correctly-rounded-divide-sqrt-fp-math"="false" "disable-tail-calls"="false" "less-precise-fpmad"="false" "min-legal-vector-width"="0" "no-frame-pointer-elim"="false" "no-infs-fp-math"="true" "no-jump-tables"="false" "no-nans-fp-math"="true" "no-signed-zeros-fp-math"="true" "no-trapping-math"="true" "stack-protector-buffer-size"="8" "target-cpu"="x86-64" "target-features"="+cx8,+fxsr,+mmx,+sse,+sse2,+x87" "unsafe-fp-math"="true" "use-soft-float"="false" } +attributes #1 = { nofree nounwind "correctly-rounded-divide-sqrt-fp-math"="false" "disable-tail-calls"="false" "less-precise-fpmad"="false" "no-frame-pointer-elim"="false" "no-infs-fp-math"="true" "no-nans-fp-math"="true" "no-signed-zeros-fp-math"="true" "no-trapping-math"="true" "stack-protector-buffer-size"="8" "target-cpu"="x86-64" "target-features"="+cx8,+fxsr,+mmx,+sse,+sse2,+x87" "unsafe-fp-math"="true" "use-soft-float"="false" } +attributes #2 = { "correctly-rounded-divide-sqrt-fp-math"="false" "disable-tail-calls"="false" "less-precise-fpmad"="false" "no-frame-pointer-elim"="false" "no-infs-fp-math"="true" "no-nans-fp-math"="true" "no-signed-zeros-fp-math"="true" "no-trapping-math"="true" "stack-protector-buffer-size"="8" "target-cpu"="x86-64" "target-features"="+cx8,+fxsr,+mmx,+sse,+sse2,+x87" "unsafe-fp-math"="true" "use-soft-float"="false" } +attributes #3 = { nounwind } +attributes #4 = { nounwind uwtable "correctly-rounded-divide-sqrt-fp-math"="false" "disable-tail-calls"="false" "less-precise-fpmad"="false" "min-legal-vector-width"="0" "no-frame-pointer-elim"="false" "no-infs-fp-math"="true" "no-jump-tables"="false" "no-nans-fp-math"="true" "no-signed-zeros-fp-math"="true" "no-trapping-math"="true" "stack-protector-buffer-size"="8" "target-cpu"="x86-64" "target-features"="+cx8,+fxsr,+mmx,+sse,+sse2,+x87" "unsafe-fp-math"="true" "use-soft-float"="false" } + +!llvm.module.flags = !{!0} +!llvm.ident = !{!1} +!hpvm_hint_promise = !{!2, !3} +!hpvm_hint_gpu = !{} +!hpvm_hint_cpu = !{} +!hpvm_hint_cpu_gpu = !{} +!hpvm_hint_cudnn = !{} + +!0 = !{i32 1, !"wchar_size", i32 4} +!1 = !{!"clang version 9.0.0 (https://gitlab.engr.illinois.edu/llvm/hpvm.git 5ccb2a532b5a0d82cee5c0d29a629a29dec2307c)"} +!2 = !{%struct.out._Z4rootPvmS_mS_m (i8*, i64, i8*, i64, i8*, i64)* @_Z4rootPvmS_mS_m_cloned} +!3 = !{%struct.out._Z13pool_max_nodePvm (i8*, i64, i8*, i64, i8*, i64)* @_Z9conv_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned__Z13pool_max_nodePvm_cloned} +!4 = !{!5, !6, i64 0} +!5 = !{!"_ZTS6RootIn", !6, i64 0, !9, i64 8, !6, i64 16, !9, i64 24, !6, i64 32, !9, i64 40, !10, i64 48} +!6 = !{!"any pointer", !7, i64 0} +!7 = !{!"omnipotent char", !8, i64 0} +!8 = !{!"Simple C++ TBAA"} +!9 = !{!"long", !7, i64 0} +!10 = !{!"_ZTS5ret_t", !6, i64 0, !9, i64 8} diff --git a/hpvm/test/regressionTests/DFG2LLVM_WrapperAPI/wrapMatMul.ll b/hpvm/test/regressionTests/DFG2LLVM_WrapperAPI/wrapMatMul.ll new file mode 100644 index 0000000000..ce901a6d41 --- /dev/null +++ b/hpvm/test/regressionTests/DFG2LLVM_WrapperAPI/wrapMatMul.ll @@ -0,0 +1,153 @@ +; RUN: opt -load LLVMBuildDFG.so -load LLVMInPlaceDFGAnalysis.so -load LLVMDFG2LLVM_WrapperAPI.so -S -inplace -dfg2llvm-wrapperapi -quantization-levels-filename=q_file.txt --configuration-inputs-filename=conf_file.txt < %s | FileCheck %s + +; ModuleID = 'fuseMatMul.hpvm.ll' +source_filename = "fuseMatMul.cpp" +target datalayout = "e-m:e-i64:64-f80:128-n8:16:32:64-S128" +target triple = "x86_64-unknown-linux-gnu" + +%struct.out._Z13pool_max_nodePvm = type <{ i8*, i64 }> +%struct.out._Z4rootPvmS_mS_m = type <{ i8*, i64 }> +%struct.out._Z9tanh_nodePvm = type <{ i8*, i64 }> + +; CHECK-LABEL: i32 @main( +; CHECK: call void @llvm_hpvm_initApproxhpvmRt(i32 0) +; CHECK-NEXT: call void @llvm_hpvm_initializeRuntimeController( +; CHECK-NEXT: call void @llvm.hpvm.init() +; CHECK: call void @hpvm_request_tensor( +; CHECK: call void @llvm_hpvm_cleanupApproxhpvmRt() +; CHECK: call void @llvm_hpvm_clearRuntimeController() +; CHECK-NEXT: call void @llvm.hpvm.cleanup() + + +; CHECK-LABEL: @_Z11matmul_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned_wrapper_api( +; CHECK: call void @hpvm_request_tensor( +; CHECK: call void @hpvm_request_tensor( +; CHECK: call void @hpvm_request_tensor( +; CHECK: call i8* @wrapper_FCLayer( +; CHECK: ret + +; CHECK-LABEL: @_Z13pool_max_nodePvm_cloned_wrapper_api( +; CHECK: call void @hpvm_request_tensor(i8* %t1, i32 1) +; CHECK: call i8* @wrapper_tensorPooling( +; CHECK: ret + + + +; Function Attrs: norecurse nounwind uwtable +define dso_local i32 @main() local_unnamed_addr #0 { +entry: + call void @llvm.hpvm.init() + %call = tail call noalias i8* @malloc(i64 64) #3 + %graphID = call i8* @llvm.hpvm.launch(i8* bitcast (%struct.out._Z4rootPvmS_mS_m (i8*, i64, i8*, i64, i8*, i64)* @_Z4rootPvmS_mS_m_cloned to i8*), i8* %call, i1 false) + call void @llvm.hpvm.wait(i8* %graphID) + %input = bitcast i8* %call to i8** + %0 = load i8*, i8** %input, align 1, !tbaa !5 + tail call void @hpvm_request_tensor(i8* %0, i32 1) #3 + call void @llvm.hpvm.cleanup() + ret i32 0 +} + +; Function Attrs: nofree nounwind +declare dso_local noalias i8* @malloc(i64) local_unnamed_addr #1 + +declare dso_local void @hpvm_request_tensor(i8*, i32) local_unnamed_addr #2 + +; Function Attrs: nounwind +declare i8* @llvm.hpvm.tensor.mul(i8*, i8*) #3 + +; Function Attrs: nounwind +declare i8* @llvm.hpvm.tensor.add(i8*, i8*) #3 + +; Function Attrs: nounwind +declare i8* @llvm.hpvm.tensor.tanh(i8*) #3 + +; Function Attrs: nounwind +declare i8* @llvm.hpvm.tensor.pool.max(i8*, i32, i32, i32, i32, i32, i32) #3 + +; Function Attrs: nounwind uwtable +define dso_local %struct.out._Z13pool_max_nodePvm @_Z13pool_max_nodePvm_cloned(i8* in %t1, i64 %bytes_t1) #4 { +entry: + %call1 = call i8* @llvm.hpvm.tensor.pool.max(i8* %t1, i32 2, i32 2, i32 0, i32 0, i32 2, i32 2) + %returnStruct = insertvalue %struct.out._Z13pool_max_nodePvm undef, i8* %call1, 0 + %returnStruct2 = insertvalue %struct.out._Z13pool_max_nodePvm %returnStruct, i64 0, 1 + ret %struct.out._Z13pool_max_nodePvm %returnStruct2 +} + +; Function Attrs: nounwind +declare i8* @llvm.hpvm.createNode(i8*) #3 + +; Function Attrs: nounwind +declare void @llvm.hpvm.bind.input(i8*, i32, i32, i1) #3 + +; Function Attrs: nounwind +declare i8* @llvm.hpvm.createEdge(i8*, i8*, i1, i32, i32, i1) #3 + +; Function Attrs: nounwind +declare void @llvm.hpvm.bind.output(i8*, i32, i32, i1) #3 + +; Function Attrs: nounwind uwtable +define dso_local %struct.out._Z4rootPvmS_mS_m @_Z4rootPvmS_mS_m_cloned(i8* in %input, i64 %input_bytes, i8* in %matmul2d_1_w, i64 %matmul2d_1_w_bytes, i8* in %matmul2d_1_b, i64 %matmul2d_1_b_bytes) #4 { +entry: + %_Z11matmul_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned.node = call i8* @llvm.hpvm.createNode(i8* bitcast (%struct.out._Z9tanh_nodePvm (i8*, i64, i8*, i64, i8*, i64)* @_Z11matmul_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned to i8*)) + call void @llvm.hpvm.bind.input(i8* %_Z11matmul_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned.node, i32 0, i32 0, i1 false) + call void @llvm.hpvm.bind.input(i8* %_Z11matmul_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned.node, i32 1, i32 1, i1 false) + call void @llvm.hpvm.bind.input(i8* %_Z11matmul_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned.node, i32 2, i32 2, i1 false) + call void @llvm.hpvm.bind.input(i8* %_Z11matmul_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned.node, i32 3, i32 3, i1 false) + call void @llvm.hpvm.bind.input(i8* %_Z11matmul_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned.node, i32 4, i32 4, i1 false) + call void @llvm.hpvm.bind.input(i8* %_Z11matmul_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned.node, i32 5, i32 5, i1 false) + %_Z13pool_max_nodePvm_cloned.node = call i8* @llvm.hpvm.createNode(i8* bitcast (%struct.out._Z13pool_max_nodePvm (i8*, i64)* @_Z13pool_max_nodePvm_cloned to i8*)) + %output4 = call i8* @llvm.hpvm.createEdge(i8* %_Z11matmul_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned.node, i8* %_Z13pool_max_nodePvm_cloned.node, i1 true, i32 0, i32 0, i1 false) + %output5 = call i8* @llvm.hpvm.createEdge(i8* %_Z11matmul_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned.node, i8* %_Z13pool_max_nodePvm_cloned.node, i1 true, i32 1, i32 1, i1 false) + call void @llvm.hpvm.bind.output(i8* %_Z13pool_max_nodePvm_cloned.node, i32 0, i32 0, i1 false) + call void @llvm.hpvm.bind.output(i8* %_Z13pool_max_nodePvm_cloned.node, i32 1, i32 1, i1 false) + ret %struct.out._Z4rootPvmS_mS_m undef +} + +; Function Attrs: nounwind +declare void @llvm.hpvm.init() #3 + +; Function Attrs: nounwind +declare i8* @llvm.hpvm.launch(i8*, i8*, i1) #3 + +; Function Attrs: nounwind +declare void @llvm.hpvm.wait(i8*) #3 + +; Function Attrs: nounwind +declare void @llvm.hpvm.cleanup() #3 + +define %struct.out._Z9tanh_nodePvm @_Z11matmul_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned(i8* in %s_s_t1, i64 %s_s_bytes_t1, i8* in %s_s_t2, i64 %s_s_bytes_t2, i8* in %s_d_t2, i64 %s_d_bytes_t2) { +entry: + %s_s_call1 = call i8* @llvm.hpvm.tensor.mul(i8* %s_s_t1, i8* %s_s_t2) + %s_call1 = call i8* @llvm.hpvm.tensor.add(i8* %s_s_call1, i8* %s_d_t2) + %call1 = call i8* @llvm.hpvm.tensor.tanh(i8* %s_call1) + %returnStruct = insertvalue %struct.out._Z9tanh_nodePvm undef, i8* %call1, 0 + %returnStruct2 = insertvalue %struct.out._Z9tanh_nodePvm %returnStruct, i64 0, 1 + ret %struct.out._Z9tanh_nodePvm %returnStruct2 +} + +attributes #0 = { norecurse nounwind uwtable "correctly-rounded-divide-sqrt-fp-math"="false" "disable-tail-calls"="false" "less-precise-fpmad"="false" "min-legal-vector-width"="0" "no-frame-pointer-elim"="false" "no-infs-fp-math"="true" "no-jump-tables"="false" "no-nans-fp-math"="true" "no-signed-zeros-fp-math"="true" "no-trapping-math"="true" "stack-protector-buffer-size"="8" "target-cpu"="x86-64" "target-features"="+cx8,+fxsr,+mmx,+sse,+sse2,+x87" "unsafe-fp-math"="true" "use-soft-float"="false" } +attributes #1 = { nofree nounwind "correctly-rounded-divide-sqrt-fp-math"="false" "disable-tail-calls"="false" "less-precise-fpmad"="false" "no-frame-pointer-elim"="false" "no-infs-fp-math"="true" "no-nans-fp-math"="true" "no-signed-zeros-fp-math"="true" "no-trapping-math"="true" "stack-protector-buffer-size"="8" "target-cpu"="x86-64" "target-features"="+cx8,+fxsr,+mmx,+sse,+sse2,+x87" "unsafe-fp-math"="true" "use-soft-float"="false" } +attributes #2 = { "correctly-rounded-divide-sqrt-fp-math"="false" "disable-tail-calls"="false" "less-precise-fpmad"="false" "no-frame-pointer-elim"="false" "no-infs-fp-math"="true" "no-nans-fp-math"="true" "no-signed-zeros-fp-math"="true" "no-trapping-math"="true" "stack-protector-buffer-size"="8" "target-cpu"="x86-64" "target-features"="+cx8,+fxsr,+mmx,+sse,+sse2,+x87" "unsafe-fp-math"="true" "use-soft-float"="false" } +attributes #3 = { nounwind } +attributes #4 = { nounwind uwtable "correctly-rounded-divide-sqrt-fp-math"="false" "disable-tail-calls"="false" "less-precise-fpmad"="false" "min-legal-vector-width"="0" "no-frame-pointer-elim"="false" "no-infs-fp-math"="true" "no-jump-tables"="false" "no-nans-fp-math"="true" "no-signed-zeros-fp-math"="true" "no-trapping-math"="true" "stack-protector-buffer-size"="8" "target-cpu"="x86-64" "target-features"="+cx8,+fxsr,+mmx,+sse,+sse2,+x87" "unsafe-fp-math"="true" "use-soft-float"="false" } + +!llvm.module.flags = !{!0} +!llvm.ident = !{!1} +!hpvm_hint_promise = !{!2, !3, !4} +!hpvm_hint_gpu = !{} +!hpvm_hint_cpu = !{} +!hpvm_hint_cpu_gpu = !{} +!hpvm_hint_cudnn = !{} + +!0 = !{i32 1, !"wchar_size", i32 4} +!1 = !{!"clang version 9.0.0 (https://gitlab.engr.illinois.edu/llvm/hpvm.git 5ccb2a532b5a0d82cee5c0d29a629a29dec2307c)"} +!2 = !{%struct.out._Z13pool_max_nodePvm (i8*, i64)* @_Z13pool_max_nodePvm_cloned} +!3 = !{%struct.out._Z4rootPvmS_mS_m (i8*, i64, i8*, i64, i8*, i64)* @_Z4rootPvmS_mS_m_cloned} +!4 = !{%struct.out._Z9tanh_nodePvm (i8*, i64, i8*, i64, i8*, i64)* @_Z11matmul_nodePvmS_m_cloned__Z13bias_add_nodePvmS_m_cloned__Z9tanh_nodePvm_cloned} +!5 = !{!6, !7, i64 0} +!6 = !{!"_ZTS6RootIn", !7, i64 0, !10, i64 8, !7, i64 16, !10, i64 24, !7, i64 32, !10, i64 40, !11, i64 48} +!7 = !{!"any pointer", !8, i64 0} +!8 = !{!"omnipotent char", !9, i64 0} +!9 = !{!"Simple C++ TBAA"} +!10 = !{!"long", !8, i64 0} +!11 = !{!"_ZTS5ret_t", !7, i64 0, !10, i64 8} -- GitLab