diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/alexnet2_cifar10/alexnet2_cifar10.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/alexnet2_cifar10/alexnet2_cifar10.txt
deleted file mode 100644
index 6ec4a06d3dbd2e088d6db287d23dd3bd5aad7ddb..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/alexnet2_cifar10/alexnet2_cifar10.txt
+++ /dev/null
@@ -1,419 +0,0 @@
-1114.3009809999999
-+++++
-conf1 1 1 84.98 0.0
-1 gpu conv fp32 11 add fp32 1 tanh fp32 1
-2 gpu conv fp32 11 add fp32 1 tanh fp32 1 pool_max fp32 1
-3 gpu conv fp32 11 add fp32 1 tanh fp32 1
-4 gpu conv fp32 11 add fp32 1 tanh fp32 1 pool_max fp32 1
-5 gpu conv fp32 11 add fp32 1 tanh fp32 1
-6 gpu conv fp32 11 add fp32 1 tanh fp32 1 pool_max fp32 1
-7 gpu mul fp32 11 add fp32 1
-8 gpu softmax fp32 1
------
-+++++
-conf2 2.4248748377353113 2.0815908534183163 84.5 0.480000000000004
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 157 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf3 2.4055188425519614 2.0586265720811823 84.48 0.5
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 266 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 269 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 157 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf4 2.4156140842962985 2.0617867479342706 84.28 0.7000000000000028
-1 gpu conv perf_fp16 157 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 163 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf5 2.396416918342732 2.0506214971794585 84.02 0.960000000000008
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 151 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 157 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf6 2.463002582910052 2.1171077568609458 83.84 1.1400000000000006
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 167 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf7 2.360283215266004 2.0255245321874304 83.78 1.2000000000000028
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 157 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf8 2.4140791541736157 2.0671513522247653 83.74000000000001 1.2399999999999949
-1 gpu conv fp16 11 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 160 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf9 2.457753689612079 2.1086250651240137 83.7 1.2800000000000011
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 163 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf10 2.459170454055443 2.1111925341396343 83.7 1.2800000000000011
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 164 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf11 2.4135986141645764 2.060453960420927 83.62 1.3599999999999994
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 157 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf12 2.4631278039012106 2.1092094797926637 83.58 1.4000000000000057
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 157 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf13 2.535761391794481 2.16998336112692 83.58 1.4000000000000057
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 266 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 157 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf14 2.289006193945062 1.961240158652051 83.54 1.4399999999999977
-1 gpu conv perf_fp16 167 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 155 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 157 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf15 2.4257674844112573 2.0808440756495563 83.5 1.480000000000004
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 161 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf16 2.458122368488622 2.109531159729078 83.48 1.5
-1 gpu conv fp16 11 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 162 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf17 2.281072202152105 1.9539314420536427 83.46000000000001 1.519999999999996
-1 gpu conv fp16 11 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 160 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 160 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf18 2.4572171342078444 2.1088933553775697 83.46000000000001 1.519999999999996
-1 gpu conv fp16 11 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 163 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf19 2.3017607719030058 1.9782265708150768 83.42 1.5600000000000023
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 162 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 162 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf20 2.379206814483014 2.047909200292713 83.39999999999999 1.5800000000000125
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 151 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf21 2.4636282705302537 2.1162281156388527 83.39999999999999 1.5800000000000125
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 160 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf22 2.461590101374146 2.1108493881199184 83.22 1.7600000000000051
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 161 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf23 2.537054645442804 2.167568834938183 83.22 1.7600000000000051
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 266 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 157 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf24 2.4631604723407885 2.1099694757102845 83.17999999999999 1.8000000000000114
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 157 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf25 2.4636282705302537 2.1162281156388527 83.14 1.8400000000000034
-1 gpu conv fp16 11 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 160 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf26 2.462588899729088 2.109477918791931 83.14 1.8400000000000034
-1 gpu conv fp16 11 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 157 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf27 2.4638085754689025 2.1071960926343603 83.1 1.8800000000000097
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 157 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf28 2.4640079766123635 2.110326453157297 83.08 1.9000000000000057
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf29 2.459337622764853 2.107249218450713 83.06 1.9200000000000017
-1 gpu conv fp16 11 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 162 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf30 2.538176340059405 2.173287257415721 83.02000000000001 1.9599999999999937
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 164 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf31 2.3905426931959846 2.044333576277581 83.02000000000001 1.9599999999999937
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 160 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 157 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf32 2.459337622764853 2.107249218450713 83.0 1.980000000000004
-1 gpu conv fp16 11 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 162 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf33 2.458968579288317 2.1063450826631396 82.89999999999999 2.0800000000000125
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 163 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf34 2.2912974651603877 1.9670210508860688 82.8 2.180000000000007
-1 gpu conv perf_fp16 168 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 155 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 160 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf35 2.4648489763056327 2.113931670664391 82.66 2.3200000000000074
-1 gpu conv fp16 11 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 160 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf36 2.4599076869402854 2.1077397371200193 82.6 2.3800000000000097
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 162 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf37 2.4636282705302537 2.1162281156388527 82.54 2.4399999999999977
-1 gpu conv fp16 11 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 160 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
-+++++
-conf38 2.591814267389778 2.222680944458784 82.26 2.719999999999999
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv perf_fp16 154 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv perf_fp16 157 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/alexnet_cifar10/alexnet_cifar10.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/alexnet_cifar10/alexnet_cifar10.txt
deleted file mode 100644
index a9ccba6eb63f620c0e3b6f95fd7c50892018f00f..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/alexnet_cifar10/alexnet_cifar10.txt
+++ /dev/null
@@ -1,511 +0,0 @@
-2592.187221
-+++++
-conf1 1 1 79.28 0.0
-1 gpu conv fp32 11 add fp32 1 tanh fp32 1 pool_max fp32 1
-2 gpu conv fp32 11 add fp32 1 tanh fp32 1 pool_max fp32 1
-3 gpu conv fp32 11 add fp32 1 tanh fp32 1
-4 gpu conv fp32 11 add fp32 1 tanh fp32 1
-5 gpu conv fp32 11 add fp32 1 tanh fp32 1 pool_max fp32 1
-6 gpu mul fp32 11 add fp32 1
-7 gpu softmax fp32 1
------
-+++++
-conf2 1.7593976485873195 1.6193399031642917 79.23 0.04999999999999716
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf3 2.092625440752526 1.9139078015388271 78.96 0.3200000000000074
-1 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 266 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf4 1.8870195448805414 1.7296919053025768 78.8 0.480000000000004
-1 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf5 2.1184804041774554 1.9598989563949536 78.75999999999999 0.5200000000000102
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf6 2.1184804041774554 1.9598989563949536 78.75999999999999 0.5200000000000102
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf7 2.0933825381386364 1.9150743378318535 78.64 0.6400000000000006
-1 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf8 2.081712090729918 1.9102226906341664 78.5 0.7800000000000011
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf9 2.081712090729918 1.9102226906341664 78.5 0.7800000000000011
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf10 2.2662606588487595 2.066560750795139 78.48 0.7999999999999972
-1 gpu conv samp_fp16 264 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf11 2.121684761285686 1.966318179285323 78.48 0.7999999999999972
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf12 2.3417491169395532 2.1355030360671465 78.38000000000001 0.8999999999999915
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf13 2.2247938983110425 2.060416584958474 78.38000000000001 0.8999999999999915
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf14 2.2247938983110425 2.060416584958474 78.38000000000001 0.8999999999999915
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf15 2.2247938983110425 2.060416584958474 78.38000000000001 0.8999999999999915
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf16 2.2627828537139263 2.065683616898884 78.32000000000001 0.9599999999999937
-1 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf17 2.2627828537139263 2.065683616898884 78.32000000000001 0.9599999999999937
-1 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf18 2.2627828537139263 2.065683616898884 78.32000000000001 0.9599999999999937
-1 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf19 2.146571989407323 1.95711703610764 78.18 1.0999999999999943
-1 gpu conv samp_fp16 264 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf20 2.303316973793268 2.1036463961913276 78.10000000000001 1.1799999999999926
-1 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf21 2.436875653706139 2.2434837737118056 78.08 1.2000000000000028
-1 gpu conv samp_fp16 264 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf22 2.436875653706139 2.2434837737118056 78.08 1.2000000000000028
-1 gpu conv samp_fp16 264 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf23 2.436875653706139 2.2434837737118056 78.08 1.2000000000000028
-1 gpu conv samp_fp16 264 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf24 2.1106508925330925 1.9419233584234938 78.06 1.2199999999999989
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf25 2.3203534290038634 2.116965679235447 78.06 1.2199999999999989
-1 gpu conv samp_fp16 264 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf26 2.3527290658539215 2.145832257234814 78.03999999999999 1.240000000000009
-1 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf27 2.3527290658539215 2.145832257234814 78.03999999999999 1.240000000000009
-1 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf28 2.432854949808342 2.2424500615508003 78.0 1.2800000000000011
-1 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf29 2.432854949808342 2.2424500615508003 78.0 1.2800000000000011
-1 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf30 2.432854949808342 2.2424500615508003 78.0 1.2800000000000011
-1 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf31 2.3137982135449207 2.1281257317083417 77.84 1.4399999999999977
-1 gpu conv samp_fp16 264 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 265 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf32 2.1198074418988333 1.9522214255218437 77.82 1.460000000000008
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf33 2.246924974355375 2.065289762405701 77.8 1.480000000000004
-1 gpu conv samp_fp16 264 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 269 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf34 2.263614734554485 2.090777846534249 77.74 1.5400000000000063
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf35 2.263614734554485 2.090777846534249 77.74 1.5400000000000063
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf36 2.263614734554485 2.090777846534249 77.74 1.5400000000000063
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf37 2.5289288699015304 2.334007588396142 77.72 1.5600000000000023
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf38 2.5289288699015304 2.334007588396142 77.72 1.5600000000000023
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf39 2.3117594882585775 2.1152397180868943 77.56 1.7199999999999989
-1 gpu conv samp_fp16 264 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf40 2.452732477854469 2.264573687601476 77.56 1.7199999999999989
-1 gpu conv perf_fp16 167 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf41 2.452732477854469 2.264573687601476 77.56 1.7199999999999989
-1 gpu conv perf_fp16 167 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf42 2.382518688546389 2.178614303992064 77.5 1.7800000000000011
-1 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf43 2.382518688546389 2.178614303992064 77.5 1.7800000000000011
-1 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf44 2.3900667100485924 2.188128526401265 77.48 1.7999999999999972
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf45 2.3900667100485924 2.188128526401265 77.48 1.7999999999999972
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf46 2.3900667100485924 2.188128526401265 77.48 1.7999999999999972
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf47 2.4835281673276515 2.279527076032239 77.3 1.980000000000004
-1 gpu conv samp_fp16 264 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf48 2.4835281673276515 2.279527076032239 77.3 1.980000000000004
-1 gpu conv samp_fp16 264 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf49 2.1553694968551302 1.9959124044028933 77.18 2.0999999999999943
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 265 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf50 2.5877520959724816 2.3763616521050364 77.03999999999999 2.240000000000009
-1 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf51 2.5877520959724816 2.3763616521050364 77.03999999999999 2.240000000000009
-1 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/alexnet_imagenet/alexnet_imagenet.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/alexnet_imagenet/alexnet_imagenet.txt
deleted file mode 100644
index b0e42a5aaa5d7b5a06b6422a5c33a0047b6eff8d..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/alexnet_imagenet/alexnet_imagenet.txt
+++ /dev/null
@@ -1,229 +0,0 @@
-2739.950736
-+++++
-conf1 1 1 56.3 0.0
-1 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-2 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-3 gpu conv fp32 11 add fp32 1 relu fp32 1
-4 gpu conv fp32 11 add fp32 1 relu fp32 1
-5 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-6 gpu mul fp32 11 add fp32 1 relu fp32 1
-7 gpu mul fp32 11 add fp32 1 relu fp32 1
-8 gpu mul fp32 11 add fp32 1
-9 gpu softmax fp32 1
------
-+++++
-conf2 1.802133644103582 1.8186433204507424 55.76 0.5399999999999991
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf3 1.7574572103878898 1.7673706184460103 55.58 0.7199999999999989
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1
-5 gpu conv samp_fp16 268 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf4 2.0227701930718065 2.043112495268932 55.42 0.8799999999999955
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 166 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf5 1.9872634777043927 2.002789650227035 55.120000000000005 1.1799999999999926
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 164 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf6 1.8204253918445088 1.843736069756362 54.84 1.4599999999999937
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 164 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 154 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf7 1.9308336510645352 1.934889049414224 54.74 1.5599999999999952
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 168 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf8 2.0146435217865446 2.0367475358800102 54.58 1.7199999999999989
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 162 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf9 2.0101709494490696 2.0329911158023064 54.400000000000006 1.8999999999999915
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 164 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf10 2.0052132441967916 2.0284931705407003 54.300000000000004 1.999999999999993
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 168 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf11 2.010827434817262 2.036001862538864 54.2 2.0999999999999943
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 164 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 154 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf12 2.019868378233057 2.0433540129730265 54.17999999999999 2.1200000000000045
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 162 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf13 1.9923471030291253 2.009177323959059 54.120000000000005 2.1799999999999926
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 164 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf14 1.9923471030291253 2.009177323959059 54.120000000000005 2.1799999999999926
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 164 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf15 2.028037341700216 2.049760395549724 54.0 2.299999999999997
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 166 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf16 1.9910730364852436 2.006510848093771 53.54 2.759999999999998
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 164 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 164 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf17 2.1567475543719614 2.159142310265706 53.300000000000004 2.999999999999993
-1 gpu conv perf_fp16 164 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 166 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf18 2.1567475543719614 2.159142310265706 53.300000000000004 2.999999999999993
-1 gpu conv perf_fp16 164 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 166 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf19 2.0232690820426464 2.0527698121318476 53.300000000000004 2.999999999999993
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 168 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 11 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/lenet_keras/lenet_keras.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/lenet_keras/lenet_keras.txt
deleted file mode 100644
index b4e51dff426f4d3c5cb7b9572e6aa5940212acbd..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/lenet_keras/lenet_keras.txt
+++ /dev/null
@@ -1,409 +0,0 @@
-282.5141369999999
-+++++
-conf1 1 1 98.7 0.0
-1 gpu conv fp32 11 add fp32 1 pool_max fp32 1 tanh fp32 1
-2 gpu conv fp32 11 add fp32 1 pool_max fp32 1 tanh fp32 1
-3 gpu mul fp32 11 add fp32 1 tanh fp32 1
-4 gpu mul fp32 11 add fp32 1 tanh fp32 1
-5 gpu softmax fp32 1
------
-+++++
-conf2 1.828613181003043 2.071721708828981 98.65 0.04999999999999716
-1 gpu conv perf_fp16 156 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf3 1.8936889628815377 2.139779619692146 98.65 0.04999999999999716
-1 gpu conv perf_fp16 152 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf4 1.8936889628815377 2.139779619692146 98.65 0.04999999999999716
-1 gpu conv perf_fp16 152 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf5 1.8936889628815377 2.139779619692146 98.65 0.04999999999999716
-1 gpu conv perf_fp16 152 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf6 1.8247639611533713 2.0227145446958756 98.64 0.060000000000002274
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf7 1.8247639611533713 2.0227145446958756 98.64 0.060000000000002274
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf8 1.8406161850501603 2.037849502542524 98.64 0.060000000000002274
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf9 1.8406161850501603 2.037849502542524 98.64 0.060000000000002274
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf10 1.8406161850501603 2.037849502542524 98.64 0.060000000000002274
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf11 1.8663357888260776 2.115790921611576 98.64 0.060000000000002274
-1 gpu conv perf_fp16 155 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf12 1.8663357888260776 2.115790921611576 98.64 0.060000000000002274
-1 gpu conv perf_fp16 155 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf13 1.8663357888260776 2.115790921611576 98.64 0.060000000000002274
-1 gpu conv perf_fp16 155 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf14 1.8645645142051612 2.1037012333044935 98.61999999999999 0.0800000000000125
-1 gpu conv perf_fp16 167 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf15 1.8645645142051612 2.1037012333044935 98.61999999999999 0.0800000000000125
-1 gpu conv perf_fp16 167 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf16 1.8645645142051612 2.1037012333044935 98.61999999999999 0.0800000000000125
-1 gpu conv perf_fp16 167 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf17 2.2168527051833635 2.453341076720038 98.61999999999999 0.0800000000000125
-1 gpu conv samp_fp16 264 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf18 2.2168527051833635 2.453341076720038 98.61999999999999 0.0800000000000125
-1 gpu conv samp_fp16 264 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf19 2.2168527051833635 2.453341076720038 98.61999999999999 0.0800000000000125
-1 gpu conv samp_fp16 264 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf20 1.8406161850501603 2.037849502542524 98.6 0.10000000000000853
-1 gpu conv fp16 12 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf21 1.8406161850501603 2.037849502542524 98.6 0.10000000000000853
-1 gpu conv fp16 12 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf22 1.8406161850501603 2.037849502542524 98.6 0.10000000000000853
-1 gpu conv fp16 12 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf23 1.8406161850501603 2.037849502542524 98.6 0.10000000000000853
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf24 1.8406161850501603 2.037849502542524 98.6 0.10000000000000853
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf25 1.8406161850501603 2.037849502542524 98.6 0.10000000000000853
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf26 2.200653361151419 2.425091789360736 98.6 0.10000000000000853
-1 gpu conv samp_fp16 266 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf27 2.200653361151419 2.425091789360736 98.6 0.10000000000000853
-1 gpu conv samp_fp16 266 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf28 1.8406161850501603 2.037849502542524 98.58 0.12000000000000455
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf29 1.8406161850501603 2.037849502542524 98.58 0.12000000000000455
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf30 1.8406161850501603 2.037849502542524 98.58 0.12000000000000455
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf31 1.8445326456180258 2.087601822059355 98.58 0.12000000000000455
-1 gpu conv perf_fp16 156 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf32 1.8445326456180258 2.087601822059355 98.58 0.12000000000000455
-1 gpu conv perf_fp16 156 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf33 1.8445326456180258 2.087601822059355 98.58 0.12000000000000455
-1 gpu conv perf_fp16 156 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf34 1.8916677984300285 2.155437579874673 98.58 0.12000000000000455
-1 gpu conv perf_fp16 158 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf35 1.8916677984300285 2.155437579874673 98.58 0.12000000000000455
-1 gpu conv perf_fp16 158 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf36 1.8916677984300285 2.155437579874673 98.58 0.12000000000000455
-1 gpu conv perf_fp16 158 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf37 1.8649226857257986 2.1076025277601325 98.56 0.14000000000000057
-1 gpu conv perf_fp16 168 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf38 1.8649226857257986 2.1076025277601325 98.56 0.14000000000000057
-1 gpu conv perf_fp16 168 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf39 1.8649226857257986 2.1076025277601325 98.56 0.14000000000000057
-1 gpu conv perf_fp16 168 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf40 1.8463058650555446 2.067271423078985 98.56 0.14000000000000057
-1 gpu conv perf_fp16 157 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf41 1.8463058650555446 2.067271423078985 98.56 0.14000000000000057
-1 gpu conv perf_fp16 157 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf42 1.8463058650555446 2.067271423078985 98.56 0.14000000000000057
-1 gpu conv perf_fp16 157 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf43 1.9234076467497994 2.1864740913112275 98.56 0.14000000000000057
-1 gpu conv perf_fp16 153 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf44 1.9234076467497994 2.1864740913112275 98.56 0.14000000000000057
-1 gpu conv perf_fp16 153 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf45 1.9234076467497994 2.1864740913112275 98.56 0.14000000000000057
-1 gpu conv perf_fp16 153 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf46 1.8698191484268973 2.13979218727595 98.54 0.1599999999999966
-1 gpu conv perf_fp16 159 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf47 1.8698191484268973 2.13979218727595 98.54 0.1599999999999966
-1 gpu conv perf_fp16 159 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf48 1.8575043605938137 2.092057786757256 98.52 0.18000000000000682
-1 gpu conv perf_fp16 165 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf49 1.8575043605938137 2.092057786757256 98.52 0.18000000000000682
-1 gpu conv perf_fp16 165 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf50 1.8575043605938137 2.092057786757256 98.52 0.18000000000000682
-1 gpu conv perf_fp16 165 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf51 1.8534621507951072 2.1231113105788597 98.44000000000001 0.2599999999999909
-1 gpu conv perf_fp16 159 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/mobilenet_cifar10/mobilenet_cifar10.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/mobilenet_cifar10/mobilenet_cifar10.txt
deleted file mode 100644
index b4d8bd893c8d9395fce6a3484d75f543f1e72da2..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/mobilenet_cifar10/mobilenet_cifar10.txt
+++ /dev/null
@@ -1,3220 +0,0 @@
-4077.307063200001
-+++++
-conf1 1 1 84.42 0.0
-1 gpu conv fp32 11
-2 gpu batchnorm fp32 11
-3 gpu relu fp32 11
-4 gpu group_conv fp32 11
-5 gpu batchnorm fp32 11
-6 gpu relu fp32 11
-7 gpu conv fp32 11
-8 gpu batchnorm fp32 11
-9 gpu relu fp32 11
-10 gpu group_conv fp32 11
-11 gpu batchnorm fp32 11
-12 gpu relu fp32 11
-13 gpu conv fp32 11
-14 gpu batchnorm fp32 11
-15 gpu relu fp32 11
-16 gpu group_conv fp32 11
-17 gpu batchnorm fp32 11
-18 gpu relu fp32 11
-19 gpu conv fp32 11
-20 gpu batchnorm fp32 11
-21 gpu relu fp32 11
-22 gpu group_conv fp32 11
-23 gpu batchnorm fp32 11
-24 gpu relu fp32 11
-25 gpu conv fp32 11
-26 gpu batchnorm fp32 11
-27 gpu relu fp32 11
-28 gpu group_conv fp32 11
-29 gpu batchnorm fp32 11
-30 gpu relu fp32 11
-31 gpu conv fp32 11
-32 gpu batchnorm fp32 11
-33 gpu relu fp32 11
-34 gpu group_conv fp32 11
-35 gpu batchnorm fp32 11
-36 gpu relu fp32 11
-37 gpu conv fp32 11
-38 gpu batchnorm fp32 11
-39 gpu relu fp32 11
-40 gpu group_conv fp32 11
-41 gpu batchnorm fp32 11
-42 gpu relu fp32 11
-43 gpu conv fp32 11
-44 gpu batchnorm fp32 11
-45 gpu relu fp32 11
-46 gpu group_conv fp32 11
-47 gpu batchnorm fp32 11
-48 gpu relu fp32 11
-49 gpu conv fp32 11
-50 gpu batchnorm fp32 11
-51 gpu relu fp32 11
-52 gpu group_conv fp32 11
-53 gpu batchnorm fp32 11
-54 gpu relu fp32 11
-55 gpu conv fp32 11
-56 gpu batchnorm fp32 11
-57 gpu relu fp32 11
-58 gpu group_conv fp32 11
-59 gpu batchnorm fp32 11
-60 gpu relu fp32 11
-61 gpu conv fp32 11
-62 gpu batchnorm fp32 11
-63 gpu relu fp32 11
-64 gpu group_conv fp32 11
-65 gpu batchnorm fp32 11
-66 gpu relu fp32 11
-67 gpu conv fp32 11
-68 gpu batchnorm fp32 11
-69 gpu relu fp32 11
-70 gpu group_conv fp32 11
-71 gpu batchnorm fp32 11
-72 gpu relu fp32 11
-73 gpu conv fp32 11
-74 gpu batchnorm fp32 11
-75 gpu relu fp32 11
-76 gpu group_conv fp32 11
-77 gpu batchnorm fp32 11
-78 gpu relu fp32 11
-79 gpu conv fp32 11
-80 gpu batchnorm fp32 11
-81 gpu relu fp32 11
-82 gpu pool_mean fp32 11
-83 gpu mul fp32 11 add fp32 1
-84 gpu softmax fp32 1
------
-+++++
-conf2 1.4930855091460031 1.447990050940341 83.72 0.7000000000000028
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv fp16 12
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf3 1.493397883226807 1.449591062426989 83.72 0.7000000000000028
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 163
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf4 1.4934429016801338 1.4500582352111675 83.72 0.7000000000000028
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 168
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf5 1.4938214813031556 1.450038222978811 83.72 0.7000000000000028
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 157
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf6 1.4933879828131855 1.449975636202813 83.72 0.7000000000000028
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 160
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf7 1.492663093331302 1.4487067754520524 83.7 0.7199999999999989
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 12
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 167
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf8 1.495724395088184 1.4507925552157772 83.56 0.8599999999999994
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 12
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 162
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf9 1.496506307637598 1.4521705950285135 83.36 1.0600000000000023
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 162
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf10 1.496532672928805 1.4521696542076958 83.36 1.0600000000000023
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 156
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf11 1.4988418058849937 1.4555327556053628 83.28 1.1400000000000006
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 168
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 12
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 164
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 158
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf12 1.4994289979945077 1.4562439330251535 83.28 1.1400000000000006
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 168
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 12
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 152
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 153
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf13 1.4952028793065038 1.450369851058777 83.14 1.2800000000000011
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 162
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 151
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 155
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 156
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 152
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 152
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf14 1.4933978285280285 1.448265686258097 83.12 1.2999999999999972
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv fp16 12
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 158
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf15 1.491958833559989 1.4459262032919467 83.08 1.3400000000000034
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 12
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 157
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 155
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv fp16 12
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 152
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 152
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf16 1.4937317297990984 1.4498121856525021 83.02000000000001 1.3999999999999915
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 156
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 158
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf17 1.4963413808686974 1.4522391736954623 82.86 1.5600000000000023
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 165
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf18 1.4942172827099065 1.4504631324933321 82.86 1.5600000000000023
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 157
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 158
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf19 1.4963964073376739 1.4525461321361477 82.86 1.5600000000000023
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 158
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf20 1.4932583049858652 1.4472547227714012 82.84 1.5799999999999983
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv samp_fp16 266
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf21 1.4964326545281064 1.4526263046333605 82.82000000000001 1.5999999999999943
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 152
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 158
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf22 1.4966042483929347 1.4527859961226985 82.82000000000001 1.5999999999999943
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 152
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 153
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf23 1.4966008974318024 1.4527415844509437 82.78 1.6400000000000006
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 155
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 158
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf24 1.4932738366973777 1.448820445466833 82.64 1.7800000000000011
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 164
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 151
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 155
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 156
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 157
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 152
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf25 1.4940402684133964 1.447332235394843 82.48 1.9399999999999977
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 12
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv samp_fp16 261
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf26 1.4981764588414919 1.4530714150549078 82.39999999999999 2.0200000000000102
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 152
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 151
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 161
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 156
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 152
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 152
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf27 1.5004334658773033 1.4549115105608688 82.3 2.1200000000000045
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 152
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 151
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 155
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 156
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 152
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 156
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf28 1.5006808163336343 1.4553824345285296 82.3 2.1200000000000045
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 152
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 151
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 155
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 156
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 152
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 152
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf29 1.4999870719460484 1.4571625511374704 82.28 2.1400000000000006
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 168
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 152
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 165
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf30 1.500042366879961 1.4574715946270216 82.28 2.1400000000000006
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 168
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 152
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 158
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf31 1.500214789632402 1.4576323532660131 82.28 2.1400000000000006
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 168
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 163
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 164
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 152
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 153
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf32 1.4927009086066445 1.4484049211953174 82.26 2.1599999999999966
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 164
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 151
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 161
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 156
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 152
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 152
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf33 1.5003438014588875 1.4538240352408085 82.22 2.200000000000003
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 152
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 151
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 155
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv fp16 12
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 152
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 152
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf34 1.5041587978616728 1.4610492456195174 82.02000000000001 2.3999999999999915
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 168
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 152
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 161
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 155
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 152
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 152
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 158
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf35 1.5000040131742656 1.4555601139156464 81.88 2.5400000000000063
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 152
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 151
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 155
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 152
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 167
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf36 1.4950571524902583 1.451478376045808 81.84 2.5799999999999983
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 164
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 12
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv perf_fp16 161
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 161
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 155
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 155
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 152
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 153
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf37 1.4975271575548847 1.4532126224638244 81.44 2.980000000000004
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 164
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 12
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 11
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 155
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 155
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 152
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 153
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/resnet18_cifar10/resnet18_cifar10.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/resnet18_cifar10/resnet18_cifar10.txt
deleted file mode 100644
index 654cffbf632686dca6310a93ecf56b6521e32039..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/resnet18_cifar10/resnet18_cifar10.txt
+++ /dev/null
@@ -1,2296 +0,0 @@
-2484.981244
-+++++
-conf1 1 1 89.56 0.0
-1 gpu conv fp32 11 add fp32 1 relu fp32 1
-2 gpu conv fp32 11 add fp32 1 relu fp32 1
-3 gpu conv fp32 11 add fp32 1
-4 gpu add fp32 11
-5 gpu relu fp32 11
-6 gpu conv fp32 11 add fp32 1 relu fp32 1
-7 gpu conv fp32 11 add fp32 1
-8 gpu add fp32 11
-9 gpu relu fp32 11
-10 gpu conv fp32 11 add fp32 1 relu fp32 1
-11 gpu conv fp32 11 add fp32 1
-12 gpu add fp32 11
-13 gpu relu fp32 11
-14 gpu conv fp32 11 add fp32 1 relu fp32 1
-15 gpu conv fp32 11 add fp32 1
-16 gpu conv fp32 11 add fp32 1
-17 gpu add fp32 11
-18 gpu relu fp32 11
-19 gpu conv fp32 11 add fp32 1 relu fp32 1
-20 gpu conv fp32 11 add fp32 1
-21 gpu add fp32 11
-22 gpu relu fp32 11
-23 gpu conv fp32 11 add fp32 1 relu fp32 1
-24 gpu conv fp32 11 add fp32 1
-25 gpu add fp32 11
-26 gpu relu fp32 11
-27 gpu conv fp32 11 add fp32 1 relu fp32 1
-28 gpu conv fp32 11 add fp32 1
-29 gpu conv fp32 11 add fp32 1
-30 gpu add fp32 11
-31 gpu relu fp32 11
-32 gpu conv fp32 11 add fp32 1 relu fp32 1
-33 gpu conv fp32 11 add fp32 1
-34 gpu add fp32 11
-35 gpu relu fp32 11
-36 gpu conv fp32 11 add fp32 1 relu fp32 1
-37 gpu conv fp32 11 add fp32 1
-38 gpu add fp32 11
-39 gpu relu fp32 11
-40 gpu pool_mean fp32 11
-41 gpu mul fp32 11 add fp32 1
-42 gpu softmax fp32 1
------
-+++++
-conf2 1.767527790869615 1.7962938589450996 88.96 0.6000000000000085
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 162 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 167 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 167 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 160 add fp16 1 relu fp16 1
-20 gpu conv perf_fp16 155 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv fp16 12 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 160 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 155 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf3 1.7676486174436143 1.7967155014984917 88.78 0.7800000000000011
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 162 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 167 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 167 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv perf_fp16 160 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 160 add fp16 1 relu fp16 1
-20 gpu conv perf_fp16 155 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv fp16 12 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 160 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 155 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf4 1.7674352647250422 1.792910560846682 88.7 0.8599999999999994
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 162 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 167 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 167 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 168 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv fp16 12 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 160 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 155 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf5 1.8655703338511067 1.8930089896922888 88.53999999999999 1.0200000000000102
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 167 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 158 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv perf_fp16 159 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 165 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 157 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf6 1.9070428103729684 1.9172857853336078 88.38000000000001 1.1799999999999926
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 160 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 157 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 152 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv samp_fp16 266 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-20 gpu conv perf_fp16 152 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv samp_fp16 261 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 155 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf7 1.769778590701739 1.7956222622694236 88.24 1.3200000000000074
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv fp16 12 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv samp_fp16 268 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf8 1.841404652091802 1.8677947628418006 88.24 1.3200000000000074
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 168 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 162 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf9 1.8679349428783487 1.8995927920729931 88.22 1.3400000000000034
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 159 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 160 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 168 add fp16 1 relu fp16 1
-20 gpu conv perf_fp16 161 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf10 1.876937310100899 1.9041581451399825 88.1 1.460000000000008
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 158 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf11 1.842140004857965 1.8673692956620238 88.06 1.5
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 166 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 167 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf12 1.9070567138857761 1.9165525910492667 88.02 1.5400000000000063
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 160 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 157 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 152 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv samp_fp16 266 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 261 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 152 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 155 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf13 1.9185835698271805 1.9328202469403 87.98 1.5799999999999983
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 160 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 157 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 152 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv samp_fp16 266 add fp16 1
-16 gpu conv perf_fp16 160 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-20 gpu conv perf_fp16 152 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 152 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 155 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf14 1.781744853993609 1.8082995958456516 87.92 1.6400000000000006
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 166 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 160 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 168 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv perf_fp16 159 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 165 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv samp_fp16 265 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv samp_fp16 268 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf15 1.9185835698271805 1.9328202469403 87.92 1.6400000000000006
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 160 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 157 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 152 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv samp_fp16 266 add fp16 1
-16 gpu conv perf_fp16 160 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-20 gpu conv perf_fp16 152 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 152 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 12 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 155 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf16 1.875261840315855 1.8986912653657988 87.88 1.6800000000000068
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 159 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 156 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 12 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf17 1.9013559086026153 1.9230901214481015 87.86 1.7000000000000028
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf18 1.9185835698271805 1.9328202469403 87.83999999999999 1.720000000000013
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 160 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 157 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 152 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv samp_fp16 266 add fp16 1
-16 gpu conv perf_fp16 160 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-20 gpu conv perf_fp16 152 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 152 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 155 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf19 1.8770503055325798 1.9007923328014182 87.82 1.740000000000009
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 162 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 158 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 151 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf20 1.8774136276932418 1.90365663123621 87.82 1.740000000000009
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 158 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf21 1.943143041264842 1.9591958561422729 87.82 1.740000000000009
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf22 1.870789918969847 1.8863625217899933 87.8 1.7600000000000051
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 264 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf23 1.7445941809066292 1.7754934270309912 87.78 1.7800000000000011
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 162 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 167 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 167 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv perf_fp16 160 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 160 add fp16 1 relu fp16 1
-20 gpu conv perf_fp16 155 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv fp16 12 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 160 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv perf_fp16 166 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 155 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf24 1.9065930313550916 1.928938946228637 87.78 1.7800000000000011
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 167 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf25 1.9021824494907031 1.9237134505552098 87.78 1.7800000000000011
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 154 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf26 1.9017271009017505 1.9211078231701697 87.78 1.7800000000000011
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 162 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf27 1.8187224917656395 1.820406007609536 87.76 1.7999999999999972
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv samp_fp16 264 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf28 1.9070855899343322 1.9285210655709735 87.76 1.7999999999999972
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv samp_fp16 268 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf29 1.9013559086026153 1.9230901214481015 87.74 1.8200000000000074
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf30 1.8772990284718367 1.9022146647342513 87.72 1.8400000000000034
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 162 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 158 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf31 1.9013559086026153 1.9230901214481015 87.68 1.8799999999999955
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf32 1.9020502478364545 1.923319572598976 87.66000000000001 1.8999999999999915
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf33 1.7516394053514481 1.7809034526471939 87.62 1.9399999999999977
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 157 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 162 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 167 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 167 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv perf_fp16 160 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 160 add fp16 1 relu fp16 1
-20 gpu conv perf_fp16 155 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv fp16 12 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 160 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv perf_fp16 166 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 155 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf34 1.7814953252955337 1.8122658147993431 87.62 1.9399999999999977
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 162 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 167 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 167 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv perf_fp16 160 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 160 add fp16 1 relu fp16 1
-20 gpu conv perf_fp16 155 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv fp16 12 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 160 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv perf_fp16 166 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 155 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf35 1.887538247557846 1.9103369445911678 87.62 1.9399999999999977
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 158 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 159 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf36 1.9107566783735581 1.9273803227885578 87.6 1.960000000000008
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 157 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf37 1.9013559086026153 1.9230901214481015 87.58 1.980000000000004
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 12 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf38 1.8984089819969947 1.9195632881772446 87.58 1.980000000000004
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 160 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf39 1.9020502478364545 1.923319572598976 87.52 2.0400000000000063
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf40 1.9020502478364545 1.923319572598976 87.52 2.0400000000000063
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf41 1.9013559086026153 1.9230901214481015 87.5 2.0600000000000023
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf42 1.9013559086026153 1.9230901214481015 87.46000000000001 2.0999999999999943
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 11 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf43 1.9196179152539186 1.9443459719929068 87.44 2.1200000000000045
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 153 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf44 1.9020502478364545 1.923319572598976 87.4 2.1599999999999966
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf45 1.9152817031040366 1.9357432559063958 87.4 2.1599999999999966
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf46 1.915754791147898 1.9373322475753219 87.4 2.1599999999999966
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf47 1.9130551004051772 1.9409232417921056 87.38 2.180000000000007
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv perf_fp16 153 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf48 1.9421147660673033 1.9584555432766413 87.38 2.180000000000007
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf49 1.9052849920081363 1.9300100333661123 87.32 2.240000000000009
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 153 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf50 1.9154322863033566 1.934908329027621 87.3 2.260000000000005
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv perf_fp16 151 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
-+++++
-conf51 1.9079703554020564 1.9287218218306195 86.96000000000001 2.5999999999999943
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv perf_fp16 153 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 161 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv perf_fp16 154 add fp16 1 relu fp16 1
-11 gpu conv perf_fp16 151 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 11 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-20 gpu conv samp_fp16 262 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv perf_fp16 158 add fp16 1 relu fp16 1
-24 gpu conv perf_fp16 153 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv samp_fp16 261 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv perf_fp16 152 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/resnet50_imagenet/resnet50_imagenet.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/resnet50_imagenet/resnet50_imagenet.txt
deleted file mode 100644
index 094eed413b520f9dd661797b96735438861d1c08..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/resnet50_imagenet/resnet50_imagenet.txt
+++ /dev/null
@@ -1,1057 +0,0 @@
-7161.053769000008
-+++++
-conf1 1 1 75.7 0.0
-1 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-2 gpu batchnorm fp32 11
-3 gpu conv fp32 11 add fp32 1
-4 gpu batchnorm fp32 11
-5 gpu relu fp32 11
-6 gpu conv fp32 11 add fp32 1
-7 gpu batchnorm fp32 11
-8 gpu relu fp32 11
-9 gpu conv fp32 11 add fp32 1
-10 gpu batchnorm fp32 11
-11 gpu conv fp32 11 add fp32 1
-12 gpu batchnorm fp32 11
-13 gpu add fp32 11
-14 gpu relu fp32 11
-15 gpu conv fp32 11 add fp32 1
-16 gpu batchnorm fp32 11
-17 gpu relu fp32 11
-18 gpu conv fp32 11 add fp32 1
-19 gpu batchnorm fp32 11
-20 gpu relu fp32 11
-21 gpu conv fp32 11 add fp32 1
-22 gpu batchnorm fp32 11
-23 gpu add fp32 11
-24 gpu relu fp32 11
-25 gpu conv fp32 11 add fp32 1
-26 gpu batchnorm fp32 11
-27 gpu relu fp32 11
-28 gpu conv fp32 11 add fp32 1
-29 gpu batchnorm fp32 11
-30 gpu relu fp32 11
-31 gpu conv fp32 11 add fp32 1
-32 gpu batchnorm fp32 11
-33 gpu add fp32 11
-34 gpu relu fp32 11
-35 gpu conv fp32 11 add fp32 1
-36 gpu batchnorm fp32 11
-37 gpu relu fp32 11
-38 gpu conv fp32 11 add fp32 1
-39 gpu batchnorm fp32 11
-40 gpu relu fp32 11
-41 gpu conv fp32 11 add fp32 1
-42 gpu batchnorm fp32 11
-43 gpu conv fp32 11 add fp32 1
-44 gpu batchnorm fp32 11
-45 gpu add fp32 11
-46 gpu relu fp32 11
-47 gpu conv fp32 11 add fp32 1
-48 gpu batchnorm fp32 11
-49 gpu relu fp32 11
-50 gpu conv fp32 11 add fp32 1
-51 gpu batchnorm fp32 11
-52 gpu relu fp32 11
-53 gpu conv fp32 11 add fp32 1
-54 gpu batchnorm fp32 11
-55 gpu add fp32 11
-56 gpu relu fp32 11
-57 gpu conv fp32 11 add fp32 1
-58 gpu batchnorm fp32 11
-59 gpu relu fp32 11
-60 gpu conv fp32 11 add fp32 1
-61 gpu batchnorm fp32 11
-62 gpu relu fp32 11
-63 gpu conv fp32 11 add fp32 1
-64 gpu batchnorm fp32 11
-65 gpu add fp32 11
-66 gpu relu fp32 11
-67 gpu conv fp32 11 add fp32 1
-68 gpu batchnorm fp32 11
-69 gpu relu fp32 11
-70 gpu conv fp32 11 add fp32 1
-71 gpu batchnorm fp32 11
-72 gpu relu fp32 11
-73 gpu conv fp32 11 add fp32 1
-74 gpu batchnorm fp32 11
-75 gpu add fp32 11
-76 gpu relu fp32 11
-77 gpu conv fp32 11 add fp32 1
-78 gpu batchnorm fp32 11
-79 gpu relu fp32 11
-80 gpu conv fp32 11 add fp32 1
-81 gpu batchnorm fp32 11
-82 gpu relu fp32 11
-83 gpu conv fp32 11 add fp32 1
-84 gpu batchnorm fp32 11
-85 gpu conv fp32 11 add fp32 1
-86 gpu batchnorm fp32 11
-87 gpu add fp32 11
-88 gpu relu fp32 11
-89 gpu conv fp32 11 add fp32 1
-90 gpu batchnorm fp32 11
-91 gpu relu fp32 11
-92 gpu conv fp32 11 add fp32 1
-93 gpu batchnorm fp32 11
-94 gpu relu fp32 11
-95 gpu conv fp32 11 add fp32 1
-96 gpu batchnorm fp32 11
-97 gpu add fp32 11
-98 gpu relu fp32 11
-99 gpu conv fp32 11 add fp32 1
-100 gpu batchnorm fp32 11
-101 gpu relu fp32 11
-102 gpu conv fp32 11 add fp32 1
-103 gpu batchnorm fp32 11
-104 gpu relu fp32 11
-105 gpu conv fp32 11 add fp32 1
-106 gpu batchnorm fp32 11
-107 gpu add fp32 11
-108 gpu relu fp32 11
-109 gpu conv fp32 11 add fp32 1
-110 gpu batchnorm fp32 11
-111 gpu relu fp32 11
-112 gpu conv fp32 11 add fp32 1
-113 gpu batchnorm fp32 11
-114 gpu relu fp32 11
-115 gpu conv fp32 11 add fp32 1
-116 gpu batchnorm fp32 11
-117 gpu add fp32 11
-118 gpu relu fp32 11
-119 gpu conv fp32 11 add fp32 1
-120 gpu batchnorm fp32 11
-121 gpu relu fp32 11
-122 gpu conv fp32 11 add fp32 1
-123 gpu batchnorm fp32 11
-124 gpu relu fp32 11
-125 gpu conv fp32 11 add fp32 1
-126 gpu batchnorm fp32 11
-127 gpu add fp32 11
-128 gpu relu fp32 11
-129 gpu conv fp32 11 add fp32 1
-130 gpu batchnorm fp32 11
-131 gpu relu fp32 11
-132 gpu conv fp32 11 add fp32 1
-133 gpu batchnorm fp32 11
-134 gpu relu fp32 11
-135 gpu conv fp32 11 add fp32 1
-136 gpu batchnorm fp32 11
-137 gpu add fp32 11
-138 gpu relu fp32 11
-139 gpu conv fp32 11 add fp32 1
-140 gpu batchnorm fp32 11
-141 gpu relu fp32 11
-142 gpu conv fp32 11 add fp32 1
-143 gpu batchnorm fp32 11
-144 gpu relu fp32 11
-145 gpu conv fp32 11 add fp32 1
-146 gpu batchnorm fp32 11
-147 gpu conv fp32 11 add fp32 1
-148 gpu batchnorm fp32 11
-149 gpu add fp32 11
-150 gpu relu fp32 11
-151 gpu conv fp32 11 add fp32 1
-152 gpu batchnorm fp32 11
-153 gpu relu fp32 11
-154 gpu conv fp32 11 add fp32 1
-155 gpu batchnorm fp32 11
-156 gpu relu fp32 11
-157 gpu conv fp32 11 add fp32 1
-158 gpu batchnorm fp32 11
-159 gpu add fp32 11
-160 gpu relu fp32 11
-161 gpu conv fp32 11 add fp32 1
-162 gpu batchnorm fp32 11
-163 gpu relu fp32 11
-164 gpu conv fp32 11 add fp32 1
-165 gpu batchnorm fp32 11
-166 gpu relu fp32 11
-167 gpu conv fp32 11 add fp32 1
-168 gpu batchnorm fp32 11
-169 gpu add fp32 11
-170 gpu relu fp32 11
-171 gpu pool_max fp32 11
-172 gpu mul fp32 11 add fp32 1
-173 gpu softmax fp32 1
------
-+++++
-conf2 1.8254789092281507 1.4527803526239977 75.7 0.0
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu batchnorm fp16 12
-3 gpu conv fp16 12 add fp16 1
-4 gpu batchnorm fp16 12
-5 gpu relu fp16 12
-6 gpu conv fp16 12 add fp16 1
-7 gpu batchnorm fp16 12
-8 gpu relu fp16 12
-9 gpu conv fp16 12 add fp16 1
-10 gpu batchnorm fp16 12
-11 gpu conv fp16 12 add fp16 1
-12 gpu batchnorm fp16 12
-13 gpu add fp16 12
-14 gpu relu fp16 12
-15 gpu conv fp16 12 add fp16 1
-16 gpu batchnorm fp16 12
-17 gpu relu fp16 12
-18 gpu conv fp16 12 add fp16 1
-19 gpu batchnorm fp16 12
-20 gpu relu fp16 12
-21 gpu conv fp16 12 add fp16 1
-22 gpu batchnorm fp16 12
-23 gpu add fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12 add fp16 1
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu conv fp16 12 add fp16 1
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 12 add fp16 1
-32 gpu batchnorm fp16 12
-33 gpu add fp16 12
-34 gpu relu fp16 12
-35 gpu conv fp16 12 add fp16 1
-36 gpu batchnorm fp16 12
-37 gpu relu fp16 12
-38 gpu conv fp16 12 add fp16 1
-39 gpu batchnorm fp16 12
-40 gpu relu fp16 12
-41 gpu conv fp16 12 add fp16 1
-42 gpu batchnorm fp16 12
-43 gpu conv fp16 12 add fp16 1
-44 gpu batchnorm fp16 12
-45 gpu add fp16 12
-46 gpu relu fp16 12
-47 gpu conv fp16 12 add fp16 1
-48 gpu batchnorm fp16 12
-49 gpu relu fp16 12
-50 gpu conv fp16 12 add fp16 1
-51 gpu batchnorm fp16 12
-52 gpu relu fp16 12
-53 gpu conv fp16 12 add fp16 1
-54 gpu batchnorm fp16 12
-55 gpu add fp16 12
-56 gpu relu fp16 12
-57 gpu conv fp16 12 add fp16 1
-58 gpu batchnorm fp16 12
-59 gpu relu fp16 12
-60 gpu conv fp16 12 add fp16 1
-61 gpu batchnorm fp16 12
-62 gpu relu fp16 12
-63 gpu conv fp16 12 add fp16 1
-64 gpu batchnorm fp16 12
-65 gpu add fp16 12
-66 gpu relu fp16 12
-67 gpu conv fp16 12 add fp16 1
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu conv fp16 12 add fp16 1
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv fp16 12 add fp16 1
-74 gpu batchnorm fp16 12
-75 gpu add fp16 12
-76 gpu relu fp16 12
-77 gpu conv fp16 12 add fp16 1
-78 gpu batchnorm fp16 12
-79 gpu relu fp16 12
-80 gpu conv fp16 12 add fp16 1
-81 gpu batchnorm fp16 12
-82 gpu relu fp16 12
-83 gpu conv fp16 12 add fp16 1
-84 gpu batchnorm fp16 12
-85 gpu conv fp16 12 add fp16 1
-86 gpu batchnorm fp16 12
-87 gpu add fp16 12
-88 gpu relu fp16 12
-89 gpu conv fp16 12 add fp16 1
-90 gpu batchnorm fp16 12
-91 gpu relu fp16 12
-92 gpu conv fp16 12 add fp16 1
-93 gpu batchnorm fp16 12
-94 gpu relu fp16 12
-95 gpu conv fp16 12 add fp16 1
-96 gpu batchnorm fp16 12
-97 gpu add fp16 12
-98 gpu relu fp16 12
-99 gpu conv fp16 12 add fp16 1
-100 gpu batchnorm fp16 12
-101 gpu relu fp16 12
-102 gpu conv fp16 12 add fp16 1
-103 gpu batchnorm fp16 12
-104 gpu relu fp16 12
-105 gpu conv fp16 12 add fp16 1
-106 gpu batchnorm fp16 12
-107 gpu add fp16 12
-108 gpu relu fp16 12
-109 gpu conv fp16 12 add fp16 1
-110 gpu batchnorm fp16 12
-111 gpu relu fp16 12
-112 gpu conv fp16 12 add fp16 1
-113 gpu batchnorm fp16 12
-114 gpu relu fp16 12
-115 gpu conv fp16 12 add fp16 1
-116 gpu batchnorm fp16 12
-117 gpu add fp16 12
-118 gpu relu fp16 12
-119 gpu conv fp16 12 add fp16 1
-120 gpu batchnorm fp16 12
-121 gpu relu fp16 12
-122 gpu conv fp16 12 add fp16 1
-123 gpu batchnorm fp16 12
-124 gpu relu fp16 12
-125 gpu conv fp16 12 add fp16 1
-126 gpu batchnorm fp16 12
-127 gpu add fp16 12
-128 gpu relu fp16 12
-129 gpu conv fp16 12 add fp16 1
-130 gpu batchnorm fp16 12
-131 gpu relu fp16 12
-132 gpu conv fp16 12 add fp16 1
-133 gpu batchnorm fp16 12
-134 gpu relu fp16 12
-135 gpu conv fp16 12 add fp16 1
-136 gpu batchnorm fp16 12
-137 gpu add fp16 12
-138 gpu relu fp16 12
-139 gpu conv fp16 12 add fp16 1
-140 gpu batchnorm fp16 12
-141 gpu relu fp16 12
-142 gpu conv fp16 12 add fp16 1
-143 gpu batchnorm fp16 12
-144 gpu relu fp16 12
-145 gpu conv fp16 12 add fp16 1
-146 gpu batchnorm fp16 12
-147 gpu conv fp16 12 add fp16 1
-148 gpu batchnorm fp16 12
-149 gpu add fp16 12
-150 gpu relu fp16 12
-151 gpu conv fp16 12 add fp16 1
-152 gpu batchnorm fp16 12
-153 gpu relu fp16 12
-154 gpu conv fp16 12 add fp16 1
-155 gpu batchnorm fp16 12
-156 gpu relu fp16 12
-157 gpu conv fp16 12 add fp16 1
-158 gpu batchnorm fp16 12
-159 gpu add fp16 12
-160 gpu relu fp16 12
-161 gpu conv fp16 12 add fp16 1
-162 gpu batchnorm fp16 12
-163 gpu relu fp16 12
-164 gpu conv fp16 12 add fp16 1
-165 gpu batchnorm fp16 12
-166 gpu relu fp16 12
-167 gpu conv fp16 12 add fp16 1
-168 gpu batchnorm fp16 12
-169 gpu add fp16 12
-170 gpu relu fp16 12
-171 gpu pool_max fp16 12
-172 gpu mul fp16 12 add fp16 1
-173 gpu softmax fp32 1
------
-+++++
-conf3 1.8254789092281507 1.4527803526239977 75.7 0.0
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu batchnorm fp16 12
-3 gpu conv fp16 12 add fp16 1
-4 gpu batchnorm fp16 12
-5 gpu relu fp16 12
-6 gpu conv fp16 12 add fp16 1
-7 gpu batchnorm fp16 12
-8 gpu relu fp16 12
-9 gpu conv fp16 12 add fp16 1
-10 gpu batchnorm fp16 12
-11 gpu conv fp16 12 add fp16 1
-12 gpu batchnorm fp16 12
-13 gpu add fp16 12
-14 gpu relu fp16 12
-15 gpu conv fp16 12 add fp16 1
-16 gpu batchnorm fp16 12
-17 gpu relu fp16 12
-18 gpu conv fp16 12 add fp16 1
-19 gpu batchnorm fp16 12
-20 gpu relu fp16 12
-21 gpu conv fp16 12 add fp16 1
-22 gpu batchnorm fp16 12
-23 gpu add fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12 add fp16 1
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu conv fp16 12 add fp16 1
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 12 add fp16 1
-32 gpu batchnorm fp16 12
-33 gpu add fp16 12
-34 gpu relu fp16 12
-35 gpu conv fp16 12 add fp16 1
-36 gpu batchnorm fp16 12
-37 gpu relu fp16 12
-38 gpu conv fp16 12 add fp16 1
-39 gpu batchnorm fp16 12
-40 gpu relu fp16 12
-41 gpu conv fp16 12 add fp16 1
-42 gpu batchnorm fp16 12
-43 gpu conv fp16 12 add fp16 1
-44 gpu batchnorm fp16 12
-45 gpu add fp16 12
-46 gpu relu fp16 12
-47 gpu conv fp16 12 add fp16 1
-48 gpu batchnorm fp16 12
-49 gpu relu fp16 12
-50 gpu conv fp16 12 add fp16 1
-51 gpu batchnorm fp16 12
-52 gpu relu fp16 12
-53 gpu conv fp16 12 add fp16 1
-54 gpu batchnorm fp16 12
-55 gpu add fp16 12
-56 gpu relu fp16 12
-57 gpu conv fp16 12 add fp16 1
-58 gpu batchnorm fp16 12
-59 gpu relu fp16 12
-60 gpu conv fp16 12 add fp16 1
-61 gpu batchnorm fp16 12
-62 gpu relu fp16 12
-63 gpu conv fp16 12 add fp16 1
-64 gpu batchnorm fp16 12
-65 gpu add fp16 12
-66 gpu relu fp16 12
-67 gpu conv fp16 12 add fp16 1
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu conv fp16 12 add fp16 1
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv fp16 12 add fp16 1
-74 gpu batchnorm fp16 12
-75 gpu add fp16 12
-76 gpu relu fp16 12
-77 gpu conv fp16 12 add fp16 1
-78 gpu batchnorm fp16 12
-79 gpu relu fp16 12
-80 gpu conv fp16 12 add fp16 1
-81 gpu batchnorm fp16 12
-82 gpu relu fp16 12
-83 gpu conv fp16 12 add fp16 1
-84 gpu batchnorm fp16 12
-85 gpu conv fp16 12 add fp16 1
-86 gpu batchnorm fp16 12
-87 gpu add fp16 12
-88 gpu relu fp16 12
-89 gpu conv fp16 12 add fp16 1
-90 gpu batchnorm fp16 12
-91 gpu relu fp16 12
-92 gpu conv fp16 12 add fp16 1
-93 gpu batchnorm fp16 12
-94 gpu relu fp16 12
-95 gpu conv fp16 12 add fp16 1
-96 gpu batchnorm fp16 12
-97 gpu add fp16 12
-98 gpu relu fp16 12
-99 gpu conv fp16 12 add fp16 1
-100 gpu batchnorm fp16 12
-101 gpu relu fp16 12
-102 gpu conv fp16 12 add fp16 1
-103 gpu batchnorm fp16 12
-104 gpu relu fp16 12
-105 gpu conv fp16 12 add fp16 1
-106 gpu batchnorm fp16 12
-107 gpu add fp16 12
-108 gpu relu fp16 12
-109 gpu conv fp16 12 add fp16 1
-110 gpu batchnorm fp16 12
-111 gpu relu fp16 12
-112 gpu conv fp16 12 add fp16 1
-113 gpu batchnorm fp16 12
-114 gpu relu fp16 12
-115 gpu conv fp16 12 add fp16 1
-116 gpu batchnorm fp16 12
-117 gpu add fp16 12
-118 gpu relu fp16 12
-119 gpu conv fp16 12 add fp16 1
-120 gpu batchnorm fp16 12
-121 gpu relu fp16 12
-122 gpu conv fp16 12 add fp16 1
-123 gpu batchnorm fp16 12
-124 gpu relu fp16 12
-125 gpu conv fp16 12 add fp16 1
-126 gpu batchnorm fp16 12
-127 gpu add fp16 12
-128 gpu relu fp16 12
-129 gpu conv fp16 12 add fp16 1
-130 gpu batchnorm fp16 12
-131 gpu relu fp16 12
-132 gpu conv fp16 12 add fp16 1
-133 gpu batchnorm fp16 12
-134 gpu relu fp16 12
-135 gpu conv fp16 12 add fp16 1
-136 gpu batchnorm fp16 12
-137 gpu add fp16 12
-138 gpu relu fp16 12
-139 gpu conv fp16 12 add fp16 1
-140 gpu batchnorm fp16 12
-141 gpu relu fp16 12
-142 gpu conv fp16 12 add fp16 1
-143 gpu batchnorm fp16 12
-144 gpu relu fp16 12
-145 gpu conv fp16 12 add fp16 1
-146 gpu batchnorm fp16 12
-147 gpu conv fp16 12 add fp16 1
-148 gpu batchnorm fp16 12
-149 gpu add fp16 12
-150 gpu relu fp16 12
-151 gpu conv fp16 12 add fp16 1
-152 gpu batchnorm fp16 12
-153 gpu relu fp16 12
-154 gpu conv fp16 12 add fp16 1
-155 gpu batchnorm fp16 12
-156 gpu relu fp16 12
-157 gpu conv fp16 12 add fp16 1
-158 gpu batchnorm fp16 12
-159 gpu add fp16 12
-160 gpu relu fp16 12
-161 gpu conv fp16 12 add fp16 1
-162 gpu batchnorm fp16 12
-163 gpu relu fp16 12
-164 gpu conv fp16 12 add fp16 1
-165 gpu batchnorm fp16 12
-166 gpu relu fp16 12
-167 gpu conv fp16 12 add fp16 1
-168 gpu batchnorm fp16 12
-169 gpu add fp16 12
-170 gpu relu fp16 12
-171 gpu pool_max fp16 12
-172 gpu mul fp16 12 add fp16 1
-173 gpu softmax fp32 1
------
-+++++
-conf4 1.8254789092281507 1.4527803526239977 75.7 0.0
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu batchnorm fp16 12
-3 gpu conv fp16 12 add fp16 1
-4 gpu batchnorm fp16 12
-5 gpu relu fp16 12
-6 gpu conv fp16 12 add fp16 1
-7 gpu batchnorm fp16 12
-8 gpu relu fp16 12
-9 gpu conv fp16 12 add fp16 1
-10 gpu batchnorm fp16 12
-11 gpu conv fp16 12 add fp16 1
-12 gpu batchnorm fp16 12
-13 gpu add fp16 12
-14 gpu relu fp16 12
-15 gpu conv fp16 12 add fp16 1
-16 gpu batchnorm fp16 12
-17 gpu relu fp16 12
-18 gpu conv fp16 12 add fp16 1
-19 gpu batchnorm fp16 12
-20 gpu relu fp16 12
-21 gpu conv fp16 12 add fp16 1
-22 gpu batchnorm fp16 12
-23 gpu add fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12 add fp16 1
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu conv fp16 12 add fp16 1
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 12 add fp16 1
-32 gpu batchnorm fp16 12
-33 gpu add fp16 12
-34 gpu relu fp16 12
-35 gpu conv fp16 12 add fp16 1
-36 gpu batchnorm fp16 12
-37 gpu relu fp16 12
-38 gpu conv fp16 12 add fp16 1
-39 gpu batchnorm fp16 12
-40 gpu relu fp16 12
-41 gpu conv fp16 12 add fp16 1
-42 gpu batchnorm fp16 12
-43 gpu conv fp16 12 add fp16 1
-44 gpu batchnorm fp16 12
-45 gpu add fp16 12
-46 gpu relu fp16 12
-47 gpu conv fp16 12 add fp16 1
-48 gpu batchnorm fp16 12
-49 gpu relu fp16 12
-50 gpu conv fp16 12 add fp16 1
-51 gpu batchnorm fp16 12
-52 gpu relu fp16 12
-53 gpu conv fp16 12 add fp16 1
-54 gpu batchnorm fp16 12
-55 gpu add fp16 12
-56 gpu relu fp16 12
-57 gpu conv fp16 12 add fp16 1
-58 gpu batchnorm fp16 12
-59 gpu relu fp16 12
-60 gpu conv fp16 12 add fp16 1
-61 gpu batchnorm fp16 12
-62 gpu relu fp16 12
-63 gpu conv fp16 12 add fp16 1
-64 gpu batchnorm fp16 12
-65 gpu add fp16 12
-66 gpu relu fp16 12
-67 gpu conv fp16 12 add fp16 1
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu conv fp16 12 add fp16 1
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv fp16 12 add fp16 1
-74 gpu batchnorm fp16 12
-75 gpu add fp16 12
-76 gpu relu fp16 12
-77 gpu conv fp16 12 add fp16 1
-78 gpu batchnorm fp16 12
-79 gpu relu fp16 12
-80 gpu conv fp16 12 add fp16 1
-81 gpu batchnorm fp16 12
-82 gpu relu fp16 12
-83 gpu conv fp16 12 add fp16 1
-84 gpu batchnorm fp16 12
-85 gpu conv fp16 12 add fp16 1
-86 gpu batchnorm fp16 12
-87 gpu add fp16 12
-88 gpu relu fp16 12
-89 gpu conv fp16 12 add fp16 1
-90 gpu batchnorm fp16 12
-91 gpu relu fp16 12
-92 gpu conv fp16 12 add fp16 1
-93 gpu batchnorm fp16 12
-94 gpu relu fp16 12
-95 gpu conv fp16 12 add fp16 1
-96 gpu batchnorm fp16 12
-97 gpu add fp16 12
-98 gpu relu fp16 12
-99 gpu conv fp16 12 add fp16 1
-100 gpu batchnorm fp16 12
-101 gpu relu fp16 12
-102 gpu conv fp16 12 add fp16 1
-103 gpu batchnorm fp16 12
-104 gpu relu fp16 12
-105 gpu conv fp16 12 add fp16 1
-106 gpu batchnorm fp16 12
-107 gpu add fp16 12
-108 gpu relu fp16 12
-109 gpu conv fp16 12 add fp16 1
-110 gpu batchnorm fp16 12
-111 gpu relu fp16 12
-112 gpu conv fp16 12 add fp16 1
-113 gpu batchnorm fp16 12
-114 gpu relu fp16 12
-115 gpu conv fp16 12 add fp16 1
-116 gpu batchnorm fp16 12
-117 gpu add fp16 12
-118 gpu relu fp16 12
-119 gpu conv fp16 12 add fp16 1
-120 gpu batchnorm fp16 12
-121 gpu relu fp16 12
-122 gpu conv fp16 12 add fp16 1
-123 gpu batchnorm fp16 12
-124 gpu relu fp16 12
-125 gpu conv fp16 12 add fp16 1
-126 gpu batchnorm fp16 12
-127 gpu add fp16 12
-128 gpu relu fp16 12
-129 gpu conv fp16 12 add fp16 1
-130 gpu batchnorm fp16 12
-131 gpu relu fp16 12
-132 gpu conv fp16 12 add fp16 1
-133 gpu batchnorm fp16 12
-134 gpu relu fp16 12
-135 gpu conv fp16 12 add fp16 1
-136 gpu batchnorm fp16 12
-137 gpu add fp16 12
-138 gpu relu fp16 12
-139 gpu conv fp16 12 add fp16 1
-140 gpu batchnorm fp16 12
-141 gpu relu fp16 12
-142 gpu conv fp16 12 add fp16 1
-143 gpu batchnorm fp16 12
-144 gpu relu fp16 12
-145 gpu conv fp16 12 add fp16 1
-146 gpu batchnorm fp16 12
-147 gpu conv fp16 12 add fp16 1
-148 gpu batchnorm fp16 12
-149 gpu add fp16 12
-150 gpu relu fp16 12
-151 gpu conv fp16 12 add fp16 1
-152 gpu batchnorm fp16 12
-153 gpu relu fp16 12
-154 gpu conv fp16 12 add fp16 1
-155 gpu batchnorm fp16 12
-156 gpu relu fp16 12
-157 gpu conv fp16 12 add fp16 1
-158 gpu batchnorm fp16 12
-159 gpu add fp16 12
-160 gpu relu fp16 12
-161 gpu conv fp16 12 add fp16 1
-162 gpu batchnorm fp16 12
-163 gpu relu fp16 12
-164 gpu conv fp16 12 add fp16 1
-165 gpu batchnorm fp16 12
-166 gpu relu fp16 12
-167 gpu conv fp16 12 add fp16 1
-168 gpu batchnorm fp16 12
-169 gpu add fp16 12
-170 gpu relu fp16 12
-171 gpu pool_max fp16 12
-172 gpu mul fp16 12 add fp16 1
-173 gpu softmax fp32 1
------
-+++++
-conf5 1.8323072136026506 1.457112696128105 74.76 0.9399999999999977
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu batchnorm fp16 12
-3 gpu conv fp16 12 add fp16 1
-4 gpu batchnorm fp16 12
-5 gpu relu fp16 12
-6 gpu conv fp16 12 add fp16 1
-7 gpu batchnorm fp16 12
-8 gpu relu fp16 12
-9 gpu conv fp16 12 add fp16 1
-10 gpu batchnorm fp16 12
-11 gpu conv fp16 12 add fp16 1
-12 gpu batchnorm fp16 12
-13 gpu add fp16 12
-14 gpu relu fp16 12
-15 gpu conv fp16 12 add fp16 1
-16 gpu batchnorm fp16 12
-17 gpu relu fp16 12
-18 gpu conv fp16 12 add fp16 1
-19 gpu batchnorm fp16 12
-20 gpu relu fp16 12
-21 gpu conv fp16 12 add fp16 1
-22 gpu batchnorm fp16 12
-23 gpu add fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12 add fp16 1
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu conv fp16 12 add fp16 1
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 12 add fp16 1
-32 gpu batchnorm fp16 12
-33 gpu add fp16 12
-34 gpu relu fp16 12
-35 gpu conv fp16 12 add fp16 1
-36 gpu batchnorm fp16 12
-37 gpu relu fp16 12
-38 gpu conv fp16 12 add fp16 1
-39 gpu batchnorm fp16 12
-40 gpu relu fp16 12
-41 gpu conv fp16 12 add fp16 1
-42 gpu batchnorm fp16 12
-43 gpu conv fp16 12 add fp16 1
-44 gpu batchnorm fp16 12
-45 gpu add fp16 12
-46 gpu relu fp16 12
-47 gpu conv fp16 12 add fp16 1
-48 gpu batchnorm fp16 12
-49 gpu relu fp16 12
-50 gpu conv fp16 12 add fp16 1
-51 gpu batchnorm fp16 12
-52 gpu relu fp16 12
-53 gpu conv fp16 12 add fp16 1
-54 gpu batchnorm fp16 12
-55 gpu add fp16 12
-56 gpu relu fp16 12
-57 gpu conv fp16 12 add fp16 1
-58 gpu batchnorm fp16 12
-59 gpu relu fp16 12
-60 gpu conv fp16 12 add fp16 1
-61 gpu batchnorm fp16 12
-62 gpu relu fp16 12
-63 gpu conv fp16 12 add fp16 1
-64 gpu batchnorm fp16 12
-65 gpu add fp16 12
-66 gpu relu fp16 12
-67 gpu conv fp16 12 add fp16 1
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu conv fp16 12 add fp16 1
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv fp16 12 add fp16 1
-74 gpu batchnorm fp16 12
-75 gpu add fp16 12
-76 gpu relu fp16 12
-77 gpu conv fp16 12 add fp16 1
-78 gpu batchnorm fp16 12
-79 gpu relu fp16 12
-80 gpu conv fp16 12 add fp16 1
-81 gpu batchnorm fp16 12
-82 gpu relu fp16 12
-83 gpu conv fp16 12 add fp16 1
-84 gpu batchnorm fp16 12
-85 gpu conv fp16 12 add fp16 1
-86 gpu batchnorm fp16 12
-87 gpu add fp16 12
-88 gpu relu fp16 12
-89 gpu conv fp16 12 add fp16 1
-90 gpu batchnorm fp16 12
-91 gpu relu fp16 12
-92 gpu conv fp16 12 add fp16 1
-93 gpu batchnorm fp16 12
-94 gpu relu fp16 12
-95 gpu conv fp16 12 add fp16 1
-96 gpu batchnorm fp16 12
-97 gpu add fp16 12
-98 gpu relu fp16 12
-99 gpu conv perf_fp16 157 add fp16 1
-100 gpu batchnorm fp16 12
-101 gpu relu fp16 12
-102 gpu conv fp16 12 add fp16 1
-103 gpu batchnorm fp16 12
-104 gpu relu fp16 12
-105 gpu conv fp16 12 add fp16 1
-106 gpu batchnorm fp16 12
-107 gpu add fp16 12
-108 gpu relu fp16 12
-109 gpu conv fp16 12 add fp16 1
-110 gpu batchnorm fp16 12
-111 gpu relu fp16 12
-112 gpu conv fp16 12 add fp16 1
-113 gpu batchnorm fp16 12
-114 gpu relu fp16 12
-115 gpu conv fp16 12 add fp16 1
-116 gpu batchnorm fp16 12
-117 gpu add fp16 12
-118 gpu relu fp16 12
-119 gpu conv fp16 12 add fp16 1
-120 gpu batchnorm fp16 12
-121 gpu relu fp16 12
-122 gpu conv fp16 12 add fp16 1
-123 gpu batchnorm fp16 12
-124 gpu relu fp16 12
-125 gpu conv fp16 12 add fp16 1
-126 gpu batchnorm fp16 12
-127 gpu add fp16 12
-128 gpu relu fp16 12
-129 gpu conv fp16 12 add fp16 1
-130 gpu batchnorm fp16 12
-131 gpu relu fp16 12
-132 gpu conv fp16 12 add fp16 1
-133 gpu batchnorm fp16 12
-134 gpu relu fp16 12
-135 gpu conv fp16 12 add fp16 1
-136 gpu batchnorm fp16 12
-137 gpu add fp16 12
-138 gpu relu fp16 12
-139 gpu conv fp16 12 add fp16 1
-140 gpu batchnorm fp16 12
-141 gpu relu fp16 12
-142 gpu conv fp16 12 add fp16 1
-143 gpu batchnorm fp16 12
-144 gpu relu fp16 12
-145 gpu conv fp16 12 add fp16 1
-146 gpu batchnorm fp16 12
-147 gpu conv fp16 12 add fp16 1
-148 gpu batchnorm fp16 12
-149 gpu add fp16 12
-150 gpu relu fp16 12
-151 gpu conv fp16 12 add fp16 1
-152 gpu batchnorm fp16 12
-153 gpu relu fp16 12
-154 gpu conv fp16 12 add fp16 1
-155 gpu batchnorm fp16 12
-156 gpu relu fp16 12
-157 gpu conv fp16 12 add fp16 1
-158 gpu batchnorm fp16 12
-159 gpu add fp16 12
-160 gpu relu fp16 12
-161 gpu conv fp16 12 add fp16 1
-162 gpu batchnorm fp16 12
-163 gpu relu fp16 12
-164 gpu conv perf_fp16 152 add fp16 1
-165 gpu batchnorm fp16 12
-166 gpu relu fp16 12
-167 gpu conv fp16 12 add fp16 1
-168 gpu batchnorm fp16 12
-169 gpu add fp16 12
-170 gpu relu fp16 12
-171 gpu pool_max fp16 12
-172 gpu mul fp16 12 add fp16 1
-173 gpu softmax fp32 1
------
-+++++
-conf6 1.8333922701839533 1.4589203187717397 74.53999999999999 1.1600000000000108
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu batchnorm fp16 12
-3 gpu conv fp16 12 add fp16 1
-4 gpu batchnorm fp16 12
-5 gpu relu fp16 12
-6 gpu conv fp16 12 add fp16 1
-7 gpu batchnorm fp16 12
-8 gpu relu fp16 12
-9 gpu conv fp16 12 add fp16 1
-10 gpu batchnorm fp16 12
-11 gpu conv fp16 12 add fp16 1
-12 gpu batchnorm fp16 12
-13 gpu add fp16 12
-14 gpu relu fp16 12
-15 gpu conv fp16 12 add fp16 1
-16 gpu batchnorm fp16 12
-17 gpu relu fp16 12
-18 gpu conv fp16 12 add fp16 1
-19 gpu batchnorm fp16 12
-20 gpu relu fp16 12
-21 gpu conv fp16 12 add fp16 1
-22 gpu batchnorm fp16 12
-23 gpu add fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12 add fp16 1
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu conv fp16 12 add fp16 1
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 12 add fp16 1
-32 gpu batchnorm fp16 12
-33 gpu add fp16 12
-34 gpu relu fp16 12
-35 gpu conv fp16 12 add fp16 1
-36 gpu batchnorm fp16 12
-37 gpu relu fp16 12
-38 gpu conv fp16 12 add fp16 1
-39 gpu batchnorm fp16 12
-40 gpu relu fp16 12
-41 gpu conv fp16 12 add fp16 1
-42 gpu batchnorm fp16 12
-43 gpu conv fp16 12 add fp16 1
-44 gpu batchnorm fp16 12
-45 gpu add fp16 12
-46 gpu relu fp16 12
-47 gpu conv fp16 12 add fp16 1
-48 gpu batchnorm fp16 12
-49 gpu relu fp16 12
-50 gpu conv fp16 12 add fp16 1
-51 gpu batchnorm fp16 12
-52 gpu relu fp16 12
-53 gpu conv fp16 12 add fp16 1
-54 gpu batchnorm fp16 12
-55 gpu add fp16 12
-56 gpu relu fp16 12
-57 gpu conv fp16 12 add fp16 1
-58 gpu batchnorm fp16 12
-59 gpu relu fp16 12
-60 gpu conv fp16 12 add fp16 1
-61 gpu batchnorm fp16 12
-62 gpu relu fp16 12
-63 gpu conv fp16 12 add fp16 1
-64 gpu batchnorm fp16 12
-65 gpu add fp16 12
-66 gpu relu fp16 12
-67 gpu conv fp16 12 add fp16 1
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu conv fp16 12 add fp16 1
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv fp16 12 add fp16 1
-74 gpu batchnorm fp16 12
-75 gpu add fp16 12
-76 gpu relu fp16 12
-77 gpu conv fp16 12 add fp16 1
-78 gpu batchnorm fp16 12
-79 gpu relu fp16 12
-80 gpu conv fp16 12 add fp16 1
-81 gpu batchnorm fp16 12
-82 gpu relu fp16 12
-83 gpu conv fp16 12 add fp16 1
-84 gpu batchnorm fp16 12
-85 gpu conv fp16 12 add fp16 1
-86 gpu batchnorm fp16 12
-87 gpu add fp16 12
-88 gpu relu fp16 12
-89 gpu conv fp16 12 add fp16 1
-90 gpu batchnorm fp16 12
-91 gpu relu fp16 12
-92 gpu conv fp16 12 add fp16 1
-93 gpu batchnorm fp16 12
-94 gpu relu fp16 12
-95 gpu conv fp16 12 add fp16 1
-96 gpu batchnorm fp16 12
-97 gpu add fp16 12
-98 gpu relu fp16 12
-99 gpu conv perf_fp16 157 add fp16 1
-100 gpu batchnorm fp16 12
-101 gpu relu fp16 12
-102 gpu conv samp_fp16 267 add fp16 1
-103 gpu batchnorm fp16 12
-104 gpu relu fp16 12
-105 gpu conv fp16 12 add fp16 1
-106 gpu batchnorm fp16 12
-107 gpu add fp16 12
-108 gpu relu fp16 12
-109 gpu conv fp16 12 add fp16 1
-110 gpu batchnorm fp16 12
-111 gpu relu fp16 12
-112 gpu conv fp16 12 add fp16 1
-113 gpu batchnorm fp16 12
-114 gpu relu fp16 12
-115 gpu conv fp16 12 add fp16 1
-116 gpu batchnorm fp16 12
-117 gpu add fp16 12
-118 gpu relu fp16 12
-119 gpu conv fp16 12 add fp16 1
-120 gpu batchnorm fp16 12
-121 gpu relu fp16 12
-122 gpu conv fp16 12 add fp16 1
-123 gpu batchnorm fp16 12
-124 gpu relu fp16 12
-125 gpu conv fp16 12 add fp16 1
-126 gpu batchnorm fp16 12
-127 gpu add fp16 12
-128 gpu relu fp16 12
-129 gpu conv fp16 12 add fp16 1
-130 gpu batchnorm fp16 12
-131 gpu relu fp16 12
-132 gpu conv fp16 12 add fp16 1
-133 gpu batchnorm fp16 12
-134 gpu relu fp16 12
-135 gpu conv fp16 12 add fp16 1
-136 gpu batchnorm fp16 12
-137 gpu add fp16 12
-138 gpu relu fp16 12
-139 gpu conv fp16 12 add fp16 1
-140 gpu batchnorm fp16 12
-141 gpu relu fp16 12
-142 gpu conv fp16 12 add fp16 1
-143 gpu batchnorm fp16 12
-144 gpu relu fp16 12
-145 gpu conv fp16 12 add fp16 1
-146 gpu batchnorm fp16 12
-147 gpu conv fp16 12 add fp16 1
-148 gpu batchnorm fp16 12
-149 gpu add fp16 12
-150 gpu relu fp16 12
-151 gpu conv fp16 12 add fp16 1
-152 gpu batchnorm fp16 12
-153 gpu relu fp16 12
-154 gpu conv fp16 12 add fp16 1
-155 gpu batchnorm fp16 12
-156 gpu relu fp16 12
-157 gpu conv fp16 12 add fp16 1
-158 gpu batchnorm fp16 12
-159 gpu add fp16 12
-160 gpu relu fp16 12
-161 gpu conv fp16 12 add fp16 1
-162 gpu batchnorm fp16 12
-163 gpu relu fp16 12
-164 gpu conv perf_fp16 152 add fp16 1
-165 gpu batchnorm fp16 12
-166 gpu relu fp16 12
-167 gpu conv fp16 12 add fp16 1
-168 gpu batchnorm fp16 12
-169 gpu add fp16 12
-170 gpu relu fp16 12
-171 gpu pool_max fp16 12
-172 gpu mul fp16 12 add fp16 1
-173 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/vgg16_cifar10/vgg16_cifar10.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/vgg16_cifar10/vgg16_cifar10.txt
deleted file mode 100644
index 2b325a9fe2d122e74cdd2b80e2768e68591313bf..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/vgg16_cifar10/vgg16_cifar10.txt
+++ /dev/null
@@ -1,913 +0,0 @@
-3776.508929999999
-+++++
-conf1 1 1 89.96 0.0
-1 gpu conv fp32 11 add fp32 1 relu fp32 1
-2 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-3 gpu conv fp32 11 add fp32 1 relu fp32 1
-4 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-5 gpu conv fp32 11 add fp32 1 relu fp32 1
-6 gpu conv fp32 11 add fp32 1 relu fp32 1
-7 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-8 gpu conv fp32 11 add fp32 1 relu fp32 1
-9 gpu conv fp32 11 add fp32 1 relu fp32 1
-10 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-11 gpu conv fp32 11 add fp32 1 relu fp32 1
-12 gpu conv fp32 11 add fp32 1 relu fp32 1
-13 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-14 gpu mul fp32 11 add fp32 1 relu fp32 1
-15 gpu mul fp32 11 add fp32 1
-16 gpu softmax fp32 1
------
-+++++
-conf2 2.1225958306417145 1.9771056444390926 89.91 0.04999999999999716
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 167 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 155 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 164 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 267 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 161 add fp16 1 relu fp16 1
-12 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf3 2.090180991844805 1.9532689756636086 89.82 0.14000000000000057
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 167 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 269 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 155 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 164 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 161 add fp16 1 relu fp16 1
-12 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf4 2.169931036393396 2.0048851858669283 89.53999999999999 0.4200000000000017
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 155 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv perf_fp16 162 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 264 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 269 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf5 2.1012179398201756 1.9325098819632314 89.42 0.539999999999992
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 155 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 264 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv fp16 11 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 269 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf6 2.2313002482945326 2.069581185407626 89.38000000000001 0.5799999999999841
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 158 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 269 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv fp16 12 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 269 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 264 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf7 2.143061101834193 1.9675759235961738 89.3 0.6599999999999966
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 155 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv fp16 12 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 265 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 264 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 269 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf8 2.199379444387758 2.0314348091429677 89.2 0.7599999999999909
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 155 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 264 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf9 2.3236298452294624 2.156907976575644 89.03999999999999 0.9200000000000017
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 159 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv fp16 11 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf10 2.3224369486241603 2.1560351277882046 89.03999999999999 0.9200000000000017
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 159 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv fp16 11 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf11 2.358467412507993 2.1904290636262784 89.02 0.9399999999999977
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 159 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 266 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 264 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf12 2.3633503986583126 2.1980949050120437 88.88000000000001 1.079999999999984
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 159 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 264 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf13 2.4903388172036043 2.3063593441573564 88.82 1.1400000000000006
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf14 2.508156996742662 2.3204109539869595 88.78 1.1799999999999926
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 156 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf15 2.4818531813049622 2.2910866330696744 88.75999999999999 1.2000000000000028
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 263 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf16 2.4591564896606 2.272664410995804 88.74 1.2199999999999989
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 263 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf17 2.5370582721089496 2.3464665753522405 88.72 1.2399999999999949
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf18 2.438100014978735 2.257620696759345 88.7 1.259999999999991
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 263 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf19 2.4776935382337006 2.2949598026093168 88.7 1.259999999999991
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf20 2.4380041604279596 2.254330054479329 88.68 1.279999999999987
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 157 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf21 2.4745444350223327 2.2883888475386525 88.64 1.3199999999999932
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf22 2.4136652022060625 2.2360545757445407 88.52 1.4399999999999977
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv perf_fp16 160 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf23 2.510093966915115 2.316437144001897 88.52 1.4399999999999977
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 266 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf24 2.475990790728594 2.28127562431577 88.5 1.4599999999999937
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv fp16 11 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf25 2.4761929121466926 2.290365501363375 88.5 1.4599999999999937
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf26 2.4763575559033875 2.291312348847263 88.5 1.4599999999999937
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf27 2.600249602991055 2.4123747341424644 88.06 1.8999999999999915
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 165 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf28 2.596077615026303 2.4115375655840245 88.02 1.9399999999999977
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 166 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf29 2.580888020555937 2.3840829703999833 87.88 2.0799999999999983
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv samp_fp16 269 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf30 2.556352783745439 2.3641413704751537 87.8 2.1599999999999966
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv samp_fp16 269 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf31 2.5559756082494527 2.3677471703724575 87.78 2.1799999999999926
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 11 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf32 2.597413373332546 2.4091972878097585 87.76 2.1999999999999886
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 164 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf33 2.4797467027434656 2.2874608793842612 87.74 2.219999999999999
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf34 2.593675604602072 2.400513932866452 87.7 2.259999999999991
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 156 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 264 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf35 2.6300759173431336 2.432687374579977 87.62 2.339999999999989
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 266 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf36 2.5907083037103864 2.4042762580264356 87.6 2.3599999999999994
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 156 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf37 2.6143261650366187 2.423427684623993 87.6 2.3599999999999994
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 156 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf38 2.6144436259117203 2.4231961521843344 87.6 2.3599999999999994
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 156 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf39 2.662088796913144 2.4660859696742032 87.6 2.3599999999999994
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf40 2.6210428708834517 2.423389791646294 87.58 2.3799999999999955
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 156 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 265 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf41 2.6399924349243533 2.4443864221157914 87.58 2.3799999999999955
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 156 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf42 2.616443708384916 2.4217582570150697 87.58 2.3799999999999955
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf43 2.6883473596205225 2.5036952786284137 87.5 2.4599999999999937
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv perf_fp16 166 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 156 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf44 2.6117356623585875 2.420771216556161 87.48 2.4799999999999898
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 156 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf45 2.6359174040106708 2.444231592562593 87.48 2.4799999999999898
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf46 2.56504192294198 2.371871906722655 87.44 2.519999999999996
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv fp16 11 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf47 2.5652588453899727 2.3816996471861174 87.44 2.519999999999996
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf48 2.68806951500876 2.5007647690311425 87.14 2.819999999999993
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv perf_fp16 166 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv perf_fp16 156 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/vgg16_cifar100/vgg16_cifar100.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/vgg16_cifar100/vgg16_cifar100.txt
deleted file mode 100644
index 2c29bedd096aec2c7f66afbe729353e372fac403..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/vgg16_cifar100/vgg16_cifar100.txt
+++ /dev/null
@@ -1,970 +0,0 @@
-3768.819777999999
-+++++
-conf1 1 1 66.5 0.0
-1 gpu conv fp32 11 add fp32 1 relu fp32 1
-2 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-3 gpu conv fp32 11 add fp32 1 relu fp32 1
-4 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-5 gpu conv fp32 11 add fp32 1 relu fp32 1
-6 gpu conv fp32 11 add fp32 1 relu fp32 1
-7 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-8 gpu conv fp32 11 add fp32 1 relu fp32 1
-9 gpu conv fp32 11 add fp32 1 relu fp32 1
-10 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-11 gpu conv fp32 11 add fp32 1 relu fp32 1
-12 gpu conv fp32 11 add fp32 1 relu fp32 1
-13 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-14 gpu mul fp32 11 add fp32 1 relu fp32 1
-15 gpu mul fp32 11 add fp32 1
-16 gpu softmax fp32 1
------
-+++++
-conf2 2.2877724452131787 2.08025704453875 66.45 0.04999999999999716
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 269 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 268 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 162 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 266 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf3 2.5314658805383816 2.30737681453141 66.45 0.04999999999999716
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf4 2.044123178914057 1.8616966918258782 66.32000000000001 0.1799999999999926
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 168 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 268 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 11 add fp16 1 relu fp16 1
-13 gpu conv fp16 11 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf5 2.231179358259141 2.0317825813373864 66.18 0.3199999999999932
-1 gpu conv fp16 11 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 161 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv fp16 12 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 269 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 265 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf6 2.2474834421641057 2.0338639876373272 65.88000000000001 0.6199999999999903
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 268 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 267 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 265 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 267 add fp16 1 relu fp16 1
-13 gpu conv fp16 11 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf7 2.22281439516094 2.0205460706906377 65.88000000000001 0.6199999999999903
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 155 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 268 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 161 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 167 add fp16 1 relu fp16 1
-12 gpu conv perf_fp16 161 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf8 2.1625085012968484 1.94560449637282 65.88000000000001 0.6199999999999903
-1 gpu conv fp16 11 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 266 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv fp16 11 add fp16 1 relu fp16 1
-10 gpu conv fp16 11 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 264 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 263 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf9 2.639337323402163 2.3960416499256825 65.8 0.7000000000000028
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 269 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf10 2.672718090670276 2.4276905528801507 65.68 0.8199999999999932
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf11 2.699089631751789 2.446114054498494 65.68 0.8199999999999932
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf12 2.6003752638648767 2.3553067802112344 65.64 0.8599999999999994
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv fp16 11 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf13 2.638763904718665 2.395072565223988 65.64 0.8599999999999994
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 268 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf14 2.6003752638648767 2.3553067802112344 65.64 0.8599999999999994
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv fp16 11 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf15 2.6003752638648767 2.3553067802112344 65.64 0.8599999999999994
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv fp16 11 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf16 2.6732183804279006 2.4287517162140326 65.62 0.8799999999999955
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf17 2.6728394017929027 2.428768169588016 65.60000000000001 0.8999999999999915
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf18 2.4549989178389238 2.2406620346549433 65.56 0.9399999999999977
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 156 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf19 2.673556689244081 2.429092581627209 65.52 0.980000000000004
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf20 2.6525635304451756 2.406830663552284 65.5 1.0
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 263 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf21 2.6692288605087553 2.423462800937785 65.5 1.0
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf22 2.583650505571873 2.3471533059252194 65.48 1.019999999999996
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 160 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 263 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf23 2.6474572655420125 2.400471260394867 65.48 1.019999999999996
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 265 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf24 2.4710116424304736 2.2555966923178996 65.46 1.0400000000000063
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 161 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 266 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf25 2.557911102074785 2.3292661683311526 65.46 1.0400000000000063
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 156 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf26 2.6032957018479532 2.367574146141511 65.44 1.0600000000000023
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf27 2.6029968728098916 2.3672068592437223 65.44 1.0600000000000023
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 164 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf28 2.602540311129756 2.3691028781436954 65.44 1.0600000000000023
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 167 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf29 2.602756708588441 2.3708111025211718 65.44 1.0600000000000023
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 168 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf30 2.603240857443844 2.3662875785790183 65.44 1.0600000000000023
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 157 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf31 2.602882717372841 2.368011704225619 65.44 1.0600000000000023
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 160 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf32 2.67999343314603 2.4305182001043826 65.4 1.0999999999999943
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf33 2.670314990364046 2.4275308713267485 65.38000000000001 1.1199999999999903
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf34 2.650982630033638 2.405821467700663 65.36 1.1400000000000006
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 263 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf35 2.6507266317871756 2.405938171802741 65.36 1.1400000000000006
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 265 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf36 2.6523068534836174 2.406695716686769 65.34 1.1599999999999966
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 264 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf37 2.6533198495191073 2.4077689394073865 65.34 1.1599999999999966
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 264 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf38 2.64630900155657 2.4073892305914986 65.32 1.1800000000000068
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf39 2.6725522534379413 2.42903505877629 65.32 1.1800000000000068
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf40 2.6435249267602225 2.403536258709464 65.3 1.2000000000000028
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 161 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf41 2.6442059720503557 2.4037376163252024 65.3 1.2000000000000028
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf42 2.6536933126724027 2.4077527693156053 65.3 1.2000000000000028
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 264 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf43 2.6442798101298948 2.4056031584129225 65.3 1.2000000000000028
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf44 2.603921271336049 2.3665955131107683 65.28 1.2199999999999989
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 157 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf45 2.4967248028856828 2.2748997625822716 65.25999999999999 1.240000000000009
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 157 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf46 2.4963953691980665 2.2764932409573166 65.25999999999999 1.240000000000009
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 160 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf47 2.678944927989822 2.4251978482969956 65.24 1.2600000000000051
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 264 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf48 2.6727135417173904 2.428897140422096 65.22 1.2800000000000011
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf49 2.600256135586627 2.355428067042657 65.16 1.3400000000000034
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-12 gpu conv fp16 11 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf50 2.264460006128871 2.058037581586567 64.9 1.5999999999999943
-1 gpu conv fp16 11 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 155 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv perf_fp16 165 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv fp16 12 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 269 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 164 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 263 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 265 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf51 2.2817447204106736 2.0758846029697513 64.84 1.6599999999999966
-1 gpu conv fp16 11 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 155 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv perf_fp16 165 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv fp16 12 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 267 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 160 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 265 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/vgg16_imagenet/vgg16_imagenet.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/vgg16_imagenet/vgg16_imagenet.txt
deleted file mode 100644
index 108a101c810f4ebe488e6f2029be4d970d7869a2..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/dev_time/vgg16_imagenet/vgg16_imagenet.txt
+++ /dev/null
@@ -1,561 +0,0 @@
-19194.623482
-+++++
-conf1 1 1 72.84 0.0
-1 gpu conv fp32 11 add fp32 1 relu fp32 1
-2 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-3 gpu conv fp32 11 add fp32 1 relu fp32 1
-4 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-5 gpu conv fp32 11 add fp32 1 relu fp32 1
-6 gpu conv fp32 11 add fp32 1 relu fp32 1
-7 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-8 gpu conv fp32 11 add fp32 1 relu fp32 1
-9 gpu conv fp32 11 add fp32 1 relu fp32 1
-10 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-11 gpu conv fp32 11 add fp32 1 relu fp32 1
-12 gpu conv fp32 11 add fp32 1 relu fp32 1
-13 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-14 gpu mul fp32 11 add fp32 1 relu fp32 1
-15 gpu mul fp32 11 add fp32 1 relu fp32 1
-16 gpu mul fp32 11 add fp32 1
-17 gpu softmax fp32 1
------
-+++++
-conf2 2.0787477568568082 1.7725701909562666 72.76 0.0799999999999983
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 267 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv fp16 12 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv fp16 12 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf3 2.2877881266029436 1.9268677640464096 72.04 0.7999999999999972
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv fp16 12 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf4 2.493698381711785 2.0336802939709626 72.02 0.8200000000000074
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 162 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv fp16 12 add fp16 1 relu fp16 1
-9 gpu conv fp16 12 add fp16 1 relu fp16 1
-10 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf5 2.164723960411776 1.8442442134020163 71.94 0.9000000000000057
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 267 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf6 2.53794461743687 2.069640641367895 71.67999999999999 1.1600000000000108
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 162 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv fp16 12 add fp16 1 relu fp16 1
-9 gpu conv fp16 12 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 156 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf7 1.7943268128686711 1.6103705347377417 71.58 1.2600000000000051
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv fp16 12 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv fp16 12 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 266 add fp16 1 relu fp16 1
-10 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv fp16 11 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf8 1.8143284638396158 1.6288620764171362 71.5 1.3400000000000034
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv fp16 12 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv fp16 12 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 266 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 162 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv fp16 11 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf9 2.5462742331906263 2.076061630349781 71.48 1.3599999999999994
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 167 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv fp16 12 add fp16 1 relu fp16 1
-9 gpu conv fp16 12 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 156 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf10 2.526515422129153 2.063839193109964 71.39999999999999 1.440000000000012
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 162 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv fp16 12 add fp16 1 relu fp16 1
-10 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv fp16 11 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf11 2.1596661517243856 1.8351710968407349 71.34 1.5
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 267 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 268 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 156 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf12 2.3444383477958337 1.981259839350623 71.22 1.6200000000000045
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv fp16 12 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf13 1.8402020049200172 1.652343405000522 71.2 1.6400000000000006
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv fp16 12 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 266 add fp16 1 relu fp16 1
-10 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-13 gpu conv fp16 11 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf14 2.6420417968257306 2.167425635999969 71.12 1.7199999999999989
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 167 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv fp16 12 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 155 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf15 2.543198098440602 2.0805826545876145 71.1 1.740000000000009
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 162 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv fp16 12 add fp16 1 relu fp16 1
-10 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf16 2.6224991911009328 2.1476958232678807 70.89999999999999 1.940000000000012
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 167 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv fp16 12 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf17 2.5978010917593752 2.131515210392801 70.8 2.0400000000000063
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 162 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv fp16 12 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf18 2.623210258119482 2.156636511928761 70.76 2.0799999999999983
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv fp16 12 add fp16 1 relu fp16 1
-9 gpu conv fp16 12 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf19 2.598187894495609 2.1322228990374104 70.76 2.0799999999999983
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 162 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv fp16 12 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf20 2.640464221374653 2.1682626030871295 70.76 2.0799999999999983
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 167 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv fp16 12 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf21 2.659563405662692 2.1881035849678936 70.54 2.299999999999997
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv fp16 12 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf22 2.636584103560761 2.1652496021557557 70.39999999999999 2.440000000000012
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 165 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv fp16 12 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf23 2.6315080449303547 2.161259580137757 70.38 2.460000000000008
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 162 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv fp16 12 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf24 2.7367939789033153 2.263326406058847 70.34 2.5
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 160 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf25 2.712182817327382 2.2404693918737233 70.24000000000001 2.5999999999999943
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 168 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf26 2.660510795888948 2.187299344706456 70.22 2.6200000000000045
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 159 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv fp16 12 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf27 2.457573203839654 2.0936930776435383 70.1 2.740000000000009
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv fp16 12 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
-+++++
-conf28 2.7452293174567757 2.2593302388139347 69.92 2.9200000000000017
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 159 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 266 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv fp16 12 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/alexnet2_cifar10/alexnet2_cifar10.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/alexnet2_cifar10/alexnet2_cifar10.txt
deleted file mode 100644
index 208f154e02ef37a6ae87904844c826ce72012b32..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/alexnet2_cifar10/alexnet2_cifar10.txt
+++ /dev/null
@@ -1,23 +0,0 @@
-1114.3009809999999
-+++++
-conf1 1 1 84.76 0.0
-1 gpu conv fp32 11 add fp32 1 tanh fp32 1
-2 gpu conv fp32 11 add fp32 1 tanh fp32 1 pool_max fp32 1
-3 gpu conv fp32 11 add fp32 1 tanh fp32 1
-4 gpu conv fp32 11 add fp32 1 tanh fp32 1 pool_max fp32 1
-5 gpu conv fp32 11 add fp32 1 tanh fp32 1
-6 gpu conv fp32 11 add fp32 1 tanh fp32 1 pool_max fp32 1
-7 gpu mul fp32 11 add fp32 1
-8 gpu softmax fp32 1
------
-+++++
-conf2 1.678391931801309 1.4393008204786808 84.76 0.0
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1
-2 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1
-6 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-7 gpu mul fp16 12 add fp16 1
-8 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/alexnet_cifar10/alexnet_cifar10.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/alexnet_cifar10/alexnet_cifar10.txt
deleted file mode 100644
index eba22e3f01e227041fcb406f87a996837cd5fa2b..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/alexnet_cifar10/alexnet_cifar10.txt
+++ /dev/null
@@ -1,421 +0,0 @@
-2592.187221
-+++++
-conf1 1 1 78.78 0.0
-1 gpu conv fp32 11 add fp32 1 tanh fp32 1 pool_max fp32 1
-2 gpu conv fp32 11 add fp32 1 tanh fp32 1 pool_max fp32 1
-3 gpu conv fp32 11 add fp32 1 tanh fp32 1
-4 gpu conv fp32 11 add fp32 1 tanh fp32 1
-5 gpu conv fp32 11 add fp32 1 tanh fp32 1 pool_max fp32 1
-6 gpu mul fp32 11 add fp32 1
-7 gpu softmax fp32 1
------
-+++++
-conf2 1.7593976485873195 1.6193399031642917 78.78 0.0
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf3 2.081712090729918 1.9102226906341664 78.53999999999999 0.2400000000000091
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf4 2.081712090729918 1.9102226906341664 78.53999999999999 0.2400000000000091
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf5 2.2627828537139263 2.065683616898884 78.34 0.4399999999999977
-1 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf6 2.3527290658539215 2.145832257234814 78.10000000000001 0.6799999999999926
-1 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf7 2.3527290658539215 2.145832257234814 78.10000000000001 0.6799999999999926
-1 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf8 2.3527290658539215 2.145832257234814 78.10000000000001 0.6799999999999926
-1 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf9 2.2247938983110425 2.060416584958474 77.98 0.7999999999999972
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf10 2.2247938983110425 2.060416584958474 77.98 0.7999999999999972
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf11 2.4370818494175888 2.250857540113024 77.98 0.7999999999999972
-1 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf12 2.432854949808342 2.2424500615508003 77.9 0.8799999999999955
-1 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf13 2.432854949808342 2.2424500615508003 77.9 0.8799999999999955
-1 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf14 2.432854949808342 2.2424500615508003 77.9 0.8799999999999955
-1 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf15 2.228328207535687 2.0675123320068267 77.82 0.960000000000008
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf16 2.228328207535687 2.0675123320068267 77.82 0.960000000000008
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf17 2.3417491169395532 2.1355030360671465 77.78 1.0
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf18 2.3417491169395532 2.1355030360671465 77.78 1.0
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf19 2.3417491169395532 2.1355030360671465 77.78 1.0
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv fp16 12 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf20 2.5243776633638846 2.324968713897418 77.78 1.0
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf21 2.5243776633638846 2.324968713897418 77.78 1.0
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf22 2.5243776633638846 2.324968713897418 77.78 1.0
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf23 2.5371416718362823 2.3372173527293847 77.56 1.2199999999999989
-1 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf24 2.5371416718362823 2.3372173527293847 77.56 1.2199999999999989
-1 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf25 2.472472828611022 2.286262888143739 77.48 1.2999999999999972
-1 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf26 2.574475112841438 2.3637004022727544 77.4 1.3799999999999955
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 267 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf27 2.1200397577541747 1.951741010849448 77.3 1.480000000000004
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf28 2.1200397577541747 1.951741010849448 77.3 1.480000000000004
-1 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf29 2.5289288699015304 2.334007588396142 77.2 1.5799999999999983
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf30 2.5289288699015304 2.334007588396142 77.2 1.5799999999999983
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf31 2.5289288699015304 2.334007588396142 77.2 1.5799999999999983
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf32 2.541739061163583 2.3463519042470864 77.18 1.5999999999999943
-1 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf33 2.541739061163583 2.3463519042470864 77.18 1.5999999999999943
-1 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf34 2.580258965052788 2.3848508703934153 76.96 1.8200000000000074
-1 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf35 2.580258965052788 2.3848508703934153 76.96 1.8200000000000074
-1 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf36 2.4768386387310675 2.295002745725082 76.94 1.8400000000000034
-1 gpu conv samp_fp16 263 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-5 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf37 2.5713008246729716 2.3684101116633007 76.94 1.8400000000000034
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 269 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf38 2.5713008246729716 2.3684101116633007 76.94 1.8400000000000034
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 269 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf39 2.5670585645212847 2.3720992406158463 76.92 1.8599999999999994
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf40 2.5670585645212847 2.3720992406158463 76.92 1.8599999999999994
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1
-5 gpu conv samp_fp16 268 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf41 2.5760229577267673 2.3777906009584133 76.9 1.8799999999999955
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 269 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
-+++++
-conf42 2.5760229577267673 2.3777906009584133 76.9 1.8799999999999955
-1 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 tanh fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 269 add fp16 1 tanh fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 tanh fp16 1
-5 gpu conv fp16 12 add fp16 1 tanh fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1
-7 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/alexnet_imagenet/alexnet_imagenet.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/alexnet_imagenet/alexnet_imagenet.txt
deleted file mode 100644
index 8ae986b90ce53e80d10e19525a51ec32f51397d8..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/alexnet_imagenet/alexnet_imagenet.txt
+++ /dev/null
@@ -1,289 +0,0 @@
-2739.950736
-+++++
-conf1 1 1 56.3 0.0
-1 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-2 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-3 gpu conv fp32 11 add fp32 1 relu fp32 1
-4 gpu conv fp32 11 add fp32 1 relu fp32 1
-5 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-6 gpu mul fp32 11 add fp32 1 relu fp32 1
-7 gpu mul fp32 11 add fp32 1 relu fp32 1
-8 gpu mul fp32 11 add fp32 1
-9 gpu softmax fp32 1
------
-+++++
-conf2 1.802133644103582 1.8186433204507424 55.76 0.5399999999999991
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf3 2.0227701930718065 2.043112495268932 55.42 0.8799999999999955
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 166 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf4 1.8063132288735129 1.8239088223620996 54.96 1.3399999999999963
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf5 1.8063132288735129 1.8239088223620996 54.96 1.3399999999999963
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf6 1.8063132288735129 1.8239088223620996 54.96 1.3399999999999963
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf7 2.085011755614172 2.122606306624671 54.92 1.3799999999999955
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 159 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf8 2.085011755614172 2.122606306624671 54.92 1.3799999999999955
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 159 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf9 1.8052659214923805 1.8217111622759978 54.82 1.4799999999999969
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 164 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf10 2.0146435217865446 2.0367475358800102 54.58 1.7199999999999989
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 162 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf11 1.9101312060368951 1.9552389688678584 54.24 2.059999999999995
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 157 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf12 1.9101312060368951 1.9552389688678584 54.24 2.059999999999995
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 157 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf13 1.9101312060368951 1.9552389688678584 54.24 2.059999999999995
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 157 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf14 2.019868378233057 2.0433540129730265 54.17999999999999 2.1200000000000045
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 162 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf15 2.019868378233057 2.0433540129730265 54.17999999999999 2.1200000000000045
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 162 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf16 2.028037341700216 2.049760395549724 53.98 2.3200000000000003
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 166 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf17 2.028037341700216 2.049760395549724 53.98 2.3200000000000003
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 166 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf18 2.028037341700216 2.049760395549724 53.98 2.3200000000000003
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 166 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 163 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf19 1.8052659214923805 1.8217111622759978 53.879999999999995 2.4200000000000017
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 11 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 164 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf20 1.8052659214923805 1.8217111622759978 53.879999999999995 2.4200000000000017
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 11 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 164 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf21 2.0267172350289036 2.046985186681549 53.86 2.4399999999999977
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 166 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 164 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf22 2.0267172350289036 2.046985186681549 53.86 2.4399999999999977
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 166 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 164 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf23 2.0267172350289036 2.046985186681549 53.86 2.4399999999999977
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 166 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 164 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
-+++++
-conf24 2.0185588815268836 2.0405961127674277 53.559999999999995 2.740000000000002
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu conv perf_fp16 162 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv perf_fp16 164 add fp16 1 relu fp16 1
-5 gpu conv perf_fp16 157 add fp16 1 relu fp16 1 pool_max fp16 1
-6 gpu mul fp16 12 add fp16 1 relu fp16 1
-7 gpu mul fp16 12 add fp16 1 relu fp16 1
-8 gpu mul fp16 12 add fp16 1
-9 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/lenet_keras/lenet_keras.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/lenet_keras/lenet_keras.txt
deleted file mode 100644
index da88f7cd26b049fd18644a834e4d34b944149cb2..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/lenet_keras/lenet_keras.txt
+++ /dev/null
@@ -1,409 +0,0 @@
-282.5141369999999
-+++++
-conf1 1 1 98.7 0.0
-1 gpu conv fp32 11 add fp32 1 pool_max fp32 1 tanh fp32 1
-2 gpu conv fp32 11 add fp32 1 pool_max fp32 1 tanh fp32 1
-3 gpu mul fp32 11 add fp32 1 tanh fp32 1
-4 gpu mul fp32 11 add fp32 1 tanh fp32 1
-5 gpu softmax fp32 1
------
-+++++
-conf2 1.9343699741206566 2.1183040240042 98.68 0.01999999999999602
-1 gpu conv samp_fp16 264 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 265 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf3 1.9343699741206566 2.1183040240042 98.68 0.01999999999999602
-1 gpu conv samp_fp16 264 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 265 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf4 1.8936889628815377 2.139779619692146 98.68 0.01999999999999602
-1 gpu conv perf_fp16 152 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf5 1.8936889628815377 2.139779619692146 98.68 0.01999999999999602
-1 gpu conv perf_fp16 152 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf6 1.6415764141643088 1.8012120076077847 98.66 0.04000000000000625
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 265 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf7 1.9358279784215788 2.1233340385374495 98.66 0.04000000000000625
-1 gpu conv samp_fp16 264 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 264 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf8 1.9358279784215788 2.1233340385374495 98.66 0.04000000000000625
-1 gpu conv samp_fp16 264 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 264 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf9 1.6319327047042609 1.8046853367113418 98.64 0.060000000000002274
-1 gpu conv fp16 12 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 263 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf10 1.6319327047042609 1.8046853367113418 98.64 0.060000000000002274
-1 gpu conv fp16 12 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 263 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf11 1.6319327047042609 1.8046853367113418 98.64 0.060000000000002274
-1 gpu conv fp16 12 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 263 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf12 1.6319327047042609 1.8046853367113418 98.64 0.060000000000002274
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 263 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf13 1.6319327047042609 1.8046853367113418 98.64 0.060000000000002274
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 263 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf14 1.5602284338468988 1.7102497386784767 98.61999999999999 0.0800000000000125
-1 gpu conv fp16 12 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 267 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf15 1.5602284338468988 1.7102497386784767 98.61999999999999 0.0800000000000125
-1 gpu conv fp16 12 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 267 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf16 1.5602284338468988 1.7102497386784767 98.61999999999999 0.0800000000000125
-1 gpu conv fp16 12 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 267 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf17 1.8224050632690918 1.9936046569348063 98.61999999999999 0.0800000000000125
-1 gpu conv samp_fp16 264 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 267 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf18 1.8224050632690918 1.9936046569348063 98.61999999999999 0.0800000000000125
-1 gpu conv samp_fp16 264 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 267 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf19 1.8224050632690918 1.9936046569348063 98.61999999999999 0.0800000000000125
-1 gpu conv samp_fp16 264 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 267 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf20 2.2168527051833635 2.453341076720038 98.61999999999999 0.0800000000000125
-1 gpu conv samp_fp16 264 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf21 2.2168527051833635 2.453341076720038 98.61999999999999 0.0800000000000125
-1 gpu conv samp_fp16 264 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf22 1.9040998718547615 2.1501783570812565 98.61999999999999 0.0800000000000125
-1 gpu conv perf_fp16 151 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf23 1.9040998718547615 2.1501783570812565 98.61999999999999 0.0800000000000125
-1 gpu conv perf_fp16 151 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf24 1.5630416487818 1.7451546885860074 98.6 0.10000000000000853
-1 gpu conv perf_fp16 156 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 267 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf25 1.5630416487818 1.7451546885860074 98.6 0.10000000000000853
-1 gpu conv perf_fp16 156 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 267 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf26 1.8406161850501603 2.037849502542524 98.6 0.10000000000000853
-1 gpu conv fp16 12 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf27 1.8406161850501603 2.037849502542524 98.6 0.10000000000000853
-1 gpu conv fp16 12 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf28 1.8406161850501603 2.037849502542524 98.6 0.10000000000000853
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf29 1.8406161850501603 2.037849502542524 98.6 0.10000000000000853
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf30 2.1941568976363475 2.4445764373737644 98.6 0.10000000000000853
-1 gpu conv samp_fp16 269 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf31 2.1941568976363475 2.4445764373737644 98.6 0.10000000000000853
-1 gpu conv samp_fp16 269 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf32 1.5602284338468988 1.7102497386784767 98.58 0.12000000000000455
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 267 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf33 1.5602284338468988 1.7102497386784767 98.58 0.12000000000000455
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 267 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf34 1.5602284338468988 1.7102497386784767 98.58 0.12000000000000455
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 267 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf35 1.9209933607603906 2.123109543083542 98.58 0.12000000000000455
-1 gpu conv samp_fp16 264 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 263 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf36 1.9209933607603906 2.123109543083542 98.58 0.12000000000000455
-1 gpu conv samp_fp16 264 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 263 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf37 1.9209933607603906 2.123109543083542 98.58 0.12000000000000455
-1 gpu conv samp_fp16 264 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 263 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf38 1.8406161850501603 2.037849502542524 98.58 0.12000000000000455
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf39 1.8406161850501603 2.037849502542524 98.58 0.12000000000000455
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf40 1.8445326456180258 2.087601822059355 98.58 0.12000000000000455
-1 gpu conv perf_fp16 156 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf41 1.8445326456180258 2.087601822059355 98.58 0.12000000000000455
-1 gpu conv perf_fp16 156 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf42 1.8649226857257986 2.1076025277601325 98.56 0.14000000000000057
-1 gpu conv perf_fp16 168 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf43 1.8649226857257986 2.1076025277601325 98.56 0.14000000000000057
-1 gpu conv perf_fp16 168 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf44 1.8463058650555446 2.067271423078985 98.56 0.14000000000000057
-1 gpu conv perf_fp16 157 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf45 1.8463058650555446 2.067271423078985 98.56 0.14000000000000057
-1 gpu conv perf_fp16 157 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf46 1.9234076467497994 2.1864740913112275 98.56 0.14000000000000057
-1 gpu conv perf_fp16 153 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf47 1.9234076467497994 2.1864740913112275 98.56 0.14000000000000057
-1 gpu conv perf_fp16 153 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 262 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf48 1.6319327047042609 1.8046853367113418 98.54 0.1599999999999966
-1 gpu conv fp16 11 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 263 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf49 1.6350106933897723 1.8435952834193967 98.52 0.18000000000000682
-1 gpu conv perf_fp16 156 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 263 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf50 1.6350106933897723 1.8435952834193967 98.52 0.18000000000000682
-1 gpu conv perf_fp16 156 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 263 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
-+++++
-conf51 1.6510114896409525 1.8591762752048948 98.48 0.21999999999999886
-1 gpu conv perf_fp16 168 add fp16 1 pool_max fp16 1 tanh fp16 1
-2 gpu conv samp_fp16 263 add fp16 1 pool_max fp16 1 tanh fp16 1
-3 gpu mul fp16 12 add fp16 1 tanh fp16 1
-4 gpu mul fp16 12 add fp16 1 tanh fp16 1
-5 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/mobilenet_cifar10/mobilenet_cifar10.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/mobilenet_cifar10/mobilenet_cifar10.txt
deleted file mode 100644
index 93ca37c00a73f1a1cfc72bf58e8067906269d813..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/mobilenet_cifar10/mobilenet_cifar10.txt
+++ /dev/null
@@ -1,871 +0,0 @@
-4077.307063200001
-+++++
-conf1 1 1 84.42 0.0
-1 gpu conv fp32 11
-2 gpu batchnorm fp32 11
-3 gpu relu fp32 11
-4 gpu group_conv fp32 11
-5 gpu batchnorm fp32 11
-6 gpu relu fp32 11
-7 gpu conv fp32 11
-8 gpu batchnorm fp32 11
-9 gpu relu fp32 11
-10 gpu group_conv fp32 11
-11 gpu batchnorm fp32 11
-12 gpu relu fp32 11
-13 gpu conv fp32 11
-14 gpu batchnorm fp32 11
-15 gpu relu fp32 11
-16 gpu group_conv fp32 11
-17 gpu batchnorm fp32 11
-18 gpu relu fp32 11
-19 gpu conv fp32 11
-20 gpu batchnorm fp32 11
-21 gpu relu fp32 11
-22 gpu group_conv fp32 11
-23 gpu batchnorm fp32 11
-24 gpu relu fp32 11
-25 gpu conv fp32 11
-26 gpu batchnorm fp32 11
-27 gpu relu fp32 11
-28 gpu group_conv fp32 11
-29 gpu batchnorm fp32 11
-30 gpu relu fp32 11
-31 gpu conv fp32 11
-32 gpu batchnorm fp32 11
-33 gpu relu fp32 11
-34 gpu group_conv fp32 11
-35 gpu batchnorm fp32 11
-36 gpu relu fp32 11
-37 gpu conv fp32 11
-38 gpu batchnorm fp32 11
-39 gpu relu fp32 11
-40 gpu group_conv fp32 11
-41 gpu batchnorm fp32 11
-42 gpu relu fp32 11
-43 gpu conv fp32 11
-44 gpu batchnorm fp32 11
-45 gpu relu fp32 11
-46 gpu group_conv fp32 11
-47 gpu batchnorm fp32 11
-48 gpu relu fp32 11
-49 gpu conv fp32 11
-50 gpu batchnorm fp32 11
-51 gpu relu fp32 11
-52 gpu group_conv fp32 11
-53 gpu batchnorm fp32 11
-54 gpu relu fp32 11
-55 gpu conv fp32 11
-56 gpu batchnorm fp32 11
-57 gpu relu fp32 11
-58 gpu group_conv fp32 11
-59 gpu batchnorm fp32 11
-60 gpu relu fp32 11
-61 gpu conv fp32 11
-62 gpu batchnorm fp32 11
-63 gpu relu fp32 11
-64 gpu group_conv fp32 11
-65 gpu batchnorm fp32 11
-66 gpu relu fp32 11
-67 gpu conv fp32 11
-68 gpu batchnorm fp32 11
-69 gpu relu fp32 11
-70 gpu group_conv fp32 11
-71 gpu batchnorm fp32 11
-72 gpu relu fp32 11
-73 gpu conv fp32 11
-74 gpu batchnorm fp32 11
-75 gpu relu fp32 11
-76 gpu group_conv fp32 11
-77 gpu batchnorm fp32 11
-78 gpu relu fp32 11
-79 gpu conv fp32 11
-80 gpu batchnorm fp32 11
-81 gpu relu fp32 11
-82 gpu pool_mean fp32 11
-83 gpu mul fp32 11 add fp32 1
-84 gpu softmax fp32 1
------
-+++++
-conf2 1.504059255565631 1.4598468219902432 81.86 2.5600000000000023
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 161
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 152
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 152
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 152
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 152
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf3 1.5040783418076804 1.459845395800413 81.86 2.5600000000000023
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 161
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 152
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 152
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 152
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 152
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf4 1.5042737817275433 1.4598464522370567 81.74 2.680000000000007
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 161
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 152
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 155
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 152
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 152
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 152
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf5 1.5042737817275433 1.4598464522370567 81.74 2.680000000000007
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 161
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 152
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 155
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 152
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 152
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 152
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf6 1.5070383438802568 1.463241585164149 81.69999999999999 2.720000000000013
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 168
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 152
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 152
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 151
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 152
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 152
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf7 1.5070575058058588 1.463240152333617 81.58 2.8400000000000034
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 168
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 152
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 152
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 152
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 152
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf8 1.5039678813445672 1.4598454486222088 81.56 2.8599999999999994
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 161
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 152
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 152
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 152
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 152
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 153
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf9 1.5038655354281372 1.4599130636549171 81.46 2.960000000000008
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv perf_fp16 161
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv perf_fp16 152
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv perf_fp16 161
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv perf_fp16 155
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv perf_fp16 152
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv perf_fp16 152
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv perf_fp16 151
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv perf_fp16 151
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv perf_fp16 153
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
-+++++
-conf10 1.4785375660713596 1.4280520288797043 84.42 0.0
-1 gpu conv fp16 12
-2 gpu batchnorm fp16 12
-3 gpu relu fp16 12
-4 gpu group_conv fp16 12
-5 gpu batchnorm fp16 12
-6 gpu relu fp16 12
-7 gpu conv fp16 12
-8 gpu batchnorm fp16 12
-9 gpu relu fp16 12
-10 gpu group_conv fp16 12
-11 gpu batchnorm fp16 12
-12 gpu relu fp16 12
-13 gpu conv fp16 12
-14 gpu batchnorm fp16 12
-15 gpu relu fp16 12
-16 gpu group_conv fp16 12
-17 gpu batchnorm fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12
-20 gpu batchnorm fp16 12
-21 gpu relu fp16 12
-22 gpu group_conv fp16 12
-23 gpu batchnorm fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu group_conv fp16 12
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 12
-32 gpu batchnorm fp16 12
-33 gpu relu fp16 12
-34 gpu group_conv fp16 12
-35 gpu batchnorm fp16 12
-36 gpu relu fp16 12
-37 gpu conv fp16 12
-38 gpu batchnorm fp16 12
-39 gpu relu fp16 12
-40 gpu group_conv fp16 12
-41 gpu batchnorm fp16 12
-42 gpu relu fp16 12
-43 gpu conv fp16 12
-44 gpu batchnorm fp16 12
-45 gpu relu fp16 12
-46 gpu group_conv fp16 12
-47 gpu batchnorm fp16 12
-48 gpu relu fp16 12
-49 gpu conv fp16 12
-50 gpu batchnorm fp16 12
-51 gpu relu fp16 12
-52 gpu group_conv fp16 12
-53 gpu batchnorm fp16 12
-54 gpu relu fp16 12
-55 gpu conv fp16 12
-56 gpu batchnorm fp16 12
-57 gpu relu fp16 12
-58 gpu group_conv fp16 12
-59 gpu batchnorm fp16 12
-60 gpu relu fp16 12
-61 gpu conv fp16 12
-62 gpu batchnorm fp16 12
-63 gpu relu fp16 12
-64 gpu group_conv fp16 12
-65 gpu batchnorm fp16 12
-66 gpu relu fp16 12
-67 gpu conv fp16 12
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu group_conv fp16 12
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv fp16 12
-74 gpu batchnorm fp16 12
-75 gpu relu fp16 12
-76 gpu group_conv fp16 12
-77 gpu batchnorm fp16 12
-78 gpu relu fp16 12
-79 gpu conv fp16 12
-80 gpu batchnorm fp16 12
-81 gpu relu fp16 12
-82 gpu pool_mean fp16 12
-83 gpu mul fp16 12 add fp16 1
-84 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/resnet18_cifar10/resnet18_cifar10.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/resnet18_cifar10/resnet18_cifar10.txt
deleted file mode 100644
index d1d75a011e9ada7994dcd5a31ee5d56fc2ee3e2f..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/resnet18_cifar10/resnet18_cifar10.txt
+++ /dev/null
@@ -1,91 +0,0 @@
-2484.981244
-+++++
-conf1 1 1 89.42 0.0
-1 gpu conv fp32 11 add fp32 1 relu fp32 1
-2 gpu conv fp32 11 add fp32 1 relu fp32 1
-3 gpu conv fp32 11 add fp32 1
-4 gpu add fp32 11
-5 gpu relu fp32 11
-6 gpu conv fp32 11 add fp32 1 relu fp32 1
-7 gpu conv fp32 11 add fp32 1
-8 gpu add fp32 11
-9 gpu relu fp32 11
-10 gpu conv fp32 11 add fp32 1 relu fp32 1
-11 gpu conv fp32 11 add fp32 1
-12 gpu add fp32 11
-13 gpu relu fp32 11
-14 gpu conv fp32 11 add fp32 1 relu fp32 1
-15 gpu conv fp32 11 add fp32 1
-16 gpu conv fp32 11 add fp32 1
-17 gpu add fp32 11
-18 gpu relu fp32 11
-19 gpu conv fp32 11 add fp32 1 relu fp32 1
-20 gpu conv fp32 11 add fp32 1
-21 gpu add fp32 11
-22 gpu relu fp32 11
-23 gpu conv fp32 11 add fp32 1 relu fp32 1
-24 gpu conv fp32 11 add fp32 1
-25 gpu add fp32 11
-26 gpu relu fp32 11
-27 gpu conv fp32 11 add fp32 1 relu fp32 1
-28 gpu conv fp32 11 add fp32 1
-29 gpu conv fp32 11 add fp32 1
-30 gpu add fp32 11
-31 gpu relu fp32 11
-32 gpu conv fp32 11 add fp32 1 relu fp32 1
-33 gpu conv fp32 11 add fp32 1
-34 gpu add fp32 11
-35 gpu relu fp32 11
-36 gpu conv fp32 11 add fp32 1 relu fp32 1
-37 gpu conv fp32 11 add fp32 1
-38 gpu add fp32 11
-39 gpu relu fp32 11
-40 gpu pool_mean fp32 11
-41 gpu mul fp32 11 add fp32 1
-42 gpu softmax fp32 1
------
-+++++
-conf2 1.3617910209460897 1.3866827244386561 89.42 0.0
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1
-3 gpu conv fp16 12 add fp16 1
-4 gpu add fp16 12
-5 gpu relu fp16 12
-6 gpu conv fp16 12 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1
-8 gpu add fp16 12
-9 gpu relu fp16 12
-10 gpu conv fp16 12 add fp16 1 relu fp16 1
-11 gpu conv fp16 12 add fp16 1
-12 gpu add fp16 12
-13 gpu relu fp16 12
-14 gpu conv fp16 12 add fp16 1 relu fp16 1
-15 gpu conv fp16 12 add fp16 1
-16 gpu conv fp16 12 add fp16 1
-17 gpu add fp16 12
-18 gpu relu fp16 12
-19 gpu conv fp16 12 add fp16 1 relu fp16 1
-20 gpu conv fp16 12 add fp16 1
-21 gpu add fp16 12
-22 gpu relu fp16 12
-23 gpu conv fp16 12 add fp16 1 relu fp16 1
-24 gpu conv fp16 12 add fp16 1
-25 gpu add fp16 12
-26 gpu relu fp16 12
-27 gpu conv fp16 12 add fp16 1 relu fp16 1
-28 gpu conv fp16 12 add fp16 1
-29 gpu conv fp16 12 add fp16 1
-30 gpu add fp16 12
-31 gpu relu fp16 12
-32 gpu conv fp16 12 add fp16 1 relu fp16 1
-33 gpu conv fp16 12 add fp16 1
-34 gpu add fp16 12
-35 gpu relu fp16 12
-36 gpu conv fp16 12 add fp16 1 relu fp16 1
-37 gpu conv fp16 12 add fp16 1
-38 gpu add fp16 12
-39 gpu relu fp16 12
-40 gpu pool_mean fp16 12
-41 gpu mul fp16 12 add fp16 1
-42 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/resnet50_imagenet/resnet50_imagenet.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/resnet50_imagenet/resnet50_imagenet.txt
deleted file mode 100644
index a045011580adb912289364d35fb85668e74261e7..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/resnet50_imagenet/resnet50_imagenet.txt
+++ /dev/null
@@ -1,1233 +0,0 @@
-7161.053769000008
-+++++
-conf1 1 1 75.7 0.0
-1 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-2 gpu batchnorm fp32 11
-3 gpu conv fp32 11 add fp32 1
-4 gpu batchnorm fp32 11
-5 gpu relu fp32 11
-6 gpu conv fp32 11 add fp32 1
-7 gpu batchnorm fp32 11
-8 gpu relu fp32 11
-9 gpu conv fp32 11 add fp32 1
-10 gpu batchnorm fp32 11
-11 gpu conv fp32 11 add fp32 1
-12 gpu batchnorm fp32 11
-13 gpu add fp32 11
-14 gpu relu fp32 11
-15 gpu conv fp32 11 add fp32 1
-16 gpu batchnorm fp32 11
-17 gpu relu fp32 11
-18 gpu conv fp32 11 add fp32 1
-19 gpu batchnorm fp32 11
-20 gpu relu fp32 11
-21 gpu conv fp32 11 add fp32 1
-22 gpu batchnorm fp32 11
-23 gpu add fp32 11
-24 gpu relu fp32 11
-25 gpu conv fp32 11 add fp32 1
-26 gpu batchnorm fp32 11
-27 gpu relu fp32 11
-28 gpu conv fp32 11 add fp32 1
-29 gpu batchnorm fp32 11
-30 gpu relu fp32 11
-31 gpu conv fp32 11 add fp32 1
-32 gpu batchnorm fp32 11
-33 gpu add fp32 11
-34 gpu relu fp32 11
-35 gpu conv fp32 11 add fp32 1
-36 gpu batchnorm fp32 11
-37 gpu relu fp32 11
-38 gpu conv fp32 11 add fp32 1
-39 gpu batchnorm fp32 11
-40 gpu relu fp32 11
-41 gpu conv fp32 11 add fp32 1
-42 gpu batchnorm fp32 11
-43 gpu conv fp32 11 add fp32 1
-44 gpu batchnorm fp32 11
-45 gpu add fp32 11
-46 gpu relu fp32 11
-47 gpu conv fp32 11 add fp32 1
-48 gpu batchnorm fp32 11
-49 gpu relu fp32 11
-50 gpu conv fp32 11 add fp32 1
-51 gpu batchnorm fp32 11
-52 gpu relu fp32 11
-53 gpu conv fp32 11 add fp32 1
-54 gpu batchnorm fp32 11
-55 gpu add fp32 11
-56 gpu relu fp32 11
-57 gpu conv fp32 11 add fp32 1
-58 gpu batchnorm fp32 11
-59 gpu relu fp32 11
-60 gpu conv fp32 11 add fp32 1
-61 gpu batchnorm fp32 11
-62 gpu relu fp32 11
-63 gpu conv fp32 11 add fp32 1
-64 gpu batchnorm fp32 11
-65 gpu add fp32 11
-66 gpu relu fp32 11
-67 gpu conv fp32 11 add fp32 1
-68 gpu batchnorm fp32 11
-69 gpu relu fp32 11
-70 gpu conv fp32 11 add fp32 1
-71 gpu batchnorm fp32 11
-72 gpu relu fp32 11
-73 gpu conv fp32 11 add fp32 1
-74 gpu batchnorm fp32 11
-75 gpu add fp32 11
-76 gpu relu fp32 11
-77 gpu conv fp32 11 add fp32 1
-78 gpu batchnorm fp32 11
-79 gpu relu fp32 11
-80 gpu conv fp32 11 add fp32 1
-81 gpu batchnorm fp32 11
-82 gpu relu fp32 11
-83 gpu conv fp32 11 add fp32 1
-84 gpu batchnorm fp32 11
-85 gpu conv fp32 11 add fp32 1
-86 gpu batchnorm fp32 11
-87 gpu add fp32 11
-88 gpu relu fp32 11
-89 gpu conv fp32 11 add fp32 1
-90 gpu batchnorm fp32 11
-91 gpu relu fp32 11
-92 gpu conv fp32 11 add fp32 1
-93 gpu batchnorm fp32 11
-94 gpu relu fp32 11
-95 gpu conv fp32 11 add fp32 1
-96 gpu batchnorm fp32 11
-97 gpu add fp32 11
-98 gpu relu fp32 11
-99 gpu conv fp32 11 add fp32 1
-100 gpu batchnorm fp32 11
-101 gpu relu fp32 11
-102 gpu conv fp32 11 add fp32 1
-103 gpu batchnorm fp32 11
-104 gpu relu fp32 11
-105 gpu conv fp32 11 add fp32 1
-106 gpu batchnorm fp32 11
-107 gpu add fp32 11
-108 gpu relu fp32 11
-109 gpu conv fp32 11 add fp32 1
-110 gpu batchnorm fp32 11
-111 gpu relu fp32 11
-112 gpu conv fp32 11 add fp32 1
-113 gpu batchnorm fp32 11
-114 gpu relu fp32 11
-115 gpu conv fp32 11 add fp32 1
-116 gpu batchnorm fp32 11
-117 gpu add fp32 11
-118 gpu relu fp32 11
-119 gpu conv fp32 11 add fp32 1
-120 gpu batchnorm fp32 11
-121 gpu relu fp32 11
-122 gpu conv fp32 11 add fp32 1
-123 gpu batchnorm fp32 11
-124 gpu relu fp32 11
-125 gpu conv fp32 11 add fp32 1
-126 gpu batchnorm fp32 11
-127 gpu add fp32 11
-128 gpu relu fp32 11
-129 gpu conv fp32 11 add fp32 1
-130 gpu batchnorm fp32 11
-131 gpu relu fp32 11
-132 gpu conv fp32 11 add fp32 1
-133 gpu batchnorm fp32 11
-134 gpu relu fp32 11
-135 gpu conv fp32 11 add fp32 1
-136 gpu batchnorm fp32 11
-137 gpu add fp32 11
-138 gpu relu fp32 11
-139 gpu conv fp32 11 add fp32 1
-140 gpu batchnorm fp32 11
-141 gpu relu fp32 11
-142 gpu conv fp32 11 add fp32 1
-143 gpu batchnorm fp32 11
-144 gpu relu fp32 11
-145 gpu conv fp32 11 add fp32 1
-146 gpu batchnorm fp32 11
-147 gpu conv fp32 11 add fp32 1
-148 gpu batchnorm fp32 11
-149 gpu add fp32 11
-150 gpu relu fp32 11
-151 gpu conv fp32 11 add fp32 1
-152 gpu batchnorm fp32 11
-153 gpu relu fp32 11
-154 gpu conv fp32 11 add fp32 1
-155 gpu batchnorm fp32 11
-156 gpu relu fp32 11
-157 gpu conv fp32 11 add fp32 1
-158 gpu batchnorm fp32 11
-159 gpu add fp32 11
-160 gpu relu fp32 11
-161 gpu conv fp32 11 add fp32 1
-162 gpu batchnorm fp32 11
-163 gpu relu fp32 11
-164 gpu conv fp32 11 add fp32 1
-165 gpu batchnorm fp32 11
-166 gpu relu fp32 11
-167 gpu conv fp32 11 add fp32 1
-168 gpu batchnorm fp32 11
-169 gpu add fp32 11
-170 gpu relu fp32 11
-171 gpu pool_max fp32 11
-172 gpu mul fp32 11 add fp32 1
-173 gpu softmax fp32 1
------
-+++++
-conf2 1.8254789092281507 1.4527803526239977 75.7 0.0
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu batchnorm fp16 12
-3 gpu conv fp16 12 add fp16 1
-4 gpu batchnorm fp16 12
-5 gpu relu fp16 12
-6 gpu conv fp16 12 add fp16 1
-7 gpu batchnorm fp16 12
-8 gpu relu fp16 12
-9 gpu conv fp16 12 add fp16 1
-10 gpu batchnorm fp16 12
-11 gpu conv fp16 12 add fp16 1
-12 gpu batchnorm fp16 12
-13 gpu add fp16 12
-14 gpu relu fp16 12
-15 gpu conv fp16 12 add fp16 1
-16 gpu batchnorm fp16 12
-17 gpu relu fp16 12
-18 gpu conv fp16 12 add fp16 1
-19 gpu batchnorm fp16 12
-20 gpu relu fp16 12
-21 gpu conv fp16 12 add fp16 1
-22 gpu batchnorm fp16 12
-23 gpu add fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12 add fp16 1
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu conv fp16 12 add fp16 1
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 12 add fp16 1
-32 gpu batchnorm fp16 12
-33 gpu add fp16 12
-34 gpu relu fp16 12
-35 gpu conv fp16 12 add fp16 1
-36 gpu batchnorm fp16 12
-37 gpu relu fp16 12
-38 gpu conv fp16 12 add fp16 1
-39 gpu batchnorm fp16 12
-40 gpu relu fp16 12
-41 gpu conv fp16 12 add fp16 1
-42 gpu batchnorm fp16 12
-43 gpu conv fp16 12 add fp16 1
-44 gpu batchnorm fp16 12
-45 gpu add fp16 12
-46 gpu relu fp16 12
-47 gpu conv fp16 12 add fp16 1
-48 gpu batchnorm fp16 12
-49 gpu relu fp16 12
-50 gpu conv fp16 12 add fp16 1
-51 gpu batchnorm fp16 12
-52 gpu relu fp16 12
-53 gpu conv fp16 12 add fp16 1
-54 gpu batchnorm fp16 12
-55 gpu add fp16 12
-56 gpu relu fp16 12
-57 gpu conv fp16 12 add fp16 1
-58 gpu batchnorm fp16 12
-59 gpu relu fp16 12
-60 gpu conv fp16 12 add fp16 1
-61 gpu batchnorm fp16 12
-62 gpu relu fp16 12
-63 gpu conv fp16 12 add fp16 1
-64 gpu batchnorm fp16 12
-65 gpu add fp16 12
-66 gpu relu fp16 12
-67 gpu conv fp16 12 add fp16 1
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu conv fp16 12 add fp16 1
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv fp16 12 add fp16 1
-74 gpu batchnorm fp16 12
-75 gpu add fp16 12
-76 gpu relu fp16 12
-77 gpu conv fp16 12 add fp16 1
-78 gpu batchnorm fp16 12
-79 gpu relu fp16 12
-80 gpu conv fp16 12 add fp16 1
-81 gpu batchnorm fp16 12
-82 gpu relu fp16 12
-83 gpu conv fp16 12 add fp16 1
-84 gpu batchnorm fp16 12
-85 gpu conv fp16 12 add fp16 1
-86 gpu batchnorm fp16 12
-87 gpu add fp16 12
-88 gpu relu fp16 12
-89 gpu conv fp16 12 add fp16 1
-90 gpu batchnorm fp16 12
-91 gpu relu fp16 12
-92 gpu conv fp16 12 add fp16 1
-93 gpu batchnorm fp16 12
-94 gpu relu fp16 12
-95 gpu conv fp16 12 add fp16 1
-96 gpu batchnorm fp16 12
-97 gpu add fp16 12
-98 gpu relu fp16 12
-99 gpu conv fp16 12 add fp16 1
-100 gpu batchnorm fp16 12
-101 gpu relu fp16 12
-102 gpu conv fp16 12 add fp16 1
-103 gpu batchnorm fp16 12
-104 gpu relu fp16 12
-105 gpu conv fp16 12 add fp16 1
-106 gpu batchnorm fp16 12
-107 gpu add fp16 12
-108 gpu relu fp16 12
-109 gpu conv fp16 12 add fp16 1
-110 gpu batchnorm fp16 12
-111 gpu relu fp16 12
-112 gpu conv fp16 12 add fp16 1
-113 gpu batchnorm fp16 12
-114 gpu relu fp16 12
-115 gpu conv fp16 12 add fp16 1
-116 gpu batchnorm fp16 12
-117 gpu add fp16 12
-118 gpu relu fp16 12
-119 gpu conv fp16 12 add fp16 1
-120 gpu batchnorm fp16 12
-121 gpu relu fp16 12
-122 gpu conv fp16 12 add fp16 1
-123 gpu batchnorm fp16 12
-124 gpu relu fp16 12
-125 gpu conv fp16 12 add fp16 1
-126 gpu batchnorm fp16 12
-127 gpu add fp16 12
-128 gpu relu fp16 12
-129 gpu conv fp16 12 add fp16 1
-130 gpu batchnorm fp16 12
-131 gpu relu fp16 12
-132 gpu conv fp16 12 add fp16 1
-133 gpu batchnorm fp16 12
-134 gpu relu fp16 12
-135 gpu conv fp16 12 add fp16 1
-136 gpu batchnorm fp16 12
-137 gpu add fp16 12
-138 gpu relu fp16 12
-139 gpu conv fp16 12 add fp16 1
-140 gpu batchnorm fp16 12
-141 gpu relu fp16 12
-142 gpu conv fp16 12 add fp16 1
-143 gpu batchnorm fp16 12
-144 gpu relu fp16 12
-145 gpu conv fp16 12 add fp16 1
-146 gpu batchnorm fp16 12
-147 gpu conv fp16 12 add fp16 1
-148 gpu batchnorm fp16 12
-149 gpu add fp16 12
-150 gpu relu fp16 12
-151 gpu conv fp16 12 add fp16 1
-152 gpu batchnorm fp16 12
-153 gpu relu fp16 12
-154 gpu conv fp16 12 add fp16 1
-155 gpu batchnorm fp16 12
-156 gpu relu fp16 12
-157 gpu conv fp16 12 add fp16 1
-158 gpu batchnorm fp16 12
-159 gpu add fp16 12
-160 gpu relu fp16 12
-161 gpu conv fp16 12 add fp16 1
-162 gpu batchnorm fp16 12
-163 gpu relu fp16 12
-164 gpu conv fp16 12 add fp16 1
-165 gpu batchnorm fp16 12
-166 gpu relu fp16 12
-167 gpu conv fp16 12 add fp16 1
-168 gpu batchnorm fp16 12
-169 gpu add fp16 12
-170 gpu relu fp16 12
-171 gpu pool_max fp16 12
-172 gpu mul fp16 12 add fp16 1
-173 gpu softmax fp32 1
------
-+++++
-conf3 1.8521749055745271 1.472492365706726 75.02 0.6800000000000068
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu batchnorm fp16 12
-3 gpu conv fp16 12 add fp16 1
-4 gpu batchnorm fp16 12
-5 gpu relu fp16 12
-6 gpu conv fp16 12 add fp16 1
-7 gpu batchnorm fp16 12
-8 gpu relu fp16 12
-9 gpu conv fp16 12 add fp16 1
-10 gpu batchnorm fp16 12
-11 gpu conv fp16 12 add fp16 1
-12 gpu batchnorm fp16 12
-13 gpu add fp16 12
-14 gpu relu fp16 12
-15 gpu conv fp16 12 add fp16 1
-16 gpu batchnorm fp16 12
-17 gpu relu fp16 12
-18 gpu conv fp16 12 add fp16 1
-19 gpu batchnorm fp16 12
-20 gpu relu fp16 12
-21 gpu conv fp16 12 add fp16 1
-22 gpu batchnorm fp16 12
-23 gpu add fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12 add fp16 1
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu conv perf_fp16 160 add fp16 1
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 11 add fp16 1
-32 gpu batchnorm fp16 12
-33 gpu add fp16 12
-34 gpu relu fp16 12
-35 gpu conv fp16 12 add fp16 1
-36 gpu batchnorm fp16 12
-37 gpu relu fp16 12
-38 gpu conv fp16 12 add fp16 1
-39 gpu batchnorm fp16 12
-40 gpu relu fp16 12
-41 gpu conv fp16 12 add fp16 1
-42 gpu batchnorm fp16 12
-43 gpu conv fp16 12 add fp16 1
-44 gpu batchnorm fp16 12
-45 gpu add fp16 12
-46 gpu relu fp16 12
-47 gpu conv fp16 12 add fp16 1
-48 gpu batchnorm fp16 12
-49 gpu relu fp16 12
-50 gpu conv fp16 12 add fp16 1
-51 gpu batchnorm fp16 12
-52 gpu relu fp16 12
-53 gpu conv fp16 12 add fp16 1
-54 gpu batchnorm fp16 12
-55 gpu add fp16 12
-56 gpu relu fp16 12
-57 gpu conv fp16 12 add fp16 1
-58 gpu batchnorm fp16 12
-59 gpu relu fp16 12
-60 gpu conv fp16 12 add fp16 1
-61 gpu batchnorm fp16 12
-62 gpu relu fp16 12
-63 gpu conv fp16 12 add fp16 1
-64 gpu batchnorm fp16 12
-65 gpu add fp16 12
-66 gpu relu fp16 12
-67 gpu conv fp16 12 add fp16 1
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu conv fp16 12 add fp16 1
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv fp16 12 add fp16 1
-74 gpu batchnorm fp16 12
-75 gpu add fp16 12
-76 gpu relu fp16 12
-77 gpu conv fp16 12 add fp16 1
-78 gpu batchnorm fp16 12
-79 gpu relu fp16 12
-80 gpu conv fp16 12 add fp16 1
-81 gpu batchnorm fp16 12
-82 gpu relu fp16 12
-83 gpu conv fp16 12 add fp16 1
-84 gpu batchnorm fp16 12
-85 gpu conv fp16 12 add fp16 1
-86 gpu batchnorm fp16 12
-87 gpu add fp16 12
-88 gpu relu fp16 12
-89 gpu conv fp16 12 add fp16 1
-90 gpu batchnorm fp16 12
-91 gpu relu fp16 12
-92 gpu conv fp16 12 add fp16 1
-93 gpu batchnorm fp16 12
-94 gpu relu fp16 12
-95 gpu conv fp16 11 add fp16 1
-96 gpu batchnorm fp16 12
-97 gpu add fp16 12
-98 gpu relu fp16 12
-99 gpu conv fp16 12 add fp16 1
-100 gpu batchnorm fp16 12
-101 gpu relu fp16 12
-102 gpu conv perf_fp16 164 add fp16 1
-103 gpu batchnorm fp16 12
-104 gpu relu fp16 12
-105 gpu conv fp16 12 add fp16 1
-106 gpu batchnorm fp16 12
-107 gpu add fp16 12
-108 gpu relu fp16 12
-109 gpu conv fp16 12 add fp16 1
-110 gpu batchnorm fp16 12
-111 gpu relu fp16 12
-112 gpu conv fp16 12 add fp16 1
-113 gpu batchnorm fp16 12
-114 gpu relu fp16 12
-115 gpu conv fp16 12 add fp16 1
-116 gpu batchnorm fp16 12
-117 gpu add fp16 12
-118 gpu relu fp16 12
-119 gpu conv fp16 12 add fp16 1
-120 gpu batchnorm fp16 12
-121 gpu relu fp16 12
-122 gpu conv fp16 12 add fp16 1
-123 gpu batchnorm fp16 12
-124 gpu relu fp16 12
-125 gpu conv fp16 12 add fp16 1
-126 gpu batchnorm fp16 12
-127 gpu add fp16 12
-128 gpu relu fp16 12
-129 gpu conv fp16 12 add fp16 1
-130 gpu batchnorm fp16 12
-131 gpu relu fp16 12
-132 gpu conv fp16 12 add fp16 1
-133 gpu batchnorm fp16 12
-134 gpu relu fp16 12
-135 gpu conv fp16 12 add fp16 1
-136 gpu batchnorm fp16 12
-137 gpu add fp16 12
-138 gpu relu fp16 12
-139 gpu conv fp16 12 add fp16 1
-140 gpu batchnorm fp16 12
-141 gpu relu fp16 12
-142 gpu conv fp16 12 add fp16 1
-143 gpu batchnorm fp16 12
-144 gpu relu fp16 12
-145 gpu conv fp16 12 add fp16 1
-146 gpu batchnorm fp16 12
-147 gpu conv fp16 12 add fp16 1
-148 gpu batchnorm fp16 12
-149 gpu add fp16 12
-150 gpu relu fp16 12
-151 gpu conv fp16 12 add fp16 1
-152 gpu batchnorm fp16 12
-153 gpu relu fp16 12
-154 gpu conv fp16 12 add fp16 1
-155 gpu batchnorm fp16 12
-156 gpu relu fp16 12
-157 gpu conv fp16 12 add fp16 1
-158 gpu batchnorm fp16 12
-159 gpu add fp16 12
-160 gpu relu fp16 12
-161 gpu conv fp16 12 add fp16 1
-162 gpu batchnorm fp16 12
-163 gpu relu fp16 12
-164 gpu conv fp16 12 add fp16 1
-165 gpu batchnorm fp16 12
-166 gpu relu fp16 12
-167 gpu conv fp16 12 add fp16 1
-168 gpu batchnorm fp16 12
-169 gpu add fp16 12
-170 gpu relu fp16 12
-171 gpu pool_max fp16 12
-172 gpu mul fp16 12 add fp16 1
-173 gpu softmax fp32 1
------
-+++++
-conf4 1.8509087142956673 1.4713858340895483 74.68 1.019999999999996
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu batchnorm fp16 12
-3 gpu conv fp16 12 add fp16 1
-4 gpu batchnorm fp16 12
-5 gpu relu fp16 12
-6 gpu conv fp16 12 add fp16 1
-7 gpu batchnorm fp16 12
-8 gpu relu fp16 12
-9 gpu conv fp16 12 add fp16 1
-10 gpu batchnorm fp16 12
-11 gpu conv fp16 12 add fp16 1
-12 gpu batchnorm fp16 12
-13 gpu add fp16 12
-14 gpu relu fp16 12
-15 gpu conv fp16 12 add fp16 1
-16 gpu batchnorm fp16 12
-17 gpu relu fp16 12
-18 gpu conv fp16 12 add fp16 1
-19 gpu batchnorm fp16 12
-20 gpu relu fp16 12
-21 gpu conv fp16 12 add fp16 1
-22 gpu batchnorm fp16 12
-23 gpu add fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12 add fp16 1
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu conv perf_fp16 160 add fp16 1
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 12 add fp16 1
-32 gpu batchnorm fp16 12
-33 gpu add fp16 12
-34 gpu relu fp16 12
-35 gpu conv fp16 12 add fp16 1
-36 gpu batchnorm fp16 12
-37 gpu relu fp16 12
-38 gpu conv fp16 12 add fp16 1
-39 gpu batchnorm fp16 12
-40 gpu relu fp16 12
-41 gpu conv fp16 12 add fp16 1
-42 gpu batchnorm fp16 12
-43 gpu conv fp16 12 add fp16 1
-44 gpu batchnorm fp16 12
-45 gpu add fp16 12
-46 gpu relu fp16 12
-47 gpu conv fp16 12 add fp16 1
-48 gpu batchnorm fp16 12
-49 gpu relu fp16 12
-50 gpu conv fp16 12 add fp16 1
-51 gpu batchnorm fp16 12
-52 gpu relu fp16 12
-53 gpu conv fp16 12 add fp16 1
-54 gpu batchnorm fp16 12
-55 gpu add fp16 12
-56 gpu relu fp16 12
-57 gpu conv fp16 12 add fp16 1
-58 gpu batchnorm fp16 12
-59 gpu relu fp16 12
-60 gpu conv fp16 12 add fp16 1
-61 gpu batchnorm fp16 12
-62 gpu relu fp16 12
-63 gpu conv fp16 12 add fp16 1
-64 gpu batchnorm fp16 12
-65 gpu add fp16 12
-66 gpu relu fp16 12
-67 gpu conv fp16 12 add fp16 1
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu conv fp16 12 add fp16 1
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv fp16 12 add fp16 1
-74 gpu batchnorm fp16 12
-75 gpu add fp16 12
-76 gpu relu fp16 12
-77 gpu conv fp16 12 add fp16 1
-78 gpu batchnorm fp16 12
-79 gpu relu fp16 12
-80 gpu conv fp16 12 add fp16 1
-81 gpu batchnorm fp16 12
-82 gpu relu fp16 12
-83 gpu conv fp16 12 add fp16 1
-84 gpu batchnorm fp16 12
-85 gpu conv fp16 12 add fp16 1
-86 gpu batchnorm fp16 12
-87 gpu add fp16 12
-88 gpu relu fp16 12
-89 gpu conv fp16 12 add fp16 1
-90 gpu batchnorm fp16 12
-91 gpu relu fp16 12
-92 gpu conv fp16 12 add fp16 1
-93 gpu batchnorm fp16 12
-94 gpu relu fp16 12
-95 gpu conv fp16 12 add fp16 1
-96 gpu batchnorm fp16 12
-97 gpu add fp16 12
-98 gpu relu fp16 12
-99 gpu conv fp16 12 add fp16 1
-100 gpu batchnorm fp16 12
-101 gpu relu fp16 12
-102 gpu conv fp16 12 add fp16 1
-103 gpu batchnorm fp16 12
-104 gpu relu fp16 12
-105 gpu conv fp16 12 add fp16 1
-106 gpu batchnorm fp16 12
-107 gpu add fp16 12
-108 gpu relu fp16 12
-109 gpu conv fp16 12 add fp16 1
-110 gpu batchnorm fp16 12
-111 gpu relu fp16 12
-112 gpu conv fp16 12 add fp16 1
-113 gpu batchnorm fp16 12
-114 gpu relu fp16 12
-115 gpu conv fp16 12 add fp16 1
-116 gpu batchnorm fp16 12
-117 gpu add fp16 12
-118 gpu relu fp16 12
-119 gpu conv fp16 12 add fp16 1
-120 gpu batchnorm fp16 12
-121 gpu relu fp16 12
-122 gpu conv fp16 12 add fp16 1
-123 gpu batchnorm fp16 12
-124 gpu relu fp16 12
-125 gpu conv fp16 12 add fp16 1
-126 gpu batchnorm fp16 12
-127 gpu add fp16 12
-128 gpu relu fp16 12
-129 gpu conv fp16 12 add fp16 1
-130 gpu batchnorm fp16 12
-131 gpu relu fp16 12
-132 gpu conv fp16 12 add fp16 1
-133 gpu batchnorm fp16 12
-134 gpu relu fp16 12
-135 gpu conv fp16 12 add fp16 1
-136 gpu batchnorm fp16 12
-137 gpu add fp16 12
-138 gpu relu fp16 12
-139 gpu conv fp16 12 add fp16 1
-140 gpu batchnorm fp16 12
-141 gpu relu fp16 12
-142 gpu conv fp16 12 add fp16 1
-143 gpu batchnorm fp16 12
-144 gpu relu fp16 12
-145 gpu conv fp16 12 add fp16 1
-146 gpu batchnorm fp16 12
-147 gpu conv fp16 12 add fp16 1
-148 gpu batchnorm fp16 12
-149 gpu add fp16 12
-150 gpu relu fp16 12
-151 gpu conv fp16 12 add fp16 1
-152 gpu batchnorm fp16 12
-153 gpu relu fp16 12
-154 gpu conv fp16 12 add fp16 1
-155 gpu batchnorm fp16 12
-156 gpu relu fp16 12
-157 gpu conv fp16 12 add fp16 1
-158 gpu batchnorm fp16 12
-159 gpu add fp16 12
-160 gpu relu fp16 12
-161 gpu conv fp16 12 add fp16 1
-162 gpu batchnorm fp16 12
-163 gpu relu fp16 12
-164 gpu conv fp16 12 add fp16 1
-165 gpu batchnorm fp16 12
-166 gpu relu fp16 12
-167 gpu conv fp16 12 add fp16 1
-168 gpu batchnorm fp16 12
-169 gpu add fp16 12
-170 gpu relu fp16 12
-171 gpu pool_max fp16 12
-172 gpu mul fp16 12 add fp16 1
-173 gpu softmax fp32 1
------
-+++++
-conf5 1.8538077719438253 1.4749308494814874 73.82 1.8800000000000097
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu batchnorm fp16 12
-3 gpu conv fp16 12 add fp16 1
-4 gpu batchnorm fp16 12
-5 gpu relu fp16 12
-6 gpu conv fp16 12 add fp16 1
-7 gpu batchnorm fp16 12
-8 gpu relu fp16 12
-9 gpu conv fp16 12 add fp16 1
-10 gpu batchnorm fp16 12
-11 gpu conv fp16 12 add fp16 1
-12 gpu batchnorm fp16 12
-13 gpu add fp16 12
-14 gpu relu fp16 12
-15 gpu conv fp16 12 add fp16 1
-16 gpu batchnorm fp16 12
-17 gpu relu fp16 12
-18 gpu conv fp16 12 add fp16 1
-19 gpu batchnorm fp16 12
-20 gpu relu fp16 12
-21 gpu conv fp16 12 add fp16 1
-22 gpu batchnorm fp16 12
-23 gpu add fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12 add fp16 1
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu conv perf_fp16 160 add fp16 1
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 11 add fp16 1
-32 gpu batchnorm fp16 12
-33 gpu add fp16 12
-34 gpu relu fp16 12
-35 gpu conv fp16 12 add fp16 1
-36 gpu batchnorm fp16 12
-37 gpu relu fp16 12
-38 gpu conv fp16 12 add fp16 1
-39 gpu batchnorm fp16 12
-40 gpu relu fp16 12
-41 gpu conv fp16 12 add fp16 1
-42 gpu batchnorm fp16 12
-43 gpu conv fp16 12 add fp16 1
-44 gpu batchnorm fp16 12
-45 gpu add fp16 12
-46 gpu relu fp16 12
-47 gpu conv fp16 12 add fp16 1
-48 gpu batchnorm fp16 12
-49 gpu relu fp16 12
-50 gpu conv fp16 12 add fp16 1
-51 gpu batchnorm fp16 12
-52 gpu relu fp16 12
-53 gpu conv perf_fp16 153 add fp16 1
-54 gpu batchnorm fp16 12
-55 gpu add fp16 12
-56 gpu relu fp16 12
-57 gpu conv fp16 12 add fp16 1
-58 gpu batchnorm fp16 12
-59 gpu relu fp16 12
-60 gpu conv fp16 12 add fp16 1
-61 gpu batchnorm fp16 12
-62 gpu relu fp16 12
-63 gpu conv fp16 12 add fp16 1
-64 gpu batchnorm fp16 12
-65 gpu add fp16 12
-66 gpu relu fp16 12
-67 gpu conv fp16 12 add fp16 1
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu conv fp16 12 add fp16 1
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv fp16 12 add fp16 1
-74 gpu batchnorm fp16 12
-75 gpu add fp16 12
-76 gpu relu fp16 12
-77 gpu conv fp16 12 add fp16 1
-78 gpu batchnorm fp16 12
-79 gpu relu fp16 12
-80 gpu conv fp16 12 add fp16 1
-81 gpu batchnorm fp16 12
-82 gpu relu fp16 12
-83 gpu conv fp16 12 add fp16 1
-84 gpu batchnorm fp16 12
-85 gpu conv fp16 12 add fp16 1
-86 gpu batchnorm fp16 12
-87 gpu add fp16 12
-88 gpu relu fp16 12
-89 gpu conv fp16 12 add fp16 1
-90 gpu batchnorm fp16 12
-91 gpu relu fp16 12
-92 gpu conv fp16 12 add fp16 1
-93 gpu batchnorm fp16 12
-94 gpu relu fp16 12
-95 gpu conv fp16 11 add fp16 1
-96 gpu batchnorm fp16 12
-97 gpu add fp16 12
-98 gpu relu fp16 12
-99 gpu conv fp16 12 add fp16 1
-100 gpu batchnorm fp16 12
-101 gpu relu fp16 12
-102 gpu conv perf_fp16 164 add fp16 1
-103 gpu batchnorm fp16 12
-104 gpu relu fp16 12
-105 gpu conv fp16 12 add fp16 1
-106 gpu batchnorm fp16 12
-107 gpu add fp16 12
-108 gpu relu fp16 12
-109 gpu conv fp16 12 add fp16 1
-110 gpu batchnorm fp16 12
-111 gpu relu fp16 12
-112 gpu conv fp16 12 add fp16 1
-113 gpu batchnorm fp16 12
-114 gpu relu fp16 12
-115 gpu conv fp16 12 add fp16 1
-116 gpu batchnorm fp16 12
-117 gpu add fp16 12
-118 gpu relu fp16 12
-119 gpu conv samp_fp16 268 add fp16 1
-120 gpu batchnorm fp16 12
-121 gpu relu fp16 12
-122 gpu conv fp16 12 add fp16 1
-123 gpu batchnorm fp16 12
-124 gpu relu fp16 12
-125 gpu conv fp16 12 add fp16 1
-126 gpu batchnorm fp16 12
-127 gpu add fp16 12
-128 gpu relu fp16 12
-129 gpu conv fp16 12 add fp16 1
-130 gpu batchnorm fp16 12
-131 gpu relu fp16 12
-132 gpu conv fp16 12 add fp16 1
-133 gpu batchnorm fp16 12
-134 gpu relu fp16 12
-135 gpu conv fp16 11 add fp16 1
-136 gpu batchnorm fp16 12
-137 gpu add fp16 12
-138 gpu relu fp16 12
-139 gpu conv fp16 12 add fp16 1
-140 gpu batchnorm fp16 12
-141 gpu relu fp16 12
-142 gpu conv fp16 12 add fp16 1
-143 gpu batchnorm fp16 12
-144 gpu relu fp16 12
-145 gpu conv fp16 12 add fp16 1
-146 gpu batchnorm fp16 12
-147 gpu conv fp16 12 add fp16 1
-148 gpu batchnorm fp16 12
-149 gpu add fp16 12
-150 gpu relu fp16 12
-151 gpu conv fp16 12 add fp16 1
-152 gpu batchnorm fp16 12
-153 gpu relu fp16 12
-154 gpu conv fp16 12 add fp16 1
-155 gpu batchnorm fp16 12
-156 gpu relu fp16 12
-157 gpu conv fp16 11 add fp16 1
-158 gpu batchnorm fp16 12
-159 gpu add fp16 12
-160 gpu relu fp16 12
-161 gpu conv fp16 12 add fp16 1
-162 gpu batchnorm fp16 12
-163 gpu relu fp16 12
-164 gpu conv fp16 12 add fp16 1
-165 gpu batchnorm fp16 12
-166 gpu relu fp16 12
-167 gpu conv fp16 12 add fp16 1
-168 gpu batchnorm fp16 12
-169 gpu add fp16 12
-170 gpu relu fp16 12
-171 gpu pool_max fp16 12
-172 gpu mul fp16 12 add fp16 1
-173 gpu softmax fp32 1
------
-+++++
-conf6 1.8538077719438253 1.4749308494814874 73.7 2.0
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu batchnorm fp16 12
-3 gpu conv fp16 12 add fp16 1
-4 gpu batchnorm fp16 12
-5 gpu relu fp16 12
-6 gpu conv fp16 12 add fp16 1
-7 gpu batchnorm fp16 12
-8 gpu relu fp16 12
-9 gpu conv fp16 12 add fp16 1
-10 gpu batchnorm fp16 12
-11 gpu conv fp16 12 add fp16 1
-12 gpu batchnorm fp16 12
-13 gpu add fp16 12
-14 gpu relu fp16 12
-15 gpu conv fp16 12 add fp16 1
-16 gpu batchnorm fp16 12
-17 gpu relu fp16 12
-18 gpu conv fp16 12 add fp16 1
-19 gpu batchnorm fp16 12
-20 gpu relu fp16 12
-21 gpu conv fp16 12 add fp16 1
-22 gpu batchnorm fp16 12
-23 gpu add fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12 add fp16 1
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu conv perf_fp16 160 add fp16 1
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 11 add fp16 1
-32 gpu batchnorm fp16 12
-33 gpu add fp16 12
-34 gpu relu fp16 12
-35 gpu conv fp16 12 add fp16 1
-36 gpu batchnorm fp16 12
-37 gpu relu fp16 12
-38 gpu conv fp16 12 add fp16 1
-39 gpu batchnorm fp16 12
-40 gpu relu fp16 12
-41 gpu conv fp16 12 add fp16 1
-42 gpu batchnorm fp16 12
-43 gpu conv fp16 12 add fp16 1
-44 gpu batchnorm fp16 12
-45 gpu add fp16 12
-46 gpu relu fp16 12
-47 gpu conv fp16 12 add fp16 1
-48 gpu batchnorm fp16 12
-49 gpu relu fp16 12
-50 gpu conv fp16 12 add fp16 1
-51 gpu batchnorm fp16 12
-52 gpu relu fp16 12
-53 gpu conv perf_fp16 153 add fp16 1
-54 gpu batchnorm fp16 12
-55 gpu add fp16 12
-56 gpu relu fp16 12
-57 gpu conv fp16 12 add fp16 1
-58 gpu batchnorm fp16 12
-59 gpu relu fp16 12
-60 gpu conv fp16 12 add fp16 1
-61 gpu batchnorm fp16 12
-62 gpu relu fp16 12
-63 gpu conv fp16 12 add fp16 1
-64 gpu batchnorm fp16 12
-65 gpu add fp16 12
-66 gpu relu fp16 12
-67 gpu conv fp16 12 add fp16 1
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu conv fp16 12 add fp16 1
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv fp16 12 add fp16 1
-74 gpu batchnorm fp16 12
-75 gpu add fp16 12
-76 gpu relu fp16 12
-77 gpu conv fp16 12 add fp16 1
-78 gpu batchnorm fp16 12
-79 gpu relu fp16 12
-80 gpu conv fp16 12 add fp16 1
-81 gpu batchnorm fp16 12
-82 gpu relu fp16 12
-83 gpu conv fp16 12 add fp16 1
-84 gpu batchnorm fp16 12
-85 gpu conv fp16 12 add fp16 1
-86 gpu batchnorm fp16 12
-87 gpu add fp16 12
-88 gpu relu fp16 12
-89 gpu conv fp16 12 add fp16 1
-90 gpu batchnorm fp16 12
-91 gpu relu fp16 12
-92 gpu conv fp16 12 add fp16 1
-93 gpu batchnorm fp16 12
-94 gpu relu fp16 12
-95 gpu conv fp16 11 add fp16 1
-96 gpu batchnorm fp16 12
-97 gpu add fp16 12
-98 gpu relu fp16 12
-99 gpu conv fp16 12 add fp16 1
-100 gpu batchnorm fp16 12
-101 gpu relu fp16 12
-102 gpu conv perf_fp16 164 add fp16 1
-103 gpu batchnorm fp16 12
-104 gpu relu fp16 12
-105 gpu conv fp16 12 add fp16 1
-106 gpu batchnorm fp16 12
-107 gpu add fp16 12
-108 gpu relu fp16 12
-109 gpu conv fp16 12 add fp16 1
-110 gpu batchnorm fp16 12
-111 gpu relu fp16 12
-112 gpu conv fp16 12 add fp16 1
-113 gpu batchnorm fp16 12
-114 gpu relu fp16 12
-115 gpu conv fp16 12 add fp16 1
-116 gpu batchnorm fp16 12
-117 gpu add fp16 12
-118 gpu relu fp16 12
-119 gpu conv samp_fp16 268 add fp16 1
-120 gpu batchnorm fp16 12
-121 gpu relu fp16 12
-122 gpu conv fp16 12 add fp16 1
-123 gpu batchnorm fp16 12
-124 gpu relu fp16 12
-125 gpu conv fp16 12 add fp16 1
-126 gpu batchnorm fp16 12
-127 gpu add fp16 12
-128 gpu relu fp16 12
-129 gpu conv fp16 12 add fp16 1
-130 gpu batchnorm fp16 12
-131 gpu relu fp16 12
-132 gpu conv fp16 12 add fp16 1
-133 gpu batchnorm fp16 12
-134 gpu relu fp16 12
-135 gpu conv fp16 12 add fp16 1
-136 gpu batchnorm fp16 12
-137 gpu add fp16 12
-138 gpu relu fp16 12
-139 gpu conv fp16 12 add fp16 1
-140 gpu batchnorm fp16 12
-141 gpu relu fp16 12
-142 gpu conv fp16 12 add fp16 1
-143 gpu batchnorm fp16 12
-144 gpu relu fp16 12
-145 gpu conv fp16 12 add fp16 1
-146 gpu batchnorm fp16 12
-147 gpu conv fp16 12 add fp16 1
-148 gpu batchnorm fp16 12
-149 gpu add fp16 12
-150 gpu relu fp16 12
-151 gpu conv fp16 12 add fp16 1
-152 gpu batchnorm fp16 12
-153 gpu relu fp16 12
-154 gpu conv fp16 12 add fp16 1
-155 gpu batchnorm fp16 12
-156 gpu relu fp16 12
-157 gpu conv fp16 12 add fp16 1
-158 gpu batchnorm fp16 12
-159 gpu add fp16 12
-160 gpu relu fp16 12
-161 gpu conv fp16 12 add fp16 1
-162 gpu batchnorm fp16 12
-163 gpu relu fp16 12
-164 gpu conv fp16 12 add fp16 1
-165 gpu batchnorm fp16 12
-166 gpu relu fp16 12
-167 gpu conv fp16 12 add fp16 1
-168 gpu batchnorm fp16 12
-169 gpu add fp16 12
-170 gpu relu fp16 12
-171 gpu pool_max fp16 12
-172 gpu mul fp16 12 add fp16 1
-173 gpu softmax fp32 1
------
-+++++
-conf7 1.8577902325643394 1.478552049679054 72.82 2.8800000000000097
-1 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-2 gpu batchnorm fp16 12
-3 gpu conv fp16 12 add fp16 1
-4 gpu batchnorm fp16 12
-5 gpu relu fp16 12
-6 gpu conv fp16 12 add fp16 1
-7 gpu batchnorm fp16 12
-8 gpu relu fp16 12
-9 gpu conv fp16 12 add fp16 1
-10 gpu batchnorm fp16 12
-11 gpu conv fp16 12 add fp16 1
-12 gpu batchnorm fp16 12
-13 gpu add fp16 12
-14 gpu relu fp16 12
-15 gpu conv fp16 12 add fp16 1
-16 gpu batchnorm fp16 12
-17 gpu relu fp16 12
-18 gpu conv fp16 12 add fp16 1
-19 gpu batchnorm fp16 12
-20 gpu relu fp16 12
-21 gpu conv fp16 12 add fp16 1
-22 gpu batchnorm fp16 12
-23 gpu add fp16 12
-24 gpu relu fp16 12
-25 gpu conv fp16 12 add fp16 1
-26 gpu batchnorm fp16 12
-27 gpu relu fp16 12
-28 gpu conv perf_fp16 160 add fp16 1
-29 gpu batchnorm fp16 12
-30 gpu relu fp16 12
-31 gpu conv fp16 11 add fp16 1
-32 gpu batchnorm fp16 12
-33 gpu add fp16 12
-34 gpu relu fp16 12
-35 gpu conv fp16 12 add fp16 1
-36 gpu batchnorm fp16 12
-37 gpu relu fp16 12
-38 gpu conv fp16 12 add fp16 1
-39 gpu batchnorm fp16 12
-40 gpu relu fp16 12
-41 gpu conv fp16 12 add fp16 1
-42 gpu batchnorm fp16 12
-43 gpu conv fp16 12 add fp16 1
-44 gpu batchnorm fp16 12
-45 gpu add fp16 12
-46 gpu relu fp16 12
-47 gpu conv fp16 12 add fp16 1
-48 gpu batchnorm fp16 12
-49 gpu relu fp16 12
-50 gpu conv fp16 12 add fp16 1
-51 gpu batchnorm fp16 12
-52 gpu relu fp16 12
-53 gpu conv fp16 11 add fp16 1
-54 gpu batchnorm fp16 12
-55 gpu add fp16 12
-56 gpu relu fp16 12
-57 gpu conv samp_fp16 268 add fp16 1
-58 gpu batchnorm fp16 12
-59 gpu relu fp16 12
-60 gpu conv fp16 12 add fp16 1
-61 gpu batchnorm fp16 12
-62 gpu relu fp16 12
-63 gpu conv fp16 12 add fp16 1
-64 gpu batchnorm fp16 12
-65 gpu add fp16 12
-66 gpu relu fp16 12
-67 gpu conv fp16 12 add fp16 1
-68 gpu batchnorm fp16 12
-69 gpu relu fp16 12
-70 gpu conv fp16 12 add fp16 1
-71 gpu batchnorm fp16 12
-72 gpu relu fp16 12
-73 gpu conv fp16 12 add fp16 1
-74 gpu batchnorm fp16 12
-75 gpu add fp16 12
-76 gpu relu fp16 12
-77 gpu conv fp16 12 add fp16 1
-78 gpu batchnorm fp16 12
-79 gpu relu fp16 12
-80 gpu conv fp16 12 add fp16 1
-81 gpu batchnorm fp16 12
-82 gpu relu fp16 12
-83 gpu conv fp16 12 add fp16 1
-84 gpu batchnorm fp16 12
-85 gpu conv fp16 12 add fp16 1
-86 gpu batchnorm fp16 12
-87 gpu add fp16 12
-88 gpu relu fp16 12
-89 gpu conv fp16 12 add fp16 1
-90 gpu batchnorm fp16 12
-91 gpu relu fp16 12
-92 gpu conv fp16 12 add fp16 1
-93 gpu batchnorm fp16 12
-94 gpu relu fp16 12
-95 gpu conv fp16 11 add fp16 1
-96 gpu batchnorm fp16 12
-97 gpu add fp16 12
-98 gpu relu fp16 12
-99 gpu conv fp16 12 add fp16 1
-100 gpu batchnorm fp16 12
-101 gpu relu fp16 12
-102 gpu conv perf_fp16 164 add fp16 1
-103 gpu batchnorm fp16 12
-104 gpu relu fp16 12
-105 gpu conv fp16 12 add fp16 1
-106 gpu batchnorm fp16 12
-107 gpu add fp16 12
-108 gpu relu fp16 12
-109 gpu conv fp16 12 add fp16 1
-110 gpu batchnorm fp16 12
-111 gpu relu fp16 12
-112 gpu conv fp16 12 add fp16 1
-113 gpu batchnorm fp16 12
-114 gpu relu fp16 12
-115 gpu conv fp16 12 add fp16 1
-116 gpu batchnorm fp16 12
-117 gpu add fp16 12
-118 gpu relu fp16 12
-119 gpu conv samp_fp16 268 add fp16 1
-120 gpu batchnorm fp16 12
-121 gpu relu fp16 12
-122 gpu conv fp16 12 add fp16 1
-123 gpu batchnorm fp16 12
-124 gpu relu fp16 12
-125 gpu conv fp16 12 add fp16 1
-126 gpu batchnorm fp16 12
-127 gpu add fp16 12
-128 gpu relu fp16 12
-129 gpu conv fp16 12 add fp16 1
-130 gpu batchnorm fp16 12
-131 gpu relu fp16 12
-132 gpu conv fp16 12 add fp16 1
-133 gpu batchnorm fp16 12
-134 gpu relu fp16 12
-135 gpu conv perf_fp16 158 add fp16 1
-136 gpu batchnorm fp16 12
-137 gpu add fp16 12
-138 gpu relu fp16 12
-139 gpu conv fp16 12 add fp16 1
-140 gpu batchnorm fp16 12
-141 gpu relu fp16 12
-142 gpu conv fp16 12 add fp16 1
-143 gpu batchnorm fp16 12
-144 gpu relu fp16 12
-145 gpu conv fp16 12 add fp16 1
-146 gpu batchnorm fp16 12
-147 gpu conv fp16 12 add fp16 1
-148 gpu batchnorm fp16 12
-149 gpu add fp16 12
-150 gpu relu fp16 12
-151 gpu conv fp16 12 add fp16 1
-152 gpu batchnorm fp16 12
-153 gpu relu fp16 12
-154 gpu conv fp16 12 add fp16 1
-155 gpu batchnorm fp16 12
-156 gpu relu fp16 12
-157 gpu conv fp16 11 add fp16 1
-158 gpu batchnorm fp16 12
-159 gpu add fp16 12
-160 gpu relu fp16 12
-161 gpu conv fp16 12 add fp16 1
-162 gpu batchnorm fp16 12
-163 gpu relu fp16 12
-164 gpu conv fp16 12 add fp16 1
-165 gpu batchnorm fp16 12
-166 gpu relu fp16 12
-167 gpu conv fp16 12 add fp16 1
-168 gpu batchnorm fp16 12
-169 gpu add fp16 12
-170 gpu relu fp16 12
-171 gpu pool_max fp16 12
-172 gpu mul fp16 12 add fp16 1
-173 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/vgg16_cifar10/vgg16_cifar10.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/vgg16_cifar10/vgg16_cifar10.txt
deleted file mode 100644
index f4e185f358dbd2282b14c0865d829903d2d270e9..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/vgg16_cifar10/vgg16_cifar10.txt
+++ /dev/null
@@ -1,58 +0,0 @@
-3776.508929999999
-+++++
-conf1 1 1 89.96 0.0
-1 gpu conv fp32 11 add fp32 1 relu fp32 1
-2 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-3 gpu conv fp32 11 add fp32 1 relu fp32 1
-4 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-5 gpu conv fp32 11 add fp32 1 relu fp32 1
-6 gpu conv fp32 11 add fp32 1 relu fp32 1
-7 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-8 gpu conv fp32 11 add fp32 1 relu fp32 1
-9 gpu conv fp32 11 add fp32 1 relu fp32 1
-10 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-11 gpu conv fp32 11 add fp32 1 relu fp32 1
-12 gpu conv fp32 11 add fp32 1 relu fp32 1
-13 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-14 gpu mul fp32 11 add fp32 1 relu fp32 1
-15 gpu mul fp32 11 add fp32 1
-16 gpu softmax fp32 1
------
-+++++
-conf2 2.4192803184847484 2.2393153800931898 89.22 0.7399999999999949
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 266 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 152 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-12 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf3 2.1240075032467187 1.9749367321301132 88.64 1.3199999999999932
-1 gpu conv fp16 11 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 167 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 152 add fp16 1 relu fp16 1
-9 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-10 gpu conv fp16 11 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-12 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 269 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/vgg16_cifar100/vgg16_cifar100.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/vgg16_cifar100/vgg16_cifar100.txt
deleted file mode 100644
index b55bb668b140ebcc9ee911f728726afed7274f85..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/vgg16_cifar100/vgg16_cifar100.txt
+++ /dev/null
@@ -1,77 +0,0 @@
-3768.819777999999
-+++++
-conf1 1 1 66.5 0.0
-1 gpu conv fp32 11 add fp32 1 relu fp32 1
-2 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-3 gpu conv fp32 11 add fp32 1 relu fp32 1
-4 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-5 gpu conv fp32 11 add fp32 1 relu fp32 1
-6 gpu conv fp32 11 add fp32 1 relu fp32 1
-7 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-8 gpu conv fp32 11 add fp32 1 relu fp32 1
-9 gpu conv fp32 11 add fp32 1 relu fp32 1
-10 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-11 gpu conv fp32 11 add fp32 1 relu fp32 1
-12 gpu conv fp32 11 add fp32 1 relu fp32 1
-13 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-14 gpu mul fp32 11 add fp32 1 relu fp32 1
-15 gpu mul fp32 11 add fp32 1
-16 gpu softmax fp32 1
------
-+++++
-conf2 2.2793321208062913 2.0502797911533945 66.42 0.0799999999999983
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 269 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 268 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv fp16 12 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 267 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 268 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-12 gpu conv fp16 11 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf3 2.2793321208062913 2.0502797911533945 66.42 0.0799999999999983
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 269 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 268 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv fp16 12 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 267 add fp16 1 relu fp16 1
-10 gpu conv samp_fp16 268 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-12 gpu conv fp16 11 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
-+++++
-conf4 2.664296720624579 2.427276363573644 64.7 1.7999999999999972
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv perf_fp16 153 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-4 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-7 gpu conv samp_fp16 261 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv perf_fp16 155 add fp16 1 relu fp16 1
-9 gpu conv samp_fp16 262 add fp16 1 relu fp16 1
-10 gpu conv perf_fp16 151 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv perf_fp16 151 add fp16 1 relu fp16 1
-12 gpu conv samp_fp16 261 add fp16 1 relu fp16 1
-13 gpu conv samp_fp16 262 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1
-16 gpu softmax fp32 1
------
diff --git a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/vgg16_imagenet/vgg16_imagenet.txt b/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/vgg16_imagenet/vgg16_imagenet.txt
deleted file mode 100644
index d0a23ffb10367c45ab76e4477f29932a5431e68b..0000000000000000000000000000000000000000
--- a/llvm/projects/hpvm-tensor-rt/PPoPP_results/soc_sim_results/emp_time/vgg16_imagenet/vgg16_imagenet.txt
+++ /dev/null
@@ -1,41 +0,0 @@
-19194.623482
-+++++
-conf1 1 1 72.84 0.0
-1 gpu conv fp32 11 add fp32 1 relu fp32 1
-2 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-3 gpu conv fp32 11 add fp32 1 relu fp32 1
-4 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-5 gpu conv fp32 11 add fp32 1 relu fp32 1
-6 gpu conv fp32 11 add fp32 1 relu fp32 1
-7 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-8 gpu conv fp32 11 add fp32 1 relu fp32 1
-9 gpu conv fp32 11 add fp32 1 relu fp32 1
-10 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-11 gpu conv fp32 11 add fp32 1 relu fp32 1
-12 gpu conv fp32 11 add fp32 1 relu fp32 1
-13 gpu conv fp32 11 add fp32 1 relu fp32 1 pool_max fp32 1
-14 gpu mul fp32 11 add fp32 1 relu fp32 1
-15 gpu mul fp32 11 add fp32 1 relu fp32 1
-16 gpu mul fp32 11 add fp32 1
-17 gpu softmax fp32 1
------
-+++++
-conf2 1.7719381411481732 1.5850925672384186 72.84 0.0
-1 gpu conv fp16 12 add fp16 1 relu fp16 1
-2 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-3 gpu conv fp16 12 add fp16 1 relu fp16 1
-4 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-5 gpu conv fp16 12 add fp16 1 relu fp16 1
-6 gpu conv fp16 12 add fp16 1 relu fp16 1
-7 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-8 gpu conv fp16 12 add fp16 1 relu fp16 1
-9 gpu conv fp16 12 add fp16 1 relu fp16 1
-10 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-11 gpu conv fp16 12 add fp16 1 relu fp16 1
-12 gpu conv fp16 12 add fp16 1 relu fp16 1
-13 gpu conv fp16 12 add fp16 1 relu fp16 1 pool_max fp16 1
-14 gpu mul fp16 12 add fp16 1 relu fp16 1
-15 gpu mul fp16 12 add fp16 1 relu fp16 1
-16 gpu mul fp16 12 add fp16 1
-17 gpu softmax fp32 1
------