From a42f98e92c74a10d8c511b528618b6a5cd4d8956 Mon Sep 17 00:00:00 2001 From: Hashim Sharif <hsharif3@miranda.cs.illinois.edu> Date: Mon, 25 Jan 2021 00:25:44 -0600 Subject: [PATCH] Moving prints and INFOS to DEBUG messages --- .../tensor_runtime/src/approx_simulation.cu | 2 +- hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/error.cu | 2 +- .../hpvm-tensor-rt/tensor_runtime/src/init_api.cc | 8 ++++---- .../hpvm-tensor-rt/tensor_runtime/src/tensor_utils.cu | 2 +- 4 files changed, 7 insertions(+), 7 deletions(-) diff --git a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/approx_simulation.cu b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/approx_simulation.cu index 1f69bb74d1..9a3c9ca848 100644 --- a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/approx_simulation.cu +++ b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/approx_simulation.cu @@ -931,7 +931,7 @@ int getSwing(int swing){ void initializeAutotuner(){ - printf("initializing tuner .... \n"); + DEBUG("initializing tuner .... \n"); sampParamSet = new SampParamSet; perfParamSet = new PerfParamSet; diff --git a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/error.cu b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/error.cu index 6b8ee15a42..7a700b435e 100644 --- a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/error.cu +++ b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/error.cu @@ -55,7 +55,7 @@ void readOpenTunerFlags(const char* file_name){ FILE* fp = fopen(file_name, "r"); if(fp == NULL){ - INFO("\nWARNING: File 'opentuner_flags' not found \n\n\n"); + DEBUG("\n WARNING: File 'opentuner_flags' not found \n\n\n"); return; } diff --git a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/init_api.cc b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/init_api.cc index 284a75c444..8b5c4aaf93 100644 --- a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/init_api.cc +++ b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/init_api.cc @@ -31,14 +31,14 @@ void llvm_hpvm_initTensorRt(int gpuid) { if (!runtime_initialized) { - printf("INITIALIZING GPU %d \n", gpuid); + INFO("INITIALIZING GPU %d \n", gpuid); // NOTE: Setting the target GPU. Can we use multiple GPUs? checkCudaErrors(cudaSetDevice(gpuid)); // Initializing cuDNN and cuBlas handles checkCudaErrors(cublasCreate(&cublasHandle)); checkCUDNN(cudnnCreate(&cudnnHandle)); - printf("CREATED HANDLES %d \n", gpuid); + DEBUG("CREATED HANDLES %d \n", gpuid); #ifdef PROMISE_TUNER_ENABLED // readOpenTunerFlags("opentuner_flags"); @@ -46,7 +46,7 @@ void llvm_hpvm_initTensorRt(int gpuid) { readOpenTunerFlags("promise_flags"); initializeAutotuner(); - printf("Read PROMISE FLAGS %d \n", gpuid); + DEBUG("Read PROMISE FLAGS %d \n", gpuid); #endif @@ -57,7 +57,7 @@ void llvm_hpvm_initTensorRt(int gpuid) { runtime_initialized = true; } - printf("DONE INTIALIZING GPU %d \n", gpuid); + INFO("DONE INTIALIZING GPU %d \n\n", gpuid); } void llvm_hpvm_cleanupTensorRt() { diff --git a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/tensor_utils.cu b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/tensor_utils.cu index 95a423da7e..2bc62057b5 100644 --- a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/tensor_utils.cu +++ b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/tensor_utils.cu @@ -426,7 +426,7 @@ void setTensorDescriptor(struct Tensor* tensor, int num_dims, if(tensor == NULL) return; - printf("**** cur_type = %d , half_type = %d \n", tensor->cur_type, half_type); + //printf("**** cur_type = %d , half_type = %d \n", tensor->cur_type, half_type); if (ONLINE_PROFILING){ if (tensor->cur_type == half_type) -- GitLab