diff --git a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/approx_simulation.cu b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/approx_simulation.cu index 1f69bb74d184f59896494dbc9ea5b5a934f2ec67..9a3c9ca848d443a20f1dcbb98fb3eda52ee15945 100644 --- a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/approx_simulation.cu +++ b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/approx_simulation.cu @@ -931,7 +931,7 @@ int getSwing(int swing){ void initializeAutotuner(){ - printf("initializing tuner .... \n"); + DEBUG("initializing tuner .... \n"); sampParamSet = new SampParamSet; perfParamSet = new PerfParamSet; diff --git a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/error.cu b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/error.cu index 6b8ee15a42106b2d6857065941324e50157763d5..7a700b435efe464153fbba7997662c7dfa970385 100644 --- a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/error.cu +++ b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/error.cu @@ -55,7 +55,7 @@ void readOpenTunerFlags(const char* file_name){ FILE* fp = fopen(file_name, "r"); if(fp == NULL){ - INFO("\nWARNING: File 'opentuner_flags' not found \n\n\n"); + DEBUG("\n WARNING: File 'opentuner_flags' not found \n\n\n"); return; } diff --git a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/init_api.cc b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/init_api.cc index 284a75c444f54a0f3aa3412c8cd177d4ebad4e2e..8b5c4aaf93db40c038c4a9a30569318ae00d6be1 100644 --- a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/init_api.cc +++ b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/init_api.cc @@ -31,14 +31,14 @@ void llvm_hpvm_initTensorRt(int gpuid) { if (!runtime_initialized) { - printf("INITIALIZING GPU %d \n", gpuid); + INFO("INITIALIZING GPU %d \n", gpuid); // NOTE: Setting the target GPU. Can we use multiple GPUs? checkCudaErrors(cudaSetDevice(gpuid)); // Initializing cuDNN and cuBlas handles checkCudaErrors(cublasCreate(&cublasHandle)); checkCUDNN(cudnnCreate(&cudnnHandle)); - printf("CREATED HANDLES %d \n", gpuid); + DEBUG("CREATED HANDLES %d \n", gpuid); #ifdef PROMISE_TUNER_ENABLED // readOpenTunerFlags("opentuner_flags"); @@ -46,7 +46,7 @@ void llvm_hpvm_initTensorRt(int gpuid) { readOpenTunerFlags("promise_flags"); initializeAutotuner(); - printf("Read PROMISE FLAGS %d \n", gpuid); + DEBUG("Read PROMISE FLAGS %d \n", gpuid); #endif @@ -57,7 +57,7 @@ void llvm_hpvm_initTensorRt(int gpuid) { runtime_initialized = true; } - printf("DONE INTIALIZING GPU %d \n", gpuid); + INFO("DONE INTIALIZING GPU %d \n\n", gpuid); } void llvm_hpvm_cleanupTensorRt() { diff --git a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/tensor_utils.cu b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/tensor_utils.cu index 95a423da7ec92e44fd192b4d4f91bb42312d9595..2bc62057b5c13161475b50b4a750da49146b97ce 100644 --- a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/tensor_utils.cu +++ b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/tensor_utils.cu @@ -426,7 +426,7 @@ void setTensorDescriptor(struct Tensor* tensor, int num_dims, if(tensor == NULL) return; - printf("**** cur_type = %d , half_type = %d \n", tensor->cur_type, half_type); + //printf("**** cur_type = %d , half_type = %d \n", tensor->cur_type, half_type); if (ONLINE_PROFILING){ if (tensor->cur_type == half_type)