diff --git a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/approx_simulation.cu b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/approx_simulation.cu
index 1f69bb74d184f59896494dbc9ea5b5a934f2ec67..9a3c9ca848d443a20f1dcbb98fb3eda52ee15945 100644
--- a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/approx_simulation.cu
+++ b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/approx_simulation.cu
@@ -931,7 +931,7 @@ int getSwing(int swing){
 
 void initializeAutotuner(){
 
-  printf("initializing tuner .... \n");
+  DEBUG("initializing tuner .... \n");
   
   sampParamSet = new SampParamSet;
   perfParamSet = new PerfParamSet;  
diff --git a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/error.cu b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/error.cu
index 6b8ee15a42106b2d6857065941324e50157763d5..7a700b435efe464153fbba7997662c7dfa970385 100644
--- a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/error.cu
+++ b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/error.cu
@@ -55,7 +55,7 @@ void readOpenTunerFlags(const char* file_name){
   
   FILE* fp = fopen(file_name, "r");
   if(fp == NULL){
-    INFO("\nWARNING: File 'opentuner_flags' not found \n\n\n");
+    DEBUG("\n WARNING: File 'opentuner_flags' not found \n\n\n");
     return;
   }
     
diff --git a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/init_api.cc b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/init_api.cc
index 284a75c444f54a0f3aa3412c8cd177d4ebad4e2e..8b5c4aaf93db40c038c4a9a30569318ae00d6be1 100644
--- a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/init_api.cc
+++ b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/init_api.cc
@@ -31,14 +31,14 @@ void llvm_hpvm_initTensorRt(int gpuid) {
 
   if (!runtime_initialized) {
 
-    printf("INITIALIZING GPU %d \n", gpuid);
+    INFO("INITIALIZING GPU %d \n", gpuid);
     // NOTE: Setting the target GPU. Can we use multiple GPUs?
     checkCudaErrors(cudaSetDevice(gpuid));
     // Initializing cuDNN and cuBlas handles
     checkCudaErrors(cublasCreate(&cublasHandle));
     checkCUDNN(cudnnCreate(&cudnnHandle));
 
-    printf("CREATED HANDLES %d \n", gpuid);
+    DEBUG("CREATED HANDLES %d \n", gpuid);
 
 #ifdef PROMISE_TUNER_ENABLED
     //    readOpenTunerFlags("opentuner_flags");
@@ -46,7 +46,7 @@ void llvm_hpvm_initTensorRt(int gpuid) {
     readOpenTunerFlags("promise_flags");
     initializeAutotuner();
 
-    printf("Read PROMISE FLAGS %d \n", gpuid);
+    DEBUG("Read PROMISE FLAGS %d \n", gpuid);
 
 #endif
 
@@ -57,7 +57,7 @@ void llvm_hpvm_initTensorRt(int gpuid) {
     runtime_initialized = true;
   }
 
-  printf("DONE INTIALIZING GPU %d \n", gpuid);
+  INFO("DONE INTIALIZING GPU %d \n\n", gpuid);
 }
 
 void llvm_hpvm_cleanupTensorRt() {
diff --git a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/tensor_utils.cu b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/tensor_utils.cu
index 95a423da7ec92e44fd192b4d4f91bb42312d9595..2bc62057b5c13161475b50b4a750da49146b97ce 100644
--- a/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/tensor_utils.cu
+++ b/hpvm/projects/hpvm-tensor-rt/tensor_runtime/src/tensor_utils.cu
@@ -426,7 +426,7 @@ void setTensorDescriptor(struct Tensor* tensor, int num_dims,
   if(tensor == NULL)
     return;
   
-  printf("**** cur_type = %d , half_type = %d \n", tensor->cur_type, half_type);
+  //printf("**** cur_type = %d , half_type = %d \n", tensor->cur_type, half_type);
 
   if (ONLINE_PROFILING){
     if (tensor->cur_type == half_type)