diff --git a/llvm/projects/soc_simulator/src/driver.py b/llvm/projects/soc_simulator/src/driver.py
index 85f17d687d645f99c6c95bf33a87b63004635b29..4eabc33818936c6bb932e0f57acb8c20d1b31e14 100644
--- a/llvm/projects/soc_simulator/src/driver.py
+++ b/llvm/projects/soc_simulator/src/driver.py
@@ -121,7 +121,7 @@ def is_promise(config_layer):
 
 
 def quantize(curr_layer, prev_layer, h2f_f2h_operation_ind, layer_data):
-    print(get_approx_type(curr_layer), get_approx_type(prev_layer))
+    #print(get_approx_type(curr_layer), get_approx_type(prev_layer))
     if curr_layer == prev_layer or curr_layer == ApproxTypes.PROMISE \
                 or prev_layer == ApproxTypes.PROMISE: # No quantization needed
         return 0.0, 0.0
@@ -149,9 +149,6 @@ def run_promise_simulation(swing, layer_data):
     patch_factor = 1 
 
     if is_conv(layer_name): 
-        print("is conv")
-        print(layer_data["N"], layer_data["H"], layer_data["W"], layer_data["Sh"], layer_data["Sw"], layer_data["Cin"], layer_data["Kh"], layer_data["Kw"], layer_data["Cout"])
-
         rows_a = layer_data["N"] * layer_data["H"] * layer_data["W"] \
 				/ (layer_data["Sh"] * layer_data["Sw"])
         cols_a = layer_data["Cin"] * layer_data["Kh"] * layer_data["Kw"]
@@ -159,7 +156,6 @@ def run_promise_simulation(swing, layer_data):
         cols_b = layer_data["Cout"]
         patch_factor = layer_data["Kh"] * layer_data["Kw"]
     elif is_fc(layer_name):
-        print("is fc")
         rows_a = layer_data["RA"] 
         cols_a = layer_data["CA"]
         rows_b = cols_a
@@ -232,7 +228,7 @@ def run_simulations(config_filename):
                 aggregate_results[results_energy_key][config_count] += energy 
 
             else:
-                print("Running layer %s on GPU" % layer_name)
+                print("Running layer %s on the GPU" % layer_name)
                 tensor_ops = config_layer.split(' ')
 
                 total_time = 0
@@ -298,4 +294,4 @@ if __name__ == "__main__":
     parse_tensor_layer_file("/home/nvidia/Gitlab/hpvm/llvm/projects/hpvm-tensor-rt/build_mobilenet/mobilenet_layers.txt")
     parse_tensor_table("/home/nvidia/Gitlab/hpvm/llvm/projects/hpvm-tensor-rt/build_pldi/mobilenet_results/mobilenet_tensors.txt")
     run_simulations("/home/nvidia/Gitlab/hpvm/llvm/projects/hpvm-tensor-rt/build_mobilenet/mobilenet_conf2.txt")
-    #display_results("blah.txt")
+    display_results("blah.txt")