From 3cb06397636c82d518370b38494b45b97df560cb Mon Sep 17 00:00:00 2001 From: Elizabeth <hashim.sharif91@gmail.com> Date: Sat, 16 Nov 2019 11:32:09 -0600 Subject: [PATCH] Added debug prints --- .../src/driver_new_config_fp16_repl.py | 19 +++++++++++-------- 1 file changed, 11 insertions(+), 8 deletions(-) diff --git a/llvm/projects/soc_simulator/src/driver_new_config_fp16_repl.py b/llvm/projects/soc_simulator/src/driver_new_config_fp16_repl.py index ad60768cdd..28a883d160 100644 --- a/llvm/projects/soc_simulator/src/driver_new_config_fp16_repl.py +++ b/llvm/projects/soc_simulator/src/driver_new_config_fp16_repl.py @@ -220,6 +220,7 @@ class Driver: op_type = layer_as_lst[i] precision_type = layer_as_lst[i + 1] op_number = layer_as_lst[i + 2] + print(' '.join(layer_as_lst[i : i + 3])) approx_type = None if line.find("fp16") != -1: @@ -238,7 +239,7 @@ class Driver: curr_layer = Driver.PrecisionTypes.FP32 quant_time, quant_energy = self.__quantize(precision_type, op_number, curr_layer, prev_layer, tensor_count, layer_table_data) if quant_time != 0: - assert i == 2 and layer_ind == 0 + assert i == 2 #and layer_ind == 0 conv_time, conv_energy = self.__run_gpu_simulation(curr_layer, layer_name, \ tensor_count, approx_type, op_number) layer_results.append((quant_time + conv_time, quant_energy + conv_energy, ' '.join(layer_as_lst[i : i + 3]))) @@ -262,9 +263,8 @@ class Driver: layer_name = layer_table_data["Name"] for tensor_ind, (op_time, op_energy, tensor_op) in enumerate(layer): - # for each operation --> include quantization time quant_time, quant_energy = 0,0 - if layer_ind == 0: + if layer_ind == 0: # force quantization quant_time, quant_energy = self.__quantize("fp16", "1", Driver.PrecisionTypes.FP16, None, 0, layer_table_data) print("FP16 QUANT: ", quant_time, quant_energy) tensor_info = self.__tensor_table[layer_name][tensor_ind] @@ -420,7 +420,7 @@ class Driver: return curr_time, curr_energy def get_final_times_energies_conf(curr_conf, curr_conf_name): - print("_____________ NEW CONFIGURATION ___________") + #print("_____________ NEW CONFIGURATION ___________") final_time = final_energy = 0 final_conf = [] # List (conf) of lists (layers) of tuples (operation data) @@ -437,12 +437,15 @@ class Driver: baseline_time = self.fp16_baseline[layer_ind][1][tensor_ind][0] baseline_energy = self.fp16_baseline[layer_ind][1][tensor_ind][1] baseline_op = self.fp16_baseline[layer_ind][1][tensor_ind][2] - print(baseline_time, baseline_energy, baseline_op) - print(op_time, tensor_op) + #print(baseline_time, baseline_energy, baseline_op) + #print(op_time, tensor_op) final_tensor_op = tensor_op #print(op_time > baseline_time) if op_time > baseline_time: print("**************** BIGGER ******************") + print(curr_conf_name) + print(baseline_time, baseline_energy, baseline_op) + print(op_time, tensor_op) final_time += baseline_time final_energy += baseline_energy final_tensor_op = baseline_op @@ -468,8 +471,8 @@ class Driver: write_conf_to_file(conf_name, baseline_conf, 1, 1) else: curr_conf = self.__conf_results[conf_index] #conf_name] - final_time, final_energy, = get_baseline_times_energies(curr_conf) - #final_time, final_energy, curr_conf = get_final_times_energies_conf(curr_conf, conf_name) + #final_time, final_energy, = get_baseline_times_energies(curr_conf) + final_time, final_energy, curr_conf = get_final_times_energies_conf(curr_conf, conf_name) write_conf_to_file(conf_name, curr_conf, baseline_total_time / final_time, baseline_total_energy / final_energy) conf_index += 1 results_file.close() -- GitLab