From 15b34fcfc8cfe0b042b93244ee0276106d2a3f8b Mon Sep 17 00:00:00 2001 From: Russel Arbore <russel.jma@gmail.com> Date: Mon, 3 Feb 2025 09:52:45 -0600 Subject: [PATCH 1/3] persist tmp dir if clang/nvcc fail --- juno_scheduler/src/pm.rs | 34 +++++++++++++++++++++++++++++----- 1 file changed, 29 insertions(+), 5 deletions(-) diff --git a/juno_scheduler/src/pm.rs b/juno_scheduler/src/pm.rs index 9c818707..20a3dba8 100644 --- a/juno_scheduler/src/pm.rs +++ b/juno_scheduler/src/pm.rs @@ -780,8 +780,15 @@ impl PassManager { .stdin(Stdio::piped()) .stdout(Stdio::piped()) .spawn() - .expect("Error running clang. Is it installed?"); - assert!(clang_process.wait().unwrap().success()); + .expect("PANIC: Error running clang. Is it installed?"); + if clang_process + .wait() + .map(|status| !status.success()) + .unwrap_or(false) + { + let path = tmp_dir.into_path(); + panic!("PANIC: Clang failed to compile the LLVM IR module. Persisting temporary directory ({}).", path.display()); + } let mut ar_args = vec!["crus", &output_archive, &llvm_object]; @@ -806,8 +813,15 @@ impl PassManager { .arg(&cuda_object) .arg(&cuda_path) .spawn() - .expect("Error running nvcc. Is it installed?"); - assert!(nvcc_process.wait().unwrap().success()); + .expect("PANIC: Error running NVCC. Is it installed?"); + if nvcc_process + .wait() + .map(|status| !status.success()) + .unwrap_or(false) + { + let path = tmp_dir.into_path(); + panic!("PANIC: NVCC failed to compile the CUDA module. Persisting temporary directory ({}).", path.display()); + } ar_args.push(&cuda_object); } @@ -816,7 +830,17 @@ impl PassManager { .args(&ar_args) .spawn() .expect("Error running ar. Is it installed?"); - assert!(ar_process.wait().unwrap().success()); + if ar_process + .wait() + .map(|status| !status.success()) + .unwrap_or(false) + { + let path = tmp_dir.into_path(); + panic!( + "PANIC: Ar failed to create a static library. Persisting temporary directory ({}).", + path.display() + ); + } // Write the Rust runtime into a file. let output_rt = format!("{}/rt_{}.hrt", output_dir, module_name); -- GitLab From a78fb8360bd51e20f1eead2af416f83cff8169c1 Mon Sep 17 00:00:00 2001 From: Russel Arbore <russel.jma@gmail.com> Date: Mon, 3 Feb 2025 10:33:23 -0600 Subject: [PATCH 2/3] add new test --- hercules_opt/src/forkify.rs | 30 ++++++++++++++----- .../fork_join_tests/src/fork_join_tests.jn | 13 ++++++++ juno_samples/fork_join_tests/src/gpu.sch | 3 ++ juno_samples/fork_join_tests/src/main.rs | 5 ++++ juno_scheduler/src/compile.rs | 1 + 5 files changed, 45 insertions(+), 7 deletions(-) diff --git a/hercules_opt/src/forkify.rs b/hercules_opt/src/forkify.rs index ec4e9fbc..65ccb586 100644 --- a/hercules_opt/src/forkify.rs +++ b/hercules_opt/src/forkify.rs @@ -10,8 +10,8 @@ use hercules_ir::*; use crate::*; -/* - * TODO: Forkify currently makes a bunch of small edits - this needs to be +/* + * TODO: Forkify currently makes a bunch of small edits - this needs to be * changed so that every loop that gets forkified corresponds to a single edit * + sub-edits. This would allow us to run forkify on a subset of a function. */ @@ -349,6 +349,23 @@ pub fn forkify_loop( }; let reduce_id = edit.add_node(reduce); + if edit.get_schedule(init).contains(&Schedule::ParallelReduce) + || edit + .get_schedule(continue_latch) + .contains(&Schedule::ParallelReduce) + { + edit = edit.add_schedule(reduce_id, Schedule::ParallelReduce)?; + } + if edit + .get_schedule(init) + .contains(&Schedule::TightAssociative) + || edit + .get_schedule(continue_latch) + .contains(&Schedule::TightAssociative) + { + edit = edit.add_schedule(reduce_id, Schedule::TightAssociative)?; + } + edit = edit.replace_all_uses_where(phi, reduce_id, |usee| *usee != reduce_id)?; edit = edit.replace_all_uses_where(continue_latch, reduce_id, |usee| { !loop_nodes.contains(usee) && *usee != reduce_id @@ -414,9 +431,8 @@ pub fn analyze_phis<'a>( phis: &'a [NodeID], loop_nodes: &'a HashSet<NodeID>, ) -> impl Iterator<Item = LoopPHI> + 'a { - - // Find data cycles within the loop of this phi, - // Start from the phis loop_continue_latch, and walk its uses until we find the original phi. + // Find data cycles within the loop of this phi, + // Start from the phis loop_continue_latch, and walk its uses until we find the original phi. phis.into_iter().map(move |phi| { let stop_on: HashSet<NodeID> = editor @@ -465,7 +481,7 @@ pub fn analyze_phis<'a>( .unwrap(); let loop_continue_latch = editor.node(phi).try_phi().unwrap().1[continue_idx]; - + let uses = walk_all_uses_stop_on(loop_continue_latch, editor, stop_on.clone()); let users = walk_all_users_stop_on(*phi, editor, stop_on.clone()); @@ -513,7 +529,7 @@ pub fn analyze_phis<'a>( // If some other node in the cycle is used, there is not a valid node to assign it after making the cycle a reduce. if intersection .iter() - .filter(|node| **node != loop_continue_latch ) + .filter(|node| **node != loop_continue_latch) .any(|data_node| { editor .get_users(*data_node) diff --git a/juno_samples/fork_join_tests/src/fork_join_tests.jn b/juno_samples/fork_join_tests/src/fork_join_tests.jn index 55e0a37e..bcfa1d25 100644 --- a/juno_samples/fork_join_tests/src/fork_join_tests.jn +++ b/juno_samples/fork_join_tests/src/fork_join_tests.jn @@ -59,3 +59,16 @@ fn test4(input : i32) -> i32[4, 4] { } return arr; } + +#[entry] +fn test5(input : i32) -> i32[4] { + let arr1 : i32[4]; + for i = 0 to 4 { + let red = arr1[i]; + for k = 0 to 3 { + red += k as i32; + } + @reduce arr1[i] = red + input; + } + return arr1; +} diff --git a/juno_samples/fork_join_tests/src/gpu.sch b/juno_samples/fork_join_tests/src/gpu.sch index 701c347c..6e2d6845 100644 --- a/juno_samples/fork_join_tests/src/gpu.sch +++ b/juno_samples/fork_join_tests/src/gpu.sch @@ -1,3 +1,5 @@ +parallel-reduce(test5@reduce); + gvn(*); phi-elim(*); dce(*); @@ -7,6 +9,7 @@ gpu(out.test1); gpu(out.test2); gpu(out.test3); gpu(out.test4); +gpu(out.test5); ip-sroa(*); sroa(*); diff --git a/juno_samples/fork_join_tests/src/main.rs b/juno_samples/fork_join_tests/src/main.rs index cbd42c50..4b6aba65 100644 --- a/juno_samples/fork_join_tests/src/main.rs +++ b/juno_samples/fork_join_tests/src/main.rs @@ -37,6 +37,11 @@ fn main() { let output = r.run(9).await; let correct = vec![63i32; 16]; assert(correct, output); + + let mut r = runner!(test5); + let output = r.run(4).await; + let correct = vec![7i32; 4]; + assert(correct, output); }); } diff --git a/juno_scheduler/src/compile.rs b/juno_scheduler/src/compile.rs index 11a8ec53..80cf2cb4 100644 --- a/juno_scheduler/src/compile.rs +++ b/juno_scheduler/src/compile.rs @@ -130,6 +130,7 @@ impl FromStr for Appliable { "parallel-fork" => Ok(Appliable::Schedule(Schedule::ParallelFork)), "parallel-reduce" => Ok(Appliable::Schedule(Schedule::ParallelReduce)), "vectorize" => Ok(Appliable::Schedule(Schedule::Vectorizable)), + "no-memset" | "no-reset" => Ok(Appliable::Schedule(Schedule::NoResetConstant)), _ => Err(s.to_string()), } -- GitLab From 40b1a7051700dca032b2747ce9fe57486089ad4e Mon Sep 17 00:00:00 2001 From: Russel Arbore <russel.jma@gmail.com> Date: Mon, 3 Feb 2025 10:37:12 -0600 Subject: [PATCH 3/3] fix --- hercules_opt/src/forkify.rs | 24 ++++++++++++++---------- 1 file changed, 14 insertions(+), 10 deletions(-) diff --git a/hercules_opt/src/forkify.rs b/hercules_opt/src/forkify.rs index 65ccb586..082f1ae9 100644 --- a/hercules_opt/src/forkify.rs +++ b/hercules_opt/src/forkify.rs @@ -349,19 +349,23 @@ pub fn forkify_loop( }; let reduce_id = edit.add_node(reduce); - if edit.get_schedule(init).contains(&Schedule::ParallelReduce) - || edit - .get_schedule(continue_latch) - .contains(&Schedule::ParallelReduce) + if (!edit.get_node(init).is_reduce() + && edit.get_schedule(init).contains(&Schedule::ParallelReduce)) + || (!edit.get_node(continue_latch).is_reduce() + && edit + .get_schedule(continue_latch) + .contains(&Schedule::ParallelReduce)) { edit = edit.add_schedule(reduce_id, Schedule::ParallelReduce)?; } - if edit - .get_schedule(init) - .contains(&Schedule::TightAssociative) - || edit - .get_schedule(continue_latch) - .contains(&Schedule::TightAssociative) + if (!edit.get_node(init).is_reduce() + && edit + .get_schedule(init) + .contains(&Schedule::TightAssociative)) + || (!edit.get_node(continue_latch).is_reduce() + && edit + .get_schedule(continue_latch) + .contains(&Schedule::TightAssociative)) { edit = edit.add_schedule(reduce_id, Schedule::TightAssociative)?; } -- GitLab