diff --git a/Cargo.lock b/Cargo.lock
index 27bf3c1e6319f1652f967c878d8ab43d9fe65ee2..8b9e4c9e4d4ee280978801aed1b4b1ddc404c05a 100644
--- a/Cargo.lock
+++ b/Cargo.lock
@@ -395,6 +395,16 @@ dependencies = [
  "syn",
 ]
 
+[[package]]
+name = "dot"
+version = "0.1.0"
+dependencies = [
+ "async-std",
+ "clap",
+ "hercules_rt",
+ "rand",
+]
+
 [[package]]
 name = "either"
 version = "1.13.0"
@@ -450,6 +460,16 @@ dependencies = [
  "pin-project-lite",
 ]
 
+[[package]]
+name = "fac"
+version = "0.1.0"
+dependencies = [
+ "async-std",
+ "clap",
+ "hercules_rt",
+ "rand",
+]
+
 [[package]]
 name = "fastrand"
 version = "2.1.1"
@@ -592,16 +612,6 @@ dependencies = [
  "serde",
 ]
 
-[[package]]
-name = "hercules_dot"
-version = "0.1.0"
-dependencies = [
- "async-std",
- "clap",
- "hercules_rt",
- "rand",
-]
-
 [[package]]
 name = "hercules_driver"
 version = "0.1.0"
@@ -638,16 +648,6 @@ dependencies = [
  "serde",
 ]
 
-[[package]]
-name = "hercules_matmul"
-version = "0.1.0"
-dependencies = [
- "async-std",
- "clap",
- "hercules_rt",
- "rand",
-]
-
 [[package]]
 name = "hercules_opt"
 version = "0.1.0"
@@ -862,6 +862,16 @@ dependencies = [
  "vob",
 ]
 
+[[package]]
+name = "matmul"
+version = "0.1.0"
+dependencies = [
+ "async-std",
+ "clap",
+ "hercules_rt",
+ "rand",
+]
+
 [[package]]
 name = "memchr"
 version = "2.7.2"
diff --git a/Cargo.toml b/Cargo.toml
index 31639dce49115fca9a53fbd56a5d10cbb355b7be..322aa33dc141e68fc046f7c095a4b7062869df2d 100644
--- a/Cargo.toml
+++ b/Cargo.toml
@@ -11,11 +11,10 @@ members = [
 	"hercules_test/hercules_tests",
 	
 	"hercules_tools/hercules_driver",
-	#"hercules_tools/hercules_hbin_dump",
 
 	"juno_frontend",
 
 	"hercules_samples/dot",
 	"hercules_samples/matmul",
-	#"hercules_samples/task_parallel"
+	"hercules_samples/fac",
 ]
diff --git a/hercules_cg/src/sched_gen.rs b/hercules_cg/src/sched_gen.rs
index 758eb786cc42eb423df06c2188c21e50205456a1..4f9decce5d66e44e3017090d574c38b7d7813793 100644
--- a/hercules_cg/src/sched_gen.rs
+++ b/hercules_cg/src/sched_gen.rs
@@ -948,20 +948,30 @@ impl<'a> FunctionContext<'a> {
                             // of the SFunction corresponding to this phi.
                             found_in_partition_predecessor = true;
                             Some((*block_id, get_svalue(*data_id).clone()))
-                        } else {
-                            // Don't add multiple inputs for block #0.
+                        } else if let Some(param_idx) = manifest.partitions[partition_idx]
+                            .parameters
+                            .iter()
+                            .position(|(_, kind)| *kind == ParameterKind::DataInput(id))
+                        {
+                            // This input to the phi is corresponds to all of
+                            // the inputs from control locations outside this
+                            // partition. This does *not* include constant nodes
+                            // in other partitions - those get propagated (see
+                            // below). Don't add multiple inputs for block #0.
                             if found_out_of_partition_predecessor {
                                 return None;
                             }
-                            // This predecessor for the phi gets passed in via a
-                            // parameter set up for this phi.
+                            // This predecessor for the phi gets passed in
+                            // via a parameter set up for this phi.
                             found_out_of_partition_predecessor = true;
-                            let param_idx = manifest.partitions[partition_idx]
-                                .parameters
-                                .iter()
-                                .position(|(_, kind)| *kind == ParameterKind::DataInput(id))
-                                .unwrap();
                             Some((BlockID::new(0), SValue::VirtualRegister(param_idx)))
+                        } else {
+                            // This input to the phi is a constant located
+                            // outside this partition these get propagated in
+                            // schedule IR.
+                            found_in_partition_predecessor = true;
+                            let svalue = get_svalue(*data_id).clone();
+                            Some((BlockID::new(0), svalue))
                         }
                     })
                     .collect();
diff --git a/hercules_ir/src/dot.rs b/hercules_ir/src/dot.rs
index 156433885ea09a3dc9f907fdbb6cc729d28f3b4c..97520382fb5ae63d1c6e2fa5b50632e09a9db8e2 100644
--- a/hercules_ir/src/dot.rs
+++ b/hercules_ir/src/dot.rs
@@ -27,7 +27,7 @@ pub fn xdot_module(
     let mut rng = rand::thread_rng();
     let num: u64 = rng.gen();
     tmp_path.push(format!("hercules_dot_{}.dot", num));
-    let mut file = File::create(tmp_path.clone()).expect("PANIC: Unable to open output file.");
+    let mut file = File::create(&tmp_path).expect("PANIC: Unable to open output file.");
     let mut contents = String::new();
     write_dot(
         &module,
@@ -77,6 +77,7 @@ pub fn write_dot<W: Write>(
         // Step 1: draw IR graph itself. This includes all IR nodes and all edges
         // between IR nodes.
         for partition_idx in 0..plan.map_or(1, |plan| plan.num_partitions) {
+            // First, write all the nodes in their subgraph.
             let partition_color = plan.map(|plan| match plan.partition_devices[partition_idx] {
                 Device::CPU => "lightblue",
                 Device::GPU => "darkseagreen",
@@ -84,7 +85,6 @@ pub fn write_dot<W: Write>(
             if let Some(partition_color) = partition_color {
                 write_partition_header(function_id, partition_idx, module, partition_color, w)?;
             }
-
             let nodes_ids = if let Some(partition_to_node_map) = &mut partition_to_node_map {
                 let mut empty = vec![];
                 std::mem::swap(&mut partition_to_node_map[partition_idx], &mut empty);
@@ -109,7 +109,15 @@ pub fn write_dot<W: Write>(
                     plan.map_or(&vec![], |plan| &plan.schedules[node_id.idx()]),
                     w,
                 )?;
+            }
+            if plans.is_some() {
+                write_graph_footer(w)?;
+            }
 
+            // Second, write all the edges coming out of a node.
+            for node_id in nodes_ids.iter() {
+                let node = &function.nodes[node_id.idx()];
+                let dst_control = node.is_control();
                 for u in def_use::get_uses(&node).as_ref() {
                     let src_control = function.nodes[u.idx()].is_control();
 
@@ -151,9 +159,6 @@ pub fn write_dot<W: Write>(
                     )?;
                 }
             }
-            if plans.is_some() {
-                write_graph_footer(w)?;
-            }
         }
 
         // Step 2: draw dominance edges in dark green. Don't draw post dominance
diff --git a/hercules_ir/src/schedule.rs b/hercules_ir/src/schedule.rs
index 2c6335147658dac3d590458c5585f92956c93d42..7c9fdc1058aa30601b04f3e4f7de37c85784c8dc 100644
--- a/hercules_ir/src/schedule.rs
+++ b/hercules_ir/src/schedule.rs
@@ -1,5 +1,4 @@
-use std::collections::HashMap;
-use std::collections::VecDeque;
+use std::collections::{HashMap, VecDeque};
 use std::iter::{repeat, zip};
 
 use crate::*;
@@ -288,6 +287,9 @@ impl Plan {
         let partition_graph = partition_graph(function, def_use, self);
         let dom = dominator(&partition_graph, NodeID::new(self.partitions[0].idx()));
         for id in (0..function.nodes.len()).map(NodeID::new) {
+            if function.nodes[id.idx()].is_control() {
+                continue;
+            }
             let part = self.partitions[id.idx()];
 
             // Check condition #1.
@@ -416,17 +418,19 @@ impl Plan {
             let data_inputs = &mut data_inputs[self.partitions[id.idx()].idx()];
             let uses = get_uses(&function.nodes[id.idx()]);
             for use_id in uses.as_ref() {
-                // For every phi node, if any one of its uses is defined in a
-                // different partition, then the phi node itself, not its
-                // outside uses, is considered a data input. This is because a
-                // phi node whose uses are all in a different partition should
-                // be lowered to a single parameter to the corresponding simple
-                // IR function. Note that for a phi node with some uses outside
-                // and some uses inside the partition, the uses outside the
-                // partition become a single parameter to the schedule IR
-                // function, and that parameter and all of the "inside" uses
-                // become the inputs to a phi inside the simple IR function.
+                // For every phi node, if any one of its non-constant uses is
+                // defined in a different partition, then the phi node itself,
+                // not its outside uses, is considered a data input. This is
+                // because a phi node whose uses are all in a different
+                // partition should be lowered to a single parameter to the
+                // corresponding simple IR function. Note that for a phi node
+                // with some uses outside and some uses inside the partition,
+                // the uses outside the partition become a single parameter to
+                // the schedule IR function, and that parameter and all of the
+                // "inside" uses become the inputs to a phi inside the simple IR
+                // function.
                 if self.partitions[id.idx()] != self.partitions[use_id.idx()]
+                    && !function.nodes[use_id.idx()].is_constant()
                     && !data_inputs.contains(&id)
                 {
                     data_inputs.push(id);
@@ -484,62 +488,30 @@ impl Plan {
         }
         data_outputs
     }
-}
 
-impl GraveUpdatable for Plan {
-    fn fix_gravestones(&mut self, grave_mapping: &Vec<NodeID>) {
-        self.schedules.fix_gravestones(grave_mapping);
-        self.partitions.fix_gravestones(grave_mapping);
-        self.partition_devices.fix_gravestones(grave_mapping);
+    pub fn renumber_partitions(&mut self) {
         let mut renumber_partitions = HashMap::new();
         for id in self.partitions.iter_mut() {
             let next_id = PartitionID::new(renumber_partitions.len());
-            *id = *renumber_partitions.entry(*id).or_insert(next_id);
+            let old_id = *id;
+            let new_id = *renumber_partitions.entry(old_id).or_insert(next_id);
+            *id = new_id;
+        }
+        let mut new_devices = vec![Device::CPU; renumber_partitions.len()];
+        for (old_id, new_id) in renumber_partitions.iter() {
+            new_devices[new_id.idx()] = self.partition_devices[old_id.idx()];
         }
+        self.partition_devices = new_devices;
         self.num_partitions = renumber_partitions.len();
     }
 }
 
-/*
- * A "default" plan should be available, where few schedules are used and
- * conservative partitioning is enacted. Only schedules that can be proven safe
- * by the compiler should be included.
- */
-pub fn default_plan(
-    function: &Function,
-    dynamic_constants: &Vec<DynamicConstant>,
-    def_use: &ImmutableDefUseMap,
-    reverse_postorder: &Vec<NodeID>,
-    fork_join_map: &HashMap<NodeID, NodeID>,
-    fork_join_nesting: &HashMap<NodeID, Vec<NodeID>>,
-    bbs: &Vec<NodeID>,
-) -> Plan {
-    // Start by creating a completely bare-bones plan doing nothing interesting.
-    let mut plan = Plan {
-        schedules: vec![vec![]; function.nodes.len()],
-        partitions: vec![PartitionID::new(0); function.nodes.len()],
-        partition_devices: vec![Device::CPU; 1],
-        num_partitions: 1,
-    };
-
-    // Infer schedules.
-    infer_parallel_reduce(function, fork_join_map, &mut plan);
-    infer_parallel_fork(
-        function,
-        def_use,
-        fork_join_map,
-        fork_join_nesting,
-        &mut plan,
-    );
-    infer_vectorizable(function, dynamic_constants, fork_join_map, &mut plan);
-    infer_associative(function, &mut plan);
-
-    // Infer a partitioning.
-    partition_out_forks(function, reverse_postorder, fork_join_map, bbs, &mut plan);
-    // TODO: uncomment once GPU backend is implemented.
-    // place_fork_partitions_on_gpu(function, &mut plan);
-
-    plan
+impl GraveUpdatable for Plan {
+    fn fix_gravestones(&mut self, grave_mapping: &Vec<NodeID>) {
+        self.schedules.fix_gravestones(grave_mapping);
+        self.partitions.fix_gravestones(grave_mapping);
+        self.renumber_partitions();
+    }
 }
 
 /*
diff --git a/hercules_opt/src/editor.rs b/hercules_opt/src/editor.rs
index 6b26e493ddb25ea56bd67871808a8e95e3e50e8d..0bd4397be4f40e29d2ef2ac765ea06129e1bafb8 100644
--- a/hercules_opt/src/editor.rs
+++ b/hercules_opt/src/editor.rs
@@ -3,7 +3,8 @@ extern crate either;
 extern crate hercules_ir;
 extern crate itertools;
 
-use std::collections::{BTreeMap, HashSet};
+use std::collections::{BTreeMap, HashMap, HashSet, VecDeque};
+use std::iter::FromIterator;
 use std::mem::take;
 
 use self::bitvec::prelude::*;
@@ -383,14 +384,18 @@ pub fn repair_plan(plan: &mut Plan, new_function: &Function, edits: &[Edit]) {
         .map(|part| Some(part))
         .collect();
     new_partitions.resize(new_function.nodes.len(), None);
-    // Iterate the added control nodes in reverse postorder.
-    for control_id in added_control_nodes {
+    // Iterate the added control nodes using a worklist.
+    let mut worklist = VecDeque::from(added_control_nodes);
+    while let Some(control_id) = worklist.pop_front() {
         let node = &new_function.nodes[control_id.idx()];
-        // There are three cases where this control node needs to start a new
+        // There are five cases where this control node needs to start a new
         // partition:
-        // 1. It's a top-level fork.
-        // 2. One of its control predecessors is a top-level join.
-        // 3. It's a region node where not every predecessor is in the same
+        // 1. It's a non-gravestone start node. This is any start node visited
+        //    by the reverse postorder.
+        // 2. It's a return node.
+        // 3. It's a top-level fork.
+        // 4. One of its control predecessors is a top-level join.
+        // 5. It's a region node where not every predecessor is in the same
         //    partition (equivalently, not every predecessor is in the same
         //    partition - only region nodes can have multiple predecessors).
         let top_level_fork = node.is_fork() && fork_join_nesting[&control_id].len() == 1;
@@ -404,7 +409,12 @@ pub fn repair_plan(plan: &mut Plan, new_function: &Function, edits: &[Edit]) {
             .map(|pred| new_partitions[pred.idx()])
             .all_equal();
 
-        if top_level_fork || top_level_join || multi_pred_region {
+        if node.is_start()
+            || node.is_return()
+            || top_level_fork
+            || top_level_join
+            || multi_pred_region
+        {
             // This control node goes in a new partition.
             let part_id = PartitionID::new(plan.num_partitions);
             plan.num_partitions += 1;
@@ -412,10 +422,12 @@ pub fn repair_plan(plan: &mut Plan, new_function: &Function, edits: &[Edit]) {
         } else {
             // This control node goes in the partition of any one of its
             // predecessors. They're all the same by condition 3 above.
-            new_partitions[control_id.idx()] = control_subgraph
-                .preds(control_id)
-                .filter_map(|pred_id| new_partitions[pred_id.idx()])
-                .next();
+            let any_pred = control_subgraph.preds(control_id).next().unwrap();
+            if new_partitions[any_pred.idx()].is_some() {
+                new_partitions[control_id.idx()] = new_partitions[any_pred.idx()];
+            } else {
+                worklist.push_back(control_id);
+            }
         }
     }
 
@@ -439,7 +451,56 @@ pub fn repair_plan(plan: &mut Plan, new_function: &Function, edits: &[Edit]) {
     // Step 6: wrap everything up.
     plan.partitions = new_partitions.into_iter().map(|id| id.unwrap()).collect();
     plan.partition_devices
-        .resize(new_function.nodes.len(), Device::CPU);
+        .resize(plan.num_partitions, Device::CPU);
+}
+
+/*
+ * Default plans can be constructed by conservatively inferring schedules and
+ * creating partitions by "repairing" a partition where the edit is adding every
+ * node in the function.
+ */
+pub fn default_plan(
+    function: &Function,
+    dynamic_constants: &Vec<DynamicConstant>,
+    def_use: &ImmutableDefUseMap,
+    reverse_postorder: &Vec<NodeID>,
+    fork_join_map: &HashMap<NodeID, NodeID>,
+    fork_join_nesting: &HashMap<NodeID, Vec<NodeID>>,
+    bbs: &Vec<NodeID>,
+) -> Plan {
+    // Start by creating a completely bare-bones plan doing nothing interesting.
+    let mut plan = Plan {
+        schedules: vec![vec![]; function.nodes.len()],
+        partitions: vec![],
+        partition_devices: vec![],
+        num_partitions: 1,
+    };
+
+    // Infer a partitioning by using `repair_plan`, where the "edit" is creating
+    // the entire function.
+    let edit = (
+        HashSet::new(),
+        HashSet::from_iter((0..function.nodes.len()).map(NodeID::new)),
+    );
+    repair_plan(&mut plan, function, &[edit]);
+    plan.renumber_partitions();
+
+    // Infer schedules.
+    infer_parallel_reduce(function, fork_join_map, &mut plan);
+    infer_parallel_fork(
+        function,
+        def_use,
+        fork_join_map,
+        fork_join_nesting,
+        &mut plan,
+    );
+    infer_vectorizable(function, dynamic_constants, fork_join_map, &mut plan);
+    infer_associative(function, &mut plan);
+
+    // TODO: uncomment once GPU backend is implemented.
+    // place_fork_partitions_on_gpu(function, &mut plan);
+
+    plan
 }
 
 #[cfg(test)]
diff --git a/hercules_samples/dot/Cargo.toml b/hercules_samples/dot/Cargo.toml
index fe5873abd363aea21b8c4e24bdc497c26ecd27f1..ab10aedaeb3c680c1c26f3bc425178f34c6485c9 100644
--- a/hercules_samples/dot/Cargo.toml
+++ b/hercules_samples/dot/Cargo.toml
@@ -1,5 +1,5 @@
 [package]
-name = "hercules_dot"
+name = "dot"
 version = "0.1.0"
 authors = ["Russel Arbore <rarbore2@illinois.edu>"]
 edition = "2021"
diff --git a/hercules_samples/dot/src/main.rs b/hercules_samples/dot/src/main.rs
index 624c193761b2c1996dcc91eee3e182775103ddaa..dbf18713465c10f7fafc7de0e180a6d43f0003a6 100644
--- a/hercules_samples/dot/src/main.rs
+++ b/hercules_samples/dot/src/main.rs
@@ -3,9 +3,9 @@ extern crate clap;
 extern crate hercules_rt;
 
 // To compile currently, run from the Hercules project root directory:
-// cargo run --bin hercules_driver hercules_samples/dot/dot.hir "Codegen"
+// cargo run --bin hercules_driver hercules_samples/dot/dot.hir "Codegen(\"hercules_samples/dot\",\"dot\")"
 // Then, you can execute this example with:
-// cargo run --bin hercules_dot
+// cargo run --bin dot
 hercules_rt::use_hman!("hercules_samples/dot/dot.hman");
 
 fn main() {
diff --git a/hercules_samples/task_parallel/Cargo.toml b/hercules_samples/fac/Cargo.toml
similarity index 87%
rename from hercules_samples/task_parallel/Cargo.toml
rename to hercules_samples/fac/Cargo.toml
index 76b678cc3b4ac042f0f48bb247c7ed58f4dc5efa..bef11a91f027edf84f1e719538168e4ce8fd21e7 100644
--- a/hercules_samples/task_parallel/Cargo.toml
+++ b/hercules_samples/fac/Cargo.toml
@@ -1,5 +1,5 @@
 [package]
-name = "hercules_task_parallel"
+name = "fac"
 version = "0.1.0"
 authors = ["Russel Arbore <rarbore2@illinois.edu>"]
 edition = "2021"
@@ -7,4 +7,5 @@ edition = "2021"
 [dependencies]
 clap = { version = "*", features = ["derive"] }
 hercules_rt = { path = "../../hercules_rt" }
+rand = "*"
 async-std = "*"
diff --git a/hercules_samples/fac/build.rs b/hercules_samples/fac/build.rs
new file mode 100644
index 0000000000000000000000000000000000000000..f43f9fbd956813de7f530d93c3afbb94a29995c6
--- /dev/null
+++ b/hercules_samples/fac/build.rs
@@ -0,0 +1,12 @@
+use std::env::current_dir;
+
+fn main() {
+    println!(
+        "cargo::rustc-link-search=native={}",
+        current_dir().unwrap().display()
+    );
+    println!("cargo::rustc-link-lib=static=fac");
+
+    println!("cargo::rerun-if-changed=fac.hman");
+    println!("cargo::rerun-if-changed=libfac.a");
+}
diff --git a/hercules_samples/fac/fac.hir b/hercules_samples/fac/fac.hir
new file mode 100644
index 0000000000000000000000000000000000000000..e43dd8cae1a605bca7c3ceac4eb7c029665e86e6
--- /dev/null
+++ b/hercules_samples/fac/fac.hir
@@ -0,0 +1,13 @@
+fn fac(x: i32) -> i32
+  zero = constant(i32, 0)
+  one = constant(i32, 1)
+  loop = region(start, if_true)
+  idx = phi(loop, zero, idx_inc)
+  idx_inc = add(idx, one)
+  fac = phi(loop, one, fac_acc)
+  fac_acc = mul(fac, idx_inc)
+  in_bounds = lt(idx_inc, x)
+  if = if(loop, in_bounds)
+  if_false = projection(if, 0)
+  if_true = projection(if, 1)
+  r = return(if_false, fac_acc)
diff --git a/hercules_samples/fac/src/main.rs b/hercules_samples/fac/src/main.rs
new file mode 100644
index 0000000000000000000000000000000000000000..5e5b4ddefb73835c45b399009e9709b06e056217
--- /dev/null
+++ b/hercules_samples/fac/src/main.rs
@@ -0,0 +1,16 @@
+extern crate async_std;
+extern crate clap;
+extern crate hercules_rt;
+
+// To compile currently, run from the Hercules project root directory:
+// cargo run --bin hercules_driver hercules_samples/fac/fac.hir "Codegen(\"hercules_samples/fac\",\"fac\")"
+// Then, you can execute this example with:
+// cargo run --bin fac
+hercules_rt::use_hman!("hercules_samples/fac/fac.hman");
+
+fn main() {
+    async_std::task::block_on(async {
+        let f = unsafe { fac(8).await };
+        println!("{}", f);
+    });
+}
diff --git a/hercules_samples/invalid/bad_phi.hir b/hercules_samples/invalid/bad_phi.hir
deleted file mode 100644
index f04a020276657a0bc0e8bee9844df29711f9552a..0000000000000000000000000000000000000000
--- a/hercules_samples/invalid/bad_phi.hir
+++ /dev/null
@@ -1,14 +0,0 @@
-fn func(x: i32) -> i32
-   zero = constant(i32, 0)
-   cond = lt(x, zero)
-   if1 = if(start, cond)
-   a1 = read_prod(if1, 0)
-   b1 = read_prod(if1, 1)
-   reg1 = region(a1, b1)
-   phi1 = phi(reg1, x, phi2)
-   if2 = if(reg1, cond)
-   a2 = read_prod(if2, 0)
-   b2 = read_prod(if2, 1)
-   reg2 = region(a2, b2)
-   phi2 = phi(reg2, phi1, x)
-   r = return(reg2, x)
\ No newline at end of file
diff --git a/hercules_samples/invalid/bad_phi2.hir b/hercules_samples/invalid/bad_phi2.hir
deleted file mode 100644
index c03628cb01c55c9b49c7d85d9d81849edaf2de0e..0000000000000000000000000000000000000000
--- a/hercules_samples/invalid/bad_phi2.hir
+++ /dev/null
@@ -1,18 +0,0 @@
-fn tricky(x: i32) -> i32
-  one = constant(i32, 1)
-  two = constant(i32, 2)
-  loop = region(start, if2_true)
-  idx = phi(loop, x, idx_dec)
-  val = phi(loop, later_val, one)
-  b = ne(one, val)
-  if1 = if(loop, b)
-  if1_false = read_prod(if1, 0)
-  if1_true = read_prod(if1, 1)
-  middle = region(if1_false, if1_true)
-  later_val = phi(middle, val, two)
-  idx_dec = sub(idx, one)
-  cond = gte(idx_dec, one)
-  if2 = if(middle, cond)
-  if2_false = read_prod(if2, 0)
-  if2_true = read_prod(if2, 1)
-  r = return(if2_false, later_val)
diff --git a/hercules_samples/matmul/Cargo.toml b/hercules_samples/matmul/Cargo.toml
index 723e9b523c693657a48a5a0213deab99ab875852..744b3f596a1007d0e22f55e2975b0357f70c663e 100644
--- a/hercules_samples/matmul/Cargo.toml
+++ b/hercules_samples/matmul/Cargo.toml
@@ -1,5 +1,5 @@
 [package]
-name = "hercules_matmul"
+name = "matmul"
 version = "0.1.0"
 authors = ["Russel Arbore <rarbore2@illinois.edu>"]
 edition = "2021"
diff --git a/hercules_samples/matmul/src/main.rs b/hercules_samples/matmul/src/main.rs
index 1fcee18cb1648ebfe13b4620caa458fdc1d203c2..ac5e96b26e65aa16adc266a784a9a727ce5d76e0 100644
--- a/hercules_samples/matmul/src/main.rs
+++ b/hercules_samples/matmul/src/main.rs
@@ -7,7 +7,7 @@ extern crate hercules_rt;
 // To compile currently, run from the Hercules project root directory:
 // cargo run --bin hercules_driver hercules_samples/matmul/matmul.hir "Codegen(\"hercules_samples/matmul\",\"matmul\")"
 // Then, you can execute this example with:
-// cargo run --bin hercules_matmul
+// cargo run --bin matmul
 hercules_rt::use_hman!("hercules_samples/matmul/matmul.hman");
 
 fn main() {
diff --git a/hercules_samples/task_parallel/src/main.rs b/hercules_samples/task_parallel/src/main.rs
deleted file mode 100644
index 01480279fa5b9f94e42f6d3b274dad82baeb393d..0000000000000000000000000000000000000000
--- a/hercules_samples/task_parallel/src/main.rs
+++ /dev/null
@@ -1,11 +0,0 @@
-extern crate async_std;
-extern crate clap;
-extern crate hercules_rt;
-
-hercules_rt::use_hbin!("task_parallel.hbin");
-
-fn main() {
-    async_std::task::block_on(async {
-        println!("{}", task_parallel(16).await);
-    });
-}
diff --git a/hercules_samples/task_parallel/task_parallel.hir b/hercules_samples/task_parallel/task_parallel.hir
deleted file mode 100644
index 6386d5ec0dd09132250d00be3affe77a04613394..0000000000000000000000000000000000000000
--- a/hercules_samples/task_parallel/task_parallel.hir
+++ /dev/null
@@ -1,14 +0,0 @@
-fn task_parallel<1>() -> u64
-  f_ctrl1 = fork(start, #0)
-  j_ctrl1 = join(f_ctrl1)
-  zero = constant(u64, 0)
-  x1 = thread_id(f_ctrl1)
-  data1 = reduce(j_ctrl1, zero, sum1)
-  sum1 = add(data1, x1)
-  f_ctrl2 = fork(j_ctrl1, #0)
-  j_ctrl2 = join(f_ctrl2)
-  x2 = thread_id(f_ctrl2)
-  data2 = reduce(j_ctrl2, zero, sum2)
-  sum2 = add(data2, x2)
-  final = add(data1, data2)
-  r = return(j_ctrl2, final)
diff --git a/hercules_tools/hercules_hbin_dump/Cargo.toml b/hercules_tools/hercules_hbin_dump/Cargo.toml
deleted file mode 100644
index 44504cb58c8f008d88632dd26cbd3ff1745b6109..0000000000000000000000000000000000000000
--- a/hercules_tools/hercules_hbin_dump/Cargo.toml
+++ /dev/null
@@ -1,11 +0,0 @@
-[package]
-name = "hercules_hbin_dump"
-version = "0.1.0"
-authors = ["Russel Arbore <rarbore2@illinois.edu>"]
-edition = "2021"
-
-[dependencies]
-clap = { version = "*", features = ["derive"] }
-postcard = { version = "*", features = ["alloc"] }
-serde = { version = "*", features = ["derive"] }
-hercules_ir = { path = "../../hercules_ir" }
\ No newline at end of file
diff --git a/hercules_tools/hercules_hbin_dump/src/main.rs b/hercules_tools/hercules_hbin_dump/src/main.rs
deleted file mode 100644
index 0b258ab4e534b189f86b53d03dc60a040de87be0..0000000000000000000000000000000000000000
--- a/hercules_tools/hercules_hbin_dump/src/main.rs
+++ /dev/null
@@ -1,28 +0,0 @@
-extern crate clap;
-extern crate hercules_ir;
-extern crate postcard;
-
-use std::fs::File;
-use std::io::prelude::*;
-
-use clap::Parser;
-
-#[derive(Parser, Debug)]
-#[command(author, version, about, long_about = None)]
-struct Args {
-    hbin_file: String,
-}
-
-fn main() {
-    let args = Args::parse();
-    if !args.hbin_file.ends_with(".hbin") {
-        eprintln!("WARNING: Running hercules_hbin_dump on a file without a .hbin extension - interpreting as a Hercules binary file.");
-    }
-
-    let mut f = File::open(args.hbin_file).unwrap();
-    let mut buffer = vec![];
-    f.read_to_end(&mut buffer).unwrap();
-    let (manifest, _): (hercules_ir::ModuleManifest, Vec<u8>) =
-        postcard::from_bytes(&buffer).unwrap();
-    println!("{:#?}", manifest);
-}