Skip to content

Commit

Permalink
cleanup
Browse files Browse the repository at this point in the history
  • Loading branch information
romnn committed Mar 9, 2024
1 parent 6b4f3c8 commit 8fc8e64
Show file tree
Hide file tree
Showing 3 changed files with 3 additions and 1,223 deletions.
2 changes: 1 addition & 1 deletion src/interconn.rs
Original file line number Diff line number Diff line change
Expand Up @@ -417,7 +417,7 @@ mod tests {
Ok(())
}

#[skip = "bridged box interconnect implementation can segfault"]
#[ignore = "bridged box interconnect implementation can segfault"]
#[test]
fn test_box_interconnect() -> eyre::Result<()> {
use playground::interconnect::{BoxInterconnect, Interconnect};
Expand Down
113 changes: 0 additions & 113 deletions src/parallel/deterministic.rs
Original file line number Diff line number Diff line change
Expand Up @@ -22,14 +22,6 @@ where
);
eprintln!();

// let cores: Vec<Vec<Arc<_>>> = self
// let cores: Vec<Vec<_>> = self
// .clusters
// .iter()
// // .map(|cluster| cluster.cores.clone())
// .map(|cluster| cluster.cores.iter().collect())
// .collect();

let mut active_clusters = utils::box_slice![false; self.clusters.len()];

let log_every = 10_000;
Expand All @@ -52,40 +44,13 @@ where
last_time = std::time::Instant::now()
}

// if self.reached_limit(cycle) || !self.active() {
if self.reached_limit(cycle) {
break;
}

crate::timeit!("serial::cycle", self.serial_cycle(cycle));

let kernels_completed = self.kernel_manager.all_kernels_completed();
// let kernels_completed = self
// .running_kernels
// // .try_read()
// .iter()
// .filter_map(Option::as_ref)
// .all(|(_, k)| k.no_more_blocks_to_run());

// this works but is not as efficient..
// for cluster_id in 0..self.clusters.len() {
// let cores_completed = self.clusters[cluster_id].num_active_threads() == 0;
// let cluster_active = !(cores_completed && kernels_completed);
// active_clusters[cluster_id] = cluster_active;
//
// if !cluster_active {
// continue;
// }
//
// rayon::scope(|core_scope| {
// for core in self.clusters.get_mut(cluster_id).unwrap().cores.iter_mut()
// {
// core_scope.spawn(move |_| {
// crate::timeit!("core::cycle", core.cycle(cycle));
// });
// }
// });
// }

rayon::scope(|core_scope| {
for cluster in self.clusters.iter_mut() {
Expand All @@ -107,36 +72,6 @@ where
}
});

// run cores in any order
// rayon::scope(|core_scope| {
// let kernels_completed = self
// .running_kernels
// // .try_read()
// .iter()
// .filter_map(Option::as_ref)
// .all(|(_, k)| k.no_more_blocks_to_run());
//
// for (cluster_id, cluster) in self.clusters.iter().enumerate() {
// let cores_completed = cluster.num_active_threads() == 0;
// let cluster_active = !(cores_completed && kernels_completed);
// active_clusters[cluster_id] = cluster_active;
//
// if !cluster_active {
// continue;
// }
//
// // for core in cores[cluster_id].iter() {
// for core in self.clusters[cluster_id].cores.iter_mut() {
// // for core in cores[cluster_id].iter().cloned() {
// // let core: Arc<RwLock<_>> = core;
// core_scope.spawn(move |_| {
// // crate::timeit!("core::cycle", core.write().cycle(cycle));
// crate::timeit!("core::cycle", core.cycle(cycle));
// });
// }
// }
// });

// collect the core packets pushed to the interconn
for (cluster_id, active) in active_clusters.iter().enumerate() {
if !active {
Expand All @@ -148,12 +83,7 @@ where
let mut core_sim_order = cluster.core_sim_order.try_lock();
for core_id in &*core_sim_order {
let core = &cluster.cores[*core_id];
// let mut core = core.try_write();
let mut core = core.try_lock();
// let core = cluster.cores[*core_id].try_read();
// let core = &mut cluster.cores[*core_id];
// let mut port = core.mem_port.lock();
// let mem_port = &mut core.mem_port;
for ic::Packet {
fetch: (dest, fetch, size),
time,
Expand All @@ -176,49 +106,7 @@ where
self.issue_block_to_core_deterministic(cycle);
self.flush_caches(cycle);

// let mut all_threads_complete = true;
// if self.config.flush_l1_cache {
// for cluster in &mut self.clusters {
// if cluster.num_active_threads() == 0 {
// cluster.cache_invalidate();
// } else {
// all_threads_complete = false;
// }
// }
// }
//
// if self.config.flush_l2_cache {
// if !self.config.flush_l1_cache {
// for cluster in &mut self.clusters {
// if cluster.num_active_threads() > 0 {
// all_threads_complete = false;
// break;
// }
// }
// }
//
// if let Some(l2_config) = &self.config.data_cache_l2 {
// if all_threads_complete {
// log::debug!("flushed L2 caches...");
// if l2_config.inner.total_lines() > 0 {
// for (i, mem_sub) in
// self.mem_sub_partitions.iter_mut().enumerate()
// {
// let mut mem_sub = mem_sub.try_lock();
// let num_dirty_lines_flushed = mem_sub.flush_l2();
// log::debug!(
// "dirty lines flushed from L2 {} is {:?}",
// i,
// num_dirty_lines_flushed
// );
// }
// }
// }
// }
// }

cycle += 1;
// self.set_cycle(cycle);

if !self.active() {
finished_kernel = self.kernel_manager.get_finished_kernel();
Expand All @@ -242,7 +130,6 @@ where
})?;

self.stats.no_kernel.sim.cycles = cycle;
// self.stats.lock().no_kernel.sim.cycles = cycle;
if let Some(log_after_cycle) = self.log_after_cycle {
if log_after_cycle >= cycle {
eprintln!("WARNING: log after {log_after_cycle} cycles but simulation ended after {cycle} cycles");
Expand Down
Loading

0 comments on commit 8fc8e64

Please sign in to comment.