| use crate::frozen::Frozen; |
| use crate::fx::{FxHashSet, FxIndexSet}; |
| use rustc_index::bit_set::BitMatrix; |
| use std::fmt::Debug; |
| use std::hash::Hash; |
| use std::mem; |
| use std::ops::Deref; |
| |
| #[cfg(test)] |
| mod tests; |
| |
| #[derive(Clone, Debug)] |
| pub struct TransitiveRelationBuilder<T> { |
| // List of elements. This is used to map from a T to a usize. |
| elements: FxIndexSet<T>, |
| |
| // List of base edges in the graph. Require to compute transitive |
| // closure. |
| edges: FxHashSet<Edge>, |
| } |
| |
| #[derive(Debug)] |
| pub struct TransitiveRelation<T> { |
| // Frozen transitive relation elements and edges. |
| builder: Frozen<TransitiveRelationBuilder<T>>, |
| |
| // Cached transitive closure derived from the edges. |
| closure: Frozen<BitMatrix<usize, usize>>, |
| } |
| |
| impl<T> Deref for TransitiveRelation<T> { |
| type Target = Frozen<TransitiveRelationBuilder<T>>; |
| |
| fn deref(&self) -> &Self::Target { |
| &self.builder |
| } |
| } |
| |
| impl<T: Clone> Clone for TransitiveRelation<T> { |
| fn clone(&self) -> Self { |
| TransitiveRelation { |
| builder: Frozen::freeze(self.builder.deref().clone()), |
| closure: Frozen::freeze(self.closure.deref().clone()), |
| } |
| } |
| } |
| |
| // HACK(eddyb) manual impl avoids `Default` bound on `T`. |
| impl<T: Eq + Hash> Default for TransitiveRelationBuilder<T> { |
| fn default() -> Self { |
| TransitiveRelationBuilder { elements: Default::default(), edges: Default::default() } |
| } |
| } |
| |
| #[derive(Copy, Clone, PartialEq, Eq, PartialOrd, Debug, Hash)] |
| struct Index(usize); |
| |
| #[derive(Clone, PartialEq, Eq, Debug, Hash)] |
| struct Edge { |
| source: Index, |
| target: Index, |
| } |
| |
| impl<T: Eq + Hash + Copy> TransitiveRelationBuilder<T> { |
| pub fn is_empty(&self) -> bool { |
| self.edges.is_empty() |
| } |
| |
| pub fn elements(&self) -> impl Iterator<Item = &T> { |
| self.elements.iter() |
| } |
| |
| fn index(&self, a: T) -> Option<Index> { |
| self.elements.get_index_of(&a).map(Index) |
| } |
| |
| fn add_index(&mut self, a: T) -> Index { |
| let (index, _added) = self.elements.insert_full(a); |
| Index(index) |
| } |
| |
| /// Applies the (partial) function to each edge and returns a new |
| /// relation builder. If `f` returns `None` for any end-point, |
| /// returns `None`. |
| pub fn maybe_map<F, U>(&self, mut f: F) -> Option<TransitiveRelationBuilder<U>> |
| where |
| F: FnMut(T) -> Option<U>, |
| U: Clone + Debug + Eq + Hash + Copy, |
| { |
| let mut result = TransitiveRelationBuilder::default(); |
| for edge in &self.edges { |
| result.add(f(self.elements[edge.source.0])?, f(self.elements[edge.target.0])?); |
| } |
| Some(result) |
| } |
| |
| /// Indicate that `a < b` (where `<` is this relation) |
| pub fn add(&mut self, a: T, b: T) { |
| let a = self.add_index(a); |
| let b = self.add_index(b); |
| let edge = Edge { source: a, target: b }; |
| self.edges.insert(edge); |
| } |
| |
| /// Compute the transitive closure derived from the edges, and converted to |
| /// the final result. After this, all elements will be immutable to maintain |
| /// the correctness of the result. |
| pub fn freeze(self) -> TransitiveRelation<T> { |
| let mut matrix = BitMatrix::new(self.elements.len(), self.elements.len()); |
| let mut changed = true; |
| while changed { |
| changed = false; |
| for edge in &self.edges { |
| // add an edge from S -> T |
| changed |= matrix.insert(edge.source.0, edge.target.0); |
| |
| // add all outgoing edges from T into S |
| changed |= matrix.union_rows(edge.target.0, edge.source.0); |
| } |
| } |
| TransitiveRelation { builder: Frozen::freeze(self), closure: Frozen::freeze(matrix) } |
| } |
| } |
| |
| impl<T: Eq + Hash + Copy> TransitiveRelation<T> { |
| /// Applies the (partial) function to each edge and returns a new |
| /// relation including transitive closures. |
| pub fn maybe_map<F, U>(&self, f: F) -> Option<TransitiveRelation<U>> |
| where |
| F: FnMut(T) -> Option<U>, |
| U: Clone + Debug + Eq + Hash + Copy, |
| { |
| Some(self.builder.maybe_map(f)?.freeze()) |
| } |
| |
| /// Checks whether `a < target` (transitively) |
| pub fn contains(&self, a: T, b: T) -> bool { |
| match (self.index(a), self.index(b)) { |
| (Some(a), Some(b)) => self.with_closure(|closure| closure.contains(a.0, b.0)), |
| (None, _) | (_, None) => false, |
| } |
| } |
| |
| /// Thinking of `x R y` as an edge `x -> y` in a graph, this |
| /// returns all things reachable from `a`. |
| /// |
| /// Really this probably ought to be `impl Iterator<Item = &T>`, but |
| /// I'm too lazy to make that work, and -- given the caching |
| /// strategy -- it'd be a touch tricky anyhow. |
| pub fn reachable_from(&self, a: T) -> Vec<T> { |
| match self.index(a) { |
| Some(a) => { |
| self.with_closure(|closure| closure.iter(a.0).map(|i| self.elements[i]).collect()) |
| } |
| None => vec![], |
| } |
| } |
| |
| /// Picks what I am referring to as the "postdominating" |
| /// upper-bound for `a` and `b`. This is usually the least upper |
| /// bound, but in cases where there is no single least upper |
| /// bound, it is the "mutual immediate postdominator", if you |
| /// imagine a graph where `a < b` means `a -> b`. |
| /// |
| /// This function is needed because region inference currently |
| /// requires that we produce a single "UB", and there is no best |
| /// choice for the LUB. Rather than pick arbitrarily, I pick a |
| /// less good, but predictable choice. This should help ensure |
| /// that region inference yields predictable results (though it |
| /// itself is not fully sufficient). |
| /// |
| /// Examples are probably clearer than any prose I could write |
| /// (there are corresponding tests below, btw). In each case, |
| /// the query is `postdom_upper_bound(a, b)`: |
| /// |
| /// ```text |
| /// // Returns Some(x), which is also LUB. |
| /// a -> a1 -> x |
| /// ^ |
| /// | |
| /// b -> b1 ---+ |
| /// |
| /// // Returns `Some(x)`, which is not LUB (there is none) |
| /// // diagonal edges run left-to-right. |
| /// a -> a1 -> x |
| /// \/ ^ |
| /// /\ | |
| /// b -> b1 ---+ |
| /// |
| /// // Returns `None`. |
| /// a -> a1 |
| /// b -> b1 |
| /// ``` |
| pub fn postdom_upper_bound(&self, a: T, b: T) -> Option<T> { |
| let mubs = self.minimal_upper_bounds(a, b); |
| self.mutual_immediate_postdominator(mubs) |
| } |
| |
| /// Viewing the relation as a graph, computes the "mutual |
| /// immediate postdominator" of a set of points (if one |
| /// exists). See `postdom_upper_bound` for details. |
| pub fn mutual_immediate_postdominator(&self, mut mubs: Vec<T>) -> Option<T> { |
| loop { |
| match mubs.len() { |
| 0 => return None, |
| 1 => return Some(mubs[0]), |
| _ => { |
| let m = mubs.pop().unwrap(); |
| let n = mubs.pop().unwrap(); |
| mubs.extend(self.minimal_upper_bounds(n, m)); |
| } |
| } |
| } |
| } |
| |
| /// Returns the set of bounds `X` such that: |
| /// |
| /// - `a < X` and `b < X` |
| /// - there is no `Y != X` such that `a < Y` and `Y < X` |
| /// - except for the case where `X < a` (i.e., a strongly connected |
| /// component in the graph). In that case, the smallest |
| /// representative of the SCC is returned (as determined by the |
| /// internal indices). |
| /// |
| /// Note that this set can, in principle, have any size. |
| pub fn minimal_upper_bounds(&self, a: T, b: T) -> Vec<T> { |
| let (Some(mut a), Some(mut b)) = (self.index(a), self.index(b)) else { |
| return vec![]; |
| }; |
| |
| // in some cases, there are some arbitrary choices to be made; |
| // it doesn't really matter what we pick, as long as we pick |
| // the same thing consistently when queried, so ensure that |
| // (a, b) are in a consistent relative order |
| if a > b { |
| mem::swap(&mut a, &mut b); |
| } |
| |
| let lub_indices = self.with_closure(|closure| { |
| // Easy case is when either a < b or b < a: |
| if closure.contains(a.0, b.0) { |
| return vec![b.0]; |
| } |
| if closure.contains(b.0, a.0) { |
| return vec![a.0]; |
| } |
| |
| // Otherwise, the tricky part is that there may be some c |
| // where a < c and b < c. In fact, there may be many such |
| // values. So here is what we do: |
| // |
| // 1. Find the vector `[X | a < X && b < X]` of all values |
| // `X` where `a < X` and `b < X`. In terms of the |
| // graph, this means all values reachable from both `a` |
| // and `b`. Note that this vector is also a set, but we |
| // use the term vector because the order matters |
| // to the steps below. |
| // - This vector contains upper bounds, but they are |
| // not minimal upper bounds. So you may have e.g. |
| // `[x, y, tcx, z]` where `x < tcx` and `y < tcx` and |
| // `z < x` and `z < y`: |
| // |
| // z --+---> x ----+----> tcx |
| // | | |
| // | | |
| // +---> y ----+ |
| // |
| // In this case, we really want to return just `[z]`. |
| // The following steps below achieve this by gradually |
| // reducing the list. |
| // 2. Pare down the vector using `pare_down`. This will |
| // remove elements from the vector that can be reached |
| // by an earlier element. |
| // - In the example above, this would convert `[x, y, |
| // tcx, z]` to `[x, y, z]`. Note that `x` and `y` are |
| // still in the vector; this is because while `z < x` |
| // (and `z < y`) holds, `z` comes after them in the |
| // vector. |
| // 3. Reverse the vector and repeat the pare down process. |
| // - In the example above, we would reverse to |
| // `[z, y, x]` and then pare down to `[z]`. |
| // 4. Reverse once more just so that we yield a vector in |
| // increasing order of index. Not necessary, but why not. |
| // |
| // I believe this algorithm yields a minimal set. The |
| // argument is that, after step 2, we know that no element |
| // can reach its successors (in the vector, not the graph). |
| // After step 3, we know that no element can reach any of |
| // its predecessors (because of step 2) nor successors |
| // (because we just called `pare_down`) |
| // |
| // This same algorithm is used in `parents` below. |
| |
| let mut candidates = closure.intersect_rows(a.0, b.0); // (1) |
| pare_down(&mut candidates, closure); // (2) |
| candidates.reverse(); // (3a) |
| pare_down(&mut candidates, closure); // (3b) |
| candidates |
| }); |
| |
| lub_indices |
| .into_iter() |
| .rev() // (4) |
| .map(|i| self.elements[i]) |
| .collect() |
| } |
| |
| /// Given an element A, returns the maximal set {B} of elements B |
| /// such that |
| /// |
| /// - A != B |
| /// - A R B is true |
| /// - for each i, j: `B[i]` R `B[j]` does not hold |
| /// |
| /// The intuition is that this moves "one step up" through a lattice |
| /// (where the relation is encoding the `<=` relation for the lattice). |
| /// So e.g., if the relation is `->` and we have |
| /// |
| /// ```text |
| /// a -> b -> d -> f |
| /// | ^ |
| /// +--> c -> e ---+ |
| /// ``` |
| /// |
| /// then `parents(a)` returns `[b, c]`. The `postdom_parent` function |
| /// would further reduce this to just `f`. |
| pub fn parents(&self, a: T) -> Vec<T> { |
| let Some(a) = self.index(a) else { |
| return vec![]; |
| }; |
| |
| // Steal the algorithm for `minimal_upper_bounds` above, but |
| // with a slight tweak. In the case where `a R a`, we remove |
| // that from the set of candidates. |
| let ancestors = self.with_closure(|closure| { |
| let mut ancestors = closure.intersect_rows(a.0, a.0); |
| |
| // Remove anything that can reach `a`. If this is a |
| // reflexive relation, this will include `a` itself. |
| ancestors.retain(|&e| !closure.contains(e, a.0)); |
| |
| pare_down(&mut ancestors, closure); // (2) |
| ancestors.reverse(); // (3a) |
| pare_down(&mut ancestors, closure); // (3b) |
| ancestors |
| }); |
| |
| ancestors |
| .into_iter() |
| .rev() // (4) |
| .map(|i| self.elements[i]) |
| .collect() |
| } |
| |
| fn with_closure<OP, R>(&self, op: OP) -> R |
| where |
| OP: FnOnce(&BitMatrix<usize, usize>) -> R, |
| { |
| op(&self.closure) |
| } |
| |
| /// Lists all the base edges in the graph: the initial _non-transitive_ set of element |
| /// relations, which will be later used as the basis for the transitive closure computation. |
| pub fn base_edges(&self) -> impl Iterator<Item = (T, T)> + '_ { |
| self.edges |
| .iter() |
| .map(move |edge| (self.elements[edge.source.0], self.elements[edge.target.0])) |
| } |
| } |
| |
| /// Pare down is used as a step in the LUB computation. It edits the |
| /// candidates array in place by removing any element j for which |
| /// there exists an earlier element i<j such that i -> j. That is, |
| /// after you run `pare_down`, you know that for all elements that |
| /// remain in candidates, they cannot reach any of the elements that |
| /// come after them. |
| /// |
| /// Examples follow. Assume that a -> b -> c and x -> y -> z. |
| /// |
| /// - Input: `[a, b, x]`. Output: `[a, x]`. |
| /// - Input: `[b, a, x]`. Output: `[b, a, x]`. |
| /// - Input: `[a, x, b, y]`. Output: `[a, x]`. |
| fn pare_down(candidates: &mut Vec<usize>, closure: &BitMatrix<usize, usize>) { |
| let mut i = 0; |
| while let Some(&candidate_i) = candidates.get(i) { |
| i += 1; |
| |
| let mut j = i; |
| let mut dead = 0; |
| while let Some(&candidate_j) = candidates.get(j) { |
| if closure.contains(candidate_i, candidate_j) { |
| // If `i` can reach `j`, then we can remove `j`. So just |
| // mark it as dead and move on; subsequent indices will be |
| // shifted into its place. |
| dead += 1; |
| } else { |
| candidates[j - dead] = candidate_j; |
| } |
| j += 1; |
| } |
| candidates.truncate(j - dead); |
| } |
| } |