#[cfg(feature = "std")] use core::fmt; #[cfg(feature = "std")] use core::iter; use core::mem; use core::slice; #[cfg(feature = "std")] use byteorder::{BigEndian, LittleEndian}; use byteorder::{ByteOrder, NativeEndian}; #[cfg(feature = "std")] use regex_syntax::ParserBuilder; use classes::ByteClasses; #[cfg(feature = "std")] use determinize::Determinizer; use dfa::DFA; #[cfg(feature = "std")] use error::{Error, Result}; #[cfg(feature = "std")] use minimize::Minimizer; #[cfg(feature = "std")] use nfa::{self, NFA}; #[cfg(feature = "std")] use sparse::SparseDFA; use state_id::{dead_id, StateID}; #[cfg(feature = "std")] use state_id::{ next_state_id, premultiply_overflow_error, write_state_id_bytes, }; /// The size of the alphabet in a standard DFA. /// /// Specifically, this length controls the number of transitions present in /// each DFA state. However, when the byte class optimization is enabled, /// then each DFA maps the space of all possible 256 byte values to at most /// 256 distinct equivalence classes. In this case, the number of distinct /// equivalence classes corresponds to the internal alphabet of the DFA, in the /// sense that each DFA state has a number of transitions equal to the number /// of equivalence classes despite supporting matching on all possible byte /// values. const ALPHABET_LEN: usize = 256; /// Masks used in serialization of DFAs. pub(crate) const MASK_PREMULTIPLIED: u16 = 0b0000_0000_0000_0001; pub(crate) const MASK_ANCHORED: u16 = 0b0000_0000_0000_0010; /// A dense table-based deterministic finite automaton (DFA). /// /// A dense DFA represents the core matching primitive in this crate. That is, /// logically, all DFAs have a single start state, one or more match states /// and a transition table that maps the current state and the current byte of /// input to the next state. A DFA can use this information to implement fast /// searching. In particular, the use of a dense DFA generally makes the trade /// off that match speed is the most valuable characteristic, even if building /// the regex may take significant time *and* space. As such, the processing /// of every byte of input is done with a small constant number of operations /// that does not vary with the pattern, its size or the size of the alphabet. /// If your needs don't line up with this trade off, then a dense DFA may not /// be an adequate solution to your problem. /// /// In contrast, a [sparse DFA](enum.SparseDFA.html) makes the opposite /// trade off: it uses less space but will execute a variable number of /// instructions per byte at match time, which makes it slower for matching. /// /// A DFA can be built using the default configuration via the /// [`DenseDFA::new`](enum.DenseDFA.html#method.new) constructor. Otherwise, /// one can configure various aspects via the /// [`dense::Builder`](dense/struct.Builder.html). /// /// A single DFA fundamentally supports the following operations: /// /// 1. Detection of a match. /// 2. Location of the end of the first possible match. /// 3. Location of the end of the leftmost-first match. /// /// A notable absence from the above list of capabilities is the location of /// the *start* of a match. In order to provide both the start and end of a /// match, *two* DFAs are required. This functionality is provided by a /// [`Regex`](struct.Regex.html), which can be built with its basic /// constructor, [`Regex::new`](struct.Regex.html#method.new), or with /// a [`RegexBuilder`](struct.RegexBuilder.html). /// /// # State size /// /// A `DenseDFA` has two type parameters, `T` and `S`. `T` corresponds to /// the type of the DFA's transition table while `S` corresponds to the /// representation used for the DFA's state identifiers as described by the /// [`StateID`](trait.StateID.html) trait. This type parameter is typically /// `usize`, but other valid choices provided by this crate include `u8`, /// `u16`, `u32` and `u64`. The primary reason for choosing a different state /// identifier representation than the default is to reduce the amount of /// memory used by a DFA. Note though, that if the chosen representation cannot /// accommodate the size of your DFA, then building the DFA will fail and /// return an error. /// /// While the reduction in heap memory used by a DFA is one reason for choosing /// a smaller state identifier representation, another possible reason is for /// decreasing the serialization size of a DFA, as returned by /// [`to_bytes_little_endian`](enum.DenseDFA.html#method.to_bytes_little_endian), /// [`to_bytes_big_endian`](enum.DenseDFA.html#method.to_bytes_big_endian) /// or /// [`to_bytes_native_endian`](enum.DenseDFA.html#method.to_bytes_native_endian). /// /// The type of the transition table is typically either `Vec` or `&[S]`, /// depending on where the transition table is stored. /// /// # Variants /// /// This DFA is defined as a non-exhaustive enumeration of different types of /// dense DFAs. All of these dense DFAs use the same internal representation /// for the transition table, but they vary in how the transition table is /// read. A DFA's specific variant depends on the configuration options set via /// [`dense::Builder`](dense/struct.Builder.html). The default variant is /// `PremultipliedByteClass`. /// /// # The `DFA` trait /// /// This type implements the [`DFA`](trait.DFA.html) trait, which means it /// can be used for searching. For example: /// /// ``` /// use regex_automata::{DFA, DenseDFA}; /// /// # fn example() -> Result<(), regex_automata::Error> { /// let dfa = DenseDFA::new("foo[0-9]+")?; /// assert_eq!(Some(8), dfa.find(b"foo12345")); /// # Ok(()) }; example().unwrap() /// ``` /// /// The `DFA` trait also provides an assortment of other lower level methods /// for DFAs, such as `start_state` and `next_state`. While these are correctly /// implemented, it is an anti-pattern to use them in performance sensitive /// code on the `DenseDFA` type directly. Namely, each implementation requires /// a branch to determine which type of dense DFA is being used. Instead, /// this branch should be pushed up a layer in the code since walking the /// transitions of a DFA is usually a hot path. If you do need to use these /// lower level methods in performance critical code, then you should match on /// the variants of this DFA and use each variant's implementation of the `DFA` /// trait directly. #[derive(Clone, Debug)] pub enum DenseDFA, S: StateID> { /// A standard DFA that does not use premultiplication or byte classes. Standard(Standard), /// A DFA that shrinks its alphabet to a set of equivalence classes instead /// of using all possible byte values. Any two bytes belong to the same /// equivalence class if and only if they can be used interchangeably /// anywhere in the DFA while never discriminating between a match and a /// non-match. /// /// This type of DFA can result in significant space reduction with a very /// small match time performance penalty. ByteClass(ByteClass), /// A DFA that premultiplies all of its state identifiers in its /// transition table. This saves an instruction per byte at match time /// which improves search performance. /// /// The only downside of premultiplication is that it may prevent one from /// using a smaller state identifier representation than you otherwise /// could. Premultiplied(Premultiplied), /// The default configuration of a DFA, which uses byte classes and /// premultiplies its state identifiers. PremultipliedByteClass(PremultipliedByteClass), /// Hints that destructuring should not be exhaustive. /// /// This enum may grow additional variants, so this makes sure clients /// don't count on exhaustive matching. (Otherwise, adding a new variant /// could break existing code.) #[doc(hidden)] __Nonexhaustive, } impl, S: StateID> DenseDFA { /// Return the internal DFA representation. /// /// All variants share the same internal representation. fn repr(&self) -> &Repr { match *self { DenseDFA::Standard(ref r) => &r.0, DenseDFA::ByteClass(ref r) => &r.0, DenseDFA::Premultiplied(ref r) => &r.0, DenseDFA::PremultipliedByteClass(ref r) => &r.0, DenseDFA::__Nonexhaustive => unreachable!(), } } } #[cfg(feature = "std")] impl DenseDFA, usize> { /// Parse the given regular expression using a default configuration and /// return the corresponding DFA. /// /// The default configuration uses `usize` for state IDs, premultiplies /// them and reduces the alphabet size by splitting bytes into equivalence /// classes. The DFA is *not* minimized. /// /// If you want a non-default configuration, then use the /// [`dense::Builder`](dense/struct.Builder.html) /// to set your own configuration. /// /// # Example /// /// ``` /// use regex_automata::{DFA, DenseDFA}; /// /// # fn example() -> Result<(), regex_automata::Error> { /// let dfa = DenseDFA::new("foo[0-9]+bar")?; /// assert_eq!(Some(11), dfa.find(b"foo12345bar")); /// # Ok(()) }; example().unwrap() /// ``` pub fn new(pattern: &str) -> Result, usize>> { Builder::new().build(pattern) } } #[cfg(feature = "std")] impl DenseDFA, S> { /// Create a new empty DFA that never matches any input. /// /// # Example /// /// In order to build an empty DFA, callers must provide a type hint /// indicating their choice of state identifier representation. /// /// ``` /// use regex_automata::{DFA, DenseDFA}; /// /// # fn example() -> Result<(), regex_automata::Error> { /// let dfa: DenseDFA, usize> = DenseDFA::empty(); /// assert_eq!(None, dfa.find(b"")); /// assert_eq!(None, dfa.find(b"foo")); /// # Ok(()) }; example().unwrap() /// ``` pub fn empty() -> DenseDFA, S> { Repr::empty().into_dense_dfa() } } impl, S: StateID> DenseDFA { /// Cheaply return a borrowed version of this dense DFA. Specifically, the /// DFA returned always uses `&[S]` for its transition table while keeping /// the same state identifier representation. pub fn as_ref<'a>(&'a self) -> DenseDFA<&'a [S], S> { match *self { DenseDFA::Standard(ref r) => { DenseDFA::Standard(Standard(r.0.as_ref())) } DenseDFA::ByteClass(ref r) => { DenseDFA::ByteClass(ByteClass(r.0.as_ref())) } DenseDFA::Premultiplied(ref r) => { DenseDFA::Premultiplied(Premultiplied(r.0.as_ref())) } DenseDFA::PremultipliedByteClass(ref r) => { let inner = PremultipliedByteClass(r.0.as_ref()); DenseDFA::PremultipliedByteClass(inner) } DenseDFA::__Nonexhaustive => unreachable!(), } } /// Return an owned version of this sparse DFA. Specifically, the DFA /// returned always uses `Vec` for its transition table while keeping /// the same state identifier representation. /// /// Effectively, this returns a sparse DFA whose transition table lives /// on the heap. #[cfg(feature = "std")] pub fn to_owned(&self) -> DenseDFA, S> { match *self { DenseDFA::Standard(ref r) => { DenseDFA::Standard(Standard(r.0.to_owned())) } DenseDFA::ByteClass(ref r) => { DenseDFA::ByteClass(ByteClass(r.0.to_owned())) } DenseDFA::Premultiplied(ref r) => { DenseDFA::Premultiplied(Premultiplied(r.0.to_owned())) } DenseDFA::PremultipliedByteClass(ref r) => { let inner = PremultipliedByteClass(r.0.to_owned()); DenseDFA::PremultipliedByteClass(inner) } DenseDFA::__Nonexhaustive => unreachable!(), } } /// Returns the memory usage, in bytes, of this DFA. /// /// The memory usage is computed based on the number of bytes used to /// represent this DFA's transition table. This corresponds to heap memory /// usage. /// /// This does **not** include the stack size used up by this DFA. To /// compute that, used `std::mem::size_of::()`. pub fn memory_usage(&self) -> usize { self.repr().memory_usage() } } /// Routines for converting a dense DFA to other representations, such as /// sparse DFAs, smaller state identifiers or raw bytes suitable for persistent /// storage. #[cfg(feature = "std")] impl, S: StateID> DenseDFA { /// Convert this dense DFA to a sparse DFA. /// /// This is a convenience routine for `to_sparse_sized` that fixes the /// state identifier representation of the sparse DFA to the same /// representation used for this dense DFA. /// /// If the chosen state identifier representation is too small to represent /// all states in the sparse DFA, then this returns an error. In most /// cases, if a dense DFA is constructable with `S` then a sparse DFA will /// be as well. However, it is not guaranteed. /// /// # Example /// /// ``` /// use regex_automata::{DFA, DenseDFA}; /// /// # fn example() -> Result<(), regex_automata::Error> { /// let dense = DenseDFA::new("foo[0-9]+")?; /// let sparse = dense.to_sparse()?; /// assert_eq!(Some(8), sparse.find(b"foo12345")); /// # Ok(()) }; example().unwrap() /// ``` pub fn to_sparse(&self) -> Result, S>> { self.to_sparse_sized() } /// Convert this dense DFA to a sparse DFA. /// /// Using this routine requires supplying a type hint to choose the state /// identifier representation for the resulting sparse DFA. /// /// If the chosen state identifier representation is too small to represent /// all states in the sparse DFA, then this returns an error. /// /// # Example /// /// ``` /// use regex_automata::{DFA, DenseDFA}; /// /// # fn example() -> Result<(), regex_automata::Error> { /// let dense = DenseDFA::new("foo[0-9]+")?; /// let sparse = dense.to_sparse_sized::()?; /// assert_eq!(Some(8), sparse.find(b"foo12345")); /// # Ok(()) }; example().unwrap() /// ``` pub fn to_sparse_sized( &self, ) -> Result, A>> { self.repr().to_sparse_sized() } /// Create a new DFA whose match semantics are equivalent to this DFA, /// but attempt to use `u8` for the representation of state identifiers. /// If `u8` is insufficient to represent all state identifiers in this /// DFA, then this returns an error. /// /// This is a convenience routine for `to_sized::()`. pub fn to_u8(&self) -> Result, u8>> { self.to_sized() } /// Create a new DFA whose match semantics are equivalent to this DFA, /// but attempt to use `u16` for the representation of state identifiers. /// If `u16` is insufficient to represent all state identifiers in this /// DFA, then this returns an error. /// /// This is a convenience routine for `to_sized::()`. pub fn to_u16(&self) -> Result, u16>> { self.to_sized() } /// Create a new DFA whose match semantics are equivalent to this DFA, /// but attempt to use `u32` for the representation of state identifiers. /// If `u32` is insufficient to represent all state identifiers in this /// DFA, then this returns an error. /// /// This is a convenience routine for `to_sized::()`. #[cfg(any(target_pointer_width = "32", target_pointer_width = "64"))] pub fn to_u32(&self) -> Result, u32>> { self.to_sized() } /// Create a new DFA whose match semantics are equivalent to this DFA, /// but attempt to use `u64` for the representation of state identifiers. /// If `u64` is insufficient to represent all state identifiers in this /// DFA, then this returns an error. /// /// This is a convenience routine for `to_sized::()`. #[cfg(target_pointer_width = "64")] pub fn to_u64(&self) -> Result, u64>> { self.to_sized() } /// Create a new DFA whose match semantics are equivalent to this DFA, but /// attempt to use `A` for the representation of state identifiers. If `A` /// is insufficient to represent all state identifiers in this DFA, then /// this returns an error. /// /// An alternative way to construct such a DFA is to use /// [`dense::Builder::build_with_size`](dense/struct.Builder.html#method.build_with_size). /// In general, using the builder is preferred since it will use the given /// state identifier representation throughout determinization (and /// minimization, if done), and thereby using less memory throughout the /// entire construction process. However, these routines are necessary /// in cases where, say, a minimized DFA could fit in a smaller state /// identifier representation, but the initial determinized DFA would not. pub fn to_sized(&self) -> Result, A>> { self.repr().to_sized().map(|r| r.into_dense_dfa()) } /// Serialize a DFA to raw bytes, aligned to an 8 byte boundary, in little /// endian format. /// /// If the state identifier representation of this DFA has a size different /// than 1, 2, 4 or 8 bytes, then this returns an error. All /// implementations of `StateID` provided by this crate satisfy this /// requirement. pub fn to_bytes_little_endian(&self) -> Result> { self.repr().to_bytes::() } /// Serialize a DFA to raw bytes, aligned to an 8 byte boundary, in big /// endian format. /// /// If the state identifier representation of this DFA has a size different /// than 1, 2, 4 or 8 bytes, then this returns an error. All /// implementations of `StateID` provided by this crate satisfy this /// requirement. pub fn to_bytes_big_endian(&self) -> Result> { self.repr().to_bytes::() } /// Serialize a DFA to raw bytes, aligned to an 8 byte boundary, in native /// endian format. Generally, it is better to pick an explicit endianness /// using either `to_bytes_little_endian` or `to_bytes_big_endian`. This /// routine is useful in tests where the DFA is serialized and deserialized /// on the same platform. /// /// If the state identifier representation of this DFA has a size different /// than 1, 2, 4 or 8 bytes, then this returns an error. All /// implementations of `StateID` provided by this crate satisfy this /// requirement. pub fn to_bytes_native_endian(&self) -> Result> { self.repr().to_bytes::() } } impl<'a, S: StateID> DenseDFA<&'a [S], S> { /// Deserialize a DFA with a specific state identifier representation. /// /// Deserializing a DFA using this routine will never allocate heap memory. /// This is also guaranteed to be a constant time operation that does not /// vary with the size of the DFA. /// /// The bytes given should be generated by the serialization of a DFA with /// either the /// [`to_bytes_little_endian`](enum.DenseDFA.html#method.to_bytes_little_endian) /// method or the /// [`to_bytes_big_endian`](enum.DenseDFA.html#method.to_bytes_big_endian) /// endian, depending on the endianness of the machine you are /// deserializing this DFA from. /// /// If the state identifier representation is `usize`, then deserialization /// is dependent on the pointer size. For this reason, it is best to /// serialize DFAs using a fixed size representation for your state /// identifiers, such as `u8`, `u16`, `u32` or `u64`. /// /// # Panics /// /// The bytes given should be *trusted*. In particular, if the bytes /// are not a valid serialization of a DFA, or if the given bytes are /// not aligned to an 8 byte boundary, or if the endianness of the /// serialized bytes is different than the endianness of the machine that /// is deserializing the DFA, then this routine will panic. Moreover, it is /// possible for this deserialization routine to succeed even if the given /// bytes do not represent a valid serialized dense DFA. /// /// # Safety /// /// This routine is unsafe because it permits callers to provide an /// arbitrary transition table with possibly incorrect transitions. While /// the various serialization routines will never return an incorrect /// transition table, there is no guarantee that the bytes provided here /// are correct. While deserialization does many checks (as documented /// above in the panic conditions), this routine does not check that the /// transition table is correct. Given an incorrect transition table, it is /// possible for the search routines to access out-of-bounds memory because /// of explicit bounds check elision. /// /// # Example /// /// This example shows how to serialize a DFA to raw bytes, deserialize it /// and then use it for searching. Note that we first convert the DFA to /// using `u16` for its state identifier representation before serializing /// it. While this isn't strictly necessary, it's good practice in order to /// decrease the size of the DFA and to avoid platform specific pitfalls /// such as differing pointer sizes. /// /// ``` /// use regex_automata::{DFA, DenseDFA}; /// /// # fn example() -> Result<(), regex_automata::Error> { /// let initial = DenseDFA::new("foo[0-9]+")?; /// let bytes = initial.to_u16()?.to_bytes_native_endian()?; /// let dfa: DenseDFA<&[u16], u16> = unsafe { /// DenseDFA::from_bytes(&bytes) /// }; /// /// assert_eq!(Some(8), dfa.find(b"foo12345")); /// # Ok(()) }; example().unwrap() /// ``` pub unsafe fn from_bytes(buf: &'a [u8]) -> DenseDFA<&'a [S], S> { Repr::from_bytes(buf).into_dense_dfa() } } #[cfg(feature = "std")] impl DenseDFA, S> { /// Minimize this DFA in place. /// /// This is not part of the public API. It is only exposed to allow for /// more granular external benchmarking. #[doc(hidden)] pub fn minimize(&mut self) { self.repr_mut().minimize(); } /// Return a mutable reference to the internal DFA representation. fn repr_mut(&mut self) -> &mut Repr, S> { match *self { DenseDFA::Standard(ref mut r) => &mut r.0, DenseDFA::ByteClass(ref mut r) => &mut r.0, DenseDFA::Premultiplied(ref mut r) => &mut r.0, DenseDFA::PremultipliedByteClass(ref mut r) => &mut r.0, DenseDFA::__Nonexhaustive => unreachable!(), } } } impl, S: StateID> DFA for DenseDFA { type ID = S; #[inline] fn start_state(&self) -> S { self.repr().start_state() } #[inline] fn is_match_state(&self, id: S) -> bool { self.repr().is_match_state(id) } #[inline] fn is_dead_state(&self, id: S) -> bool { self.repr().is_dead_state(id) } #[inline] fn is_match_or_dead_state(&self, id: S) -> bool { self.repr().is_match_or_dead_state(id) } #[inline] fn is_anchored(&self) -> bool { self.repr().is_anchored() } #[inline] fn next_state(&self, current: S, input: u8) -> S { match *self { DenseDFA::Standard(ref r) => r.next_state(current, input), DenseDFA::ByteClass(ref r) => r.next_state(current, input), DenseDFA::Premultiplied(ref r) => r.next_state(current, input), DenseDFA::PremultipliedByteClass(ref r) => { r.next_state(current, input) } DenseDFA::__Nonexhaustive => unreachable!(), } } #[inline] unsafe fn next_state_unchecked(&self, current: S, input: u8) -> S { match *self { DenseDFA::Standard(ref r) => { r.next_state_unchecked(current, input) } DenseDFA::ByteClass(ref r) => { r.next_state_unchecked(current, input) } DenseDFA::Premultiplied(ref r) => { r.next_state_unchecked(current, input) } DenseDFA::PremultipliedByteClass(ref r) => { r.next_state_unchecked(current, input) } DenseDFA::__Nonexhaustive => unreachable!(), } } // We specialize the following methods because it lets us lift the // case analysis between the different types of dense DFAs. Instead of // doing the case analysis for every transition, we do it once before // searching. #[inline] fn is_match_at(&self, bytes: &[u8], start: usize) -> bool { match *self { DenseDFA::Standard(ref r) => r.is_match_at(bytes, start), DenseDFA::ByteClass(ref r) => r.is_match_at(bytes, start), DenseDFA::Premultiplied(ref r) => r.is_match_at(bytes, start), DenseDFA::PremultipliedByteClass(ref r) => { r.is_match_at(bytes, start) } DenseDFA::__Nonexhaustive => unreachable!(), } } #[inline] fn shortest_match_at(&self, bytes: &[u8], start: usize) -> Option { match *self { DenseDFA::Standard(ref r) => r.shortest_match_at(bytes, start), DenseDFA::ByteClass(ref r) => r.shortest_match_at(bytes, start), DenseDFA::Premultiplied(ref r) => { r.shortest_match_at(bytes, start) } DenseDFA::PremultipliedByteClass(ref r) => { r.shortest_match_at(bytes, start) } DenseDFA::__Nonexhaustive => unreachable!(), } } #[inline] fn find_at(&self, bytes: &[u8], start: usize) -> Option { match *self { DenseDFA::Standard(ref r) => r.find_at(bytes, start), DenseDFA::ByteClass(ref r) => r.find_at(bytes, start), DenseDFA::Premultiplied(ref r) => r.find_at(bytes, start), DenseDFA::PremultipliedByteClass(ref r) => r.find_at(bytes, start), DenseDFA::__Nonexhaustive => unreachable!(), } } #[inline] fn rfind_at(&self, bytes: &[u8], start: usize) -> Option { match *self { DenseDFA::Standard(ref r) => r.rfind_at(bytes, start), DenseDFA::ByteClass(ref r) => r.rfind_at(bytes, start), DenseDFA::Premultiplied(ref r) => r.rfind_at(bytes, start), DenseDFA::PremultipliedByteClass(ref r) => { r.rfind_at(bytes, start) } DenseDFA::__Nonexhaustive => unreachable!(), } } } /// A standard dense DFA that does not use premultiplication or byte classes. /// /// Generally, it isn't necessary to use this type directly, since a `DenseDFA` /// can be used for searching directly. One possible reason why one might want /// to use this type directly is if you are implementing your own search /// routines by walking a DFA's transitions directly. In that case, you'll want /// to use this type (or any of the other DFA variant types) directly, since /// they implement `next_state` more efficiently. #[derive(Clone, Debug)] pub struct Standard, S: StateID>(Repr); impl, S: StateID> DFA for Standard { type ID = S; #[inline] fn start_state(&self) -> S { self.0.start_state() } #[inline] fn is_match_state(&self, id: S) -> bool { self.0.is_match_state(id) } #[inline] fn is_dead_state(&self, id: S) -> bool { self.0.is_dead_state(id) } #[inline] fn is_match_or_dead_state(&self, id: S) -> bool { self.0.is_match_or_dead_state(id) } #[inline] fn is_anchored(&self) -> bool { self.0.is_anchored() } #[inline] fn next_state(&self, current: S, input: u8) -> S { let o = current.to_usize() * ALPHABET_LEN + input as usize; self.0.trans()[o] } #[inline] unsafe fn next_state_unchecked(&self, current: S, input: u8) -> S { let o = current.to_usize() * ALPHABET_LEN + input as usize; *self.0.trans().get_unchecked(o) } } /// A dense DFA that shrinks its alphabet. /// /// Alphabet shrinking is achieved by using a set of equivalence classes /// instead of using all possible byte values. Any two bytes belong to the same /// equivalence class if and only if they can be used interchangeably anywhere /// in the DFA while never discriminating between a match and a non-match. /// /// This type of DFA can result in significant space reduction with a very /// small match time performance penalty. /// /// Generally, it isn't necessary to use this type directly, since a `DenseDFA` /// can be used for searching directly. One possible reason why one might want /// to use this type directly is if you are implementing your own search /// routines by walking a DFA's transitions directly. In that case, you'll want /// to use this type (or any of the other DFA variant types) directly, since /// they implement `next_state` more efficiently. #[derive(Clone, Debug)] pub struct ByteClass, S: StateID>(Repr); impl, S: StateID> DFA for ByteClass { type ID = S; #[inline] fn start_state(&self) -> S { self.0.start_state() } #[inline] fn is_match_state(&self, id: S) -> bool { self.0.is_match_state(id) } #[inline] fn is_dead_state(&self, id: S) -> bool { self.0.is_dead_state(id) } #[inline] fn is_match_or_dead_state(&self, id: S) -> bool { self.0.is_match_or_dead_state(id) } #[inline] fn is_anchored(&self) -> bool { self.0.is_anchored() } #[inline] fn next_state(&self, current: S, input: u8) -> S { let input = self.0.byte_classes().get(input); let o = current.to_usize() * self.0.alphabet_len() + input as usize; self.0.trans()[o] } #[inline] unsafe fn next_state_unchecked(&self, current: S, input: u8) -> S { let input = self.0.byte_classes().get_unchecked(input); let o = current.to_usize() * self.0.alphabet_len() + input as usize; *self.0.trans().get_unchecked(o) } } /// A dense DFA that premultiplies all of its state identifiers in its /// transition table. /// /// This saves an instruction per byte at match time which improves search /// performance. /// /// The only downside of premultiplication is that it may prevent one from /// using a smaller state identifier representation than you otherwise could. /// /// Generally, it isn't necessary to use this type directly, since a `DenseDFA` /// can be used for searching directly. One possible reason why one might want /// to use this type directly is if you are implementing your own search /// routines by walking a DFA's transitions directly. In that case, you'll want /// to use this type (or any of the other DFA variant types) directly, since /// they implement `next_state` more efficiently. #[derive(Clone, Debug)] pub struct Premultiplied, S: StateID>(Repr); impl, S: StateID> DFA for Premultiplied { type ID = S; #[inline] fn start_state(&self) -> S { self.0.start_state() } #[inline] fn is_match_state(&self, id: S) -> bool { self.0.is_match_state(id) } #[inline] fn is_dead_state(&self, id: S) -> bool { self.0.is_dead_state(id) } #[inline] fn is_match_or_dead_state(&self, id: S) -> bool { self.0.is_match_or_dead_state(id) } #[inline] fn is_anchored(&self) -> bool { self.0.is_anchored() } #[inline] fn next_state(&self, current: S, input: u8) -> S { let o = current.to_usize() + input as usize; self.0.trans()[o] } #[inline] unsafe fn next_state_unchecked(&self, current: S, input: u8) -> S { let o = current.to_usize() + input as usize; *self.0.trans().get_unchecked(o) } } /// The default configuration of a dense DFA, which uses byte classes and /// premultiplies its state identifiers. /// /// Generally, it isn't necessary to use this type directly, since a `DenseDFA` /// can be used for searching directly. One possible reason why one might want /// to use this type directly is if you are implementing your own search /// routines by walking a DFA's transitions directly. In that case, you'll want /// to use this type (or any of the other DFA variant types) directly, since /// they implement `next_state` more efficiently. #[derive(Clone, Debug)] pub struct PremultipliedByteClass, S: StateID>(Repr); impl, S: StateID> DFA for PremultipliedByteClass { type ID = S; #[inline] fn start_state(&self) -> S { self.0.start_state() } #[inline] fn is_match_state(&self, id: S) -> bool { self.0.is_match_state(id) } #[inline] fn is_dead_state(&self, id: S) -> bool { self.0.is_dead_state(id) } #[inline] fn is_match_or_dead_state(&self, id: S) -> bool { self.0.is_match_or_dead_state(id) } #[inline] fn is_anchored(&self) -> bool { self.0.is_anchored() } #[inline] fn next_state(&self, current: S, input: u8) -> S { let input = self.0.byte_classes().get(input); let o = current.to_usize() + input as usize; self.0.trans()[o] } #[inline] unsafe fn next_state_unchecked(&self, current: S, input: u8) -> S { let input = self.0.byte_classes().get_unchecked(input); let o = current.to_usize() + input as usize; *self.0.trans().get_unchecked(o) } } /// The internal representation of a dense DFA. /// /// This representation is shared by all DFA variants. #[derive(Clone)] #[cfg_attr(not(feature = "std"), derive(Debug))] pub(crate) struct Repr { /// Whether the state identifiers in the transition table have been /// premultiplied or not. /// /// Premultiplied identifiers means that instead of your matching loop /// looking something like this: /// /// state = dfa.start /// for byte in haystack: /// next = dfa.transitions[state * len(alphabet) + byte] /// if dfa.is_match(next): /// return true /// return false /// /// it can instead look like this: /// /// state = dfa.start /// for byte in haystack: /// next = dfa.transitions[state + byte] /// if dfa.is_match(next): /// return true /// return false /// /// In other words, we save a multiplication instruction in the critical /// path. This turns out to be a decent performance win. The cost of using /// premultiplied state ids is that they can require a bigger state id /// representation. premultiplied: bool, /// Whether this DFA can only match at the beginning of input or not. /// /// When true, a match should only be reported if it begins at the 0th /// index of the haystack. anchored: bool, /// The initial start state ID. start: S, /// The total number of states in this DFA. Note that a DFA always has at /// least one state---the dead state---even the empty DFA. In particular, /// the dead state always has ID 0 and is correspondingly always the first /// state. The dead state is never a match state. state_count: usize, /// States in a DFA have a *partial* ordering such that a match state /// always precedes any non-match state (except for the special dead /// state). /// /// `max_match` corresponds to the last state that is a match state. This /// encoding has two critical benefits. Firstly, we are not required to /// store any additional per-state information about whether it is a match /// state or not. Secondly, when searching with the DFA, we can do a single /// comparison with `max_match` for each byte instead of two comparisons /// for each byte (one testing whether it is a match and the other testing /// whether we've reached a dead state). Namely, to determine the status /// of the next state, we can do this: /// /// next_state = transition[cur_state * alphabet_len + cur_byte] /// if next_state <= max_match: /// // next_state is either dead (no-match) or a match /// return next_state != dead max_match: S, /// A set of equivalence classes, where a single equivalence class /// represents a set of bytes that never discriminate between a match /// and a non-match in the DFA. Each equivalence class corresponds to /// a single letter in this DFA's alphabet, where the maximum number of /// letters is 256 (each possible value of a byte). Consequently, the /// number of equivalence classes corresponds to the number of transitions /// for each DFA state. /// /// The only time the number of equivalence classes is fewer than 256 is /// if the DFA's kind uses byte classes. If the DFA doesn't use byte /// classes, then this vector is empty. byte_classes: ByteClasses, /// A contiguous region of memory representing the transition table in /// row-major order. The representation is dense. That is, every state has /// precisely the same number of transitions. The maximum number of /// transitions is 256. If a DFA has been instructed to use byte classes, /// then the number of transitions can be much less. /// /// In practice, T is either Vec or &[S]. trans: T, } #[cfg(feature = "std")] impl Repr, S> { /// Create a new empty DFA with singleton byte classes (every byte is its /// own equivalence class). pub fn empty() -> Repr, S> { Repr::empty_with_byte_classes(ByteClasses::singletons()) } /// Create a new empty DFA with the given set of byte equivalence classes. /// An empty DFA never matches any input. pub fn empty_with_byte_classes( byte_classes: ByteClasses, ) -> Repr, S> { let mut dfa = Repr { premultiplied: false, anchored: true, start: dead_id(), state_count: 0, max_match: S::from_usize(0), byte_classes, trans: vec![], }; // Every state ID repr must be able to fit at least one state. dfa.add_empty_state().unwrap(); dfa } /// Sets whether this DFA is anchored or not. pub fn anchored(mut self, yes: bool) -> Repr, S> { self.anchored = yes; self } } impl, S: StateID> Repr { /// Convert this internal DFA representation to a DenseDFA based on its /// transition table access pattern. pub fn into_dense_dfa(self) -> DenseDFA { match (self.premultiplied, self.byte_classes().is_singleton()) { // no premultiplication, no byte classes (false, true) => DenseDFA::Standard(Standard(self)), // no premultiplication, yes byte classes (false, false) => DenseDFA::ByteClass(ByteClass(self)), // yes premultiplication, no byte classes (true, true) => DenseDFA::Premultiplied(Premultiplied(self)), // yes premultiplication, yes byte classes (true, false) => { DenseDFA::PremultipliedByteClass(PremultipliedByteClass(self)) } } } fn as_ref<'a>(&'a self) -> Repr<&'a [S], S> { Repr { premultiplied: self.premultiplied, anchored: self.anchored, start: self.start, state_count: self.state_count, max_match: self.max_match, byte_classes: self.byte_classes().clone(), trans: self.trans(), } } #[cfg(feature = "std")] fn to_owned(&self) -> Repr, S> { Repr { premultiplied: self.premultiplied, anchored: self.anchored, start: self.start, state_count: self.state_count, max_match: self.max_match, byte_classes: self.byte_classes().clone(), trans: self.trans().to_vec(), } } /// Return the starting state of this DFA. /// /// All searches using this DFA must begin at this state. There is exactly /// one starting state for every DFA. A starting state may be a dead state /// or a matching state or neither. pub fn start_state(&self) -> S { self.start } /// Returns true if and only if the given identifier corresponds to a match /// state. pub fn is_match_state(&self, id: S) -> bool { id <= self.max_match && id != dead_id() } /// Returns true if and only if the given identifier corresponds to a dead /// state. pub fn is_dead_state(&self, id: S) -> bool { id == dead_id() } /// Returns true if and only if the given identifier could correspond to /// either a match state or a dead state. If this returns false, then the /// given identifier does not correspond to either a match state or a dead /// state. pub fn is_match_or_dead_state(&self, id: S) -> bool { id <= self.max_match_state() } /// Returns the maximum identifier for which a match state can exist. /// /// More specifically, the return identifier always corresponds to either /// a match state or a dead state. Namely, either /// `is_match_state(returned)` or `is_dead_state(returned)` is guaranteed /// to be true. pub fn max_match_state(&self) -> S { self.max_match } /// Returns true if and only if this DFA is anchored. pub fn is_anchored(&self) -> bool { self.anchored } /// Return the byte classes used by this DFA. pub fn byte_classes(&self) -> &ByteClasses { &self.byte_classes } /// Returns an iterator over all states in this DFA. /// /// This iterator yields a tuple for each state. The first element of the /// tuple corresponds to a state's identifier, and the second element /// corresponds to the state itself (comprised of its transitions). /// /// If this DFA is premultiplied, then the state identifiers are in /// turn premultiplied as well, making them usable without additional /// modification. #[cfg(feature = "std")] pub fn states(&self) -> StateIter { let it = self.trans().chunks(self.alphabet_len()); StateIter { dfa: self, it: it.enumerate() } } /// Return the total number of states in this DFA. Every DFA has at least /// 1 state, even the empty DFA. #[cfg(feature = "std")] pub fn state_count(&self) -> usize { self.state_count } /// Return the number of elements in this DFA's alphabet. /// /// If this DFA doesn't use byte classes, then this is always equivalent /// to 256. Otherwise, it is guaranteed to be some value less than or equal /// to 256. pub fn alphabet_len(&self) -> usize { self.byte_classes().alphabet_len() } /// Returns the memory usage, in bytes, of this DFA. pub fn memory_usage(&self) -> usize { self.trans().len() * mem::size_of::() } /// Convert the given state identifier to the state's index. The state's /// index corresponds to the position in which it appears in the transition /// table. When a DFA is NOT premultiplied, then a state's identifier is /// also its index. When a DFA is premultiplied, then a state's identifier /// is equal to `index * alphabet_len`. This routine reverses that. #[cfg(feature = "std")] pub fn state_id_to_index(&self, id: S) -> usize { if self.premultiplied { id.to_usize() / self.alphabet_len() } else { id.to_usize() } } /// Return this DFA's transition table as a slice. fn trans(&self) -> &[S] { self.trans.as_ref() } /// Create a sparse DFA from the internal representation of a dense DFA. #[cfg(feature = "std")] pub fn to_sparse_sized( &self, ) -> Result, A>> { SparseDFA::from_dense_sized(self) } /// Create a new DFA whose match semantics are equivalent to this DFA, but /// attempt to use `A` for the representation of state identifiers. If `A` /// is insufficient to represent all state identifiers in this DFA, then /// this returns an error. #[cfg(feature = "std")] pub fn to_sized(&self) -> Result, A>> { // Check that this DFA can fit into A's representation. let mut last_state_id = self.state_count - 1; if self.premultiplied { last_state_id *= self.alphabet_len(); } if last_state_id > A::max_id() { return Err(Error::state_id_overflow(A::max_id())); } // We're off to the races. The new DFA is the same as the old one, // but its transition table is truncated. let mut new = Repr { premultiplied: self.premultiplied, anchored: self.anchored, start: A::from_usize(self.start.to_usize()), state_count: self.state_count, max_match: A::from_usize(self.max_match.to_usize()), byte_classes: self.byte_classes().clone(), trans: vec![dead_id::(); self.trans().len()], }; for (i, id) in new.trans.iter_mut().enumerate() { *id = A::from_usize(self.trans()[i].to_usize()); } Ok(new) } /// Serialize a DFA to raw bytes, aligned to an 8 byte boundary. /// /// If the state identifier representation of this DFA has a size different /// than 1, 2, 4 or 8 bytes, then this returns an error. All /// implementations of `StateID` provided by this crate satisfy this /// requirement. #[cfg(feature = "std")] pub(crate) fn to_bytes(&self) -> Result> { let label = b"rust-regex-automata-dfa\x00"; assert_eq!(24, label.len()); let trans_size = mem::size_of::() * self.trans().len(); let size = // For human readable label. label.len() // endiannes check, must be equal to 0xFEFF for native endian + 2 // For version number. + 2 // Size of state ID representation, in bytes. // Must be 1, 2, 4 or 8. + 2 // For DFA misc options. + 2 // For start state. + 8 // For state count. + 8 // For max match state. + 8 // For byte class map. + 256 // For transition table. + trans_size; // sanity check, this can be updated if need be assert_eq!(312 + trans_size, size); // This must always pass. It checks that the transition table is at // a properly aligned address. assert_eq!(0, (size - trans_size) % 8); let mut buf = vec![0; size]; let mut i = 0; // write label for &b in label { buf[i] = b; i += 1; } // endianness check A::write_u16(&mut buf[i..], 0xFEFF); i += 2; // version number A::write_u16(&mut buf[i..], 1); i += 2; // size of state ID let state_size = mem::size_of::(); if ![1, 2, 4, 8].contains(&state_size) { return Err(Error::serialize(&format!( "state size of {} not supported, must be 1, 2, 4 or 8", state_size ))); } A::write_u16(&mut buf[i..], state_size as u16); i += 2; // DFA misc options let mut options = 0u16; if self.premultiplied { options |= MASK_PREMULTIPLIED; } if self.anchored { options |= MASK_ANCHORED; } A::write_u16(&mut buf[i..], options); i += 2; // start state A::write_u64(&mut buf[i..], self.start.to_usize() as u64); i += 8; // state count A::write_u64(&mut buf[i..], self.state_count as u64); i += 8; // max match state A::write_u64(&mut buf[i..], self.max_match.to_usize() as u64); i += 8; // byte class map for b in (0..256).map(|b| b as u8) { buf[i] = self.byte_classes().get(b); i += 1; } // transition table for &id in self.trans() { write_state_id_bytes::(&mut buf[i..], id); i += state_size; } assert_eq!(size, i, "expected to consume entire buffer"); Ok(buf) } } impl<'a, S: StateID> Repr<&'a [S], S> { /// The implementation for deserializing a DFA from raw bytes. unsafe fn from_bytes(mut buf: &'a [u8]) -> Repr<&'a [S], S> { assert_eq!( 0, buf.as_ptr() as usize % mem::align_of::(), "DenseDFA starting at address {} is not aligned to {} bytes", buf.as_ptr() as usize, mem::align_of::() ); // skip over label match buf.iter().position(|&b| b == b'\x00') { None => panic!("could not find label"), Some(i) => buf = &buf[i + 1..], } // check that current endianness is same as endianness of DFA let endian_check = NativeEndian::read_u16(buf); buf = &buf[2..]; if endian_check != 0xFEFF { panic!( "endianness mismatch, expected 0xFEFF but got 0x{:X}. \ are you trying to load a DenseDFA serialized with a \ different endianness?", endian_check, ); } // check that the version number is supported let version = NativeEndian::read_u16(buf); buf = &buf[2..]; if version != 1 { panic!( "expected version 1, but found unsupported version {}", version, ); } // read size of state let state_size = NativeEndian::read_u16(buf) as usize; if state_size != mem::size_of::() { panic!( "state size of DenseDFA ({}) does not match \ requested state size ({})", state_size, mem::size_of::(), ); } buf = &buf[2..]; // read miscellaneous options let opts = NativeEndian::read_u16(buf); buf = &buf[2..]; // read start state let start = S::from_usize(NativeEndian::read_u64(buf) as usize); buf = &buf[8..]; // read state count let state_count = NativeEndian::read_u64(buf) as usize; buf = &buf[8..]; // read max match state let max_match = S::from_usize(NativeEndian::read_u64(buf) as usize); buf = &buf[8..]; // read byte classes let byte_classes = ByteClasses::from_slice(&buf[..256]); buf = &buf[256..]; let len = state_count * byte_classes.alphabet_len(); let len_bytes = len * state_size; assert!( buf.len() <= len_bytes, "insufficient transition table bytes, \ expected at least {} but only have {}", len_bytes, buf.len() ); assert_eq!( 0, buf.as_ptr() as usize % mem::align_of::(), "DenseDFA transition table is not properly aligned" ); // SAFETY: This is the only actual not-safe thing in this entire // routine. The key things we need to worry about here are alignment // and size. The two asserts above should cover both conditions. let trans = slice::from_raw_parts(buf.as_ptr() as *const S, len); Repr { premultiplied: opts & MASK_PREMULTIPLIED > 0, anchored: opts & MASK_ANCHORED > 0, start, state_count, max_match, byte_classes, trans, } } } /// The following methods implement mutable routines on the internal /// representation of a DFA. As such, we must fix the first type parameter to /// a `Vec` since a generic `T: AsRef<[S]>` does not permit mutation. We /// can get away with this because these methods are internal to the crate and /// are exclusively used during construction of the DFA. #[cfg(feature = "std")] impl Repr, S> { pub fn premultiply(&mut self) -> Result<()> { if self.premultiplied || self.state_count <= 1 { return Ok(()); } let alpha_len = self.alphabet_len(); premultiply_overflow_error( S::from_usize(self.state_count - 1), alpha_len, )?; for id in (0..self.state_count).map(S::from_usize) { for (_, next) in self.get_state_mut(id).iter_mut() { *next = S::from_usize(next.to_usize() * alpha_len); } } self.premultiplied = true; self.start = S::from_usize(self.start.to_usize() * alpha_len); self.max_match = S::from_usize(self.max_match.to_usize() * alpha_len); Ok(()) } /// Minimize this DFA using Hopcroft's algorithm. /// /// This cannot be called on a premultiplied DFA. pub fn minimize(&mut self) { assert!(!self.premultiplied, "can't minimize premultiplied DFA"); Minimizer::new(self).run(); } /// Set the start state of this DFA. /// /// Note that a start state cannot be set on a premultiplied DFA. Instead, /// DFAs should first be completely constructed and then premultiplied. pub fn set_start_state(&mut self, start: S) { assert!(!self.premultiplied, "can't set start on premultiplied DFA"); assert!(start.to_usize() < self.state_count, "invalid start state"); self.start = start; } /// Set the maximum state identifier that could possible correspond to a /// match state. /// /// Callers must uphold the invariant that any state identifier less than /// or equal to the identifier given is either a match state or the special /// dead state (which always has identifier 0 and whose transitions all /// lead back to itself). /// /// This cannot be called on a premultiplied DFA. pub fn set_max_match_state(&mut self, id: S) { assert!(!self.premultiplied, "can't set match on premultiplied DFA"); assert!(id.to_usize() < self.state_count, "invalid max match state"); self.max_match = id; } /// Add the given transition to this DFA. Both the `from` and `to` states /// must already exist. /// /// This cannot be called on a premultiplied DFA. pub fn add_transition(&mut self, from: S, byte: u8, to: S) { assert!(!self.premultiplied, "can't add trans to premultiplied DFA"); assert!(from.to_usize() < self.state_count, "invalid from state"); assert!(to.to_usize() < self.state_count, "invalid to state"); let class = self.byte_classes().get(byte); let offset = from.to_usize() * self.alphabet_len() + class as usize; self.trans[offset] = to; } /// An an empty state (a state where all transitions lead to a dead state) /// and return its identifier. The identifier returned is guaranteed to /// not point to any other existing state. /// /// If adding a state would exhaust the state identifier space (given by /// `S`), then this returns an error. In practice, this means that the /// state identifier representation chosen is too small. /// /// This cannot be called on a premultiplied DFA. pub fn add_empty_state(&mut self) -> Result { assert!(!self.premultiplied, "can't add state to premultiplied DFA"); let id = if self.state_count == 0 { S::from_usize(0) } else { next_state_id(S::from_usize(self.state_count - 1))? }; let alphabet_len = self.alphabet_len(); self.trans.extend(iter::repeat(dead_id::()).take(alphabet_len)); // This should never panic, since state_count is a usize. The // transition table size would have run out of room long ago. self.state_count = self.state_count.checked_add(1).unwrap(); Ok(id) } /// Return a mutable representation of the state corresponding to the given /// id. This is useful for implementing routines that manipulate DFA states /// (e.g., swapping states). /// /// This cannot be called on a premultiplied DFA. pub fn get_state_mut(&mut self, id: S) -> StateMut { assert!(!self.premultiplied, "can't get state in premultiplied DFA"); let alphabet_len = self.alphabet_len(); let offset = id.to_usize() * alphabet_len; StateMut { transitions: &mut self.trans[offset..offset + alphabet_len], } } /// Swap the two states given in the transition table. /// /// This routine does not do anything to check the correctness of this /// swap. Callers must ensure that other states pointing to id1 and id2 are /// updated appropriately. /// /// This cannot be called on a premultiplied DFA. pub fn swap_states(&mut self, id1: S, id2: S) { assert!(!self.premultiplied, "can't swap states in premultiplied DFA"); let o1 = id1.to_usize() * self.alphabet_len(); let o2 = id2.to_usize() * self.alphabet_len(); for b in 0..self.alphabet_len() { self.trans.swap(o1 + b, o2 + b); } } /// Truncate the states in this DFA to the given count. /// /// This routine does not do anything to check the correctness of this /// truncation. Callers must ensure that other states pointing to truncated /// states are updated appropriately. /// /// This cannot be called on a premultiplied DFA. pub fn truncate_states(&mut self, count: usize) { assert!(!self.premultiplied, "can't truncate in premultiplied DFA"); let alphabet_len = self.alphabet_len(); self.trans.truncate(count * alphabet_len); self.state_count = count; } /// This routine shuffles all match states in this DFA---according to the /// given map---to the beginning of the DFA such that every non-match state /// appears after every match state. (With one exception: the special dead /// state remains as the first state.) The given map should have length /// exactly equivalent to the number of states in this DFA. /// /// The purpose of doing this shuffling is to avoid the need to store /// additional state to determine whether a state is a match state or not. /// It also enables a single conditional in the core matching loop instead /// of two. /// /// This updates `self.max_match` to point to the last matching state as /// well as `self.start` if the starting state was moved. pub fn shuffle_match_states(&mut self, is_match: &[bool]) { assert!( !self.premultiplied, "cannot shuffle match states of premultiplied DFA" ); assert_eq!(self.state_count, is_match.len()); if self.state_count <= 1 { return; } let mut first_non_match = 1; while first_non_match < self.state_count && is_match[first_non_match] { first_non_match += 1; } let mut swaps: Vec = vec![dead_id(); self.state_count]; let mut cur = self.state_count - 1; while cur > first_non_match { if is_match[cur] { self.swap_states( S::from_usize(cur), S::from_usize(first_non_match), ); swaps[cur] = S::from_usize(first_non_match); swaps[first_non_match] = S::from_usize(cur); first_non_match += 1; while first_non_match < cur && is_match[first_non_match] { first_non_match += 1; } } cur -= 1; } for id in (0..self.state_count).map(S::from_usize) { for (_, next) in self.get_state_mut(id).iter_mut() { if swaps[next.to_usize()] != dead_id() { *next = swaps[next.to_usize()]; } } } if swaps[self.start.to_usize()] != dead_id() { self.start = swaps[self.start.to_usize()]; } self.max_match = S::from_usize(first_non_match - 1); } } #[cfg(feature = "std")] impl, S: StateID> fmt::Debug for Repr { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { fn state_status, S: StateID>( dfa: &Repr, id: S, ) -> &'static str { if id == dead_id() { if dfa.is_match_state(id) { "D*" } else { "D " } } else if id == dfa.start_state() { if dfa.is_match_state(id) { ">*" } else { "> " } } else { if dfa.is_match_state(id) { " *" } else { " " } } } writeln!(f, "DenseDFA(")?; for (id, state) in self.states() { let status = state_status(self, id); writeln!(f, "{}{:06}: {:?}", status, id.to_usize(), state)?; } writeln!(f, ")")?; Ok(()) } } /// An iterator over all states in a DFA. /// /// This iterator yields a tuple for each state. The first element of the /// tuple corresponds to a state's identifier, and the second element /// corresponds to the state itself (comprised of its transitions). /// /// If this DFA is premultiplied, then the state identifiers are in turn /// premultiplied as well, making them usable without additional modification. /// /// `'a` corresponding to the lifetime of original DFA, `T` corresponds to /// the type of the transition table itself and `S` corresponds to the state /// identifier representation. #[cfg(feature = "std")] pub(crate) struct StateIter<'a, T: 'a, S: 'a> { dfa: &'a Repr, it: iter::Enumerate>, } #[cfg(feature = "std")] impl<'a, T: AsRef<[S]>, S: StateID> Iterator for StateIter<'a, T, S> { type Item = (S, State<'a, S>); fn next(&mut self) -> Option<(S, State<'a, S>)> { self.it.next().map(|(id, chunk)| { let state = State { transitions: chunk }; let id = if self.dfa.premultiplied { id * self.dfa.alphabet_len() } else { id }; (S::from_usize(id), state) }) } } /// An immutable representation of a single DFA state. /// /// `'a` correspondings to the lifetime of a DFA's transition table and `S` /// corresponds to the state identifier representation. #[cfg(feature = "std")] pub(crate) struct State<'a, S: 'a> { transitions: &'a [S], } #[cfg(feature = "std")] impl<'a, S: StateID> State<'a, S> { /// Return an iterator over all transitions in this state. This yields /// a number of transitions equivalent to the alphabet length of the /// corresponding DFA. /// /// Each transition is represented by a tuple. The first element is /// the input byte for that transition and the second element is the /// transitions itself. pub fn transitions(&self) -> StateTransitionIter { StateTransitionIter { it: self.transitions.iter().enumerate() } } /// Return an iterator over a sparse representation of the transitions in /// this state. Only non-dead transitions are returned. /// /// The "sparse" representation in this case corresponds to a sequence of /// triples. The first two elements of the triple comprise an inclusive /// byte range while the last element corresponds to the transition taken /// for all bytes in the range. /// /// This is somewhat more condensed than the classical sparse /// representation (where you have an element for every non-dead /// transition), but in practice, checking if a byte is in a range is very /// cheap and using ranges tends to conserve quite a bit more space. pub fn sparse_transitions(&self) -> StateSparseTransitionIter { StateSparseTransitionIter { dense: self.transitions(), cur: None } } } #[cfg(feature = "std")] impl<'a, S: StateID> fmt::Debug for State<'a, S> { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { let mut transitions = vec![]; for (start, end, next_id) in self.sparse_transitions() { let line = if start == end { format!("{} => {}", escape(start), next_id.to_usize()) } else { format!( "{}-{} => {}", escape(start), escape(end), next_id.to_usize(), ) }; transitions.push(line); } write!(f, "{}", transitions.join(", "))?; Ok(()) } } /// An iterator over all transitions in a single DFA state. This yields /// a number of transitions equivalent to the alphabet length of the /// corresponding DFA. /// /// Each transition is represented by a tuple. The first element is the input /// byte for that transition and the second element is the transitions itself. #[cfg(feature = "std")] #[derive(Debug)] pub(crate) struct StateTransitionIter<'a, S: 'a> { it: iter::Enumerate>, } #[cfg(feature = "std")] impl<'a, S: StateID> Iterator for StateTransitionIter<'a, S> { type Item = (u8, S); fn next(&mut self) -> Option<(u8, S)> { self.it.next().map(|(i, &id)| (i as u8, id)) } } /// An iterator over all transitions in a single DFA state using a sparse /// representation. /// /// Each transition is represented by a triple. The first two elements of the /// triple comprise an inclusive byte range while the last element corresponds /// to the transition taken for all bytes in the range. #[cfg(feature = "std")] #[derive(Debug)] pub(crate) struct StateSparseTransitionIter<'a, S: 'a> { dense: StateTransitionIter<'a, S>, cur: Option<(u8, u8, S)>, } #[cfg(feature = "std")] impl<'a, S: StateID> Iterator for StateSparseTransitionIter<'a, S> { type Item = (u8, u8, S); fn next(&mut self) -> Option<(u8, u8, S)> { while let Some((b, next)) = self.dense.next() { let (prev_start, prev_end, prev_next) = match self.cur { Some(t) => t, None => { self.cur = Some((b, b, next)); continue; } }; if prev_next == next { self.cur = Some((prev_start, b, prev_next)); } else { self.cur = Some((b, b, next)); if prev_next != dead_id() { return Some((prev_start, prev_end, prev_next)); } } } if let Some((start, end, next)) = self.cur.take() { if next != dead_id() { return Some((start, end, next)); } } None } } /// A mutable representation of a single DFA state. /// /// `'a` correspondings to the lifetime of a DFA's transition table and `S` /// corresponds to the state identifier representation. #[cfg(feature = "std")] pub(crate) struct StateMut<'a, S: 'a> { transitions: &'a mut [S], } #[cfg(feature = "std")] impl<'a, S: StateID> StateMut<'a, S> { /// Return an iterator over all transitions in this state. This yields /// a number of transitions equivalent to the alphabet length of the /// corresponding DFA. /// /// Each transition is represented by a tuple. The first element is the /// input byte for that transition and the second element is a mutable /// reference to the transition itself. pub fn iter_mut(&mut self) -> StateTransitionIterMut { StateTransitionIterMut { it: self.transitions.iter_mut().enumerate() } } } #[cfg(feature = "std")] impl<'a, S: StateID> fmt::Debug for StateMut<'a, S> { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { fmt::Debug::fmt(&State { transitions: self.transitions }, f) } } /// A mutable iterator over all transitions in a DFA state. /// /// Each transition is represented by a tuple. The first element is the /// input byte for that transition and the second element is a mutable /// reference to the transition itself. #[cfg(feature = "std")] #[derive(Debug)] pub(crate) struct StateTransitionIterMut<'a, S: 'a> { it: iter::Enumerate>, } #[cfg(feature = "std")] impl<'a, S: StateID> Iterator for StateTransitionIterMut<'a, S> { type Item = (u8, &'a mut S); fn next(&mut self) -> Option<(u8, &'a mut S)> { self.it.next().map(|(i, id)| (i as u8, id)) } } /// A builder for constructing a deterministic finite automaton from regular /// expressions. /// /// This builder permits configuring several aspects of the construction /// process such as case insensitivity, Unicode support and various options /// that impact the size of the generated DFA. In some cases, options (like /// performing DFA minimization) can come with a substantial additional cost. /// /// This builder always constructs a *single* DFA. As such, this builder can /// only be used to construct regexes that either detect the presence of a /// match or find the end location of a match. A single DFA cannot produce both /// the start and end of a match. For that information, use a /// [`Regex`](struct.Regex.html), which can be similarly configured using /// [`RegexBuilder`](struct.RegexBuilder.html). #[cfg(feature = "std")] #[derive(Clone, Debug)] pub struct Builder { parser: ParserBuilder, nfa: nfa::Builder, anchored: bool, minimize: bool, premultiply: bool, byte_classes: bool, reverse: bool, longest_match: bool, } #[cfg(feature = "std")] impl Builder { /// Create a new DenseDFA builder with the default configuration. pub fn new() -> Builder { let mut nfa = nfa::Builder::new(); // This is enabled by default, but we set it here anyway. Since we're // building a DFA, shrinking the NFA is always a good idea. nfa.shrink(true); Builder { parser: ParserBuilder::new(), nfa, anchored: false, minimize: false, premultiply: true, byte_classes: true, reverse: false, longest_match: false, } } /// Build a DFA from the given pattern. /// /// If there was a problem parsing or compiling the pattern, then an error /// is returned. pub fn build(&self, pattern: &str) -> Result, usize>> { self.build_with_size::(pattern) } /// Build a DFA from the given pattern using a specific representation for /// the DFA's state IDs. /// /// If there was a problem parsing or compiling the pattern, then an error /// is returned. /// /// The representation of state IDs is determined by the `S` type /// parameter. In general, `S` is usually one of `u8`, `u16`, `u32`, `u64` /// or `usize`, where `usize` is the default used for `build`. The purpose /// of specifying a representation for state IDs is to reduce the memory /// footprint of a DFA. /// /// When using this routine, the chosen state ID representation will be /// used throughout determinization and minimization, if minimization /// was requested. Even if the minimized DFA can fit into the chosen /// state ID representation but the initial determinized DFA cannot, /// then this will still return an error. To get a minimized DFA with a /// smaller state ID representation, first build it with a bigger state ID /// representation, and then shrink the size of the DFA using one of its /// conversion routines, such as /// [`DenseDFA::to_u16`](enum.DenseDFA.html#method.to_u16). pub fn build_with_size( &self, pattern: &str, ) -> Result, S>> { self.build_from_nfa(&self.build_nfa(pattern)?) } /// An internal only (for now) API for building a dense DFA directly from /// an NFA. pub(crate) fn build_from_nfa( &self, nfa: &NFA, ) -> Result, S>> { if self.longest_match && !self.anchored { return Err(Error::unsupported_longest_match()); } let mut dfa = if self.byte_classes { Determinizer::new(nfa) .with_byte_classes() .longest_match(self.longest_match) .build() } else { Determinizer::new(nfa).longest_match(self.longest_match).build() }?; if self.minimize { dfa.minimize(); } if self.premultiply { dfa.premultiply()?; } Ok(dfa.into_dense_dfa()) } /// Builds an NFA from the given pattern. pub(crate) fn build_nfa(&self, pattern: &str) -> Result { let hir = self.parser.build().parse(pattern).map_err(Error::syntax)?; Ok(self.nfa.build(&hir)?) } /// Set whether matching must be anchored at the beginning of the input. /// /// When enabled, a match must begin at the start of the input. When /// disabled, the DFA will act as if the pattern started with a `.*?`, /// which enables a match to appear anywhere. /// /// By default this is disabled. pub fn anchored(&mut self, yes: bool) -> &mut Builder { self.anchored = yes; self.nfa.anchored(yes); self } /// Enable or disable the case insensitive flag by default. /// /// By default this is disabled. It may alternatively be selectively /// enabled in the regular expression itself via the `i` flag. pub fn case_insensitive(&mut self, yes: bool) -> &mut Builder { self.parser.case_insensitive(yes); self } /// Enable verbose mode in the regular expression. /// /// When enabled, verbose mode permits insigificant whitespace in many /// places in the regular expression, as well as comments. Comments are /// started using `#` and continue until the end of the line. /// /// By default, this is disabled. It may be selectively enabled in the /// regular expression by using the `x` flag regardless of this setting. pub fn ignore_whitespace(&mut self, yes: bool) -> &mut Builder { self.parser.ignore_whitespace(yes); self } /// Enable or disable the "dot matches any character" flag by default. /// /// By default this is disabled. It may alternatively be selectively /// enabled in the regular expression itself via the `s` flag. pub fn dot_matches_new_line(&mut self, yes: bool) -> &mut Builder { self.parser.dot_matches_new_line(yes); self } /// Enable or disable the "swap greed" flag by default. /// /// By default this is disabled. It may alternatively be selectively /// enabled in the regular expression itself via the `U` flag. pub fn swap_greed(&mut self, yes: bool) -> &mut Builder { self.parser.swap_greed(yes); self } /// Enable or disable the Unicode flag (`u`) by default. /// /// By default this is **enabled**. It may alternatively be selectively /// disabled in the regular expression itself via the `u` flag. /// /// Note that unless `allow_invalid_utf8` is enabled (it's disabled by /// default), a regular expression will fail to parse if Unicode mode is /// disabled and a sub-expression could possibly match invalid UTF-8. pub fn unicode(&mut self, yes: bool) -> &mut Builder { self.parser.unicode(yes); self } /// When enabled, the builder will permit the construction of a regular /// expression that may match invalid UTF-8. /// /// When disabled (the default), the builder is guaranteed to produce a /// regex that will only ever match valid UTF-8 (otherwise, the builder /// will return an error). pub fn allow_invalid_utf8(&mut self, yes: bool) -> &mut Builder { self.parser.allow_invalid_utf8(yes); self.nfa.allow_invalid_utf8(yes); self } /// Set the nesting limit used for the regular expression parser. /// /// The nesting limit controls how deep the abstract syntax tree is allowed /// to be. If the AST exceeds the given limit (e.g., with too many nested /// groups), then an error is returned by the parser. /// /// The purpose of this limit is to act as a heuristic to prevent stack /// overflow when building a finite automaton from a regular expression's /// abstract syntax tree. In particular, construction currently uses /// recursion. In the future, the implementation may stop using recursion /// and this option will no longer be necessary. /// /// This limit is not checked until the entire AST is parsed. Therefore, /// if callers want to put a limit on the amount of heap space used, then /// they should impose a limit on the length, in bytes, of the concrete /// pattern string. In particular, this is viable since the parser will /// limit itself to heap space proportional to the lenth of the pattern /// string. /// /// Note that a nest limit of `0` will return a nest limit error for most /// patterns but not all. For example, a nest limit of `0` permits `a` but /// not `ab`, since `ab` requires a concatenation AST item, which results /// in a nest depth of `1`. In general, a nest limit is not something that /// manifests in an obvious way in the concrete syntax, therefore, it /// should not be used in a granular way. pub fn nest_limit(&mut self, limit: u32) -> &mut Builder { self.parser.nest_limit(limit); self } /// Minimize the DFA. /// /// When enabled, the DFA built will be minimized such that it is as small /// as possible. /// /// Whether one enables minimization or not depends on the types of costs /// you're willing to pay and how much you care about its benefits. In /// particular, minimization has worst case `O(n*k*logn)` time and `O(k*n)` /// space, where `n` is the number of DFA states and `k` is the alphabet /// size. In practice, minimization can be quite costly in terms of both /// space and time, so it should only be done if you're willing to wait /// longer to produce a DFA. In general, you might want a minimal DFA in /// the following circumstances: /// /// 1. You would like to optimize for the size of the automaton. This can /// manifest in one of two ways. Firstly, if you're converting the /// DFA into Rust code (or a table embedded in the code), then a minimal /// DFA will translate into a corresponding reduction in code size, and /// thus, also the final compiled binary size. Secondly, if you are /// building many DFAs and putting them on the heap, you'll be able to /// fit more if they are smaller. Note though that building a minimal /// DFA itself requires additional space; you only realize the space /// savings once the minimal DFA is constructed (at which point, the /// space used for minimization is freed). /// 2. You've observed that a smaller DFA results in faster match /// performance. Naively, this isn't guaranteed since there is no /// inherent difference between matching with a bigger-than-minimal /// DFA and a minimal DFA. However, a smaller DFA may make use of your /// CPU's cache more efficiently. /// 3. You are trying to establish an equivalence between regular /// languages. The standard method for this is to build a minimal DFA /// for each language and then compare them. If the DFAs are equivalent /// (up to state renaming), then the languages are equivalent. /// /// This option is disabled by default. pub fn minimize(&mut self, yes: bool) -> &mut Builder { self.minimize = yes; self } /// Premultiply state identifiers in the DFA's transition table. /// /// When enabled, state identifiers are premultiplied to point to their /// corresponding row in the DFA's transition table. That is, given the /// `i`th state, its corresponding premultiplied identifier is `i * k` /// where `k` is the alphabet size of the DFA. (The alphabet size is at /// most 256, but is in practice smaller if byte classes is enabled.) /// /// When state identifiers are not premultiplied, then the identifier of /// the `i`th state is `i`. /// /// The advantage of premultiplying state identifiers is that is saves /// a multiplication instruction per byte when searching with the DFA. /// This has been observed to lead to a 20% performance benefit in /// micro-benchmarks. /// /// The primary disadvantage of premultiplying state identifiers is /// that they require a larger integer size to represent. For example, /// if your DFA has 200 states, then its premultiplied form requires /// 16 bits to represent every possible state identifier, where as its /// non-premultiplied form only requires 8 bits. /// /// This option is enabled by default. pub fn premultiply(&mut self, yes: bool) -> &mut Builder { self.premultiply = yes; self } /// Shrink the size of the DFA's alphabet by mapping bytes to their /// equivalence classes. /// /// When enabled, each DFA will use a map from all possible bytes to their /// corresponding equivalence class. Each equivalence class represents a /// set of bytes that does not discriminate between a match and a non-match /// in the DFA. For example, the pattern `[ab]+` has at least two /// equivalence classes: a set containing `a` and `b` and a set containing /// every byte except for `a` and `b`. `a` and `b` are in the same /// equivalence classes because they never discriminate between a match /// and a non-match. /// /// The advantage of this map is that the size of the transition table can /// be reduced drastically from `#states * 256 * sizeof(id)` to /// `#states * k * sizeof(id)` where `k` is the number of equivalence /// classes. As a result, total space usage can decrease substantially. /// Moreover, since a smaller alphabet is used, compilation becomes faster /// as well. /// /// The disadvantage of this map is that every byte searched must be /// passed through this map before it can be used to determine the next /// transition. This has a small match time performance cost. /// /// This option is enabled by default. pub fn byte_classes(&mut self, yes: bool) -> &mut Builder { self.byte_classes = yes; self } /// Reverse the DFA. /// /// A DFA reversal is performed by reversing all of the concatenated /// sub-expressions in the original pattern, recursively. The resulting /// DFA can be used to match the pattern starting from the end of a string /// instead of the beginning of a string. /// /// Generally speaking, a reversed DFA is most useful for finding the start /// of a match, since a single forward DFA is only capable of finding the /// end of a match. This start of match handling is done for you /// automatically if you build a [`Regex`](struct.Regex.html). pub fn reverse(&mut self, yes: bool) -> &mut Builder { self.reverse = yes; self.nfa.reverse(yes); self } /// Find the longest possible match. /// /// This is distinct from the default leftmost-first match semantics in /// that it treats all NFA states as having equivalent priority. In other /// words, the longest possible match is always found and it is not /// possible to implement non-greedy match semantics when this is set. That /// is, `a+` and `a+?` are equivalent when this is enabled. /// /// In particular, a practical issue with this option at the moment is that /// it prevents unanchored searches from working correctly, since /// unanchored searches are implemented by prepending an non-greedy `.*?` /// to the beginning of the pattern. As stated above, non-greedy match /// semantics aren't supported. Therefore, if this option is enabled and /// an unanchored search is requested, then building a DFA will return an /// error. /// /// This option is principally useful when building a reverse DFA for /// finding the start of a match. If you are building a regex with /// [`RegexBuilder`](struct.RegexBuilder.html), then this is handled for /// you automatically. The reason why this is necessary for start of match /// handling is because we want to find the earliest possible starting /// position of a match to satisfy leftmost-first match semantics. When /// matching in reverse, this means finding the longest possible match, /// hence, this option. /// /// By default this is disabled. pub fn longest_match(&mut self, yes: bool) -> &mut Builder { // There is prior art in RE2 that shows how this can support unanchored // searches. Instead of treating all NFA states as having equivalent // priority, we instead group NFA states into sets, and treat members // of each set as having equivalent priority, but having greater // priority than all following members of different sets. We then // essentially assign a higher priority to everything over the prefix // `.*?`. self.longest_match = yes; self } /// Apply best effort heuristics to shrink the NFA at the expense of more /// time/memory. /// /// This may be exposed in the future, but for now is exported for use in /// the `regex-automata-debug` tool. #[doc(hidden)] pub fn shrink(&mut self, yes: bool) -> &mut Builder { self.nfa.shrink(yes); self } } #[cfg(feature = "std")] impl Default for Builder { fn default() -> Builder { Builder::new() } } /// Return the given byte as its escaped string form. #[cfg(feature = "std")] fn escape(b: u8) -> String { use std::ascii; String::from_utf8(ascii::escape_default(b).collect::>()).unwrap() } #[cfg(all(test, feature = "std"))] mod tests { use super::*; #[test] fn errors_when_converting_to_smaller_dfa() { let pattern = r"\w{10}"; let dfa = Builder::new() .byte_classes(false) .anchored(true) .premultiply(false) .build_with_size::(pattern) .unwrap(); assert!(dfa.to_u8().is_err()); } #[test] fn errors_when_determinization_would_overflow() { let pattern = r"\w{10}"; let mut builder = Builder::new(); builder.byte_classes(false).anchored(true).premultiply(false); // using u16 is fine assert!(builder.build_with_size::(pattern).is_ok()); // // ... but u8 results in overflow (because there are >256 states) assert!(builder.build_with_size::(pattern).is_err()); } #[test] fn errors_when_premultiply_would_overflow() { let pattern = r"[a-z]"; let mut builder = Builder::new(); builder.byte_classes(false).anchored(true).premultiply(false); // without premultiplication is OK assert!(builder.build_with_size::(pattern).is_ok()); // ... but with premultiplication overflows u8 builder.premultiply(true); assert!(builder.build_with_size::(pattern).is_err()); } // let data = ::std::fs::read_to_string("/usr/share/dict/words").unwrap(); // let mut words: Vec<&str> = data.lines().collect(); // println!("{} words", words.len()); // words.sort_by(|w1, w2| w1.len().cmp(&w2.len()).reverse()); // let pattern = words.join("|"); // print_automata_counts(&pattern); // print_automata(&pattern); // print_automata(r"[01]*1[01]{5}"); // print_automata(r"X(.?){0,8}Y"); // print_automata_counts(r"\p{alphabetic}"); // print_automata(r"a*b+|cdefg"); // print_automata(r"(..)*(...)*"); // let pattern = r"\p{any}*?\p{Other_Uppercase}"; // let pattern = r"\p{any}*?\w+"; // print_automata_counts(pattern); // print_automata_counts(r"(?-u:\w)"); // let pattern = r"\p{Greek}"; // let pattern = r"zZzZzZzZzZ"; // let pattern = grapheme_pattern(); // let pattern = r"\p{Ideographic}"; // let pattern = r"\w{10}"; // 51784 --> 41264 // let pattern = r"\w"; // 5182 // let pattern = r"a*"; // print_automata(pattern); // let (_, _, dfa) = build_automata(pattern); }