use rayon::prelude::*; const OCTILLION: u128 = 1_000_000_000_000_000_000_000_000_000; /// Produce a parallel iterator for 0u128..10²⁷ fn octillion() -> rayon::range::Iter { (0..OCTILLION).into_par_iter() } /// Produce a parallel iterator for 0u128..=10²⁷ fn octillion_inclusive() -> rayon::range_inclusive::Iter { (0..=OCTILLION).into_par_iter() } /// Produce a parallel iterator for 0u128..10²⁷ using `flat_map` fn octillion_flat() -> impl ParallelIterator { (0u32..1_000_000_000) .into_par_iter() .with_max_len(1_000) .map(|i| u64::from(i) * 1_000_000_000) .flat_map(|i| { (0u32..1_000_000_000) .into_par_iter() .with_max_len(1_000) .map(move |j| i + u64::from(j)) }) .map(|i| u128::from(i) * 1_000_000_000) .flat_map(|i| { (0u32..1_000_000_000) .into_par_iter() .with_max_len(1_000) .map(move |j| i + u128::from(j)) }) } // NOTE: `find_first` and `find_last` currently take too long on 32-bit targets, // because the `AtomicUsize` match position has much too limited resolution. #[test] #[cfg_attr(not(target_pointer_width = "64"), ignore)] fn find_first_octillion() { let x = octillion().find_first(|_| true); assert_eq!(x, Some(0)); } #[test] #[cfg_attr(not(target_pointer_width = "64"), ignore)] fn find_first_octillion_inclusive() { let x = octillion_inclusive().find_first(|_| true); assert_eq!(x, Some(0)); } #[test] #[cfg_attr(not(target_pointer_width = "64"), ignore)] fn find_first_octillion_flat() { let x = octillion_flat().find_first(|_| true); assert_eq!(x, Some(0)); } fn two_threads R, R: Send>(f: F) -> R { // FIXME: If we don't use at least two threads, then we end up walking // through the entire iterator sequentially, without the benefit of any // short-circuiting. We probably don't want testing to wait that long. ;) let builder = rayon::ThreadPoolBuilder::new().num_threads(2); let pool = builder.build().unwrap(); pool.install(f) } #[test] #[cfg_attr(not(target_pointer_width = "64"), ignore)] fn find_last_octillion() { // It would be nice if `find_last` could prioritize the later splits, // basically flipping the `join` args, without needing indexed `rev`. // (or could we have an unindexed `rev`?) let x = two_threads(|| octillion().find_last(|_| true)); assert_eq!(x, Some(OCTILLION - 1)); } #[test] #[cfg_attr(not(target_pointer_width = "64"), ignore)] fn find_last_octillion_inclusive() { let x = two_threads(|| octillion_inclusive().find_last(|_| true)); assert_eq!(x, Some(OCTILLION)); } #[test] #[cfg_attr(not(target_pointer_width = "64"), ignore)] fn find_last_octillion_flat() { let x = two_threads(|| octillion_flat().find_last(|_| true)); assert_eq!(x, Some(OCTILLION - 1)); } #[test] fn find_any_octillion() { let x = two_threads(|| octillion().find_any(|x| *x > OCTILLION / 2)); assert!(x.is_some()); } #[test] fn find_any_octillion_flat() { let x = two_threads(|| octillion_flat().find_any(|x| *x > OCTILLION / 2)); assert!(x.is_some()); } #[test] fn filter_find_any_octillion() { let x = two_threads(|| { octillion() .filter(|x| *x > OCTILLION / 2) .find_any(|_| true) }); assert!(x.is_some()); } #[test] fn filter_find_any_octillion_flat() { let x = two_threads(|| { octillion_flat() .filter(|x| *x > OCTILLION / 2) .find_any(|_| true) }); assert!(x.is_some()); } #[test] fn fold_find_any_octillion_flat() { let x = two_threads(|| octillion_flat().fold(|| (), |_, _| ()).find_any(|_| true)); assert!(x.is_some()); }