//! Parallel iterator types for [slices][std::slice] //! //! You will rarely need to interact with this module directly unless you need //! to name one of the iterator types. //! //! [std::slice]: https://doc.rust-lang.org/stable/std/slice/ mod mergesort; mod quicksort; mod test; use self::mergesort::par_mergesort; use self::quicksort::par_quicksort; use crate::iter::plumbing::*; use crate::iter::*; use crate::split_producer::*; use std::cmp; use std::cmp::Ordering; use std::fmt::{self, Debug}; use super::math::div_round_up; /// Parallel extensions for slices. pub trait ParallelSlice { /// Returns a plain slice, which is used to implement the rest of the /// parallel methods. fn as_parallel_slice(&self) -> &[T]; /// Returns a parallel iterator over subslices separated by elements that /// match the separator. /// /// # Examples /// /// ``` /// use rayon::prelude::*; /// let smallest = [1, 2, 3, 0, 2, 4, 8, 0, 3, 6, 9] /// .par_split(|i| *i == 0) /// .map(|numbers| numbers.iter().min().unwrap()) /// .min(); /// assert_eq!(Some(&1), smallest); /// ``` fn par_split

(&self, separator: P) -> Split<'_, T, P> where P: Fn(&T) -> bool + Sync + Send, { Split { slice: self.as_parallel_slice(), separator, } } /// Returns a parallel iterator over all contiguous windows of length /// `window_size`. The windows overlap. /// /// # Examples /// /// ``` /// use rayon::prelude::*; /// let windows: Vec<_> = [1, 2, 3].par_windows(2).collect(); /// assert_eq!(vec![[1, 2], [2, 3]], windows); /// ``` fn par_windows(&self, window_size: usize) -> Windows<'_, T> { Windows { window_size, slice: self.as_parallel_slice(), } } /// Returns a parallel iterator over at most `chunk_size` elements of /// `self` at a time. The chunks do not overlap. /// /// If the number of elements in the iterator is not divisible by /// `chunk_size`, the last chunk may be shorter than `chunk_size`. All /// other chunks will have that exact length. /// /// # Examples /// /// ``` /// use rayon::prelude::*; /// let chunks: Vec<_> = [1, 2, 3, 4, 5].par_chunks(2).collect(); /// assert_eq!(chunks, vec![&[1, 2][..], &[3, 4], &[5]]); /// ``` fn par_chunks(&self, chunk_size: usize) -> Chunks<'_, T> { assert!(chunk_size != 0, "chunk_size must not be zero"); Chunks { chunk_size, slice: self.as_parallel_slice(), } } /// Returns a parallel iterator over `chunk_size` elements of /// `self` at a time. The chunks do not overlap. /// /// If `chunk_size` does not divide the length of the slice, then the /// last up to `chunk_size-1` elements will be omitted and can be /// retrieved from the remainder function of the iterator. /// /// # Examples /// /// ``` /// use rayon::prelude::*; /// let chunks: Vec<_> = [1, 2, 3, 4, 5].par_chunks_exact(2).collect(); /// assert_eq!(chunks, vec![&[1, 2][..], &[3, 4]]); /// ``` fn par_chunks_exact(&self, chunk_size: usize) -> ChunksExact<'_, T> { assert!(chunk_size != 0, "chunk_size must not be zero"); let slice = self.as_parallel_slice(); let rem = slice.len() % chunk_size; let len = slice.len() - rem; let (fst, snd) = slice.split_at(len); ChunksExact { chunk_size, slice: fst, rem: snd, } } } impl ParallelSlice for [T] { #[inline] fn as_parallel_slice(&self) -> &[T] { self } } /// Parallel extensions for mutable slices. pub trait ParallelSliceMut { /// Returns a plain mutable slice, which is used to implement the rest of /// the parallel methods. fn as_parallel_slice_mut(&mut self) -> &mut [T]; /// Returns a parallel iterator over mutable subslices separated by /// elements that match the separator. /// /// # Examples /// /// ``` /// use rayon::prelude::*; /// let mut array = [1, 2, 3, 0, 2, 4, 8, 0, 3, 6, 9]; /// array.par_split_mut(|i| *i == 0) /// .for_each(|slice| slice.reverse()); /// assert_eq!(array, [3, 2, 1, 0, 8, 4, 2, 0, 9, 6, 3]); /// ``` fn par_split_mut

(&mut self, separator: P) -> SplitMut<'_, T, P> where P: Fn(&T) -> bool + Sync + Send, { SplitMut { slice: self.as_parallel_slice_mut(), separator, } } /// Returns a parallel iterator over at most `chunk_size` elements of /// `self` at a time. The chunks are mutable and do not overlap. /// /// If the number of elements in the iterator is not divisible by /// `chunk_size`, the last chunk may be shorter than `chunk_size`. All /// other chunks will have that exact length. /// /// # Examples /// /// ``` /// use rayon::prelude::*; /// let mut array = [1, 2, 3, 4, 5]; /// array.par_chunks_mut(2) /// .for_each(|slice| slice.reverse()); /// assert_eq!(array, [2, 1, 4, 3, 5]); /// ``` fn par_chunks_mut(&mut self, chunk_size: usize) -> ChunksMut<'_, T> { assert!(chunk_size != 0, "chunk_size must not be zero"); ChunksMut { chunk_size, slice: self.as_parallel_slice_mut(), } } /// Returns a parallel iterator over `chunk_size` elements of /// `self` at a time. The chunks are mutable and do not overlap. /// /// If `chunk_size` does not divide the length of the slice, then the /// last up to `chunk_size-1` elements will be omitted and can be /// retrieved from the remainder function of the iterator. /// /// # Examples /// /// ``` /// use rayon::prelude::*; /// let mut array = [1, 2, 3, 4, 5]; /// array.par_chunks_exact_mut(3) /// .for_each(|slice| slice.reverse()); /// assert_eq!(array, [3, 2, 1, 4, 5]); /// ``` fn par_chunks_exact_mut(&mut self, chunk_size: usize) -> ChunksExactMut<'_, T> { assert!(chunk_size != 0, "chunk_size must not be zero"); let slice = self.as_parallel_slice_mut(); let rem = slice.len() % chunk_size; let len = slice.len() - rem; let (fst, snd) = slice.split_at_mut(len); ChunksExactMut { chunk_size, slice: fst, rem: snd, } } /// Sorts the slice in parallel. /// /// This sort is stable (i.e. does not reorder equal elements) and `O(n log n)` worst-case. /// /// When applicable, unstable sorting is preferred because it is generally faster than stable /// sorting and it doesn't allocate auxiliary memory. /// See [`par_sort_unstable`](#method.par_sort_unstable). /// /// # Current implementation /// /// The current algorithm is an adaptive merge sort inspired by /// [timsort](https://en.wikipedia.org/wiki/Timsort). /// It is designed to be very fast in cases where the slice is nearly sorted, or consists of /// two or more sorted sequences concatenated one after another. /// /// Also, it allocates temporary storage the same size as `self`, but for very short slices a /// non-allocating insertion sort is used instead. /// /// In order to sort the slice in parallel, the slice is first divided into smaller chunks and /// all chunks are sorted in parallel. Then, adjacent chunks that together form non-descending /// or descending runs are concatenated. Finally, the remaining chunks are merged together using /// parallel subdivision of chunks and parallel merge operation. /// /// # Examples /// /// ``` /// use rayon::prelude::*; /// /// let mut v = [-5, 4, 1, -3, 2]; /// /// v.par_sort(); /// assert_eq!(v, [-5, -3, 1, 2, 4]); /// ``` fn par_sort(&mut self) where T: Ord, { par_mergesort(self.as_parallel_slice_mut(), T::lt); } /// Sorts the slice in parallel with a comparator function. /// /// This sort is stable (i.e. does not reorder equal elements) and `O(n log n)` worst-case. /// /// When applicable, unstable sorting is preferred because it is generally faster than stable /// sorting and it doesn't allocate auxiliary memory. /// See [`par_sort_unstable_by`](#method.par_sort_unstable_by). /// /// # Current implementation /// /// The current algorithm is an adaptive merge sort inspired by /// [timsort](https://en.wikipedia.org/wiki/Timsort). /// It is designed to be very fast in cases where the slice is nearly sorted, or consists of /// two or more sorted sequences concatenated one after another. /// /// Also, it allocates temporary storage the same size as `self`, but for very short slices a /// non-allocating insertion sort is used instead. /// /// In order to sort the slice in parallel, the slice is first divided into smaller chunks and /// all chunks are sorted in parallel. Then, adjacent chunks that together form non-descending /// or descending runs are concatenated. Finally, the remaining chunks are merged together using /// parallel subdivision of chunks and parallel merge operation. /// /// # Examples /// /// ``` /// use rayon::prelude::*; /// /// let mut v = [5, 4, 1, 3, 2]; /// v.par_sort_by(|a, b| a.cmp(b)); /// assert_eq!(v, [1, 2, 3, 4, 5]); /// /// // reverse sorting /// v.par_sort_by(|a, b| b.cmp(a)); /// assert_eq!(v, [5, 4, 3, 2, 1]); /// ``` fn par_sort_by(&mut self, compare: F) where F: Fn(&T, &T) -> Ordering + Sync, { par_mergesort(self.as_parallel_slice_mut(), |a, b| { compare(a, b) == Ordering::Less }); } /// Sorts the slice in parallel with a key extraction function. /// /// This sort is stable (i.e. does not reorder equal elements) and `O(n log n)` worst-case. /// /// When applicable, unstable sorting is preferred because it is generally faster than stable /// sorting and it doesn't allocate auxiliary memory. /// See [`par_sort_unstable_by_key`](#method.par_sort_unstable_by_key). /// /// # Current implementation /// /// The current algorithm is an adaptive merge sort inspired by /// [timsort](https://en.wikipedia.org/wiki/Timsort). /// It is designed to be very fast in cases where the slice is nearly sorted, or consists of /// two or more sorted sequences concatenated one after another. /// /// Also, it allocates temporary storage the same size as `self`, but for very short slices a /// non-allocating insertion sort is used instead. /// /// In order to sort the slice in parallel, the slice is first divided into smaller chunks and /// all chunks are sorted in parallel. Then, adjacent chunks that together form non-descending /// or descending runs are concatenated. Finally, the remaining chunks are merged together using /// parallel subdivision of chunks and parallel merge operation. /// /// # Examples /// /// ``` /// use rayon::prelude::*; /// /// let mut v = [-5i32, 4, 1, -3, 2]; /// /// v.par_sort_by_key(|k| k.abs()); /// assert_eq!(v, [1, 2, -3, 4, -5]); /// ``` fn par_sort_by_key(&mut self, f: F) where B: Ord, F: Fn(&T) -> B + Sync, { par_mergesort(self.as_parallel_slice_mut(), |a, b| f(a).lt(&f(b))); } /// Sorts the slice in parallel, but may not preserve the order of equal elements. /// /// This sort is unstable (i.e. may reorder equal elements), in-place (i.e. does not allocate), /// and `O(n log n)` worst-case. /// /// # Current implementation /// /// The current algorithm is based on Orson Peters' [pattern-defeating quicksort][pdqsort], /// which is a quicksort variant designed to be very fast on certain kinds of patterns, /// sometimes achieving linear time. It is randomized but deterministic, and falls back to /// heapsort on degenerate inputs. /// /// It is generally faster than stable sorting, except in a few special cases, e.g. when the /// slice consists of several concatenated sorted sequences. /// /// All quicksorts work in two stages: partitioning into two halves followed by recursive /// calls. The partitioning phase is sequential, but the two recursive calls are performed in /// parallel. /// /// [pdqsort]: https://github.com/orlp/pdqsort /// /// # Examples /// /// ``` /// use rayon::prelude::*; /// /// let mut v = [-5, 4, 1, -3, 2]; /// /// v.par_sort_unstable(); /// assert_eq!(v, [-5, -3, 1, 2, 4]); /// ``` fn par_sort_unstable(&mut self) where T: Ord, { par_quicksort(self.as_parallel_slice_mut(), T::lt); } /// Sorts the slice in parallel with a comparator function, but may not preserve the order of /// equal elements. /// /// This sort is unstable (i.e. may reorder equal elements), in-place (i.e. does not allocate), /// and `O(n log n)` worst-case. /// /// # Current implementation /// /// The current algorithm is based on Orson Peters' [pattern-defeating quicksort][pdqsort], /// which is a quicksort variant designed to be very fast on certain kinds of patterns, /// sometimes achieving linear time. It is randomized but deterministic, and falls back to /// heapsort on degenerate inputs. /// /// It is generally faster than stable sorting, except in a few special cases, e.g. when the /// slice consists of several concatenated sorted sequences. /// /// All quicksorts work in two stages: partitioning into two halves followed by recursive /// calls. The partitioning phase is sequential, but the two recursive calls are performed in /// parallel. /// /// [pdqsort]: https://github.com/orlp/pdqsort /// /// # Examples /// /// ``` /// use rayon::prelude::*; /// /// let mut v = [5, 4, 1, 3, 2]; /// v.par_sort_unstable_by(|a, b| a.cmp(b)); /// assert_eq!(v, [1, 2, 3, 4, 5]); /// /// // reverse sorting /// v.par_sort_unstable_by(|a, b| b.cmp(a)); /// assert_eq!(v, [5, 4, 3, 2, 1]); /// ``` fn par_sort_unstable_by(&mut self, compare: F) where F: Fn(&T, &T) -> Ordering + Sync, { par_quicksort(self.as_parallel_slice_mut(), |a, b| { compare(a, b) == Ordering::Less }); } /// Sorts the slice in parallel with a key extraction function, but may not preserve the order /// of equal elements. /// /// This sort is unstable (i.e. may reorder equal elements), in-place (i.e. does not allocate), /// and `O(n log n)` worst-case. /// /// # Current implementation /// /// The current algorithm is based on Orson Peters' [pattern-defeating quicksort][pdqsort], /// which is a quicksort variant designed to be very fast on certain kinds of patterns, /// sometimes achieving linear time. It is randomized but deterministic, and falls back to /// heapsort on degenerate inputs. /// /// It is generally faster than stable sorting, except in a few special cases, e.g. when the /// slice consists of several concatenated sorted sequences. /// /// All quicksorts work in two stages: partitioning into two halves followed by recursive /// calls. The partitioning phase is sequential, but the two recursive calls are performed in /// parallel. /// /// [pdqsort]: https://github.com/orlp/pdqsort /// /// # Examples /// /// ``` /// use rayon::prelude::*; /// /// let mut v = [-5i32, 4, 1, -3, 2]; /// /// v.par_sort_unstable_by_key(|k| k.abs()); /// assert_eq!(v, [1, 2, -3, 4, -5]); /// ``` fn par_sort_unstable_by_key(&mut self, f: F) where B: Ord, F: Fn(&T) -> B + Sync, { par_quicksort(self.as_parallel_slice_mut(), |a, b| f(a).lt(&f(b))); } } impl ParallelSliceMut for [T] { #[inline] fn as_parallel_slice_mut(&mut self) -> &mut [T] { self } } impl<'data, T: Sync + 'data> IntoParallelIterator for &'data [T] { type Item = &'data T; type Iter = Iter<'data, T>; fn into_par_iter(self) -> Self::Iter { Iter { slice: self } } } impl<'data, T: Sync + 'data> IntoParallelIterator for &'data Vec { type Item = &'data T; type Iter = Iter<'data, T>; fn into_par_iter(self) -> Self::Iter { Iter { slice: self } } } impl<'data, T: Send + 'data> IntoParallelIterator for &'data mut [T] { type Item = &'data mut T; type Iter = IterMut<'data, T>; fn into_par_iter(self) -> Self::Iter { IterMut { slice: self } } } impl<'data, T: Send + 'data> IntoParallelIterator for &'data mut Vec { type Item = &'data mut T; type Iter = IterMut<'data, T>; fn into_par_iter(self) -> Self::Iter { IterMut { slice: self } } } /// Parallel iterator over immutable items in a slice #[derive(Debug)] pub struct Iter<'data, T: Sync> { slice: &'data [T], } impl<'data, T: Sync> Clone for Iter<'data, T> { fn clone(&self) -> Self { Iter { ..*self } } } impl<'data, T: Sync + 'data> ParallelIterator for Iter<'data, T> { type Item = &'data T; fn drive_unindexed(self, consumer: C) -> C::Result where C: UnindexedConsumer, { bridge(self, consumer) } fn opt_len(&self) -> Option { Some(self.len()) } } impl<'data, T: Sync + 'data> IndexedParallelIterator for Iter<'data, T> { fn drive(self, consumer: C) -> C::Result where C: Consumer, { bridge(self, consumer) } fn len(&self) -> usize { self.slice.len() } fn with_producer(self, callback: CB) -> CB::Output where CB: ProducerCallback, { callback.callback(IterProducer { slice: self.slice }) } } struct IterProducer<'data, T: Sync> { slice: &'data [T], } impl<'data, T: 'data + Sync> Producer for IterProducer<'data, T> { type Item = &'data T; type IntoIter = ::std::slice::Iter<'data, T>; fn into_iter(self) -> Self::IntoIter { self.slice.iter() } fn split_at(self, index: usize) -> (Self, Self) { let (left, right) = self.slice.split_at(index); (IterProducer { slice: left }, IterProducer { slice: right }) } } /// Parallel iterator over immutable non-overlapping chunks of a slice #[derive(Debug)] pub struct Chunks<'data, T: Sync> { chunk_size: usize, slice: &'data [T], } impl<'data, T: Sync> Clone for Chunks<'data, T> { fn clone(&self) -> Self { Chunks { ..*self } } } impl<'data, T: Sync + 'data> ParallelIterator for Chunks<'data, T> { type Item = &'data [T]; fn drive_unindexed(self, consumer: C) -> C::Result where C: UnindexedConsumer, { bridge(self, consumer) } fn opt_len(&self) -> Option { Some(self.len()) } } impl<'data, T: Sync + 'data> IndexedParallelIterator for Chunks<'data, T> { fn drive(self, consumer: C) -> C::Result where C: Consumer, { bridge(self, consumer) } fn len(&self) -> usize { div_round_up(self.slice.len(), self.chunk_size) } fn with_producer(self, callback: CB) -> CB::Output where CB: ProducerCallback, { callback.callback(ChunksProducer { chunk_size: self.chunk_size, slice: self.slice, }) } } struct ChunksProducer<'data, T: Sync> { chunk_size: usize, slice: &'data [T], } impl<'data, T: 'data + Sync> Producer for ChunksProducer<'data, T> { type Item = &'data [T]; type IntoIter = ::std::slice::Chunks<'data, T>; fn into_iter(self) -> Self::IntoIter { self.slice.chunks(self.chunk_size) } fn split_at(self, index: usize) -> (Self, Self) { let elem_index = cmp::min(index * self.chunk_size, self.slice.len()); let (left, right) = self.slice.split_at(elem_index); ( ChunksProducer { chunk_size: self.chunk_size, slice: left, }, ChunksProducer { chunk_size: self.chunk_size, slice: right, }, ) } } /// Parallel iterator over immutable non-overlapping chunks of a slice #[derive(Debug)] pub struct ChunksExact<'data, T: Sync> { chunk_size: usize, slice: &'data [T], rem: &'data [T], } impl<'data, T: Sync> ChunksExact<'data, T> { /// Return the remainder of the original slice that is not going to be /// returned by the iterator. The returned slice has at most `chunk_size-1` /// elements. pub fn remainder(&self) -> &'data [T] { self.rem } } impl<'data, T: Sync> Clone for ChunksExact<'data, T> { fn clone(&self) -> Self { ChunksExact { ..*self } } } impl<'data, T: Sync + 'data> ParallelIterator for ChunksExact<'data, T> { type Item = &'data [T]; fn drive_unindexed(self, consumer: C) -> C::Result where C: UnindexedConsumer, { bridge(self, consumer) } fn opt_len(&self) -> Option { Some(self.len()) } } impl<'data, T: Sync + 'data> IndexedParallelIterator for ChunksExact<'data, T> { fn drive(self, consumer: C) -> C::Result where C: Consumer, { bridge(self, consumer) } fn len(&self) -> usize { self.slice.len() / self.chunk_size } fn with_producer(self, callback: CB) -> CB::Output where CB: ProducerCallback, { callback.callback(ChunksExactProducer { chunk_size: self.chunk_size, slice: self.slice, }) } } struct ChunksExactProducer<'data, T: Sync> { chunk_size: usize, slice: &'data [T], } impl<'data, T: 'data + Sync> Producer for ChunksExactProducer<'data, T> { type Item = &'data [T]; type IntoIter = ::std::slice::ChunksExact<'data, T>; fn into_iter(self) -> Self::IntoIter { self.slice.chunks_exact(self.chunk_size) } fn split_at(self, index: usize) -> (Self, Self) { let elem_index = index * self.chunk_size; let (left, right) = self.slice.split_at(elem_index); ( ChunksExactProducer { chunk_size: self.chunk_size, slice: left, }, ChunksExactProducer { chunk_size: self.chunk_size, slice: right, }, ) } } /// Parallel iterator over immutable overlapping windows of a slice #[derive(Debug)] pub struct Windows<'data, T: Sync> { window_size: usize, slice: &'data [T], } impl<'data, T: Sync> Clone for Windows<'data, T> { fn clone(&self) -> Self { Windows { ..*self } } } impl<'data, T: Sync + 'data> ParallelIterator for Windows<'data, T> { type Item = &'data [T]; fn drive_unindexed(self, consumer: C) -> C::Result where C: UnindexedConsumer, { bridge(self, consumer) } fn opt_len(&self) -> Option { Some(self.len()) } } impl<'data, T: Sync + 'data> IndexedParallelIterator for Windows<'data, T> { fn drive(self, consumer: C) -> C::Result where C: Consumer, { bridge(self, consumer) } fn len(&self) -> usize { assert!(self.window_size >= 1); self.slice.len().saturating_sub(self.window_size - 1) } fn with_producer(self, callback: CB) -> CB::Output where CB: ProducerCallback, { callback.callback(WindowsProducer { window_size: self.window_size, slice: self.slice, }) } } struct WindowsProducer<'data, T: Sync> { window_size: usize, slice: &'data [T], } impl<'data, T: 'data + Sync> Producer for WindowsProducer<'data, T> { type Item = &'data [T]; type IntoIter = ::std::slice::Windows<'data, T>; fn into_iter(self) -> Self::IntoIter { self.slice.windows(self.window_size) } fn split_at(self, index: usize) -> (Self, Self) { let left_index = cmp::min(self.slice.len(), index + (self.window_size - 1)); let left = &self.slice[..left_index]; let right = &self.slice[index..]; ( WindowsProducer { window_size: self.window_size, slice: left, }, WindowsProducer { window_size: self.window_size, slice: right, }, ) } } /// Parallel iterator over mutable items in a slice #[derive(Debug)] pub struct IterMut<'data, T: Send> { slice: &'data mut [T], } impl<'data, T: Send + 'data> ParallelIterator for IterMut<'data, T> { type Item = &'data mut T; fn drive_unindexed(self, consumer: C) -> C::Result where C: UnindexedConsumer, { bridge(self, consumer) } fn opt_len(&self) -> Option { Some(self.len()) } } impl<'data, T: Send + 'data> IndexedParallelIterator for IterMut<'data, T> { fn drive(self, consumer: C) -> C::Result where C: Consumer, { bridge(self, consumer) } fn len(&self) -> usize { self.slice.len() } fn with_producer(self, callback: CB) -> CB::Output where CB: ProducerCallback, { callback.callback(IterMutProducer { slice: self.slice }) } } struct IterMutProducer<'data, T: Send> { slice: &'data mut [T], } impl<'data, T: 'data + Send> Producer for IterMutProducer<'data, T> { type Item = &'data mut T; type IntoIter = ::std::slice::IterMut<'data, T>; fn into_iter(self) -> Self::IntoIter { self.slice.iter_mut() } fn split_at(self, index: usize) -> (Self, Self) { let (left, right) = self.slice.split_at_mut(index); ( IterMutProducer { slice: left }, IterMutProducer { slice: right }, ) } } /// Parallel iterator over mutable non-overlapping chunks of a slice #[derive(Debug)] pub struct ChunksMut<'data, T: Send> { chunk_size: usize, slice: &'data mut [T], } impl<'data, T: Send + 'data> ParallelIterator for ChunksMut<'data, T> { type Item = &'data mut [T]; fn drive_unindexed(self, consumer: C) -> C::Result where C: UnindexedConsumer, { bridge(self, consumer) } fn opt_len(&self) -> Option { Some(self.len()) } } impl<'data, T: Send + 'data> IndexedParallelIterator for ChunksMut<'data, T> { fn drive(self, consumer: C) -> C::Result where C: Consumer, { bridge(self, consumer) } fn len(&self) -> usize { div_round_up(self.slice.len(), self.chunk_size) } fn with_producer(self, callback: CB) -> CB::Output where CB: ProducerCallback, { callback.callback(ChunksMutProducer { chunk_size: self.chunk_size, slice: self.slice, }) } } struct ChunksMutProducer<'data, T: Send> { chunk_size: usize, slice: &'data mut [T], } impl<'data, T: 'data + Send> Producer for ChunksMutProducer<'data, T> { type Item = &'data mut [T]; type IntoIter = ::std::slice::ChunksMut<'data, T>; fn into_iter(self) -> Self::IntoIter { self.slice.chunks_mut(self.chunk_size) } fn split_at(self, index: usize) -> (Self, Self) { let elem_index = cmp::min(index * self.chunk_size, self.slice.len()); let (left, right) = self.slice.split_at_mut(elem_index); ( ChunksMutProducer { chunk_size: self.chunk_size, slice: left, }, ChunksMutProducer { chunk_size: self.chunk_size, slice: right, }, ) } } /// Parallel iterator over mutable non-overlapping chunks of a slice #[derive(Debug)] pub struct ChunksExactMut<'data, T: Send> { chunk_size: usize, slice: &'data mut [T], rem: &'data mut [T], } impl<'data, T: Send> ChunksExactMut<'data, T> { /// Return the remainder of the original slice that is not going to be /// returned by the iterator. The returned slice has at most `chunk_size-1` /// elements. /// /// Note that this has to consume `self` to return the original lifetime of /// the data, which prevents this from actually being used as a parallel /// iterator since that also consumes. This method is provided for parity /// with `std::iter::ChunksExactMut`, but consider calling `remainder()` or /// `take_remainder()` as alternatives. pub fn into_remainder(self) -> &'data mut [T] { self.rem } /// Return the remainder of the original slice that is not going to be /// returned by the iterator. The returned slice has at most `chunk_size-1` /// elements. /// /// Consider `take_remainder()` if you need access to the data with its /// original lifetime, rather than borrowing through `&mut self` here. pub fn remainder(&mut self) -> &mut [T] { self.rem } /// Return the remainder of the original slice that is not going to be /// returned by the iterator. The returned slice has at most `chunk_size-1` /// elements. Subsequent calls will return an empty slice. pub fn take_remainder(&mut self) -> &'data mut [T] { std::mem::replace(&mut self.rem, &mut []) } } impl<'data, T: Send + 'data> ParallelIterator for ChunksExactMut<'data, T> { type Item = &'data mut [T]; fn drive_unindexed(self, consumer: C) -> C::Result where C: UnindexedConsumer, { bridge(self, consumer) } fn opt_len(&self) -> Option { Some(self.len()) } } impl<'data, T: Send + 'data> IndexedParallelIterator for ChunksExactMut<'data, T> { fn drive(self, consumer: C) -> C::Result where C: Consumer, { bridge(self, consumer) } fn len(&self) -> usize { self.slice.len() / self.chunk_size } fn with_producer(self, callback: CB) -> CB::Output where CB: ProducerCallback, { callback.callback(ChunksExactMutProducer { chunk_size: self.chunk_size, slice: self.slice, }) } } struct ChunksExactMutProducer<'data, T: Send> { chunk_size: usize, slice: &'data mut [T], } impl<'data, T: 'data + Send> Producer for ChunksExactMutProducer<'data, T> { type Item = &'data mut [T]; type IntoIter = ::std::slice::ChunksExactMut<'data, T>; fn into_iter(self) -> Self::IntoIter { self.slice.chunks_exact_mut(self.chunk_size) } fn split_at(self, index: usize) -> (Self, Self) { let elem_index = index * self.chunk_size; let (left, right) = self.slice.split_at_mut(elem_index); ( ChunksExactMutProducer { chunk_size: self.chunk_size, slice: left, }, ChunksExactMutProducer { chunk_size: self.chunk_size, slice: right, }, ) } } /// Parallel iterator over slices separated by a predicate pub struct Split<'data, T, P> { slice: &'data [T], separator: P, } impl<'data, T, P: Clone> Clone for Split<'data, T, P> { fn clone(&self) -> Self { Split { separator: self.separator.clone(), ..*self } } } impl<'data, T: Debug, P> Debug for Split<'data, T, P> { fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { f.debug_struct("Split").field("slice", &self.slice).finish() } } impl<'data, T, P> ParallelIterator for Split<'data, T, P> where P: Fn(&T) -> bool + Sync + Send, T: Sync, { type Item = &'data [T]; fn drive_unindexed(self, consumer: C) -> C::Result where C: UnindexedConsumer, { let producer = SplitProducer::new(self.slice, &self.separator); bridge_unindexed(producer, consumer) } } /// Implement support for `SplitProducer`. impl<'data, T, P> Fissile

for &'data [T] where P: Fn(&T) -> bool, { fn length(&self) -> usize { self.len() } fn midpoint(&self, end: usize) -> usize { end / 2 } fn find(&self, separator: &P, start: usize, end: usize) -> Option { self[start..end].iter().position(separator) } fn rfind(&self, separator: &P, end: usize) -> Option { self[..end].iter().rposition(separator) } fn split_once(self, index: usize) -> (Self, Self) { let (left, right) = self.split_at(index); (left, &right[1..]) // skip the separator } fn fold_splits(self, separator: &P, folder: F, skip_last: bool) -> F where F: Folder, Self: Send, { let mut split = self.split(separator); if skip_last { split.next_back(); } folder.consume_iter(split) } } /// Parallel iterator over mutable slices separated by a predicate pub struct SplitMut<'data, T, P> { slice: &'data mut [T], separator: P, } impl<'data, T: Debug, P> Debug for SplitMut<'data, T, P> { fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { f.debug_struct("SplitMut") .field("slice", &self.slice) .finish() } } impl<'data, T, P> ParallelIterator for SplitMut<'data, T, P> where P: Fn(&T) -> bool + Sync + Send, T: Send, { type Item = &'data mut [T]; fn drive_unindexed(self, consumer: C) -> C::Result where C: UnindexedConsumer, { let producer = SplitProducer::new(self.slice, &self.separator); bridge_unindexed(producer, consumer) } } /// Implement support for `SplitProducer`. impl<'data, T, P> Fissile

for &'data mut [T] where P: Fn(&T) -> bool, { fn length(&self) -> usize { self.len() } fn midpoint(&self, end: usize) -> usize { end / 2 } fn find(&self, separator: &P, start: usize, end: usize) -> Option { self[start..end].iter().position(separator) } fn rfind(&self, separator: &P, end: usize) -> Option { self[..end].iter().rposition(separator) } fn split_once(self, index: usize) -> (Self, Self) { let (left, right) = self.split_at_mut(index); (left, &mut right[1..]) // skip the separator } fn fold_splits(self, separator: &P, folder: F, skip_last: bool) -> F where F: Folder, Self: Send, { let mut split = self.split_mut(separator); if skip_last { split.next_back(); } folder.consume_iter(split) } }