diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-17 12:20:29 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-17 12:20:29 +0000 |
commit | 631cd5845e8de329d0e227aaa707d7ea228b8f8f (patch) | |
tree | a1b87c8f8cad01cf18f7c5f57a08f102771ed303 /vendor/threadpool | |
parent | Adding debian version 1.69.0+dfsg1-1. (diff) | |
download | rustc-631cd5845e8de329d0e227aaa707d7ea228b8f8f.tar.xz rustc-631cd5845e8de329d0e227aaa707d7ea228b8f8f.zip |
Merging upstream version 1.70.0+dfsg1.
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'vendor/threadpool')
-rw-r--r-- | vendor/threadpool/.cargo-checksum.json | 1 | ||||
-rw-r--r-- | vendor/threadpool/CHANGES.md | 72 | ||||
-rw-r--r-- | vendor/threadpool/Cargo.toml | 27 | ||||
-rw-r--r-- | vendor/threadpool/LICENSE-APACHE | 201 | ||||
-rw-r--r-- | vendor/threadpool/LICENSE-MIT | 25 | ||||
-rw-r--r-- | vendor/threadpool/src/lib.rs | 1329 |
6 files changed, 1655 insertions, 0 deletions
diff --git a/vendor/threadpool/.cargo-checksum.json b/vendor/threadpool/.cargo-checksum.json new file mode 100644 index 000000000..a92778aa9 --- /dev/null +++ b/vendor/threadpool/.cargo-checksum.json @@ -0,0 +1 @@ +{"files":{"CHANGES.md":"b2c93083ee90cb9ec57185763a5cd2909fd6e048f90f652fb29e932c083337ee","Cargo.toml":"16d9b37ffdb58a7f77fd89742597eb35ab7aa2d0840f4f086a5580d95949849b","LICENSE-APACHE":"a60eea817514531668d7e00765731449fe14d059d3249e0bc93b36de45f759f2","LICENSE-MIT":"6485b8ed310d3f0340bf1ad1f47645069ce4069dcc6bb46c7d5c6faf41de1fdb","src/lib.rs":"ca973733471524ebca5c1c975f8a10a0c2394008b90e24db2fe37b0be7943b20"},"package":"d050e60b33d41c19108b32cea32164033a9013fe3b46cbd4457559bfbf77afaa"}
\ No newline at end of file diff --git a/vendor/threadpool/CHANGES.md b/vendor/threadpool/CHANGES.md new file mode 100644 index 000000000..1df287700 --- /dev/null +++ b/vendor/threadpool/CHANGES.md @@ -0,0 +1,72 @@ +# Changes + +## 1.8.1 + +* [Fix a typo](https://github.com/rust-threadpool/rust-threadpool/pull/107) + +## 1.8.0 + +* [Raise minimal rustc version to 1.13.0](https://github.com/rust-threadpool/rust-threadpool/pull/99) +* [Update num_cpus to 1.13.0](https://github.com/rust-threadpool/rust-threadpool/pull/105) + +## 1.7.1 + +* [Join waves](https://github.com/rust-threadpool/rust-threadpool/pull/89) + +## 1.7.0 + +* [Introduce `threadpool::Builder`](https://github.com/rust-threadpool/rust-threadpool/pull/83) +* [Add more hyperlinks to documentation](https://github.com/rust-threadpool/rust-threadpool/pull/87) +* [Add keywords and categories to Cargo.toml](https://github.com/rust-threadpool/rust-threadpool/pull/88) + +## 1.6.0 + +* [Implement `PartialEq` and `Eq` for `ThreadPool`](https://github.com/rust-threadpool/rust-threadpool/pull/81) + +## 1.5.0 + +* [Implement `Default` for `ThreadPool` use 'num_cpus' crate.](https://github.com/rust-threadpool/rust-threadpool/pull/72) + +## 1.4.1 + +* [Introduce `with_name`, deprecate `new_with_name`](https://github.com/rust-threadpool/rust-threadpool/pull/73) +* [Documentation improvements](https://github.com/rust-threadpool/rust-threadpool/pull/71) + +## 1.4.0 + +* [Implementation of the `join` operation](https://github.com/rust-threadpool/rust-threadpool/pull/63) + +## 1.3.2 + +* [Enable `#[deprecated]` doc, requires Rust 1.9](https://github.com/rust-threadpool/rust-threadpool/pull/38) + +## 1.3.1 + +* [Implement std::fmt::Debug for ThreadPool](https://github.com/rust-threadpool/rust-threadpool/pull/50) + +## 1.3.0 + +* [Add barrier sync example](https://github.com/rust-threadpool/rust-threadpool/pull/35) +* [Rename `threads` method/params to `num_threads`, deprecate old usage](https://github.com/rust-threadpool/rust-threadpool/pull/34) +* [Stop using deprecated `sleep_ms` function in tests](https://github.com/rust-threadpool/rust-threadpool/pull/33) + +## 1.2.0 + +* [New method to determine number of panicked threads](https://github.com/rust-threadpool/rust-threadpool/pull/31) + +## 1.1.1 + +* [Silence warning related to unused result](https://github.com/rust-threadpool/rust-threadpool/pull/30) +* [Minor doc improvements](https://github.com/rust-threadpool/rust-threadpool/pull/30) + +## 1.1.0 + +* [New constructor for specifying thread names for a thread pool](https://github.com/rust-threadpool/rust-threadpool/pull/28) + +## 1.0.2 + +* [Use atomic counters](https://github.com/rust-threadpool/rust-threadpool/pull/25) + +## 1.0.1 + +* [Switch active_count from Mutex to RwLock for more performance](https://github.com/rust-threadpool/rust-threadpool/pull/23) diff --git a/vendor/threadpool/Cargo.toml b/vendor/threadpool/Cargo.toml new file mode 100644 index 000000000..f1ace53ee --- /dev/null +++ b/vendor/threadpool/Cargo.toml @@ -0,0 +1,27 @@ +# THIS FILE IS AUTOMATICALLY GENERATED BY CARGO +# +# When uploading crates to the registry Cargo will automatically +# "normalize" Cargo.toml files for maximal compatibility +# with all versions of Cargo and also rewrite `path` dependencies +# to registry (e.g., crates.io) dependencies +# +# If you believe there's an error in this file please file an +# issue against the rust-lang/cargo repository. If you're +# editing this file be aware that the upstream Cargo.toml +# will likely look very different (and much more reasonable) + +[package] +name = "threadpool" +version = "1.8.1" +authors = ["The Rust Project Developers", "Corey Farwell <coreyf@rwell.org>", "Stefan Schindler <dns2utf8@estada.ch>"] +include = ["**/*.rs", "Cargo.toml", "CHANGES.md", "LICENSE-APACHE", "LICENSE-MIT"] +description = "A thread pool for running a number of jobs on a fixed set of worker threads.\n" +homepage = "https://github.com/rust-threadpool/rust-threadpool" +documentation = "https://docs.rs/threadpool" +readme = "README.md" +keywords = ["threadpool", "thread", "pool", "threading", "parallelism"] +categories = ["concurrency", "os"] +license = "MIT/Apache-2.0" +repository = "https://github.com/rust-threadpool/rust-threadpool" +[dependencies.num_cpus] +version = "1.13" diff --git a/vendor/threadpool/LICENSE-APACHE b/vendor/threadpool/LICENSE-APACHE new file mode 100644 index 000000000..16fe87b06 --- /dev/null +++ b/vendor/threadpool/LICENSE-APACHE @@ -0,0 +1,201 @@ + Apache License + Version 2.0, January 2004 + http://www.apache.org/licenses/ + +TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION + +1. Definitions. + + "License" shall mean the terms and conditions for use, reproduction, + and distribution as defined by Sections 1 through 9 of this document. + + "Licensor" shall mean the copyright owner or entity authorized by + the copyright owner that is granting the License. + + "Legal Entity" shall mean the union of the acting entity and all + other entities that control, are controlled by, or are under common + control with that entity. For the purposes of this definition, + "control" means (i) the power, direct or indirect, to cause the + direction or management of such entity, whether by contract or + otherwise, or (ii) ownership of fifty percent (50%) or more of the + outstanding shares, or (iii) beneficial ownership of such entity. + + "You" (or "Your") shall mean an individual or Legal Entity + exercising permissions granted by this License. + + "Source" form shall mean the preferred form for making modifications, + including but not limited to software source code, documentation + source, and configuration files. + + "Object" form shall mean any form resulting from mechanical + transformation or translation of a Source form, including but + not limited to compiled object code, generated documentation, + and conversions to other media types. + + "Work" shall mean the work of authorship, whether in Source or + Object form, made available under the License, as indicated by a + copyright notice that is included in or attached to the work + (an example is provided in the Appendix below). + + "Derivative Works" shall mean any work, whether in Source or Object + form, that is based on (or derived from) the Work and for which the + editorial revisions, annotations, elaborations, or other modifications + represent, as a whole, an original work of authorship. For the purposes + of this License, Derivative Works shall not include works that remain + separable from, or merely link (or bind by name) to the interfaces of, + the Work and Derivative Works thereof. + + "Contribution" shall mean any work of authorship, including + the original version of the Work and any modifications or additions + to that Work or Derivative Works thereof, that is intentionally + submitted to Licensor for inclusion in the Work by the copyright owner + or by an individual or Legal Entity authorized to submit on behalf of + the copyright owner. For the purposes of this definition, "submitted" + means any form of electronic, verbal, or written communication sent + to the Licensor or its representatives, including but not limited to + communication on electronic mailing lists, source code control systems, + and issue tracking systems that are managed by, or on behalf of, the + Licensor for the purpose of discussing and improving the Work, but + excluding communication that is conspicuously marked or otherwise + designated in writing by the copyright owner as "Not a Contribution." + + "Contributor" shall mean Licensor and any individual or Legal Entity + on behalf of whom a Contribution has been received by Licensor and + subsequently incorporated within the Work. + +2. Grant of Copyright License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + copyright license to reproduce, prepare Derivative Works of, + publicly display, publicly perform, sublicense, and distribute the + Work and such Derivative Works in Source or Object form. + +3. Grant of Patent License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + (except as stated in this section) patent license to make, have made, + use, offer to sell, sell, import, and otherwise transfer the Work, + where such license applies only to those patent claims licensable + by such Contributor that are necessarily infringed by their + Contribution(s) alone or by combination of their Contribution(s) + with the Work to which such Contribution(s) was submitted. If You + institute patent litigation against any entity (including a + cross-claim or counterclaim in a lawsuit) alleging that the Work + or a Contribution incorporated within the Work constitutes direct + or contributory patent infringement, then any patent licenses + granted to You under this License for that Work shall terminate + as of the date such litigation is filed. + +4. Redistribution. You may reproduce and distribute copies of the + Work or Derivative Works thereof in any medium, with or without + modifications, and in Source or Object form, provided that You + meet the following conditions: + + (a) You must give any other recipients of the Work or + Derivative Works a copy of this License; and + + (b) You must cause any modified files to carry prominent notices + stating that You changed the files; and + + (c) You must retain, in the Source form of any Derivative Works + that You distribute, all copyright, patent, trademark, and + attribution notices from the Source form of the Work, + excluding those notices that do not pertain to any part of + the Derivative Works; and + + (d) If the Work includes a "NOTICE" text file as part of its + distribution, then any Derivative Works that You distribute must + include a readable copy of the attribution notices contained + within such NOTICE file, excluding those notices that do not + pertain to any part of the Derivative Works, in at least one + of the following places: within a NOTICE text file distributed + as part of the Derivative Works; within the Source form or + documentation, if provided along with the Derivative Works; or, + within a display generated by the Derivative Works, if and + wherever such third-party notices normally appear. The contents + of the NOTICE file are for informational purposes only and + do not modify the License. You may add Your own attribution + notices within Derivative Works that You distribute, alongside + or as an addendum to the NOTICE text from the Work, provided + that such additional attribution notices cannot be construed + as modifying the License. + + You may add Your own copyright statement to Your modifications and + may provide additional or different license terms and conditions + for use, reproduction, or distribution of Your modifications, or + for any such Derivative Works as a whole, provided Your use, + reproduction, and distribution of the Work otherwise complies with + the conditions stated in this License. + +5. Submission of Contributions. Unless You explicitly state otherwise, + any Contribution intentionally submitted for inclusion in the Work + by You to the Licensor shall be under the terms and conditions of + this License, without any additional terms or conditions. + Notwithstanding the above, nothing herein shall supersede or modify + the terms of any separate license agreement you may have executed + with Licensor regarding such Contributions. + +6. Trademarks. This License does not grant permission to use the trade + names, trademarks, service marks, or product names of the Licensor, + except as required for reasonable and customary use in describing the + origin of the Work and reproducing the content of the NOTICE file. + +7. Disclaimer of Warranty. Unless required by applicable law or + agreed to in writing, Licensor provides the Work (and each + Contributor provides its Contributions) on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or + implied, including, without limitation, any warranties or conditions + of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A + PARTICULAR PURPOSE. You are solely responsible for determining the + appropriateness of using or redistributing the Work and assume any + risks associated with Your exercise of permissions under this License. + +8. Limitation of Liability. In no event and under no legal theory, + whether in tort (including negligence), contract, or otherwise, + unless required by applicable law (such as deliberate and grossly + negligent acts) or agreed to in writing, shall any Contributor be + liable to You for damages, including any direct, indirect, special, + incidental, or consequential damages of any character arising as a + result of this License or out of the use or inability to use the + Work (including but not limited to damages for loss of goodwill, + work stoppage, computer failure or malfunction, or any and all + other commercial damages or losses), even if such Contributor + has been advised of the possibility of such damages. + +9. Accepting Warranty or Additional Liability. While redistributing + the Work or Derivative Works thereof, You may choose to offer, + and charge a fee for, acceptance of support, warranty, indemnity, + or other liability obligations and/or rights consistent with this + License. However, in accepting such obligations, You may act only + on Your own behalf and on Your sole responsibility, not on behalf + of any other Contributor, and only if You agree to indemnify, + defend, and hold each Contributor harmless for any liability + incurred by, or claims asserted against, such Contributor by reason + of your accepting any such warranty or additional liability. + +END OF TERMS AND CONDITIONS + +APPENDIX: How to apply the Apache License to your work. + + To apply the Apache License to your work, attach the following + boilerplate notice, with the fields enclosed by brackets "[]" + replaced with your own identifying information. (Don't include + the brackets!) The text should be enclosed in the appropriate + comment syntax for the file format. We also recommend that a + file or class name and description of purpose be included on the + same "printed page" as the copyright notice for easier + identification within third-party archives. + +Copyright [yyyy] [name of copyright owner] + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. diff --git a/vendor/threadpool/LICENSE-MIT b/vendor/threadpool/LICENSE-MIT new file mode 100644 index 000000000..39d4bdb5a --- /dev/null +++ b/vendor/threadpool/LICENSE-MIT @@ -0,0 +1,25 @@ +Copyright (c) 2014 The Rust Project Developers + +Permission is hereby granted, free of charge, to any +person obtaining a copy of this software and associated +documentation files (the "Software"), to deal in the +Software without restriction, including without +limitation the rights to use, copy, modify, merge, +publish, distribute, sublicense, and/or sell copies of +the Software, and to permit persons to whom the Software +is furnished to do so, subject to the following +conditions: + +The above copyright notice and this permission notice +shall be included in all copies or substantial portions +of the Software. + +THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF +ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED +TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A +PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT +SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY +CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION +OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR +IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER +DEALINGS IN THE SOFTWARE. diff --git a/vendor/threadpool/src/lib.rs b/vendor/threadpool/src/lib.rs new file mode 100644 index 000000000..d8b8c29aa --- /dev/null +++ b/vendor/threadpool/src/lib.rs @@ -0,0 +1,1329 @@ +// Copyright 2014 The Rust Project Developers. See the COPYRIGHT +// file at the top-level directory of this distribution and at +// http://rust-lang.org/COPYRIGHT. +// +// Licensed under the Apache License, Version 2.0 <LICENSE-APACHE or +// http://www.apache.org/licenses/LICENSE-2.0> or the MIT license +// <LICENSE-MIT or http://opensource.org/licenses/MIT>, at your +// option. This file may not be copied, modified, or distributed +// except according to those terms. + +//! A thread pool used to execute functions in parallel. +//! +//! Spawns a specified number of worker threads and replenishes the pool if any worker threads +//! panic. +//! +//! # Examples +//! +//! ## Synchronized with a channel +//! +//! Every thread sends one message over the channel, which then is collected with the `take()`. +//! +//! ``` +//! use threadpool::ThreadPool; +//! use std::sync::mpsc::channel; +//! +//! let n_workers = 4; +//! let n_jobs = 8; +//! let pool = ThreadPool::new(n_workers); +//! +//! let (tx, rx) = channel(); +//! for _ in 0..n_jobs { +//! let tx = tx.clone(); +//! pool.execute(move|| { +//! tx.send(1).expect("channel will be there waiting for the pool"); +//! }); +//! } +//! +//! assert_eq!(rx.iter().take(n_jobs).fold(0, |a, b| a + b), 8); +//! ``` +//! +//! ## Synchronized with a barrier +//! +//! Keep in mind, if a barrier synchronizes more jobs than you have workers in the pool, +//! you will end up with a [deadlock](https://en.wikipedia.org/wiki/Deadlock) +//! at the barrier which is [not considered unsafe]( +//! https://doc.rust-lang.org/reference/behavior-not-considered-unsafe.html). +//! +//! ``` +//! use threadpool::ThreadPool; +//! use std::sync::{Arc, Barrier}; +//! use std::sync::atomic::{AtomicUsize, Ordering}; +//! +//! // create at least as many workers as jobs or you will deadlock yourself +//! let n_workers = 42; +//! let n_jobs = 23; +//! let pool = ThreadPool::new(n_workers); +//! let an_atomic = Arc::new(AtomicUsize::new(0)); +//! +//! assert!(n_jobs <= n_workers, "too many jobs, will deadlock"); +//! +//! // create a barrier that waits for all jobs plus the starter thread +//! let barrier = Arc::new(Barrier::new(n_jobs + 1)); +//! for _ in 0..n_jobs { +//! let barrier = barrier.clone(); +//! let an_atomic = an_atomic.clone(); +//! +//! pool.execute(move|| { +//! // do the heavy work +//! an_atomic.fetch_add(1, Ordering::Relaxed); +//! +//! // then wait for the other threads +//! barrier.wait(); +//! }); +//! } +//! +//! // wait for the threads to finish the work +//! barrier.wait(); +//! assert_eq!(an_atomic.load(Ordering::SeqCst), /* n_jobs = */ 23); +//! ``` + +extern crate num_cpus; + +use std::fmt; +use std::sync::atomic::{AtomicUsize, Ordering}; +use std::sync::mpsc::{channel, Receiver, Sender}; +use std::sync::{Arc, Condvar, Mutex}; +use std::thread; + +trait FnBox { + fn call_box(self: Box<Self>); +} + +impl<F: FnOnce()> FnBox for F { + fn call_box(self: Box<F>) { + (*self)() + } +} + +type Thunk<'a> = Box<FnBox + Send + 'a>; + +struct Sentinel<'a> { + shared_data: &'a Arc<ThreadPoolSharedData>, + active: bool, +} + +impl<'a> Sentinel<'a> { + fn new(shared_data: &'a Arc<ThreadPoolSharedData>) -> Sentinel<'a> { + Sentinel { + shared_data: shared_data, + active: true, + } + } + + /// Cancel and destroy this sentinel. + fn cancel(mut self) { + self.active = false; + } +} + +impl<'a> Drop for Sentinel<'a> { + fn drop(&mut self) { + if self.active { + self.shared_data.active_count.fetch_sub(1, Ordering::SeqCst); + if thread::panicking() { + self.shared_data.panic_count.fetch_add(1, Ordering::SeqCst); + } + self.shared_data.no_work_notify_all(); + spawn_in_pool(self.shared_data.clone()) + } + } +} + +/// [`ThreadPool`] factory, which can be used in order to configure the properties of the +/// [`ThreadPool`]. +/// +/// The three configuration options available: +/// +/// * `num_threads`: maximum number of threads that will be alive at any given moment by the built +/// [`ThreadPool`] +/// * `thread_name`: thread name for each of the threads spawned by the built [`ThreadPool`] +/// * `thread_stack_size`: stack size (in bytes) for each of the threads spawned by the built +/// [`ThreadPool`] +/// +/// [`ThreadPool`]: struct.ThreadPool.html +/// +/// # Examples +/// +/// Build a [`ThreadPool`] that uses a maximum of eight threads simultaneously and each thread has +/// a 8 MB stack size: +/// +/// ``` +/// let pool = threadpool::Builder::new() +/// .num_threads(8) +/// .thread_stack_size(8_000_000) +/// .build(); +/// ``` +#[derive(Clone, Default)] +pub struct Builder { + num_threads: Option<usize>, + thread_name: Option<String>, + thread_stack_size: Option<usize>, +} + +impl Builder { + /// Initiate a new [`Builder`]. + /// + /// [`Builder`]: struct.Builder.html + /// + /// # Examples + /// + /// ``` + /// let builder = threadpool::Builder::new(); + /// ``` + pub fn new() -> Builder { + Builder { + num_threads: None, + thread_name: None, + thread_stack_size: None, + } + } + + /// Set the maximum number of worker-threads that will be alive at any given moment by the built + /// [`ThreadPool`]. If not specified, defaults the number of threads to the number of CPUs. + /// + /// [`ThreadPool`]: struct.ThreadPool.html + /// + /// # Panics + /// + /// This method will panic if `num_threads` is 0. + /// + /// # Examples + /// + /// No more than eight threads will be alive simultaneously for this pool: + /// + /// ``` + /// use std::thread; + /// + /// let pool = threadpool::Builder::new() + /// .num_threads(8) + /// .build(); + /// + /// for _ in 0..100 { + /// pool.execute(|| { + /// println!("Hello from a worker thread!") + /// }) + /// } + /// ``` + pub fn num_threads(mut self, num_threads: usize) -> Builder { + assert!(num_threads > 0); + self.num_threads = Some(num_threads); + self + } + + /// Set the thread name for each of the threads spawned by the built [`ThreadPool`]. If not + /// specified, threads spawned by the thread pool will be unnamed. + /// + /// [`ThreadPool`]: struct.ThreadPool.html + /// + /// # Examples + /// + /// Each thread spawned by this pool will have the name "foo": + /// + /// ``` + /// use std::thread; + /// + /// let pool = threadpool::Builder::new() + /// .thread_name("foo".into()) + /// .build(); + /// + /// for _ in 0..100 { + /// pool.execute(|| { + /// assert_eq!(thread::current().name(), Some("foo")); + /// }) + /// } + /// ``` + pub fn thread_name(mut self, name: String) -> Builder { + self.thread_name = Some(name); + self + } + + /// Set the stack size (in bytes) for each of the threads spawned by the built [`ThreadPool`]. + /// If not specified, threads spawned by the threadpool will have a stack size [as specified in + /// the `std::thread` documentation][thread]. + /// + /// [thread]: https://doc.rust-lang.org/nightly/std/thread/index.html#stack-size + /// [`ThreadPool`]: struct.ThreadPool.html + /// + /// # Examples + /// + /// Each thread spawned by this pool will have a 4 MB stack: + /// + /// ``` + /// let pool = threadpool::Builder::new() + /// .thread_stack_size(4_000_000) + /// .build(); + /// + /// for _ in 0..100 { + /// pool.execute(|| { + /// println!("This thread has a 4 MB stack size!"); + /// }) + /// } + /// ``` + pub fn thread_stack_size(mut self, size: usize) -> Builder { + self.thread_stack_size = Some(size); + self + } + + /// Finalize the [`Builder`] and build the [`ThreadPool`]. + /// + /// [`Builder`]: struct.Builder.html + /// [`ThreadPool`]: struct.ThreadPool.html + /// + /// # Examples + /// + /// ``` + /// let pool = threadpool::Builder::new() + /// .num_threads(8) + /// .thread_stack_size(4_000_000) + /// .build(); + /// ``` + pub fn build(self) -> ThreadPool { + let (tx, rx) = channel::<Thunk<'static>>(); + + let num_threads = self.num_threads.unwrap_or_else(num_cpus::get); + + let shared_data = Arc::new(ThreadPoolSharedData { + name: self.thread_name, + job_receiver: Mutex::new(rx), + empty_condvar: Condvar::new(), + empty_trigger: Mutex::new(()), + join_generation: AtomicUsize::new(0), + queued_count: AtomicUsize::new(0), + active_count: AtomicUsize::new(0), + max_thread_count: AtomicUsize::new(num_threads), + panic_count: AtomicUsize::new(0), + stack_size: self.thread_stack_size, + }); + + // Threadpool threads + for _ in 0..num_threads { + spawn_in_pool(shared_data.clone()); + } + + ThreadPool { + jobs: tx, + shared_data: shared_data, + } + } +} + +struct ThreadPoolSharedData { + name: Option<String>, + job_receiver: Mutex<Receiver<Thunk<'static>>>, + empty_trigger: Mutex<()>, + empty_condvar: Condvar, + join_generation: AtomicUsize, + queued_count: AtomicUsize, + active_count: AtomicUsize, + max_thread_count: AtomicUsize, + panic_count: AtomicUsize, + stack_size: Option<usize>, +} + +impl ThreadPoolSharedData { + fn has_work(&self) -> bool { + self.queued_count.load(Ordering::SeqCst) > 0 || self.active_count.load(Ordering::SeqCst) > 0 + } + + /// Notify all observers joining this pool if there is no more work to do. + fn no_work_notify_all(&self) { + if !self.has_work() { + *self + .empty_trigger + .lock() + .expect("Unable to notify all joining threads"); + self.empty_condvar.notify_all(); + } + } +} + +/// Abstraction of a thread pool for basic parallelism. +pub struct ThreadPool { + // How the threadpool communicates with subthreads. + // + // This is the only such Sender, so when it is dropped all subthreads will + // quit. + jobs: Sender<Thunk<'static>>, + shared_data: Arc<ThreadPoolSharedData>, +} + +impl ThreadPool { + /// Creates a new thread pool capable of executing `num_threads` number of jobs concurrently. + /// + /// # Panics + /// + /// This function will panic if `num_threads` is 0. + /// + /// # Examples + /// + /// Create a new thread pool capable of executing four jobs concurrently: + /// + /// ``` + /// use threadpool::ThreadPool; + /// + /// let pool = ThreadPool::new(4); + /// ``` + pub fn new(num_threads: usize) -> ThreadPool { + Builder::new().num_threads(num_threads).build() + } + + /// Creates a new thread pool capable of executing `num_threads` number of jobs concurrently. + /// Each thread will have the [name][thread name] `name`. + /// + /// # Panics + /// + /// This function will panic if `num_threads` is 0. + /// + /// # Examples + /// + /// ```rust + /// use std::thread; + /// use threadpool::ThreadPool; + /// + /// let pool = ThreadPool::with_name("worker".into(), 2); + /// for _ in 0..2 { + /// pool.execute(|| { + /// assert_eq!( + /// thread::current().name(), + /// Some("worker") + /// ); + /// }); + /// } + /// pool.join(); + /// ``` + /// + /// [thread name]: https://doc.rust-lang.org/std/thread/struct.Thread.html#method.name + pub fn with_name(name: String, num_threads: usize) -> ThreadPool { + Builder::new() + .num_threads(num_threads) + .thread_name(name) + .build() + } + + /// **Deprecated: Use [`ThreadPool::with_name`](#method.with_name)** + #[inline(always)] + #[deprecated(since = "1.4.0", note = "use ThreadPool::with_name")] + pub fn new_with_name(name: String, num_threads: usize) -> ThreadPool { + Self::with_name(name, num_threads) + } + + /// Executes the function `job` on a thread in the pool. + /// + /// # Examples + /// + /// Execute four jobs on a thread pool that can run two jobs concurrently: + /// + /// ``` + /// use threadpool::ThreadPool; + /// + /// let pool = ThreadPool::new(2); + /// pool.execute(|| println!("hello")); + /// pool.execute(|| println!("world")); + /// pool.execute(|| println!("foo")); + /// pool.execute(|| println!("bar")); + /// pool.join(); + /// ``` + pub fn execute<F>(&self, job: F) + where + F: FnOnce() + Send + 'static, + { + self.shared_data.queued_count.fetch_add(1, Ordering::SeqCst); + self.jobs + .send(Box::new(job)) + .expect("ThreadPool::execute unable to send job into queue."); + } + + /// Returns the number of jobs waiting to executed in the pool. + /// + /// # Examples + /// + /// ``` + /// use threadpool::ThreadPool; + /// use std::time::Duration; + /// use std::thread::sleep; + /// + /// let pool = ThreadPool::new(2); + /// for _ in 0..10 { + /// pool.execute(|| { + /// sleep(Duration::from_secs(100)); + /// }); + /// } + /// + /// sleep(Duration::from_secs(1)); // wait for threads to start + /// assert_eq!(8, pool.queued_count()); + /// ``` + pub fn queued_count(&self) -> usize { + self.shared_data.queued_count.load(Ordering::Relaxed) + } + + /// Returns the number of currently active threads. + /// + /// # Examples + /// + /// ``` + /// use threadpool::ThreadPool; + /// use std::time::Duration; + /// use std::thread::sleep; + /// + /// let pool = ThreadPool::new(4); + /// for _ in 0..10 { + /// pool.execute(move || { + /// sleep(Duration::from_secs(100)); + /// }); + /// } + /// + /// sleep(Duration::from_secs(1)); // wait for threads to start + /// assert_eq!(4, pool.active_count()); + /// ``` + pub fn active_count(&self) -> usize { + self.shared_data.active_count.load(Ordering::SeqCst) + } + + /// Returns the maximum number of threads the pool will execute concurrently. + /// + /// # Examples + /// + /// ``` + /// use threadpool::ThreadPool; + /// + /// let mut pool = ThreadPool::new(4); + /// assert_eq!(4, pool.max_count()); + /// + /// pool.set_num_threads(8); + /// assert_eq!(8, pool.max_count()); + /// ``` + pub fn max_count(&self) -> usize { + self.shared_data.max_thread_count.load(Ordering::Relaxed) + } + + /// Returns the number of panicked threads over the lifetime of the pool. + /// + /// # Examples + /// + /// ``` + /// use threadpool::ThreadPool; + /// + /// let pool = ThreadPool::new(4); + /// for n in 0..10 { + /// pool.execute(move || { + /// // simulate a panic + /// if n % 2 == 0 { + /// panic!() + /// } + /// }); + /// } + /// pool.join(); + /// + /// assert_eq!(5, pool.panic_count()); + /// ``` + pub fn panic_count(&self) -> usize { + self.shared_data.panic_count.load(Ordering::Relaxed) + } + + /// **Deprecated: Use [`ThreadPool::set_num_threads`](#method.set_num_threads)** + #[deprecated(since = "1.3.0", note = "use ThreadPool::set_num_threads")] + pub fn set_threads(&mut self, num_threads: usize) { + self.set_num_threads(num_threads) + } + + /// Sets the number of worker-threads to use as `num_threads`. + /// Can be used to change the threadpool size during runtime. + /// Will not abort already running or waiting threads. + /// + /// # Panics + /// + /// This function will panic if `num_threads` is 0. + /// + /// # Examples + /// + /// ``` + /// use threadpool::ThreadPool; + /// use std::time::Duration; + /// use std::thread::sleep; + /// + /// let mut pool = ThreadPool::new(4); + /// for _ in 0..10 { + /// pool.execute(move || { + /// sleep(Duration::from_secs(100)); + /// }); + /// } + /// + /// sleep(Duration::from_secs(1)); // wait for threads to start + /// assert_eq!(4, pool.active_count()); + /// assert_eq!(6, pool.queued_count()); + /// + /// // Increase thread capacity of the pool + /// pool.set_num_threads(8); + /// + /// sleep(Duration::from_secs(1)); // wait for new threads to start + /// assert_eq!(8, pool.active_count()); + /// assert_eq!(2, pool.queued_count()); + /// + /// // Decrease thread capacity of the pool + /// // No active threads are killed + /// pool.set_num_threads(4); + /// + /// assert_eq!(8, pool.active_count()); + /// assert_eq!(2, pool.queued_count()); + /// ``` + pub fn set_num_threads(&mut self, num_threads: usize) { + assert!(num_threads >= 1); + let prev_num_threads = self + .shared_data + .max_thread_count + .swap(num_threads, Ordering::Release); + if let Some(num_spawn) = num_threads.checked_sub(prev_num_threads) { + // Spawn new threads + for _ in 0..num_spawn { + spawn_in_pool(self.shared_data.clone()); + } + } + } + + /// Block the current thread until all jobs in the pool have been executed. + /// + /// Calling `join` on an empty pool will cause an immediate return. + /// `join` may be called from multiple threads concurrently. + /// A `join` is an atomic point in time. All threads joining before the join + /// event will exit together even if the pool is processing new jobs by the + /// time they get scheduled. + /// + /// Calling `join` from a thread within the pool will cause a deadlock. This + /// behavior is considered safe. + /// + /// # Examples + /// + /// ``` + /// use threadpool::ThreadPool; + /// use std::sync::Arc; + /// use std::sync::atomic::{AtomicUsize, Ordering}; + /// + /// let pool = ThreadPool::new(8); + /// let test_count = Arc::new(AtomicUsize::new(0)); + /// + /// for _ in 0..42 { + /// let test_count = test_count.clone(); + /// pool.execute(move || { + /// test_count.fetch_add(1, Ordering::Relaxed); + /// }); + /// } + /// + /// pool.join(); + /// assert_eq!(42, test_count.load(Ordering::Relaxed)); + /// ``` + pub fn join(&self) { + // fast path requires no mutex + if self.shared_data.has_work() == false { + return (); + } + + let generation = self.shared_data.join_generation.load(Ordering::SeqCst); + let mut lock = self.shared_data.empty_trigger.lock().unwrap(); + + while generation == self.shared_data.join_generation.load(Ordering::Relaxed) + && self.shared_data.has_work() + { + lock = self.shared_data.empty_condvar.wait(lock).unwrap(); + } + + // increase generation if we are the first thread to come out of the loop + self.shared_data.join_generation.compare_and_swap( + generation, + generation.wrapping_add(1), + Ordering::SeqCst, + ); + } +} + +impl Clone for ThreadPool { + /// Cloning a pool will create a new handle to the pool. + /// The behavior is similar to [Arc](https://doc.rust-lang.org/stable/std/sync/struct.Arc.html). + /// + /// We could for example submit jobs from multiple threads concurrently. + /// + /// ``` + /// use threadpool::ThreadPool; + /// use std::thread; + /// use std::sync::mpsc::channel; + /// + /// let pool = ThreadPool::with_name("clone example".into(), 2); + /// + /// let results = (0..2) + /// .map(|i| { + /// let pool = pool.clone(); + /// thread::spawn(move || { + /// let (tx, rx) = channel(); + /// for i in 1..12 { + /// let tx = tx.clone(); + /// pool.execute(move || { + /// tx.send(i).expect("channel will be waiting"); + /// }); + /// } + /// drop(tx); + /// if i == 0 { + /// rx.iter().fold(0, |accumulator, element| accumulator + element) + /// } else { + /// rx.iter().fold(1, |accumulator, element| accumulator * element) + /// } + /// }) + /// }) + /// .map(|join_handle| join_handle.join().expect("collect results from threads")) + /// .collect::<Vec<usize>>(); + /// + /// assert_eq!(vec![66, 39916800], results); + /// ``` + fn clone(&self) -> ThreadPool { + ThreadPool { + jobs: self.jobs.clone(), + shared_data: self.shared_data.clone(), + } + } +} + +/// Create a thread pool with one thread per CPU. +/// On machines with hyperthreading, +/// this will create one thread per hyperthread. +impl Default for ThreadPool { + fn default() -> Self { + ThreadPool::new(num_cpus::get()) + } +} + +impl fmt::Debug for ThreadPool { + fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { + f.debug_struct("ThreadPool") + .field("name", &self.shared_data.name) + .field("queued_count", &self.queued_count()) + .field("active_count", &self.active_count()) + .field("max_count", &self.max_count()) + .finish() + } +} + +impl PartialEq for ThreadPool { + /// Check if you are working with the same pool + /// + /// ``` + /// use threadpool::ThreadPool; + /// + /// let a = ThreadPool::new(2); + /// let b = ThreadPool::new(2); + /// + /// assert_eq!(a, a); + /// assert_eq!(b, b); + /// + /// # // TODO: change this to assert_ne in the future + /// assert!(a != b); + /// assert!(b != a); + /// ``` + fn eq(&self, other: &ThreadPool) -> bool { + let a: &ThreadPoolSharedData = &*self.shared_data; + let b: &ThreadPoolSharedData = &*other.shared_data; + a as *const ThreadPoolSharedData == b as *const ThreadPoolSharedData + // with rust 1.17 and late: + // Arc::ptr_eq(&self.shared_data, &other.shared_data) + } +} +impl Eq for ThreadPool {} + +fn spawn_in_pool(shared_data: Arc<ThreadPoolSharedData>) { + let mut builder = thread::Builder::new(); + if let Some(ref name) = shared_data.name { + builder = builder.name(name.clone()); + } + if let Some(ref stack_size) = shared_data.stack_size { + builder = builder.stack_size(stack_size.to_owned()); + } + builder + .spawn(move || { + // Will spawn a new thread on panic unless it is cancelled. + let sentinel = Sentinel::new(&shared_data); + + loop { + // Shutdown this thread if the pool has become smaller + let thread_counter_val = shared_data.active_count.load(Ordering::Acquire); + let max_thread_count_val = shared_data.max_thread_count.load(Ordering::Relaxed); + if thread_counter_val >= max_thread_count_val { + break; + } + let message = { + // Only lock jobs for the time it takes + // to get a job, not run it. + let lock = shared_data + .job_receiver + .lock() + .expect("Worker thread unable to lock job_receiver"); + lock.recv() + }; + + let job = match message { + Ok(job) => job, + // The ThreadPool was dropped. + Err(..) => break, + }; + // Do not allow IR around the job execution + shared_data.active_count.fetch_add(1, Ordering::SeqCst); + shared_data.queued_count.fetch_sub(1, Ordering::SeqCst); + + job.call_box(); + + shared_data.active_count.fetch_sub(1, Ordering::SeqCst); + shared_data.no_work_notify_all(); + } + + sentinel.cancel(); + }) + .unwrap(); +} + +#[cfg(test)] +mod test { + use super::{Builder, ThreadPool}; + use std::sync::atomic::{AtomicUsize, Ordering}; + use std::sync::mpsc::{channel, sync_channel}; + use std::sync::{Arc, Barrier}; + use std::thread::{self, sleep}; + use std::time::Duration; + + const TEST_TASKS: usize = 4; + + #[test] + fn test_set_num_threads_increasing() { + let new_thread_amount = TEST_TASKS + 8; + let mut pool = ThreadPool::new(TEST_TASKS); + for _ in 0..TEST_TASKS { + pool.execute(move || sleep(Duration::from_secs(23))); + } + sleep(Duration::from_secs(1)); + assert_eq!(pool.active_count(), TEST_TASKS); + + pool.set_num_threads(new_thread_amount); + + for _ in 0..(new_thread_amount - TEST_TASKS) { + pool.execute(move || sleep(Duration::from_secs(23))); + } + sleep(Duration::from_secs(1)); + assert_eq!(pool.active_count(), new_thread_amount); + + pool.join(); + } + + #[test] + fn test_set_num_threads_decreasing() { + let new_thread_amount = 2; + let mut pool = ThreadPool::new(TEST_TASKS); + for _ in 0..TEST_TASKS { + pool.execute(move || { + assert_eq!(1, 1); + }); + } + pool.set_num_threads(new_thread_amount); + for _ in 0..new_thread_amount { + pool.execute(move || sleep(Duration::from_secs(23))); + } + sleep(Duration::from_secs(1)); + assert_eq!(pool.active_count(), new_thread_amount); + + pool.join(); + } + + #[test] + fn test_active_count() { + let pool = ThreadPool::new(TEST_TASKS); + for _ in 0..2 * TEST_TASKS { + pool.execute(move || loop { + sleep(Duration::from_secs(10)) + }); + } + sleep(Duration::from_secs(1)); + let active_count = pool.active_count(); + assert_eq!(active_count, TEST_TASKS); + let initialized_count = pool.max_count(); + assert_eq!(initialized_count, TEST_TASKS); + } + + #[test] + fn test_works() { + let pool = ThreadPool::new(TEST_TASKS); + + let (tx, rx) = channel(); + for _ in 0..TEST_TASKS { + let tx = tx.clone(); + pool.execute(move || { + tx.send(1).unwrap(); + }); + } + + assert_eq!(rx.iter().take(TEST_TASKS).fold(0, |a, b| a + b), TEST_TASKS); + } + + #[test] + #[should_panic] + fn test_zero_tasks_panic() { + ThreadPool::new(0); + } + + #[test] + fn test_recovery_from_subtask_panic() { + let pool = ThreadPool::new(TEST_TASKS); + + // Panic all the existing threads. + for _ in 0..TEST_TASKS { + pool.execute(move || panic!("Ignore this panic, it must!")); + } + pool.join(); + + assert_eq!(pool.panic_count(), TEST_TASKS); + + // Ensure new threads were spawned to compensate. + let (tx, rx) = channel(); + for _ in 0..TEST_TASKS { + let tx = tx.clone(); + pool.execute(move || { + tx.send(1).unwrap(); + }); + } + + assert_eq!(rx.iter().take(TEST_TASKS).fold(0, |a, b| a + b), TEST_TASKS); + } + + #[test] + fn test_should_not_panic_on_drop_if_subtasks_panic_after_drop() { + let pool = ThreadPool::new(TEST_TASKS); + let waiter = Arc::new(Barrier::new(TEST_TASKS + 1)); + + // Panic all the existing threads in a bit. + for _ in 0..TEST_TASKS { + let waiter = waiter.clone(); + pool.execute(move || { + waiter.wait(); + panic!("Ignore this panic, it should!"); + }); + } + + drop(pool); + + // Kick off the failure. + waiter.wait(); + } + + #[test] + fn test_massive_task_creation() { + let test_tasks = 4_200_000; + + let pool = ThreadPool::new(TEST_TASKS); + let b0 = Arc::new(Barrier::new(TEST_TASKS + 1)); + let b1 = Arc::new(Barrier::new(TEST_TASKS + 1)); + + let (tx, rx) = channel(); + + for i in 0..test_tasks { + let tx = tx.clone(); + let (b0, b1) = (b0.clone(), b1.clone()); + + pool.execute(move || { + // Wait until the pool has been filled once. + if i < TEST_TASKS { + b0.wait(); + // wait so the pool can be measured + b1.wait(); + } + + tx.send(1).is_ok(); + }); + } + + b0.wait(); + assert_eq!(pool.active_count(), TEST_TASKS); + b1.wait(); + + assert_eq!(rx.iter().take(test_tasks).fold(0, |a, b| a + b), test_tasks); + pool.join(); + + let atomic_active_count = pool.active_count(); + assert!( + atomic_active_count == 0, + "atomic_active_count: {}", + atomic_active_count + ); + } + + #[test] + fn test_shrink() { + let test_tasks_begin = TEST_TASKS + 2; + + let mut pool = ThreadPool::new(test_tasks_begin); + let b0 = Arc::new(Barrier::new(test_tasks_begin + 1)); + let b1 = Arc::new(Barrier::new(test_tasks_begin + 1)); + + for _ in 0..test_tasks_begin { + let (b0, b1) = (b0.clone(), b1.clone()); + pool.execute(move || { + b0.wait(); + b1.wait(); + }); + } + + let b2 = Arc::new(Barrier::new(TEST_TASKS + 1)); + let b3 = Arc::new(Barrier::new(TEST_TASKS + 1)); + + for _ in 0..TEST_TASKS { + let (b2, b3) = (b2.clone(), b3.clone()); + pool.execute(move || { + b2.wait(); + b3.wait(); + }); + } + + b0.wait(); + pool.set_num_threads(TEST_TASKS); + + assert_eq!(pool.active_count(), test_tasks_begin); + b1.wait(); + + b2.wait(); + assert_eq!(pool.active_count(), TEST_TASKS); + b3.wait(); + } + + #[test] + fn test_name() { + let name = "test"; + let mut pool = ThreadPool::with_name(name.to_owned(), 2); + let (tx, rx) = sync_channel(0); + + // initial thread should share the name "test" + for _ in 0..2 { + let tx = tx.clone(); + pool.execute(move || { + let name = thread::current().name().unwrap().to_owned(); + tx.send(name).unwrap(); + }); + } + + // new spawn thread should share the name "test" too. + pool.set_num_threads(3); + let tx_clone = tx.clone(); + pool.execute(move || { + let name = thread::current().name().unwrap().to_owned(); + tx_clone.send(name).unwrap(); + panic!(); + }); + + // recover thread should share the name "test" too. + pool.execute(move || { + let name = thread::current().name().unwrap().to_owned(); + tx.send(name).unwrap(); + }); + + for thread_name in rx.iter().take(4) { + assert_eq!(name, thread_name); + } + } + + #[test] + fn test_debug() { + let pool = ThreadPool::new(4); + let debug = format!("{:?}", pool); + assert_eq!( + debug, + "ThreadPool { name: None, queued_count: 0, active_count: 0, max_count: 4 }" + ); + + let pool = ThreadPool::with_name("hello".into(), 4); + let debug = format!("{:?}", pool); + assert_eq!( + debug, + "ThreadPool { name: Some(\"hello\"), queued_count: 0, active_count: 0, max_count: 4 }" + ); + + let pool = ThreadPool::new(4); + pool.execute(move || sleep(Duration::from_secs(5))); + sleep(Duration::from_secs(1)); + let debug = format!("{:?}", pool); + assert_eq!( + debug, + "ThreadPool { name: None, queued_count: 0, active_count: 1, max_count: 4 }" + ); + } + + #[test] + fn test_repeate_join() { + let pool = ThreadPool::with_name("repeate join test".into(), 8); + let test_count = Arc::new(AtomicUsize::new(0)); + + for _ in 0..42 { + let test_count = test_count.clone(); + pool.execute(move || { + sleep(Duration::from_secs(2)); + test_count.fetch_add(1, Ordering::Release); + }); + } + + println!("{:?}", pool); + pool.join(); + assert_eq!(42, test_count.load(Ordering::Acquire)); + + for _ in 0..42 { + let test_count = test_count.clone(); + pool.execute(move || { + sleep(Duration::from_secs(2)); + test_count.fetch_add(1, Ordering::Relaxed); + }); + } + pool.join(); + assert_eq!(84, test_count.load(Ordering::Relaxed)); + } + + #[test] + fn test_multi_join() { + use std::sync::mpsc::TryRecvError::*; + + // Toggle the following lines to debug the deadlock + fn error(_s: String) { + //use ::std::io::Write; + //let stderr = ::std::io::stderr(); + //let mut stderr = stderr.lock(); + //stderr.write(&_s.as_bytes()).is_ok(); + } + + let pool0 = ThreadPool::with_name("multi join pool0".into(), 4); + let pool1 = ThreadPool::with_name("multi join pool1".into(), 4); + let (tx, rx) = channel(); + + for i in 0..8 { + let pool1 = pool1.clone(); + let pool0_ = pool0.clone(); + let tx = tx.clone(); + pool0.execute(move || { + pool1.execute(move || { + error(format!("p1: {} -=- {:?}\n", i, pool0_)); + pool0_.join(); + error(format!("p1: send({})\n", i)); + tx.send(i).expect("send i from pool1 -> main"); + }); + error(format!("p0: {}\n", i)); + }); + } + drop(tx); + + assert_eq!(rx.try_recv(), Err(Empty)); + error(format!("{:?}\n{:?}\n", pool0, pool1)); + pool0.join(); + error(format!("pool0.join() complete =-= {:?}", pool1)); + pool1.join(); + error("pool1.join() complete\n".into()); + assert_eq!( + rx.iter().fold(0, |acc, i| acc + i), + 0 + 1 + 2 + 3 + 4 + 5 + 6 + 7 + ); + } + + #[test] + fn test_empty_pool() { + // Joining an empty pool must return imminently + let pool = ThreadPool::new(4); + + pool.join(); + + assert!(true); + } + + #[test] + fn test_no_fun_or_joy() { + // What happens when you keep adding jobs after a join + + fn sleepy_function() { + sleep(Duration::from_secs(6)); + } + + let pool = ThreadPool::with_name("no fun or joy".into(), 8); + + pool.execute(sleepy_function); + + let p_t = pool.clone(); + thread::spawn(move || { + (0..23).map(|_| p_t.execute(sleepy_function)).count(); + }); + + pool.join(); + } + + #[test] + fn test_clone() { + let pool = ThreadPool::with_name("clone example".into(), 2); + + // This batch of jobs will occupy the pool for some time + for _ in 0..6 { + pool.execute(move || { + sleep(Duration::from_secs(2)); + }); + } + + // The following jobs will be inserted into the pool in a random fashion + let t0 = { + let pool = pool.clone(); + thread::spawn(move || { + // wait for the first batch of tasks to finish + pool.join(); + + let (tx, rx) = channel(); + for i in 0..42 { + let tx = tx.clone(); + pool.execute(move || { + tx.send(i).expect("channel will be waiting"); + }); + } + drop(tx); + rx.iter() + .fold(0, |accumulator, element| accumulator + element) + }) + }; + let t1 = { + let pool = pool.clone(); + thread::spawn(move || { + // wait for the first batch of tasks to finish + pool.join(); + + let (tx, rx) = channel(); + for i in 1..12 { + let tx = tx.clone(); + pool.execute(move || { + tx.send(i).expect("channel will be waiting"); + }); + } + drop(tx); + rx.iter() + .fold(1, |accumulator, element| accumulator * element) + }) + }; + + assert_eq!( + 861, + t0.join() + .expect("thread 0 will return after calculating additions",) + ); + assert_eq!( + 39916800, + t1.join() + .expect("thread 1 will return after calculating multiplications",) + ); + } + + #[test] + fn test_sync_shared_data() { + fn assert_sync<T: Sync>() {} + assert_sync::<super::ThreadPoolSharedData>(); + } + + #[test] + fn test_send_shared_data() { + fn assert_send<T: Send>() {} + assert_send::<super::ThreadPoolSharedData>(); + } + + #[test] + fn test_send() { + fn assert_send<T: Send>() {} + assert_send::<ThreadPool>(); + } + + #[test] + fn test_cloned_eq() { + let a = ThreadPool::new(2); + + assert_eq!(a, a.clone()); + } + + #[test] + /// The scenario is joining threads should not be stuck once their wave + /// of joins has completed. So once one thread joining on a pool has + /// succeded other threads joining on the same pool must get out even if + /// the thread is used for other jobs while the first group is finishing + /// their join + /// + /// In this example this means the waiting threads will exit the join in + /// groups of four because the waiter pool has four workers. + fn test_join_wavesurfer() { + let n_cycles = 4; + let n_workers = 4; + let (tx, rx) = channel(); + let builder = Builder::new() + .num_threads(n_workers) + .thread_name("join wavesurfer".into()); + let p_waiter = builder.clone().build(); + let p_clock = builder.build(); + + let barrier = Arc::new(Barrier::new(3)); + let wave_clock = Arc::new(AtomicUsize::new(0)); + let clock_thread = { + let barrier = barrier.clone(); + let wave_clock = wave_clock.clone(); + thread::spawn(move || { + barrier.wait(); + for wave_num in 0..n_cycles { + wave_clock.store(wave_num, Ordering::SeqCst); + sleep(Duration::from_secs(1)); + } + }) + }; + + { + let barrier = barrier.clone(); + p_clock.execute(move || { + barrier.wait(); + // this sleep is for stabilisation on weaker platforms + sleep(Duration::from_millis(100)); + }); + } + + // prepare three waves of jobs + for i in 0..3 * n_workers { + let p_clock = p_clock.clone(); + let tx = tx.clone(); + let wave_clock = wave_clock.clone(); + p_waiter.execute(move || { + let now = wave_clock.load(Ordering::SeqCst); + p_clock.join(); + // submit jobs for the second wave + p_clock.execute(|| sleep(Duration::from_secs(1))); + let clock = wave_clock.load(Ordering::SeqCst); + tx.send((now, clock, i)).unwrap(); + }); + } + println!("all scheduled at {}", wave_clock.load(Ordering::SeqCst)); + barrier.wait(); + + p_clock.join(); + //p_waiter.join(); + + drop(tx); + let mut hist = vec![0; n_cycles]; + let mut data = vec![]; + for (now, after, i) in rx.iter() { + let mut dur = after - now; + if dur >= n_cycles - 1 { + dur = n_cycles - 1; + } + hist[dur] += 1; + + data.push((now, after, i)); + } + for (i, n) in hist.iter().enumerate() { + println!( + "\t{}: {} {}", + i, + n, + &*(0..*n).fold("".to_owned(), |s, _| s + "*") + ); + } + assert!(data.iter().all(|&(cycle, stop, i)| if i < n_workers { + cycle == stop + } else { + cycle < stop + })); + + clock_thread.join().unwrap(); + } +} |