| //===- llvm/Support/Parallel.cpp - Parallel algorithms --------------------===// | 
 | // | 
 | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. | 
 | // See https://llvm.org/LICENSE.txt for license information. | 
 | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception | 
 | // | 
 | //===----------------------------------------------------------------------===// | 
 |  | 
 | #include "llvm/Support/Parallel.h" | 
 | #include "llvm/Config/llvm-config.h" | 
 | #include "llvm/Support/ManagedStatic.h" | 
 |  | 
 | #if LLVM_ENABLE_THREADS | 
 |  | 
 | #include "llvm/Support/Threading.h" | 
 |  | 
 | #include <atomic> | 
 | #include <future> | 
 | #include <stack> | 
 | #include <thread> | 
 | #include <vector> | 
 |  | 
 | namespace llvm { | 
 | namespace parallel { | 
 | namespace detail { | 
 |  | 
 | namespace { | 
 |  | 
 | /// An abstract class that takes closures and runs them asynchronously. | 
 | class Executor { | 
 | public: | 
 |   virtual ~Executor() = default; | 
 |   virtual void add(std::function<void()> func) = 0; | 
 |  | 
 |   static Executor *getDefaultExecutor(); | 
 | }; | 
 |  | 
 | /// An implementation of an Executor that runs closures on a thread pool | 
 | ///   in filo order. | 
 | class ThreadPoolExecutor : public Executor { | 
 | public: | 
 |   explicit ThreadPoolExecutor(unsigned ThreadCount = hardware_concurrency()) { | 
 |     // Spawn all but one of the threads in another thread as spawning threads | 
 |     // can take a while. | 
 |     Threads.reserve(ThreadCount); | 
 |     Threads.resize(1); | 
 |     std::lock_guard<std::mutex> Lock(Mutex); | 
 |     Threads[0] = std::thread([&, ThreadCount] { | 
 |       for (unsigned i = 1; i < ThreadCount; ++i) { | 
 |         Threads.emplace_back([=] { work(); }); | 
 |         if (Stop) | 
 |           break; | 
 |       } | 
 |       ThreadsCreated.set_value(); | 
 |       work(); | 
 |     }); | 
 |   } | 
 |  | 
 |   void stop() { | 
 |     { | 
 |       std::lock_guard<std::mutex> Lock(Mutex); | 
 |       if (Stop) | 
 |         return; | 
 |       Stop = true; | 
 |     } | 
 |     Cond.notify_all(); | 
 |     ThreadsCreated.get_future().wait(); | 
 |   } | 
 |  | 
 |   ~ThreadPoolExecutor() override { | 
 |     stop(); | 
 |     std::thread::id CurrentThreadId = std::this_thread::get_id(); | 
 |     for (std::thread &T : Threads) | 
 |       if (T.get_id() == CurrentThreadId) | 
 |         T.detach(); | 
 |       else | 
 |         T.join(); | 
 |   } | 
 |  | 
 |   struct Deleter { | 
 |     static void call(void *Ptr) { ((ThreadPoolExecutor *)Ptr)->stop(); } | 
 |   }; | 
 |  | 
 |   void add(std::function<void()> F) override { | 
 |     { | 
 |       std::lock_guard<std::mutex> Lock(Mutex); | 
 |       WorkStack.push(F); | 
 |     } | 
 |     Cond.notify_one(); | 
 |   } | 
 |  | 
 | private: | 
 |   void work() { | 
 |     while (true) { | 
 |       std::unique_lock<std::mutex> Lock(Mutex); | 
 |       Cond.wait(Lock, [&] { return Stop || !WorkStack.empty(); }); | 
 |       if (Stop) | 
 |         break; | 
 |       auto Task = WorkStack.top(); | 
 |       WorkStack.pop(); | 
 |       Lock.unlock(); | 
 |       Task(); | 
 |     } | 
 |   } | 
 |  | 
 |   std::atomic<bool> Stop{false}; | 
 |   std::stack<std::function<void()>> WorkStack; | 
 |   std::mutex Mutex; | 
 |   std::condition_variable Cond; | 
 |   std::promise<void> ThreadsCreated; | 
 |   std::vector<std::thread> Threads; | 
 | }; | 
 |  | 
 | Executor *Executor::getDefaultExecutor() { | 
 |   // The ManagedStatic enables the ThreadPoolExecutor to be stopped via | 
 |   // llvm_shutdown() which allows a "clean" fast exit, e.g. via _exit(). This | 
 |   // stops the thread pool and waits for any worker thread creation to complete | 
 |   // but does not wait for the threads to finish. The wait for worker thread | 
 |   // creation to complete is important as it prevents intermittent crashes on | 
 |   // Windows due to a race condition between thread creation and process exit. | 
 |   // | 
 |   // The ThreadPoolExecutor will only be destroyed when the static unique_ptr to | 
 |   // it is destroyed, i.e. in a normal full exit. The ThreadPoolExecutor | 
 |   // destructor ensures it has been stopped and waits for worker threads to | 
 |   // finish. The wait is important as it prevents intermittent crashes on | 
 |   // Windows when the process is doing a full exit. | 
 |   // | 
 |   // The Windows crashes appear to only occur with the MSVC static runtimes and | 
 |   // are more frequent with the debug static runtime. | 
 |   // | 
 |   // This also prevents intermittent deadlocks on exit with the MinGW runtime. | 
 |   static ManagedStatic<ThreadPoolExecutor, object_creator<ThreadPoolExecutor>, | 
 |                        ThreadPoolExecutor::Deleter> | 
 |       ManagedExec; | 
 |   static std::unique_ptr<ThreadPoolExecutor> Exec(&(*ManagedExec)); | 
 |   return Exec.get(); | 
 | } | 
 | } // namespace | 
 |  | 
 | static std::atomic<int> TaskGroupInstances; | 
 |  | 
 | // Latch::sync() called by the dtor may cause one thread to block. If is a dead | 
 | // lock if all threads in the default executor are blocked. To prevent the dead | 
 | // lock, only allow the first TaskGroup to run tasks parallelly. In the scenario | 
 | // of nested parallel_for_each(), only the outermost one runs parallelly. | 
 | TaskGroup::TaskGroup() : Parallel(TaskGroupInstances++ == 0) {} | 
 | TaskGroup::~TaskGroup() { --TaskGroupInstances; } | 
 |  | 
 | void TaskGroup::spawn(std::function<void()> F) { | 
 |   if (Parallel) { | 
 |     L.inc(); | 
 |     Executor::getDefaultExecutor()->add([&, F] { | 
 |       F(); | 
 |       L.dec(); | 
 |     }); | 
 |   } else { | 
 |     F(); | 
 |   } | 
 | } | 
 |  | 
 | } // namespace detail | 
 | } // namespace parallel | 
 | } // namespace llvm | 
 | #endif // LLVM_ENABLE_THREADS |