Spaces:
Build error
Build error
/* | |
* SPDX-FileCopyrightText: Copyright (c) 2025 NVIDIA CORPORATION & AFFILIATES. All rights reserved. | |
* SPDX-License-Identifier: Apache-2.0 | |
* | |
* Licensed under the Apache License, Version 2.0 (the "License"); | |
* you may not use this file except in compliance with the License. | |
* You may obtain a copy of the License at | |
* | |
* http://www.apache.org/licenses/LICENSE-2.0 | |
* | |
* Unless required by applicable law or agreed to in writing, software | |
* distributed under the License is distributed on an "AS IS" BASIS, | |
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |
* See the License for the specific language governing permissions and | |
* limitations under the License. | |
*/ | |
// This file was taken from the tev image viewer and is re-released here | |
// under the NVIDIA Source Code License with permission from the author. | |
namespace ngp { | |
ThreadPool::ThreadPool() | |
: ThreadPool{std::thread::hardware_concurrency()} {} | |
ThreadPool::ThreadPool(size_t max_num_threads, bool force) { | |
if (!force) { | |
max_num_threads = std::min((size_t)std::thread::hardware_concurrency(), max_num_threads); | |
} | |
start_threads(max_num_threads); | |
} | |
ThreadPool::~ThreadPool() { | |
wait_until_queue_completed(); | |
shutdown_threads(m_threads.size()); | |
} | |
void ThreadPool::start_threads(size_t num) { | |
m_num_threads += num; | |
for (size_t i = m_threads.size(); i < m_num_threads; ++i) { | |
m_threads.emplace_back([this, i] { | |
while (true) { | |
std::unique_lock<std::mutex> lock{m_task_queue_mutex}; | |
// look for a work item | |
while (i < m_num_threads && m_task_queue.empty()) { | |
// if there are none, signal that the queue is completed | |
// and wait for notification of new work items. | |
m_task_queue_completed_condition.notify_all(); | |
m_worker_condition.wait(lock); | |
} | |
if (i >= m_num_threads) { | |
break; | |
} | |
std::function<void()> task{std::move(m_task_queue.front())}; | |
m_task_queue.pop_front(); | |
// Unlock the lock, so we can process the task without blocking other threads | |
lock.unlock(); | |
task(); | |
} | |
}); | |
} | |
} | |
void ThreadPool::shutdown_threads(size_t num) { | |
auto num_to_close = std::min(num, m_num_threads); | |
{ | |
std::lock_guard<std::mutex> lock{m_task_queue_mutex}; | |
m_num_threads -= num_to_close; | |
} | |
// Wake up all the threads to have them quit | |
m_worker_condition.notify_all(); | |
for (auto i = 0u; i < num_to_close; ++i) { | |
m_threads.back().join(); | |
m_threads.pop_back(); | |
} | |
} | |
void ThreadPool::set_n_threads(size_t num) { | |
if (m_num_threads > num) { | |
shutdown_threads(m_num_threads - num); | |
} else if (m_num_threads < num) { | |
start_threads(num - m_num_threads); | |
} | |
} | |
void ThreadPool::wait_until_queue_completed() { | |
std::unique_lock<std::mutex> lock{m_task_queue_mutex}; | |
m_task_queue_completed_condition.wait(lock, [this]() { return m_task_queue.empty(); }); | |
} | |
void ThreadPool::flush_queue() { | |
std::lock_guard<std::mutex> lock{m_task_queue_mutex}; | |
m_task_queue.clear(); | |
} | |
} | |