third-party/tbb/rfcs/proposed/numa_support/README.md
In Non-Uniform Memory Access (NUMA) systems, the cost of memory accesses depends on the nearness of the processor to the memory resource on which the accessed data resides. While oneTBB has core support that enables developers to tune for Non-Uniform Memory Access (NUMA) systems, we believe this support can be simplified and improved to provide an improved user experience.
This RFC acts as an umbrella for sub-proposals that address four areas for improvement:
We expect that this draft proposal will spawn sub-proposals that will progress independently based on feedback and prioritization of the suggested features.
The features for NUMA tuning already available in the oneTBB 1.3 specification include:
tbb::info namespace [info_namespace]
std::vector<numa_node_id> numa_nodes()int default_concurrency(numa_node_id id = oneapi::tbb::task_arena::automatic)tbb::task_arena::constraints in [scheduler.task_arena]Below is the example based on existing oneTBB documentation that demonstrates the use of these APIs
to pin threads to different arenas to each of the NUMA nodes available on a system, submit work
across those task_arena objects and into associated task_group objects, and then wait for work
again using both the task_arena and task_group objects.
void constrain_for_numa_nodes() {
std::vector<tbb::numa_node_id> numa_nodes = tbb::info::numa_nodes();
std::vector<tbb::task_arena> arenas(numa_nodes.size());
std::vector<tbb::task_group> task_groups(numa_nodes.size());
// initialize each arena, each constrained to a different NUMA node
for (int i = 0; i < numa_nodes.size(); i++)
arenas[i].initialize(tbb::task_arena::constraints(numa_nodes[i]), 0);
// enqueue work to all but the first arena, using the task_group to track work
// by using defer, the task_group reference count is incremented immediately
for (int i = 1; i < numa_nodes.size(); i++)
arenas[i].enqueue(
task_groups[i].defer([] {
tbb::parallel_for(0, N, [](int j) { f(w); });
})
);
// directly execute the work to completion in the remaining arena
arenas[0].execute([] {
tbb::parallel_for(0, N, [](int j) { f(w); });
});
// join the other arenas to wait on their task_groups
for (int i = 1; i < numa_nodes.size(); i++)
arenas[i].execute([&task_groups, i] { task_groups[i].wait(); });
}
In general when tuning a parallel application for NUMA systems, the goal is to expose sufficient parallelism while minimizing (or at least controlling) data access and communication costs. The tradeoffs involved in this tuning often rely on application-specific knowledge.
In particular, NUMA tuning typically involves:
As shown in the previous example, the oneTBB 1.3 specification only provides low-level
support for NUMA optimization. The tbb::info namespace provides topology discovery. And the
combination of task_arena, task_arena::constraints and task_group provide a mechanism for
placing tasks onto specific processors. There is no high-level support for memory allocation
or placement, or for guiding the task distribution of algorithms.
The behavior of existing features is not always predictable. There is a note in
section [info_namespace] of the oneTBB specification that describes
the function std::vector<numa_node_id> numa_nodes(), "If error occurs during system topology
parsing, returns vector containing single element that equals to task_arena::automatic."
In practice, the error can occurs because HWLOC is not detected on the system. While the
oneTBB documentation states in several places that HWLOC is required for NUMA support and
even provides guidance on
how to check for HWLOC,
the inability to resolve HWLOC at runtime silently returns a default of task_arena::automatic. This
default does not pin threads to NUMA nodes. It is too easy to write code similar to the preceding
example and be unaware that a HWLOC installation error (or lack of HWLOC) has undone all your effort.
Getting good performance using these tools requires notable manual coding effort by users. As we
can see in the preceding example, if we want to spread work across the NUMA nodes in
a system we might need to query the topology using functions in the tbb::info namespace, create
one task_arena per NUMA node, along with one task_group per NUMA node, and then add an
extra loop that iterates over these task_arena and task_group objects to execute the
work on the desired NUMA nodes. We also need to handle all container allocations using OS-specific
APIs (or behaviors, such as first-touch) to allocator or place them on the appropriate NUMA nodes.
The out-of-the-box performance of the generic TBB APIs on NUMA systems is not good enough. Should the oneTBB library do anything special by default if the system is a NUMA system? Or should regular random stealing distribute the work across all of the cores, regardless of which NUMA first touched the data?
Is it reasonable for a developer to expect that a series of loops, such as the ones that follow, will
try to create a NUMA-friendly distribution of tasks so that accesses to the same elements of b and c
in the two loops are from the same NUMA nodes? Or is this too much to expect without providing hints?
tbb::parallel_for(0, N,
[](int i) {
b[i] = f(i);
c[i] = g(i);
});
tbb::parallel_for(0, N,
[](int i) {
a[i] = b[i] + c[i];
});
See sub-RFC for increased availability of NUMA API
See sub-RFC for creation and use of NUMA-constrained arenas
Define allocators or other features that simplify the process of allocating or placing data onto specific NUMA nodes.
As discussed earlier, NUMA-aware allocation is just the first step in optimizing for NUMA architectures.
We also need to deliver mechanisms to guide task distribution so that tasks are executed on execution
resources that are near to the data they access. oneTBB already provides low-level support through
tbb::info and tbb::task_arena, but we should up-level this support into the high-level algorithms,
flow graph and containers where appropriate.
For high-level oneTBB features that are modified to provide improved NUMA support, we can try to align default behaviors for those features with user-expectations when used on NUMA systems.