C++ Parallelization and Synchronization

undefined
C
+
+
 
-
 
p
a
r
a
l
l
e
l
i
z
a
t
i
o
n
 
a
n
d
s
y
n
c
h
r
o
n
i
z
a
t
i
o
n
Jakub Yaghob
Martin 
Kruliš
 
T
h
e
 
p
r
o
b
l
e
m
Race conditions
Separate threads with shared state
Result of computation depends on OS scheduling
 
R
a
c
e
 
c
o
n
d
i
t
i
o
n
s
 
 
s
i
m
p
l
e
d
e
m
o
Linked list
Shared state
List lst;
Thread A
lst.push_front(A);
Thread B
lst.push_front(B);
lst
X
Y
lst
X
Y
B
A
Initial state
 
Correct state
lst
X
Y
A
B
 
Another correct state
lst
X
Y
B
A
 
Incorrect state
 
R
a
c
e
 
c
o
n
d
i
t
i
o
n
s
 
 
a
d
v
a
n
c
e
d
d
e
m
o
struct
 Counter {
  Counter():value(0) { }
  
int value;
  void increment()
{
 
++value;
 }
  
void decrement()
{
 
--value;
 
}
  int get()
{
 
return value
; 
}
};
Shared state
Counter c;
Thread A
c.increment();
cout << c.get();
Thread B
c.increment();
cout << c.get();
Possible outputs
12, 21, 
01
, 
10
 
C
+
+
 
1
1
 
f
e
a
t
u
r
e
s
Atomic operations
Low-level threads
High-level futures
Synchronization primitives
Thread-local storage
 
C
+
+
 
1
1
 
 
a
t
o
m
i
c
 
o
p
e
r
a
t
i
o
n
s
Atomic operations
Header <atomics>
Allows creating portable lock-free algorithms and
data structures
Memory ordering
Fences
Lock-free operations, algorithms, data-structures
 
C
+
+
 
1
1
 
 
a
t
o
m
i
c
 
o
p
e
r
a
t
i
o
n
s
Memory ordering
 
enum memory_order;
memory_order_seq_cst
Sequentially consistent, most restrictive memory model
memory_order_relaxed
Totally relaxed memory model, allows best freedom for
CPU and compiler optimizations
memory_order_acquire
, 
memory_order_release
,
memory_order_acq_rel
, 
memory_order_consume
Additional barriers, weaker then sequentially consistent,
stronger the relaxed
 
C
+
+
 
1
1
 
 
a
t
o
m
i
c
 
o
p
e
r
a
t
i
o
n
s
Barriers
Acquire barrier
All loads read after acquire will perform after it (loads do not
overtake acquire)
Release barrier
All stores written before release are committed before the
release (writes do not delay)
Consume
Weaker version of acquire barrier
Loads after barrier are not reordered before it (like acquire,
but only in cases, where they depend on that value)
Perform consume barrier, then load a pointer, then
dereference the pointer; the dereference will not move before
consume barrier
 
C
+
+
 
1
1
 
 
a
t
o
m
i
c
 
o
p
e
r
a
t
i
o
n
s
Easy way to make the demo safe
#include <atomic>
struct Counter {
  std::atomic<
int
>
 value;
  void increment(){
 
++value;
 }
  
void decrement(){
 
--value;
 
}
  int get(){
 
return value
.load(); 
}
};
 
C
+
+
 
1
1
 
 
a
t
o
m
i
c
 
o
p
e
r
a
t
i
o
n
s
Template atomic
Defined for any type
Load, store, compare_exchange
Specialized for bool, all integral types, and
pointers
Load, store, compare_exchange
Arithmetic and bitwise operations
fetch_add
 
C
+
+
 
1
1
 
 
a
t
o
m
i
c
 
o
p
e
r
a
t
i
o
n
s
Fences
Explicit memory barrier
void atomic_thread_fence(memory_order
order) noexcept;
memory_order_relaxed
No effect
memory_order_acquire
, 
memory_order_consume
An acquire fence
memory_order_release
A release fence
memory_order_acq_rel
Both an acquire and a release fence
memory_order_seq_cst
Sequentially consistent
 
C
+
+
 
1
1
 
 
l
o
c
k
-
f
r
e
e
p
r
o
g
r
a
m
m
i
n
g
 
C
+
+
 
1
1
 
 
l
o
c
k
-
f
r
e
e
p
r
o
g
r
a
m
m
i
n
g
 
C
+
+
 
1
1
 
 
t
h
r
e
a
d
s
Low-level threads
Header <thread>
thread
 class
Fork-join paradigm
Namespace 
this_thread
 
C
+
+
 
1
1
 
 
t
h
r
e
a
d
s
Class thread
Constructor
template <class F, class ...Args>
explicit thread(F&& f, Args&&... args);
Destructor
If joinable() then 
terminate()
bool joinable() const noexcept;
void join();
Blocks, until the thread *this has completed
void detach();
id get_id() const noexcept;
 
C
+
+
 
1
1
 
 
t
h
r
e
a
d
s
Namespace 
this_thread
thread::id get_id() noexcept;
Unique ID of the current thread
void yield() noexcept;
Opportunity to reschedule
sleep_for
, 
sleep_until
Blocks the thread for relative/absolute timeout
 
C
+
+
 
1
1
 
 
t
h
r
e
a
d
s
Demo
#include <iostream>
#include <thread>
void thread_fn() { std::cout << “Hello from thread” <<
std::endl; }
int main(int argc, char **argv) {
  std::thread thr(&thread_fn);
  std::cout << “Hello from main” << std::endl;
  thr.join();
  return 0;
}
 
C
+
+
 
1
1
 
 
t
h
r
e
a
d
s
fork
join
“Hello from main”
“Hello from thread”
 
C
+
+
 
1
1
 
 
t
h
r
e
a
d
s
fork
blocked on join
“Hello from main”
“Hello from thread”
thread creation
overhead
 
C
+
+
 
1
1
 
 
t
h
r
e
a
d
s
fork
barrier
 
C
+
+
 
1
1
 
 
t
h
r
e
a
d
s
Demo
#include <iostream>
#include <thread>
#include <vector>
int main(int argc, char **argv) {
  std::vector<std::thread> workers;
  for(int i=0;i<10;++i)
    workers.push_back(std::thread([i]() {
      std::cout << “Hello from thread “ << i << std::endl;
    }));
  std::cout << “Hello from main” << std::endl;
  for(auto &t : workers)
    t.join();
  return 0;
}
 
C
+
+
 
1
1
 
 
t
h
r
e
a
d
s
Passing arguments to threads
By value
Safe, but you MUST make deep copy
By move (rvalue reference)
Safe, as long as strict (deep) adherence to move
semantics
By const reference
Safe, as long as object is guaranteed deep-immutable
By non-const reference
Safe, as long as the object is monitor
 
C
+
+
 
1
1
 
 
f
u
t
u
r
e
s
Futures
Header <future>
High-level asynchronous execution
Future
Promise
Async
Error handling
 
C
+
+
 
1
1
 
 
f
u
t
u
r
e
s
Shared state
Consist of
Some state information and some (possibly not yet
evaluated) result, which can be a (possibly void) value or an
exception
Asynchronous return object
Object, that reads results from an shared state
Waiting function
Potentially blocks to wait for the shared state to be made
ready
Asynchronous provider
Object that provides a result to a shared state
 
C
+
+
 
1
1
 
 
f
u
t
u
r
e
s
Future
std::future<T>
Future value of type T
Retrieve value via 
get()
Waits until the shared state is ready
wait()
, 
wait_for()
, 
wait_until()
std::shared_future<T>
Value can be read by more then one thread
 
C
+
+
 
1
1
 
 
f
u
t
u
r
e
s
Async
std::async
Higher-level convenience utility
Launches a function potentially in a new thread
Async usage
int foo(double, char, bool);
auto fut = std::async(foo, 1.5, 'x', false);
auto res = fut.get();
 
C
+
+
 
1
1
 
 
f
u
t
u
r
e
s
Packaged task
std::packaged_task
How to implement async with more control
Wraps a function and provides a future for the
function result value, but the object itself is
callable
 
C
+
+
 
1
1
 
 
f
u
t
u
r
e
s
Packaged task usage
std::packaged_task<int(double, char, bool)>
tsk(foo);
auto fut = tsk.get_future();
std::thread thr(std::move(tsk), 1.5, 'x', false);
auto res = fut.get();
 
C
+
+
 
1
1
 
 
f
u
t
u
r
e
s
Promise
std::promise<T>
Lowest-level
Steps
Calling thread makes a promise
Calling thread obtains a future from the promise
The promise, along with function arguments, are
moved into a separate thread
The new thread executes the function and fulfills the
promise
The original thread retrieves the result
 
C
+
+
 
1
1
 
 
f
u
t
u
r
e
s
Promise usage
Thread A
std::promise<int> prm;
auto fut = prm.get_future();
std::thread thr(thr_fnc, std::move(prm));
auto res = fut.get();
Thread B
void thr_fnc(std::promise<int> &&prm) {
  prm.set_value(123);
}
 
C
+
+
 
1
1
 
 
f
u
t
u
r
e
s
Constraints
A default-constructed promise is inactive
Can die without consequence
A promise becomes active, when a future is obtained via
get_future()
Only one future may be obtained
A promise must either be satisfied via 
set_value()
, or
have an exception set via 
set_exception()
A satisfied promise can die without consequence
get()
 becomes available on the future
A promise with an exception will raise the stored exception
upon call of 
get()
 on the future
A promise with neither value nor exception will raise “broken
promise” exception
 
C
+
+
 
1
1
 
 
f
u
t
u
r
e
s
Exceptions
All exceptions of type 
std::future_error
Has error code with enum type 
std::future_errc
 inactive promise
std::promise<int> pr;
// fine, no problem
 active promise, unused
std::promise<int> pr;
auto fut = pr.get_future();
// fine, no problem
// fut.get() blocks indefinitely
 too many futures
std::promise<int> pr;
auto fut1 = pr.get_future();
auto fut2 = pr.get_future();
// error “Future already
retrieved”
 
C
+
+
 
1
1
 
 
f
u
t
u
r
e
s
 satisfied promise
std::promise<int> pr;
auto fut = pr.get_future();
{
 
std::promise<int>
pr2(std::move(pr));
  
pr2.set_value(10);
} 
auto r = 
fut.get(); 
// fine, return 10
 too much satisfaction
std::promise<int> pr;
auto fut = pr.get_future();
{
 
std::promise<int>
pr2(std::move(pr));
  
pr2.set_value(10);
  
pr2.set_value(1
1
);
// error “Promise already
satisfied”
} 
auto r = 
fut.get();
 
C
+
+
 
1
1
 
 
f
u
t
u
r
e
s
 exception
std::promise<int> pr;
auto fut = pr.get_future();
{
 
std::promise<int> pr2(std::move(pr));
  p
r2.set_
exception
(
    std::make_exception_ptr(
      std::runtime_error(“bububu”))
);
} 
auto r = 
fut.get();
// throws the runtime_error
 
C
+
+
 
1
1
 
 
f
u
t
u
r
e
s
 broken promise
std::promise<int> pr;
auto fut = pr.get_future();
{
 
std::promise<int> pr2(std::move(pr));
  
// error “Broken promise”
} 
auto r = 
fut.get();
 
C
+
+
 
1
1
 
 
s
y
n
c
h
r
o
n
i
z
a
t
i
o
n
p
r
i
m
i
t
i
v
e
s
Synchronization primitives
Mutual exclusion
Header <mutex>
Condition variables
Header <condition_variable>
 
C
+
+
 
1
1
 
 
m
u
t
e
x
Mutex
A synchronization primitive that can be used to protect
shared data from being simultaneously accessed by
multiple threads
mutex
 offers exclusive, non-recursive ownership semantics
A calling thread 
owns
 a 
mutex
 from the time that it successfully
calls either 
lock
 or 
try_lock
 until it calls 
unlock
When a thread owns a 
mutex
, all other threads will block (for
calls to 
lock
) or receive a false return value (for 
try_lock
) if
they attempt to claim ownership of the 
mutex
A calling thread must not own the 
mutex
 prior to calling 
lock
or 
try_lock
The behavior of a program is undefined if a 
mutex
 is
destroyed while still owned by some thread
 
C
+
+
 
1
1
 
 
m
u
t
e
x
 
e
x
a
m
p
l
e
Shared state
List lst;
std::mutex mtx;
Thread A
mtx.lock();
lst.push_front(A);
mtx.unlock();
Thread B
mtx.lock();
lst.push_front(B);
mtx.unlock();
 
C
+
+
 
1
1
 
 
m
u
t
e
x
 
v
a
r
i
a
n
t
s
Other 
mutex
 variants
timed_mutex
In addition, 
timed_mutex
 provides the ability to attempt to claim ownership
of a 
timed_mutex
 with a timeout via the 
try_lock_for
 and
try_lock_until
recursive_mutex
exclusive, recursive ownership semantics
A calling thread 
owns
 a 
recursive_mutex
 for a period of time that starts when it
successfully calls either 
lock
 or 
try_lock
. During this period, the thread may make
additional calls to 
lock
 or 
try_lock
. The period of ownership ends when the thread
makes a matching number of calls to 
unlock
When a thread owns a 
recursive_mutex
, all other threads will block (for calls to
lock
) or receive a false return value (for 
try_lock
) if they attempt to claim
ownership of the 
recursive_mutex
The maximum number of times that a 
recursive_mutex
 may be locked is
unspecified, but after that number is reached, calls to 
lock
 will throw
std::system_error 
and calls to 
try_lock
 will return false
recursive_timed_mutex
Combination
 
C
+
+
 
1
1
 
 
m
u
t
e
x
 
w
r
a
p
p
e
r
s
std::unique_lock
Lock class with more features
Timed wait, deferred lock
std::lock_guard
Scope based lock (RAII)
Linked list demo, code for one thread
{
std::lock_guard<std::mutex> lk(mtx);
lst.push_front(X);
}
 
C
+
+
 
1
4
 
 
m
u
t
e
x
 
v
a
r
i
a
n
t
s
 
a
n
d
w
r
a
p
p
e
r
s
Other mutex variants in C++ 14
std::shared_timed_mutex
Multiple threads can make shared lock using
lock_shared()
Additional wrapper
std::shared_lock
Calls 
lock_shared
 for the given mutex
 
C
+
+
 
1
1
 
 
l
o
c
k
i
n
g
 
a
l
g
o
r
i
t
h
m
s
std::lock
locks specified mutexes, blocks if any are
unavailable
std::try_lock
attempts to obtain ownership of mutexes via
repeated calls to 
try_lock
// don't actually take the locks yet
std::unique_lock<std::mutex> lock1(mtx1, std::defer_lock);
std::unique_lock<std::mutex> lock2(mtx2, std::defer_lock);
// lock both unique_locks without deadlock
std::lock(lock1, lock2);
 
C
+
+
 
1
1
 
 
c
a
l
l
 
o
n
c
e
std::once_flag
Helper object for 
std::call_once
std::call_once
invokes a function only once even if called from
multiple threads
std::once_flag flag;
void do_once() {
   std::call_once(flag, [](){ do something only once }); }
std::thread t1(do_once);
std::thread t2(do_once);
 
C
+
+
 
1
1
 
 
c
o
n
d
i
t
i
o
n
 
v
a
r
i
a
b
l
e
std::condition_variable
Can be used to block a thread, or multiple threads
at the same time, until
a notification is received from another thread
a timeout expires, or
a spurious wakeup occurs
Appears to be signaled, although the condition is not valid
Verify the condition after the thread has finished waiting
Works with 
std::unique_lock
Atomically manipulates mutex
 
C
+
+
1
1
 
 
c
o
n
d
i
t
i
o
n
 
v
a
r
i
a
b
l
e
e
x
a
m
p
l
e
Producer
for () {
  // produce something
  
std::lock_guard<std::mutex> lock(m);
  queue
.push(i
tem
);
  
notified = true;
  
cond_var.notify_one();
}
std::lock_guard<std::mutex> lock(m);
notified = true;
done = true;
cond_var.notify_one();
Consumer
std::unique_lock<std::mutex> lock(m);
while(!done) {
  
while (!notified) {
    
// loop to avoid spurious wakeups
    
cond_var.wait(lock);
  
}
  
while(!produced_nums.empty()) {
    // consume
    
produced_nums.pop();
  
}
  
notified = false;
}
std::mutex m;
std::condition_variable cond_var;
bool done = false; bool notified = false;
 
C
+
+
 
1
1
 
 
t
h
r
e
a
d
-
l
o
c
a
l
 
s
t
o
r
a
g
e
Thread-local storage
Added a new storage-class
Use keyword 
thread_local
Must be present in all declarations of a variable
Only for namespace or block scope variables and to
the names of static data members
For block scope variables 
static
 is implied
Storage of a variable lasts for the duration of a
thread in which it is created
 
C
+
+
 
e
x
t
e
n
s
i
o
n
s
 
 
p
a
r
a
l
l
e
l
i
s
m
Parallelism
TS v2, v1 will be adopted in C++ 17
Parallel algorithms
Execution policy
Sequential, parallel, parallel_vector
seq
, 
par
, 
par_vec
Dynamic execution policy controlled by execution_policy
for_each
reduce
, 
scan
, 
transform_reduce
, 
transform_scan
Inclusive scan – includes i-th input element in the i-th sum
Exclusive scan – excludes i-th input element from the i-th sum
Parallel exceptions
Multiple exceptions in one parallel algorithm
 
C
+
+
 
e
x
t
e
n
s
i
o
n
s
 
 
p
a
r
a
l
l
e
l
i
s
m
Parallel algorithms
Not all algorithms have parallel version
a
d
j
a
c
e
n
t
_
d
i
f
f
e
r
e
n
c
e
,
 
a
d
j
a
c
e
n
t
_
f
i
n
d
,
 
a
l
l
_
o
f
,
 
a
n
y
_
o
f
,
 
c
o
p
y
,
 
c
o
p
y
_
i
f
,
c
o
p
y
_
n
,
 
c
o
u
n
t
,
 
c
o
u
n
t
_
i
f
,
 
e
q
u
a
l
,
 
e
x
c
l
u
s
i
v
e
_
s
c
a
n
,
 
f
i
l
l
,
 
f
i
l
l
_
n
,
 
f
i
n
d
,
f
i
n
d
_
e
n
d
,
 
f
i
n
d
_
f
i
r
s
t
_
o
f
,
 
f
i
n
d
_
i
f
,
 
f
i
n
d
_
i
f
_
n
o
t
,
 
f
o
r
_
e
a
c
h
,
f
o
r
_
e
a
c
h
_
n
,
 
g
e
n
e
r
a
t
e
,
 
g
e
n
e
r
a
t
e
_
n
,
 
i
n
c
l
u
d
e
s
,
 
i
n
c
l
u
s
i
v
e
_
s
c
a
n
,
i
n
n
e
r
_
p
r
o
d
u
c
t
,
 
i
n
p
l
a
c
e
_
m
e
r
g
e
,
 
i
s
_
h
e
a
p
,
 
i
s
_
h
e
a
p
_
u
n
t
i
l
,
i
s
_
p
a
r
t
i
t
i
o
n
e
d
,
 
i
s
_
s
o
r
t
e
d
,
 
i
s
_
s
o
r
t
e
d
_
u
n
t
i
l
,
l
e
x
i
c
o
g
r
a
p
h
i
c
a
l
_
c
o
m
p
a
r
e
,
 
m
a
x
_
e
l
e
m
e
n
t
,
 
m
e
r
g
e
,
 
m
i
n
_
e
l
e
m
e
n
t
,
m
i
n
m
a
x
_
e
l
e
m
e
n
t
,
 
m
i
s
m
a
t
c
h
,
 
m
o
v
e
,
 
n
o
n
e
_
o
f
,
 
n
t
h
_
e
l
e
m
e
n
t
,
p
a
r
t
i
a
l
_
s
o
r
t
,
 
p
a
r
t
i
a
l
_
s
o
r
t
_
c
o
p
y
,
 
p
a
r
t
i
t
i
o
n
,
 
p
a
r
t
i
t
i
o
n
_
c
o
p
y
,
r
e
d
u
c
e
,
 
r
e
m
o
v
e
,
 
r
e
m
o
v
e
_
c
o
p
y
,
 
r
e
m
o
v
e
_
c
o
p
y
_
i
f
,
 
r
e
m
o
v
e
_
i
f
,
 
r
e
p
l
a
c
e
,
r
e
p
l
a
c
e
_
c
o
p
y
,
 
r
e
p
l
a
c
e
_
c
o
p
y
_
i
f
,
 
r
e
p
l
a
c
e
_
i
f
,
 
r
e
v
e
r
s
e
,
r
e
v
e
r
s
e
_
c
o
p
y
,
 
r
o
t
a
t
e
,
 
r
o
t
a
t
e
_
c
o
p
y
,
 
s
e
a
r
c
h
,
 
s
e
a
r
c
h
_
n
,
s
e
t
_
d
i
f
f
e
r
e
n
c
e
,
 
s
e
t
_
i
n
t
e
r
s
e
c
t
i
o
n
,
 
s
e
t
_
s
y
m
m
e
t
r
i
c
_
d
i
f
f
e
r
e
n
c
e
,
s
e
t
_
u
n
i
o
n
,
 
s
o
r
t
,
 
s
t
a
b
l
e
_
p
a
r
t
i
t
i
o
n
,
 
s
t
a
b
l
e
_
s
o
r
t
,
 
s
w
a
p
_
r
a
n
g
e
s
,
t
r
a
n
s
f
o
r
m
,
 
t
r
a
n
s
f
o
r
m
_
e
x
c
l
u
s
i
v
e
_
s
c
a
n
,
 
t
r
a
n
s
f
o
r
m
_
i
n
c
l
u
s
i
v
e
_
s
c
a
n
,
t
r
a
n
s
f
o
r
m
_
r
e
d
u
c
e
,
 
u
n
i
n
i
t
i
a
l
i
z
e
d
_
c
o
p
y
,
 
u
n
i
n
i
t
i
a
l
i
z
e
d
_
c
o
p
y
_
n
,
u
n
i
n
i
t
i
a
l
i
z
e
d
_
f
i
l
l
,
 
u
n
i
n
i
t
i
a
l
i
z
e
d
_
f
i
l
l
_
n
,
 
u
n
i
q
u
e
,
 
u
n
i
q
u
e
_
c
o
p
y
 
C
+
+
 
e
x
t
e
n
s
i
o
n
s
 
 
p
a
r
a
l
l
e
l
i
s
m
Task block
Support for fork-join paradigm
Spawn other task_blocks and wait for their
completion
Exceptions
Each task_block has an exception list
Exceptions from forked task_blocks are stored in the
exception list
Exceptions are invoked when task_block finishes
 
C
+
+
 
e
x
t
e
n
s
i
o
n
s
 
 
c
o
n
c
u
r
r
e
n
c
y
Concurrency
TS, won’t be adopted in C++ 17
Improvements to future
future<T2> then(F &&f)
Execute asynchronously a function f when the future is ready
Latches
Thread coordination mechanism
Block one or more threads until an operation is completed
Single use
Barriers
Thread coordination mechanism
Reusable
Multiple barrier types
barrier
flex_barrier – calls a function in a completion phase
 
C
+
+
 
e
x
t
e
n
s
i
o
n
 
 
t
r
a
n
s
a
c
t
i
o
n
a
l
m
e
m
o
r
y
TS v1, won’t be adopted in C++ 17
Transactional memory
Added several keywords for statements and declarations
s
y
n
c
h
r
o
n
i
z
e
d
 
c
o
m
p
o
u
n
d
-
s
t
a
t
e
m
e
n
t
Synchronized with other synchronized blocks
One global lock for all synchronized blocks
Atomic blocks
Execute atomically and not concurrently with synchronized blocks
Can execute concurrently with other atomic blocks if no conflicts
Differs in behavior with exceptions
a
t
o
m
i
c
_
n
o
e
x
c
e
p
t
 
c
o
m
p
o
u
n
d
-
s
t
a
t
e
m
e
n
t
Escaping exception causes undefined behavior
a
t
o
m
i
c
_
c
a
n
c
e
l
 
c
o
m
p
o
u
n
d
-
s
t
a
t
e
m
e
n
t
Escaping exception rolls back the transaction, but must be transaction safe
F
u
n
c
t
i
o
n
s
 
c
a
n
 
b
e
 
d
e
c
l
a
r
e
d
 
t
r
a
n
s
a
c
t
i
o
n
_
s
a
f
e
a
t
o
m
i
c
_
c
o
m
m
i
t
 
c
o
m
p
o
u
n
d
-
s
t
a
t
e
m
e
n
t
Escaping exception commits the transaction
Slide Note
Embed
Share

Explore the challenges of race conditions in C++ multithreading, from basic demonstrations to advanced scenarios. Delve into C++11 features like atomic operations, memory ordering, and synchronization primitives to create efficient and thread-safe applications.

  • C++
  • Multithreading
  • Synchronization
  • Atomic Operations
  • Memory Ordering

Uploaded on Sep 09, 2024 | 0 Views


Download Presentation

Please find below an Image/Link to download the presentation.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author. Download presentation by click this link. If you encounter any issues during the download, it is possible that the publisher has removed the file from their server.

E N D

Presentation Transcript


  1. C++ - parallelization and synchronization Jakub Yaghob Martin Kruli

  2. The problem Race conditions Separate threads with shared state Result of computation depends on OS scheduling

  3. Race conditions simple demo Initial state Linked list Shared state List lst; Thread A lst.push_front(A); Thread B lst.push_front(B); lst X Y Correct state lst B A X Y Another correct state lst A B X Y Incorrect state A lst X Y B

  4. Race conditions advanced demo Shared state Counter c; Thread A c.increment(); cout << c.get(); Thread B c.increment(); cout << c.get(); Possible outputs 12, 21, 01, 10 struct Counter { Counter():value(0) { } int value; void increment() { ++value; } void decrement() { --value; } int get() { return value; } };

  5. C++ 11 features Atomic operations Low-level threads High-level futures Synchronization primitives Thread-local storage

  6. C++ 11 atomic operations Atomic operations Header <atomics> Allows creating portable lock-free algorithms and data structures Memory ordering Fences Lock-free operations, algorithms, data-structures

  7. C++ 11 atomic operations Memory ordering enum memory_order; memory_order_seq_cst Sequentially consistent, most restrictive memory model memory_order_relaxed Totally relaxed memory model, allows best freedom for CPU and compiler optimizations memory_order_acquire, memory_order_release, memory_order_acq_rel, memory_order_consume Additional barriers, weaker then sequentially consistent, stronger the relaxed

  8. C++ 11 atomic operations Barriers Acquire barrier All loads read after acquire will perform after it (loads do not overtake acquire) Release barrier All stores written before release are committed before the release (writes do not delay) Consume Weaker version of acquire barrier Loads after barrier are not reordered before it (like acquire, but only in cases, where they depend on that value) Perform consume barrier, then load a pointer, then dereference the pointer; the dereference will not move before consume barrier

  9. C++ 11 atomic operations Easy way to make the demo safe #include <atomic> struct Counter { std::atomic<int> value; void increment(){ ++value; } void decrement(){ --value; } int get(){ return value.load(); } };

  10. C++ 11 atomic operations Template atomic Defined for any type Load, store, compare_exchange Specialized for bool, all integral types, and pointers Load, store, compare_exchange Arithmetic and bitwise operations fetch_add

  11. C++ 11 atomic operations Fences Explicit memory barrier void atomic_thread_fence(memory_order order) noexcept; memory_order_relaxed No effect memory_order_acquire, memory_order_consume An acquire fence memory_order_release A release fence memory_order_acq_rel Both an acquire and a release fence memory_order_seq_cst Sequentially consistent

  12. C++ 11 lock-free programming

  13. C++ 11 lock-free programming

  14. C++ 11 threads Low-level threads Header <thread> thread class Fork-join paradigm Namespace this_thread

  15. C++ 11 threads Class thread Constructor template <class F, class ...Args> explicit thread(F&& f, Args&&... args); Destructor If joinable() then terminate() bool joinable() const noexcept; void join(); Blocks, until the thread *this has completed void detach(); id get_id() const noexcept;

  16. C++ 11 threads Namespace this_thread thread::id get_id() noexcept; Unique ID of the current thread void yield() noexcept; Opportunity to reschedule sleep_for, sleep_until Blocks the thread for relative/absolute timeout

  17. C++ 11 threads Demo #include <iostream> #include <thread> void thread_fn() { std::cout << Hello from thread << std::endl; } int main(int argc, char **argv) { std::thread thr(&thread_fn); std::cout << Hello from main << std::endl; thr.join(); return 0; }

  18. C++ 11 threads fork Hello from main Hello from thread join

  19. C++ 11 threads fork thread creation overhead Hello from main blocked on join Hello from thread

  20. C++ 11 threads fork barrier

  21. C++ 11 threads Demo #include <iostream> #include <thread> #include <vector> int main(int argc, char **argv) { std::vector<std::thread> workers; for(int i=0;i<10;++i) workers.push_back(std::thread([i]() { std::cout << Hello from thread << i << std::endl; })); std::cout << Hello from main << std::endl; for(auto &t : workers) t.join(); return 0; }

  22. C++ 11 threads Passing arguments to threads By value Safe, but you MUST make deep copy By move (rvalue reference) Safe, as long as strict (deep) adherence to move semantics By const reference Safe, as long as object is guaranteed deep-immutable By non-const reference Safe, as long as the object is monitor

  23. C++ 11 futures Futures Header <future> High-level asynchronous execution Future Promise Async Error handling

  24. C++ 11 futures Shared state Consist of Some state information and some (possibly not yet evaluated) result, which can be a (possibly void) value or an exception Asynchronous return object Object, that reads results from an shared state Waiting function Potentially blocks to wait for the shared state to be made ready Asynchronous provider Object that provides a result to a shared state

  25. C++ 11 futures Future std::future<T> Future value of type T Retrieve value via get() Waits until the shared state is ready wait(), wait_for(), wait_until() std::shared_future<T> Value can be read by more then one thread

  26. C++ 11 futures Async std::async Higher-level convenience utility Launches a function potentially in a new thread Async usage int foo(double, char, bool); auto fut = std::async(foo, 1.5, 'x', false); auto res = fut.get();

  27. C++ 11 futures Packaged task std::packaged_task How to implement async with more control Wraps a function and provides a future for the function result value, but the object itself is callable

  28. C++ 11 futures Packaged task usage std::packaged_task<int(double, char, bool)> tsk(foo); auto fut = tsk.get_future(); std::thread thr(std::move(tsk), 1.5, 'x', false); auto res = fut.get();

  29. C++ 11 futures Promise std::promise<T> Lowest-level Steps Calling thread makes a promise Calling thread obtains a future from the promise The promise, along with function arguments, are moved into a separate thread The new thread executes the function and fulfills the promise The original thread retrieves the result

  30. C++ 11 futures Promise usage Thread A std::promise<int> prm; auto fut = prm.get_future(); std::thread thr(thr_fnc, std::move(prm)); auto res = fut.get(); Thread B void thr_fnc(std::promise<int> &&prm) { prm.set_value(123); }

  31. C++ 11 futures Constraints A default-constructed promise is inactive Can die without consequence A promise becomes active, when a future is obtained via get_future() Only one future may be obtained A promise must either be satisfied via set_value(), or have an exception set via set_exception() A satisfied promise can die without consequence get() becomes available on the future A promise with an exception will raise the stored exception upon call of get() on the future A promise with neither value nor exception will raise broken promise exception

  32. C++ 11 futures Exceptions All exceptions of type std::future_error Has error code with enum type std::future_errc inactive promise std::promise<int> pr; // fine, no problem too many futures std::promise<int> pr; auto fut1 = pr.get_future(); auto fut2 = pr.get_future(); // error Future already retrieved active promise, unused std::promise<int> pr; auto fut = pr.get_future(); // fine, no problem // fut.get() blocks indefinitely

  33. C++ 11 futures satisfied promise std::promise<int> pr; auto fut = pr.get_future(); { std::promise<int> pr2(std::move(pr)); pr2.set_value(10); } auto r = fut.get(); // fine, return 10 too much satisfaction std::promise<int> pr; auto fut = pr.get_future(); { std::promise<int> pr2(std::move(pr)); pr2.set_value(10); pr2.set_value(11); // error Promise already satisfied } auto r = fut.get();

  34. C++ 11 futures exception std::promise<int> pr; auto fut = pr.get_future(); { std::promise<int> pr2(std::move(pr)); pr2.set_exception( std::make_exception_ptr( std::runtime_error( bububu ))); } auto r = fut.get(); // throws the runtime_error

  35. C++ 11 futures broken promise std::promise<int> pr; auto fut = pr.get_future(); { std::promise<int> pr2(std::move(pr)); // error Broken promise } auto r = fut.get();

  36. C++ 11 synchronization primitives Synchronization primitives Mutual exclusion Header <mutex> Condition variables Header <condition_variable>

  37. C++ 11 mutex Mutex A synchronization primitive that can be used to protect shared data from being simultaneously accessed by multiple threads mutex offers exclusive, non-recursive ownership semantics A calling thread owns a mutex from the time that it successfully calls either lock or try_lock until it calls unlock When a thread owns a mutex, all other threads will block (for calls to lock) or receive a false return value (for try_lock) if they attempt to claim ownership of the mutex A calling thread must not own the mutex prior to calling lock or try_lock The behavior of a program is undefined if a mutex is destroyed while still owned by some thread

  38. C++ 11 mutex example Shared state List lst; std::mutex mtx; Thread A mtx.lock(); lst.push_front(A); mtx.unlock(); Thread B mtx.lock(); lst.push_front(B); mtx.unlock();

  39. C++ 11 mutex variants Other mutex variants timed_mutex In addition, timed_mutex provides the ability to attempt to claim ownership of a timed_mutex with a timeout via the try_lock_for and try_lock_until recursive_mutex exclusive, recursive ownership semantics A calling thread owns a recursive_mutex for a period of time that starts when it successfully calls either lock or try_lock. During this period, the thread may make additional calls to lock or try_lock. The period of ownership ends when the thread makes a matching number of calls to unlock When a thread owns a recursive_mutex, all other threads will block (for calls to lock) or receive a false return value (for try_lock) if they attempt to claim ownership of the recursive_mutex The maximum number of times that a recursive_mutex may be locked is unspecified, but after that number is reached, calls to lock will throw std::system_error and calls to try_lock will return false recursive_timed_mutex Combination

  40. C++ 11 mutex wrappers std::unique_lock Lock class with more features Timed wait, deferred lock std::lock_guard Scope based lock (RAII) Linked list demo, code for one thread { std::lock_guard<std::mutex> lk(mtx); lst.push_front(X); }

  41. C++ 14 mutex variants and wrappers Other mutex variants in C++ 14 std::shared_timed_mutex Multiple threads can make shared lock using lock_shared() Additional wrapper std::shared_lock Calls lock_shared for the given mutex

  42. C++ 11 locking algorithms std::lock locks specified mutexes, blocks if any are unavailable std::try_lock attempts to obtain ownership of mutexes via repeated calls to try_lock // don't actually take the locks yet std::unique_lock<std::mutex> lock1(mtx1, std::defer_lock); std::unique_lock<std::mutex> lock2(mtx2, std::defer_lock); // lock both unique_locks without deadlock std::lock(lock1, lock2);

  43. C++ 11 call once std::once_flag Helper object for std::call_once std::call_once invokes a function only once even if called from multiple threads std::once_flag flag; void do_once() { std::call_once(flag, [](){ do something only once }); } std::thread t1(do_once); std::thread t2(do_once);

  44. C++ 11 condition variable std::condition_variable Can be used to block a thread, or multiple threads at the same time, until a notification is received from another thread a timeout expires, or a spurious wakeup occurs Appears to be signaled, although the condition is not valid Verify the condition after the thread has finished waiting Works with std::unique_lock Atomically manipulates mutex

  45. C++11 condition variable example std::mutex m; std::condition_variable cond_var; bool done = false; bool notified = false; Producer for () { // produce something std::lock_guard<std::mutex> lock(m); queue.push(item); notified = true; cond_var.notify_one(); } std::lock_guard<std::mutex> lock(m); notified = true; done = true; cond_var.notify_one(); Consumer std::unique_lock<std::mutex> lock(m); while(!done) { while (!notified) { // loop to avoid spurious wakeups cond_var.wait(lock); } while(!produced_nums.empty()) { // consume produced_nums.pop(); } notified = false; }

  46. C++ 11 thread-local storage Thread-local storage Added a new storage-class Use keyword thread_local Must be present in all declarations of a variable Only for namespace or block scope variables and to the names of static data members For block scope variables static is implied Storage of a variable lasts for the duration of a thread in which it is created

  47. C++ extensions parallelism Parallelism TS v2, v1 will be adopted in C++ 17 Parallel algorithms Execution policy Sequential, parallel, parallel_vector seq, par, par_vec Dynamic execution policy controlled by execution_policy for_each reduce, scan, transform_reduce, transform_scan Inclusive scan includes i-th input element in the i-th sum Exclusive scan excludes i-th input element from the i-th sum Parallel exceptions Multiple exceptions in one parallel algorithm

  48. C++ extensions parallelism Parallel algorithms Not all algorithms have parallel version adjacent_difference, adjacent_find, all_of, any_of, copy, copy_if, copy_n, count, count_if, equal, exclusive_scan, fill, fill_n, find, find_end, find_first_of, find_if, find_if_not, for_each, for_each_n, generate, generate_n, includes, inclusive_scan, inner_product, inplace_merge, is_heap, is_heap_until, is_partitioned, is_sorted, is_sorted_until, lexicographical_compare, max_element, merge, min_element, minmax_element, mismatch, move, none_of, nth_element, partial_sort, partial_sort_copy, partition, partition_copy, reduce, remove, remove_copy, remove_copy_if, remove_if, replace, replace_copy, replace_copy_if, replace_if, reverse, reverse_copy, rotate, rotate_copy, search, search_n, set_difference, set_intersection, set_symmetric_difference, set_union, sort, stable_partition, stable_sort, swap_ranges, transform, transform_exclusive_scan, transform_inclusive_scan, transform_reduce, uninitialized_copy, uninitialized_copy_n, uninitialized_fill, uninitialized_fill_n, unique, unique_copy

  49. C++ extensions parallelism Task block Support for fork-join paradigm Spawn other task_blocks and wait for their completion Exceptions Each task_block has an exception list Exceptions from forked task_blocks are stored in the exception list Exceptions are invoked when task_block finishes

  50. C++ extensions concurrency Concurrency TS, won t be adopted in C++ 17 Improvements to future future<T2> then(F &&f) Execute asynchronously a function f when the future is ready Latches Thread coordination mechanism Block one or more threads until an operation is completed Single use Barriers Thread coordination mechanism Reusable Multiple barrier types barrier flex_barrier calls a function in a completion phase

Related


More Related Content

giItT1WQy@!-/#giItT1WQy@!-/#giItT1WQy@!-/#giItT1WQy@!-/#giItT1WQy@!-/#