Upgrade to Pro — share decks privately, control downloads, hide ads and more …

Tim Mattson (Intel) - OpenMP workshop at Multic...

Tim Mattson (Intel) - OpenMP workshop at Multicore World 2013

A tutorial introduction to Parallel programming using OpenMP by Intel's Principle Engineer Tim Mattson.

Hands-on workshop collocated with Multicore World 2013, Monday 18 February 2013, Wellington, New Zealand.

Courtesy of Open Parallel Ltd

Multicore World 2013

February 19, 2013
Tweet

More Decks by Multicore World 2013

Other Decks in Programming

Transcript

  1. 1 A  “Hands-on”  Introduction  to  to   parallel Programming (with

    OpenMP*) Tim Mattson Intel Corporation [email protected] * The  name  “OpenMP”  is  the  property  of  the  OpenMP  Architecture  Review  Board.
  2. 2 Preliminaries: part 1  Disclosures The views expressed in

    this tutorial are my own. – I am not speaking for my employer. – I am not speaking for the OpenMP ARB  I take my tutorials VERY seriously: Help  me  improve  …  let  me  know  if  you  have  ideas   on how to improve this content.
  3. 3 Preliminaries: Part 2  Our plan for the day

    .. Active learning! We will mix short lectures with short exercises.  Please follow these simple rules Do the exercises I assign and then change things around and experiment. – Embrace active learning! Don’t  cheat: Do Not look at the solutions before you  complete  an  exercise  …  even  if  you  get  really   frustrated.
  4. 4 Our Plan for the day Topic Exercise concepts Intro

    to parallel programming No exercise Basic concepts and the jargon of parallel programming OMP Intro Install sw, hello_world Parallel regions Creating threads Pi_spmd_simple Parallel, default data environment, runtime library calls Synchronization Pi_spmd_final False sharing, critical, atomic Parallel loops Pi_loop, Matmul For, schedule, reduction, The rest of worksharing and synchronization No exercise Single, sections, master, runtime libraries, environment variables, synchronization, etc. Data Environment Mandelbrot set area Data environment details, software optimization Break
  5. 5 Extra Content for self-study (in backup) Topic Exercise concepts

    OpenMP Recursive matrix multiply OpenMP tasks Memory model Producer-Consumer The need for flush Threadprivate Monte Carlo PI Thread private variables and modular programming A survey of algorithms and parallel programming models No exercise Design patterns, Cilk, MPI, OpenCL, TBB, etc.
  6. 6 Outline  Introduction to parallel computing  Introduction to

    OpenMP  Creating Threads  Synchronization  Parallel Loops  Synchronize single masters and stuff  Data environment  Tasks  Memory model  Threadprivate Data
  7. Agenda – parallel theory  How to sound like a

    parallel programmer  An overly brief look at parallel architecture
  8. Concurrency vs. Parallelism  Two important definitions:  Concurrency: A

    condition of a system in which multiple tasks are logically active at one time.  Parallelism: A condition of a system in which multiple tasks are actually active at one time. Figure  from  “An  Introduction  to  Concurrency  in  Programming  Languages”  by  J.  Sottile,  Timothy  G.  Mattson,  and  Craig  E  Rasmussen, 2010
  9. Concurrency vs. Parallelism Figure  from  “An  Introduction  to  Concurrency  in

     Programming  Languages”  by  J.  Sottile,  Timothy  G.  Mattson,  and  Craig  E  Rasmussen, 2010  Two important definitions:  Concurrency: A condition of a system in which multiple tasks are logically active at one time.  Parallelism: A condition of a system in which multiple tasks are actually active at one time.
  10. 10 An Example of Parallel Computing Load Data Compute T1

    Consume Results Compute TN … Timeseq (1) = Tload + N*Ttask + Tconsume Compute N independent tasks on one processor Ideally Cut runtime by ~1/P (Note: Parallelism only speeds-up the concurrent part) … Timepar (P) = Tload + (N/P)*Ttask + Tconsume Compute N independent tasks with P processors Load Data Compute T1 Compute TN Consume Results
  11. 11 Parallel Performance  MP Linpack benchmark, order 1000 matrix

    (solve a dense system of linear  equations  …  the  dense  linear  algebra  computational  pattern). 0 0.5 1 1.5 2 2.5 3 3.5 4 0 10 20 30 40 50 GFlops # cores  Intel SCC 48 processor, 500 MHz core, 1 GHz router, DDR3 at 800 MHz.
  12. 12 Talking about performance ) ( ) 1 ( )

    ( P Time Time P S par seq  P P S  ) (  Perfect Linear Speedup: happens when no parallel overhead and algorithm is 100% parallel.  Speedup: the increased performance from running on P processors
  13. 13 Performance scalability  MP Linpack benchmark, order 1000 matrix

    (solve a dense system of linear  equations  …  the  dense  linear  algebra  computational  pattern). Intel SCC 48 processor, 500 Mhz core, 1 Ghz router, DDR3 at 800 Mhz. 0 20 40 60 80 100 120 140 0 10 20 30 40 50 60 Cores Speedup The 48-core  SCC  processor:  the  programmer’s  view, T, G. Mattson, R. F. Van der Wijngaart, M. Riepen, T. Lehnig, P. Brett, W. Haas, P. Kennedy, J. Howard, S. Vangal, N. Borkar, G. Ruhl, S. Dighe, Proceedings SC10, New Orleans 2010
  14. 14 Talking about performance ) ( ) 1 ( )

    ( P Time Time P S par seq  P P S  ) ( P P S  ) (  Perfect Linear Speedup: happens when no parallel overhead and algorithm is 100% parallel.  Super-linear Speedup: Speed grows faster than the number of processing elements  Speedup: the increased performance from running on P processors
  15. 15 SuperLlnear Speedup  MP Linpack benchmark, order 1000 matrix

    (solve a dense system of linear  equations  …  the  dense  linear  algebra  computational  pattern). 0 20 40 60 80 100 120 140 0 10 20 30 40 50 60 Cores Speedup 0.03 GFs 0.245 GFs 0.78 GFs 3.45 GFs 2.4 GFs 1.6 GFs Why is this number so small? Intel SCC 48 processor, 500 Mhz core, 1 Ghz router, DDR3 at 800 Mhz.
  16. 16 Why the Superlinear speedup? R = router, MC =

    Memory Controller, P54C 16KB L1-D$ 16KB L1-I$ 256KB unified L2$ Mesh I/F To Router P54C 16KB L1-D$ 16KB L1-I$ 256KB unified L2$ Message Passing Buffer 16 KB R R Tile Tile Tile Tile Tile Tile Tile Tile R Tile Tile R Tile Tile R Tile Tile Tile R Tile R to PCI Tile Tile R Tile Tile R Tile Tile Tile R Tile R R R R R R R R R R R R R R MC MC MC MC • Intel SCC 48 core research chip • SCC  caches  are  so  small,  even  a  small  portion  of  our  O(1000)  matrices  won’t  fit.      Hence the single node performance measures memory overhead. • As you add more cores, the aggregate cache size grows.  Eventually the tiles of the matrices being processed fits in the caches and performance sharply increases  superlinear speedup. P54C = second generation Pentium® core, The 48-core  SCC  processor:  the  programmer’s  view, T, G. Mattson, R. F. Van der Wijngaart, M. Riepen, T. Lehnig, P. Brett, W. Haas, P. Kennedy, J. Howard, S. Vangal, N. Borkar, G. Ruhl, S. Dighe, Proceedings SC10, New Orleans 2010
  17. 17 A more typical speedup plot  CHARMM molecular dynamics

    program running the myoglobin benchmark on an Intel Paragon XP/S supercomputer with 32 Mbyte nodes running OSF R 1.2. (The nbody computational pattern). Speedup relative to running the parallel program on one node. 0 20 40 60 80 100 120 140 160 0 100 200 300 400 500 600 Nodes Speedup Porting Applications to the MP-Paragon Supercomputer: The CHARMM Molecular Dynamics program, T.G.  Mattson,  Intel  Supercomputers  User’s  Group  meeting,  1995. Strong scaling …  the   speedup trends for a fixed size problem.
  18. 18 Efficiency  Efficiency measures  how  well  the  parallel  system’s

     resources  are   being utilized.  Where P is the number of nodes and T is the elapsed runtime. P P S P Time P Time par seq ) ( ) ( *   
  19. 19 Efficiency  CHARMM molecular dynamics program running the myoglobin

    benchmark on an Intel Paragon XP/S supercomputer with 32 Mbyte nodes running OSF R 1.2. (The nbody computational pattern). Speedup relative to running the parallel program on one node. Nodes Efficiency Porting Applications to the MP-Paragon Supercomputer: The CHARMM Molecular Dynamics program, T.G. Mattson, Intel Supercomputers  User’s  Group  meeting,  1995. 0 0.2 0.4 0.6 0.8 1 1.2 0 100 200 300 400 500 600
  20. 20 Amdahl's Law  Gene Amdahl (1967) argued that parallel

    computing would be of limited value.  He pointed out that signficant portions of many programs are not parallelizable. Parallel hardware does not help serial code: Runtime = 3 s Runtime = 2.25 s The  “middle   second”  runs   perfectly parallel on 4 threads
  21. 21 Amdahl’s  Law  What is the maximum speedup you

    can expect from a parallel program?  Approximate the runtime as a part that can be sped up with additional processors and a part that is fundamentally serial. seq par Time P fraction parallel fraction serial P Time * ) _ _ ( ) (    If you had an unlimited number of processors:  If serial_fraction is  and parallel_fraction is (1- ) then the speedup is: P Time P Time P S seq seq           1 1 ) 1 ( * ) 1 ( ) 1 ( ) (   P  The maximum possible speedup is:  1  S Amdahl’s   Law
  22. 22 Amdahl's Law and the CHARMM MD program  We

    Profiled CHARMM running on the Paragon XPS to find the time spent  in  code  that  was  not  parallelized  …  concluded  that  CHARMM   has a serial fraction of ~0.003.  The maximum possible speedup is: S= 1/0.003 = 333 0 50 100 150 200 250 0 100 200 300 400 500 600 Est. from serial fraction Observed Nodes Speedup
  23. 23 Limitations to scalability  Remember the speedup plot we

    discussed from last time? Why does the app. Scale worse than we’d  expect  from   Amdahl’s  law?
  24. 24 Molecular dynamics  The potential energy, U(r), is divided

    into two parts:  Bonded terms – Groups of atoms connected by chemical bonds.  Non-bonded terms – longer range forces (e.g. electrostatic). • An N-body problem …  i.e.  every   atom depends on every other atom, an O(N2) problem. Bonds, angles and torsions Source: Izaguirre, Ma and Skeel,  SAC’03  slides,  March  10  2003  Models motion of atoms in molecular systems by solving Newton’s  equations  of  motion:
  25. 25 Bulk Synchronous Processing  Many message passing applications have

    few (if any) sends and receives.  They  use  a  design  pattern  called  “Bulk Synchronous Processing”.  Uses the Single Program Multiple Data pattern  Each process maintains a local view of the global data  A problem broken down into phases each composed of two subphases: • Compute on local view of data • Communicate to update global view on all processes (collective communication).  Continue phases until complete Collective comm. Collective comm.
  26. 26 Source: CS267 Lecture 7 More Collective Data Movement A

    B D C A B C D A B C D A B C D A B C D Allgather P0 P1 P2 P3 P0 P1 P2 P3 A B D C ABCD AllReduce ABCD ABCD ABCD
  27. 27 Molecular dynamics simulation real atoms(3,N) real force(3,N) int neighbors(MX,N)

    // Every PE has a copy of atoms and force loop over time steps parallel loop over atoms Compute neighbor list (for my atoms) Compute nonbonded forces (my atoms and neighbors) Barrier All reduce (Sum force arrays, each PE gets a copy) Compute bonded forces (for my atoms) Integrate to Update position (for my atoms) All_gather(update atoms array) end loop over atoms end loop We  used  a  cutoff  method  …  the  potential   energy drops off quickly so atoms beyond a neighborhood can be ignored in the nonbonded force calc.
  28. 28 Molecular dynamics simulation real atoms(3,N) real force(3,N) int neighbors(MX,N)

    //MX = max neighbors an atom may have // Every PE has a copy of atoms and force loop over time steps parallel loop over atoms Compute neighbor list (for my atoms) Compute long range forces (my atoms and neighbors) Barrier All reduce (Sum force arrays, each PE gets a copy) Compute bonded forces (for my atoms) Integrate to Update position (for my atoms) All_gather(update atoms array) end loop over atoms end loop synchronization Collective Communication
  29. 29 Limitations to scalability  Remember the speedup plot we

    discussed earlier? Why does the app. Scale worse than we’d  expect  from   Amdahl’s  law?
  30. 30 CHARMM Myoglobin Benchmark  Percent of runtime for the

    different phases of the computation 0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% integ list comm wait Ebond Enon 512 256 128 64 148 32 16 8 1 Number of Nodes Percent of total runtime CHARMM running on a distributed memory, MPP supercomputer using a message passing library (NX)
  31. 31 Charm Myoglobin Benchmark  Percent of runtime for the

    different phases of the computation 0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% integ list comm wait Ebond Enon 512 256 128 64 148 32 16 8 1 Number of Nodes Percent of total runtime Enon (the n-body term) scales better than the other computational terms. This was taken into account in the Serial fraction estimate  for  the  Amdahl’s  law  analysis O(N) O(N) O(N2) O(MX*N )
  32. 32 Charm Myoglobin Benchmark  Percent of runtime for the

    different phases of the computation 0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% integ list comm wait Ebond Enon 512 256 128 64 148 32 16 8 1 Number of Nodes Percent of total runtime The fraction of time spent waiting grows with the number of nodes due to two factors: (1) the cost of the barrier grows with the number of nodes, and (2) variation in the work for each node increases as node count  grows  …  load  imbalance.
  33. 33 Synchronization overhead  Processes finish their work and must

    assure that all processes are finished before the results are combined into the global force array.  This  is  parallel  overhead  since  this  doesn’t  occur  in  a  serial   program.  The synchronization construct itself takes time and in some cases (such as a barrier) the cost grows with the number of nodes. CPU 3 CPU 2 CPU 1 CPU 0 CPU 3 CPU 2 CPU 1 CPU 0 Time
  34. 34 CPU 3 CPU 2 CPU 1 Load imbalance 

    If some processes finish their share of the computation early, the time spent waiting for other processors is wasted.  This is an example of Load Imbalance Time CPU 0 CPU 3 CPU 2 CPU 1 CPU 0
  35. 35 Charm Myoglobin Benchmark  Percent of runtime for the

    different phases of the computation 0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% integ list comm wait Ebond Enon 512 256 128 64 148 32 16 8 1 Number of Nodes Percent of total runtime The communication growth is the chief culprit limiting performance in this case.
  36. 36 Communication  On distributed-memory machines (e.g. a cluster), communication

    can only occur by sending discrete messages over a network  The sending processor marshals the shared data from the application's data structures into a message buffer  The receiving processor must wait for the message to arrive ...  ... and un-pack the data back into data structures Time CPU 0 CPU 3
  37. 37 Communication  On distributed-memory machines (e.g. a cluster), communication

    can only occur by sending discrete messages over a network  The sending processor marshals the shared data from the application's data structures into a message buffer  The receiving processor must wait for the message to arrive ...  ... and un-pack the data back into data structures  If the communication protocol is synchronous, then the sending processor must wait for acknowledgement that the message was received Time CPU 0 CPU 3
  38. 38 Charm Myoglobin Benchmark  Percent of runtime for the

    different phases of the computation 0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% integ list comm wait Ebond Enon 512 256 128 64 148 32 16 8 1 Number of Nodes Percent of total runtime Remember these are collective comms. Composed of multiple messages each of which incur these overheads
  39. 39 Limitations to scalability  Remember the speedup plot we

    discussed from last time? Sync, wait, and comm. overheads combined explain this gap
  40. Agenda – parallel theory  How to sound like a

    parallel programmer  An overly brief look at parallel architecture
  41. 41 How do we connect cores together?  A symmetric

    multiprocessor (SMP) consists of a collection of processors that share a single address space:  Multiple processing elements.  A  shared  address  space  with  “equal-time”  access  for  each  processor.  The OS treats every processor the same Proc3 Proc2 Proc1 ProcN Shared Address Space
  42. 42 How realistic is this model?  Some of the

    old supercomputer mainframes followed this model,  But as soon as we added caches to CPUs, the SMP model fell apart.  Caches  …  all  memory  is  equal,  but   some memory is more equal than others. A CPU with lots of cache …
  43. 43 NUMA issues on a Multicore Machine 2-socket Clovertown Dell

    PE1950 2 threads, 2 cores, sharing a cache 2 thrds, 2 cores, 1 sock, no shared cache A single quad-core chip is a NUMA machine! Source Dieter an Mey, IWOMP’07 face to face meeting 2 thrds, 2 cores, 2 sockets $ $ Xeon® 5300 Processor block diagram Third party names are the property of their owners.
  44. 44 Put these into a larger system and it only

     get’s  worse Proc0 Proc1 Memory (0) Proc2 Proc3 Memory (1 ) NODE 0 NODE 1 • Memory access takes longer if memory is remote. • For example, on an SGI Altix: •Proc0 to local memory (0) 207 cycles •Proc0 to remote memory (1) 409 cycles Source: J. Marathe & F. Mueller, Gelato ICE, April 2007. • Consider a typical NUMA computer:
  45. Modern GPGPU Architecture • Generic many core GPU • Less

    space devoted to control logic and caches • Large register files to support multiple thread contexts • Low latency hardware managed thread switching • Large number of ALU per “core”  with  small  user   managed cache per core • Memory bus optimized for bandwidth On Board System Memory 150 GBPS 150 GBPS Simple ALUs Cache Source: Advanced topics in heterogeneous computing with OpenCL, SC10 Benedict R. Gaster
  46. 46 Moving  “beyond  the  single  board”  Parallel computers are

    classified in terms of streams of data and streams of instructions:  MIMD Computers: Multiple streams of instructions acting on multiple streams of data.  SIMD Computers: A single stream of instructions acting on multiple streams of data.  Parallel Hardware comes in many forms:  On chip: Instruction level parallelism (e.g. IPF)  Multiprocessor: Multiple processors inside a single computer.  Multicomputer: networks of computers working together.
  47. 47 Hardware for parallel computing Symmetric Multiprocessor (SMP) Non-uniform Memory

    Architecture (NUMA) Massively Parallel Processor (MPP) Cluster Single Instruction Multiple Data (SIMD)* *SIMD has failed as a way to organize large scale computers with multiple processors. It has succeeded, however, as a mechanism to increase instruction level parallelism in modern microprocessors (MMX, SSE, AVX, etc.). Multiple Instruction Multiple Data (MIMD) Parallel Computers Shared Address Space Disjoint Address Space Distributed Computing
  48. 48 Examples: SIMD MPP Thinking machines CM-2: The Classic Symmetric

    SIMD supercomputer (mid- 80’s): Description: Up to 64K bit- serial processing elements. Strength: Supports deterministic programming models  …  single  thread  of   control for ease of understanding. Weakness: Poor floating point performance. Programming model was not general enough. TMC struggled throughout  the  90’s  and  filed   for bankruptcy in 1994. Third party names are the property of their owners. “…  we  want  to  build  a  computer  that   will  be  proud  of  us”,  Danny  Hillis
  49. 49 Examples: Symmetric Multi-Processor Cray 2: The Classic Symmetric Multi-

    Processor (mid-80’s): Description: multiple Vector processors connected to a custom high speed memory. 500 MFLOP processors. Strength: Simple memory architecture makes this the easiest supercomputer in history to program. Truly an SMP (no caches). Weakness: Poor scalability. VERY expensive due to the fact that everything (memory to processors) were custom. Third party names are the property of their owners.
  50. 50 Examples: Massively Parallel Processors Paragon MP: The Classic MPP

    (early-90’s): Description:  3  i860  CPU’s  (a  vector   inspired microprocessor) connected by a custom mesh interconnect. 40 MFLOP processors*. Strength: A massively scalable machine (3000+ processors). The lights were pretty, but useful helping to show bottlenecks in the code. Weakness: Hard to program (NX message passing and later MPI). Expensive due to low volume microprocessor, custom back- plane and packaging. Third party names are the property of their owners.
  51. 51 Examples: Cluster NCSA’s  Itanium  cluster:   (early-00’s): Description: 160

    dual IPF nodes connected by a Myracom network. 3.2 GFLOP per processors. Strength: Highly scalable, nothing custom so hardware costs are reasonable. Weakness: Hard to program ( MPI). Lack of application software. Cluster middleware is fragile and still evolving. Third party names are the property of their owners.
  52. 52 Examples: distributed computing (e.g. GRID) Intel’s  Cure@home   program.

    Description: Thousands of home PC’s  donated  to  solve  important   problems. Strength: Highly scalable – the ultimiate costs performance since it uses compute-cycles that would otherwise be wasted. Weakness: Only coarse grained embarrassingly parallel algorithms can be used. Security constraints difficult to enforce. Target Protein Third party names are the property of their owners.
  53. 53 Outline  Introduction to parallel computing  Introduction to

    OpenMP  Creating Threads  Synchronization  Parallel Loops  Synchronize single masters and stuff  Data environment
  54. 54 OpenMP* Overview: omp_set_lock(lck) #pragma omp parallel for private(A, B)

    #pragma omp critical C$OMP parallel do shared(a, b, c) C$OMP PARALLEL REDUCTION (+: A, B) call OMP_INIT_LOCK (ilok) call omp_test_lock(jlok) setenv  OMP_SCHEDULE  “dynamic” CALL OMP_SET_NUM_THREADS(10) C$OMP DO lastprivate(XX) C$OMP ORDERED C$OMP SINGLE PRIVATE(X) C$OMP SECTIONS C$OMP MASTER C$OMP ATOMIC C$OMP FLUSH C$OMP PARALLEL DO ORDERED PRIVATE (A, B, C) C$OMP THREADPRIVATE(/ABC/) C$OMP PARALLEL COPYIN(/blk/) Nthrds = OMP_GET_NUM_PROCS() !$OMP BARRIER OpenMP: An API for Writing Multithreaded Applications A set of compiler directives and library routines for parallel application programmers Greatly simplifies writing multi-threaded (MT) programs in Fortran, C and C++ Standardizes last 20 years of SMP practice * The  name  “OpenMP”  is  the  property  of  the  OpenMP  Architecture  Review  Board.
  55. 55 OpenMP Basic Defs: Solution Stack OpenMP Runtime library OS/system

    support for shared memory and threading Directives, Compiler OpenMP library Environment variables Application End User Shared Address Space Proc3 Proc2 Proc1 ProcN
  56. 56 OpenMP core syntax  Most of the constructs in

    OpenMP are compiler directives. #pragma omp construct  [clause  [clause]…] Example #pragma omp parallel num_threads(4)  Function prototypes and types in the file: #include <omp.h>  Most OpenMP* constructs apply to a “structured  block”. Structured block: a block of one or more statements with one point of entry at the top and one point of exit at the bottom. It’s  OK  to  have  an  exit()  within  the  structured  block.
  57. 57 Exercise 1, Part A: Hello world Verify that your

    environment works  Write  a  program  that  prints  “hello  world”. int main() { int ID = 0; printf(“  hello(%d)  ”,  ID);; printf(“  world(%d)  \n”,  ID);; }
  58. 58 Exercise 1, Part B: Hello world Verify that your

    OpenMP environment works  Write  a  multithreaded  program  that  prints  “hello  world”. void main() { int ID = 0; printf(“  hello(%d)  ”,  ID);; printf(“  world(%d)  \n”,  ID);; } #pragma omp parallel { } #include <omp.h> Switches for compiling and linking gcc -fopenmp gcc pgcc -mp pgi icl /Qopenmp intel (windows) icc –openmp intel (linux)
  59. 59 Exercise 1: Solution A multi-threaded  “Hello  world”  program 

    Write a multithreaded program where each thread  prints  “hello  world”. #include  “omp.h” void main() { #pragma omp parallel { int ID = omp_get_thread_num(); printf(“  hello(%d)  ”,  ID);; printf(“  world(%d)  \n”,  ID);; } } Sample Output: hello(1) hello(0) world(1) world(0) hello (3) hello(2) world(3) world(2) OpenMP include file Parallel region with default number of threads Runtime library function to return a thread ID. End of the Parallel region
  60. 60 OpenMP Overview: How do threads interact?  OpenMP is

    a multi-threading, shared address model. – Threads communicate by sharing variables.  Unintended sharing of data causes race conditions: – race  condition:  when  the  program’s  outcome   changes as the threads are scheduled differently.  To control race conditions: – Use synchronization to protect data conflicts.  Synchronization is expensive so: – Change how data is accessed to minimize the need for synchronization.
  61. 61 Outline  Introduction to parallel computing  Introduction to

    OpenMP  Creating Threads  Synchronization  Parallel Loops  Synchronize single masters and stuff  Data environment
  62. 62 OpenMP Programming Model: Fork-Join Parallelism: Master thread spawns a

    team of threads as needed. Parallelism added incrementally until performance goals are met: i.e. the sequential program evolves into a parallel program. Parallel Regions Master Thread in red A Nested Parallel region Sequential Parts
  63. 63 Thread Creation: Parallel Regions  You create threads in

    OpenMP* with the parallel construct.  For example, To create a 4 thread Parallel region: double A[1000]; omp_set_num_threads(4); #pragma omp parallel { int ID = omp_get_thread_num(); pooh(ID,A); }  Each thread calls pooh(ID,A) for ID = 0 to 3 Each thread executes a copy of the code within the structured block Runtime function to request a certain number of threads Runtime function returning a thread ID * The  name  “OpenMP”  is  the  property  of  the  OpenMP  Architecture  Review  Board
  64. 64 Thread Creation: Parallel Regions  You create threads in

    OpenMP* with the parallel construct.  For example, To create a 4 thread Parallel region: double A[1000]; #pragma omp parallel num_threads(4) { int ID = omp_get_thread_num(); pooh(ID,A); }  Each thread calls pooh(ID,A) for ID = 0 to 3 Each thread executes a copy of the code within the structured block clause to request a certain number of threads Runtime function returning a thread ID * The  name  “OpenMP”  is  the  property  of  the  OpenMP  Architecture  Review  Board
  65. 65 Thread Creation: Parallel Regions example  Each thread executes

    the same code redundantly. double A[1000]; omp_set_num_threads(4); #pragma omp parallel { int ID = omp_get_thread_num(); pooh(ID, A); } printf(“all  done\n”);; omp_set_num_threads(4) pooh(1,A) pooh(2,A) pooh(3,A) printf(“all  done\n”);; pooh(0,A) double A[1000]; A single copy of A is shared between all threads. Threads wait here for all threads to finish before proceeding (i.e. a barrier) * The  name  “OpenMP”  is  the  property  of  the  OpenMP  Architecture  Review  Board
  66. 66 Exercises 2 to 4: Numerical Integration  4.0 (1+x2)

    dx =  0 1  F(xi )x   i = 0 N Mathematically, we know that: We can approximate the integral as a sum of rectangles: Where each rectangle has width x and height F(xi ) at the middle of interval i. 4.0 2.0 1.0 X 0.0
  67. 67 Exercises 2 to 4: Serial PI Program static long

    num_steps = 100000; double step; void main () { int i; double x, pi, sum = 0.0; step = 1.0/(double) num_steps; for (i=0;i< num_steps; i++){ x = (i+0.5)*step; sum = sum + 4.0/(1.0+x*x); } pi = step * sum; }
  68. 68 Exercise 2  Create a parallel version of the

    pi program using a parallel construct.  Pay close attention to shared versus private variables.  In addition to a parallel construct, you will need the runtime library routines int omp_get_num_threads(); int omp_get_thread_num(); double omp_get_wtime(); Time in Seconds since a fixed point in the past Thread ID or rank Number of threads in the team
  69. 69 Outline  Introduction to parallel computing  Introduction to

    OpenMP  Creating Threads  Synchronization  Parallel Loops  Synchronize single masters and stuff  Data environment
  70. 70 Discussed later Synchronization  High level synchronization: – critical

    – atomic – barrier – ordered  Low level synchronization – flush – locks (both simple and nested) Synchronization is used to impose order constraints and to protect access to shared data
  71. 71 Synchronization: critical  Mutual exclusion: Only one thread at

    a time can enter a critical region. float res; #pragma omp parallel { float B; int i, id, nthrds; id = omp_get_thread_num(); nthrds = omp_get_num_threads(); for(i=id;i<niters;i+=nthrds){ B = big_job(i); #pragma omp critical res += consume (B); } } Threads wait their turn – only one at a time calls consume()
  72. 72 Synchronization: Atomic  Atomic provides mutual exclusion but only

    applies to the update of a memory location (the update of X in the following example) #pragma omp parallel { double tmp, B; B = DOIT(); #pragma omp atomic X += big_ugly(B); } #pragma omp parallel { double tmp, B; B = DOIT(); tmp = big_ugly(B); #pragma omp atomic X += tmp; } Atomic only protects the read/update of X
  73. 73 Exercise 3  In exercise 2, you probably used

    an array to create space for each thread to store its partial sum.  If array elements happen to share a cache line, this leads to false sharing. – Non-shared data in the same cache line so each update  invalidates  the  cache  line  …  in  essence   “sloshing  independent  data”  back  and  forth   between threads.  Modify  your  “pi  program”  from  exercise  2  to   avoid false sharing due to the sum array.
  74. 74 Outline  Introduction to parallel computing  Introduction to

    OpenMP  Creating Threads  Synchronization  Parallel Loops  Synchronize single masters and stuff  Data environment
  75. 75 Discussed later SPMD vs. worksharing  A parallel construct

    by itself creates an SPMD or    “Single  Program  Multiple  Data”  program  …   i.e., each thread redundantly executes the same code.  How do you split up pathways through the code between threads within a team? This is called worksharing – Loop construct – Sections/section constructs – Single construct – Task construct
  76. 76 The loop worksharing Constructs  The loop worksharing construct

    splits up loop iterations among the threads in a team #pragma omp parallel { #pragma omp for for (I=0;I<N;I++){ NEAT_STUFF(I); } } Loop construct name: •C/C++: for •Fortran: do The  variable  I  is  made  “private”  to  each   thread by default. You could do this explicitly  with  a  “private(I)”  clause
  77. 77 Loop worksharing Constructs A motivating example for(i=0;i<N;i++) { a[i]

    = a[i] + b[i];} #pragma omp parallel { int id, i, Nthrds, istart, iend; id = omp_get_thread_num(); Nthrds = omp_get_num_threads(); istart = id * N / Nthrds; iend = (id+1) * N / Nthrds; if (id == Nthrds-1)iend = N; for(i=istart;i<iend;i++) { a[i] = a[i] + b[i];} } #pragma omp parallel #pragma omp for for(i=0;i<N;i++) { a[i] = a[i] + b[i];} Sequential code OpenMP parallel region OpenMP parallel region and a worksharing for construct
  78. 78 loop worksharing constructs: The schedule clause  The schedule

    clause affects how loop iterations are mapped onto threads  schedule(static [,chunk]) – Deal-out  blocks  of  iterations  of  size  “chunk”  to  each  thread.  schedule(dynamic[,chunk]) – Each  thread  grabs  “chunk”  iterations  off  a  queue  until  all  iterations   have been handled.  schedule(guided[,chunk]) – Threads dynamically grab blocks of iterations. The size of the block starts  large  and  shrinks  down  to  size  “chunk”  as  the  calculation   proceeds.  schedule(runtime) – Schedule and chunk size taken from the OMP_SCHEDULE environment variable (or the runtime library).  schedule(auto) – Schedule is left up to the runtime to choose (does not have to be any of the above).
  79. 79 Schedule Clause When To Use STATIC Pre-determined and predictable

    by the programmer DYNAMIC Unpredictable, highly variable work per iteration GUIDED Special case of dynamic to reduce scheduling overhead AUTO When the runtime can “learn”  from  previous   executions of the same loop loop work-sharing constructs: The schedule clause Least work at runtime : scheduling done at compile-time Most work at runtime : complex scheduling logic used at run-time
  80. 80 Combined parallel/worksharing construct  OpenMP  shortcut:  Put  the  “parallel”

     and  the   worksharing directive on the same line double res[MAX]; int i; #pragma omp parallel { #pragma omp for for (i=0;i< MAX; i++) { res[i] = huge(); } } These are equivalent double res[MAX]; int i; #pragma omp parallel for for (i=0;i< MAX; i++) { res[i] = huge(); }
  81. 81 Working with loops  Basic approach Find compute intensive

    loops Make the loop iterations independent .. So they can safely execute in any order without loop-carried dependencies Place the appropriate OpenMP directive and test int i, j, A[MAX]; j = 5; for (i=0;i< MAX; i++) { j +=2; A[i] = big(j); } int i, A[MAX]; #pragma omp parallel for for (i=0;i< MAX; i++) { int j = 5 + 2*(i+1); A[i] = big(j); } Remove loop carried dependence Note: loop index “i”  is  private  by   default
  82. #pragma omp parallel for collapse(2) for (int i=0; i<N; i++)

    { for (int j=0; j<M; j++) { ..... } } 82 Nested loops  Will form a single loop of length NxM and then parallelize that.  Useful if N is O(no. of threads) so parallelizing the outer loop makes balancing the load difficult. Number of loops to be parallelized, counting from the outside  For perfectly nested rectangular loops we can parallelize multiple loops in the nest with the collapse clause:
  83. 83 Reduction  We are combining values into a single

    accumulation variable  (ave)  …  there  is  a  true  dependence  between   loop  iterations  that  can’t  be  trivially  removed  This  is  a  very  common  situation  …  it  is  called  a   “reduction”.  Support for reduction operations is included in most parallel programming environments. double ave=0.0, A[MAX]; int i; for (i=0;i< MAX; i++) { ave + = A[i]; } ave = ave/MAX;  How do we handle this case?
  84. 84 Reduction  OpenMP reduction clause: reduction (op : list)

     Inside a parallel or a work-sharing construct: – A local copy of each list variable is made and initialized depending  on  the  “op”  (e.g.  0  for  “+”). – Updates occur on the local copy. – Local copies are reduced into a single value and combined with the original global value.  The  variables  in  “list”  must  be  shared  in  the  enclosing   parallel region. double ave=0.0, A[MAX]; int i; #pragma omp parallel for reduction (+:ave) for (i=0;i< MAX; i++) { ave + = A[i]; } ave = ave/MAX;
  85. 85 OpenMP: Reduction operands/initial-values  Many different associative operands can

    be used with reduction:  Initial values are the ones that make sense mathematically. Operator Initial value + 0 * 1 - 0 min Largest pos. number max Most neg. number C/C++ only Operator Initial value & ~0 | 0 ^ 0 && 1 || 0 Fortran Only Operator Initial value .AND. .true. .OR. .false. .NEQV. .false. .IEOR. 0 .IOR. 0 .IAND. All bits on .EQV. .true.
  86. 86 Exercise 4: Pi with loops  Go back to

    the serial pi program and parallelize it with a loop construct  Your goal is to minimize the number of changes made to the serial program.
  87. 87 Exercise 5: Optimizing loops  Parallelize the matrix multiplication

    program in the file matmul.c  Can you optimize the program by playing with how the loops are scheduled?
  88. 88 Outline  Introduction to parallel computing  Introduction to

    OpenMP  Creating Threads  Synchronization  Parallel Loops  Synchronize single masters and stuff  Data environment
  89. 89 Synchronization: Barrier  Barrier: Each thread waits until all

    threads arrive. // declare and initialize arrays A, B and C #pragma omp parallel { int id=omp_get_thread_num(); A[id] = big_calc1(id); #pragma omp barrier #pragma omp for for(i=0;i<N;i++){C[i]=big_calc3(i,A);} #pragma omp for nowait for(i=0;i<N;i++){ B[i]=big_calc2(C, i); } A[id] = big_calc4(id); } implicit barrier at the end of a parallel region implicit barrier at the end of a for worksharing construct no implicit barrier due to nowait
  90. 90 Master Construct  The master construct denotes a structured

    block that is only executed by the master thread.  The other threads just skip it (no synchronization is implied). #pragma omp parallel { do_many_things(); #pragma omp master { exchange_boundaries(); } #pragma omp barrier do_many_other_things(); }
  91. 91 Single worksharing Construct  The single construct denotes a

    block of code that is executed by only one thread (not necessarily the master thread).  A barrier is implied at the end of the single block (can remove the barrier with a nowait clause). #pragma omp parallel { do_many_things(); #pragma omp single { exchange_boundaries(); } do_many_other_things(); }
  92. 92 Sections worksharing Construct  The Sections worksharing construct gives

    a different structured block to each thread. #pragma omp parallel { #pragma omp sections { #pragma omp section X_calculation(); #pragma omp section y_calculation(); #pragma omp section z_calculation(); } } By  default,  there  is  a  barrier  at  the  end  of  the  “omp   sections”.    Use  the  “nowait”  clause  to  turn  off  the  barrier.
  93. 93 Synchronization: Lock routines  Simple Lock routines: A simple

    lock is available if it is unset. –omp_init_lock(), omp_set_lock(), omp_unset_lock(), omp_test_lock(), omp_destroy_lock()  Nested Locks A nested lock is available if it is unset or if it is set but owned by the thread executing the nested lock function –omp_init_nest_lock(), omp_set_nest_lock(), omp_unset_nest_lock(), omp_test_nest_lock(), omp_destroy_nest_lock() Note: a thread always accesses the most recent copy of the lock,  so  you  don’t  need  to  use  a  flush  on  the  lock  variable. A lock implies a memory fence (a “flush”) of all thread visible variables
  94. 94 Synchronization: Simple Locks  Example: conflicts are rare, but

    to play it safe, we must assure mutual exclusion for updates to histogram elements. #pragma omp parallel for for(i=0;i<NBUCKETS; i++){ omp_init_lock(&hist_locks[i]); hist[i] = 0; } #pragma omp parallel for for(i=0;i<NVALS;i++){ ival = (int) sample(arr[i]); omp_set_lock(&hist_locks[ival]); hist[ival]++; omp_unset_lock(&hist_locks[ival]); } for(i=0;i<NBUCKETS; i++) omp_destroy_lock(&hist_locks[i]); Free-up storage when done. One lock per element of hist Enforce mutual exclusion on update to hist array
  95. 95 Runtime Library routines  Runtime environment routines: – Modify/Check

    the number of threads – omp_set_num_threads(), omp_get_num_threads(), omp_get_thread_num(), omp_get_max_threads() – Are we in an active parallel region? – omp_in_parallel() – Do you want the system to dynamically vary the number of threads from one parallel construct to another? – omp_set_dynamic, omp_get_dynamic(); – How many processors in the system? – omp_num_procs() …plus a few less commonly used routines.
  96. 96 Runtime Library routines  To use a known, fixed

    number of threads in a program, (1)  tell  the  system  that  you  don’t  want  dynamic  adjustment  of   the number of threads, (2) set the number of threads, then (3) save the number you got. #include <omp.h> void main() { int num_threads; omp_set_dynamic( 0 ); omp_set_num_threads( omp_get_num_procs() ); #pragma omp parallel { int id=omp_get_thread_num(); #pragma omp single num_threads = omp_get_num_threads(); do_lots_of_stuff(id); } } Protect this op since Memory stores are not atomic Request as many threads as you have processors. Disable dynamic adjustment of the number of threads. Even in this case, the system may give you fewer threads than requested. If the precise # of threads matters, test for it and respond accordingly.
  97. 97 Environment Variables  Set the default number of threads

    to use. – OMP_NUM_THREADS int_literal  Control  how  “omp for  schedule(RUNTIME)”   loop iterations are scheduled. – OMP_SCHEDULE  “schedule[,  chunk_size]”  Process binding is enabled if this variable is true  …  i.e.  if  true  the  runtime  will  not  move   threads around between processors. – OMP_PROC_BIND true | false … Plus several less commonly used environment variables.
  98. 98 Outline  Introduction to parallel computing  Introduction to

    OpenMP  Creating Threads  Synchronization  Parallel Loops  Synchronize single masters and stuff  Data environment
  99. 99 Data environment: Default storage attributes  Shared Memory programming

    model: – Most variables are shared by default  Global variables are SHARED among threads – Fortran: COMMON blocks, SAVE variables, MODULE variables – C: File scope variables, static – Both: dynamically allocated memory (ALLOCATE, malloc, new)  But not everything is shared... – Stack variables in subprograms(Fortran) or functions(C) called from parallel regions are PRIVATE – Automatic variables within a statement block are PRIVATE.
  100. 100 double A[10]; int main() { int index[10]; #pragma omp

    parallel work(index); printf(“%d\n”,  index[0]);; } extern double A[10]; void work(int *index) { double temp[10]; static int count; ... } Data sharing: Examples temp A, index, count temp temp A, index, count A, index and count are shared by all threads. temp is local to each thread
  101. 101 Data sharing: Changing storage attributes  One can selectively

    change storage attributes for constructs using the following clauses* – SHARED – PRIVATE – FIRSTPRIVATE  The final value of a private inside a parallel loop can be transmitted to the shared variable outside the loop with: – LASTPRIVATE  The default attributes can be overridden with: – DEFAULT (PRIVATE | SHARED | NONE) All the clauses on this page apply to the OpenMP construct NOT to the entire region. *All data clauses apply to parallel constructs and worksharing constructs except  “shared”  which  only  applies  to  parallel  constructs. DEFAULT(PRIVATE) is Fortran only
  102. 102 Data Sharing: Private Clause void wrong() { int tmp

    = 0; #pragma omp parallel for private(tmp) for (int j = 0; j < 1000; ++j) tmp += j; printf(“%d\n”,  tmp); }  private(var) creates a new local copy of var for each thread. – The value of the private copies is uninitialized – The value of the original variable is unchanged after the region tmp was not initialized tmp is 0 here
  103. 103 Data Sharing: Private Clause When is the original variable

    valid? int tmp; void danger() { tmp = 0; #pragma omp parallel private(tmp) work(); printf(“%d\n”,  tmp);; }  The  original  variable’s  value  is  unspecified  if  it  is   referenced outside of the construct – Implementations may reference the original variable or a copy  …..  a  dangerous  programming  practice! extern int tmp; void work() { tmp = 5; } unspecified which copy of tmp tmp has unspecified value
  104. Firstprivate Clause  Variables initialized from shared variable  C++

    objects are copy-constructed 104 incr = 0; #pragma omp parallel for firstprivate(incr) for (i = 0; i <= MAX; i++) { if ((i%2)==0) incr++; A[i] = incr; } Each thread gets its own copy of incr with an initial value of 0
  105. Lastprivate Clause  Variables update shared variable using value from

    last iteration  C++ objects are updated as if by assignment void sq2(int n, double *lastterm) { double x; int i; #pragma omp parallel for lastprivate(x) for (i = 0; i < n; i++){ x = a[i]*a[i] + b[i]*b[i]; b[i] = sqrt(x); } *lastterm = x; } 105 “x”  has  the  value  it  held   for  the  “last  sequential”   iteration (i.e., for i=(n-1))
  106. 106 Data Sharing: A data environment test  Consider this

    example of PRIVATE and FIRSTPRIVATE  Are A,B,C local to each thread or shared inside the parallel region?  What are their initial values inside and values after the parallel region? variables A,B, and C = 1 #pragma omp parallel private(B) firstprivate(C) Inside this parallel region ...  “A”  is  shared  by  all  threads;;  equals  1  “B”  and  “C”  are  local  to  each  thread. – B’s  initial  value  is  undefined – C’s  initial  value  equals    1 Outside this parallel region ...  The  values  of  “B”  and  “C”  are  unspecified  if  referenced  in  the   region but outside the construct.
  107. 107 Data Sharing: Default Clause  Note that the default

    storage attribute is DEFAULT(SHARED) (so no need to use it)  Exception: #pragma omp task  To change default: DEFAULT(PRIVATE)  each variable in the construct is made private as if specified in a private clause  mostly saves typing  DEFAULT(NONE): no default for variables in static extent. Must list storage attribute for each variable in static extent. Good programming practice! Only the Fortran API supports default(private). C/C++ only has default(shared) or default(none).
  108. 108 Data Sharing: Default Clause Example itotal = 1000 C$OMP

    PARALLEL DEFAULT(PRIVATE) SHARED(itotal) np = omp_get_num_threads() each = itotal/np ……… C$OMP END PARALLEL itotal = 1000 C$OMP PARALLEL PRIVATE(np, each) np = omp_get_num_threads() each = itotal/np ……… C$OMP END PARALLEL These two code fragments are equivalent
  109. 109 Exercise 6: Mandelbrot set area  The supplied program

    (mandel.c) computes the area of a Mandelbrot set.  The program has been parallelized with OpenMP,  but  we  were  lazy  and  didn’t  do  it   right.  Find  and  fix  the  errors  (hint  …  the  problem  is   with the data environment).
  110. 110 Exercise 6 (cont.)  Once you have a working

    version, try to optimize the program? Try different schedules on the parallel loop. Try different mechanisms to support mutual exclusion  …  do  the  efficiencies  change?
  111. 111 Conclusion  We have now covered the most commonly

    used constructs in OpenMP. We’ve  skipped  tasks,  thread  private  data,  and  the   subtle details of the memory model needed by advanced programmers to build their own synchronization constructs.  Download  the  spec  to  learn  more  …  the  spec  is   filled with examples to support your continuing education. www.openmp.org  Get involved: get your organization to join the OpenMP ARB. Work with us through Compunity.
  112. 112 Extra content: (what  you’re  missing  from  the  full  day

     version  of  the  course).  Tasks  Memory model  Threadprivate Data  A survey of algorithms and programming models
  113. What are tasks?  Tasks are independent units of work

     Threads are assigned to perform the work of each task  Tasks may be deferred  Tasks may be executed immediately  The runtime system decides which of the above  Tasks are composed of: – code to execute – data environment – internal control variables (ICV) Serial Parallel
  114. Task Construct – Explicit Task View  A team of

    threads is created at the omp parallel construct  A single thread is chosen to execute the while loop – lets call  this  thread  “L”  Thread L operates the while loop, creates tasks, and fetches next pointers  Each time L crosses the omp task construct it generates a new task and has a thread assigned to it  Each task runs in its own thread  All tasks complete at the barrier at the end of the parallel  region’s  single   construct #pragma omp parallel { #pragma omp single { // block 1 node * p = head; while (p) { //block 2 #pragma omp task firstprivate(p) process(p); p = p->next; //block 3 } } }
  115. #pragma omp parallel num_threads(8) // assume 8 threads { #pragma

    omp single private(p) { … while (p) { #pragma omp task { processwork(p); } p = p->next; } } } Simple Task Example A pool of 8 threads is created here One thread gets to execute the while loop The  single  “while  loop”   thread creates a task for each instance of processwork()
  116. Why are tasks useful? #pragma omp parallel { #pragma omp

    single { // block 1 node * p = head; while (p) { //block 2 #pragma omp task process(p); p = p->next; //block 3 } } } Have potential to parallelize irregular patterns and recursive function calls Block 1 Block 2 Task 1 Block 2 Task 2 Block 2 Task 3 Block 3 Block 3 Time Single Threaded Block 1 Block 3 Block 3 Thr1 Thr2 Thr3 Thr4 Block 2 Task 2 Block 2 Task 1 Block 2 Task 3 Time Saved Idle Idle
  117. When are tasks guaranteed to complete  Tasks are guaranteed

    to be complete at thread barriers: #pragma omp barrier  …    or  task  barriers #pragma omp taskwait 117
  118. Task Completion Example #pragma omp parallel { #pragma omp task

    foo(); #pragma omp barrier #pragma omp single { #pragma omp task bar(); } } Multiple foo tasks created here – one for each thread All foo tasks guaranteed to be completed here One bar task created here bar task guaranteed to be completed here
  119. int fib ( int n ) { int x,y; if

    ( n < 2 ) return n; #pragma omp task x = fib(n-1); #pragma omp task y = fib(n-2); #pragma omp taskwait return x+y } Data Scoping with tasks: Fibonacci example. n is private in both tasks What’s  wrong  here? Can’t  use  private  variables   outside of tasks x is a private variable y is a private variable
  120. int fib ( int n ) { int x,y; if

    ( n < 2 ) return n; #pragma omp task shared (x) x = fib(n-1); #pragma omp task shared(y) y = fib(n-2); #pragma omp taskwait return x+y } Data Scoping with tasks: Fibonacci example. n is private in both tasks What’s  wrong  here? x & y are shared Good solution we need both values to compute the sum
  121. List ml; //my_list Element *e; #pragma omp parallel #pragma omp

    single { for(e=ml->first;e;e=e->next) #pragma omp task process(e); } Data Scoping with tasks: List Traversal example What’s  wrong  here? Possible data race ! Shared variable e updated by multiple tasks
  122. List ml; //my_list Element *e; #pragma omp parallel #pragma omp

    single { for(e=ml->first;e;e=e->next) #pragma omp task firstprivate(e) process(e); } Data Scoping with tasks: List Traversal example Good solution – e is firstprivate
  123. List ml; //my_list Element *e; #pragma omp parallel #pragma omp

    single private(e) { for(e=ml->first;e;e=e->next) #pragma omp task process(e); } Data Scoping with tasks: List Traversal example Good solution – e is private
  124. Recursive matrix multiplication  Quarter each input matrix and output

    matrix  Treat each submatrix as a single element and multiply  8 submatrix multiplications, 4 additions A B C A1,1 A1,2 A2,1 A2,2 B1,1 B1,2 B2,1 B2,2 C1,1 C1,2 C2,1 C2,2 C1,1 = A1,1 ·B1,1 + A1,2 ·B2,1 C2,1 = A2,1 ·B1,1 + A2,2 ·B2,1 C1,2 = A1,1 ·B1,2 + A1,2 ·B2,2 C2,2 = A2,1 ·B1,2 + A2,2 ·B2,2 124
  125. How to multiply submatrices?  Use the same routine that

    is computing the full matrix multiplication  Quarter each input submatrix and output submatrix  Treat each sub-submatrix as a single element and multiply A B C A1,1 A1,2 A2,1 A2,2 B1,1 B1,2 B2,1 B2,2 C1,1 C1,2 C2,1 C2,2 C111,1 = A111,1 ·B111,1 + A111,2 ·B112,1 + A121,1 ·B211,1 + A121,2 ·B212,1 C1,1 = A1,1 ·B1,1 + A1,2 ·B2,1 125 A1,1 A111,1 A111,2 A112,1 A112,2 B1,1 B111,1 B111,2 B112,1 B112,2 C1,1 C111,1 C111,2 C112,1 C112,2
  126. C1,1 = A1,1 ·B1,1 + A1,2 ·B2,1 C2,1 = A2,1

    ·B1,1 + A2,2 ·B2,1 C1,2 = A1,1 ·B1,2 + A1,2 ·B2,2 C2,2 = A2,1 ·B1,2 + A2,2 ·B2,2 Recursively multiply submatrices  Also need stopping criteria for recursion 126 void matmultrec(int mf, int ml, int nf, int nl, int pf, int pl, double **A, double **B, double **C) {// Dimensions: A[mf..ml][pf..pl] B[pf..pl][nf..nl] C[mf..ml][nf..nl] // C11 += A11*B11 matmultrec(mf, mf+(ml-mf)/2, nf, nf+(nl-nf)/2, pf, pf+(pl-pf)/2, A, B, C); // C11 += A12*B21 matmultrec(mf, mf+(ml-mf)/2, nf, nf+(nl-nf)/2, pf+(pl-pf)/2, pl, A, B, C); . . . }  Need range of indices to define each submatrix to be used
  127. 127 Exercise 7: Recursive matrix multiplication  Consider the program

    matmul_recur.c. This program implements a recursive algorithm for multiply two square matrices. Parallelize this program using OpenMP tasks.
  128. 128 Extra content: (what  you’re  missing  from  the  full  day

     version  of  the  course).  Tasks  Memory model  Threadprivate Data  A survey of algorithms and programming models
  129. 129 OpenMP memory model  Multiple copies of data may

    be present in various levels of cache, or in registers.  OpenMP supports a shared memory model.  All threads share an address space, but it can get complicated: proc1 proc2 proc3 procN Shared memory cache1 cache2 cache3 cacheN a a . . .
  130. 130 OpenMP and Relaxed Consistency  OpenMP supports a relaxed-consistency

    shared memory model. Threads can maintain a temporary view of shared memory which is not consistent with that of other threads. These temporary views are made consistent only at certain points in the program. The operation which enforces consistency is called the flush operation
  131. 131 Flush operation  Defines a sequence point at which

    a thread is guaranteed to see a consistent view of memory All previous read/writes by this thread have completed and are visible to other threads No subsequent read/writes by this thread have occurred A flush operation is analogous to a fence in other shared  memory  API’s
  132. 132 Synchronization: flush example  Flush forces data to be

    updated in memory so other threads see the most recent value double A; A = compute(); #pragma omp flush(A) // flush to memory to make sure other // threads can pick up the right value Note: OpenMP’s flush is analogous to a fence in other  shared  memory  API’s.
  133. 133 Flush and synchronization  A flush operation is implied

    by OpenMP synchronizations, e.g. at entry/exit of parallel regions at implicit and explicit barriers at entry/exit of critical regions whenever a lock is set or unset …. (but not at entry to worksharing regions or entry/exit of master regions)
  134. 134 What is the Big Deal with Flush?  Compilers

    routinely reorder instructions implementing a program  This helps better exploit the functional units, keep machine busy, hide memory latencies, etc.  Compiler generally cannot move instructions:  past a barrier  past a flush on all variables  But it can move them past a flush with a list of variables so long as those variables are not accessed  Keeping track of consistency when flushes are used can  be  confusing  …  especially  if  “flush(list)”  is  used. Note: the flush operation does not actually synchronize different threads. It just ensures that a thread’s values are made consistent with main memory.
  135. 135 Example: prod_cons.c int main() { double *A, sum, runtime;

    int flag = 0; A = (double *)malloc(N*sizeof(double)); runtime = omp_get_wtime(); fill_rand(N, A); // Producer: fill an array of data sum = Sum_array(N, A); // Consumer: sum the array runtime = omp_get_wtime() - runtime; printf(" In %lf secs, The sum is %lf \n",runtime,sum); } • Parallelize a producer consumer program – One thread produces values that another thread consumes. – The key is to implement pairwise synchronization between threads. – Often used with a stream of produced values to implement “pipeline   parallelism”
  136. 136 Pair wise synchronizaion in OpenMP  OpenMP lacks synchronization

    constructs that work between pairs of threads.  When this is needed you have to build it yourself.  Pair wise synchronization Use a shared flag variable Reader spins waiting for the new flag value Use flushes to force updates to and from memory
  137. 137 Example: producer consumer int main() { double *A, sum,

    runtime; int numthreads, flag = 0; A = (double *)malloc(N*sizeof(double)); #pragma omp parallel sections { #pragma omp section { fill_rand(N, A); #pragma omp flush flag = 1; #pragma omp flush (flag) } #pragma omp section { #pragma omp flush (flag) while (flag == 0){ #pragma omp flush (flag) } #pragma omp flush sum = Sum_array(N, A); } } } Use flag to Signal when the “produced” value is ready Flush forces refresh to memory. Guarantees that the other thread sees the new value of A Notice you must put the flush inside the while loop to make sure the updated flag variable is seen Flush needed on both “reader” and “writer” sides of the communication
  138. The OpenMP 3.1 atomics (1 of 2)  Atomic was

    expanded to cover the full range of common scenarios where you need to protect a memory operation so it occurs atomically: # pragma omp atomic [read | write | update | capture] 138 • Atomic can protect loads # pragma omp atomic read v = x; • Atomic can protect stores # pragma omp atomic write x = expr; • Atomic can protect updates to a storage location (this is the default behavior  …  i.e.  when  you  don’t  provide  a  clause) # pragma omp atomic update x++; or ++x; or x--; or –x; or x binop= expr; or x = x binop expr; This is the original OpenMP atomic
  139. The OpenMP 3.1 atomics (2 of 2)  Atomic can

    protect the assignment of a value (its capture) AND an associated update operation: # pragma omp atomic capture statement or structured block 139 • Where the statement is one of the following forms: v = x++; v = ++x; v = x--; v = –x; v = x binop expr; • Where the structured block is one of the following forms: {v = x; x binop = expr;} {x binop = expr; v = x;} {v=x; x=x binop expr;} {X = x binop expr; v = x;} {v = x; x++;} {v=x; ++x:} {++x; v=x:} {x++; v = x;} {v = x; x--;} {v= x; --x;} {--x; v = x;} {x--; v = x;} The capture semantics in atomic were added to map onto common hardware supported atomic ops and to support modern lock free algorithms.
  140. Atomics and synchronization flags 140 int main() { double *A,

    sum, runtime; int numthreads, flag = 0, flg_tmp; A = (double *)malloc(N*sizeof(double)); #pragma omp parallel sections { #pragma omp section { fill_rand(N, A); #pragma omp flush #pragma atomic write flag = 1; #pragma omp flush (flag) } #pragma omp section { while (1){ #pragma omp flush(flag) #pragma omp atomic read flg_tmp= flag; if (flg_tmp==1) break; } #pragma omp flush sum = Sum_array(N, A); } } } This program is truly race  free  …  the  reads   and writes of flag are protected so the two threads can not conflict.
  141. 141 Extra content: (what  you’re  missing  from  the  full  day

     version  of  the  course).  Tasks  Memory model  Threadprivate Data  A survey of algorithms and programming models
  142. 142 Data sharing: Threadprivate  Makes global data private to

    a thread  Fortran: COMMON blocks  C: File scope and static variables, static class members  Different from making them PRIVATE  with PRIVATE global variables are masked.  THREADPRIVATE preserves global scope within each thread  Threadprivate variables can be initialized using COPYIN or at time of definition (using language- defined initialization capabilities).
  143. 143 A threadprivate example (C) int counter = 0; #pragma

    omp threadprivate(counter) int increment_counter() { counter++; return (counter); } Use threadprivate to create a counter for each thread.
  144. 144 Data Copying: Copyin parameter (N=1000) common/buf/A(N) !$OMP THREADPRIVATE(/buf/) C

    Initialize the A array call init_data(N,A) !$OMP PARALLEL COPYIN(A) …  Now  each  thread  sees  threadprivate array A initialied …  to  the  global  value  set  in  the  subroutine  init_data() !$OMP END PARALLEL end You initialize threadprivate data using a copyin clause.
  145. 145 Data Copying: Copyprivate #include <omp.h> void input_parameters (int, int);

    // fetch values of input parameters void do_work(int, int); void main() { int Nsize, choice; #pragma omp parallel private (Nsize, choice) { #pragma omp single copyprivate (Nsize, choice) input_parameters (Nsize, choice); do_work(Nsize, choice); } } Used with a single region to broadcast values of privates from one member of a team to the rest of the team.
  146. 146 Exercise 8: Monte Carlo Calculations Using Random numbers to

    solve tough problems  Sample a problem domain to estimate areas, compute probabilities, find optimal values, etc.  Example: Computing π with a digital dart board:  Throw darts at the circle/square.  Chance of falling in circle is proportional to ratio of areas: Ac = r2 * π As = (2*r) * (2*r) = 4 * r2 P = Ac /As = π /4  Compute π by randomly choosing points, count the fraction that falls in the circle, compute pi. 2 * r N= 10 π = 2.8 N=100 π = 3.16 N= 1000 π = 3.148
  147. 147 Exercise 8  We provide three files for this

    exercise  pi_mc.c: the monte carlo method pi program  random.c: a simple random number generator  random.h: include file for random number generator  Create a parallel version of this program without changing the interfaces to functions in random.c  This  is  an  exercise  in  modular  software  …  why  should  a  user   of your parallel random number generator have to know any details of the generator or make any changes to how the generator is called?  The random number generator must be threadsafe.  Extra Credit:  Make your random number generator numerically correct (non- overlapping sequences of pseudo-random numbers).
  148. 148 Conclusion  We have now covered the full sweep

    of the OpenMP specification. We’ve  left  off  some  minor  details,  but  we’ve  covered   all  the  major  topics  …  remaining  content  you  can   pick up on your own.  Download  the  spec  to  learn  more  …  the  spec  is   filled with examples to support your continuing education. www.openmp.org  Get involved: get your organization to join the OpenMP ARB. Work with us through Compunity.
  149. 149 Extra content: (what  you’re  missing  from  the  full  day

     version  of  the  course).  Tasks  Memory model  Threadprivate Data  A survey of algorithms and programming models
  150. © 2009 Mathew J. Sottile, Timothy G. Mattson, and Craig

    E Rasmussen 151 Getting started with parallel algorithms • Concurrency is a general concept – …  multiple  activities  that  can  occur  and  make  progress   at the same time. • A parallel algorithm is any algorithm that uses concurrency to solve a problem of a given size in less time • Scientific programmers have been working with parallelism  since  the  early  80’s – Hence we have almost 30 years of experience to draw on to help us understand parallel algorithms.
  151. © 2009 Mathew J. Sottile, Timothy G. Mattson, and Craig

    E Rasmussen 152 A formal discipline of design • Christopher  Alexander’s  approach  to  (civil)   architecture: – A design pattern “describes  a  problem  which   occurs over and over again in our environment, and then describes the core of the solution to that problem, in such a way that you can use this solution a million times over, without ever doing it the  same  way  twice.“  Page x, A Pattern Language, Christopher Alexander • A pattern language is an organized way of tackling an architectural problem using patterns • The gang of 4 used patterns to bring order to the chaos of object oriented design. • The  book  “over  night”  turned  object   oriented  design  from  “an  art”  to  a   systematic design discipline.
  152. © 2009 Mathew J. Sottile, Timothy G. Mattson, and Craig

    E Rasmussen 153 153 Can Design patterns bring order to parallel programming? • The  book  “Patterns  for  Parallel   Programming”  contains  a   design pattern language to capture how experts think about parallel programming. • It is an attempt to be to parallel programming what the GOF book was to object oriented programming. • The patterns were mined from established practice in scientific  computing  …  hence   its a useful set of patterns but not complete (e.g. its weak on graph algorithms).
  153. © 2009 Mathew J. Sottile, Timothy G. Mattson, and Craig

    E Rasmussen 154 Basic approach from the book • Identify the concurrency in your problem: – Find the tasks, data dependencies and any other constraints. • Develop a strategy to exploit this concurrency: – Which elements of the parallel design will be used to organize your approach to exploiting concurrency. • Identify and use the right algorithm pattern to turn your strategy into the design of a specific algorithm. • Choose the supporting patterns to move your design into source code. – This step is heavily influenced by the target platform
  154. © 2009 Mathew J. Sottile, Timothy G. Mattson, and Craig

    E Rasmussen 155 Original Problem Tasks, shared and local data Find Concurrency Supporting patterns Corresponding source code Program SPMD_Emb_Par () { TYPE *tmp, *func(); global_array Data(TYPE); global_array Res(TYPE); int N = get_num_procs(); int id = get_proc_id(); if (id==0) setup_problem(N,DATA); for (int I= 0; I<N;I=I+Num){ tmp = func(I); Res.accumulate( tmp); } } Program SPMD_Emb_Par () { TYPE *tmp, *func(); global_array Data(TYPE); global_array Res(TYPE); int N = get_num_procs(); int id = get_proc_id(); if (id==0) setup_problem(N,DATA); for (int I= 0; I<N;I=I+Num){ tmp = func(I); Res.accumulate( tmp); } } Program SPMD_Emb_Par () { TYPE *tmp, *func(); global_array Data(TYPE); global_array Res(TYPE); int N = get_num_procs(); int id = get_proc_id(); if (id==0) setup_problem(N,DATA); for (int I= 0; I<N;I=I+Num){ tmp = func(I); Res.accumulate( tmp); } } Program SPMD_Emb_Par () { TYPE *tmp, *func(); global_array Data(TYPE); global_array Res(TYPE); int Num = get_num_procs(); int id = get_proc_id(); if (id==0) setup_problem(N, Data); for (int I= ID; I<N;I=I+Num){ tmp = func(I, Data); Res.accumulate( tmp); } } Units of execution + new shared data for extracted dependencies Concurrency in Parallel software:
  155. © 2009 Mathew J. Sottile, Timothy G. Mattson, and Craig

    E Rasmussen 156 Strategies for exploiting concurrency • Given  the  results  from  your  “finding  concurrency”   analysis, there are many different ways to turn them into a parallel algorithm. • In most cases, one of three Distinct Strategies are used – Agenda parallelism: The collection of tasks that are to be computed. – Result parallelism: Updates to the data. – Specialist parallelism: The flow of data between a fixed set of tasks. Ref: N. Carriero and D. Gelernter, How to Write Parallel Programs: A First Course, 1990.
  156. © 2009 Mathew J. Sottile, Timothy G. Mattson, and Craig

    E Rasmussen 157 The Algorithm Design Patterns Result Parallelism Geometric Decomposition Task Parallelism Divide and Conquer Recursive Data Specialist Parallelism Pipeline Event Based Coordination Agenda Parallelism Data Parallelism Embarrassingly Parallel Separable Dependencies Start with a basic concurrency decomposition • A problem decomposed into a set of tasks • A  data  decomposition  aligned  with  the  set  of  tasks  …  designed  to  minimize   interactions between tasks and make concurrent updates to data safe. • Dependencies and ordering constraints between groups of tasks.
  157. 158 Implementation strategy Patterns (Supporting Structures) Patterns that support Implementing

    Algorithm strategies as code. Program Structure Task-queue SPMD Loop Parallelism Fork/Join Data Structures Shared Data Shared Queue Partitioned Array Index-map Actors Partitioned Graph Shared Map
  158. 159 Implementation strategy Patterns (Supporting Structures) Patterns that support Implementing

    Algorithm strategies as code. Program Structure Task-queue SPMD Loop Parallelism Fork/Join Data Structures Shared Data Shared Queue Partitioned Array Index-map Actors Partitioned Graph Shared Map We will only have time to consider these 4 patterns.
  159. 160 The role for Parallel programming languages  A pattern

    is text that explains a design concept. It is not a working program.  We need to move from a design expressed as a composition of  patterns  into  source  code  …  ultimately  you  need  a   programming language/API.  In the following slides we will survey some key patterns and introduce key programming languages along the way:  OpenMP: Directive driven parallelism for shared memory computers.  Cilk: fine grained fork-join for shared memory  Thread libs (pthreads or windows threads): low level multi-threading  MPI: The Message passing standard from the HPC world.  OpenCL: New standard for heterogeneous platforms 3rd party names are the property of their owners.
  160. 161 A simple Example: The PI program Numerical Integration 

    4.0 (1+x2) dx =  0 1  F(xi )x   i = 0 N Mathematically, we know that: We can approximate the integral as a sum of rectangles: Where each rectangle has width x and height F(xi ) at the middle of interval i. 4.0 2.0 1.0 X 0.0
  161. 162 PI Program: The sequential program static long num_steps =

    100000; double step; void main () { int i; double x, pi, sum = 0.0; step = 1.0/(double) num_steps; for (i=1;i<= num_steps; i++){ x = (i-0.5)*step; sum = sum + 4.0/(1.0+x*x); } pi = step * sum; }
  162. 163 Task Parallelism Pattern  Use when:  The problem

    naturally decomposes into a distinct collection of tasks  Solution  Define the set of tasks and a way to detect when the computation is done.  Manage  (or  “remove”)  dependencies  so  the  correct  answer  is   produced regardless of the details of how the tasks execute.  Schedule the tasks for execution in a way that keeps the work balanced between the processing elements of the parallel computer and  Note:  This pattern includes the well known embarrassingly parallel pattern (no dependencies).
  163. 164 PI Program: Task Parallelism static long num_steps = 100000;

    double step; void main () { int i; double x, pi, sum = 0.0; step = 1.0/(double) num_steps; for (i=1;i<= num_steps; i++){ x = (i-0.5)*step; sum = sum + 4.0/(1.0+x*x); } pi = step * sum; } Modify loop to make iterations independent and define blocks of loop iterations as an explicit task Can support the execution of the resulting tasks with the fork-join, loop-level parallelism, or SPMD patterns
  164. 165 Native Thread Libraries  Linux and Windows both include

    native threads for shared address space programming  API provides:  Thread creation (fork)  Thread destruction (join)  Synchronization.  Programmer  is  in  control  …  these  are  very  general.  Downside: programmer MUST control everything. 3rd party names are the property of their owners.
  165. Fork Join Pattern: Win32 API Fork/join Pattern: (1) Package concurrent

    tasks in a function, (2) fork threads to run each function, (3) join when done, and (4) manage dependencies. #include <windows.h> #define NUM_THREADS 2 HANDLE thread_handles[NUM_THREADS]; CRITICAL_SECTION hUpdateMutex; static long num_steps = 100000; double step; double global_sum = 0.0; void Pi (void *arg) { int i, start; double x, sum = 0.0; start = *(int *) arg; step = 1.0/(double) num_steps; for (i=start;i<= num_steps; i=i+NUM_THREADS){ x = (i-0.5)*step; sum = sum + 4.0/(1.0+x*x); } EnterCriticalSection(&hUpdateMutex); global_sum += sum; LeaveCriticalSection(&hUpdateMutex); } void main () { double pi; int i; DWORD threadID; int threadArg[NUM_THREADS]; for(i=0; i<NUM_THREADS; i++) threadArg[i] = i+1; InitializeCriticalSection(&hUpdateMutex); for (i=0; i<NUM_THREADS; i++){ thread_handles[i] = CreateThread(0, 0, (LPTHREAD_START_ROUTINE) Pi, &threadArg[i], 0, &threadID); } WaitForMultipleObjects(NUM_THREADS, thread_handles, TRUE,INFINITE); pi = global_sum * step; printf(" pi is %f \n",pi); } Put work into a function Launch threads to execute the function Wait until the threads are done Protect update to shared data
  166. 167 OpenMP Programming Model: Fork-Join Parallelism:  Master thread spawns

    a team of threads as needed.  Parallelism added incrementally until performance goals are met: i.e. the sequential program evolves into a parallel program. Parallel Regions Master Thread in red A Nested Parallel region Sequential Parts 3rd party names are the property of their owners.
  167. 168 OpenMP PI Program: Loop level parallelism pattern #include <omp.h>

    static long num_steps = 100000; double step; #define NUM_THREADS 2 void main () { int i; double x, pi, sum =0.0; step = 1.0/(double) num_steps; omp_set_num_threads(NUM_THREADS); #pragma omp parallel for private(x) reduction (+:sum) for (i=0;i< num_steps; i++){ x = (i+0.5)*step; sum += 4.0/(1.0+x*x); } pi = sum[i] * step; } Loop Level Parallelism: Parallelism expressed solely by (1) exposing concurrency, (2) managing dependencies, and (3) splitting up loops . 3rd party names are the property of their owners.
  168. 169 Parallel  API’s:  MPI the Message Passing Interface omp_set_lock(lck) MPI_Bsend_init

    MPI_Pack MPI_Sendrecv_replace MPI_Recv_init MPI_Allgatherv MPI_Unpack MPI_Sendrecv MPI_Bcast MPI_Ssend C$OMP ORDERED MPI_Startall MPI_Test_cancelled MPI_Type_free MPI_Type_contiguous MPI_Barrier MPI_Start MPI_COMM_WORLD MPI_Recv MPI_Send MPI_Waitall MPI_Reduce MPI_Alltoallv MPI_Group_compare MPI_Scan MPI_Group_size MPI_Errhandler_create MPI: An API for Writing Clustered Applications  A library of routines to coordinate the execution of multiple processes.  Provides point to point and collective communication in Fortran, C and C++  Unifies last 15 years of cluster computing and MPP practice Third party names are the property of their owners.
  169. 170 MPI and The SPMD Model Replicate the program. Add

    glue code Break up the data A sequential program working on a data set •A parallel program working on a decomposed data set. • Coordination by passing messages. MPI is a standard API for passing messages between threads and processes. Works on distributed and shared memory systems. The number one API for parallel programming. 3rd party names are the property of their owners.
  170. 171 MPI Pi program: SPMD pattern SPMD Programs: Each thread

    runs the same code with the thread ID selecting thread-specific behavior. #include <mpi.h> void main (int argc, char *argv[]) { int i, id, numprocs; double x, pi, step, sum = 0.0, step1, stepN ; step = 1.0/(double) num_steps ; MPI_Init(&argc, &argv) ; MPI_Comm_Rank(MPI_COMM_WORLD, &id) ; MPI_Comm_Size(MPI_COMM_WORLD, &numprocs) ; step1 = id *num_steps/numprocs ; stepN = (id+1)*num_steps/numprocs; if (stepN > num_steps) stepN = num_steps; for (i=step1; i<stepN; i++) { x = (i+0.5)*step; sum += 4.0/(1.0+x*x); } sum *= step ; MPI_Reduce(&sum, &pi, 1, MPI_DOUBLE, MPI_SUM, 0, MPI_COMM_WORLD) ; }
  171. 172 Divide and Conquer Pattern  Use when:  A

    problem includes a method to divide into subproblems and a way to recombine solutions of subproblems into a global solution.  Solution  Define a split operation  Continue to split the problem until subproblems are small enough to solve directly.  Recombine solutions to subproblems to solve original global problem.  Note:  Computing may occur at each phase (split, leaves, recombine).
  172. 173 Divide and conquer  Split the problem into smaller

    sub-problems. Continue until the sub-problems can be solve directly.  3 Options:  Do work as you split into sub-problems.  Do work only at the leaves.  Do work as you recombine.
  173. 174 Cilk: divide and conquer meets fork-join  Extends C

    to create a parallel language but maintains serial semantics.  A fork-join style task oriented programming model perfect for recursive algorithms (e.g. branch-and-bound)  …  shared  memory  machines  only!  Solid  theoretical  foundation  …  can  prove  performance  theorems.   cilk Marks  a  function  as  a  “cilk”  function  that  can  be  spawned spawn Spawns  a  cilk  function  …  only  2  to  5  times  the  cost  of  a   regular function call sync Wait until immediate children spawned functions return  “Advanced”  key  words inlet Define a function to handle return values from a cilk task cilk_fence A portable memory fence. abort Terminate all currently existing spawned tasks  Includes locks and a few other odds and ends. 3rd party names are the property of their owners.
  174. 175 © 2009 Matthew J. Sottile, Timothy G. Mattson, and

    Craig E Rasmussen 175 PI Program: Cilk (divide and conquer implemented with fork-join) static long num_steps =  1073741824;;        //  I’m  lazy  …  make  it  a  power  of  2 double step = 1.0/(double) num_steps; cilk double pi_comp(int istep, int nstep){ double x, sum; if(nstep < MIN_SIZE) for (int i=istep, sum=0.0; i<= nstep; i++){ x = (i+0.5)*step; sum += 4.0/(1.0+x*x); } return sum; else { sum1 = spawn pi_comp(istep, nstep/2); sum2 = spawn pi_comp(istep+nstep/2, nstep/2); } sync; return sum1+sum2; } int main () double pi, sum = spawn pi_comp(0,num_steps); sync; pi = step * sum; } Recursively split range of the loop until its small enough to just directly compute Wait until child tasks are done then return the sum …  implements  a  balanced  binary  tree  reduction!
  175. © 2009 Matthew J. Sottile, Timothy G. Mattson, and Craig

    E Rasmussen 176 Cilk and OpenMP • With the task construct in OpenMP 3.0, we can use the Cilk style of programming in OpenMP. • In the following examples, we show an OpenMP program to count instances of a key in an array.
  176. © 2009 Matthew J. Sottile, Timothy G. Mattson, and Craig

    E Rasmussen 177 Count keys: Main program #define N 131072 int main() { long a[N]; int i; long key = 42, nkey=0; // fill the array and make sure it has a few instances of the key for (i=0;i<N;i++) a[i] = random()%N; a[N%43] = key; a[N%73] = key; a[N%3] = key; // count key in a with geometric decomposition nkey = search_geom(N, a, key); // count key in a with recursive splitting nkey = search_recur(N, a, key); }
  177. © 2009 Matthew J. Sottile, Timothy G. Mattson, and Craig

    E Rasmussen 178 Count keys: with OpenMP long search_recur(long Nlen, long *a, long key) { long count = 0; #pragma omp parallel reduction(+:count) { #pragma omp single count = search(Nlen, a, key); } return count; } • Design Patterns used: - Divide and conquer with fork=join long search(long Nlen, long *a, long key) { long count1=0, count2=0, Nl2; if (Nlen == 2){ if (*(a) == key) count1=1; if (*(a+1) == key) count2=1; return count1+count2; } else { Nl2 = Nlen/2; #pragma omp task shared(count1) count1 = search(Nl2, a, key); #pragma omp task shared(count2) count2 = search(Nl2, a+Nl2, key); #pragma omp taskwait return count1+count2; } }
  178. © 2009 Matthew J. Sottile, Timothy G. Mattson, and Craig

    E Rasmussen 179 Count keys: Random number generator // A simple random number generator static unsigned long long MULTIPLIER = 764261123; static unsigned long long PMOD = 2147483647; unsigned long long random_last = 42; long random() { unsigned long long random_next; // // compute an integer random number from zero to pmod // random_next = (unsigned long long)(( MULTIPLIER * random_last)% PMOD); random_last = random_next; return (long) random_next; }
  179. 180 Data Parallelism Pattern  Use when:  Your problem

    is defined in terms of collections of data elements operated on by a similar (if not identical) sequence of instructions; i.e. the concurrency is in the data.  Solution  Define collections of data elements that can be updated in parallel.  Define computation as a sequence of collective operations applied together to each data element. Data 1 Data 2 Data 3 Data n Tasks …… Often supported with the data-parallel/Index-map pattern.
  180. - Page 181 OpenCL: a language designed for the data-

    parallel index-map pattern CPUs Multiple cores driving performance increases GPUs Increasingly general purpose data-parallel computing Graphics APIs and Shading Languages Multi-processor programming – e.g. OpenMP Emerging Intersection Heterogeneous Computing OpenCL – Open Computing Language Open, royalty-free standard for portable, parallel programming of heterogeneous parallel computing CPUs, GPUs, and other processors 3rd party names are the property of their owners. Source: SC09 OpenCL tutorial
  181. - Page 182 OpenCL Platform Model • One Host +

    one or more Compute Devices - Each Compute Device is composed of one or more Compute Units - Each Compute Unit is further divided into one or more Processing Elements Source: SC09 OpenCL tutorial
  182. - Page 183 The data parallel index-map pattern: • define

    a problem domain in terms of an index-map and execute a kernel invocation for each point in the domain - E.g., process a 1024 x 1024 image: Global problem dimensions: 1024 x 1024 = 1 kernel execution per pixel: 1,048,576 total kernel executions void scalar_mul(int n, const float *a, const float *b, float *c) { int i; for (i=0; i<n; i++) c[i] = a[i] * b[i]; } Scalar kernel void dp_mul(global const float *a, global const float *b, global float *c) { int id = get_global_id(0); c[id] = a[id] * b[id]; }  //  execute  over  “n”  work-items Data Parallel Source: SC09 OpenCL tutorial
  183. - Page 184 An N-dimension domain of work-items • Global

    Dimensions: 1024 x 1024 (whole problem space) • Local Dimensions: 128 x 128 (work group … executes together) 1024 1024 Synchronization between work-items possible only within workgroups: barriers and memory fences Cannot synchronize outside of a workgroup • Choose the dimensions that are “best” for your algorithm
  184. - Page 185 Vector Addition - Host Program // create

    the OpenCL context on a GPU device cl_context = clCreateContextFromType(0, CL_DEVICE_TYPE_GPU, NULL, NULL, NULL); // get the list of GPU devices associated with context clGetContextInfo(context, CL_CONTEXT_DEVICES, 0, NULL, &cb); devices = malloc(cb); clGetContextInfo(context, CL_CONTEXT_DEVICES, cb, devices, NULL); // create a command-queue cmd_queue = clCreateCommandQueue(context, devices[0], 0, NULL); // allocate the buffer memory objects memobjs[0] = clCreateBuffer(context, CL_MEM_READ_ONLY | CL_MEM_COPY_HOST_PTR, sizeof(cl_float)*n, srcA, NULL);} memobjs[1] = clCreateBuffer(context,CL_MEM_READ_ONLY | CL_MEM_COPY_HOST_PTR, sizeof(cl_float)*n, srcB, NULL); memobjs[2] = clCreateBuffer(context,CL_MEM_WRITE_ONLY, sizeof(cl_float)*n, NULL, NULL); // create the program program = clCreateProgramWithSource(context, 1, &program_source, NULL, NULL); // build the program err = clBuildProgram(program, 0, NULL, NULL, NULL, NULL); // create the kernel kernel  =  clCreateKernel(program,  “vec_add”,  NULL);; // set the args values err = clSetKernelArg(kernel, 0, (void *) &memobjs[0], sizeof(cl_mem)); err |= clSetKernelArg(kernel, 1, (void *)&memobjs[1], sizeof(cl_mem)); err |= clSetKernelArg(kernel, 2, (void *)&memobjs[2], sizeof(cl_mem)); // set work-item dimensions global_work_size[0] = n; // execute kernel err = clEnqueueNDRangeKernel(cmd_queue, kernel, 1, NULL, global_work_size, NULL, 0, NULL, NULL); // read output array err = clEnqueueReadBuffer(context, memobjs[2], CL_TRUE, 0, n*sizeof(cl_float), dst, 0, NULL, NULL); Define platform and queues Define Memory objects Create the program Build the program Create and setup kernel Execute the kernel Read results on the host It’s complicated, but most of this is “boilerplate” and not as bad as it looks.
  185. Conclusion  While there are vast numbers of different parallel

    algorithms  “out  there”  most  are  composed  of  a  modest   number of design patterns.  We can use these patterns to help understand these algorithms and how they map onto different systems.  There are hundreds (thousands?) of programming models in existence. But only a small number are actually used:  OpenMP  Pthreads  MPI  TBB  Cilk  OpenCL  CUDA 186
  186. If you want to learn more about parallel programming  languages

     …   187 A look at the history of concurrency in programming languages and a summary of key concurrent programming languages in use today. This is the greatest computer science book ever written. Everyone should own  a  copy  of  this  book  …  it’s  really   that good!!! -- Tim  Mattson,  ESC’2012 We even have PowerPoint slides online so lazy professors can use it as a text book. (see www.parlang.com).
  187. 188 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  188. 189 OpenMP Organizations  OpenMP architecture review board URL, the

     “owner”  of  the  OpenMP  specification: www.openmp.org  OpenMP  User’s  Group  (cOMPunity)  URL: www.compunity.org Get involved, join compunity and help define the future of OpenMP
  189. 190 Books about OpenMP  A new book about OpenMP

    2.5 by a team of authors at the forefront of OpenMP’s  evolution.  A book about how to “think  parallel”  with   examples in OpenMP, MPI and java
  190. 191 OpenMP Papers  Sosa CP, Scalmani C, Gomperts R,

    Frisch MJ. Ab initio quantum chemistry on a ccNUMA architecture using OpenMP. III. Parallel Computing, vol.26, no.7-8, July 2000, pp.843-56. Publisher: Elsevier, Netherlands.  Couturier R, Chipot C. Parallel molecular dynamics using OPENMP on a shared memory machine. Computer Physics Communications, vol.124, no.1, Jan. 2000, pp.49-59. Publisher: Elsevier, Netherlands.  Bentz  J.,  Kendall  R.,  “Parallelization  of  General  Matrix  Multiply  Routines  Using   OpenMP”,  Shared  Memory  Parallel  Programming  with  OpenMP,  Lecture  notes  in   Computer Science, Vol. 3349, P. 1, 2005  Bova SW, Breshearsz CP, Cuicchi CE, Demirbilek Z, Gabb HA. Dual-level parallel analysis of harbor wave response using MPI and OpenMP. International Journal of High Performance Computing Applications, vol.14, no.1, Spring 2000, pp.49-64. Publisher: Sage Science Press, USA.  Ayguade E, Martorell X, Labarta J, Gonzalez M, Navarro N. Exploiting multiple levels of parallelism in OpenMP: a case study. Proceedings of the 1999 International Conference on Parallel Processing. IEEE Comput. Soc. 1999, pp.172-80. Los Alamitos, CA, USA.  Bova SW, Breshears CP, Cuicchi C, Demirbilek Z, Gabb H. Nesting OpenMP in an MPI application. Proceedings of the ISCA 12th International Conference. Parallel and Distributed Systems. ISCA. 1999, pp.566-71. Cary, NC, USA.
  191. 192 OpenMP Papers (continued)  Jost G., Labarta J., Gimenez

    J., What Multilevel Parallel Programs do when you are not watching: a Performance analysis case study comparing MPI/OpenMP, MLP, and Nested OpenMP, Shared Memory Parallel Programming with OpenMP, Lecture notes in Computer Science, Vol. 3349, P. 29, 2005  Gonzalez M, Serra A, Martorell X, Oliver J, Ayguade E, Labarta J, Navarro N. Applying interposition techniques for performance analysis of OPENMP parallel applications. Proceedings 14th International Parallel and Distributed Processing Symposium. IPDPS 2000. IEEE Comput. Soc. 2000, pp.235-40.  Chapman B, Mehrotra P, Zima H. Enhancing OpenMP with features for locality control. Proceedings of Eighth ECMWF Workshop on the Use of Parallel Processors in Meteorology. Towards Teracomputing. World Scientific Publishing. 1999, pp.301-13. Singapore.  Steve W. Bova, Clay P. Breshears, Henry Gabb, Rudolf Eigenmann, Greg Gaertner, Bob Kuhn, Bill Magro, Stefano Salvini. Parallel Programming with Message Passing and Directives; SIAM News, Volume 32, No 9, Nov. 1999.  Cappello F, Richard O, Etiemble D. Performance of the NAS benchmarks on a cluster of SMP PCs using a parallelization of the MPI programs with OpenMP. Lecture Notes in Computer Science Vol.1662. Springer-Verlag. 1999, pp.339-50.  Liu Z., Huang L., Chapman B., Weng T., Efficient Implementationi of OpenMP for Clusters with Implicit Data Distribution, Shared Memory Parallel Programming with OpenMP, Lecture notes in Computer Science, Vol. 3349, P. 121, 2005
  192. 193 OpenMP Papers (continued)  B.  Chapman,  F.  Bregier,  A.

     Patil,  A.  Prabhakar,  “Achieving   performance under OpenMP on ccNUMA and software distributed shared  memory  systems,”  Concurrency and Computation: Practice and Experience. 14(8-9): 713-739, 2002.  J. M. Bull and M. E. Kambites. JOMP: an OpenMP-like interface for Java. Proceedings of the ACM 2000 conference on Java Grande, 2000, Pages 44 - 53.  L.  Adhianto  and  B.  Chapman,  “Performance  modeling  of   communication and computation in hybrid MPI and OpenMP applications, Simulation Modeling Practice and Theory, vol 15, p. 481- 491, 2007.  Shah S, Haab G, Petersen P, Throop J. Flexible control structures for parallelism in OpenMP; Concurrency: Practice and Experience, 2000; 12:1219-1239. Publisher John Wiley & Sons, Ltd.  Mattson, T.G., How Good is OpenMP? Scientific Programming, Vol. 11, Number 2, p.81-93, 2003.  Duran  A.,  Silvera  R.,  Corbalan  J.,  Labarta  J.,  “Runtime  Adjustment  of   Parallel  Nested  Loops”,    Shared  Memory  Parallel  Programming  with   OpenMP, Lecture notes in Computer Science, Vol. 3349, P. 137, 2005
  193. 194 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  194. 195 OpenMP pre-history  OpenMP based upon SMP directive standardization

    efforts PCF and aborted ANSI X3H5 – late  80’s Nobody fully implemented either standard Only a couple of partial implementations  Vendors  considered  proprietary  API’s  to  be  a   competitive feature: Every vendor had proprietary directives sets Even  KAP,  a  “portable”  multi-platform parallelization tool used different directives on each platform PCF – Parallel computing forum KAP – parallelization tool from KAI.
  195. 196 History of OpenMP SGI Cray Merged, needed commonality across

    products KAI ISV - needed larger market was tired of recoding for SMPs. Urged vendors to standardize. ASCI Wrote a rough draft straw man SMP API DEC IBM Intel HP Other vendors invited to join 1997
  196. 197 OpenMP Release History OpenMP Fortran 1.1 OpenMP C/C++ 1.0

    OpenMP Fortran 2.0 OpenMP C/C++ 2.0 1998 2000 1999 2002 OpenMP Fortran 1.0 1997 OpenMP 2.5 2005 A single specification for Fortran, C and C++ OpenMP 3.0 Tasking, other new features 2008 OpenMP 3.1 2011 A few more features and bug fixes
  197. 198 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  198. 199 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  199. 200 Exercise 1: Solution A multi-threaded  “Hello  world”  program 

    Write a multithreaded program where each thread  prints  “hello  world”. #include  “omp.h” void main() { #pragma omp parallel { int ID = omp_get_thread_num(); printf(“  hello(%d)  ”,  ID);; printf(“  world(%d)  \n”,  ID);; } } Sample Output: hello(1) hello(0) world(1) world(0) hello (3) hello(2) world(3) world(2) OpenMP include file Parallel region with default number of threads Runtime library function to return a thread ID. End of the Parallel region
  200. 201 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  201. 202 The SPMD pattern  The most common approach for

    parallel algorithms is the SPMD or Single Program Multiple Data pattern.  Each thread runs the same program (Single Program), but using the thread ID, they operate on different data (Multiple Data) or take slightly different paths through the code.  In OpenMP this means: A  parallel  region  “near  the  top  of  the  code”. Pick up thread ID and num_threads. Use them to split up loops and select different blocks of data to work on.
  202. 203 #include <omp.h> static long num_steps = 100000; double step;

    #define NUM_THREADS 2 void main () { int i, nthreads; double pi, sum[NUM_THREADS]; step = 1.0/(double) num_steps; omp_set_num_threads(NUM_THREADS); #pragma omp parallel { int i, id,nthrds; double x; id = omp_get_thread_num(); nthrds = omp_get_num_threads(); if (id == 0) nthreads = nthrds; for (i=id, sum[id]=0.0;i< num_steps; i=i+nthrds) { x = (i+0.5)*step; sum[id] += 4.0/(1.0+x*x); } } for(i=0, pi=0.0;i<nthreads;i++)pi += sum[i] * step; } Exercise 2: A simple SPMD pi program Promote scalar to an array dimensioned by number of threads to avoid race condition. This is a common trick in SPMD programs to create a cyclic distribution of loop iterations Only one thread should copy the number of threads to the global value to make sure multiple threads writing to the same address don’t conflict.
  203. 204 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  204. 205 False sharing  If independent data elements happen to

    sit on the same cache line, each update will cause the cache lines to “slosh  back  and  forth”  between  threads.  This  is  called  “false  sharing”.  If you promote scalars to an array to support creation of an SPMD program, the array elements are contiguous in memory and hence share cache lines.  Result  …  poor  scalability  Solution:  When updates to an item are frequent, work with local copies of data instead of an array indexed by the thread ID.  Pad arrays so elements you use are on distinct cache lines.
  205. 206 #include <omp.h> static long num_steps = 100000; double step;

    #define NUM_THREADS 2 void main () { double pi = 0.0; step = 1.0/(double) num_steps; int nthreads; omp_set_num_threads(NUM_THREADS); #pragma omp parallel { int i, id,nthrds; double x, sum; id = omp_get_thread_num(); nthrds = omp_get_num_threads(); if (id == 0) nthreads = nthrds; for (i=id, sum=0.0;i< num_steps; i=i+nthrds){ x = (i+0.5)*step; sum += 4.0/(1.0+x*x); } #pragma omp critical pi += sum * step; } } Exercise 3: SPMD Pi without false sharing Sum goes “out of scope” beyond the parallel region … so you must sum it in here. Must protect summation into pi in a critical region so updates don’t conflict No array, so no false sharing. Create a scalar local to each thread to accumulate partial sums.
  206. 207 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  207. 208 Exercise 4: solution #include <omp.h> static long num_steps =

    100000; double step; void main () { int i; double x, pi, sum = 0.0; step = 1.0/(double) num_steps; #pragma omp parallel { double x; #pragma omp for reduction(+:sum) for (i=0;i< num_steps; i++){ x = (i+0.5)*step; sum = sum + 4.0/(1.0+x*x); } } pi = step * sum; }
  208. 209 Exercise 4: solution #include <omp.h> static long num_steps =

    100000; double step; void main () { int i; double x, pi, sum = 0.0; step = 1.0/(double) num_steps; #pragma omp parallel for private(x) reduction(+:sum) for (i=0;i< num_steps; i++){ x = (i+0.5)*step; sum = sum + 4.0/(1.0+x*x); } pi = step * sum; } Note: we created a parallel program without changing any code and by adding 2 simple lines of text! i private by default For good OpenMP implementations, reduction is more scalable than critical.
  209. 210 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  210. 211 Matrix multiplication #pragma omp parallel for private(tmp, i, j,

    k) for (i=0; i<Ndim; i++){ for (j=0; j<Mdim; j++){ tmp = 0.0; for(k=0;k<Pdim;k++){ /* C(i,j) = sum(over k) A(i,k) * B(k,j) */ tmp += *(A+(i*Ndim+k)) * *(B+(k*Pdim+j)); } *(C+(i*Ndim+j)) = tmp; } } •On a dual core laptop •13.2 seconds 153 Mflops one thread •7.5 seconds 270 Mflops two threads Results on an Intel dual core 1.83 GHz CPU, Intel IA-32 compiler 10.1 build 2
  211. 212 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  212. Exercise 6: Area of a Mandelbrot set  Solution is

    in the file mandel_par.c  Errors: Eps is private but uninitialized. Two solutions – It’s  read-only so you can make it shared. – Make it firstprivate The loop index variable j is shared by default. Make it private. The  variable  c  has  global  scope  so  “testpoint”  may   pick up the global value rather than the private value in  the  loop.    Solution  …  pass  C  as  an  arg to testpoint Updates  to  “numoutside”  are  a  race.    Protect  with  an   atomic. 213
  213. 214 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  214. #define THRESHOLD 32768 // product size below which simple matmult

    code is called void matmultrec(int mf, int ml, int nf, int nl, int pf, int pl, double **A, double **B, double **C) // Dimensions: A[mf..ml][pf..pl] B[pf..pl][nf..nl] C[mf..ml][nf..nl] { if ((ml-mf)*(nl-nf)*(pl-pf) < THRESHOLD) matmult (mf, ml, nf, nl, pf, pl, A, B, C); else { #pragma omp task { matmultrec(mf, mf+(ml-mf)/2, nf, nf+(nl-nf)/2, pf, pf+(pl-pf)/2, A, B, C); // C11 += A11*B11 matmultrec(mf, mf+(ml-mf)/2, nf, nf+(nl-nf)/2, pf+(pl-pf)/2, pl, A, B, C); // C11 += A12*B21 } #pragma omp task { matmultrec(mf, mf+(ml-mf)/2, nf+(nl-nf)/2, nl, pf, pf+(pl-pf)/2, A, B, C); // C12 += A11*B12 matmultrec(mf, mf+(ml-mf)/2, nf+(nl-nf)/2, nl, pf+(pl-pf)/2, pl, A, B, C); // C12 += A12*B22 } #pragma omp task { matmultrec(mf+(ml-mf)/2, ml, nf, nf+(nl-nf)/2, pf, pf+(pl-pf)/2, A, B, C); // C21 += A21*B11 matmultrec(mf+(ml-mf)/2, ml, nf, nf+(nl-nf)/2, pf+(pl-pf)/2, pl, A, B, C); // C21 += A22*B21 } #pragma omp task { matmultrec(mf+(ml-mf)/2, ml, nf+(nl-nf)/2, nl, pf, pf+(pl-pf)/2, A, B, C); // C22 += A21*B12 matmultrec(mf+(ml-mf)/2, ml, nf+(nl-nf)/2, nl, pf+(pl-pf)/2, pl, A, B, C); // C22 += A22*B22 } #pragma omp taskwait } } Recursive Solution 215  Could be executed in parallel as 4 tasks  Each task executes the two calls for the same output submatrix of C  However, the same number of multiplication operations needed
  215. Solution:  Strassen’s  Algorithm 21 #pragma omp task {// M1 =

    (A11 + A22)*(B11 + B22) AddMatBlocks(wAM1, m2, p2, A11, A22); AddMatBlocks(wBM1, p2, n2, B11, B22); strassenMMult(m2, n2, p2, wAM1, wBM1, M1); } #pragma omp task {// M2 = (A21 + A22)*B11 AddMatBlocks(wAM2, m2, p2, A21, A22); strassenMMult(m2, n2, p2, wAM2, B11, M2); } #pragma omp task {// M3 = A11*(B12 - B22) SubMatBlocks(wBM3, p2, n2, B12, B22); strassenMMult(m2, n2, p2, A11, wBM3, M3); } #pragma omp task {// M4 = A22*(B21 - B11) SubMatBlocks(wBM4, p2, n2, B21, B11); strassenMMult(m2, n2, p2, A22, wBM4, M4); } #pragma omp task {// M5 = (A11 + A12)*B22 AddMatBlocks(wAM5, m2, p2, A11, A12); strassenMMult(m2, n2, p2, wAM5, B22, M5); } #pragma omp task {// M6 = (A21 - A11)*(B11 + B12) SubMatBlocks(wAM6, m2, p2, A21, A11); AddMatBlocks(wBM6, p2, n2, B11, B12); strassenMMult(m2, n2, p2, wAM6, wBM6, M6); } #pragma omp task {// M7 = (A12 - A22)*(B21 + B22) SubMatBlocks(wAM7, m2, p2, A12, A22); AddMatBlocks(wBM7, p2, n2, B21, B22); strassenMMult(m2, n2, p2, wAM7, wBM7, M7); } #pragma omp taskwait for (int i = 0; i < m2; i++) for (int j = 0; j < n2; j++) { C11[i][j] = M1[i][j] + M4[i][j] – M5[i][j] + M7[i][j]; C12[i][j] = M3[i][j] + M5[i][j]; C21[i][j] = M2[i][j] + M4[i][j]; C22[i][j] = M1[i][j] - M2[i][j] + M3[i][j] + M6[i][j]; }
  216. Potential Optimizations  Cut down the number of temporary matrices

    needed  Use M1-M4 to call first four tasks  When done, compute partial C results  Re-use M1-M3 for computing M5-M7  Complete C submatrix computations (C11, C12, C22)  Use C11, C12, C21 to receive M1, M3, M2, respectively  Compute M4-M7 as before  When done, C22 = C11 – C21 + C12 +M6  Then, update C11, C12, C21 with M4, M5, M7 as needed  Allocate work matrices (wAM1, wBM1, etc.) within task before usage 217 #pragma omp task {// M5 = (A11 + A12)*B22 double **wAM5 = Allocate2DArray< double >(m2, p2); AddMatBlocks(tAM5, m2, p2, A11, A12); strassenMMult(m2, n2, p2, wAM5, B22, M5); Free2DArray< double >(wAM5); }
  217. 218 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  218. 219 Computers and random numbers  We  use  “dice”  to

     make  random  numbers:    Given previous values, you cannot predict the next value.  There  are  no  patterns  in  the  series  …  and  it  goes  on  forever.  Computers  are  deterministic  machines  …  set  an  initial   state, run a sequence of predefined instructions, and you get a deterministic answer  By design, computers are not random and cannot produce random numbers.  However, with some very clever programming, we can make  “pseudo  random”  numbers  that  are  as  random  as   you  need  them  to  be  …  but  only  if  you  are  very  careful.  Why do I care? Random numbers drive statistical methods used in countless applications:  Sample a large space of alternatives to find statistically good answers (Monte Carlo methods).
  219. 220 Monte Carlo Calculations: Using Random numbers to solve tough

    problems  Sample a problem domain to estimate areas, compute probabilities, find optimal values, etc.  Example: Computing π with a digital dart board:  Throw darts at the circle/square.  Chance of falling in circle is proportional to ratio of areas: Ac = r2 * π As = (2*r) * (2*r) = 4 * r2 P = Ac /As = π /4  Compute π by randomly choosing points, count the fraction that falls in the circle, compute pi. 2 * r N= 10 π = 2.8 N=100 π = 3.16 N= 1000 π = 3.148
  220. 221 Parallel Programmers love Monte Carlo algorithms #include  “omp.h” static

    long num_trials = 10000; int main () { long i; long Ncirc = 0; double pi, x, y; double r = 1.0; // radius of circle. Side of squrare is 2*r seed(0,-r, r); // The circle and square are centered at the origin #pragma omp parallel for private (x, y) reduction (+:Ncirc) for(i=0;i<num_trials; i++) { x = random(); y = random(); if ( x*x + y*y) <= r*r) Ncirc++; } pi = 4.0 * ((double)Ncirc/(double)num_trials); printf("\n %d trials, pi is %f \n",num_trials, pi); } Embarrassingly parallel: the parallelism is so easy its embarrassing. Add two lines and you have a parallel program.
  221. 222 Linear Congruential Generator (LCG)  LCG: Easy to write,

    cheap to compute, portable, OK quality  If you pick the multiplier and addend correctly, LCG has a period of PMOD.  Picking good LCG parameters is complicated, so look it up (Numerical Recipes is a good source). I used the following:  MULTIPLIER = 1366  ADDEND = 150889  PMOD = 714025 random_next = (MULTIPLIER * random_last + ADDEND)% PMOD; random_last = random_next;
  222. 223 LCG code static long MULTIPLIER = 1366; static long

    ADDEND = 150889; static long PMOD = 714025; long random_last = 0; double random () { long random_next; random_next = (MULTIPLIER * random_last + ADDEND)% PMOD; random_last = random_next; return ((double)random_next/(double)PMOD); } Seed the pseudo random sequence by setting random_last
  223. 224 Running the PI_MC program with LCG generator 0.00001 0.0001

    0.001 0.01 0.1 1 1 2 3 4 5 6 LCG - one thread LCG, 4 threads, trail 1 LCG 4 threads, trial 2 LCG, 4 threads, trial 3 Log 10 Relative error Log10 number of samples Run the same program the same way and get different answers! That is not acceptable! Issue: my LCG generator is not threadsafe Program written using the Intel C/C++ compiler (10.0.659.2005) in Microsoft Visual studio 2005 (8.0.50727.42) and running on a dual-core laptop (Intel T2400 @ 1.83 Ghz with 2 GB RAM) running Microsoft Windows XP.
  224. 225 LCG code: threadsafe version static long MULTIPLIER = 1366;

    static long ADDEND = 150889; static long PMOD = 714025; long random_last = 0; #pragma omp threadprivate(random_last) double random () { long random_next; random_next = (MULTIPLIER * random_last + ADDEND)% PMOD; random_last = random_next; return ((double)random_next/(double)PMOD); } random_last carries state between random number computations, To make the generator threadsafe, make random_last threadprivate so each thread has its own copy.
  225. 226 Thread safe random number generators Log10 Relative error Log10

    number of samples Thread safe version gives the same answer each time you run the program. But for large number of samples, its quality is lower than the one thread result! Why? 0.00001 0.0001 0.001 0.01 0.1 1 1 2 3 4 5 6 LCG - one thread LCG 4 threads, trial 1 LCT 4 threads, trial 2 LCG 4 threads, trial 3 LCG 4 threads, thread safe
  226. 227 Pseudo Random Sequences  Random number Generators (RNGs) define

    a sequence of pseudo-random numbers of length equal to the period of the RNG  In a typical problem, you grab a subsequence of the RNG range Seed determines starting point  Grab arbitrary seeds and you may generate overlapping sequences  E.g.  three  sequences  …  last  one  wraps  at  the  end  of  the  RNG  period.  Overlapping sequences = over-sampling  and  bad  statistics  …  lower   quality or even wrong answers! Thread 1 Thread 2 Thread 3
  227. 228 Parallel random number generators  Multiple threads cooperate to

    generate and use random numbers.  Solutions:  Replicate and Pray  Give each thread a separate, independent generator  Have one thread generate all the numbers.  Leapfrog  …  deal  out  sequence  values  “round   robin”  as  if  dealing  a  deck  of  cards.  Block  method  …  pick  your  seed  so  each   threads gets a distinct contiguous block.  Other  than  “replicate  and  pray”,  these  are  difficult   to  implement.    Be  smart  …  buy  a  math  library  that   does it right. If done right, can generate the same sequence regardless of the number of threads  … Nice for debugging, but not really needed scientifically. Intel’s  Math  kernel  Library  supports   all of these methods.
  228. 229 MKL Random number generators (RNG) #define BLOCK 100 double

    buff[BLOCK]; VSLStreamStatePtr stream; vslNewStream(&ran_stream, VSL_BRNG_WH, (int)seed_val); vdRngUniform (VSL_METHOD_DUNIFORM_STD, stream, BLOCK, buff, low, hi) vslDeleteStream( &stream );  MKL includes several families of RNGs in its vector statistics library.  Specialized to efficiently generate vectors of random numbers Initialize a stream or pseudo random numbers Select type of RNG and set seed Fill buff with BLOCK pseudo rand. nums, uniformly distributed with values between lo and hi. Delete the stream when you are done
  229. 230 Wichmann-Hill generators (WH)  WH is a family of

    273 parameter sets each defining a non- overlapping and independent RNG.  Easy to use, just make each stream threadprivate and initiate RNG stream so each thread gets a unique WG RNG. VSLStreamStatePtr stream; #pragma omp threadprivate(stream) … vslNewStream(&ran_stream, VSL_BRNG_WH+Thrd_ID, (int)seed);
  230. 231 Independent Generator for each thread 0.0001 0.001 0.01 0.1

    1 1 2 3 4 5 6 WH one thread WH, 2 threads WH, 4 threads Log10 Relative error Log10 number of samples Notice that once you get beyond the high error, small sample count range, adding threads doesn’t   decrease quality of random sampling.
  231. 232 #pragma omp single { nthreads = omp_get_num_threads(); iseed =

    PMOD/MULTIPLIER; // just pick a seed pseed[0] = iseed; mult_n = MULTIPLIER; for (i = 1; i < nthreads; ++i) { iseed = (unsigned long long)((MULTIPLIER * iseed) % PMOD); pseed[i] = iseed; mult_n = (mult_n * MULTIPLIER) % PMOD; } } random_last = (unsigned long long) pseed[id]; Leap Frog method  Interleave samples in the sequence of pseudo random numbers:  Thread i starts at the ith number in the sequence  Stride through sequence, stride length = number of threads.  Result  …  the  same  sequence  of  values  regardless  of  the  number   of threads. One thread computes offsets and strided multiplier LCG with Addend = 0 just to keep things simple Each thread stores offset starting point into its threadprivate “last random” value
  232. 233 Same sequence with many threads.  We can use

    the leapfrog method to generate the same answer for any number of threads Steps One thread 2 threads 4 threads 1000 3.156 3.156 3.156 10000 3.1168 3.1168 3.1168 100000 3.13964 3.13964 3.13964 1000000 3.140348 3.140348 3.140348 10000000 3.141658 3.141658 3.141658 Used the MKL library with two generator streams per computation: one for the x values (WH) and one for the y values (WH+1). Also used the leapfrog method to deal out iterations among threads.
  233. 234 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  234. Exercise 9 solution #pragma omp parallel for default (none) \

    shared(x,f,npart,rcoff,side) \ reduction(+:epot,vir) \ schedule (static,32) for (int i=0; i<npart*3; i+=3) { ……… Loop is not well load balanced: best schedule has to be found by experiment. Compiler will warn you if you have missed some variables
  235. ........ #pragma omp atomic f[j] -= forcex; #pragma omp atomic

    f[j+1] -= forcey; #pragma omp atomic f[j+2] -= forcez; } } #pragma omp atomic f[i] += fxi; #pragma omp atomic f[i+1] += fyi; #pragma omp atomic f[i+2] += fzi; } } All updates to f must be atomic Exercise 9 solution (cont.)
  236. Exercise 9 with orphaning #pragma omp single { vir =

    0.0; epot = 0.0; } #pragma omp for reduction(+:epot,vir) \ schedule (static,32) for (int i=0; i<npart*3; i+=3) { ……… All variables which used to be shared here are now implicitly determined Implicit barrier needed to avoid race condition with update of reduction variables at end of the for construct
  237. Exercise 9 with array reduction ftemp[myid][j] -= forcex; ftemp[myid][j+1] -=

    forcey; ftemp[myid][j+2] -= forcez; } } ftemp[myid][i] += fxi; ftemp[myid][i+1] += fyi; ftemp[myid][i+2] += fzi; } Replace atomics with accumulation into array with extra dimension
  238. Exercise 9 with array reduction …. #pragma omp for for(int

    i=0;i<(npart*3);i++){ for(int id=0;id<nthreads;id++){ f[i] += ftemp[id][i]; ftemp[id][i] = 0.0; } } Reduction can be done in parallel Zero ftemp for next time round
  239. 240 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  240. 241 Exercise 10: tasks in OpenMP  Consider the program

    linked.c Traverses a linked list computing a sequence of Fibonacci numbers at each node.  Parallelize this program using tasks.  Compare  your  solution’s  complexity  to  an   approach without tasks.
  241. 242 Linked lists with tasks (intel taskq)  See the

    file Linked_intel_taskq.c #pragma omp parallel { #pragma intel omp taskq { while (p != NULL) { #pragma intel omp task captureprivate(p) processwork(p); p = p->next; } } } Array, Static, 1 Intel taskq One Thread 45 seconds 48 seconds Two Threads 28 seconds 30 seconds Results on an Intel dual core 1.83 GHz CPU, Intel IA-32 compiler 10.1 build 2
  242. 243 Linked lists with tasks (OpenMP 3)  See the

    file Linked_omp3_tasks.c #pragma omp parallel { #pragma omp single { p=head; while (p) { #pragma omp task firstprivate(p) processwork(p); p = p->next; } } } Creates a task with its own copy of “p” initialized to the value of “p” when the task is defined
  243. 244 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  244. 245 Exercise 11: linked lists the hard way  Consider

    the program linked.c Traverses a linked list computing a sequence of Fibonacci numbers at each node.  Parallelize this program using constructs defined  in  worksharing  constructs  …  i.e.  don’t   use tasks).  Once you have a correct program, optimize it.
  245. 246 Linked lists without tasks  See the file Linked_omp25.c

    while (p != NULL) { p = p->next; count++; } p = head; for(i=0; i<count; i++) { parr[i] = p; p = p->next; } #pragma omp parallel { #pragma omp for schedule(static,1) for(i=0; i<count; i++) processwork(parr[i]); } Count number of items in the linked list Copy pointer to each node into an array Process nodes in parallel with a for loop Default schedule Static,1 One Thread 48 seconds 45 seconds Two Threads 39 seconds 28 seconds Results on an Intel dual core 1.83 GHz CPU, Intel IA-32 compiler 10.1 build 2
  246. 247 Linked lists without tasks: C++ STL  See the

    file Linked_cpp.cpp std::vector<node *> nodelist; for (p = head; p != NULL; p = p->next) nodelist.push_back(p); int j = (int)nodelist.size(); #pragma omp parallel for schedule(static,1) for (int i = 0; i < j; ++i) processwork(nodelist[i]); C++, default sched. C++, (static,1) C, (static,1) One Thread 37 seconds 49 seconds 45 seconds Two Threads 47 seconds 32 seconds 28 seconds Copy pointer to each node into an array Count number of items in the linked list Process nodes in parallel with a for loop Results on an Intel dual core 1.83 GHz CPU, Intel IA-32 compiler 10.1 build 2
  247. 248 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  248. 249 OpenMP memory model proc1 proc2 proc3 procN Shared memory

    cache1 cache2 cache3 cacheN a a . . .  A memory model is defined in terms of:  Coherence: Behavior of the memory system when a single address is accessed by multiple threads.  Consistency: Orderings of reads, writes, or synchronizations (RWS) with various addresses and by multiple threads.  OpenMP supports a shared memory model.  All threads share an address space, but it can get complicated:
  249. 250 Source code Program order memory a b Commit order

    private view thread thread private view threadprivate threadprivate a a b b Wa Wb Ra Rb . . . OpenMP Memory Model: Basic Terms compiler Executable code Code order Wb Rb Wa Ra . . . RW’s in any semantically equivalent order
  250. 251 Consistency: Memory Access Re-ordering  Re-ordering:  Compiler re-orders

    program order to the code order  Machine re-orders code order to the memory commit order  At  a  given  point  in  time,  the  “private  view”  seen  by  a   thread may be different from the view in shared memory.  Consistency Models define constraints on the orders of Reads (R), Writes (W) and Synchronizations (S)  …  i.e.  how  do  the  values  “seen”  by  a  thread  change  as  you   change how ops follow (→) other ops.  Possibilities include: – R→R,    W→W,    R→W,      R→S,    S→S,    W→S
  251. 252 Consistency  Sequential Consistency: In a multi-processor, ops (R,

    W, S) are sequentially consistent if: – They remain in program order for each processor. – They are seen to be in the same overall order by each of the other processors. Program order = code order = commit order  Relaxed consistency: Remove some of the ordering constraints for memory ops (R, W, S).
  252. 253 OpenMP and Relaxed Consistency  OpenMP defines consistency as

    a variant of weak consistency: S ops must be in sequential order across threads. Can not reorder S ops with R or W ops on the same thread – Weak consistency guarantees S→W,      S→R  ,  R→S,  W→S,  S→S  The Synchronization operation relevant to this discussion is flush.
  253. 254 Flush  Defines a sequence point at which a

    thread is guaranteed to see a consistent view of memory with respect  to  the  “flush  set”.  The flush set is:  “all  thread  visible  variables”  for  a  flush  construct  without  an   argument list.  a  list  of  variables  when  the  “flush(list)”  construct  is  used.  The action of Flush is to guarantee that: – All R,W ops that overlap the flush set and occur prior to the flush complete before the flush executes – All R,W ops that overlap the flush set and occur after the flush  don’t  execute  until  after  the  flush. – Flushes with overlapping flush sets can not be reordered. Memory ops: R = Read, W = write, S = synchronization
  254. 255 Synchronization: flush example  Flush forces data to be

    updated in memory so other threads see the most recent value double A; A = compute(); flush(A); // flush to memory to make sure other // threads can pick up the right value Note:  OpenMP’s  flush  is  analogous  to  a  fence  in   other  shared  memory  API’s.
  255. 256 What is the Big Deal with Flush?  Compilers

    routinely reorder instructions implementing a program  This helps better exploit the functional units, keep machine busy, hide memory latencies, etc.  Compiler generally cannot move instructions:  past a barrier  past a flush on all variables  But it can move them past a flush with a list of variables so long as those variables are not accessed  Keeping track of consistency when flushes are used can  be  confusing  …  especially  if  “flush(list)”  is  used. Note: the flush operation does not actually synchronize different threads. It just ensures that a thread’s values are made consistent with main memory.
  256. 257 Exercise 12: producer consumer  Parallelize  the  “prod_cons.c”  program.

     This is a well known pattern called the producer consumer pattern One thread produces values that another thread consumes. Often used with a stream of produced values to implement  “pipeline  parallelism”  The key is to implement pairwise synchronization between threads.
  257. 258 Exercise 12: prod_cons.c int main() { double *A, sum,

    runtime; int flag = 0; A = (double *)malloc(N*sizeof(double)); runtime = omp_get_wtime(); fill_rand(N, A); // Producer: fill an array of data sum = Sum_array(N, A); // Consumer: sum the array runtime = omp_get_wtime() - runtime; printf(" In %lf seconds, The sum is %lf \n",runtime,sum); }
  258. 259 Pair wise synchronizaion in OpenMP  OpenMP lacks synchronization

    constructs that work between pairs of threads.  When this is needed you have to build it yourself.  Pair wise synchronization Use a shared flag variable Reader spins waiting for the new flag value Use flushes to force updates to and from memory
  259. 260 Exercise 10: producer consumer int main() { double *A,

    sum, runtime; int numthreads, flag = 0; A = (double *)malloc(N*sizeof(double)); #pragma omp parallel sections { #pragma omp section { fill_rand(N, A); #pragma omp flush flag = 1; #pragma omp flush (flag) } #pragma omp section { #pragma omp flush (flag) while (flag != 0){ #pragma omp flush (flag) } #pragma omp flush sum = Sum_array(N, A); } } } Use flag to Signal when the “produced” value is ready Flush forces refresh to memory. Guarantees that the other thread sees the new value of A Notice you must put the flush inside the while loop to make sure the updated flag variable is seen Flush needed on both “reader” and “writer” sides of the communication
  260. 261 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  261. Fortran and OpenMP  We were careful to design the

    OpenMP constructs so they cleanly map onto C, C++ and Fortran.  There are a few syntactic differences that once understood, will allow you to move back and forth between languages.  In the specification, language specific notes are included when each construct is defined. 262
  262. OpenMP: Some syntax details for Fortran programmers  Most of

    the constructs in OpenMP are compiler directives. For Fortran, the directives take one of the forms: C$OMP construct  [clause  [clause]…] !$OMP construct  [clause  [clause]…] *$OMP construct  [clause  [clause]…]  The OpenMP include file and lib module use omp_lib Include omp_lib.h
  263. OpenMP: Structured blocks (Fortran) C$OMP PARALLEL 10 wrk(id) = garbage(id)

    res(id) = wrk(id)**2 if(conv(res(id)) goto 10 C$OMP END PARALLEL print *,id Most OpenMP constructs apply to structured blocks. – Structured block: a block of code with one point of entry at the top and one point of exit at the bottom. – The  only  “branches”  allowed  are  STOP   statements in Fortran and exit() in C/C++. C$OMP PARALLEL 10 wrk(id) = garbage(id) 30 res(id)=wrk(id)**2 if(conv(res(id))goto 20 go to 10 C$OMP END PARALLEL if(not_DONE) goto 30 20 print *, id A structured block Not A structured block
  264. OpenMP: Structured Block Boundaries  In Fortran: a block is

    a single statement or a group of statements between directive/end-directive pairs. C$OMP PARALLEL 10 wrk(id) = garbage(id) res(id) = wrk(id)**2 if(conv(res(id)) goto 10 C$OMP END PARALLEL C$OMP PARALLEL DO do I=1,N res(I)=bigComp(I) end do C$OMP END PARALLEL DO  The    “construct/end  construct”  pairs  is  done  anywhere  a   structured block appears in Fortran. Some examples:  DO      …    END    DO  PARALLEL    …    END  PARREL  CRICITAL    …  END  CRITICAL  SECTION    …  END  SECTION  SECTIONS    …  END  SECTIONS  SINGLE    …    END  SINGLE  MASTER  …  END  MASTER
  265. Runtime library routines  The include file or module defines

    parameters  Integer parameter omp_locl_kind  Integer parameter omp_nest_lock_kind  Integer parameter omp_sched_kind  Integer parameter openmp_version – With  value  that  matches  C’s  _OPEMMP  macro  Fortran interfaces are similar to those used with C  Subroutine omp_set_num_threads (num_threads)  Integer function omp_get_num_threads()  Integer function omp_get_thread_num()\  Subroutine omp_init_lock(svar) – Integer(kind=omp_lock_kind) svar  Subroutine omp_destroy_lock(svar)  Subroutine omp_set_lock(svar)  Subroutine omp_unset_lock(svar) 266
  266. 267 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  267. 268 How do people mix MPI and OpenMP? Replicate the

    program. Add glue code Break up the data A sequential program working on a data set •Create the MPI program with its data decomposition. • Use OpenMP inside each MPI process.
  268. 269 Pi program with MPI and OpenMP #include <mpi.h> #include

     “omp.h” void main (int argc, char *argv[]) { int i, my_id, numprocs; double x, pi, step, sum = 0.0 ; step = 1.0/(double) num_steps ; MPI_Init(&argc, &argv) ; MPI_Comm_Rank(MPI_COMM_WORLD, &my_id) ; MPI_Comm_Size(MPI_COMM_WORLD, &numprocs) ; my_steps = num_steps/numprocs ; #pragma omp parallel for reduction(+:sum) private(x) for (i=my_id*my_steps; i<(m_id+1)*my_steps ; i++) { x = (i+0.5)*step; sum += 4.0/(1.0+x*x); } sum *= step ; MPI_Reduce(&sum, &pi, 1, MPI_DOUBLE, MPI_SUM, 0, MPI_COMM_WORLD) ; } Get the MPI part done first, then add OpenMP pragma where it makes sense to do so
  269. 270 Key issues when mixing OpenMP and MPI 1. Messages

    are sent to a process not to a particular thread.  Not all MPIs are threadsafe. MPI 2.0 defines threading modes: – MPI_Thread_Single: no support for multiple threads – MPI_Thread_Funneled: Mult threads, only master calls MPI – MPI_Thread_Serialized: Mult threads each calling MPI, but they do it one at a time. – MPI_Thread_Multiple: Multiple threads without any restrictions  Request and test thread modes with the function: MPI_init_thread(desired_mode, delivered_mode, ierr) 2. Environment variables are not propagated by mpirun. You’ll  need  to  broadcast  OpenMP  parameters  and  set   them with the library routines.
  270. 271 Dangerous Mixing of MPI and OpenMP  The following

    will work only if MPI_Thread_Multiple is supported …  a  level  of  support  I  wouldn’t  depend  on. MPI_Comm_Rank(MPI_COMM_WORLD, &mpi_id) ; #pragma omp parallel { int tag, swap_neigh, stat, omp_id = omp_thread_num(); long buffer [BUFF_SIZE], incoming [BUFF_SIZE]; big_ugly_calc1(omp_id, mpi_id, buffer); // Finds MPI id and tag so neighbor(omp_id,  mpi_id,  &swap_neigh,  &tag);;    //  messages  don’t  conflict MPI_Send (buffer, BUFF_SIZE, MPI_LONG, swap_neigh, tag, MPI_COMM_WORLD); MPI_Recv (incoming, buffer_count, MPI_LONG, swap_neigh, tag, MPI_COMM_WORLD, &stat); big_ugly_calc2(omp_id, mpi_id, incoming, buffer); #pragma critical consume(buffer, omp_id, mpi_id); }
  271. 272 Messages and threads  Keep message passing and threaded

    sections of your program separate: Setup message passing outside OpenMP parallel regions (MPI_Thread_funneled) Surround with appropriate directives (e.g. critical section or master) (MPI_Thread_Serialized) For certain applications depending on how it is designed it may not matter which thread handles a message. (MPI_Thread_Multiple) – Beware of race conditions though if two threads are probing on the same message and then racing to receive it.
  272. 273 Safe Mixing of MPI and OpenMP Put MPI in

    sequential regions MPI_Init(&argc, &argv) ; MPI_Comm_Rank(MPI_COMM_WORLD, &mpi_id) ; // a whole bunch of initializations #pragma omp parallel for for (I=0;I<N;I++) { U[I] = big_calc(I); } MPI_Send (U, BUFF_SIZE, MPI_DOUBLE, swap_neigh, tag, MPI_COMM_WORLD); MPI_Recv (incoming, buffer_count, MPI_DOUBLE, swap_neigh, tag, MPI_COMM_WORLD, &stat); #pragma omp parallel for for (I=0;I<N;I++) { U[I] = other_big_calc(I, incoming); } consume(U, mpi_id); Technically Requires MPI_Thread_funneled, but I have never had a problem with this approach … even with pre- MPI-2.0 libraries.
  273. 274 Safe Mixing of MPI and OpenMP Protect MPI calls

    inside a parallel region MPI_Init(&argc, &argv) ; MPI_Comm_Rank(MPI_COMM_WORLD, &mpi_id) ; // a whole bunch of initializations #pragma omp parallel { #pragma omp for for (I=0;I<N;I++) U[I] = big_calc(I); #pragma master { MPI_Send (U, BUFF_SIZE, MPI_DOUBLE, neigh, tag, MPI_COMM_WORLD); MPI_Recv (incoming, count, MPI_DOUBLE, neigh, tag, MPI_COMM_WORLD, &stat); } #pragma omp barrier #pragma omp for for (I=0;I<N;I++) U[I] = other_big_calc(I, incoming); #pragma omp master consume(U, mpi_id); } Technically Requires MPI_Thread_funneled, but I have never had a problem with this approach … even with pre- MPI-2.0 libraries.
  274. 275 Hybrid OpenMP/MPI works, but is it worth it? 

    Literature* is mixed on the hybrid model: sometimes its better, sometimes MPI alone is best.  There is potential for benefit to the hybrid model  MPI algorithms often require replicated data making them less memory efficient.  Fewer total MPI communicating agents means fewer messages and less overhead from message conflicts.  Algorithms with good cache efficiency should benefit from shared caches of multi-threaded programs.  The model maps perfectly with clusters of SMP nodes.  But  really,  it’s  a  case  by  case  basis  and  to  large  extent  depends   on the particular application. *L. Adhianto and Chapman, 2007
  275. 276 Appendices  Sources for Additional information  OpenMP History

     Solutions to exercises  Exercise 1: hello world  Exercise 2: Simple SPMD Pi program  Exercise 3: SPMD Pi without false sharing  Exercise 4: Loop level Pi  Exercise 5: Matrix multiplication  Exercise 6: Mandelbrot Set area  Exercise 7: Recursive matrix multiplication  Exercise 8: Monte Carlo Pi and random numbers  Exercise 9: molecular dynamics  Exercise 10: linked lists with tasks  Exercise 11: linked lists without tasks  Flush, memory models and OpenMP: producer consumer  Fortran and OpenMP  Mixing OpenMP and MPI  Compiler Notes
  276. 277 Compiler notes: Intel on Windows  Intel compiler: Launch

     SW  dev  environment  …  on  my  laptop  I  use: – start/intel software development tools/intel C++ compiler 11.0/C+ build environment for 32 bit apps cd to the directory that holds your source code Build software for program foo.c – icl /Qopenmp foo.c Set number of threads environment variable – set OMP_NUM_THREADS=4 Run your program – foo.exe To get rid of the pwd on the prompt, type prompt = %
  277. 278 Compiler notes: Visual Studio  Start  “new  project” 

    Select win 32 console project  Set name and path  On  the  next  panel,  Click  “next”  instead  of  finish  so  you  can   select an empty project on the following panel.  Drag and drop your source file into the source folder on the visual studio solution explorer  Activate OpenMP – Go to project properties/configuration properties/C.C++/language  …  and  activate  OpenMP  Set number of threads inside the program  Build the project  Run  “without  debug”  from  the  debug  menu.
  278. 279 Compiler notes: Other  Linux and OS X with

    gcc: > gcc -fopenmp foo.c > export OMP_NUM_THREADS=4 > ./a.out  Linux and OS X with PGI: > pgcc -mp foo.c > export OMP_NUM_THREADS=4 > ./a.out for the Bash shell
  279. 280 OpenMP constructs  #pragma omp parallel  #pragma omp

    for  #pragma omp critical  #pragma omp atomic  #pragma omp barrier  Data environment clauses  private (variable_list)  firstprivate (variable_list)  lastprivate (variable_list)  reduction(+:variable_list)  Tasks  (remember  …  private  data  is  made  firstprivate  by  default)  pragma omp task  pragma omp taskwait  #pragma threadprivate(variable_list) Where variable_list is a comma separated list of variables Print the value of the macro _OPENMP And its value will be yyyymm For the year and month of the spec the implementation used Put this on a line right after you define the variables in question