159 lines
		
	
	
		
			7.3 KiB
		
	
	
	
		
			HTML
		
	
	
	
	
	
		
		
			
		
	
	
			159 lines
		
	
	
		
			7.3 KiB
		
	
	
	
		
			HTML
		
	
	
	
	
	
|   | <!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN" "http://www.w3.org/TR/html4/loose.dtd"> | ||
|  | <html> | ||
|  | <!-- This manual is for FFTW
 | ||
|  | (version 3.3.10, 10 December 2020). | ||
|  | 
 | ||
|  | Copyright (C) 2003 Matteo Frigo. | ||
|  | 
 | ||
|  | Copyright (C) 2003 Massachusetts Institute of Technology. | ||
|  | 
 | ||
|  | Permission is granted to make and distribute verbatim copies of this | ||
|  | manual provided the copyright notice and this permission notice are | ||
|  | preserved on all copies. | ||
|  | 
 | ||
|  | Permission is granted to copy and distribute modified versions of this | ||
|  | manual under the conditions for verbatim copying, provided that the | ||
|  | entire resulting derived work is distributed under the terms of a | ||
|  | permission notice identical to this one. | ||
|  | 
 | ||
|  | Permission is granted to copy and distribute translations of this manual | ||
|  | into another language, under the above conditions for modified versions, | ||
|  | except that this permission notice may be stated in a translation | ||
|  | approved by the Free Software Foundation. --> | ||
|  | <!-- Created by GNU Texinfo 6.7, http://www.gnu.org/software/texinfo/ --> | ||
|  | <head> | ||
|  | <meta http-equiv="Content-Type" content="text/html; charset=utf-8"> | ||
|  | <title>Combining MPI and Threads (FFTW 3.3.10)</title> | ||
|  | 
 | ||
|  | <meta name="description" content="Combining MPI and Threads (FFTW 3.3.10)"> | ||
|  | <meta name="keywords" content="Combining MPI and Threads (FFTW 3.3.10)"> | ||
|  | <meta name="resource-type" content="document"> | ||
|  | <meta name="distribution" content="global"> | ||
|  | <meta name="Generator" content="makeinfo"> | ||
|  | <link href="index.html" rel="start" title="Top"> | ||
|  | <link href="Concept-Index.html" rel="index" title="Concept Index"> | ||
|  | <link href="index.html#SEC_Contents" rel="contents" title="Table of Contents"> | ||
|  | <link href="Distributed_002dmemory-FFTW-with-MPI.html" rel="up" title="Distributed-memory FFTW with MPI"> | ||
|  | <link href="FFTW-MPI-Reference.html" rel="next" title="FFTW MPI Reference"> | ||
|  | <link href="FFTW-MPI-Performance-Tips.html" rel="prev" title="FFTW MPI Performance Tips"> | ||
|  | <style type="text/css"> | ||
|  | <!--
 | ||
|  | a.summary-letter {text-decoration: none} | ||
|  | blockquote.indentedblock {margin-right: 0em} | ||
|  | div.display {margin-left: 3.2em} | ||
|  | div.example {margin-left: 3.2em} | ||
|  | div.lisp {margin-left: 3.2em} | ||
|  | kbd {font-style: oblique} | ||
|  | pre.display {font-family: inherit} | ||
|  | pre.format {font-family: inherit} | ||
|  | pre.menu-comment {font-family: serif} | ||
|  | pre.menu-preformatted {font-family: serif} | ||
|  | span.nolinebreak {white-space: nowrap} | ||
|  | span.roman {font-family: initial; font-weight: normal} | ||
|  | span.sansserif {font-family: sans-serif; font-weight: normal} | ||
|  | ul.no-bullet {list-style: none} | ||
|  | --> | ||
|  | </style> | ||
|  | 
 | ||
|  | 
 | ||
|  | </head> | ||
|  | 
 | ||
|  | <body lang="en"> | ||
|  | <span id="Combining-MPI-and-Threads"></span><div class="header"> | ||
|  | <p> | ||
|  | Next: <a href="FFTW-MPI-Reference.html" accesskey="n" rel="next">FFTW MPI Reference</a>, Previous: <a href="FFTW-MPI-Performance-Tips.html" accesskey="p" rel="prev">FFTW MPI Performance Tips</a>, Up: <a href="Distributed_002dmemory-FFTW-with-MPI.html" accesskey="u" rel="up">Distributed-memory FFTW with MPI</a>   [<a href="index.html#SEC_Contents" title="Table of contents" rel="contents">Contents</a>][<a href="Concept-Index.html" title="Index" rel="index">Index</a>]</p> | ||
|  | </div> | ||
|  | <hr> | ||
|  | <span id="Combining-MPI-and-Threads-1"></span><h3 class="section">6.11 Combining MPI and Threads</h3> | ||
|  | <span id="index-threads-2"></span> | ||
|  | 
 | ||
|  | <p>In certain cases, it may be advantageous to combine MPI | ||
|  | (distributed-memory) and threads (shared-memory) parallelization. | ||
|  | FFTW supports this, with certain caveats.  For example, if you have a | ||
|  | cluster of 4-processor shared-memory nodes, you may want to use | ||
|  | threads within the nodes and MPI between the nodes, instead of MPI for | ||
|  | all parallelization. | ||
|  | </p> | ||
|  | <p>In particular, it is possible to seamlessly combine the MPI FFTW | ||
|  | routines with the multi-threaded FFTW routines (see <a href="Multi_002dthreaded-FFTW.html">Multi-threaded FFTW</a>). However, some care must be taken in the initialization code, | ||
|  | which should look something like this: | ||
|  | </p> | ||
|  | <div class="example"> | ||
|  | <pre class="example">int threads_ok; | ||
|  | 
 | ||
|  | int main(int argc, char **argv) | ||
|  | { | ||
|  |     int provided; | ||
|  |     MPI_Init_thread(&argc, &argv, MPI_THREAD_FUNNELED, &provided); | ||
|  |     threads_ok = provided >= MPI_THREAD_FUNNELED; | ||
|  | 
 | ||
|  |     if (threads_ok) threads_ok = fftw_init_threads(); | ||
|  |     fftw_mpi_init(); | ||
|  | 
 | ||
|  |     ... | ||
|  |     if (threads_ok) fftw_plan_with_nthreads(...); | ||
|  |     ... | ||
|  |      | ||
|  |     MPI_Finalize(); | ||
|  | } | ||
|  | </pre></div> | ||
|  | <span id="index-fftw_005fmpi_005finit-3"></span> | ||
|  | <span id="index-fftw_005finit_005fthreads-2"></span> | ||
|  | <span id="index-fftw_005fplan_005fwith_005fnthreads-1"></span> | ||
|  | 
 | ||
|  | <p>First, note that instead of calling <code>MPI_Init</code>, you should call | ||
|  | <code>MPI_Init_threads</code>, which is the initialization routine defined | ||
|  | by the MPI-2 standard to indicate to MPI that your program will be | ||
|  | multithreaded.  We pass <code>MPI_THREAD_FUNNELED</code>, which indicates | ||
|  | that we will only call MPI routines from the main thread.  (FFTW will | ||
|  | launch additional threads internally, but the extra threads will not | ||
|  | call MPI code.)  (You may also pass <code>MPI_THREAD_SERIALIZED</code> or | ||
|  | <code>MPI_THREAD_MULTIPLE</code>, which requests additional multithreading | ||
|  | support from the MPI implementation, but this is not required by | ||
|  | FFTW.)  The <code>provided</code> parameter returns what level of threads | ||
|  | support is actually supported by your MPI implementation; this | ||
|  | <em>must</em> be at least <code>MPI_THREAD_FUNNELED</code> if you want to call | ||
|  | the FFTW threads routines, so we define a global variable | ||
|  | <code>threads_ok</code> to record this.  You should only call | ||
|  | <code>fftw_init_threads</code> or <code>fftw_plan_with_nthreads</code> if | ||
|  | <code>threads_ok</code> is true.  For more information on thread safety in | ||
|  | MPI, see the | ||
|  | <a href="http://www.mpi-forum.org/docs/mpi-20-html/node162.htm">MPI and | ||
|  | Threads</a> section of the MPI-2 standard. | ||
|  | <span id="index-thread-safety-2"></span> | ||
|  | </p> | ||
|  | 
 | ||
|  | <p>Second, we must call <code>fftw_init_threads</code> <em>before</em> | ||
|  | <code>fftw_mpi_init</code>.  This is critical for technical reasons having | ||
|  | to do with how FFTW initializes its list of algorithms. | ||
|  | </p> | ||
|  | <p>Then, if you call <code>fftw_plan_with_nthreads(N)</code>, <em>every</em> MPI | ||
|  | process will launch (up to) <code>N</code> threads to parallelize its transforms. | ||
|  | </p> | ||
|  | <p>For example, in the hypothetical cluster of 4-processor nodes, you | ||
|  | might wish to launch only a single MPI process per node, and then call | ||
|  | <code>fftw_plan_with_nthreads(4)</code> on each process to use all | ||
|  | processors in the nodes. | ||
|  | </p> | ||
|  | <p>This may or may not be faster than simply using as many MPI processes | ||
|  | as you have processors, however.  On the one hand, using threads | ||
|  | within a node eliminates the need for explicit message passing within | ||
|  | the node.  On the other hand, FFTW’s transpose routines are not | ||
|  | multi-threaded, and this means that the communications that do take | ||
|  | place will not benefit from parallelization within the node. | ||
|  | Moreover, many MPI implementations already have optimizations to | ||
|  | exploit shared memory when it is available, so adding the | ||
|  | multithreaded FFTW on top of this may be superfluous. | ||
|  | <span id="index-transpose-4"></span> | ||
|  | </p> | ||
|  | <hr> | ||
|  | <div class="header"> | ||
|  | <p> | ||
|  | Next: <a href="FFTW-MPI-Reference.html" accesskey="n" rel="next">FFTW MPI Reference</a>, Previous: <a href="FFTW-MPI-Performance-Tips.html" accesskey="p" rel="prev">FFTW MPI Performance Tips</a>, Up: <a href="Distributed_002dmemory-FFTW-with-MPI.html" accesskey="u" rel="up">Distributed-memory FFTW with MPI</a>   [<a href="index.html#SEC_Contents" title="Table of contents" rel="contents">Contents</a>][<a href="Concept-Index.html" title="Index" rel="index">Index</a>]</p> | ||
|  | </div> | ||
|  | 
 | ||
|  | 
 | ||
|  | 
 | ||
|  | </body> | ||
|  | </html> |