Statistiques
| Révision :

root / src / grid / HPL_reduce.c

Historique | Voir | Annoter | Télécharger (7,31 ko)

1 1 equemene
/*
2 1 equemene
 * -- High Performance Computing Linpack Benchmark (HPL)
3 1 equemene
 *    HPL - 2.0 - September 10, 2008
4 1 equemene
 *    Antoine P. Petitet
5 1 equemene
 *    University of Tennessee, Knoxville
6 1 equemene
 *    Innovative Computing Laboratory
7 1 equemene
 *    (C) Copyright 2000-2008 All Rights Reserved
8 1 equemene
 *
9 1 equemene
 * -- Copyright notice and Licensing terms:
10 1 equemene
 *
11 1 equemene
 * Redistribution  and  use in  source and binary forms, with or without
12 1 equemene
 * modification, are  permitted provided  that the following  conditions
13 1 equemene
 * are met:
14 1 equemene
 *
15 1 equemene
 * 1. Redistributions  of  source  code  must retain the above copyright
16 1 equemene
 * notice, this list of conditions and the following disclaimer.
17 1 equemene
 *
18 1 equemene
 * 2. Redistributions in binary form must reproduce  the above copyright
19 1 equemene
 * notice, this list of conditions,  and the following disclaimer in the
20 1 equemene
 * documentation and/or other materials provided with the distribution.
21 1 equemene
 *
22 1 equemene
 * 3. All  advertising  materials  mentioning  features  or  use of this
23 1 equemene
 * software must display the following acknowledgement:
24 1 equemene
 * This  product  includes  software  developed  at  the  University  of
25 1 equemene
 * Tennessee, Knoxville, Innovative Computing Laboratory.
26 1 equemene
 *
27 1 equemene
 * 4. The name of the  University,  the name of the  Laboratory,  or the
28 1 equemene
 * names  of  its  contributors  may  not  be used to endorse or promote
29 1 equemene
 * products  derived   from   this  software  without  specific  written
30 1 equemene
 * permission.
31 1 equemene
 *
32 1 equemene
 * -- Disclaimer:
33 1 equemene
 *
34 1 equemene
 * THIS  SOFTWARE  IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
35 1 equemene
 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES,  INCLUDING,  BUT NOT
36 1 equemene
 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
37 1 equemene
 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY
38 1 equemene
 * OR  CONTRIBUTORS  BE  LIABLE FOR ANY  DIRECT,  INDIRECT,  INCIDENTAL,
39 1 equemene
 * SPECIAL,  EXEMPLARY,  OR  CONSEQUENTIAL DAMAGES  (INCLUDING,  BUT NOT
40 1 equemene
 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
41 1 equemene
 * DATA OR PROFITS; OR BUSINESS INTERRUPTION)  HOWEVER CAUSED AND ON ANY
42 1 equemene
 * THEORY OF LIABILITY, WHETHER IN CONTRACT,  STRICT LIABILITY,  OR TORT
43 1 equemene
 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
44 1 equemene
 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
45 1 equemene
 * ---------------------------------------------------------------------
46 1 equemene
 */
47 1 equemene
/*
48 1 equemene
 * Include files
49 1 equemene
 */
50 1 equemene
#include "hpl.h"
51 1 equemene
52 1 equemene
#ifdef STDC_HEADERS
53 1 equemene
int HPL_reduce
54 1 equemene
(
55 1 equemene
   void *                           BUFFER,
56 1 equemene
   const int                        COUNT,
57 1 equemene
   const HPL_T_TYPE                 DTYPE,
58 1 equemene
   const HPL_T_OP                   OP,
59 1 equemene
   const int                        ROOT,
60 1 equemene
   MPI_Comm                         COMM
61 1 equemene
)
62 1 equemene
#else
63 1 equemene
int HPL_reduce
64 1 equemene
( BUFFER, COUNT, DTYPE, OP, ROOT, COMM )
65 1 equemene
   void *                           BUFFER;
66 1 equemene
   const int                        COUNT;
67 1 equemene
   const HPL_T_TYPE                 DTYPE;
68 1 equemene
   const HPL_T_OP                   OP;
69 1 equemene
   const int                        ROOT;
70 1 equemene
   MPI_Comm                         COMM;
71 1 equemene
#endif
72 1 equemene
{
73 1 equemene
/*
74 1 equemene
 * Purpose
75 1 equemene
 * =======
76 1 equemene
 *
77 1 equemene
 * HPL_reduce performs a global reduce operation across all processes of
78 1 equemene
 * a group.  Note that the input buffer is  used as workarray and in all
79 1 equemene
 * processes but the accumulating process corrupting the original data.
80 1 equemene
 *
81 1 equemene
 * Arguments
82 1 equemene
 * =========
83 1 equemene
 *
84 1 equemene
 * BUFFER  (local input/output)          void *
85 1 equemene
 *         On entry,  BUFFER  points to  the  buffer to be  reduced.  On
86 1 equemene
 *         exit,  and  in process of rank  ROOT  this array contains the
87 1 equemene
 *         reduced data.  This  buffer  is also used as workspace during
88 1 equemene
 *         the operation in the other processes of the group.
89 1 equemene
 *
90 1 equemene
 * COUNT   (global input)                const int
91 1 equemene
 *         On entry,  COUNT  indicates the number of entries in  BUFFER.
92 1 equemene
 *         COUNT must be at least zero.
93 1 equemene
 *
94 1 equemene
 * DTYPE   (global input)                const HPL_T_TYPE
95 1 equemene
 *         On entry,  DTYPE  specifies the type of the buffers operands.
96 1 equemene
 *
97 1 equemene
 * OP      (global input)                const HPL_T_OP
98 1 equemene
 *         On entry, OP is a pointer to the local combine function.
99 1 equemene
 *
100 1 equemene
 * ROOT    (global input)                const int
101 1 equemene
 *         On entry, ROOT is the coordinate of the accumulating process.
102 1 equemene
 *
103 1 equemene
 * COMM    (global/local input)          MPI_Comm
104 1 equemene
 *         The MPI communicator identifying the process collection.
105 1 equemene
 *
106 1 equemene
 * ---------------------------------------------------------------------
107 1 equemene
 */
108 1 equemene
/*
109 1 equemene
 * .. Local Variables ..
110 1 equemene
 */
111 1 equemene
   MPI_Status                 status;
112 1 equemene
   void                       * buffer = NULL;
113 1 equemene
   int                        hplerr=MPI_SUCCESS, d=1, i, ip2=1, mask=0,
114 1 equemene
                              mpierr, mydist, partner, rank, size,
115 1 equemene
                              tag = MSGID_BEGIN_COLL;
116 1 equemene
/* ..
117 1 equemene
 * .. Executable Statements ..
118 1 equemene
 */
119 1 equemene
   if( COUNT <= 0 ) return( MPI_SUCCESS );
120 1 equemene
   mpierr = MPI_Comm_size( COMM, &size );
121 1 equemene
   if( size  == 1 ) return( MPI_SUCCESS );
122 1 equemene
   mpierr = MPI_Comm_rank( COMM, &rank );
123 1 equemene
   i = size - 1; while( i > 1 ) { i >>= 1; d++; }
124 1 equemene
125 1 equemene
   if( DTYPE == HPL_INT )
126 1 equemene
      buffer = (void *)( (int *)   malloc( (size_t)(COUNT) *
127 1 equemene
                                           sizeof( int    ) ) );
128 1 equemene
   else
129 1 equemene
      buffer = (void *)( (double *)malloc( (size_t)(COUNT) *
130 1 equemene
                                           sizeof( double ) ) );
131 1 equemene
132 1 equemene
   if( !( buffer ) )
133 1 equemene
   { HPL_pabort( __LINE__, "HPL_reduce", "Memory allocation failed" ); }
134 1 equemene
135 1 equemene
   if( ( mydist = MModSub( rank, ROOT, size ) ) == 0 )
136 1 equemene
   {
137 1 equemene
      do
138 1 equemene
      {
139 1 equemene
         mpierr = MPI_Recv( buffer, COUNT, HPL_2_MPI_TYPE( DTYPE ),
140 1 equemene
                            MModAdd( ROOT, ip2, size ), tag, COMM,
141 1 equemene
                            &status );
142 1 equemene
         if( mpierr != MPI_SUCCESS ) hplerr = mpierr;
143 1 equemene
         OP( COUNT, buffer, BUFFER, DTYPE );
144 1 equemene
         ip2 <<= 1; d--;
145 1 equemene
      } while( d );
146 1 equemene
   }
147 1 equemene
   else
148 1 equemene
   {
149 1 equemene
      do
150 1 equemene
      {
151 1 equemene
         if( ( mydist & mask ) == 0 )
152 1 equemene
         {
153 1 equemene
            partner = mydist ^ ip2;
154 1 equemene
155 1 equemene
            if( mydist & ip2 )
156 1 equemene
            {
157 1 equemene
               partner = MModAdd( ROOT, partner, size );
158 1 equemene
               mpierr = MPI_Send( BUFFER, COUNT, HPL_2_MPI_TYPE( DTYPE ),
159 1 equemene
                                  partner, tag, COMM );
160 1 equemene
            }
161 1 equemene
            else if( partner < size )
162 1 equemene
            {
163 1 equemene
               partner = MModAdd( ROOT, partner, size );
164 1 equemene
               mpierr  = MPI_Recv( buffer, COUNT, HPL_2_MPI_TYPE( DTYPE ),
165 1 equemene
                                   partner, tag, COMM, &status );
166 1 equemene
               OP( COUNT, buffer, BUFFER, DTYPE );
167 1 equemene
            }
168 1 equemene
            if( mpierr != MPI_SUCCESS ) hplerr = mpierr;
169 1 equemene
         }
170 1 equemene
         mask ^= ip2; ip2 <<= 1; d--;
171 1 equemene
      } while( d );
172 1 equemene
   }
173 1 equemene
   if( buffer ) free( buffer );
174 1 equemene
175 1 equemene
   return( hplerr );
176 1 equemene
/*
177 1 equemene
 * End of HPL_reduce
178 1 equemene
 */
179 1 equemene
}