Skip to content
GitLab
Explore
Sign in
Register
Primary navigation
Search or go to…
Project
gmsh
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Wiki
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Snippets
Build
Pipelines
Jobs
Pipeline schedules
Artifacts
Deploy
Releases
Model registry
Operate
Environments
Monitor
Incidents
Analyze
Value stream analytics
Contributor analytics
CI/CD analytics
Repository analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
Community forum
Contribute to GitLab
Provide feedback
Terms and privacy
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
Larry Price
gmsh
Commits
a69a0e84
Commit
a69a0e84
authored
15 years ago
by
Christophe Geuzaine
Browse files
Options
Downloads
Patches
Plain Diff
fix non-mpi remote
parent
34adcf24
Branches
Branches containing commit
Tags
Tags containing commit
No related merge requests found
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
Common/GmshRemote.cpp
+30
-24
30 additions, 24 deletions
Common/GmshRemote.cpp
Common/GmshRemote.h
+0
-8
0 additions, 8 deletions
Common/GmshRemote.h
with
30 additions
and
32 deletions
Common/GmshRemote.cpp
+
30
−
24
View file @
a69a0e84
...
@@ -20,6 +20,13 @@
...
@@ -20,6 +20,13 @@
#if defined(HAVE_MPI)
#if defined(HAVE_MPI)
#include
<mpi.h>
#include
<mpi.h>
#define MPI_GMSH_COMPUTE_VIEW 1
#define MPI_GMSH_DATA_READY 2
#define MPI_GMSH_VARRAY 3
#define MPI_GMSH_VARRAY_LEN 4
#define MPI_GMSH_SHUTDOWN 5
#define MPI_GMSH_PARSE_STRING 6
#define MPI_GMSH_MERGE_FILE 7
#endif
#endif
static
void
computeAndSendVertexArrays
(
GmshClient
*
client
,
bool
compute
=
true
)
static
void
computeAndSendVertexArrays
(
GmshClient
*
client
,
bool
compute
=
true
)
...
@@ -93,8 +100,8 @@ static void computeAndSendVertexArrays()
...
@@ -93,8 +100,8 @@ static void computeAndSendVertexArrays()
}
}
// Merge the vertex arrays
// Merge the vertex arrays
void
addToVertexArrays
(
const
char
*
bytes
,
int
len
)
{
void
addToVertexArrays
(
const
char
*
bytes
,
int
len
)
{
int
is
=
sizeof
(
int
),
ds
=
sizeof
(
double
);
int
is
=
sizeof
(
int
),
ds
=
sizeof
(
double
);
std
::
string
name
;
std
::
string
name
;
...
@@ -158,8 +165,8 @@ int GmshRemote()
...
@@ -158,8 +165,8 @@ int GmshRemote()
if
(
client
&&
nbDaemon
<
2
)
computeAndSendVertexArrays
(
client
);
if
(
client
&&
nbDaemon
<
2
)
computeAndSendVertexArrays
(
client
);
while
(
1
){
while
(
1
){
// On the node with MPI rank 0, communicate through a
// socket
.
//
on the node with MPI rank 0, communicate through a
socket
if
(
rank
==
0
)
{
if
(
rank
==
0
)
{
// stop if we have no communications for 5 minutes
// stop if we have no communications for 5 minutes
int
ret
=
client
->
Select
(
300
,
0
);
int
ret
=
client
->
Select
(
300
,
0
);
...
@@ -190,27 +197,27 @@ int GmshRemote()
...
@@ -190,27 +197,27 @@ int GmshRemote()
break
;
break
;
}
}
else
if
(
type
==
GmshSocket
::
GMSH_VERTEX_ARRAY
){
else
if
(
type
==
GmshSocket
::
GMSH_VERTEX_ARRAY
){
#if !defined(HAVE_MPI)
ParseString
(
msg
);
computeAndSendVertexArrays
(
client
);
#else
// FIXME should parse options on each node before computing varrays!
#if defined(HAVE_MPI)
// tell every node to start computing
client
->
Info
(
"Sending vertex arrays..."
);
// Tell every node to start computing
int
mpi_msg
=
MPI_GMSH_COMPUTE_VIEW
;
int
mpi_msg
=
MPI_GMSH_COMPUTE_VIEW
;
MPI_Bcast
(
&
mpi_msg
,
1
,
MPI_INT
,
0
,
MPI_COMM_WORLD
);
MPI_Bcast
(
&
mpi_msg
,
1
,
MPI_INT
,
0
,
MPI_COMM_WORLD
);
// fill the arrays on the master node
// Fill the arrays on the master node
for
(
unsigned
int
i
=
0
;
i
<
PView
::
list
.
size
();
i
++
)
for
(
unsigned
int
i
=
0
;
i
<
PView
::
list
.
size
();
i
++
)
PView
::
list
[
i
]
->
fillVertexArrays
();
PView
::
list
[
i
]
->
fillVertexArrays
();
// wait and send the data from every other node
// Wait and send the data from every other node
for
(
int
i
=
0
;
i
<
nbDaemon
-
1
;
i
++
)
{
for
(
int
i
=
0
;
i
<
nbDaemon
-
1
;
i
++
)
{
int
nbArrays
;
int
nbArrays
;
MPI_Status
status
;
MPI_Status
status
;
MPI_Recv
(
&
nbArrays
,
1
,
MPI_INT
,
MPI_ANY_SOURCE
,
MPI_Recv
(
&
nbArrays
,
1
,
MPI_INT
,
MPI_ANY_SOURCE
,
MPI_GMSH_DATA_READY
,
MPI_COMM_WORLD
,
&
status
);
MPI_GMSH_DATA_READY
,
MPI_COMM_WORLD
,
&
status
);
int
source
=
status
.
MPI_SOURCE
;
int
source
=
status
.
MPI_SOURCE
;
// Get each varray in turn, then add it to the varrays of the master node
// get each varray in turn, then add it to the varrays of
// the master node
for
(
int
j
=
0
;
j
<
nbArrays
;
j
++
)
{
for
(
int
j
=
0
;
j
<
nbArrays
;
j
++
)
{
int
len
;
int
len
;
MPI_Status
status2
;
MPI_Status
status2
;
...
@@ -219,22 +226,21 @@ int GmshRemote()
...
@@ -219,22 +226,21 @@ int GmshRemote()
char
str
[
len
];
char
str
[
len
];
MPI_Recv
(
str
,
len
,
MPI_CHAR
,
status
.
MPI_SOURCE
,
MPI_Recv
(
str
,
len
,
MPI_CHAR
,
status
.
MPI_SOURCE
,
MPI_GMSH_VARRAY
,
MPI_COMM_WORLD
,
&
status2
);
MPI_GMSH_VARRAY
,
MPI_COMM_WORLD
,
&
status2
);
addToVertexArrays
(
str
,
len
);
addToVertexArrays
(
str
,
len
);
}
}
}
}
computeAndSendVertexArrays
(
client
,
false
);
computeAndSendVertexArrays
(
client
,
false
);
#endif
#endif
}
}
else
if
(
type
==
GmshSocket
::
GMSH_MERGE_FILE
){
else
if
(
type
==
GmshSocket
::
GMSH_MERGE_FILE
){
MergeFile
(
msg
);
MergeFile
(
msg
);
#if defined(HAVE_MPI)
#if !defined(HAVE_MPI)
computeAndSendVertexArrays
(
client
);
#else
int
mpi_msg
=
MPI_GMSH_MERGE_FILE
;
int
mpi_msg
=
MPI_GMSH_MERGE_FILE
;
MPI_Bcast
(
&
mpi_msg
,
1
,
MPI_INT
,
0
,
MPI_COMM_WORLD
);
MPI_Bcast
(
&
mpi_msg
,
1
,
MPI_INT
,
0
,
MPI_COMM_WORLD
);
MPI_Bcast
(
&
length
,
1
,
MPI_INT
,
0
,
MPI_COMM_WORLD
);
MPI_Bcast
(
&
length
,
1
,
MPI_INT
,
0
,
MPI_COMM_WORLD
);
MPI_Bcast
(
msg
,
length
,
MPI_CHAR
,
0
,
MPI_COMM_WORLD
);
MPI_Bcast
(
msg
,
length
,
MPI_CHAR
,
0
,
MPI_COMM_WORLD
);
#else
computeAndSendVertexArrays
(
client
);
#endif
#endif
}
}
else
if
(
type
==
GmshSocket
::
GMSH_PARSE_STRING
){
else
if
(
type
==
GmshSocket
::
GMSH_PARSE_STRING
){
...
@@ -256,9 +262,9 @@ int GmshRemote()
...
@@ -256,9 +262,9 @@ int GmshRemote()
}
}
delete
[]
msg
;
delete
[]
msg
;
}
else
{
}
else
{
// if we're not on the master node (rank != 0) wait for him...
#if defined(HAVE_MPI)
#if defined(HAVE_MPI)
// If we're not on the master node, we wait for him...
int
mpi_msg
;
int
mpi_msg
;
MPI_Bcast
(
&
mpi_msg
,
1
,
MPI_INT
,
0
,
MPI_COMM_WORLD
);
MPI_Bcast
(
&
mpi_msg
,
1
,
MPI_INT
,
0
,
MPI_COMM_WORLD
);
if
(
mpi_msg
==
MPI_GMSH_COMPUTE_VIEW
)
if
(
mpi_msg
==
MPI_GMSH_COMPUTE_VIEW
)
...
@@ -280,7 +286,7 @@ int GmshRemote()
...
@@ -280,7 +286,7 @@ int GmshRemote()
MergeFile
(
msg
);
MergeFile
(
msg
);
}
}
#endif
#endif
}
// if (rank != 0)
}
}
}
#if defined(HAVE_MPI)
#if defined(HAVE_MPI)
...
...
This diff is collapsed.
Click to expand it.
Common/GmshRemote.h
+
0
−
8
View file @
a69a0e84
...
@@ -8,14 +8,6 @@
...
@@ -8,14 +8,6 @@
#include
<string>
#include
<string>
#define MPI_GMSH_COMPUTE_VIEW 1
#define MPI_GMSH_DATA_READY 2
#define MPI_GMSH_VARRAY 3
#define MPI_GMSH_VARRAY_LEN 4
#define MPI_GMSH_SHUTDOWN 5
#define MPI_GMSH_PARSE_STRING 6
#define MPI_GMSH_MERGE_FILE 7
int
GmshRemote
();
int
GmshRemote
();
#endif
#endif
...
...
This diff is collapsed.
Click to expand it.
Preview
0%
Loading
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment