2010-05-12 13:27:55 +00:00
|
|
|
/*---------------------------------------------------------------------------*\
|
|
|
|
========= |
|
|
|
|
\\ / F ield | OpenFOAM: The Open Source CFD Toolbox
|
|
|
|
\\ / O peration |
|
|
|
|
\\ / A nd | Copyright held by original author
|
|
|
|
\\/ M anipulation |
|
|
|
|
-------------------------------------------------------------------------------
|
|
|
|
License
|
|
|
|
This file is part of OpenFOAM.
|
|
|
|
|
|
|
|
OpenFOAM is free software; you can redistribute it and/or modify it
|
|
|
|
under the terms of the GNU General Public License as published by the
|
|
|
|
Free Software Foundation; either version 2 of the License, or (at your
|
|
|
|
option) any later version.
|
|
|
|
|
|
|
|
OpenFOAM is distributed in the hope that it will be useful, but WITHOUT
|
|
|
|
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
|
|
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
|
|
for more details.
|
|
|
|
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
|
|
along with OpenFOAM; if not, write to the Free Software Foundation,
|
|
|
|
Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
|
|
|
|
|
|
|
|
Application
|
|
|
|
redistributeMeshPar
|
|
|
|
|
|
|
|
Description
|
2010-08-26 14:22:03 +00:00
|
|
|
Redistributes existing decomposed mesh and fields according to the current
|
|
|
|
settings in the decomposeParDict file.
|
|
|
|
|
|
|
|
Must be run on maximum number of source and destination processors.
|
|
|
|
Balances mesh and writes new mesh to new time directory.
|
|
|
|
|
|
|
|
Can also work like decomposePar:
|
|
|
|
|
|
|
|
# Create empty processors
|
|
|
|
mkdir processor0
|
|
|
|
..
|
|
|
|
mkdir processorN
|
|
|
|
|
|
|
|
# Copy undecomposed polyMesh
|
|
|
|
cp -r constant processor0
|
|
|
|
|
|
|
|
# Distribute
|
|
|
|
mpirun -np ddd redistributeMeshPar -parallel
|
2010-05-12 13:27:55 +00:00
|
|
|
|
|
|
|
\*---------------------------------------------------------------------------*/
|
|
|
|
|
|
|
|
#include "fvMesh.H"
|
|
|
|
#include "decompositionMethod.H"
|
|
|
|
#include "PstreamReduceOps.H"
|
|
|
|
#include "fvCFD.H"
|
|
|
|
#include "fvMeshDistribute.H"
|
|
|
|
#include "mapDistributePolyMesh.H"
|
|
|
|
#include "IOobjectList.H"
|
|
|
|
|
|
|
|
// * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * //
|
|
|
|
|
|
|
|
// Tolerance (as fraction of the bounding box). Needs to be fairly lax since
|
|
|
|
// usually meshes get written with limited precision (6 digits)
|
|
|
|
static const scalar defaultMergeTol = 1E-6;
|
|
|
|
|
|
|
|
|
|
|
|
// Read mesh if available. Otherwise create empty mesh with same non-proc
|
|
|
|
// patches as proc0 mesh. Requires all processors to have all patches
|
|
|
|
// (and in same order).
|
|
|
|
autoPtr<fvMesh> createMesh
|
|
|
|
(
|
|
|
|
const Time& runTime,
|
2010-08-26 14:22:03 +00:00
|
|
|
const word& regionName,
|
2010-05-12 13:27:55 +00:00
|
|
|
const fileName& instDir,
|
|
|
|
const bool haveMesh
|
|
|
|
)
|
|
|
|
{
|
|
|
|
Pout<< "Create mesh for time = "
|
|
|
|
<< runTime.timeName() << nl << endl;
|
|
|
|
|
2010-08-26 14:22:03 +00:00
|
|
|
IOobject io
|
2010-05-12 13:27:55 +00:00
|
|
|
(
|
2010-08-26 14:22:03 +00:00
|
|
|
regionName,
|
|
|
|
instDir,
|
|
|
|
runTime,
|
|
|
|
IOobject::MUST_READ
|
2010-05-12 13:27:55 +00:00
|
|
|
);
|
|
|
|
|
|
|
|
if (!haveMesh)
|
|
|
|
{
|
2010-08-26 14:22:03 +00:00
|
|
|
// Create dummy mesh. Only used on procs that don't have mesh.
|
|
|
|
fvMesh dummyMesh
|
|
|
|
(
|
|
|
|
io,
|
|
|
|
xferCopy(pointField()),
|
|
|
|
xferCopy(faceList()),
|
|
|
|
xferCopy(labelList()),
|
|
|
|
xferCopy(labelList()),
|
|
|
|
false
|
|
|
|
);
|
|
|
|
Pout<< "Writing dummy mesh to " << dummyMesh.polyMesh::objectPath()
|
|
|
|
<< endl;
|
2010-05-12 13:27:55 +00:00
|
|
|
dummyMesh.write();
|
|
|
|
}
|
|
|
|
|
2010-08-26 14:22:03 +00:00
|
|
|
Pout<< "Reading mesh from " << io.objectPath() << endl;
|
|
|
|
autoPtr<fvMesh> meshPtr(new fvMesh(io));
|
2010-05-12 13:27:55 +00:00
|
|
|
fvMesh& mesh = meshPtr();
|
|
|
|
|
|
|
|
|
|
|
|
// Determine patches.
|
|
|
|
if (Pstream::master())
|
|
|
|
{
|
|
|
|
// Send patches
|
|
|
|
for
|
|
|
|
(
|
|
|
|
int slave=Pstream::firstSlave();
|
|
|
|
slave<=Pstream::lastSlave();
|
|
|
|
slave++
|
|
|
|
)
|
|
|
|
{
|
|
|
|
OPstream toSlave(Pstream::blocking, slave);
|
|
|
|
toSlave << mesh.boundaryMesh();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// Receive patches
|
|
|
|
IPstream fromMaster(Pstream::blocking, Pstream::masterNo());
|
|
|
|
PtrList<entry> patchEntries(fromMaster);
|
|
|
|
|
|
|
|
if (haveMesh)
|
|
|
|
{
|
|
|
|
// Check master names against mine
|
|
|
|
|
|
|
|
const polyBoundaryMesh& patches = mesh.boundaryMesh();
|
|
|
|
|
|
|
|
forAll(patchEntries, patchI)
|
|
|
|
{
|
|
|
|
const entry& e = patchEntries[patchI];
|
|
|
|
const word type(e.dict().lookup("type"));
|
|
|
|
const word& name = e.keyword();
|
|
|
|
|
|
|
|
if (type == processorPolyPatch::typeName)
|
|
|
|
{
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (patchI >= patches.size())
|
|
|
|
{
|
|
|
|
FatalErrorIn
|
|
|
|
(
|
|
|
|
"createMesh(const Time&, const fileName&, const bool)"
|
|
|
|
) << "Non-processor patches not synchronised."
|
|
|
|
<< endl
|
|
|
|
<< "Processor " << Pstream::myProcNo()
|
|
|
|
<< " has only " << patches.size()
|
|
|
|
<< " patches, master has "
|
|
|
|
<< patchI
|
|
|
|
<< exit(FatalError);
|
|
|
|
}
|
|
|
|
|
|
|
|
if
|
|
|
|
(
|
|
|
|
type != patches[patchI].type()
|
|
|
|
|| name != patches[patchI].name()
|
|
|
|
)
|
|
|
|
{
|
|
|
|
FatalErrorIn
|
|
|
|
(
|
|
|
|
"createMesh(const Time&, const fileName&, const bool)"
|
|
|
|
) << "Non-processor patches not synchronised."
|
|
|
|
<< endl
|
|
|
|
<< "Master patch " << patchI
|
|
|
|
<< " name:" << type
|
|
|
|
<< " type:" << type << endl
|
|
|
|
<< "Processor " << Pstream::myProcNo()
|
|
|
|
<< " patch " << patchI
|
|
|
|
<< " has name:" << patches[patchI].name()
|
|
|
|
<< " type:" << patches[patchI].type()
|
|
|
|
<< exit(FatalError);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// Add patch
|
|
|
|
List<polyPatch*> patches(patchEntries.size());
|
|
|
|
label nPatches = 0;
|
|
|
|
|
|
|
|
forAll(patchEntries, patchI)
|
|
|
|
{
|
|
|
|
const entry& e = patchEntries[patchI];
|
|
|
|
const word type(e.dict().lookup("type"));
|
|
|
|
const word& name = e.keyword();
|
|
|
|
|
|
|
|
if (type == processorPolyPatch::typeName)
|
|
|
|
{
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
Pout<< "Adding patch:" << nPatches
|
|
|
|
<< " name:" << name
|
|
|
|
<< " type:" << type << endl;
|
|
|
|
|
|
|
|
dictionary patchDict(e.dict());
|
|
|
|
patchDict.remove("nFaces");
|
|
|
|
patchDict.add("nFaces", 0);
|
|
|
|
patchDict.remove("startFace");
|
|
|
|
patchDict.add("startFace", 0);
|
|
|
|
|
|
|
|
patches[patchI] = polyPatch::New
|
|
|
|
(
|
|
|
|
name,
|
|
|
|
patchDict,
|
|
|
|
nPatches++,
|
|
|
|
mesh.boundaryMesh()
|
|
|
|
).ptr();
|
|
|
|
}
|
|
|
|
patches.setSize(nPatches);
|
|
|
|
mesh.addFvPatches(patches, false); // no parallel comms
|
|
|
|
|
|
|
|
//// Write empty mesh now we have correct patches
|
|
|
|
//meshPtr().write();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!haveMesh)
|
|
|
|
{
|
|
|
|
// We created a dummy mesh file above. Delete it.
|
2010-08-26 14:22:03 +00:00
|
|
|
Pout<< "Removing dummy mesh " << io.objectPath()
|
|
|
|
<< endl;
|
|
|
|
rmDir(io.objectPath());
|
2010-05-12 13:27:55 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Force recreation of globalMeshData.
|
|
|
|
mesh.clearOut();
|
|
|
|
mesh.globalData();
|
|
|
|
|
|
|
|
return meshPtr;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Get merging distance when matching face centres
|
|
|
|
scalar getMergeDistance
|
|
|
|
(
|
|
|
|
const argList& args,
|
|
|
|
const Time& runTime,
|
|
|
|
const boundBox& bb
|
|
|
|
)
|
|
|
|
{
|
|
|
|
scalar mergeTol = defaultMergeTol;
|
2010-08-26 14:22:03 +00:00
|
|
|
args.optionReadIfPresent("mergeTol", mergeTol);
|
|
|
|
|
|
|
|
scalar writeTol =
|
2010-05-12 13:27:55 +00:00
|
|
|
Foam::pow(scalar(10.0), -scalar(IOstream::defaultPrecision()));
|
|
|
|
|
|
|
|
Info<< "Merge tolerance : " << mergeTol << nl
|
|
|
|
<< "Write tolerance : " << writeTol << endl;
|
|
|
|
|
|
|
|
if (runTime.writeFormat() == IOstream::ASCII && mergeTol < writeTol)
|
|
|
|
{
|
|
|
|
FatalErrorIn("getMergeDistance")
|
|
|
|
<< "Your current settings specify ASCII writing with "
|
|
|
|
<< IOstream::defaultPrecision() << " digits precision." << endl
|
|
|
|
<< "Your merging tolerance (" << mergeTol << ") is finer than this."
|
|
|
|
<< endl
|
|
|
|
<< "Please change your writeFormat to binary"
|
|
|
|
<< " or increase the writePrecision" << endl
|
|
|
|
<< "or adjust the merge tolerance (-mergeTol)."
|
|
|
|
<< exit(FatalError);
|
|
|
|
}
|
|
|
|
|
2010-08-26 14:22:03 +00:00
|
|
|
scalar mergeDist = mergeTol * bb.mag();
|
2010-05-12 13:27:55 +00:00
|
|
|
|
|
|
|
Info<< "Overall meshes bounding box : " << bb << nl
|
|
|
|
<< "Relative tolerance : " << mergeTol << nl
|
|
|
|
<< "Absolute matching distance : " << mergeDist << nl
|
|
|
|
<< endl;
|
|
|
|
|
|
|
|
return mergeDist;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
void printMeshData(Ostream& os, const polyMesh& mesh)
|
|
|
|
{
|
2010-08-26 14:22:03 +00:00
|
|
|
os << "Number of points: " << mesh.points().size() << nl
|
|
|
|
<< " faces: " << mesh.faces().size() << nl
|
|
|
|
<< " internal faces: " << mesh.faceNeighbour().size() << nl
|
|
|
|
<< " cells: " << mesh.cells().size() << nl
|
|
|
|
<< " boundary patches: " << mesh.boundaryMesh().size() << nl
|
|
|
|
<< " point zones: " << mesh.pointZones().size() << nl
|
|
|
|
<< " face zones: " << mesh.faceZones().size() << nl
|
|
|
|
<< " cell zones: " << mesh.cellZones().size() << nl;
|
2010-05-12 13:27:55 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Debugging: write volScalarField with decomposition for post processing.
|
|
|
|
void writeDecomposition
|
|
|
|
(
|
|
|
|
const word& name,
|
|
|
|
const fvMesh& mesh,
|
|
|
|
const labelList& decomp
|
|
|
|
)
|
|
|
|
{
|
|
|
|
Info<< "Writing wanted cell distribution to volScalarField " << name
|
2010-08-26 14:22:03 +00:00
|
|
|
<< " for postprocessing purposes." << nl << endl;
|
2010-05-12 13:27:55 +00:00
|
|
|
|
|
|
|
volScalarField procCells
|
|
|
|
(
|
|
|
|
IOobject
|
|
|
|
(
|
|
|
|
name,
|
|
|
|
mesh.time().timeName(),
|
|
|
|
mesh,
|
|
|
|
IOobject::NO_READ,
|
|
|
|
IOobject::AUTO_WRITE,
|
|
|
|
false // do not register
|
|
|
|
),
|
|
|
|
mesh,
|
|
|
|
dimensionedScalar(name, dimless, -1),
|
|
|
|
zeroGradientFvPatchScalarField::typeName
|
|
|
|
);
|
|
|
|
|
|
|
|
forAll(procCells, cI)
|
|
|
|
{
|
|
|
|
procCells[cI] = decomp[cI];
|
|
|
|
}
|
|
|
|
procCells.write();
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Read vol or surface fields
|
|
|
|
//template<class T, class Mesh>
|
|
|
|
template<class GeoField>
|
|
|
|
void readFields
|
|
|
|
(
|
|
|
|
const boolList& haveMesh,
|
|
|
|
const fvMesh& mesh,
|
|
|
|
const autoPtr<fvMeshSubset>& subsetterPtr,
|
|
|
|
IOobjectList& allObjects,
|
|
|
|
PtrList<GeoField>& fields
|
|
|
|
)
|
|
|
|
{
|
|
|
|
//typedef GeometricField<T, fvPatchField, Mesh> fldType;
|
|
|
|
|
|
|
|
// Get my objects of type
|
|
|
|
IOobjectList objects(allObjects.lookupClass(GeoField::typeName));
|
|
|
|
// Check that we all have all objects
|
|
|
|
wordList objectNames = objects.toc();
|
|
|
|
// Get master names
|
|
|
|
wordList masterNames(objectNames);
|
|
|
|
Pstream::scatter(masterNames);
|
|
|
|
|
|
|
|
if (haveMesh[Pstream::myProcNo()] && objectNames != masterNames)
|
|
|
|
{
|
|
|
|
FatalErrorIn("readFields()")
|
|
|
|
<< "differing fields of type " << GeoField::typeName
|
|
|
|
<< " on processors." << endl
|
|
|
|
<< "Master has:" << masterNames << endl
|
|
|
|
<< Pstream::myProcNo() << " has:" << objectNames
|
|
|
|
<< abort(FatalError);
|
|
|
|
}
|
|
|
|
|
|
|
|
fields.setSize(masterNames.size());
|
|
|
|
|
|
|
|
// Have master send all fields to processors that don't have a mesh
|
|
|
|
if (Pstream::master())
|
|
|
|
{
|
|
|
|
forAll(masterNames, i)
|
|
|
|
{
|
|
|
|
const word& name = masterNames[i];
|
|
|
|
IOobject& io = *objects[name];
|
|
|
|
io.writeOpt() = IOobject::AUTO_WRITE;
|
|
|
|
|
|
|
|
// Load field
|
|
|
|
fields.set(i, new GeoField(io, mesh));
|
|
|
|
|
|
|
|
// Create zero sized field and send
|
|
|
|
if (subsetterPtr.valid())
|
|
|
|
{
|
|
|
|
tmp<GeoField> tsubfld = subsetterPtr().interpolate(fields[i]);
|
|
|
|
|
|
|
|
// Send to all processors that don't have a mesh
|
|
|
|
for (label procI = 1; procI < Pstream::nProcs(); procI++)
|
|
|
|
{
|
|
|
|
if (!haveMesh[procI])
|
|
|
|
{
|
|
|
|
OPstream toProc(Pstream::blocking, procI);
|
|
|
|
toProc<< tsubfld();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else if (!haveMesh[Pstream::myProcNo()])
|
|
|
|
{
|
|
|
|
// Don't have mesh (nor fields). Receive empty field from master.
|
|
|
|
|
|
|
|
forAll(masterNames, i)
|
|
|
|
{
|
|
|
|
const word& name = masterNames[i];
|
|
|
|
|
|
|
|
// Receive field
|
|
|
|
IPstream fromMaster(Pstream::blocking, Pstream::masterNo());
|
|
|
|
|
|
|
|
fields.set
|
|
|
|
(
|
|
|
|
i,
|
|
|
|
new GeoField
|
|
|
|
(
|
|
|
|
IOobject
|
|
|
|
(
|
|
|
|
name,
|
|
|
|
mesh.time().timeName(),
|
|
|
|
mesh,
|
|
|
|
IOobject::NO_READ,
|
|
|
|
IOobject::AUTO_WRITE
|
|
|
|
),
|
|
|
|
mesh,
|
|
|
|
fromMaster
|
|
|
|
)
|
|
|
|
);
|
|
|
|
|
|
|
|
//// Write it for next time round (since mesh gets written as well)
|
|
|
|
//fields[i].write();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
// Have mesh so just try to load
|
|
|
|
forAll(masterNames, i)
|
|
|
|
{
|
|
|
|
const word& name = masterNames[i];
|
|
|
|
IOobject& io = *objects[name];
|
|
|
|
io.writeOpt() = IOobject::AUTO_WRITE;
|
|
|
|
|
|
|
|
// Load field
|
|
|
|
fields.set(i, new GeoField(io, mesh));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Debugging: compare two fields.
|
|
|
|
void compareFields
|
|
|
|
(
|
|
|
|
const scalar tolDim,
|
|
|
|
const volVectorField& a,
|
|
|
|
const volVectorField& b
|
|
|
|
)
|
|
|
|
{
|
|
|
|
forAll(a, cellI)
|
|
|
|
{
|
|
|
|
if (mag(b[cellI] - a[cellI]) > tolDim)
|
|
|
|
{
|
|
|
|
FatalErrorIn
|
|
|
|
(
|
|
|
|
"compareFields"
|
|
|
|
"(const scalar, const volVectorField&, const volVectorField&)"
|
|
|
|
) << "Did not map volVectorField correctly:" << nl
|
|
|
|
<< "cell:" << cellI
|
|
|
|
<< " transfer b:" << b[cellI]
|
|
|
|
<< " real cc:" << a[cellI]
|
|
|
|
<< abort(FatalError);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
forAll(a.boundaryField(), patchI)
|
|
|
|
{
|
|
|
|
// We have real mesh cellcentre and
|
|
|
|
// mapped original cell centre.
|
|
|
|
|
|
|
|
const fvPatchVectorField& aBoundary =
|
|
|
|
a.boundaryField()[patchI];
|
|
|
|
|
|
|
|
const fvPatchVectorField& bBoundary =
|
|
|
|
b.boundaryField()[patchI];
|
|
|
|
|
|
|
|
if (!bBoundary.coupled())
|
|
|
|
{
|
|
|
|
forAll(aBoundary, i)
|
|
|
|
{
|
|
|
|
if (mag(aBoundary[i] - bBoundary[i]) > tolDim)
|
|
|
|
{
|
|
|
|
FatalErrorIn
|
|
|
|
(
|
|
|
|
"compareFields"
|
|
|
|
"(const scalar, const volVectorField&"
|
|
|
|
", const volVectorField&)"
|
|
|
|
) << "Did not map volVectorField correctly:"
|
|
|
|
<< endl
|
|
|
|
<< "patch:" << patchI << " patchFace:" << i
|
|
|
|
<< " cc:" << endl
|
|
|
|
<< " real :" << aBoundary[i] << endl
|
|
|
|
<< " mapped :" << bBoundary[i] << endl
|
|
|
|
<< abort(FatalError);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Main program:
|
|
|
|
|
|
|
|
int main(int argc, char *argv[])
|
|
|
|
{
|
2010-08-26 14:22:03 +00:00
|
|
|
# include "addRegionOption.H"
|
2010-05-12 13:27:55 +00:00
|
|
|
argList::validOptions.insert("mergeTol", "relative merge distance");
|
2010-08-26 14:22:03 +00:00
|
|
|
|
|
|
|
// Create argList. This will check for non-existing processor dirs.
|
2010-05-12 13:27:55 +00:00
|
|
|
# include "setRootCase.H"
|
|
|
|
|
2010-08-26 14:22:03 +00:00
|
|
|
//- Not useful anymore. See above.
|
|
|
|
//// Create processor directory if non-existing
|
|
|
|
//if (!Pstream::master() && !isDir(args.path()))
|
|
|
|
//{
|
|
|
|
// Pout<< "Creating case directory " << args.path() << endl;
|
|
|
|
// mkDir(args.path());
|
|
|
|
//}
|
|
|
|
|
|
|
|
# include "createTime.H"
|
|
|
|
|
|
|
|
word regionName = polyMesh::defaultRegion;
|
|
|
|
fileName meshSubDir;
|
|
|
|
|
|
|
|
if (args.optionReadIfPresent("region", regionName))
|
2010-05-12 13:27:55 +00:00
|
|
|
{
|
2010-08-26 14:22:03 +00:00
|
|
|
meshSubDir = regionName/polyMesh::meshSubDir;
|
2010-05-12 13:27:55 +00:00
|
|
|
}
|
2010-08-26 14:22:03 +00:00
|
|
|
else
|
|
|
|
{
|
|
|
|
meshSubDir = polyMesh::meshSubDir;
|
|
|
|
}
|
|
|
|
Info<< "Using mesh subdirectory " << meshSubDir << nl << endl;
|
2010-05-12 13:27:55 +00:00
|
|
|
|
|
|
|
|
|
|
|
// Get time instance directory. Since not all processors have meshes
|
|
|
|
// just use the master one everywhere.
|
|
|
|
|
|
|
|
fileName masterInstDir;
|
|
|
|
if (Pstream::master())
|
|
|
|
{
|
2010-08-26 14:22:03 +00:00
|
|
|
masterInstDir = runTime.findInstance(meshSubDir, "points");
|
2010-05-12 13:27:55 +00:00
|
|
|
}
|
|
|
|
Pstream::scatter(masterInstDir);
|
|
|
|
|
|
|
|
// Check who has a mesh
|
2010-08-26 14:22:03 +00:00
|
|
|
const fileName meshPath = runTime.path()/masterInstDir/meshSubDir;
|
|
|
|
|
|
|
|
Info<< "Found points in " << meshPath << nl << endl;
|
|
|
|
|
2010-05-12 13:27:55 +00:00
|
|
|
|
|
|
|
boolList haveMesh(Pstream::nProcs(), false);
|
2010-08-26 14:22:03 +00:00
|
|
|
haveMesh[Pstream::myProcNo()] = isDir(meshPath);
|
2010-05-12 13:27:55 +00:00
|
|
|
Pstream::gatherList(haveMesh);
|
|
|
|
Pstream::scatterList(haveMesh);
|
|
|
|
Info<< "Per processor mesh availability : " << haveMesh << endl;
|
|
|
|
const bool allHaveMesh = (findIndex(haveMesh, false) == -1);
|
|
|
|
|
|
|
|
// Create mesh
|
|
|
|
autoPtr<fvMesh> meshPtr = createMesh
|
|
|
|
(
|
|
|
|
runTime,
|
2010-08-26 14:22:03 +00:00
|
|
|
regionName,
|
2010-05-12 13:27:55 +00:00
|
|
|
masterInstDir,
|
|
|
|
haveMesh[Pstream::myProcNo()]
|
|
|
|
);
|
|
|
|
fvMesh& mesh = meshPtr();
|
|
|
|
|
|
|
|
Pout<< "Read mesh:" << endl;
|
|
|
|
printMeshData(Pout, mesh);
|
|
|
|
Pout<< endl;
|
|
|
|
|
|
|
|
IOdictionary decompositionDict
|
|
|
|
(
|
|
|
|
IOobject
|
|
|
|
(
|
|
|
|
"decomposeParDict",
|
|
|
|
runTime.system(),
|
|
|
|
mesh,
|
|
|
|
IOobject::MUST_READ,
|
|
|
|
IOobject::NO_WRITE
|
|
|
|
)
|
|
|
|
);
|
|
|
|
|
|
|
|
labelList finalDecomp;
|
|
|
|
|
|
|
|
// Create decompositionMethod and new decomposition
|
|
|
|
{
|
|
|
|
autoPtr<decompositionMethod> decomposer
|
|
|
|
(
|
|
|
|
decompositionMethod::New
|
|
|
|
(
|
|
|
|
decompositionDict,
|
|
|
|
mesh
|
|
|
|
)
|
|
|
|
);
|
|
|
|
|
|
|
|
if (!decomposer().parallelAware())
|
|
|
|
{
|
|
|
|
WarningIn(args.executable())
|
|
|
|
<< "You have selected decomposition method "
|
|
|
|
<< decomposer().typeName
|
|
|
|
<< " which does" << endl
|
|
|
|
<< "not synchronise the decomposition across"
|
|
|
|
<< " processor patches." << endl
|
|
|
|
<< " You might want to select a decomposition method which"
|
|
|
|
<< " is aware of this. Continuing."
|
|
|
|
<< endl;
|
|
|
|
}
|
|
|
|
|
|
|
|
finalDecomp = decomposer().decompose(mesh.cellCentres());
|
|
|
|
}
|
|
|
|
|
|
|
|
// Dump decomposition to volScalarField
|
|
|
|
writeDecomposition("decomposition", mesh, finalDecomp);
|
|
|
|
|
|
|
|
|
|
|
|
// Create 0 sized mesh to do all the generation of zero sized
|
|
|
|
// fields on processors that have zero sized meshes. Note that this is
|
|
|
|
// only nessecary on master but since polyMesh construction with
|
|
|
|
// Pstream::parRun does parallel comms we have to do it on all
|
|
|
|
// processors
|
|
|
|
autoPtr<fvMeshSubset> subsetterPtr;
|
|
|
|
|
|
|
|
if (!allHaveMesh)
|
|
|
|
{
|
|
|
|
// Find last non-processor patch.
|
|
|
|
const polyBoundaryMesh& patches = mesh.boundaryMesh();
|
|
|
|
|
|
|
|
label nonProcI = -1;
|
|
|
|
|
|
|
|
forAll(patches, patchI)
|
|
|
|
{
|
|
|
|
if (isA<processorPolyPatch>(patches[patchI]))
|
|
|
|
{
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
nonProcI++;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (nonProcI == -1)
|
|
|
|
{
|
|
|
|
FatalErrorIn(args.executable())
|
|
|
|
<< "Cannot find non-processor patch on processor "
|
|
|
|
<< Pstream::myProcNo() << endl
|
|
|
|
<< " Current patches:" << patches.names() << abort(FatalError);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Subset 0 cells, no parallel comms. This is used to create zero-sized
|
|
|
|
// fields.
|
2010-08-26 14:22:03 +00:00
|
|
|
subsetterPtr.reset(new fvMeshSubset(mesh));
|
2010-05-12 13:27:55 +00:00
|
|
|
subsetterPtr().setLargeCellSubset(labelHashSet(0), nonProcI, false);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// Get original objects (before incrementing time!)
|
|
|
|
IOobjectList objects(mesh, runTime.timeName());
|
|
|
|
// We don't want to map the decomposition (mapping already tested when
|
|
|
|
// mapping the cell centre field)
|
|
|
|
IOobjectList::iterator iter = objects.find("decomposition");
|
|
|
|
if (iter != objects.end())
|
|
|
|
{
|
|
|
|
objects.erase(iter);
|
|
|
|
}
|
|
|
|
|
|
|
|
PtrList<volScalarField> volScalarFields;
|
|
|
|
readFields
|
|
|
|
(
|
|
|
|
haveMesh,
|
|
|
|
mesh,
|
|
|
|
subsetterPtr,
|
|
|
|
objects,
|
|
|
|
volScalarFields
|
|
|
|
);
|
|
|
|
|
|
|
|
PtrList<volVectorField> volVectorFields;
|
|
|
|
readFields
|
|
|
|
(
|
|
|
|
haveMesh,
|
|
|
|
mesh,
|
|
|
|
subsetterPtr,
|
|
|
|
objects,
|
|
|
|
volVectorFields
|
|
|
|
);
|
|
|
|
|
|
|
|
PtrList<volSphericalTensorField> volSphereTensorFields;
|
|
|
|
readFields
|
|
|
|
(
|
|
|
|
haveMesh,
|
|
|
|
mesh,
|
|
|
|
subsetterPtr,
|
|
|
|
objects,
|
|
|
|
volSphereTensorFields
|
|
|
|
);
|
|
|
|
|
|
|
|
PtrList<volSymmTensorField> volSymmTensorFields;
|
|
|
|
readFields
|
|
|
|
(
|
|
|
|
haveMesh,
|
|
|
|
mesh,
|
|
|
|
subsetterPtr,
|
|
|
|
objects,
|
|
|
|
volSymmTensorFields
|
|
|
|
);
|
|
|
|
|
|
|
|
PtrList<volTensorField> volTensorFields;
|
|
|
|
readFields
|
|
|
|
(
|
|
|
|
haveMesh,
|
|
|
|
mesh,
|
|
|
|
subsetterPtr,
|
|
|
|
objects,
|
|
|
|
volTensorFields
|
|
|
|
);
|
|
|
|
|
|
|
|
PtrList<surfaceScalarField> surfScalarFields;
|
|
|
|
readFields
|
|
|
|
(
|
|
|
|
haveMesh,
|
|
|
|
mesh,
|
|
|
|
subsetterPtr,
|
|
|
|
objects,
|
|
|
|
surfScalarFields
|
|
|
|
);
|
|
|
|
|
|
|
|
PtrList<surfaceVectorField> surfVectorFields;
|
|
|
|
readFields
|
|
|
|
(
|
|
|
|
haveMesh,
|
|
|
|
mesh,
|
|
|
|
subsetterPtr,
|
|
|
|
objects,
|
|
|
|
surfVectorFields
|
|
|
|
);
|
|
|
|
|
|
|
|
PtrList<surfaceSphericalTensorField> surfSphereTensorFields;
|
|
|
|
readFields
|
|
|
|
(
|
|
|
|
haveMesh,
|
|
|
|
mesh,
|
|
|
|
subsetterPtr,
|
|
|
|
objects,
|
|
|
|
surfSphereTensorFields
|
|
|
|
);
|
|
|
|
|
|
|
|
PtrList<surfaceSymmTensorField> surfSymmTensorFields;
|
|
|
|
readFields
|
|
|
|
(
|
|
|
|
haveMesh,
|
|
|
|
mesh,
|
|
|
|
subsetterPtr,
|
|
|
|
objects,
|
|
|
|
surfSymmTensorFields
|
|
|
|
);
|
|
|
|
|
|
|
|
PtrList<surfaceTensorField> surfTensorFields;
|
|
|
|
readFields
|
|
|
|
(
|
|
|
|
haveMesh,
|
|
|
|
mesh,
|
|
|
|
subsetterPtr,
|
|
|
|
objects,
|
|
|
|
surfTensorFields
|
|
|
|
);
|
|
|
|
|
|
|
|
|
|
|
|
// Debugging: Create additional volField that will be mapped.
|
|
|
|
// Used to test correctness of mapping
|
|
|
|
volVectorField mapCc("mapCc", 1*mesh.C());
|
|
|
|
|
|
|
|
// Global matching tolerance
|
|
|
|
const scalar tolDim = getMergeDistance
|
|
|
|
(
|
|
|
|
args,
|
|
|
|
runTime,
|
|
|
|
mesh.globalData().bb()
|
|
|
|
);
|
|
|
|
|
|
|
|
// Mesh distribution engine
|
|
|
|
fvMeshDistribute distributor(mesh, tolDim);
|
|
|
|
|
|
|
|
Pout<< "Wanted distribution:"
|
|
|
|
<< distributor.countCells(finalDecomp) << nl << endl;
|
|
|
|
|
|
|
|
// Do actual sending/receiving of mesh
|
|
|
|
autoPtr<mapDistributePolyMesh> map = distributor.distribute(finalDecomp);
|
|
|
|
|
|
|
|
//// Distribute any non-registered data accordingly
|
|
|
|
//map().distributeFaceData(faceCc);
|
|
|
|
|
|
|
|
|
|
|
|
// Print a bit
|
|
|
|
Pout<< "After distribution mesh:" << endl;
|
|
|
|
printMeshData(Pout, mesh);
|
|
|
|
Pout<< endl;
|
|
|
|
|
|
|
|
runTime++;
|
|
|
|
Pout<< "Writing redistributed mesh to " << runTime.timeName() << nl << endl;
|
|
|
|
mesh.write();
|
|
|
|
|
|
|
|
|
|
|
|
// Debugging: test mapped cellcentre field.
|
|
|
|
compareFields(tolDim, mesh.C(), mapCc);
|
|
|
|
|
|
|
|
// Print nice message
|
|
|
|
// ~~~~~~~~~~~~~~~~~~
|
|
|
|
|
|
|
|
// Determine which processors remain so we can print nice final message.
|
|
|
|
labelList nFaces(Pstream::nProcs());
|
|
|
|
nFaces[Pstream::myProcNo()] = mesh.nFaces();
|
|
|
|
Pstream::gatherList(nFaces);
|
|
|
|
Pstream::scatterList(nFaces);
|
|
|
|
|
|
|
|
Info<< nl
|
|
|
|
<< "You can pick up the redecomposed mesh from the polyMesh directory"
|
|
|
|
<< " in " << runTime.timeName() << "." << nl
|
|
|
|
<< "If you redecomposed the mesh to less processors you can delete"
|
|
|
|
<< nl
|
|
|
|
<< "the processor directories with 0 sized meshes in them." << nl
|
|
|
|
<< "Below is a sample set of commands to do this."
|
2010-08-26 14:22:03 +00:00
|
|
|
<< " Take care when issuing these" << nl
|
2010-05-12 13:27:55 +00:00
|
|
|
<< "commands." << nl << endl;
|
|
|
|
|
|
|
|
forAll(nFaces, procI)
|
|
|
|
{
|
|
|
|
fileName procDir = "processor" + name(procI);
|
|
|
|
|
|
|
|
if (nFaces[procI] == 0)
|
|
|
|
{
|
|
|
|
Info<< " rm -r " << procDir.c_str() << nl;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
2010-08-26 14:22:03 +00:00
|
|
|
fileName timeDir = procDir/runTime.timeName()/meshSubDir;
|
|
|
|
fileName constDir = procDir/runTime.constant()/meshSubDir;
|
2010-05-12 13:27:55 +00:00
|
|
|
|
|
|
|
Info<< " rm -r " << constDir.c_str() << nl
|
|
|
|
<< " mv " << timeDir.c_str()
|
|
|
|
<< ' ' << constDir.c_str() << nl;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
Info<< endl;
|
|
|
|
|
|
|
|
|
|
|
|
Pout<< "End\n" << endl;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// ************************************************************************* //
|