|
Isorropia: Partitioning, Load Balancing and more
|
This example shows how to use hypergraph partitioning methods with hyperedge weights.
//@HEADER //************************************************************************ // // Isorropia: Partitioning and Load Balancing Package // Copyright (2006) Sandia Corporation // //Under terms of Contract DE-AC04-94AL85000, there is a non-exclusive //license for use of this work by or on behalf of the U.S. Government. // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // 1. Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // // 2. Redistributions in binary form must reproduce the above copyright // notice, this list of conditions and the following disclaimer in the // documentation and/or other materials provided with the distribution. // // 3. Neither the name of the Corporation nor the names of the // contributors may be used to endorse or promote products derived from // this software without specific prior written permission. // // THIS SOFTWARE IS PROVIDED BY SANDIA CORPORATION "AS IS" AND ANY // EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE // IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR // PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL SANDIA CORPORATION OR THE // CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, // EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, // PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR // PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF // LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING // NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS // SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. // //************************************************************************ //@HEADER //-------------------------------------------------------------------- //This file is a self-contained example of creating an Epetra_RowMatrix //object, and using Isorropia to create a rebalanced copy of it using //Zoltan's Hypergraph partitioning. //Hypergraph edge weights are used to influence the repartitioning. //-------------------------------------------------------------------- //Include Isorropia_Exception.hpp only because the helper functions at //the bottom of this file (which create the epetra objects) can //potentially throw exceptions. #include <Isorropia_Exception.hpp> //The Isorropia symbols being demonstrated are declared //in these headers: #include <Isorropia_Epetra.hpp> #include <Isorropia_EpetraCostDescriber.hpp> #include <Isorropia_EpetraRedistributor.hpp> #include <Isorropia_EpetraPartitioner.hpp> #ifdef HAVE_MPI #include <mpi.h> #endif #ifdef HAVE_EPETRA #ifdef HAVE_MPI #include <Epetra_MpiComm.h> #else #include <Epetra_SerialComm.h> #endif #include <Epetra_Map.h> #include <Epetra_Vector.h> #include <Epetra_CrsMatrix.h> #include <Epetra_LinearProblem.h> #endif #include "ispatest_lbeval_utils.hpp" //Declaration for helper-function that creates epetra rowmatrix objects. This //function is implemented at the bottom of this file. #ifdef HAVE_EPETRA Teuchos::RCP<const Epetra_RowMatrix> create_epetra_rowmatrix(int numProcs, int localProc, int local_n); #endif int main(int argc, char** argv) { #if defined(HAVE_MPI) && defined(HAVE_EPETRA) int numProcs = 1; int localProc = 0; //first, set up our MPI environment... MPI_Init(&argc, &argv); MPI_Comm_rank(MPI_COMM_WORLD, &localProc); MPI_Comm_size(MPI_COMM_WORLD, &numProcs); int local_n = 1200; //Create a Epetra_RowMatrix object. Teuchos::RCP<const Epetra_RowMatrix> rowmatrix; try { rowmatrix = create_epetra_rowmatrix(numProcs, localProc, local_n); } catch(std::exception& exc) { std::cout << "vert_weights example: create_epetra_rowmatrix threw" << " exception '" << exc.what() << "' on proc " << localProc << std::endl; MPI_Finalize(); return(-1); } //We'll need a Teuchos::ParameterList object to pass to the //Isorropia::Epetra::Partitioner class. Teuchos::ParameterList paramlist; #ifdef HAVE_ISORROPIA_ZOLTAN // If Zoltan is available, we'll specify that the Zoltan package be // used for the partitioning operation, by creating a parameter // sublist named "Zoltan". // In the sublist, we'll set parameters that we want sent to Zoltan. // (As it turns out, Isorropia selects Zoltan's hypergraph partitioner // by default, so we don't actually need to specify it. But it's // useful for illustration...) paramlist.set("PARTITIONING METHOD", "HYPERGRAPH"); #else // If Zoltan is not available, a simple linear partitioner will be // used to partition such that the number of nonzeros is equal (or // close to equal) on each processor. No parameter is necessary to // specify this. #endif //Now we're going to create a Epetra_Vector with weights to //be used as hypergraph edge weights in the repartitioning operation. // // We think of the rows as vertices of the hypergraph, and the columns // as hyperedges. Our row matrix is square, so we can use the // the row map to indicate how the column weights should be // distributed. Teuchos::RCP<Epetra_Vector> hge_weights = Teuchos::rcp(new Epetra_Vector(rowmatrix->RowMatrixRowMap())); double* vals = hge_weights->Values(); const Epetra_BlockMap& map = rowmatrix->RowMatrixRowMap(); int num = map.NumMyElements(); //For this demo, we'll assign the weights to be elem+1, where 'elem' is //the global-id of the corresponding row. (If we don't use +1, zoltan //complains that the first one has a zero weight.) //Using these linearly-increasing weights should cause the partitioner //to put an UN-EQUAL number of rows on each processor... for(int i=0; i<num; ++i) { vals[i] = 1.0*(map.GID(i)+1); } Teuchos::RCP<Isorropia::Epetra::CostDescriber> costs = Teuchos::rcp(new Isorropia::Epetra::CostDescriber); costs->setHypergraphEdgeWeights(hge_weights); //Now create the partitioner Teuchos::RCP<Isorropia::Epetra::Partitioner> partitioner = Teuchos::rcp(new Isorropia::Epetra::Partitioner(rowmatrix, costs, paramlist)); //Next create a Redistributor object and use it to create a repartitioned //copy of the matrix. Isorropia::Epetra::Redistributor rd(partitioner); Teuchos::RCP<Epetra_CrsMatrix> bal_matrix; //Use a try-catch block because Isorropia will throw an exception //if it encounters an error. if (localProc == 0) { std::cout << " calling Isorropia::Epetra::Redistributor::redistribute..." << std::endl; } try { bal_matrix = rd.redistribute(*rowmatrix); } catch(std::exception& exc) { std::cout << "linsys example: Isorropia::Epetra::Redistributor threw " << "exception '" << exc.what() << "' on proc " << localProc << std::endl; MPI_Finalize(); return(-1); } // Results double bal0, bal1, cutn0, cutn1, cutl0, cutl1; #if 1 // Balance and cut quality before partitioning double goalWeight = 1.0 / (double)numProcs; ispatest::compute_hypergraph_metrics(*rowmatrix, *costs, goalWeight, bal0, cutn0, cutl0); // Balance and cut quality after partitioning Teuchos::RCP<Epetra_Vector> new_weights = rd.redistribute(*hge_weights); Isorropia::Epetra::CostDescriber new_costs; new_costs.setHypergraphEdgeWeights(new_weights); ispatest::compute_hypergraph_metrics(*bal_matrix, new_costs, goalWeight, bal1, cutn1, cutl1); #else std::vector<double> bal(2), cutn(2), cutl(2); Epetra_Import &importer = rd.get_importer(); costs->compareBeforeAndAfterHypergraph(*rowmatrix, *bal_matrix, importer, bal, cutn, cutl); bal0 = bal[0]; cutn0 = cutn[0]; cutl0 = cutl[0]; bal1 = bal[1]; cutn1 = cutn[1]; cutl1 = cutl[1]; #endif if (localProc == 0){ std::cout << "Before partitioning: "; std::cout << "Balance " << bal0 << " cutN " << cutn0 << " cutL " << cutl0; std::cout << std::endl; std::cout << "After partitioning: "; std::cout << "Balance " << bal1 << " cutN " << cutn1 << " cutL " << cutl1; std::cout << std::endl; } MPI_Finalize(); #else std::cout << "vert_weights: must have both MPI and EPETRA. Make sure " << "Trilinos is configured with --enable-mpi and --enable-epetra." << std::endl; #endif return(0); } //Below is the implementation of the helper-function that creates the //epetra rowmatrix for use in the above example program. #if defined(HAVE_MPI) && defined(HAVE_EPETRA) Teuchos::RCP<const Epetra_RowMatrix> create_epetra_rowmatrix(int numProcs, int localProc, int local_n) { if (localProc == 0) { std::cout << " creating Epetra_CrsMatrix with even row-distribution..." << std::endl; } //create an Epetra_CrsMatrix with rows spread evenly over //processors. Epetra_MpiComm comm(MPI_COMM_WORLD); int global_num_rows = numProcs*local_n; Epetra_Map rowmap(global_num_rows, local_n, 0, comm); int nnz_per_row = 9; Teuchos::RCP<Epetra_CrsMatrix> matrix = Teuchos::rcp(new Epetra_CrsMatrix(Copy, rowmap, nnz_per_row)); // Add rows one-at-a-time double negOne = -1.0; double posTwo = 4.0; for (int i=0; i<local_n; i++) { int GlobalRow = matrix->GRID(i); int RowLess1 = GlobalRow - 1; int RowPlus1 = GlobalRow + 1; int RowLess2 = GlobalRow - 2; int RowPlus2 = GlobalRow + 2; int RowLess3 = GlobalRow - 3; int RowPlus3 = GlobalRow + 3; int RowLess4 = GlobalRow - 4; int RowPlus4 = GlobalRow + 4; if (RowLess4>=0) { matrix->InsertGlobalValues(GlobalRow, 1, &negOne, &RowLess4); } if (RowLess3>=0) { matrix->InsertGlobalValues(GlobalRow, 1, &negOne, &RowLess3); } if (RowLess2>=0) { matrix->InsertGlobalValues(GlobalRow, 1, &negOne, &RowLess2); } if (RowLess1>=0) { matrix->InsertGlobalValues(GlobalRow, 1, &negOne, &RowLess1); } if (RowPlus1<global_num_rows) { matrix->InsertGlobalValues(GlobalRow, 1, &negOne, &RowPlus1); } if (RowPlus2<global_num_rows) { matrix->InsertGlobalValues(GlobalRow, 1, &negOne, &RowPlus2); } if (RowPlus3<global_num_rows) { matrix->InsertGlobalValues(GlobalRow, 1, &negOne, &RowPlus3); } if (RowPlus4<global_num_rows) { matrix->InsertGlobalValues(GlobalRow, 1, &negOne, &RowPlus4); } matrix->InsertGlobalValues(GlobalRow, 1, &posTwo, &GlobalRow); } int err = matrix->FillComplete(); if (err != 0) { throw Isorropia::Exception("create_epetra_matrix: error in matrix.FillComplete()"); } return(matrix); } #endif //HAVE_MPI && HAVE_EPETRA