On Wed, Jul 24, 2024 at 9:32 PM Matthew Thomas <matthew.thom...@anu.edu.au> wrote:
> Hi Matt, > > I have attached the configuration file below. > >From the log: MPI: Version: 3 mpiexec: /apps/intel-tools/intel-mpi/2021.11.0/bin/mpiexec Implementation: mpich3 I_MPI_NUMVERSION: 20211100300 MPICH_NUMVERSION: 30400002 so you want to use that mpiexec. This is what will be used if you try make check in $PETSC_DIR. Thanks, Matt > Thanks, > Matt > > > > > On 25 Jul 2024, at 11:26 AM, Matthew Knepley <knep...@gmail.com> wrote: > > On Wed, Jul 24, 2024 at 8:37 PM Matthew Thomas <matthew.thom...@anu.edu.au> > wrote: > > Hello Matt, > > Thanks for the help. I believe the problem is coming from an incorrect > linking with MPI and PETSc. > > I tried running with petscmpiexec from > $PETSC_DIR/lib/petsc/bin/petscmpiexec. This gave me the error > > Error build location not found! Please set PETSC_DIR and PETSC_ARCH > correctly for this build. > > > Naturally I have set these two values and echo $PETSC_DIR gives the path I > expect, so it seems like I am running my programs with a different version > of MPI than petsc expects which could explain the memory usage. > > Do you have any ideas how to fix this? > > > Yes. First we determine what MPI you configured with. Send configure.log, > which has this information. > > Thanks, > > Matt > > > Thanks, > Matt > > On 24 Jul 2024, at 8:41 PM, Matthew Knepley <knep...@gmail.com> wrote: > > You don't often get email from knep...@gmail.com. Learn why this is > important > <https://urldefense.us/v3/__https://aka.ms/LearnAboutSenderIdentification__;!!G_uCfscf7eWS!Z2rv8eb7Cg2HQOiE2c11H9inWPs58oj2IuqkyMjKueg_gDd5G7Ya_P8dkPggDNh_IrYKprQIK0WvlW0CILAi$ > > > On Tue, Jul 23, 2024 at 8:02 PM Matthew Thomas <matthew.thom...@anu.edu.au> > wrote: > > Hello Matt, > > I have attached the output with mat_view for 8 and 40 processors. > > I am unsure what is meant by the matrix communicator and the partitioning. > I am using the default behaviour in every case. How can I find this > information? > > > This shows that the matrix is taking the same amount of memory for 8 and > 40 procs, so that is not your problem. Also, > it is a very small amount of memory: > > 100K rows x 3 nz/row x 8 bytes/nz = 2.4 MB > > and 50% overhead for indexing, so something under 4MB. I am not sure what > is taking up the rest of the memory, but I do not > think it is PETSc from the log you included. > > Thanks, > > Matt > > > I have attached the log view as well if that helps. > > Thanks, > Matt > > > > > On 23 Jul 2024, at 9:24 PM, Matthew Knepley <knep...@gmail.com> wrote: > > You don't often get email from knep...@gmail.com. Learn why this is > important > <https://urldefense.us/v3/__https://aka.ms/LearnAboutSenderIdentification__;!!G_uCfscf7eWS!Z2rv8eb7Cg2HQOiE2c11H9inWPs58oj2IuqkyMjKueg_gDd5G7Ya_P8dkPggDNh_IrYKprQIK0WvlW0CILAi$ > > > Also, you could run with > > -mat_view ::ascii_info_detail > > and send the output for both cases. The storage of matrix values is not > redundant, so something else is > going on. First, what communicator do you use for the matrix, and what > partitioning? > > Thanks, > > Matt > > On Mon, Jul 22, 2024 at 10:27 PM Barry Smith <bsm...@petsc.dev> wrote: > > This Message Is From an External Sender > This message came from outside your organization. > > > Send the code. > > On Jul 22, 2024, at 9:18 PM, Matthew Thomas via petsc-users < > petsc-users@mcs.anl.gov> wrote: > > This Message Is From an External Sender > This message came from outside your organization. > > Hello, > > I am using petsc and slepc to solve an eigenvalue problem for sparse > matrices. When I run my code with double the number of processors, the memory > usage also doubles. > > I am able to reproduce this behaviour with ex1 of slepc’s hands on exercises. > > The issue is occurring with petsc not with slepc as this still occurs when I > remove the solve step and just create and assemble the petsc matrix. > > With n=100000, this uses ~1Gb with 8 processors, but ~5Gb with 40 processors. > > This was done with petsc 3.21.3, on linux compiled with Intel using Intel-MPI > > Is this the expected behaviour? If not, how can I bug fix this? > > > Thanks, > Matt > > > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > > https://urldefense.us/v3/__https://www.cse.buffalo.edu/*knepley/__;fg!!G_uCfscf7eWS!Z2rv8eb7Cg2HQOiE2c11H9inWPs58oj2IuqkyMjKueg_gDd5G7Ya_P8dkPggDNh_IrYKprQIK0WvlSHTmW6H$ > > <https://urldefense.us/v3/__http://www.cse.buffalo.edu/*knepley/__;fg!!G_uCfscf7eWS!Z2rv8eb7Cg2HQOiE2c11H9inWPs58oj2IuqkyMjKueg_gDd5G7Ya_P8dkPggDNh_IrYKprQIK0Wvlfv4A8zx$ > > > > > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > > https://urldefense.us/v3/__https://www.cse.buffalo.edu/*knepley/__;fg!!G_uCfscf7eWS!Z2rv8eb7Cg2HQOiE2c11H9inWPs58oj2IuqkyMjKueg_gDd5G7Ya_P8dkPggDNh_IrYKprQIK0WvlSHTmW6H$ > > <https://urldefense.us/v3/__http://www.cse.buffalo.edu/*knepley/__;fg!!G_uCfscf7eWS!Z2rv8eb7Cg2HQOiE2c11H9inWPs58oj2IuqkyMjKueg_gDd5G7Ya_P8dkPggDNh_IrYKprQIK0Wvlfv4A8zx$ > > > > > > > -- > What most experimenters take for granted before they begin their > experiments is infinitely more interesting than any results to which their > experiments lead. > -- Norbert Wiener > > https://urldefense.us/v3/__https://www.cse.buffalo.edu/*knepley/__;fg!!G_uCfscf7eWS!Z2rv8eb7Cg2HQOiE2c11H9inWPs58oj2IuqkyMjKueg_gDd5G7Ya_P8dkPggDNh_IrYKprQIK0WvlSHTmW6H$ > > <https://urldefense.us/v3/__http://www.cse.buffalo.edu/*knepley/__;fg!!G_uCfscf7eWS!Z2rv8eb7Cg2HQOiE2c11H9inWPs58oj2IuqkyMjKueg_gDd5G7Ya_P8dkPggDNh_IrYKprQIK0Wvlfv4A8zx$ > > > > > -- What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead. -- Norbert Wiener https://urldefense.us/v3/__https://www.cse.buffalo.edu/*knepley/__;fg!!G_uCfscf7eWS!Z2rv8eb7Cg2HQOiE2c11H9inWPs58oj2IuqkyMjKueg_gDd5G7Ya_P8dkPggDNh_IrYKprQIK0WvlSHTmW6H$ <https://urldefense.us/v3/__http://www.cse.buffalo.edu/*knepley/__;fg!!G_uCfscf7eWS!Z2rv8eb7Cg2HQOiE2c11H9inWPs58oj2IuqkyMjKueg_gDd5G7Ya_P8dkPggDNh_IrYKprQIK0Wvlfv4A8zx$ >