Good, some helpful information with the runs you made. In the crash below it made more progress it was able to allocate multiple regions of shared memory and access them. I don't know why it would crash later.
Can you tell me all the steps with miniforge (which seems to be related to the failure) you use? I've never used miniforge. If I can get to an environment that reproduces the problem I can debug it and fix it. Barry > On Oct 24, 2024, at 11:49 AM, Praveen C <cprav...@gmail.com> wrote: > > I get this > > $ mpiexec -n 3 ./ex89f -n 20 -mpi_linear_solver_server > -mpi_linear_solver_server -mpi_linear_solver_server_ksp_view -ksp_monitor > -ksp_converged_reason -ksp_view > -mpi_linear_solver_server_minimum_count_per_rank 5 > [0]PETSC ERROR: > ------------------------------------------------------------------------ > [0]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation, > probably memory access out of range > [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger > [0]PETSC ERROR: or see > https://urldefense.us/v3/__https://petsc.org/release/faq/*valgrind__;Iw!!G_uCfscf7eWS!a83YNiE6vfZTxyhWxnJxmh95Z63M4HapLLI6b0Q_hFwekNKgjv2j3t61lo38kvy2gLDVjWO5IScFYBVUNuMuh34$ > and > https://urldefense.us/v3/__https://petsc.org/release/faq/__;!!G_uCfscf7eWS!a83YNiE6vfZTxyhWxnJxmh95Z63M4HapLLI6b0Q_hFwekNKgjv2j3t61lo38kvy2gLDVjWO5IScFYBVUJaIxcoc$ > > [0]PETSC ERROR: configure using --with-debugging=yes, recompile, link, and > run > [0]PETSC ERROR: to get more information on the crash. > [0]PETSC ERROR: Run with -malloc_debug to check if memory corruption is > causing the crash. > -------------------------------------------------------------------------- > MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_SELF > Proc: [[47380,1],0] > Errorcode: 59 > > NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes. > You may or may not see output from other processes, depending on > exactly when Open MPI kills them. > -------------------------------------------------------------------------- > -------------------------------------------------------------------------- > prterun has exited due to process rank 0 with PID 0 on node MacMiniHome > calling > "abort". This may have caused other processes in the application to be > terminated by signals sent by prterun (as reported here). > ————————————————————————————————————— > > and I have this after the code exits > > $ ipcs -m > IPC status from <running system> as of Thu Oct 24 21:17:39 IST 2024 > T ID KEY MODE OWNER GROUP > Shared Memory: > m 1572864 0x0000000b --rw-rw-rw- praveen staff > m 524289 0x0000000c --rw-rw-rw- praveen staff > m 655362 0x0000000d --rw-rw-rw- praveen staff > m 262147 0x0000000e --rw-rw-rw- praveen staff > m 262148 0x0000000f --rw-rw-rw- praveen staff > m 393221 0x0000000a --rw-rw-rw- praveen staff > > This is with petsc installed with miniforge, which I also use with clawpack. > With spack installed petsc, I can run the ex89f example. > > Thanks > praveen > >> On 24 Oct 2024, at 7:55 PM, Barry Smith <bsm...@petsc.dev> wrote: >> >> >> Ok, super strange, I've run many times on my Mac >> >> Can you please try to remove that allocated memory with ipcrm and then >> >> cd $PETSC_DIR/src/ksp/ksp/tutorials >> make ex89f >> mpiexec -n 3 ./ex89f -n 20 -mpi_linear_solver_server >> -mpi_linear_solver_server -mpi_linear_solver_server_ksp_view -ksp_monitor >> -ksp_converged_reason -ksp_view >> -mpi_linear_solver_server_minimum_count_per_rank 5 >> >> This does the same thing as the GeoClaw code but is much simpler. >> >> Barry >> >> >> >> >>> On Oct 23, 2024, at 10:55 PM, Praveen C <cprav...@gmail.com> wrote: >>> >>> I get very similar error on my mac with >>> >>> $ gfortran -v >>> Using built-in specs. >>> COLLECT_GCC=gfortran >>> COLLECT_LTO_WRAPPER=/opt/homebrew/Caskroom/miniforge/base/envs/claw/libexec/gcc/arm64-apple-darwin20.0.0/13.2.0/lto-wrapper >>> Target: arm64-apple-darwin20.0.0 >>> Configured with: ../configure >>> --prefix=/opt/homebrew/Caskroom/miniforge/base/envs/claw >>> --build=x86_64-apple-darwin13.4.0 --host=arm64-apple-darwin20.0.0 >>> --target=arm64-apple-darwin20.0.0 >>> --with-libiconv-prefix=/opt/homebrew/Caskroom/miniforge/base/envs/claw >>> --enable-languages=fortran --disable-multilib --enable-checking=release >>> --disable-bootstrap --disable-libssp >>> --with-gmp=/opt/homebrew/Caskroom/miniforge/base/envs/claw >>> --with-mpfr=/opt/homebrew/Caskroom/miniforge/base/envs/claw >>> --with-mpc=/opt/homebrew/Caskroom/miniforge/base/envs/claw >>> --with-isl=/opt/homebrew/Caskroom/miniforge/base/envs/claw >>> --enable-darwin-at-rpath >>> Thread model: posix >>> Supported LTO compression algorithms: zlib >>> gcc version 13.2.0 (GCC) >>> >>> Before starting >>> >>> $ ipcs -m >>> IPC status from <running system> as of Thu Oct 24 08:02:11 IST 2024 >>> T ID KEY MODE OWNER GROUP >>> Shared Memory: >>> >>> and when I run the code >>> >>> Using a PETSc solver >>> Using Bouss equations from the start >>> rnode allocated... >>> node allocated... >>> listOfGrids allocated... >>> Storage allocated... >>> bndList allocated... >>> Gridding level 1 at t = 0.000000E+00: 4 grids with 10000 cells >>> Setting initial dt to 2.9999999999999999E-002 >>> max threads set to 1 >>> >>> Done reading data, starting computation ... >>> >>> Total zeta at initial time: 39269.907650665169 >>> GEOCLAW: Frame 0 output files done at time t = 0.000000D+00 >>> >>> [0]PETSC ERROR: --------------------- Error Message >>> -------------------------------------------------------------- >>> [0]PETSC ERROR: Petsc has generated inconsistent data >>> [0]PETSC ERROR: Unable to locate PCMPI allocated shared address 0x130698000 >>> [0]PETSC ERROR: WARNING! There are unused option(s) set! Could be the >>> program crashed before usage or a spelling mistake, etc! >>> [0]PETSC ERROR: Option left: name:-ksp_type value: preonly source: file >>> [0]PETSC ERROR: Option left: name:-mpi_ksp_max_it value: 200 source: file >>> [0]PETSC ERROR: Option left: name:-mpi_ksp_reuse_preconditioner (no >>> value) source: file >>> [0]PETSC ERROR: Option left: name:-mpi_ksp_rtol value: 1.e-9 source: file >>> [0]PETSC ERROR: Option left: name:-mpi_ksp_type value: gmres source: file >>> [0]PETSC ERROR: Option left: name:-mpi_linear_solver_server_view (no >>> value) source: file >>> [0]PETSC ERROR: Option left: name:-mpi_pc_gamg_sym_graph value: true >>> source: file >>> [0]PETSC ERROR: Option left: name:-mpi_pc_gamg_symmetrize_graph value: >>> true source: file >>> [0]PETSC ERROR: Option left: name:-mpi_pc_type value: gamg source: file >>> [0]PETSC ERROR: Option left: name:-pc_mpi_minimum_count_per_rank value: >>> 5000 source: file >>> [0]PETSC ERROR: Option left: name:-pc_type value: mpi source: file >>> [0]PETSC ERROR: See >>> https://urldefense.us/v3/__https://petsc.org/release/faq/__;!!G_uCfscf7eWS!a83YNiE6vfZTxyhWxnJxmh95Z63M4HapLLI6b0Q_hFwekNKgjv2j3t61lo38kvy2gLDVjWO5IScFYBVUJaIxcoc$ >>> for trouble shooting. >>> [0]PETSC ERROR: Petsc Release Version 3.22.0, Sep 28, 2024 >>> [0]PETSC ERROR: /Users/praveen/work/bouss/radial_flat/xgeoclaw with 6 MPI >>> process(es) and PETSC_ARCH on MacMiniHome.local by praveen Thu Oct 24 >>> 08:04:27 2024 >>> [0]PETSC ERROR: Configure options: AR=arm64-apple-darwin20.0.0-ar CC=mpicc >>> CXX=mpicxx FC=mpifort CFLAGS="-ftree-vectorize -fPIC >>> -fstack-protector-strong -O2 -pipe -isystem >>> /opt/homebrew/Caskroom/miniforge/base/envs/claw/include " >>> CPPFLAGS="-D_FORTIFY_SOURCE=2 -isystem >>> /opt/homebrew/Caskroom/miniforge/base/envs/claw/include >>> -mmacosx-version-min=11.0 -mmacosx-version-min=11.0" >>> CXXFLAGS="-ftree-vectorize -fPIC -fstack-protector-strong -O2 -pipe >>> -stdlib=libc++ -fvisibility-inlines-hidden -fmessage-length=0 -isystem >>> /opt/homebrew/Caskroom/miniforge/base/envs/claw/include " >>> FFLAGS="-march=armv8.3-a -ftree-vectorize -fPIC -fno-stack-protector -O2 >>> -pipe -isystem /opt/homebrew/Caskroom/miniforge/base/envs/claw/include " >>> LDFLAGS="-Wl,-headerpad_max_install_names -Wl,-dead_strip_dylibs >>> -Wl,-rpath,/opt/homebrew/Caskroom/miniforge/base/envs/claw/lib >>> -L/opt/homebrew/Caskroom/miniforge/base/envs/claw/lib" >>> LIBS="-Wl,-rpath,/opt/homebrew/Caskroom/miniforge/base/envs/claw/lib >>> -lmpi_mpifh -lgfortran" --COPTFLAGS=-O3 --CXXOPTFLAGS=-O3 --FOPTFLAGS=-O3 >>> --with-clib-autodetect=0 --with-cxxlib-autodetect=0 >>> --with-fortranlib-autodetect=0 --with-debugging=0 >>> --with-blas-lib=libblas.dylib --with-lapack-lib=liblapack.dylib >>> --with-yaml=1 --with-hdf5=1 --with-fftw=1 --with-hwloc=0 --with-hypre=1 >>> --with-metis=1 --with-mpi=1 --with-mumps=1 --with-parmetis=1 >>> --with-pthread=1 --with-ptscotch=1 --with-shared-libraries --with-ssl=0 >>> --with-scalapack=1 --with-superlu=1 --with-superlu_dist=1 >>> --with-superlu_dist-include=/opt/homebrew/Caskroom/miniforge/base/envs/claw/include/superlu-dist >>> --with-superlu_dist-lib=-lsuperlu_dist --with-suitesparse=1 >>> --with-suitesparse-dir=/opt/homebrew/Caskroom/miniforge/base/envs/claw >>> --with-x=0 --with-scalar-type=real --with-cuda=0 --with-batch >>> --prefix=/opt/homebrew/Caskroom/miniforge/base/envs/claw >>> [0]PETSC ERROR: #1 PetscShmgetMapAddresses() at >>> /Users/runner/miniforge3/conda-bld/petsc_1728030427805/work/src/sys/utils/server.c:114 >>> [0]PETSC ERROR: #2 PCMPISetMat() at >>> /Users/runner/miniforge3/conda-bld/petsc_1728030427805/work/src/ksp/pc/impls/mpi/pcmpi.c:269 >>> [0]PETSC ERROR: #3 PCSetUp_MPI() at >>> /Users/runner/miniforge3/conda-bld/petsc_1728030427805/work/src/ksp/pc/impls/mpi/pcmpi.c:853 >>> [0]PETSC ERROR: #4 PCSetUp() at >>> /Users/runner/miniforge3/conda-bld/petsc_1728030427805/work/src/ksp/pc/interface/precon.c:1071 >>> [0]PETSC ERROR: #5 KSPSetUp() at >>> /Users/runner/miniforge3/conda-bld/petsc_1728030427805/work/src/ksp/ksp/interface/itfunc.c:415 >>> [0]PETSC ERROR: #6 KSPSolve_Private() at >>> /Users/runner/miniforge3/conda-bld/petsc_1728030427805/work/src/ksp/ksp/interface/itfunc.c:826 >>> [0]PETSC ERROR: #7 KSPSolve() at >>> /Users/runner/miniforge3/conda-bld/petsc_1728030427805/work/src/ksp/ksp/interface/itfunc.c:1075 >>> >>> Code does not progress and I kill it >>> >>> ^CTraceback (most recent call last): >>> File >>> "/Users/praveen/Applications/clawpack/clawutil/src/python/clawutil/runclaw.py", >>> line 341, in <module> >>> runclaw(*args) >>> File >>> "/Users/praveen/Applications/clawpack/clawutil/src/python/clawutil/runclaw.py", >>> line 242, in runclaw >>> proc = subprocess.check_call(cmd_split, >>> ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ >>> File >>> "/opt/homebrew/Caskroom/miniforge/base/envs/claw/lib/python3.12/subprocess.py", >>> line 408, in check_call >>> retcode = call(*popenargs, **kwargs) >>> ^^^^^^^^^^^^^^^^^^^^^^^^^^ >>> File >>> "/opt/homebrew/Caskroom/miniforge/base/envs/claw/lib/python3.12/subprocess.py", >>> line 391, in call >>> return p.wait(timeout=timeout) >>> ^^^^^^^^^^^^^^^^^^^^^^^ >>> File >>> "/opt/homebrew/Caskroom/miniforge/base/envs/claw/lib/python3.12/subprocess.py", >>> line 1264, in wait >>> return self._wait(timeout=timeout) >>> ^^^^^^^^^^^^^^^^^^^^^^^^^^^ >>> File >>> "/opt/homebrew/Caskroom/miniforge/base/envs/claw/lib/python3.12/subprocess.py", >>> line 2053, in _wait >>> (pid, sts) = self._try_wait(0) >>> ^^^^^^^^^^^^^^^^^ >>> File >>> "/opt/homebrew/Caskroom/miniforge/base/envs/claw/lib/python3.12/subprocess.py", >>> line 2011, in _try_wait >>> (pid, sts) = os.waitpid(self.pid, wait_flags) >>> ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ >>> KeyboardInterrupt >>> make[1]: *** [output] Interrupt: 2 >>> make: *** [.output] Interrupt: 2 >>> >>> Now it says >>> >>> $ ipcs -m >>> IPC status from <running system> as of Thu Oct 24 08:05:06 IST 2024 >>> T ID KEY MODE OWNER GROUP >>> Shared Memory: >>> m 720896 0x0000000a --rw-rw-rw- praveen staff >>> >>> Thanks >>> praveen >>> >>>> On 23 Oct 2024, at 8:26 PM, Barry Smith <bsm...@petsc.dev> wrote: >>>> >>>> >>>> Hmm, so it is creating the first shared memory region with ID of 10 (A >>>> in hex) puts it in a linked list in PETSc but then when it tries to find >>>> it in the linked list it cannot find it. >>>> >>>> I don't know how to reproduce this or debug it remotely. >>>> >>>> Can you build on a completely different machine or with completely >>>> different compilers? >>>> >>>> Barry >>>> >>>> >>>> >>>> >>>>> On Oct 23, 2024, at 10:31 AM, Praveen C <cprav...@gmail.com> wrote: >>>>> >>>>> I get same error and now it shows >>>>> >>>>> $ ipcs -m >>>>> >>>>> ------ Shared Memory Segments -------- >>>>> key shmid owner perms bytes nattch status >>>>> >>>>> 0x0000000a 32788 praveen 666 240 6 >>>>> >>>>> Note that the code seems to be still running after printing those error >>>>> message, but it is not printing any progress which it should do. >>>>> >>>>> Thanks >>>>> praveen >>>>> >>>>>> On 23 Oct 2024, at 7:56 PM, Barry Smith <bsm...@petsc.dev> wrote: >>>>>> >>>>>> >>>>>> Try >>>>>> >>>>>> ipcrm -m 11 >>>>>> >>>>>> ipcs -m >>>>>> >>>>>> Try running the program again >>>>>> >>>>>> If failed check >>>>>> >>>>>> ipcs -m >>>>>> >>>>>> again >>>>>> >>>>>> >>>>>> >>>>>>> On Oct 23, 2024, at 10:20 AM, Praveen C <cprav...@gmail.com> wrote: >>>>>>> >>>>>>> Hello Barry >>>>>>> >>>>>>> I see this >>>>>>> >>>>>>> $ ipcs -m >>>>>>> >>>>>>> ------ Shared Memory Segments -------- >>>>>>> key shmid owner perms bytes nattch >>>>>>> status >>>>>>> 0x0000000a 11 praveen 666 240 6 >>>>>>> >>>>>>> and I am observing same error as below. >>>>>>> >>>>>>> Thanks >>>>>>> praveen >>>>>>> >>>>>>>> On 23 Oct 2024, at 7:08 PM, Barry Smith <bsm...@petsc.dev> wrote: >>>>>>>> >>>>>>>> >>>>>>>> Please take a look at the notes in >>>>>>>> https://urldefense.us/v3/__https://petsc.org/release/manualpages/Sys/PetscShmgetAllocateArray/__;!!G_uCfscf7eWS!a83YNiE6vfZTxyhWxnJxmh95Z63M4HapLLI6b0Q_hFwekNKgjv2j3t61lo38kvy2gLDVjWO5IScFYBVU4eqTfOA$ >>>>>>>> For some reason your program is not able to access/use the Unix >>>>>>>> shared memory; check if you are already using the shared memory (so it >>>>>>>> is not available for a new run) or the limits are too low to access >>>>>>>> enough memory. >>>>>>>> >>>>>>>> Barry >>>>>>>> >>>>>>>>> On Oct 23, 2024, at 8:23 AM, Praveen C <cprav...@gmail.com> wrote: >>>>>>>>> >>>>>>>>> Dear all >>>>>>>>> >>>>>>>>> I am not able to run the boussinesq example from geoclaw using >>>>>>>>> petsc@3.22.0 >>>>>>>>> >>>>>>>>> https://urldefense.us/v3/__https://github.com/clawpack/geoclaw/tree/3303883f46572c58130d161986b8a87a57ca7816/examples/bouss__;!!G_uCfscf7eWS!e3VQ4NHKmXGstRsQW5vtI7fmKfUT9zmJkMJcPbcvPyIjicyfJpNoMgx3wZ-qyGcKNSjIkNZkzilec8MnHN6PMw$ >>>>>>>>> >>>>>>>>> >>>>>>>>> It runs with petsc@3.21.6 >>>>>>>>> >>>>>>>>> The error I get is given below. After printing this, the code does >>>>>>>>> not progress. >>>>>>>>> >>>>>>>>> I use the following petsc options >>>>>>>>> >>>>>>>>> # set min numbers of matrix rows per MPI rank (default is 10000) >>>>>>>>> -mpi_linear_solve_minimum_count_per_rank 5000 >>>>>>>>> >>>>>>>>> >>>>>>>>> # Krylov linear solver: >>>>>>>>> -mpi_linear_solver_server >>>>>>>>> -mpi_linear_solver_server_view >>>>>>>>> -ksp_type gmres >>>>>>>>> -ksp_max_it 200 >>>>>>>>> -ksp_reuse_preconditioner >>>>>>>>> -ksp_rtol 1.e-9 >>>>>>>>> >>>>>>>>> # preconditioner: >>>>>>>>> -pc_type gamg >>>>>>>>> >>>>>>>>> I installed petsc and other dependencies for clawpack using miniforge. >>>>>>>>> >>>>>>>>> Thanks >>>>>>>>> pc >>>>>>>>> >>>>>>>>> ==> Use Bouss. in water deeper than 1.0000000000000000 Using >>>>>>>>> a PETSc solver >>>>>>>>> Using Bouss equations from the start >>>>>>>>> rnode allocated... >>>>>>>>> node allocated... >>>>>>>>> listOfGrids allocated... >>>>>>>>> Storage allocated... >>>>>>>>> bndList allocated... >>>>>>>>> Gridding level 1 at t = 0.000000E+00: 4 grids with 10000 >>>>>>>>> cells >>>>>>>>> Setting initial dt to 2.9999999999999999E-002 >>>>>>>>> max threads set to 6 >>>>>>>>> Done reading data, starting computation ... Total zeta at >>>>>>>>> initial time: 39269.907650665169 GEOCLAW: Frame 0 output >>>>>>>>> files done at time t = 0.000000D+00 >>>>>>>>> >>>>>>>>> [0]PETSC ERROR: --------------------- Error Message >>>>>>>>> -------------------------------------------------------------- >>>>>>>>> [0]PETSC ERROR: Petsc has generated inconsistent data >>>>>>>>> [0]PETSC ERROR: Unable to locate PCMPI allocated shared address >>>>>>>>> 0x55e6d750ae20 >>>>>>>>> [0]PETSC ERROR: WARNING! There are unused option(s) set! Could be the >>>>>>>>> program crashed before usage or a spelling mistake, etc! >>>>>>>>> [0]PETSC ERROR: Option left: name:-ksp_max_it value: 200 source: >>>>>>>>> file >>>>>>>>> [0]PETSC ERROR: Option left: name:-ksp_reuse_preconditioner (no >>>>>>>>> value) source: file >>>>>>>>> [0]PETSC ERROR: Option left: name:-ksp_rtol value: 1.e-9 source: >>>>>>>>> file >>>>>>>>> [0]PETSC ERROR: Option left: name:-ksp_type value: gmres source: >>>>>>>>> file >>>>>>>>> [0]PETSC ERROR: Option left: >>>>>>>>> name:-mpi_linear_solve_minimum_count_per_rank value: 5000 source: file >>>>>>>>> [0]PETSC ERROR: Option left: name:-mpi_linear_solver_server_view >>>>>>>>> (no value) source: file >>>>>>>>> [0]PETSC ERROR: Option left: name:-pc_type value: gamg source: file >>>>>>>>> [0]PETSC ERROR: See >>>>>>>>> https://urldefense.us/v3/__https://petsc.org/release/faq/__;!!G_uCfscf7eWS!e3VQ4NHKmXGstRsQW5vtI7fmKfUT9zmJkMJcPbcvPyIjicyfJpNoMgx3wZ-qyGcKNSjIkNZkzilec8MvNjNo7A$ >>>>>>>>> for trouble shooting. >>>>>>>>> [0]PETSC ERROR: Petsc Release Version 3.22.0, Sep 28, 2024 [0]PETSC >>>>>>>>> ERROR: /home/praveen/bouss/radial_flat/xgeoclaw with 6 MPI >>>>>>>>> process(es) and PETSC_ARCH on euler by praveen Thu Oct 17 21:49:54 >>>>>>>>> 2024 >>>>>>>>> [0]PETSC ERROR: Configure options: >>>>>>>>> AR=${PREFIX}/bin/x86_64-conda-linux-gnu-ar CC=mpicc CXX=mpicxx >>>>>>>>> FC=mpifort CFLAGS="-march=nocona -mtune=haswell -ftree-vectorize >>>>>>>>> -fPIC -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe >>>>>>>>> -isystem /opt/miniforge/envs/claw/include " CPPFLAGS="-DNDEBUG >>>>>>>>> -D_FORTIFY_SOURCE=2 -O2 -isystem /opt/miniforge/envs/claw/include" >>>>>>>>> CXXFLAGS="-fvisibility-inlines-hidden -fmessage-length=0 >>>>>>>>> -march=nocona -mtune=haswell -ftree-vectorize -fPIC >>>>>>>>> -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe >>>>>>>>> -isystem /opt/miniforge/envs/claw/include " FFLAGS="-march=nocona >>>>>>>>> -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-strong >>>>>>>>> -fno-plt -O2 -ffunction-sections -pipe -isystem >>>>>>>>> /opt/miniforge/envs/claw/include -Wl,--no-as-needed" >>>>>>>>> LDFLAGS="-pthread -fopenmp -Wl,-O2 -Wl,--sort-common -Wl,--as-needed >>>>>>>>> -Wl,-z,relro -Wl,-z,now -Wl,--disable-new-dtags -Wl,--gc-sections >>>>>>>>> -Wl,--allow-shlib-undefined -Wl,-rpath,/opt/miniforge/envs/claw/lib >>>>>>>>> -Wl,-rpath-link,/opt/miniforge/envs/claw/lib >>>>>>>>> -L/opt/miniforge/envs/claw/lib >>>>>>>>> -Wl,-rpath-link,/opt/miniforge/envs/claw/lib" >>>>>>>>> LIBS="-Wl,-rpath,/opt/miniforge/envs/claw/lib -lmpi_mpifh -lgfortran" >>>>>>>>> --COPTFLAGS=-O3 --CXXOPTFLAGS=-O3 --FOPTFLAGS=-O3 >>>>>>>>> --with-clib-autodetect=0 --with-cxxlib-autodetect=0 >>>>>>>>> --with-fortranlib-autodetect=0 --with-debugging=0 >>>>>>>>> --with-blas-lib=libblas.so --with-lapack-lib=liblapack.so >>>>>>>>> --with-yaml=1 --with-hdf5=1 --with-fftw=1 --with-hwloc=0 >>>>>>>>> --with-hypre=1 --with-metis=1 --with-mpi=1 --with-mumps=1 >>>>>>>>> --with-parmetis=1 --with-pthread=1 --with-ptscotch=1 >>>>>>>>> --with-shared-libraries --with-ssl=0 --with-scalapack=1 >>>>>>>>> --with-superlu=1 --with-superlu_dist=1 >>>>>>>>> --with-superlu_dist-include=/opt/miniforge/envs/claw/include/superlu-dist >>>>>>>>> --with-superlu_dist-lib=-lsuperlu_dist --with-suitesparse=1 >>>>>>>>> --with-suitesparse-dir=/opt/miniforge/envs/claw --with-x=0 >>>>>>>>> --with-scalar-type=real --with-cuda=0 >>>>>>>>> --prefix=/opt/miniforge/envs/claw >>>>>>>>> [0]PETSC ERROR: #1 PetscShmgetMapAddresses() at >>>>>>>>> /home/conda/feedstock_root/build_artifacts/petsc_1728030599661/work/src/sys/utils/server.c:114 >>>>>>>>> [0]PETSC ERROR: #2 PCMPISetMat() at >>>>>>>>> /home/conda/feedstock_root/build_artifacts/petsc_1728030599661/work/src/ksp/pc/impls/mpi/pcmpi.c:269 >>>>>>>>> [0]PETSC ERROR: #3 PCSetUp_MPI() at >>>>>>>>> /home/conda/feedstock_root/build_artifacts/petsc_1728030599661/work/src/ksp/pc/impls/mpi/pcmpi.c:853 >>>>>>>>> [0]PETSC ERROR: #4 PCSetUp() at >>>>>>>>> /home/conda/feedstock_root/build_artifacts/petsc_1728030599661/work/src/ksp/pc/interface/precon.c:1071 >>>>>>>>> [0]PETSC ERROR: #5 KSPSetUp() at >>>>>>>>> /home/conda/feedstock_root/build_artifacts/petsc_1728030599661/work/src/ksp/ksp/interface/itfunc.c:415 >>>>>>>>> [0]PETSC ERROR: #6 KSPSolve_Private() at >>>>>>>>> /home/conda/feedstock_root/build_artifacts/petsc_1728030599661/work/src/ksp/ksp/interface/itfunc.c:826 >>>>>>>>> [0]PETSC ERROR: #7 KSPSolve() at >>>>>>>>> /home/conda/feedstock_root/build_artifacts/petsc_1728030599661/work/src/ksp/ksp/interface/itfunc.c:1075 >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >