To that end, I now present a bug I'm getting. First, I built ScaLAPACK and BLACK with:
- Code: Select all
# ./setup.py --blaslib="-L/usr/local/atlas/lib -llapack -lf77blas -lcblas -latlas" --lapacklib="-L/usr/local/atlas/lib -llapack -lf77blas -lcblas -latlas" --downblacs
This showed no problems.
I then went into the build/scalapack-1.8.0/TESTING directory and tried to run test xdsep:
- Code: Select all
# mpirun -np 4 ./xdsep
SCALAPACK symmetric Eigendecomposition routines.
' '
Running tests of the parallel symmetric eigenvalue routine: PDSYEVX & PDSYEV & PDSYEVD.
The following scaled residual checks will be computed:
||AQ - QL|| / ((abstol + ||A|| * eps) * N)
||Q^T*Q - I|| / (N * eps)
An explanation of the input/output parameters follows:
RESULT : passed; or an indication of which eigen request test failed
N : The number of rows and columns of the matrix A.
P : The number of process rows.
Q : The number of process columns.
NB : The size of the square blocks the matrix A is split into.
THRESH : If a residual value is less than THRESH, RESULT is flagged as PASSED.
: the QTQ norm is allowed to exceed THRESH for those eigenvectors
: which could not be reorthogonalized for lack of workspace.
TYP : matrix type (see PDSEPtst.f).
SUB : Subtests (see PDSEPtst).f
CHK : ||AQ - QL|| / ((abstol + ||A|| * eps) * N)
QTQ : ||Q^T*Q - I||/ (N * eps)
: when the adjusted QTQ exceeds THRESH
the adjusted QTQ norm is printed
: otherwise the true QTQ norm is printed
If NT>1, CHK and QTQ are the max over all eigen request tests
TEST : EVX - testing PDSYEVX, EV - testing PDSYEV, EVD - testing PDSYEVD
N NB P Q TYP SUB WALL CPU CHK QTQ CHECK TEST
----- --- --- --- --- --- -------- -------- --------- --------- ----- ----
'TEST 1 - test tiny matrices - different process configurations'
0 1 1 2 8 N 0.00 -1.00 0.0 0.0 PASSED EVX
[oxygen:07445] *** Process received signal ***
[oxygen:07445] Signal: Segmentation fault (11)
[oxygen:07445] Signal code: (128)
[oxygen:07445] Failing at address: (nil)
[oxygen:07445] [ 0] /lib64/libpthread.so.0 [0x3729a0ed30]
[oxygen:07445] [ 1] /usr/local/lib/openmpi/mca_pml_ob1.so [0x2ad6aec778a9]
[oxygen:07445] [ 2] /usr/local/lib/openmpi/mca_btl_sm.so(mca_btl_sm_component_progress+0x6b9) [0x2ad6af28ed99]
[oxygen:07445] [ 3] /usr/local/lib/openmpi/mca_bml_r2.so(mca_bml_r2_progress+0x2b) [0x2ad6aee8305b]
[oxygen:07445] [ 4] /usr/local/lib/libopen-pal.so.0(opal_progress+0x4a) [0x2ad6a96a2b5a]
[oxygen:07445] [ 5] /usr/local/lib/libmpi.so.0(ompi_request_wait_all+0x1cd) [0x2ad6a91d559d]
[oxygen:07445] [ 6] /usr/local/lib/openmpi/mca_coll_tuned.so(ompi_coll_tuned_allreduce_intra_recursivedoubling+0x313) [0x2ad6afcb2f23]
[oxygen:07445] [ 7] /usr/local/lib/libmpi.so.0(ompi_comm_activate+0x90) [0x2ad6a91c49a0]
[oxygen:07445] [ 8] /usr/local/lib/libmpi.so.0(ompi_comm_create+0x174) [0x2ad6a91c4744]
[oxygen:07445] [ 9] /usr/local/lib/libmpi.so.0(MPI_Comm_create+0xc8) [0x2ad6a91ed2b8]
[oxygen:07445] [10] ./xdsep(Cblacs_gridmap+0x160) [0x4c0740]
[oxygen:07445] [11] ./xdsep(SL_Cgridreshape+0x110) [0x41fec0]
[oxygen:07445] [12] ./xdsep(pdlasizesyev_+0x262) [0x4184d2]
[oxygen:07445] [13] ./xdsep(pdsqpsubtst_+0x6e4) [0x418cb4]
[oxygen:07445] [14] ./xdsep(pdseptst_+0x607a) [0x40cf1a]
[oxygen:07445] [15] ./xdsep(pdsepreq_+0x82e) [0x416ade]
[oxygen:07445] [16] ./xdsep(MAIN__+0x15a5) [0x415ef1]
[oxygen:07445] [17] ./xdsep(main+0x2c) [0x4c469c]
[oxygen:07445] [18] /lib64/libc.so.6(__libc_start_main+0xfa) [0x3728e1e32a]
[oxygen:07445] [19] ./xdsep(dsymv_+0x79) [0x406dd9]
[oxygen:07445] *** End of error message ***
mpirun noticed that job rank 0 with PID 7445 on node oxygen.nrl.navy.mil exited on signal 11 (Segmentation fault).
3 additional processes aborted (not shown)
I am now stuck and cannot figure out how to fix this. Any help would be appreciated.

