ARSC T3D Users' Newsletter 49, August 25, 1995

Programming Environment 1.2.2 Now Default

As of 8/23/95, ARSC is running the 1.2.2 Programming Environment as the default. If you have any problems with this release on the T3D contact Mike Ess, This release has many new features and some of them were described in the CRI release letter that appeared in T3D newsletter #47 (8/11/95). Below is an additional description of what's in the 1.2.2 release that appeared in the Cray Research Service Bulletin:

  > Systems:  PVP, MPP
  > OS:       UNICOS 8.0, UNICOS 9.0,
  >           UNICOS MAX or later
  > Product:  CF77 6.2.2 PE,
  >           CF77 6.2.2 PE for MPP systems,
  >           CF90 1.2.2 PE,
  >           CF90 1.2.2 PE for MPP systems,
  >           Cray Standard C 1.2.2 PE,
  >           Cray Standard C 1.2.2 PE for MPP systems
  > Audience: All
  > Date:     July 1995
  > The following Cray Programming Environments were released in June:
  >   CF90 1.2.2 Programming Environment for both Cray PVP and Cray MPP 
  >         systems 
  >   Cray Standard C 1.2.2 Programming Environment for both Cray PVP and
  >         Cray MPP systems 
  >   CF77 6.2.2 Programming Environment for Cray PVP and Cray MPP systems 
  > The programming environments provide an integrated development environment 
  > and combine compilers with development tools and libraries. Each of these 
  > programming environments includes the CrayTools 1.3.2 package, the 
  > CrayLibs 1.2.2 package for either Cray PVP or Cray MPP systems, and one of 
  > the following compilers:
  >   CF90 1.0.3 
  >   CF90_M 0.1.2 
  >   SCC 4.0.3 
  >   SCC_M 4.0.4 
  >   CF77 6.0.4 
  >   CF77_M 6.2.1 
  > You cannot order programming environment components separately; you must 
  > order a complete programming environment. To order a programming environment,
  > contact your Cray Research service representative or complete a software 
  > order form in the CRInform service.
  > Programming environment components discussed in this article are supported on 
  > Cray PVP systems that run the UNICOS 8.0 release or later and on Cray MPP 
  > systems that run both the UNICOS MAX release or later and the 
  > UNICOS 8.0 release or later.
  > Note: You must install the CF77_M 6.2.1 compiling system and CrayLibs_M 1.2.2 
  > at the same time because dependencies exist between these products in the 
  > following shared array intrinsics, which are introduced for the first time in 
  > this release:
  >   PRESUM, PREPROD, PREMIN, and PREMAX parallel prefix operations 
  >   SCANNSUM, SCANPROD, SCANMIN, and SCANMAX segmented scan operations 
  > If programs that use these array intrinsics are compiled with CF77_M 6.2.1, 
  > you also must use CrayLibs_M 1.2.2 or later.
  > Enhancements
  > For both PVP and MPP systems, the libsci LAPACK libraries contain additional 
  > support for the following (LAPACK is a software package for solving linear 
  > systems and matrix eigenvalue problems):
  >   Factorization variant QR with pivoting 
  >   Linear least squares driver routines 
  >   Reductions to condensed form for eigenvalue computations 
  > The libraries also contain performance improvements to xLARTG and xLARFG, the 
  > auxiliary routines that generate elementary transformations.
  > For MPP systems, the programming environment 1.2.2 releases contain the 
  > following enhancements:
  >   Performance in the CF77_M compiling system was enhanced by adding 
  >   optimizations that improve cache performance for stack data. 
  >   Performance was enhanced for all MPP 1.2.2 programming environments 
  >   through optimizations that streamline inner loop computations. 
  >   The SHMEM_MY_PE() library routine, which returns the number of the 
  >   processor elements (PEs) executing, and the SHMEM_N_PES() library 
  >   routine, which returns the number of PEs available to the program, were 
  >   added. These routines are especially useful with the CF90 compiler, 
  >   which does not have intrinsics for this purpose. 
  >   Performance was improved in the SHMEM_GET library routine when transfer 
  >   sizes of one to three words are achieved. 
  >   Performance was improved for the Parallel Virtual Machine (PVM) global 
  >   operation function, pvm_reduce(). Previously, pvm_reduce() was layered 
  >   on the pvm_psend() and pvm_precv() functions, and it used the same 
  >   linear algorithm as is used in the public domain version of PVM. The 
  >   optimized version uses more direct shared memory and achieves more 
  >   logarithmic scaling. 
  >   The following Fortran shared array intrinsics were added: 
  >     PRESUM, PREPROD, PREMIN, and PREMAX parallel prefix operations 
  >     SCANNSUM, SCANPROD, SCANMIN, and SCANMAX segmented scan operations 
  > These intrinsic functions behave similarly to reduction functions, but 
  > partial results are retained in the result array. The segmented scan 
  > intrinsics also carry a mask of stop bits that tells the operation to stop 
  > and start over again.
  > Enhancements to the libsci libraries for Cray MPP systems include the 
  > following:
  >   The addition of an optimized implementation of ScaLAPACK, a set of 
  >   scalable algorithms for solving linear algebra problems. ScaLAPACK 
  >   contains several linear solvers, similarity and equivalence 
  >   reduction routines, and an Eigensystem solver for real symmetric 
  >   matrices. These algorithms use the Basic Linear Algebra Communication
  >   Subprograms (BLACS) for communication and the Parallel Basic Linear 
  >   Algebraic
  >   Subroutines (PBLAS) for computation. The PBLAS, new in this release, 
  >   provide distributed-memory software comparable to the BLAS in 
  >   shared-memory codes. 
  >   New Fast Fourier Transform (FFT) routines were added to perform 
  >   parallel 2-D and 3-D real-to-complex and complex-to-real FFTs. 
  >   The performance of the FFT subroutine, CCFFT, was improved for 
  >   prime-length vectors and for lengths that cannot be factored into 
  >   powers of 2, 3, and 5. The single-PE routines
  >   SCFFT and CSFFT have been extended to work with odd-length vectors. 
  >   Newly optimized versions of the following routines were added: 
  >   SNRM2, SCNRM2, SROTG, and CROTG from the Level 1 BLAS routines and 
  >   STRSM, STRMM, and the other Level 3 BLAS routines. 
  > Compatibilities issues
  > The user interface for loop unrolling in the CF77_M compiling system has 
  > changed. In previous releases, when a user specified the following, a total 
  > of 5 copies of the loop body were created; the original loop plus 4:
  > CDIR$ UNROLL [4]
  > This implementation was not compatible with the FPP implementation of the 
  > directive and was confusing to users. With this release, the previous example 
  > results in a total of 4 copies of the loop body; the original plus 3.
  > Documentation changes
  > In addition to numerous man pages updated for this release, the following 
  > manuals have also been updated:
  >   Cray MPP Fortran Reference Manual, publication SR-2504 6.2.2 (available 
  >   online in Docview form only) 
  >   Programming Environment Installation Guide, publication SG-5217 1.2.2 
  >   (available in printed form only) 

One Year Anniversary of the ARSC T3D Newsletter

The first ARSC T3D Newsletter was mailed out on 8/25/94 and since then there have been 49 issues (almost weekly). I thought about providing an index but that would be a lot of work and would be incomplete no matter how much effort was put into it.

What has been useful for me is to accumulate all of the newsletters in a single directory (~.5MB) and when a query comes up I just grep for the keywords. With grep there is no need for an index.

PATP Meeting at JPL

This week Ming Jiang and I will be at the PATP Scientific Conference held at JPL, August 24th and 25th. I'll have details of that meeting in future newsletters.

List of Differences Between T3D and Y-MP

The current list of differences between the T3D and the Y-MP is:
  1. Data type sizes are not the same (Newsletter #5)
  2. Uninitialized variables are different (Newsletter #6)
  3. The effect of the -a static compiler switch (Newsletter #7)
  4. There is no GETENV on the T3D (Newsletter #8)
  5. Missing routine SMACH on T3D (Newsletter #9)
  6. Different Arithmetics (Newsletter #9)
  7. Different clock granularities for gettimeofday (Newsletter #11)
  8. Restrictions on record length for direct I/O files (Newsletter #19)
  9. Implied DO loop is not "vectorized" on the T3D (Newsletter #20)
  10. Missing Linpack and Eispack routines in libsci (Newsletter #25)
  11. F90 manual for Y-MP, no manual for T3D (Newsletter #31)
  12. RANF() and its manpage differ between machines (Newsletter #37)
  13. CRAY2IEG is available only on the Y-MP (Newsletter #40)
  14. Missing sort routines on the T3D (Newsletter #41)
I encourage users to e-mail in differences that they have found, so we all can benefit from each other's experience.
Current Editors:
Ed Kornkven ARSC HPC Specialist ph: 907-450-8669
Kate Hedstrom ARSC Oceanographic Specialist ph: 907-450-8678
Arctic Region Supercomputing Center
University of Alaska Fairbanks
PO Box 756020
Fairbanks AK 99775-6020
E-mail Subscriptions: Archives:
    Back issues of the ASCII e-mail edition of the ARSC T3D/T3E/HPC Users' Newsletter are available by request. Please contact the editors.
Back to Top