Selecting one of the following will take you directly to that section:
Use C99 language features.
Chooses generally optimal flags for a processor that supports SSE capabillity.
Disable C++ exception handling support.
Disable C++ run time type information support.
Align "unconstrained" data objects of size greater than or equal to 16 bytes on cache-line boundaries. An "unconstrained" object is a variable or array that is not a member of an aggregate structure or common block, is not allocatable, and is not an automatic array. On by default on 64-bit Linux systems.
Set SSE to flush-to-zero mode; if a floating-point underflow occurs, the value is set to zero.
Treat denormalized numbers as zero. Included with "-fastsse" on Intel based systems. For AMD based systems, "-Mdaz" is not included by default with "-fastsse".
Generate code to set up a stack frame.
Eliminates operations that set up a true stack frame pointer for every function. With this option enabled, you cannot perform a traceback on the generated code and you cannot access local variables.
Instructs the compiler to use relaxed precision in the calculation of reciprocal square root (1/sqrt). Can result in improved performance at the expense of numerical accuracy.
Instructs the compiler to use relaxed precision in the calculation of square root. Can result in improved performance at the expense of numerical accuracy.
Instructs the compiler to use relaxed precision in the calculation of divides. Can result in improved performance at the expense of numerical accuracy.
Instructs the compiler to use relaxed precision in the calculation of some intrinsic functions. Can result in improved performance at the expense of numerical accuracy. The default on an AMD system is "-Mfprelaxed=rsqrt". The default on an Intel system is "-Mfprelaxed=rsqrt,sqrt,div"
Set the fetch-ahead distance for prefetch instructions to m cache lines
Set maximum number of prefetch instructions to generate for a given loop to p.
Use the prefetchnta instruction.
Use the prefetch instruction.
Use the prefetcht0 instruction.
Use the AMD-specific prefetchw instruction.
Enable generation of prefetch instructions on processors where they are supported.
Disable generation of prefetch instructions.
Use SSE/SSE2 instructions to perform scalar floating-point arithmetic on targets where these instructions are supported.
Do not use SSE/SSE2 instructions to perform scalar floating-point arithmetic; use x87 operations instead.
Instructs the compiler to extend the sign bit that is set as a result of an object's conversion from one data type to an object of a larger signed data type.
Treat individual array element references as candidates for possible loop-carried redundancy elimination. The default is to eliminate only redundant expressions involving two or more operands.
Allow expression re-association; specifying this sub-option can increase opportunities for loop-carried redundancy elimination.
Disable expression re-association.
Enables loop-carried redundancy elimination, an optimization that can reduce the number of arithmetic operations and memory references in loops.
Disable loop-carried redundancy elimination.
Instructs the compiler not to perform idiom recognition or introduce calls to hand-optimized vector functions.
Generate profile-feedback instrumentation (PFI); this includes extra code to collect run-time statistics and dump them to a trace file for use in a subsequent compilation. PFI gathers information about a program's execution and data values but does not gather information from hardware performance counters. PFI does gather data for optimizations which are unique to profile-feedback optimization.
Enable profile-feedback optimizations.
Interprocedural Analysis option: Recognize when targets of pointer dummy are aligned.
Interprocedural Analysis option: Disable recognizition when targets of pointer dummy are aligned.
Interprocedural Analysis option: Remove arguments replaced by -Mipa=ptr,const
Interprocedural Analysis option: Do not remove arguments replaced by -Mipa=ptr,const
Interprocedural Analysis option: Generate call graph information for pgicg tool.
Interprocedural Analysis option: Do not generate call graph information for pgicg tool.
Interprocedural Analysis option: Enable interprocedural constant propagation.
Interprocedural Analysis option: Disable interprocedural constant propagation.
Interprocedural Analysis option: Used with -Mipa=inline to specify functions which should not be inlined.
Instructs the compiler to perform interprocedural analysis. Equivalant to -Mipa=align,arg,const,f90ptr,shape,globals,libc,localarg,ptr,pure.
Interprocedural Analysis option: Force all objects to recompile regardless whether IPA information has changed.
Interprocedural Analysis option: Optimize references to global values.
Interprocedural Analysis option: Do not optimize references to global values.
Interprocedural Analysis option: Automatically determine which functions to inline, limit to n levels. IPA-based function inlining is performed from leaf routines upward.
Interprocedural Analysis option: Automatically determine which functions to inline. IPA-based function inlining is performed from leaf routines upward.
Interprocedural Analysis option: Allow inlining of routines from libraries.
Interprocedural Analysis option: Do not inline routines from libraries.
Interprocedural Analysis option: Used to optimize calls to certain functions in the system standard C library, libc.
Interprocedural Analysis option: Allow recompiling and optimization of routines from libraries using IPA information.
Interprocedural Analysis option: Don't optimize routines in libraries.
Interprocedural Analysis option: -Mipa=arg plus externalizes local pointer targets.
Interprocedural Analysis option: -Mipa=arg plus externalizes local pointer targets.
Interprocedural Analysis option: Do not externalize local pointer targets.
Interprocedural Analysis option: Enable pointer disambiguation across procedure calls.
Interprocedural Analysis option: Disable pointer disambiguation.
Interprocedural Analysis option: Fortran 90/95 Pointer disambiguation across calls.
Interprocedural Analysis option: Disable Fortran 90/95 pointer disambiguation
Interprocedural Analysis option: Pure function detection.
Interprocedural Analysis option: Disable pure function detection.
Interprocedural Analysis option: Perform Fortran 90 array shape propagation.
Interprocedural Analysis option: Disable Fortran 90 array shape propagation.
Interprocedural Analysis option: Remove functions that are never called.
Interprocedural Analysis option: Do not remove functions that are never called.
Enable Interprocedural Analysis.
Instructs the parallelizer to generate alternate serial code for parallelized loops. Without arguments, the parallelizer determines an appropriate cutoff length and generates serial code to be executed whenever the loop count is less than or equal to that length.
Instructs the parallelizer to generate alternate serial code for parallelized loops. With arguments, the serial altcode is executed whenever the loop count is less than or equal to n.
Always execute the parallelized version of a loop regardless of the loop count.
Disables parallelization of loops with reductions.
Assume loops containing calls are safe to parallelize and allows loops containing calls to be candidates for parallelization. Also, no minimum loop count threshold must be satisfied before parallelization will occur, and last values of scalars are assumed to be safe.
Do not assume loops containing calls are safe to parallelize.
Parallelize with block distribution. Contiguous blocks of iterations of a parallelizable loop are assigned to the available processors.
Parallelize with cyclic distribution. The outermost parallelizable loop in any loop nest is parallelized. If a parallelized loop is innermost, its iterations are allocated to processors cyclically. For example, if there are 3 processors executing a loop, processor 0 performs iterations 0, 3, 6, etc.; processor 1 performs iterations 1, 4, 7, etc.; and processor 2 performs iterations 2, 5, 8, etc.
Enable parallelization of innermost loops.
Disable parallelization of innermost loops.
Instructs the compiler to enable auto-concurrentization of loops. If -Mconcur is specified, multiple processors will be used to execute loops that the compiler determines to be parallelizable.
Instructs the inliner to inline the functions within the library filename.ext.
Instructs the inliner to inline all eligible functions except func, a function in the source text. Multiple functions can be listed, comma-separated.
Instructs the inliner to inline function func.
Instructs the inliner to inline functions with n or fewer statements.
Instructs the inliner to perform n levels of inlining.
Instructs the inliner to perform 1 level of inlining.
Adds a call to the routine "mallopt" in the main routine. This option can have a dramatic impact on the performance of programs that dynamically allocate memory, especially for those which have a few large mallocs. To be effective, this switch must be specified when compiling the file containing the Fortran, C, or C++ main routine. This is currently only effective on 64-bit Linux systems.
Assume all pointers and arrays are independent and safe for aggressive optimizations, and in particular that no pointers or arrays overlap of conflict with each other.
Instructs the compiler that arrays and pointers are treated with the same copyin and copyout semantics as Fortran dummy arguments.
Instructs the compiler that local pointers and arrays do not overlap or conflict with each other and are independent.
Instructs the compiler that local pointers and arrays do not overlap or conflict with each other and are independent.
Instructs the compiler that static pointers and arrays do not overlap or conflict with each other and are independent.
Instructs the compiler that global or external pointers and arrays do not overlap or conflict with each other and are independent.
Instructs the C/C++ compiler to override data dependencies between pointers of a given storage class.
Instructs the compiler to completely unroll loops with a constant loop count of less than or equal to m.
Instructs the compiler to unroll u times, a loop that is not completely unrolled, or has a non-constant loop count.
Invokes the loop unroller.
Disable loop unrolling.
Enable an optional post-pass instruction scheduling.
Disable an optional post-pass instruction scheduling.
Disable automatic vector pipelining.
Instructs the vectorizer to generate alternate code for vectorized loops when appropriate. For each vectorized loop the compiler decides whether to generate altcode and what type or types to generate, which may be any or all of:
The compiler also determines suitable loop count and array alignment conditions for executing the altcode.
Disables alternate code generation for vectorized loops.
Instructs the vectorizer to enable certain associativity conversions that can change the results of a computations due to roundoff error. A typical optimization is to change an arithmetic operation to an arithmetic opteration that is mathmatically correct, but can be computationally different, due to round-off error.
Instructs the vectorizer to disable associativity conversions.
Instructs the vectorizer, when performing cache tiling optimizations, to assume a cache size of n. The default size is n=262144.
Instructs the vectorizer to enable loop fusion.
Instructs the vectorizer to enable idiom recognition.
Generate vector loops for all loops where possible regardless of the number of statements in the loop. This overrides a heuristic in the vectorizer that ordinarily prevents vectorization of loops with a number of statements that exceed a certain threshold.
Instructs the vectorizer to generate prefetch instructions.
Instructs the vectorizer to search for vectorizable loops and, where possible, make use of SSE, SSE2, and prefetch instructions.
Enable automatic vector pipelining.
Disables -Ktrap=fp.
-Ktrap is only processed by the compilers when compiling main functions' programs. The options inv, denorm, divz, ovf, unf, and inexact correspond to the processor's exception mask bits invalid operation, denormalized operand, divide-by-zero, overflow, underflow, and precision, respectively. Normally, the processor's exception mask bits are on (floating-point exceptions are masked the processor recovers from the exceptions and continues). If a floating-point exception occurs and its corresponding mask bit is off (or unmasked ), execution terminates with an arithmetic exception (C's SIGFPE signal). -Ktrap=fp is equivalent to -Ktrap=inv,divz,ovf.
Enable long branches.
Link with the AMD Core Math Library. Available from www.amd.com
Use the -mp option to instruct the compiler to interpret user-inserted OpenMP shared-memory parallel programming directives and generate an executable file which will utilize multiple processors in a shared-memory parallel system. When used strictly as a linker flag, the PGI OpenMP runtime will be linked and users can use the environment variables MP_BIND and MP_BLIST to bind a serial program to a CPU.
The align sub-option to -mp forces loop iterations to be allocated to OpenMP processes using an algorithm that maximizes alignment of vector sub-sections in loops that are both parallelized and vectorized for SSE. This can improve performance in program units that include many such loops. It can result in load-balancing problems that significantly decrease performance in program units with relatively short loops that contain a large amount of work in each iteration.
The nonuma suboption to -mp tells the driver to not link with libnuma.
(For use only on 64-bit Linux targets) Generate code for the medium memory model in the linux86-64 execution environment. The default small memory model of the linux86-64 environment limits the combined area for a user's object or executable to 1GB, with the Linux kernel managing usage of the second 1GB of address for system routines, shared libraries, stacks, etc. Programs are started at a fixed address, and the program can use a single instruction to make most memory references. The medium memory model allows for larger than 2GB data areas, or .bss sections. Program units compiled using either -mcmodel=medium or -fpic require additional instructions to reference memory. The effect on performance is a function of the data-use of the application. The -mcmodel=medium switch must be used at both compile time and link time to create 64-bit executables. Program units compiled for the default small memory model can be linked into medium memory model executables as long as they are compiled -fpic, or position-independent.
Enable support for 64-bit indexing and single static data objects larger than 2GB in size. This option is default in the presence of -mcmodel=medium. Can be used separately together with the default small memory model for certain 64-bit applications that manage their own memory space.
A basic block is generated for each C statement. No scheduling is done between statements. No global optimizations are performed.
Level-one optimization specifies local optimization (-O1). The compiler performs scheduling of basic blocks as well as register allocation. This optimization level is a good choice when the code is very irregular; that is it contains many short statements containing IF statements and the program does not contain loops (DO or DO WHILE statements). For certain types of code, this optimization level may perform better than level-two (-O2) although this case rarely occurs.
The PGI compilers perform many different types of local optimizations, including but not limited to:
Level-two optimization (-O2 or -O) specifies global optimization. The -fast option generally will specify global optimization; however, the -fast switch will vary from release to release depending on a reasonable selection of switches for any one particular release. The -O or -O2 level performs all level-one local optimizations as well as global optimizations. Control flow analysis is applied and global registers are allocated for all functions and subroutines. Loop regions are given special consideration. This optimization level is a good choice when the program contains loops, the loops are short, and the structure of the code is regular.
The PGI compilers perform many different types of global optimizations, including but not limited to:
Specify the type of the target processor as AMD64 Processor 32-bit mode.
Specify the type of the target processor as AMD64 Processor 64-bit mode.
Specify the type of the target processor as Intel P7 Architecture with EM64t, 64-bit mode.
Specify the type of the target processor as Intel P7 Architecture (Pentium 4, Xeon, Centrino).
Specify the type of the target processor as Intel Core 2 EM64T or compatible architecture using 64-bit mode.
Specify the type of the target processor as Intel Core 2 or compatible architecture using 32-bit mode.
Use the unified AMD/Intel 64-bit mode.
Don't include Fortran main program object module.