# xHEMM (Hermitian Matrix-Matrix Product) VI

LabVIEW 2012 GPU Analysis Toolkit Help

Edition Date: June 2012

Part Number: 373575A-01

»View Product Info Download Help (Windows Only)

Owning Palette: LVCUBLAS

Requires: GPU Analysis Toolkit

Calculates the symmetric matrix expression alpha*A*B + beta*C for symmetric matrix A, for rectangular matrices B (if applicable) and C, and for scalars alpha and beta. When you wire data to A in, B in (if available), and C, this VI automatically selects the first available instance.

To use a different instance, you must manually select the polymorphic instance you want to use.

Use the pull-down menu to select an instance of this VI.

 Select an instance xHEMM (Hermitian Matrix-Matrix Product) (aAA + bC)xHEMM (Hermitian Matrix-Matrix Product) (aAB + bC)xHEMM (Hermitian Matrix-Matrix Product) (aBA + bC)

## xHEMM (Hermitian Matrix-Matrix Product) (aAA + bC)

The connector pane displays the default data types for this polymorphic instance.  fill mode specifies the triangular portion of the matrix A in the calculation.

 0 CUBLAS_FILL_MODE_LOWER (default) 1 CUBLAS_FILL_MODE_UPPER CUBLAS Handle in specifies the initialized CUBLAS library to use for the BLAS calculation. For example, you can wire the CUBLAS Handle output from the Initialize Library VI to specify the CUBLAS handle to the CUBLAS library you already initialized. This input also determines the device that executes the function. C specifies the Hermitian matrix C stored on the device. The diagonal elements of C have a zero imaginary component. The calculation ignores the diagonal elements. This input specifies a class that can contain the following data types:

• Complex single-precision, floating-point numeric
• Complex double-precision, floating-point numeric A in specifies the matrix A stored on the device. This input specifies a class that can contain the following data types:

• Complex single-precision, floating-point numeric
• Complex double-precision, floating-point numeric a represents alpha and specifies the scalar operand in the product alpha*A*B. The default is 1. b represents beta and specifies the scalar operand in the product beta*C. The default is 0. error in describes error conditions that occur before this node runs. This input provides standard error in functionality. m specifies the number of rows to use in A in and C. n specifies the number of columns to use in A in and C. leading dimensions specifies the column dimension to index consecutive rows. Use lda, ldb, and ldc for A, A, and C, respectively. CUBLAS Handle out returns the handle that defines the BLAS operation. aAA+bC returns a matrix with the same dimensions as C. For the elements of the first m rows and m columns of the triangular component you select, aAA+bC returns the result of the calculation. For any remaining elements, aAA+bC returns the value of the element with the same index in C. A out returns the symmetric matrix A stored on the device. error out contains error information. This output provides standard error out functionality.

## xHEMM (Hermitian Matrix-Matrix Product) (aAB + bC)

The connector pane displays the default data types for this polymorphic instance.  fill mode specifies the triangular portion of the matrix A in the calculation.

 0 CUBLAS_FILL_MODE_LOWER (default) 1 CUBLAS_FILL_MODE_UPPER CUBLAS Handle in specifies the initialized CUBLAS library to use for the BLAS calculation. For example, you can wire the CUBLAS Handle output from the Initialize Library VI to specify the CUBLAS handle to the CUBLAS library you already initialized. This input also determines the device that executes the function. C specifies the Hermitian matrix C stored on the device. The diagonal elements of C have a zero imaginary component. The calculation ignores the diagonal elements. This input specifies a class that can contain the following data types:

• Complex single-precision, floating-point numeric
• Complex double-precision, floating-point numeric A in specifies the matrix A stored on the device. This input specifies a class that can contain the following data types:

• Complex single-precision, floating-point numeric
• Complex double-precision, floating-point numeric B in specifies the matrix B stored on the device. This input specifies a class that can contain the following data types:

• Complex single-precision, floating-point numeric
• Complex double-precision, floating-point numeric a represents alpha and specifies the scalar operand in the product alpha*A*B. The default is 1. b represents beta and specifies the scalar operand in the product beta*C. The default is 0. error in describes error conditions that occur before this node runs. This input provides standard error in functionality. m specifies the number of rows to use in A in, B in, and C. n specifies the number of columns to use in B in and C. leading dimensions specifies the column dimension to index consecutive rows. Use lda, ldb, and ldc for A, B, and C, respectively. CUBLAS Handle out returns the handle that defines the BLAS operation. aAB+bC returns a matrix with the same dimensions as C. For the elements of the first m rows and n columns of the triangular component you select, aAB+bC returns the result of the calculation. For any remaining elements, aAB+bC returns the value of the element with the same index in C. A out returns the symmetric matrix A stored on the device. B out returns the matrix B stored on the device. error out contains error information. This output provides standard error out functionality.

## xHEMM (Hermitian Matrix-Matrix Product) (aBA + bC)

The connector pane displays the default data types for this polymorphic instance.  fill mode specifies the triangular portion of the matrix A in the calculation.

 0 CUBLAS_FILL_MODE_LOWER (default) 1 CUBLAS_FILL_MODE_UPPER CUBLAS Handle in specifies the initialized CUBLAS library to use for the BLAS calculation. For example, you can wire the CUBLAS Handle output from the Initialize Library VI to specify the CUBLAS handle to the CUBLAS library you already initialized. This input also determines the device that executes the function. C specifies the Hermitian matrix C stored on the device. The diagonal elements of C have a zero imaginary component. The calculation ignores the diagonal elements. This input specifies a class that can contain the following data types:

• Complex single-precision, floating-point numeric
• Complex double-precision, floating-point numeric A in specifies the matrix A stored on the device. This input specifies a class that can contain the following data types:

• Complex single-precision, floating-point numeric
• Complex double-precision, floating-point numeric B in specifies the matrix B stored on the device. This input specifies a class that can contain the following data types:

• Complex single-precision, floating-point numeric
• Complex double-precision, floating-point numeric a represents alpha and specifies the scalar operand in the product alpha*A*B. The default is 1. b represents beta and specifies the scalar operand in the product beta*C. The default is 0. error in describes error conditions that occur before this node runs. This input provides standard error in functionality. m specifies the number of rows to use in A in, B in, and C. n specifies the number of columns to use in B in and C. leading dimensions specifies the column dimension to index consecutive rows. Use lda, ldb, and ldc for A, B, and C, respectively. CUBLAS Handle out returns the handle that defines the BLAS operation. aBA+bC returns a matrix with the same dimensions as C. For the elements of the first m rows and n columns of the triangular component you select, aBA+bC returns the result of the calculation. For any remaining elements, aBA+bC returns the value of the element with the same index in C. A out returns the symmetric matrix A stored on the device. B out returns the matrix B stored on the device. error out contains error information. This output provides standard error out functionality.

## xHEMM (Hermitian Matrix-Matrix Product) Details

For more information on how to use this VI, refer to the Designing the Block Diagram to Compute on a GPU Device topic.

For more information about the CUBLAS library and BLAS operations, refer to the NVIDIA GPU Computing Documentation website at nvidia.com and download the CUBLAS Library User Guide.

Refer to the BLAS (Basic Linear Algebra Subprograms) website at netlib.org for more information on BLAS functions.