LabVIEW GPU Analysis Toolkit API Reference

Table of Contents

xHER2K (Hermitian Matrix Rank-2k Update) VI

  • Updated2023-02-21
  • 11 minute(s) read

xHER2K (Hermitian Matrix Rank-2k Update) VI

Owning Palette: LVCUBLAS

Requires: GPU Analysis Toolkit

Calculates the Hermitian rank-2k update of the matrix expression alpha*A*conj(B') + conj(alpha)*B*conj(A') + beta*C or alpha*conj(A')*B + conj(alpha)*conj(B')*A + beta*C. When you wire data to A in, B in (if available), and C, this VI automatically selects the first available instance.

To use a different instance, you must manually select the polymorphic instance you want to use.

Details  

xHER2K (Hermitian Matrix Rank-2k Update) (aAconj(A') + conj(a)Aconj(A') + bC)

The connector pane displays the default data types for this polymorphic instance.

fill mode specifies the triangular portion of the matrix A in the calculation.

0CUBLAS_FILL_MODE_LOWER (default)
1CUBLAS_FILL_MODE_UPPER
CUBLAS Handle in specifies the initialized CUBLAS library to use for the BLAS calculation. For example, you can wire the CUBLAS Handle output from the Initialize Library VI to specify the CUBLAS handle to the CUBLAS library you already initialized. This input also determines the device that executes the function.
C specifies the Hermitian matrix C stored on the device. The diagonal elements of C have a zero imaginary component. The calculation ignores the diagonal elements. This input specifies a class that can contain the following data types:

  • Complex single-precision, floating-point numeric
  • Complex double-precision, floating-point numeric
A in specifies the matrix A stored on the device. This input specifies a class that can contain the following data types:

  • Complex single-precision, floating-point numeric
  • Complex double-precision, floating-point numeric
a represents alpha and specifies the scalar operand in the product alpha*A*conj(A') and conj(alpha)*A*conj(A'). The default is 1.
b represents beta and specifies the scalar operand in the product beta*C. The default is 0.
error in describes error conditions that occur before this node runs. This input provides standard error in functionality.
n specifies the number of rows to use in matrix A.
k specifies the number of columns to use in matrix A.
leading dimensions specifies the column dimension to index consecutive rows. Use lda, ldb, and ldc for A, A, and C, respectively.
CUBLAS Handle out returns the handle that defines the BLAS operation.
aAconj(A')+conj(a)Aconj(A')+bC returns a matrix with the same dimensions as C. For the elements of the first n rows and n columns of the triangular component defined by fill mode, aAconj(A')+conj(a)Aconj(A')+bC returns the result of the calculation. For any remaining elements, aAconj(A')+conj(a)Aconj(A')+bC returns the value of the element with the same index in C.
A out returns the matrix A stored on the device.
error out contains error information. This output provides standard error out functionality.

xHER2K (Hermitian Matrix Rank-2k Update) (aconj(A')A + conj(a)conj(A')A + bC)

The connector pane displays the default data types for this polymorphic instance.

fill mode specifies the triangular portion of the matrix A in the calculation.

0CUBLAS_FILL_MODE_LOWER (default)
1CUBLAS_FILL_MODE_UPPER
CUBLAS Handle in specifies the initialized CUBLAS library to use for the BLAS calculation. For example, you can wire the CUBLAS Handle output from the Initialize Library VI to specify the CUBLAS handle to the CUBLAS library you already initialized. This input also determines the device that executes the function.
C specifies the Hermitian matrix C stored on the device. The diagonal elements of C have a zero imaginary component. The calculation ignores the diagonal elements. This input specifies a class that can contain the following data types:

  • Complex single-precision, floating-point numeric
  • Complex double-precision, floating-point numeric
A in specifies the matrix A stored on the device. This input specifies a class that can contain the following data types:

  • Complex single-precision, floating-point numeric
  • Complex double-precision, floating-point numeric
a represents alpha and specifies the scalar operand in the product alpha*conj(A')*A and conj(alpha)*conj(A')*A. The default is 1.
b represents beta and specifies the scalar operand in the product beta*C. The default is 0.
error in describes error conditions that occur before this node runs. This input provides standard error in functionality.
n specifies the number of rows to use in matrix conj(A').
k specifies the number of columns to use in matrix conj(A').
leading dimensions specifies the column dimension to index consecutive rows. Use lda, ldb, and ldc for A, A, and C, respectively.
CUBLAS Handle out returns the handle that defines the BLAS operation.
aconj(A')A+conj(a)conj(A')A+bC returns a matrix with the same dimensions as C. For the elements of the first n rows and n columns of the triangular component defined by fill mode, aconj(A')A+conj(a)conj(A')A+bC returns the result of the calculation. For any remaining elements, aconj(A')A+conj(a)conj(A')A+bC returns the value of the element with the same index in C.
A out returns the matrix A stored on the device.
error out contains error information. This output provides standard error out functionality.

xHER2K (Hermitian Matrix Rank-2k Update) (aAconj(B') + conj(a)Bconj(A') + bC)

The connector pane displays the default data types for this polymorphic instance.

fill mode specifies the triangular portion of the matrix A in the calculation.

0CUBLAS_FILL_MODE_LOWER (default)
1CUBLAS_FILL_MODE_UPPER
CUBLAS Handle in specifies the initialized CUBLAS library to use for the BLAS calculation. For example, you can wire the CUBLAS Handle output from the Initialize Library VI to specify the CUBLAS handle to the CUBLAS library you already initialized. This input also determines the device that executes the function.
C specifies the Hermitian matrix C stored on the device. The diagonal elements of C have a zero imaginary component. The calculation ignores the diagonal elements. This input specifies a class that can contain the following data types:

  • Complex single-precision, floating-point numeric
  • Complex double-precision, floating-point numeric
A in specifies the matrix A stored on the device. This input specifies a class that can contain the following data types:

  • Complex single-precision, floating-point numeric
  • Complex double-precision, floating-point numeric
B in specifies the matrix B stored on the device. This input specifies a class that can contain the following data types:

  • Complex single-precision, floating-point numeric
  • Complex double-precision, floating-point numeric
a represents alpha and specifies the scalar operand for the product alpha*A*conj(B)' and conj(alpha)*B*conj(A').
b represents beta and specifies the scalar operand in the product beta*C. The default is 0.
error in describes error conditions that occur before this node runs. This input provides standard error in functionality.
n specifies the number of rows to use in matrix A and conj(B').
k specifies the number of columns to use in matrix A and conj(B').
leading dimensions specifies the column dimension to index consecutive rows. Use lda, ldb, and ldc for A, B, and C, respectively.
CUBLAS Handle out returns the handle that defines the BLAS operation.
aAconj(B')+conj(a)Bconj(A')+bC returns a matrix with the same dimensions as C. For the elements of the first n rows and n columns of the triangular component defined by fill mode, aAconj(B')+conj(a)Bconj(A')+bC returns the result of the calculation. For any remaining elements, aAconj(B')+conj(a)Bconj(A')+bC returns the value of the element with the same index in C.
A out returns the matrix A stored on the device.
B out returns the matrix B stored on the device.
error out contains error information. This output provides standard error out functionality.

xHER2K (Hermitian Matrix Rank-2k Update) (aconj(A')B + conj(a)conj(B')A + bC)

The connector pane displays the default data types for this polymorphic instance.

fill mode specifies the triangular portion of the matrix A in the calculation.

0CUBLAS_FILL_MODE_LOWER (default)
1CUBLAS_FILL_MODE_UPPER
CUBLAS Handle in specifies the initialized CUBLAS library to use for the BLAS calculation. For example, you can wire the CUBLAS Handle output from the Initialize Library VI to specify the CUBLAS handle to the CUBLAS library you already initialized. This input also determines the device that executes the function.
C specifies the Hermitian matrix C stored on the device. The diagonal elements of C have a zero imaginary component. The calculation ignores the diagonal elements. This input specifies a class that can contain the following data types:

  • Complex single-precision, floating-point numeric
  • Complex double-precision, floating-point numeric
A in specifies the matrix A stored on the device. This input specifies a class that can contain the following data types:

  • Complex single-precision, floating-point numeric
  • Complex double-precision, floating-point numeric
B in specifies the matrix B stored on the device. This input specifies a class that can contain the following data types:

  • Complex single-precision, floating-point numeric
  • Complex double-precision, floating-point numeric
a represents alpha and specifies the scalar operand in the product alpha*conj(A')*B and conj(alpha)*conj(B')*A. The default is 1.
b represents beta and specifies the scalar operand in the product beta*C. The default is 0.
error in describes error conditions that occur before this node runs. This input provides standard error in functionality.
n specifies the number of rows to use in matrix conj(A') and B.
k specifies the number of columns to use in matrix A.
leading dimensions specifies the column dimension to index consecutive rows. Use lda, ldb, and ldc for A, B, and C, respectively.
CUBLAS Handle out returns the handle that defines the BLAS operation.
aconj(A')B+conj(a)conj(B')A+bC returns a matrix with the same dimensions as C. For the elements of the first n rows and n columns of the triangular component defined by fill mode, aconj(A')B+conj(a)conj(B')A+bC returns the result of the calculation. For any remaining elements, aconj(A')B+conj(a)conj(B')A+bC returns the value of the element with the same index in C.
A out returns the matrix A stored on the device.
B out returns the matrix B stored on the device.
error out contains error information. This output provides standard error out functionality.

xHER2K (Hermitian Matrix Rank-2k Update) Details

For more information on how to use this VI, refer to the Designing the Block Diagram to Compute on a GPU Device topic.

For more information about the CUBLAS library and BLAS operations, refer to the NVIDIA GPU Computing Documentation website at nvidia.com and download the CUBLAS Library User Guide.

Refer to the BLAS (Basic Linear Algebra Subprograms) website at netlib.org for more information on BLAS functions.

Log in to get a better experience