/* ----------------------------------------------------------------------
|
* Project: CMSIS DSP Library
|
* Title: arm_cmplx_mat_mult_q15.c
|
* Description: Q15 complex matrix multiplication
|
*
|
* $Date: 27. January 2017
|
* $Revision: V.1.5.1
|
*
|
* Target Processor: Cortex-M cores
|
* -------------------------------------------------------------------- */
|
/*
|
* Copyright (C) 2010-2017 ARM Limited or its affiliates. All rights reserved.
|
*
|
* SPDX-License-Identifier: Apache-2.0
|
*
|
* Licensed under the Apache License, Version 2.0 (the License); you may
|
* not use this file except in compliance with the License.
|
* You may obtain a copy of the License at
|
*
|
* www.apache.org/licenses/LICENSE-2.0
|
*
|
* Unless required by applicable law or agreed to in writing, software
|
* distributed under the License is distributed on an AS IS BASIS, WITHOUT
|
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
* See the License for the specific language governing permissions and
|
* limitations under the License.
|
*/
|
|
#include "arm_math.h"
|
|
/**
|
* @ingroup groupMatrix
|
*/
|
|
/**
|
* @addtogroup CmplxMatrixMult
|
* @{
|
*/
|
|
|
/**
|
* @brief Q15 Complex matrix multiplication
|
* @param[in] *pSrcA points to the first input complex matrix structure
|
* @param[in] *pSrcB points to the second input complex matrix structure
|
* @param[out] *pDst points to output complex matrix structure
|
* @param[in] *pScratch points to the array for storing intermediate results
|
* @return The function returns either
|
* <code>ARM_MATH_SIZE_MISMATCH</code> or <code>ARM_MATH_SUCCESS</code> based on the outcome of size checking.
|
*
|
* \par Conditions for optimum performance
|
* Input, output and state buffers should be aligned by 32-bit
|
*
|
* \par Restrictions
|
* If the silicon does not support unaligned memory access enable the macro UNALIGNED_SUPPORT_DISABLE
|
* In this case input, output, scratch buffers should be aligned by 32-bit
|
*
|
* @details
|
* <b>Scaling and Overflow Behavior:</b>
|
*
|
* \par
|
* The function is implemented using a 64-bit internal accumulator. The inputs to the
|
* multiplications are in 1.15 format and multiplications yield a 2.30 result.
|
* The 2.30 intermediate
|
* results are accumulated in a 64-bit accumulator in 34.30 format. This approach
|
* provides 33 guard bits and there is no risk of overflow. The 34.30 result is then
|
* truncated to 34.15 format by discarding the low 15 bits and then saturated to
|
* 1.15 format.
|
*
|
* \par
|
* Refer to <code>arm_mat_mult_fast_q15()</code> for a faster but less precise version of this function.
|
*
|
*/
|
|
|
|
|
arm_status arm_mat_cmplx_mult_q15(
|
const arm_matrix_instance_q15 * pSrcA,
|
const arm_matrix_instance_q15 * pSrcB,
|
arm_matrix_instance_q15 * pDst,
|
q15_t * pScratch)
|
{
|
/* accumulator */
|
q15_t *pSrcBT = pScratch; /* input data matrix pointer for transpose */
|
q15_t *pInA = pSrcA->pData; /* input data matrix pointer A of Q15 type */
|
q15_t *pInB = pSrcB->pData; /* input data matrix pointer B of Q15 type */
|
q15_t *px; /* Temporary output data matrix pointer */
|
uint16_t numRowsA = pSrcA->numRows; /* number of rows of input matrix A */
|
uint16_t numColsB = pSrcB->numCols; /* number of columns of input matrix B */
|
uint16_t numColsA = pSrcA->numCols; /* number of columns of input matrix A */
|
uint16_t numRowsB = pSrcB->numRows; /* number of rows of input matrix A */
|
uint16_t col, i = 0U, row = numRowsB, colCnt; /* loop counters */
|
arm_status status; /* status of matrix multiplication */
|
q63_t sumReal, sumImag;
|
|
#ifdef UNALIGNED_SUPPORT_DISABLE
|
q15_t in; /* Temporary variable to hold the input value */
|
q15_t a, b, c, d;
|
#else
|
q31_t in; /* Temporary variable to hold the input value */
|
q31_t prod1, prod2;
|
q31_t pSourceA, pSourceB;
|
#endif
|
|
#ifdef ARM_MATH_MATRIX_CHECK
|
/* Check for matrix mismatch condition */
|
if ((pSrcA->numCols != pSrcB->numRows) ||
|
(pSrcA->numRows != pDst->numRows) || (pSrcB->numCols != pDst->numCols))
|
{
|
/* Set status as ARM_MATH_SIZE_MISMATCH */
|
status = ARM_MATH_SIZE_MISMATCH;
|
}
|
else
|
#endif
|
{
|
/* Matrix transpose */
|
do
|
{
|
/* Apply loop unrolling and exchange the columns with row elements */
|
col = numColsB >> 2;
|
|
/* The pointer px is set to starting address of the column being processed */
|
px = pSrcBT + i;
|
|
/* First part of the processing with loop unrolling. Compute 4 outputs at a time.
|
** a second loop below computes the remaining 1 to 3 samples. */
|
while (col > 0U)
|
{
|
#ifdef UNALIGNED_SUPPORT_DISABLE
|
/* Read two elements from the row */
|
in = *pInB++;
|
*px = in;
|
in = *pInB++;
|
px[1] = in;
|
|
/* Update the pointer px to point to the next row of the transposed matrix */
|
px += numRowsB * 2;
|
|
/* Read two elements from the row */
|
in = *pInB++;
|
*px = in;
|
in = *pInB++;
|
px[1] = in;
|
|
/* Update the pointer px to point to the next row of the transposed matrix */
|
px += numRowsB * 2;
|
|
/* Read two elements from the row */
|
in = *pInB++;
|
*px = in;
|
in = *pInB++;
|
px[1] = in;
|
|
/* Update the pointer px to point to the next row of the transposed matrix */
|
px += numRowsB * 2;
|
|
/* Read two elements from the row */
|
in = *pInB++;
|
*px = in;
|
in = *pInB++;
|
px[1] = in;
|
|
/* Update the pointer px to point to the next row of the transposed matrix */
|
px += numRowsB * 2;
|
|
/* Decrement the column loop counter */
|
col--;
|
}
|
|
/* If the columns of pSrcB is not a multiple of 4, compute any remaining output samples here.
|
** No loop unrolling is used. */
|
col = numColsB % 0x4U;
|
|
while (col > 0U)
|
{
|
/* Read two elements from the row */
|
in = *pInB++;
|
*px = in;
|
in = *pInB++;
|
px[1] = in;
|
#else
|
|
/* Read two elements from the row */
|
in = *__SIMD32(pInB)++;
|
|
*__SIMD32(px) = in;
|
|
/* Update the pointer px to point to the next row of the transposed matrix */
|
px += numRowsB * 2;
|
|
|
/* Read two elements from the row */
|
in = *__SIMD32(pInB)++;
|
|
*__SIMD32(px) = in;
|
|
/* Update the pointer px to point to the next row of the transposed matrix */
|
px += numRowsB * 2;
|
|
/* Read two elements from the row */
|
in = *__SIMD32(pInB)++;
|
|
*__SIMD32(px) = in;
|
|
/* Update the pointer px to point to the next row of the transposed matrix */
|
px += numRowsB * 2;
|
|
/* Read two elements from the row */
|
in = *__SIMD32(pInB)++;
|
|
*__SIMD32(px) = in;
|
|
/* Update the pointer px to point to the next row of the transposed matrix */
|
px += numRowsB * 2;
|
|
/* Decrement the column loop counter */
|
col--;
|
}
|
|
/* If the columns of pSrcB is not a multiple of 4, compute any remaining output samples here.
|
** No loop unrolling is used. */
|
col = numColsB % 0x4U;
|
|
while (col > 0U)
|
{
|
/* Read two elements from the row */
|
in = *__SIMD32(pInB)++;
|
|
*__SIMD32(px) = in;
|
#endif
|
|
/* Update the pointer px to point to the next row of the transposed matrix */
|
px += numRowsB * 2;
|
|
/* Decrement the column loop counter */
|
col--;
|
}
|
|
i = i + 2U;
|
|
/* Decrement the row loop counter */
|
row--;
|
|
} while (row > 0U);
|
|
/* Reset the variables for the usage in the following multiplication process */
|
row = numRowsA;
|
i = 0U;
|
px = pDst->pData;
|
|
/* The following loop performs the dot-product of each row in pSrcA with each column in pSrcB */
|
/* row loop */
|
do
|
{
|
/* For every row wise process, the column loop counter is to be initiated */
|
col = numColsB;
|
|
/* For every row wise process, the pIn2 pointer is set
|
** to the starting address of the transposed pSrcB data */
|
pInB = pSrcBT;
|
|
/* column loop */
|
do
|
{
|
/* Set the variable sum, that acts as accumulator, to zero */
|
sumReal = 0;
|
sumImag = 0;
|
|
/* Apply loop unrolling and compute 2 MACs simultaneously. */
|
colCnt = numColsA >> 1;
|
|
/* Initiate the pointer pIn1 to point to the starting address of the column being processed */
|
pInA = pSrcA->pData + i * 2;
|
|
|
/* matrix multiplication */
|
while (colCnt > 0U)
|
{
|
/* c(m,n) = a(1,1)*b(1,1) + a(1,2) * b(2,1) + .... + a(m,p)*b(p,n) */
|
|
#ifdef UNALIGNED_SUPPORT_DISABLE
|
|
/* read real and imag values from pSrcA buffer */
|
a = *pInA;
|
b = *(pInA + 1U);
|
/* read real and imag values from pSrcB buffer */
|
c = *pInB;
|
d = *(pInB + 1U);
|
|
/* Multiply and Accumlates */
|
sumReal += (q31_t) a *c;
|
sumImag += (q31_t) a *d;
|
sumReal -= (q31_t) b *d;
|
sumImag += (q31_t) b *c;
|
|
/* read next real and imag values from pSrcA buffer */
|
a = *(pInA + 2U);
|
b = *(pInA + 3U);
|
/* read next real and imag values from pSrcB buffer */
|
c = *(pInB + 2U);
|
d = *(pInB + 3U);
|
|
/* update pointer */
|
pInA += 4U;
|
|
/* Multiply and Accumlates */
|
sumReal += (q31_t) a *c;
|
sumImag += (q31_t) a *d;
|
sumReal -= (q31_t) b *d;
|
sumImag += (q31_t) b *c;
|
/* update pointer */
|
pInB += 4U;
|
#else
|
/* read real and imag values from pSrcA and pSrcB buffer */
|
pSourceA = *__SIMD32(pInA)++;
|
pSourceB = *__SIMD32(pInB)++;
|
|
/* Multiply and Accumlates */
|
#ifdef ARM_MATH_BIG_ENDIAN
|
prod1 = -__SMUSD(pSourceA, pSourceB);
|
#else
|
prod1 = __SMUSD(pSourceA, pSourceB);
|
#endif
|
prod2 = __SMUADX(pSourceA, pSourceB);
|
sumReal += (q63_t) prod1;
|
sumImag += (q63_t) prod2;
|
|
/* read real and imag values from pSrcA and pSrcB buffer */
|
pSourceA = *__SIMD32(pInA)++;
|
pSourceB = *__SIMD32(pInB)++;
|
|
/* Multiply and Accumlates */
|
#ifdef ARM_MATH_BIG_ENDIAN
|
prod1 = -__SMUSD(pSourceA, pSourceB);
|
#else
|
prod1 = __SMUSD(pSourceA, pSourceB);
|
#endif
|
prod2 = __SMUADX(pSourceA, pSourceB);
|
sumReal += (q63_t) prod1;
|
sumImag += (q63_t) prod2;
|
|
#endif /* #ifdef UNALIGNED_SUPPORT_DISABLE */
|
|
/* Decrement the loop counter */
|
colCnt--;
|
}
|
|
/* process odd column samples */
|
if ((numColsA & 0x1U) > 0U)
|
{
|
/* c(m,n) = a(1,1)*b(1,1) + a(1,2) * b(2,1) + .... + a(m,p)*b(p,n) */
|
|
#ifdef UNALIGNED_SUPPORT_DISABLE
|
|
/* read real and imag values from pSrcA and pSrcB buffer */
|
a = *pInA++;
|
b = *pInA++;
|
c = *pInB++;
|
d = *pInB++;
|
|
/* Multiply and Accumlates */
|
sumReal += (q31_t) a *c;
|
sumImag += (q31_t) a *d;
|
sumReal -= (q31_t) b *d;
|
sumImag += (q31_t) b *c;
|
|
#else
|
/* read real and imag values from pSrcA and pSrcB buffer */
|
pSourceA = *__SIMD32(pInA)++;
|
pSourceB = *__SIMD32(pInB)++;
|
|
/* Multiply and Accumlates */
|
#ifdef ARM_MATH_BIG_ENDIAN
|
prod1 = -__SMUSD(pSourceA, pSourceB);
|
#else
|
prod1 = __SMUSD(pSourceA, pSourceB);
|
#endif
|
prod2 = __SMUADX(pSourceA, pSourceB);
|
sumReal += (q63_t) prod1;
|
sumImag += (q63_t) prod2;
|
|
#endif /* #ifdef UNALIGNED_SUPPORT_DISABLE */
|
|
}
|
|
/* Saturate and store the result in the destination buffer */
|
|
*px++ = (q15_t) (__SSAT(sumReal >> 15, 16));
|
*px++ = (q15_t) (__SSAT(sumImag >> 15, 16));
|
|
/* Decrement the column loop counter */
|
col--;
|
|
} while (col > 0U);
|
|
i = i + numColsA;
|
|
/* Decrement the row loop counter */
|
row--;
|
|
} while (row > 0U);
|
|
/* set status as ARM_MATH_SUCCESS */
|
status = ARM_MATH_SUCCESS;
|
}
|
|
/* Return to application */
|
return (status);
|
}
|
|
/**
|
* @} end of MatrixMult group
|
*/
|