/*
|
* Copyright (C) 2010-2018 Arm Limited or its affiliates. All rights reserved.
|
*
|
* SPDX-License-Identifier: Apache-2.0
|
*
|
* Licensed under the Apache License, Version 2.0 (the License); you may
|
* not use this file except in compliance with the License.
|
* You may obtain a copy of the License at
|
*
|
* www.apache.org/licenses/LICENSE-2.0
|
*
|
* Unless required by applicable law or agreed to in writing, software
|
* distributed under the License is distributed on an AS IS BASIS, WITHOUT
|
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
* See the License for the specific language governing permissions and
|
* limitations under the License.
|
*/
|
|
/* ----------------------------------------------------------------------
|
* Project: CMSIS NN Library
|
* Title: arm_fully_connected_mat_q7_vec_q15_opt.c
|
* Description: Mixed Q15-Q7 opt fully-connected layer function
|
*
|
* $Date: 17. January 2018
|
* $Revision: V.1.0.0
|
*
|
* Target Processor: Cortex-M cores
|
*
|
* -------------------------------------------------------------------- */
|
|
#include "arm_math.h"
|
#include "arm_nnfunctions.h"
|
|
/**
|
* @ingroup groupNN
|
*/
|
|
/**
|
* @addtogroup FC
|
* @{
|
*/
|
|
/**
|
* @brief Mixed Q15-Q7 opt fully-connected layer function
|
* @param[in] pV pointer to input vector
|
* @param[in] pM pointer to matrix weights
|
* @param[in] dim_vec length of the vector
|
* @param[in] num_of_rows number of rows in weight matrix
|
* @param[in] bias_shift amount of left-shift for bias
|
* @param[in] out_shift amount of right-shift for output
|
* @param[in] bias pointer to bias
|
* @param[in,out] pOut pointer to output vector
|
* @param[in,out] vec_buffer pointer to buffer space for input
|
* @return The function returns <code>ARM_MATH_SUCCESS</code>
|
*
|
* @details
|
*
|
* <b>Buffer size:</b>
|
*
|
* vec_buffer size: 0
|
*
|
* Q7_Q15 version of the fully connected layer
|
*
|
* Weights are in q7_t and Activations are in q15_t
|
*
|
* Limitation: x4 version requires weight reordering to work
|
*
|
* Here we use only one pointer to read 4 rows in the weight
|
* matrix. So if the original q7_t matrix looks like this:
|
*
|
* | a11 | a12 | a13 | a14 | a15 | a16 | a17 |
|
*
|
* | a21 | a22 | a23 | a24 | a25 | a26 | a27 |
|
*
|
* | a31 | a32 | a33 | a34 | a35 | a36 | a37 |
|
*
|
* | a41 | a42 | a43 | a44 | a45 | a46 | a47 |
|
*
|
* | a51 | a52 | a53 | a54 | a55 | a56 | a57 |
|
*
|
* | a61 | a62 | a63 | a64 | a65 | a66 | a67 |
|
*
|
* We operates on multiple-of-4 rows, so the first four rows becomes
|
*
|
* | a11 | a21 | a12 | a22 | a31 | a41 | a32 | a42 |
|
*
|
* | a13 | a23 | a14 | a24 | a33 | a43 | a34 | a44 |
|
*
|
* | a15 | a25 | a16 | a26 | a35 | a45 | a36 | a46 |
|
*
|
* The column left over will be in-order.
|
* which is:
|
* | a17 | a27 | a37 | a47 |
|
*
|
* For the left-over rows, we do 1x1 computation, so the data remains
|
* as its original order.
|
*
|
* So the stored weight matrix looks like this:
|
*
|
* | a11 | a21 | a12 | a22 | a31 | a41 |
|
*
|
* | a32 | a42 | a13 | a23 | a14 | a24 |
|
*
|
* | a33 | a43 | a34 | a44 | a15 | a25 |
|
*
|
* | a16 | a26 | a35 | a45 | a36 | a46 |
|
*
|
* | a17 | a27 | a37 | a47 | a51 | a52 |
|
*
|
* | a53 | a54 | a55 | a56 | a57 | a61 |
|
*
|
* | a62 | a63 | a64 | a65 | a66 | a67 |
|
*
|
*/
|
|
arm_status
|
arm_fully_connected_mat_q7_vec_q15_opt(const q15_t * pV,
|
const q7_t * pM,
|
const uint16_t dim_vec,
|
const uint16_t num_of_rows,
|
const uint16_t bias_shift,
|
const uint16_t out_shift, const q7_t * bias, q15_t * pOut, q15_t * vec_buffer)
|
{
|
|
#if defined (ARM_MATH_DSP)
|
/* Run the following code for Cortex-M4 and Cortex-M7 */
|
|
const q7_t *pB = pM;
|
q15_t *pO = pOut;
|
const q7_t *pBias = bias;
|
const q15_t *pA = pV;
|
|
uint16_t rowCnt = num_of_rows >> 2;
|
|
while (rowCnt)
|
{
|
q31_t sum = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift);
|
q31_t sum2 = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift);
|
q31_t sum3 = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift);
|
q31_t sum4 = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift);
|
|
uint16_t colCnt = dim_vec >> 1;
|
|
pA = pV;
|
|
#ifdef USE_INTRINSIC
|
|
#ifndef ARM_MATH_BIG_ENDIAN
|
|
while (colCnt)
|
{
|
q31_t inM11, inM12, inM13, inM14;
|
q31_t inV;
|
|
inV = *__SIMD32(pA)++;
|
inM11 = *__SIMD32(pB)++;
|
inM12 = __SXTB16(__ROR(inM11, 8));
|
inM11 = __SXTB16(inM11);
|
sum = __SMLAD(inM11, inV, sum);
|
sum2 = __SMLAD(inM12, inV, sum2);
|
inM13 = *__SIMD32(pB)++;
|
inM14 = __SXTB16(__ROR(inM13, 8));
|
inM13 = __SXTB16(inM13);
|
sum3 = __SMLAD(inM13, inV, sum3);
|
sum4 = __SMLAD(inM14, inV, sum4);
|
colCnt--;
|
}
|
|
#else
|
|
while (colCnt)
|
{
|
q31_t inM11, inM12, inM13, inM14;
|
q31_t inV;
|
|
inV = *__SIMD32(pA)++;
|
inM11 = *__SIMD32(pB)++;
|
inM12 = __SXTB16(__ROR(inM11, 8));
|
inM11 = __SXTB16(inM11);
|
sum = __SMLAD(inM12, inV, sum);
|
sum2 = __SMLAD(inM11, inV, sum2);
|
inM13 = *__SIMD32(pB)++;
|
inM14 = __SXTB16(__ROR(inM13, 8));
|
inM13 = __SXTB16(inM13);
|
sum3 = __SMLAD(inM14, inV, sum3);
|
sum4 = __SMLAD(inM13, inV, sum4);
|
colCnt--;
|
}
|
|
#endif /* ARM_MATH_BIG_ENDIAN */
|
|
#else
|
|
/*
|
* register needed:
|
* loop counter: colCnt
|
* accumulators: sum, sum2, sum3, sum4
|
* pointers: pB, pA
|
* weight data: inM11, inM12, inM13, inM14
|
* activation data: inV
|
*/
|
|
#ifndef ARM_MATH_BIG_ENDIAN
|
asm volatile ("COL_LOOP_%=:\n"
|
"ldr.w r4, [%[pA]], #4\n"
|
"ldr.w r1, [%[pB]], #8\n"
|
"mov.w r0, r1, ror #8\n"
|
"sxtb16 r0, r0\n"
|
"sxtb16 r1, r1\n"
|
"smlad %[sum], r4, r1, %[sum]\n"
|
"smlad %[sum2], r4, r0, %[sum2]\n"
|
"ldr.w r3, [%[pB], #-4]\n"
|
"mov.w r2, r3, ror #8\n"
|
"sxtb16 r2, r2\n"
|
"sxtb16 r3, r3\n"
|
"smlad %[sum3], r4, r3, %[sum3]\n"
|
"smlad %[sum4], r4, r2, %[sum4]\n"
|
"subs %[colCnt], #1\n"
|
"bne COL_LOOP_%=\n":[sum] "+r"(sum),
|
[sum2] "+r"(sum2),[sum3] "+r"(sum3),
|
[sum4] "+r"(sum4),[pB] "+r"(pB),[pA] "+r"(pA):[colCnt] "r"(colCnt):"r0", "r1", "r2", "r3", "r4");
|
#else
|
asm volatile ("COL_LOOP_%=:\n"
|
"ldr.w r4, [%[pA]], #4\n"
|
"ldr.w r1, [%[pB]], #8\n"
|
"mov.w r0, r1, ror #8\n"
|
"sxtb16 r0, r0\n"
|
"sxtb16 r1, r1\n"
|
"smlad %[sum], r4, r0, %[sum]\n"
|
"smlad %[sum2], r4, r1, %[sum2]\n"
|
"ldr.w r3, [%[pB], #-4]\n"
|
"mov.w r2, r3, ror #8\n"
|
"sxtb16 r2, r2\n"
|
"sxtb16 r3, r3\n"
|
"smlad %[sum3], r4, r2, %[sum3]\n"
|
"smlad %[sum4], r4, r3, %[sum4]\n"
|
"subs %[colCnt], #1\n"
|
"bne COL_LOOP_%=\n":[sum] "+r"(sum),
|
[sum2] "+r"(sum2),[sum3] "+r"(sum3),
|
[sum4] "+r"(sum4),[pB] "+r"(pB),[pA] "+r"(pA):[colCnt] "r"(colCnt):"r0", "r1", "r2", "r3", "r4");
|
#endif /* ARM_MATH_BIG_ENDIAN */
|
|
#endif /* USE_INTRINSIC */
|
|
colCnt = dim_vec & 0x1;
|
while (colCnt)
|
{
|
q15_t inV = *pA++;
|
q7_t inM = *pB++;
|
q7_t inM2 = *pB++;
|
q7_t inM3 = *pB++;
|
q7_t inM4 = *pB++;
|
|
sum += inV * inM;
|
sum2 += inV * inM2;
|
sum3 += inV * inM3;
|
sum4 += inV * inM4;
|
colCnt--;
|
} /* while over colCnt */
|
*pO++ = (q15_t) (__SSAT((sum >> out_shift), 16));
|
*pO++ = (q15_t) (__SSAT((sum2 >> out_shift), 16));
|
*pO++ = (q15_t) (__SSAT((sum3 >> out_shift), 16));
|
*pO++ = (q15_t) (__SSAT((sum4 >> out_shift), 16));
|
|
/* adjust the pointers and counters */
|
rowCnt--;
|
}
|
|
/* left-over part of the rows */
|
rowCnt = num_of_rows & 0x3;
|
|
while (rowCnt)
|
{
|
q31_t sum = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift);
|
|
uint16_t colCnt = dim_vec >> 2;
|
|
pA = pV;
|
|
while (colCnt)
|
{
|
q31_t inV1, inV2, inM11, inM12;
|
|
pB = (q7_t *) read_and_pad((void *)pB, &inM11, &inM12);
|
|
inV1 = *__SIMD32(pA)++;
|
sum = __SMLAD(inV1, inM11, sum);
|
|
inV2 = *__SIMD32(pA)++;
|
sum = __SMLAD(inV2, inM12, sum);
|
|
colCnt--;
|
}
|
|
/* left-over of the vector */
|
colCnt = dim_vec & 0x3;
|
while (colCnt)
|
{
|
q15_t inV = *pA++;
|
q7_t inM = *pB++;
|
sum += inV * inM;
|
colCnt--;
|
}
|
|
*pO++ = (q15_t) (__SSAT((sum >> out_shift), 16));
|
|
rowCnt--;
|
}
|
|
#else
|
/* Run the following code as reference implementation for Cortex-M0 and Cortex-M3 */
|
uint16_t rowCnt = num_of_rows >> 2;
|
const q7_t *pB = pM;
|
const q15_t *pA;
|
q15_t *pO = pOut;
|
const q7_t *pBias = bias;
|
|
while (rowCnt)
|
{
|
q31_t sum = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift);
|
q31_t sum2 = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift);
|
q31_t sum3 = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift);
|
q31_t sum4 = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift);
|
uint16_t colCnt = dim_vec >> 1;
|
|
pA = pV;
|
|
while (colCnt)
|
{
|
q15_t inA1 = *pA++;
|
q15_t inA2 = *pA++;
|
|
q7_t inB1 = *pB++;
|
q7_t inB3 = *pB++;
|
q7_t inB2 = *pB++;
|
q7_t inB4 = *pB++;
|
|
sum += inA1 * inB1 + inA2 * inB2;
|
sum2 += inA1 * inB3 + inA2 * inB4;
|
|
inB1 = *pB++;
|
inB3 = *pB++;
|
inB2 = *pB++;
|
inB4 = *pB++;
|
|
sum3 += inA1 * inB1 + inA2 * inB2;
|
sum4 += inA1 * inB3 + inA2 * inB4;
|
|
colCnt--;
|
}
|
|
colCnt = dim_vec & 0x1;
|
while (colCnt)
|
{
|
q15_t inA = *pA++;
|
q7_t inB = *pB++;
|
sum += inA * inB;
|
inB = *pB++;
|
sum2 += inA * inB;
|
inB = *pB++;
|
sum3 += inA * inB;
|
inB = *pB++;
|
sum4 += inA * inB;
|
|
colCnt--;
|
}
|
*pO++ = (q15_t) __SSAT((sum >> out_shift), 16);
|
*pO++ = (q15_t) __SSAT((sum2 >> out_shift), 16);
|
*pO++ = (q15_t) __SSAT((sum3 >> out_shift), 16);
|
*pO++ = (q15_t) __SSAT((sum4 >> out_shift), 16);
|
|
rowCnt--;
|
}
|
|
rowCnt = num_of_rows & 0x3;
|
|
while (rowCnt)
|
{
|
int ip_out = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift);
|
int j;
|
|
pA = pV;
|
for (j = 0; j < dim_vec; j++)
|
{
|
q15_t inA = *pA++;
|
q7_t inB = *pB++;
|
ip_out += inA * inB;
|
}
|
*pO++ = (q15_t) __SSAT((ip_out >> out_shift), 16);
|
|
rowCnt--;
|
}
|
|
#endif /* ARM_MATH_DSP */
|
|
/* Return to ARM_MATH_SUCCESS */
|
return (ARM_MATH_SUCCESS);
|
|
}
|
|
/**
|
* @} end of FC group
|
*/
|