/*
|
* Copyright (C) 2017 The Android Open Source Project
|
*
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
* you may not use this file except in compliance with the License.
|
* You may obtain a copy of the License at
|
*
|
* http://www.apache.org/licenses/LICENSE-2.0
|
*
|
* Unless required by applicable law or agreed to in writing, software
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
* See the License for the specific language governing permissions and
|
* limitations under the License.
|
*/
|
|
#define LOG_TAG "OperationsUtils"
|
|
#include "OperationsUtils.h"
|
#include "Operations.h"
|
#include "Utils.h"
|
|
#include <cmath>
|
|
namespace android {
|
namespace nn {
|
|
namespace {
|
|
bool validateOperandTypes(const std::vector<OperandType>& expectedTypes, const char* tag,
|
uint32_t operandCount,
|
std::function<OperandType(uint32_t)> getOperandType) {
|
NN_RET_CHECK_EQ(operandCount, expectedTypes.size());
|
for (uint32_t i = 0; i < operandCount; ++i) {
|
OperandType type = getOperandType(i);
|
NN_RET_CHECK(type == expectedTypes[i])
|
<< "Invalid " << tag << " tensor type " << toString(type) << " for " << tag << " "
|
<< i << ", expected " << toString(expectedTypes[i]);
|
}
|
return true;
|
}
|
|
} // namespace
|
|
bool validateInputTypes(const IOperationValidationContext* context,
|
const std::vector<OperandType>& expectedTypes) {
|
return validateOperandTypes(expectedTypes, "input", context->getNumInputs(),
|
[context](uint32_t index) { return context->getInputType(index); });
|
}
|
|
bool validateOutputTypes(const IOperationValidationContext* context,
|
const std::vector<OperandType>& expectedTypes) {
|
return validateOperandTypes(
|
expectedTypes, "output", context->getNumOutputs(),
|
[context](uint32_t index) { return context->getOutputType(index); });
|
}
|
|
bool validateHalVersion(const IOperationValidationContext* context,
|
HalVersion minSupportedHalVersion) {
|
if (context->getHalVersion() < minSupportedHalVersion) {
|
NN_RET_CHECK_FAIL() << "The given inputs and outputs are only supported in "
|
<< toString(minSupportedHalVersion) << " and later (validating using "
|
<< toString(context->getHalVersion()) << ")";
|
}
|
return true;
|
}
|
|
bool SameShape(const Shape& in1, const Shape& in2) {
|
if (in1.type != in2.type || in1.dimensions.size() != in2.dimensions.size()) {
|
return false;
|
}
|
for (size_t i = 0; i < in1.dimensions.size(); i++) {
|
if (in1.dimensions[i] != in2.dimensions[i]) {
|
return false;
|
}
|
}
|
return true;
|
}
|
|
bool SetShape(const Shape& in, Shape* out) {
|
if (in.type != out->type) {
|
return false;
|
}
|
out->dimensions = in.dimensions;
|
return true;
|
}
|
|
bool combineDimensions(const std::vector<uint32_t>& lhs, const std::vector<uint32_t>& rhs,
|
std::vector<uint32_t>* combined) {
|
if (rhs.empty()) {
|
*combined = lhs;
|
return true;
|
}
|
if (lhs.empty()) {
|
*combined = rhs;
|
return true;
|
}
|
NN_RET_CHECK_EQ(lhs.size(), rhs.size()) << "incompatible ranks";
|
combined->resize(lhs.size());
|
for (uint32_t i = 0; i < lhs.size(); i++) {
|
if (lhs[i] == 0) {
|
(*combined)[i] = rhs[i];
|
continue;
|
}
|
if (rhs[i] == 0) {
|
(*combined)[i] = lhs[i];
|
continue;
|
}
|
NN_RET_CHECK_EQ(lhs[i], rhs[i]) << "incompatible dimension: " << i;
|
(*combined)[i] = lhs[i];
|
}
|
return true;
|
}
|
|
uint32_t getNumberOfElements(const Shape& shape) {
|
uint32_t count = 1;
|
for (size_t i = 0; i < shape.dimensions.size(); i++) {
|
count *= shape.dimensions[i];
|
}
|
return count;
|
}
|
|
uint32_t getNumberOfElements(const Shape& shape,
|
size_t firstAxisInclusive,
|
size_t lastAxisExclusive) {
|
nnAssert(0 <= firstAxisInclusive);
|
nnAssert(firstAxisInclusive <= lastAxisExclusive);
|
nnAssert(lastAxisExclusive <= shape.dimensions.size());
|
uint32_t count = 1;
|
for (size_t i = firstAxisInclusive; i < lastAxisExclusive; i++) {
|
count *= shape.dimensions[i];
|
}
|
return count;
|
}
|
|
uint32_t getNumberOfDimensions(const Shape& shape) {
|
return shape.dimensions.size();
|
}
|
|
uint32_t getSizeOfDimension(const Shape& shape, uint32_t dimensionIdx) {
|
nnAssert(0 <= dimensionIdx && dimensionIdx < shape.dimensions.size());
|
return shape.dimensions[dimensionIdx];
|
}
|
|
bool handleNegativeAxis(int32_t numberOfDimensions, int32_t* axis) {
|
NN_CHECK(-numberOfDimensions <= *axis && *axis < numberOfDimensions);
|
if (*axis < 0) {
|
*axis += numberOfDimensions;
|
}
|
return true;
|
}
|
|
bool QuantizeMultiplier(double double_multiplier, int32_t* quantized_multiplier, int* shift) {
|
if (double_multiplier == 0.) {
|
*quantized_multiplier = 0;
|
*shift = 0;
|
return true;
|
}
|
const double q = std::frexp(double_multiplier, shift);
|
auto q_fixed = static_cast<int64_t>(std::round(q * (1ll << 31)));
|
NN_RET_CHECK(q_fixed <= (1ll << 31));
|
if (q_fixed == (1ll << 31)) {
|
q_fixed /= 2;
|
++*shift;
|
}
|
NN_RET_CHECK_LE(q_fixed, std::numeric_limits<int32_t>::max());
|
*quantized_multiplier = static_cast<int32_t>(q_fixed);
|
return true;
|
}
|
|
bool QuantizeMultiplierSmallerThanOne(double double_multiplier,
|
int32_t* quantized_multiplier,
|
int32_t* right_shift) {
|
NN_OPS_CHECK(double_multiplier >= 0.);
|
NN_OPS_CHECK(double_multiplier < 1.);
|
if (double_multiplier == 0.) {
|
*quantized_multiplier = 0;
|
*right_shift = 0;
|
return true;
|
}
|
NN_OPS_CHECK(double_multiplier > 0.);
|
const double q = std::frexp(double_multiplier, right_shift);
|
*right_shift *= -1;
|
int64_t q_fixed = static_cast<int64_t>(std::round(q * (1LL << 31)));
|
NN_OPS_CHECK(q_fixed <= (1LL << 31));
|
if (q_fixed == (1LL << 31)) {
|
q_fixed /= 2;
|
--*right_shift;
|
}
|
NN_OPS_CHECK(*right_shift >= 0);
|
NN_OPS_CHECK(q_fixed <= std::numeric_limits<int32_t>::max());
|
*quantized_multiplier = static_cast<int32_t>(q_fixed);
|
return true;
|
}
|
|
bool QuantizeMultiplierGreaterThanOne(double double_multiplier,
|
int32_t* quantized_multiplier,
|
int* left_shift) {
|
NN_OPS_CHECK(double_multiplier > 1.);
|
const double q = std::frexp(double_multiplier, left_shift);
|
int64_t q_fixed = static_cast<int64_t>(std::round(q * (1LL << 31)));
|
NN_OPS_CHECK(q_fixed <= (1LL << 31));
|
if (q_fixed == (1LL << 31)) {
|
q_fixed /= 2;
|
++*left_shift;
|
}
|
NN_OPS_CHECK(*left_shift >= 0);
|
NN_OPS_CHECK(q_fixed <= std::numeric_limits<int32_t>::max());
|
*quantized_multiplier = static_cast<int32_t>(q_fixed);
|
return true;
|
}
|
|
bool GetQuantizedConvolutionMultipler(const Shape& inputShape, const Shape& filterShape,
|
const Shape& biasShape, const Shape& outputShape,
|
double* multiplier) {
|
// Upcast bias and input_product to double
|
const double input_product_scale = inputShape.scale * filterShape.scale;
|
const double bias_scale = biasShape.scale;
|
|
// The following conditions must be guaranteed by the training pipeline.
|
NN_OPS_CHECK(std::abs(input_product_scale - bias_scale) <=
|
1e-6 * std::min(input_product_scale, bias_scale));
|
NN_OPS_CHECK(input_product_scale >= 0);
|
*multiplier = input_product_scale / outputShape.scale;
|
return true;
|
}
|
|
void CalculateActivationRangeUint8(int32_t activation,
|
const Shape& outputShape,
|
int32_t* act_min,
|
int32_t* act_max) {
|
const int32_t qmin = std::numeric_limits<uint8_t>::min();
|
const int32_t qmax = std::numeric_limits<uint8_t>::max();
|
|
const auto scale = outputShape.scale;
|
const auto zero_point = outputShape.offset;
|
|
auto quantize = [scale, zero_point](float f) {
|
return zero_point + static_cast<int32_t>(std::round(f / scale));
|
};
|
|
if (activation == kActivationRelu) {
|
*act_min = std::max(qmin, quantize(0.0));
|
*act_max = qmax;
|
} else if (activation == kActivationRelu6) {
|
*act_min = std::max(qmin, quantize(0.0));
|
*act_max = std::min(qmax, quantize(6.0));
|
} else if (activation == kActivationRelu1) {
|
*act_min = std::max(qmin, quantize(-1.0));
|
*act_max = std::min(qmax, quantize(1.0));
|
} else if (activation == kActivationNone){
|
*act_min = qmin;
|
*act_max = qmax;
|
} else {
|
LOG(ERROR) << "Unsupported fused activation function.";
|
}
|
}
|
|
void CalculateActivationRangeFloat(int32_t activation,
|
float* activation_min,
|
float* activation_max) {
|
if (activation == kActivationRelu) {
|
*activation_min = 0.f;
|
*activation_max = std::numeric_limits<float>::max();
|
} else if (activation == kActivationRelu6) {
|
*activation_min = 0.f;
|
*activation_max = 6.f;
|
} else if (activation == kActivationRelu1) {
|
*activation_min = -1.f;
|
*activation_max = 1.f;
|
} else if (activation == kActivationNone){
|
*activation_min = std::numeric_limits<float>::lowest();
|
*activation_max = std::numeric_limits<float>::max();
|
} else {
|
LOG(ERROR) << "Unsupported fused activation function.";
|
}
|
}
|
|
int32_t CalculateInputRadius(int input_integer_bits, int input_left_shift) {
|
const double max_input_rescaled = 1.0 * ((1 << input_integer_bits) - 1) *
|
(1LL << (31 - input_integer_bits)) /
|
(1LL << input_left_shift);
|
// Tighten bound using floor. Suppose that we could use the exact value.
|
// After scaling the difference, the result would be at the maximum. Thus we
|
// must ensure that our value has lower magnitude.
|
return static_cast<int32_t>(std::floor(max_input_rescaled));
|
}
|
|
void calculateExplicitPaddingImpl(int32_t in_size, int32_t stride, int32_t dilation_factor,
|
int32_t filter_size, int32_t padding_implicit,
|
bool isTransposeConv, int32_t* padding_head,
|
int32_t* padding_tail) {
|
*padding_head = 0;
|
*padding_tail = 0;
|
|
int32_t effective_filter_size = (filter_size - 1) * dilation_factor + 1;
|
|
if (padding_implicit == kPaddingSame) {
|
int32_t out_size = (in_size + stride - 1) / stride;
|
int32_t tmp = (out_size - 1) * stride + effective_filter_size;
|
if (tmp > in_size) {
|
*padding_head = (tmp - in_size) / 2;
|
*padding_tail = (tmp - in_size) - *padding_head;
|
}
|
// For transpose conv, make padding tail fit tightly to the end of the last stride.
|
if (isTransposeConv) {
|
*padding_tail = (tmp - in_size) - *padding_head;
|
}
|
}
|
}
|
|
bool calculateBroadcastedShape(const Shape& in1, const Shape& in2, Shape* out) {
|
NN_RET_CHECK(in1.type == in2.type);
|
uint32_t numberOfDims1 = getNumberOfDimensions(in1);
|
uint32_t numberOfDims2 = getNumberOfDimensions(in2);
|
uint32_t maxDims = std::max(numberOfDims1, numberOfDims2);
|
out->dimensions = std::vector<uint32_t>(maxDims);
|
for (uint32_t i = 1; i <= maxDims; i++) {
|
uint32_t dim1 = 1;
|
if (i <= numberOfDims1) {
|
dim1 = getSizeOfDimension(in1, numberOfDims1 - i);
|
}
|
uint32_t dim2 = 1;
|
if (i <= numberOfDims2) {
|
dim2 = getSizeOfDimension(in2, numberOfDims2 - i);
|
}
|
if (dim1 != dim2 && dim1 != 1 && dim2 != 1) {
|
LOG(ERROR) << "Dimensions mismatch for broadcast:\n"
|
<< "First tensor: dimension " << numberOfDims1 - i << " of size " << dim1
|
<< "\nSecond tensor: dimension " << numberOfDims2 - i << "of size " << dim2;
|
return false;
|
}
|
out->dimensions[maxDims - i] = (dim1 == 1) ? dim2 : dim1;
|
}
|
return true;
|
}
|
|
uint8_t requantize(uint8_t value, const Shape& oldShape, const Shape& newShape) {
|
double doubleValue = (value - oldShape.offset) * oldShape.scale;
|
double doubleRet = doubleValue / newShape.scale + newShape.offset;
|
if (doubleRet < 0) return 0;
|
if (doubleRet > 255) return 255;
|
return static_cast<uint8_t>(std::round(doubleRet));
|
}
|
|
bool floorPrepare(const Shape& input, Shape* output) {
|
return SetShape(input, output);
|
}
|
|
bool depthwiseConvPrepare(const Shape& input, const Shape& filter, const Shape& bias,
|
int32_t padding_left, int32_t padding_right, int32_t padding_top,
|
int32_t padding_bottom, int32_t stride_width, int32_t stride_height,
|
int32_t depth_multiplier, int32_t dilation_width_factor,
|
int32_t dilation_height_factor, Shape* output) {
|
if (filter.type == OperandType::TENSOR_QUANT8_SYMM_PER_CHANNEL) {
|
NN_OPS_CHECK(input.type == OperandType::TENSOR_QUANT8_ASYMM);
|
} else {
|
NN_OPS_CHECK(input.type == filter.type);
|
}
|
if (input.type == OperandType::TENSOR_QUANT8_ASYMM) {
|
NN_OPS_CHECK(bias.type == OperandType::TENSOR_INT32);
|
} else {
|
NN_OPS_CHECK(input.type == bias.type);
|
}
|
NN_OPS_CHECK(getNumberOfDimensions(input) == 4);
|
NN_OPS_CHECK(getNumberOfDimensions(filter) == 4);
|
NN_OPS_CHECK(getNumberOfDimensions(bias) == 1);
|
|
NN_OPS_CHECK(getSizeOfDimension(filter, 3) == getSizeOfDimension(bias, 0));
|
|
uint32_t channels_out = getSizeOfDimension(filter, 3);
|
uint32_t channels_in = getSizeOfDimension(input, 3);
|
uint32_t width = getSizeOfDimension(input, 2);
|
uint32_t height = getSizeOfDimension(input, 1);
|
uint32_t filterWidth = getSizeOfDimension(filter, 2);
|
uint32_t filterHeight = getSizeOfDimension(filter, 1);
|
uint32_t batches = getSizeOfDimension(input, 0);
|
|
NN_OPS_CHECK(depth_multiplier * channels_in == channels_out);
|
int32_t effectiveFilterWidth = (filterWidth - 1) * dilation_width_factor + 1;
|
int32_t effectiveFilterHeight = (filterHeight - 1) * dilation_height_factor + 1;
|
NN_RET_CHECK_GT(effectiveFilterWidth, padding_left);
|
NN_RET_CHECK_GT(effectiveFilterWidth, padding_right);
|
NN_RET_CHECK_GT(effectiveFilterHeight, padding_top);
|
NN_RET_CHECK_GT(effectiveFilterHeight, padding_bottom);
|
|
uint32_t outWidth = computeOutSize(width, filterWidth, stride_width, dilation_width_factor,
|
padding_left, padding_right);
|
uint32_t outHeight = computeOutSize(height, filterHeight, stride_height, dilation_height_factor,
|
padding_top, padding_bottom);
|
|
output->type = input.type;
|
output->dimensions = {batches, outHeight, outWidth, channels_out};
|
return true;
|
}
|
|
bool genericActivationPrepare(const Shape& input,
|
Shape* output) {
|
NN_OPS_CHECK(getNumberOfDimensions(input) <= 4);
|
return SetShape(input, output);
|
}
|
|
bool genericNormalizationPrepare(const Shape& input, Shape* output) {
|
return SetShape(input, output);
|
}
|
|
bool reshapePrepare(const Shape& input,
|
const int32_t* targetDims,
|
const int32_t targetDimsSize,
|
Shape* output) {
|
// Reshape allows one of the targetDims components to have the
|
// special -1 value, meaning it will be calculated automatically based on the
|
// input. Here we calculate what that dimension should be so that the number
|
// of output elements in the same as the number of input elements.
|
int32_t numInputElements = (int32_t) getNumberOfElements(input);
|
|
std::vector<uint32_t> outDims(targetDimsSize);
|
int32_t numOutputElements = 1;
|
int32_t strechDim = -1;
|
for (int32_t i = 0; i < targetDimsSize; ++i) {
|
int32_t value = targetDims[i];
|
if (value == -1) {
|
NN_OPS_CHECK(strechDim == -1);
|
strechDim = i;
|
} else {
|
numOutputElements *= value;
|
outDims[i] = (uint32_t)value;
|
}
|
}
|
if (strechDim != -1) {
|
int32_t strechValue = numInputElements / numOutputElements;
|
outDims[strechDim] = (uint32_t) strechValue;
|
numOutputElements *= strechValue;
|
}
|
|
NN_OPS_CHECK(numInputElements == numOutputElements);
|
|
output->type = input.type;
|
output->dimensions = outDims;
|
output->offset = input.offset;
|
output->scale = input.scale;
|
|
return true;
|
}
|
|
bool depthToSpacePrepare(const Shape& input,
|
int32_t blockSize,
|
Shape* output) {
|
NN_OPS_CHECK(getNumberOfDimensions(input) == 4);
|
NN_OPS_CHECK(blockSize > 0);
|
|
uint32_t batches = getSizeOfDimension(input, 0);
|
uint32_t height = getSizeOfDimension(input, 1);
|
uint32_t width = getSizeOfDimension(input, 2);
|
uint32_t channels = getSizeOfDimension(input, 3);
|
|
NN_OPS_CHECK(channels % (blockSize * blockSize) == 0);
|
output->type = input.type;
|
output->dimensions = {batches,
|
height * blockSize,
|
width * blockSize,
|
channels / (blockSize * blockSize)};
|
output->offset = input.offset;
|
output->scale = input.scale;
|
|
return true;
|
}
|
|
bool spaceToDepthPrepare(const Shape& input,
|
int32_t blockSize,
|
Shape* output) {
|
NN_OPS_CHECK(getNumberOfDimensions(input) == 4);
|
NN_OPS_CHECK(blockSize > 0);
|
|
uint32_t batches = getSizeOfDimension(input, 0);
|
uint32_t height = getSizeOfDimension(input, 1);
|
uint32_t width = getSizeOfDimension(input, 2);
|
uint32_t channels = getSizeOfDimension(input, 3);
|
|
NN_OPS_CHECK(height % blockSize == 0);
|
NN_OPS_CHECK(width % blockSize == 0);
|
|
output->type = input.type;
|
output->dimensions = {batches,
|
height / blockSize,
|
width / blockSize,
|
channels * (blockSize * blockSize)};
|
output->offset = input.offset;
|
output->scale = input.scale;
|
|
return true;
|
}
|
|
bool embeddingLookupPrepare(const Shape &valueShape,
|
const Shape &lookupShape,
|
Shape *outputShape) {
|
NN_OPS_CHECK(getNumberOfDimensions(valueShape) >= 2);
|
NN_OPS_CHECK(getNumberOfDimensions(lookupShape) == 1);
|
|
const uint32_t rows = getSizeOfDimension(valueShape, 0);
|
const uint32_t columns = getSizeOfDimension(valueShape, 1);
|
|
const uint32_t lookups = getSizeOfDimension(lookupShape, 0);
|
|
outputShape->type = valueShape.type;
|
outputShape->dimensions = { lookups, columns };
|
for (uint32_t i = 2; i < getNumberOfDimensions(valueShape); i++) {
|
outputShape->dimensions.push_back(getSizeOfDimension(valueShape, i));
|
}
|
outputShape->offset = valueShape.offset;
|
outputShape->scale = valueShape.scale;
|
|
return true;
|
}
|
|
bool hashtableLookupPrepare(const Shape &lookupShape,
|
const Shape &keyShape,
|
const Shape &valueShape,
|
Shape *outputShape,
|
Shape *hitShape) {
|
NN_OPS_CHECK(getNumberOfDimensions(lookupShape) == 1);
|
NN_OPS_CHECK(getNumberOfDimensions(keyShape) == 1);
|
NN_OPS_CHECK(getNumberOfDimensions(valueShape) >= 1);
|
|
const uint32_t lookups = getSizeOfDimension(lookupShape, 0);
|
const uint32_t keys = getSizeOfDimension(keyShape, 0);
|
const uint32_t rows = getSizeOfDimension(valueShape, 0);
|
outputShape->type = valueShape.type;
|
outputShape->dimensions = { lookups };
|
for (uint32_t i = 1; i < getNumberOfDimensions(valueShape); i++) {
|
outputShape->dimensions.push_back(getSizeOfDimension(valueShape, i));
|
}
|
outputShape->offset = valueShape.offset;
|
outputShape->scale = valueShape.scale;
|
|
hitShape->type = OperandType::TENSOR_QUANT8_ASYMM;
|
hitShape->dimensions = { lookups };
|
hitShape->offset = 0;
|
hitShape->scale = 1.f;
|
|
return true;
|
}
|
|
bool padPrepare(const Shape& input,
|
const int32_t* paddingsData,
|
const Shape& paddingsShape,
|
Shape* output) {
|
uint32_t numInputDims = getNumberOfDimensions(input);
|
|
// paddings need to be provided as a 2-D int32 tensor.
|
NN_OPS_CHECK(paddingsShape.type == OperandType::TENSOR_INT32);
|
NN_OPS_CHECK(getNumberOfDimensions(paddingsShape) == 2);
|
NN_OPS_CHECK(getSizeOfDimension(paddingsShape, 0) == numInputDims);
|
NN_OPS_CHECK(getSizeOfDimension(paddingsShape, 1) == 2);
|
|
std::vector<uint32_t> outDims(numInputDims);
|
for (uint32_t i = 0; i < numInputDims; ++i) {
|
int32_t beforePadding = *paddingsData++;
|
int32_t afterPadding = *paddingsData++;
|
// Pad value has to be greater than equal to 0.
|
NN_OPS_CHECK(beforePadding >= 0 && afterPadding >= 0);
|
outDims[i] = beforePadding + getSizeOfDimension(input, i) + afterPadding;
|
}
|
output->type = input.type;
|
output->dimensions = outDims;
|
output->offset = input.offset;
|
output->scale = input.scale;
|
|
return true;
|
}
|
|
bool batchToSpacePrepare(const Shape& input,
|
const int32_t* blockSizeData,
|
const Shape& blockSizeShape,
|
Shape* output) {
|
// Only 4D NHWC tensors are supported.
|
NN_OPS_CHECK(getNumberOfDimensions(input) == 4);
|
|
// blockSize need to be provided as a 1-D int32 tensor.
|
NN_OPS_CHECK(blockSizeShape.type == OperandType::TENSOR_INT32);
|
NN_OPS_CHECK(getNumberOfDimensions(blockSizeShape) == 1);
|
// Only applies to spatial dimensions.
|
NN_OPS_CHECK(getSizeOfDimension(blockSizeShape, 0) == 2);
|
|
uint32_t batches = getSizeOfDimension(input, 0);
|
uint32_t height = getSizeOfDimension(input, 1);
|
uint32_t width = getSizeOfDimension(input, 2);
|
uint32_t channels = getSizeOfDimension(input, 3);
|
|
NN_OPS_CHECK(batches % (blockSizeData[0] * blockSizeData[1]) == 0);
|
output->type = input.type;
|
output->dimensions = {batches / (blockSizeData[0] * blockSizeData[1]),
|
height * blockSizeData[0],
|
width * blockSizeData[1],
|
channels};
|
output->offset = input.offset;
|
output->scale = input.scale;
|
|
return true;
|
}
|
|
bool spaceToBatchPrepare(const Shape& input,
|
const int32_t* blockSizeData,
|
const Shape& blockSizeShape,
|
const int32_t* paddingsData,
|
const Shape& paddingsShape,
|
Shape* output) {
|
// Only 4D NHWC tensors are supported.
|
NN_OPS_CHECK(getNumberOfDimensions(input) == 4);
|
|
// blockSize need to be provided as a 1-D int32 tensor.
|
NN_OPS_CHECK(blockSizeShape.type == OperandType::TENSOR_INT32);
|
NN_OPS_CHECK(getNumberOfDimensions(blockSizeShape) == 1);
|
// Only applies to spatial dimensions.
|
NN_OPS_CHECK(getSizeOfDimension(blockSizeShape, 0) == 2);
|
|
// paddings need to be provided as a 2-D int32 tensor.
|
NN_OPS_CHECK(paddingsShape.type == OperandType::TENSOR_INT32);
|
NN_OPS_CHECK(getNumberOfDimensions(paddingsShape) == 2);
|
NN_OPS_CHECK(getSizeOfDimension(paddingsShape, 0) == 2);
|
NN_OPS_CHECK(getSizeOfDimension(paddingsShape, 1) == 2);
|
|
uint32_t batches = getSizeOfDimension(input, 0);
|
uint32_t height = getSizeOfDimension(input, 1);
|
uint32_t width = getSizeOfDimension(input, 2);
|
uint32_t channels = getSizeOfDimension(input, 3);
|
|
uint32_t paddedHeight = paddingsData[0] + height + paddingsData[1];
|
uint32_t paddedWidth = paddingsData[2] + width + paddingsData[3];
|
|
NN_OPS_CHECK(paddedHeight % blockSizeData[0] == 0);
|
NN_OPS_CHECK(paddedWidth % blockSizeData[1] == 0);
|
|
output->type = input.type;
|
output->dimensions = {batches * (blockSizeData[0] * blockSizeData[1]),
|
paddedHeight / blockSizeData[0],
|
paddedWidth / blockSizeData[1],
|
channels};
|
output->offset = input.offset;
|
output->scale = input.scale;
|
|
return true;
|
}
|
|
bool squeezePrepare(const Shape& input,
|
const int32_t* squeezeDims,
|
const Shape& squeezeDimsShape,
|
Shape* output) {
|
int32_t numInputDims = static_cast<int32_t>(getNumberOfDimensions(input));
|
|
// squeezeDims need to be provided as a 1-D int32 tensor.
|
NN_OPS_CHECK(squeezeDimsShape.type == OperandType::TENSOR_INT32);
|
NN_OPS_CHECK(getNumberOfDimensions(squeezeDimsShape) == 1);
|
|
int32_t squeezeDimsSize = static_cast<int32_t>(getSizeOfDimension(squeezeDimsShape, 0));
|
std::vector<bool> shouldSqueeze(numInputDims, false);
|
int32_t numDimsSqueezed = 0;
|
|
if (squeezeDimsSize == 0) {
|
// If squeezeDimsSize is 0, all dims with value 1 will be squeezed.
|
for (int32_t idx = 0; idx < numInputDims; ++idx) {
|
if (getSizeOfDimension(input, idx) == 1) {
|
shouldSqueeze[idx] = true;
|
++numDimsSqueezed;
|
}
|
}
|
} else {
|
for (int32_t idx = 0; idx < squeezeDimsSize; ++idx) {
|
int32_t current = squeezeDims[idx] < 0 ? squeezeDims[idx] + numInputDims
|
: squeezeDims[idx];
|
NN_OPS_CHECK(current >= 0 && current < numInputDims &&
|
getSizeOfDimension(input, current) == 1);
|
if (!shouldSqueeze[current]) ++numDimsSqueezed;
|
shouldSqueeze[current] = true;
|
}
|
}
|
|
// Sets output dimensions.
|
std::vector<uint32_t> outDims(numInputDims - numDimsSqueezed);
|
for (int32_t inIdx = 0, outIdx = 0; inIdx < numInputDims; ++inIdx) {
|
if (!shouldSqueeze[inIdx]) {
|
outDims[outIdx++] = getSizeOfDimension(input, inIdx);
|
}
|
}
|
|
output->type = input.type;
|
output->dimensions = outDims;
|
output->offset = input.offset;
|
output->scale = input.scale;
|
|
return true;
|
}
|
|
bool meanPrepare(const Shape& input,
|
const int32_t* axisData,
|
const Shape& axisShape,
|
bool keepDims,
|
Shape* output) {
|
|
// perm need to be provided as a 1-D int32 tensor.
|
NN_OPS_CHECK(axisShape.type == OperandType::TENSOR_INT32);
|
NN_OPS_CHECK(getNumberOfDimensions(axisShape) == 1);
|
|
int32_t numInputDims = static_cast<int32_t>(getNumberOfDimensions(input));
|
int32_t axisSize = static_cast<int32_t>(getSizeOfDimension(axisShape, 0));
|
|
// Determines size of output tensor.
|
if (keepDims) {
|
std::vector<uint32_t> outDims(numInputDims);
|
for (int32_t idx = 0; idx < numInputDims; ++idx) {
|
bool isAxis = false;
|
for (int32_t axisIdx = 0; axisIdx < axisSize; ++axisIdx) {
|
if (axisData[axisIdx] == idx || axisData[axisIdx] + numInputDims == idx) {
|
isAxis = true;
|
break;
|
}
|
}
|
if (isAxis) {
|
outDims[idx] = 1;
|
} else {
|
outDims[idx] = getSizeOfDimension(input, idx);
|
}
|
}
|
output->dimensions = outDims;
|
} else {
|
// Calculates size of reducing axis.
|
int32_t numReduceAxis = axisSize;
|
for (int32_t i = 0; i < axisSize; ++i) {
|
int32_t current = axisData[i];
|
if (current < 0) {
|
current += numInputDims;
|
}
|
NN_OPS_CHECK(current >= 0 && current < numInputDims);
|
for (int32_t j = 0; j < i; ++j) {
|
int32_t previous = axisData[j];
|
if (previous < 0) {
|
previous += numInputDims;
|
}
|
if (current == previous) {
|
--numReduceAxis;
|
break;
|
}
|
}
|
}
|
// Determines output dimensions.
|
std::vector<uint32_t> outDims(numInputDims - numReduceAxis);
|
int32_t numSkipAxis = 0;
|
for (int32_t idx = 0; idx < numInputDims; ++idx) {
|
bool isAxis = false;
|
for (int32_t axisIdx = 0; axisIdx < axisSize; ++axisIdx) {
|
if (axisData[axisIdx] == idx || axisData[axisIdx] + numInputDims == idx) {
|
++numSkipAxis;
|
isAxis = true;
|
break;
|
}
|
}
|
if (!isAxis) {
|
outDims[idx - numSkipAxis] = getSizeOfDimension(input, idx);
|
}
|
}
|
output->dimensions = outDims;
|
}
|
|
output->type = input.type;
|
output->offset = input.offset;
|
output->scale = input.scale;
|
|
return true;
|
}
|
|
bool stridedSlicePrepare(const Shape& input,
|
const int32_t* beginData, const Shape& beginShape,
|
const int32_t* endData, const Shape& endShape,
|
const int32_t* stridesData, const Shape& stridesShape,
|
int32_t beginMask, int32_t endMask, int32_t shrinkAxisMask,
|
Shape* output) {
|
uint32_t numInputDims = getNumberOfDimensions(input);
|
// StridedSlice op only supports 1D-4D input arrays.
|
NN_OPS_CHECK(numInputDims <= 4);
|
|
NN_OPS_CHECK(getNumberOfDimensions(beginShape) == 1);
|
NN_OPS_CHECK(getNumberOfDimensions(endShape) == 1);
|
NN_OPS_CHECK(getNumberOfDimensions(stridesShape) == 1);
|
|
NN_OPS_CHECK(getSizeOfDimension(beginShape, 0) == numInputDims);
|
NN_OPS_CHECK(getSizeOfDimension(endShape, 0) == numInputDims);
|
NN_OPS_CHECK(getSizeOfDimension(stridesShape, 0) == numInputDims);
|
|
NN_OPS_CHECK(beginShape.type == OperandType::TENSOR_INT32);
|
NN_OPS_CHECK(endShape.type == OperandType::TENSOR_INT32);
|
NN_OPS_CHECK(stridesShape.type == OperandType::TENSOR_INT32);
|
|
// Determine size of output tensor and map indices
|
std::vector<uint32_t> outDims;
|
for (int32_t idx = 0; idx < static_cast<int32_t>(numInputDims); idx++) {
|
int32_t dim = static_cast<int32_t>(getSizeOfDimension(input, idx));
|
int32_t stride = stridesData[idx];
|
// stride value has to be non-zero
|
NN_OPS_CHECK(stride != 0);
|
bool positiveStride = stride > 0;
|
|
int32_t begin = beginMask & (1 << idx)
|
? positiveStride ? 0 : dim - 1
|
: ClampedIndex(beginData[idx], dim, positiveStride);
|
int32_t end = endMask & (1 << idx)
|
? positiveStride ? dim : -1
|
: ClampedIndex(endData[idx], dim, positiveStride);
|
|
// This is valid for both positive and negative strides
|
int32_t outDim = ceil((end - begin) / static_cast<float>(stride));
|
outDim = outDim < 0 ? 0 : static_cast<uint32_t>(outDim);
|
if (!(shrinkAxisMask & (1 << idx))) {
|
outDims.push_back(outDim);
|
} else {
|
if (outDim != 1) {
|
LOG(ERROR) << "Outdim " << idx << " is " << outDim << ", expected 1";
|
NN_OPS_CHECK(outDim == 1);
|
}
|
}
|
}
|
|
output->type = input.type;
|
output->dimensions = outDims;
|
output->offset = input.offset;
|
output->scale = input.scale;
|
|
return true;
|
}
|
|
bool argMinMaxPrepare(const Shape& input, int32_t axis, Shape* output) {
|
NN_CHECK(handleNegativeAxis(input, &axis));
|
|
output->type = OperandType::TENSOR_INT32;
|
|
// Copy the input dimensions, omitting the axis dimension.
|
output->dimensions.clear();
|
output->dimensions.reserve(getNumberOfDimensions(input) - 1);
|
output->dimensions.insert(output->dimensions.end(),
|
input.dimensions.begin(),
|
input.dimensions.begin() + axis);
|
output->dimensions.insert(output->dimensions.end(),
|
input.dimensions.begin() + axis + 1,
|
input.dimensions.end());
|
|
return true;
|
}
|
|
bool splitPrepare(const Shape& input, int32_t axis, int32_t numOutputs,
|
std::vector<Shape>* output) {
|
NN_CHECK(handleNegativeAxis(input, &axis));
|
|
const int32_t sizeOfAxisToSplit = input.dimensions[axis];
|
NN_OPS_CHECK(sizeOfAxisToSplit % numOutputs == 0);
|
const int32_t sliceSize = sizeOfAxisToSplit / numOutputs;
|
|
for (int i = 0; i < numOutputs; ++i) {
|
output->at(i).type = input.type;
|
output->at(i).dimensions = input.dimensions;
|
output->at(i).dimensions[axis] = sliceSize;
|
output->at(i).offset = input.offset;
|
output->at(i).scale = input.scale;
|
}
|
return true;
|
}
|
|
bool groupedConvPrepare(const Shape& input, const Shape& filter, const Shape& bias,
|
int32_t padding_left, int32_t padding_right, int32_t padding_top,
|
int32_t padding_bottom, int32_t stride_width, int32_t stride_height,
|
int32_t numGroups, Shape* output) {
|
if (filter.type == OperandType::TENSOR_QUANT8_SYMM_PER_CHANNEL) {
|
NN_OPS_CHECK(input.type == OperandType::TENSOR_QUANT8_ASYMM);
|
} else {
|
NN_OPS_CHECK(input.type == filter.type);
|
}
|
if (input.type == OperandType::TENSOR_QUANT8_ASYMM) {
|
NN_OPS_CHECK(bias.type == OperandType::TENSOR_INT32);
|
} else {
|
NN_OPS_CHECK(input.type == bias.type);
|
}
|
NN_OPS_CHECK(getNumberOfDimensions(input) == 4);
|
NN_OPS_CHECK(getNumberOfDimensions(filter) == 4);
|
NN_OPS_CHECK(getNumberOfDimensions(bias) == 1);
|
|
NN_OPS_CHECK(getSizeOfDimension(filter, 0) == getSizeOfDimension(bias, 0));
|
|
NN_OPS_CHECK(getSizeOfDimension(filter, 3) * numGroups == getSizeOfDimension(input, 3));
|
NN_OPS_CHECK(getSizeOfDimension(filter, 0) % numGroups == 0);
|
|
uint32_t channels_out = getSizeOfDimension(filter, 0);
|
uint32_t width = getSizeOfDimension(input, 2);
|
uint32_t height = getSizeOfDimension(input, 1);
|
uint32_t filterWidth = getSizeOfDimension(filter, 2);
|
uint32_t filterHeight = getSizeOfDimension(filter, 1);
|
uint32_t batches = getSizeOfDimension(input, 0);
|
|
NN_RET_CHECK_GT(static_cast<int32_t>(filterWidth), padding_left);
|
NN_RET_CHECK_GT(static_cast<int32_t>(filterWidth), padding_right);
|
NN_RET_CHECK_GT(static_cast<int32_t>(filterHeight), padding_top);
|
NN_RET_CHECK_GT(static_cast<int32_t>(filterHeight), padding_bottom);
|
|
uint32_t outWidth =
|
computeOutSize(width, filterWidth, stride_width, padding_left, padding_right);
|
uint32_t outHeight =
|
computeOutSize(height, filterHeight, stride_height, padding_top, padding_bottom);
|
|
output->type = input.type;
|
output->dimensions = {batches, outHeight, outWidth, channels_out};
|
return true;
|
}
|
|
} // namespace nn
|
} // namespace android
|