source/backend/metal/MetalReduction.mm (85 lines of code) (raw):

// // MetalReduction.mm // MNN // // Created by MNN on 2019/01/30. // Copyright © 2018, Alibaba Group Holding Limited // #import "backend/metal/MetalReduction.hpp" #import "backend/metal/MNNMetalContext.h" #import "core/Macro.h" #import "core/Macro.h" #import "backend/metal/MetalBackend.hpp" #import "core/TensorUtils.hpp" #if MNN_METAL_ENABLED namespace MNN { MetalReduction::MetalReduction(Backend *backend, const ReductionParam *p, halide_type_t type) : MetalExecution(backend) { auto integer = type.code == halide_type_int; NSString *kernel; switch (p->operation()) { case ReductionType_SUM: kernel = integer ? @"reduce_sum_s" : @"reduce_sum_f"; break; case ReductionType_ASUM: case ReductionType_SUMSQ: MNN_ASSERT(false); // both un-supported break; case ReductionType_MEAN: kernel = integer ? @"reduce_mean_s" : @"reduce_mean_f"; break; case ReductionType_MAXIMUM: kernel = integer ? @"reduce_max_s" : @"reduce_max_f"; break; case ReductionType_MINIMUM: kernel = integer ? @"reduce_min_s" : @"reduce_min_f"; break; case ReductionType_PROD: kernel = integer ? @"reduce_prod_s" : @"reduce_prod_f"; break; default: break; } // The reduce after geometry compute has only one axis mAxis = p->dim()->data()[0]; auto mkbn = static_cast<MetalBackend *>(backend); auto context = (__bridge MNNMetalContext *)mkbn->context(); mConst = [context newDeviceBuffer:4 * sizeof(int) access:CPUWriteOnly]; mPipeline = [context pipelineWithName:kernel fp16:mkbn->useFp16InsteadFp32()]; } ErrorCode MetalReduction::onResize(const std::vector<Tensor *> &inputs, const std::vector<Tensor *> &outputs) { int outsideSize = 1, axisSize = 1, insideSize = 1; for (int i = 0; i < mAxis; i++) { outsideSize *= inputs[0]->length(i); } axisSize = inputs[0]->length(mAxis); for (int i = mAxis + 1; i < inputs[0]->dimensions(); i++) { insideSize *= inputs[0]->length(i); } auto backend = static_cast<MetalBackend *>(this->backend()); auto context = (__bridge MNNMetalContext *)backend->context(); ((int *)mConst.contents)[0] = outsideSize; ((int *)mConst.contents)[1] = axisSize; ((int *)mConst.contents)[2] = insideSize; ((int *)mConst.contents)[3] = axisSize * insideSize; mThreads = [context computeBestGroupAndLocal:mPipeline threads:MTLSizeMake(outsideSize, insideSize, 1)]; return NO_ERROR; } void MetalReduction::onEncode(const std::vector<Tensor *> &inputs, const std::vector<Tensor *> &outputs, id<MTLComputeCommandEncoder> encoder) { auto &input = inputs[0], &output = outputs[0]; [encoder setComputePipelineState:mPipeline]; [encoder setBuffer:(id<MTLBuffer>)((MetalRuntimeAllocator::MetalBufferAlloc *)input->deviceId())->getBuffer() offset:TensorUtils::getDescribe(input)->extra.offset atIndex:0]; [encoder setBuffer:(id<MTLBuffer>)((MetalRuntimeAllocator::MetalBufferAlloc *)output->deviceId())->getBuffer() offset:TensorUtils::getDescribe(output)->extra.offset atIndex:1]; [encoder setBuffer:mConst offset:0 atIndex:2]; [encoder dispatchThreadgroups:mThreads.first threadsPerThreadgroup:mThreads.second]; } class MetalReductionCreator : public MetalBackend::Creator { public: virtual Execution *onCreate(const std::vector<Tensor *> &inputs, const MNN::Op *op, Backend *backend, const std::vector<Tensor *>& outputs) const { auto param = op->main_as_ReductionParam(); switch (param->operation()) { case ReductionType_ALL: case ReductionType_ANY: case ReductionType_ASUM: case ReductionType_SUMSQ: return nullptr; default: break; }; return new MetalReduction(backend, op->main_as_ReductionParam(), inputs[0]->getType()); } }; REGISTER_METAL_OP_CREATOR(MetalReductionCreator, OpType_Reduction); } // namespace MNN #endif /* MNN_METAL_ENABLED */