1
// Tencent is pleased to support the open source community by making ncnn available.
3
// Copyright (C) 2019 THL A29 Limited, a Tencent company. All rights reserved.
5
// Licensed under the BSD 3-Clause License (the "License"); you may not use this file except
6
// in compliance with the License. You may obtain a copy of the License at
8
// https://opensource.org/licenses/BSD-3-Clause
10
// Unless required by applicable law or agreed to in writing, software distributed
11
// under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR
12
// CONDITIONS OF ANY KIND, either express or implied. See the License for the
13
// specific language governing permissions and limitations under the License.
15
#include "batchnorm_vulkan.h"
17
#include "layer_shader_type.h"
21
BatchNorm_vulkan::BatchNorm_vulkan()
23
support_vulkan = true;
24
support_image_storage = true;
26
pipeline_batchnorm = 0;
27
pipeline_batchnorm_pack4 = 0;
28
pipeline_batchnorm_pack8 = 0;
31
int BatchNorm_vulkan::create_pipeline(const Option& opt)
33
const Mat& shape = top_shapes.empty() ? Mat() : top_shapes[0];
35
int elempack = opt.use_shader_pack8 && channels % 8 == 0 ? 8 : channels % 4 == 0 ? 4 : 1;
38
if (opt.use_fp16_storage)
40
elemsize = elempack * 2u;
42
else if (opt.use_fp16_packed)
44
elemsize = elempack == 1 ? 4u : elempack * 2u;
48
elemsize = elempack * 4u;
52
if (shape.dims == 1) shape_packed = Mat(shape.w / elempack, (void*)0, elemsize, elempack);
53
if (shape.dims == 2) shape_packed = Mat(shape.w, shape.h / elempack, (void*)0, elemsize, elempack);
54
if (shape.dims == 3) shape_packed = Mat(shape.w, shape.h, shape.c / elempack, (void*)0, elemsize, elempack);
55
if (shape.dims == 4) shape_packed = Mat(shape.w, shape.h, shape.d, shape.c / elempack, (void*)0, elemsize, elempack);
57
std::vector<vk_specialization_type> specializations(0 + 5);
58
specializations[0 + 0].i = std::min(3, shape_packed.dims);
59
specializations[0 + 1].i = shape_packed.w;
60
specializations[0 + 2].i = shape_packed.h * shape_packed.d;
61
specializations[0 + 3].i = shape_packed.c;
62
specializations[0 + 4].i = shape_packed.cstep;
64
Mat local_size_xyz(4, 4, std::min(4, channels / elempack), (void*)0);
65
if (shape_packed.dims == 1)
67
local_size_xyz.w = std::min(64, shape_packed.w);
71
if (shape_packed.dims == 2)
73
local_size_xyz.w = std::min(8, shape_packed.w);
74
local_size_xyz.h = std::min(8, shape_packed.h);
77
if (shape_packed.dims == 3)
79
local_size_xyz.w = std::min(4, shape_packed.w);
80
local_size_xyz.h = std::min(4, shape_packed.h);
81
local_size_xyz.c = std::min(4, shape_packed.c);
83
if (shape_packed.dims == 4)
85
local_size_xyz.w = std::min(4, shape_packed.w);
86
local_size_xyz.h = std::min(4, shape_packed.h * shape_packed.d);
87
local_size_xyz.c = std::min(4, shape_packed.c);
93
pipeline_batchnorm = new Pipeline(vkdev);
94
pipeline_batchnorm->set_optimal_local_size_xyz(local_size_xyz);
95
pipeline_batchnorm->create(LayerShaderType::batchnorm, opt, specializations);
101
pipeline_batchnorm_pack4 = new Pipeline(vkdev);
102
pipeline_batchnorm_pack4->set_optimal_local_size_xyz(local_size_xyz);
103
pipeline_batchnorm_pack4->create(LayerShaderType::batchnorm_pack4, opt, specializations);
109
pipeline_batchnorm_pack8 = new Pipeline(vkdev);
110
pipeline_batchnorm_pack8->set_optimal_local_size_xyz(local_size_xyz);
111
pipeline_batchnorm_pack8->create(LayerShaderType::batchnorm_pack8, opt, specializations);
117
int BatchNorm_vulkan::destroy_pipeline(const Option& /*opt*/)
119
delete pipeline_batchnorm;
120
pipeline_batchnorm = 0;
122
delete pipeline_batchnorm_pack4;
123
pipeline_batchnorm_pack4 = 0;
125
delete pipeline_batchnorm_pack8;
126
pipeline_batchnorm_pack8 = 0;
131
int BatchNorm_vulkan::upload_model(VkTransfer& cmd, const Option& opt)
133
int elempack = opt.use_shader_pack8 && channels % 8 == 0 ? 8 : channels % 4 == 0 ? 4 : 1;
136
convert_packing(a_data, a_data_packed, elempack, opt);
138
if (opt.use_image_storage)
140
cmd.record_upload(a_data_packed, a_data_gpu_image, opt);
144
cmd.record_upload(a_data_packed, a_data_gpu, opt);
148
convert_packing(b_data, b_data_packed, elempack, opt);
150
if (opt.use_image_storage)
152
cmd.record_upload(b_data_packed, b_data_gpu_image, opt);
156
cmd.record_upload(b_data_packed, b_data_gpu, opt);
168
int BatchNorm_vulkan::forward_inplace(VkMat& bottom_top_blob, VkCompute& cmd, const Option& /*opt*/) const
170
int elempack = bottom_top_blob.elempack;
172
std::vector<VkMat> bindings(3);
173
bindings[0] = bottom_top_blob;
174
bindings[1] = a_data_gpu;
175
bindings[2] = b_data_gpu;
177
std::vector<vk_constant_type> constants(5);
178
constants[0].i = std::min(3, bottom_top_blob.dims);
179
constants[1].i = bottom_top_blob.w;
180
constants[2].i = bottom_top_blob.h * bottom_top_blob.d;
181
constants[3].i = bottom_top_blob.c;
182
constants[4].i = bottom_top_blob.cstep;
184
const Pipeline* pipeline = elempack == 8 ? pipeline_batchnorm_pack8
185
: elempack == 4 ? pipeline_batchnorm_pack4
186
: pipeline_batchnorm;
188
cmd.record_pipeline(pipeline, bindings, constants, bottom_top_blob);
193
int BatchNorm_vulkan::forward_inplace(VkImageMat& bottom_top_blob, VkCompute& cmd, const Option& /*opt*/) const
195
int elempack = bottom_top_blob.elempack;
197
std::vector<VkImageMat> bindings(4);
198
bindings[0] = bottom_top_blob;
199
bindings[1] = bottom_top_blob;
200
bindings[2] = a_data_gpu_image;
201
bindings[3] = b_data_gpu_image;
203
std::vector<vk_constant_type> constants(5);
204
constants[0].i = std::min(3, bottom_top_blob.dims);
205
constants[1].i = bottom_top_blob.w;
206
constants[2].i = bottom_top_blob.h * bottom_top_blob.d;
207
constants[3].i = bottom_top_blob.c;
208
constants[4].i = 0; //bottom_top_blob.cstep;
210
const Pipeline* pipeline = elempack == 8 ? pipeline_batchnorm_pack8
211
: elempack == 4 ? pipeline_batchnorm_pack4
212
: pipeline_batchnorm;
214
cmd.record_pipeline(pipeline, bindings, constants, bottom_top_blob);