1
// Tencent is pleased to support the open source community by making ncnn available.
3
// Copyright (C) 2019 THL A29 Limited, a Tencent company. All rights reserved.
5
// Licensed under the BSD 3-Clause License (the "License"); you may not use this file except
6
// in compliance with the License. You may obtain a copy of the License at
8
// https://opensource.org/licenses/BSD-3-Clause
10
// Unless required by applicable law or agreed to in writing, software distributed
11
// under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR
12
// CONDITIONS OF ANY KIND, either express or implied. See the License for the
13
// specific language governing permissions and limitations under the License.
18
#extension GL_EXT_shader_16bit_storage: require
20
#if NCNN_fp16_arithmetic
21
#extension GL_EXT_shader_explicit_arithmetic_types_float16: require
24
#extension GL_GOOGLE_include_directive: enable
25
#include "vulkan_activation.comp"
27
#define LOCAL_MEMORY_UNROLL_INCH 8
29
layout (constant_id = 0) const int bias_term = 0;
30
layout (constant_id = 1) const int activation_type = 0;
31
layout (constant_id = 2) const float activation_param_0 = 0;
32
layout (constant_id = 3) const float activation_param_1 = 0;
34
#define shape_constant_id_offset 4
35
layout (constant_id = shape_constant_id_offset + 0) const int w = 0;
36
layout (constant_id = shape_constant_id_offset + 1) const int h = 0;
37
layout (constant_id = shape_constant_id_offset + 2) const int c = 0;
38
layout (constant_id = shape_constant_id_offset + 3) const int cstep = 0;
40
layout (constant_id = shape_constant_id_offset + 4) const int outw = 0;
41
layout (constant_id = shape_constant_id_offset + 5) const int outh = 0;
42
layout (constant_id = shape_constant_id_offset + 6) const int outc = 0;
43
layout (constant_id = shape_constant_id_offset + 7) const int outcstep = 0;
46
layout (binding = 0) uniform unfp sampler3D bottom_blob;
47
layout (binding = 1, imfmtc4) writeonly uniform unfp image3D top_blob;
48
layout (binding = 2) uniform unfp sampler3D weight_blob;
49
layout (binding = 3) uniform unfp sampler3D bias_blob;
51
layout (binding = 0) readonly buffer bottom_blob { sfpvec4 bottom_blob_data[]; };
52
layout (binding = 1) writeonly buffer top_blob { sfpvec4 top_blob_data[]; };
53
layout (binding = 2) readonly buffer weight_blob { sfpvec4 weight_data[]; };
54
layout (binding = 3) readonly buffer bias_blob { sfpvec4 bias_data[]; };
57
layout (push_constant) uniform parameter
70
#if NCNN_shader_local_memory
71
shared lfpvec4 tmp_v[8][LOCAL_MEMORY_UNROLL_INCH][4];
72
shared lfpvec4 tmp_k[8][LOCAL_MEMORY_UNROLL_INCH][4];
77
int gx = int(gl_GlobalInvocationID.x) * 4;
78
int gy = int(gl_GlobalInvocationID.y);
81
if (gx >= psc(outw) * psc(outh) || gy >= psc(outc))
84
#if !NCNN_shader_local_memory
85
if (gx >= psc(outcstep) || gy >= psc(outc))
98
afpvec4 b = image3d_ld4(bias_blob, ivec3(gy, 0, 0));
100
afpvec4 b = buffer_ld4(bias_data, gy);
116
ivec4 gx4 = gx + ivec4(0, 1, 2, 3);
118
ivec4 sy4 = gx4 / psc(w);
119
ivec4 sx4 = gx4 % psc(w);
121
for (int z = 0; z < psc(c); z++)
123
afpvec4 v0 = image3d_ld4(bottom_blob, ivec3(sx4.r, sy4.r, z));
124
afpvec4 v1 = image3d_ld4(bottom_blob, ivec3(sx4.g, sy4.g, z));
125
afpvec4 v2 = image3d_ld4(bottom_blob, ivec3(sx4.b, sy4.b, z));
126
afpvec4 v3 = image3d_ld4(bottom_blob, ivec3(sx4.a, sy4.a, z));
129
image3d_ld4(weight_blob, ivec3(0, z, gy)),
130
image3d_ld4(weight_blob, ivec3(1, z, gy)),
131
image3d_ld4(weight_blob, ivec3(2, z, gy)),
132
image3d_ld4(weight_blob, ivec3(3, z, gy))
141
int w_offset = gy * psc(c) * 4;
144
#if NCNN_shader_local_memory
145
const int lx = int(gl_LocalInvocationID.x);
146
const int ly = int(gl_LocalInvocationID.y);
149
for (; z + (LOCAL_MEMORY_UNROLL_INCH - 1) < psc(c); z += LOCAL_MEMORY_UNROLL_INCH)
153
for (int z4 = 0; z4 < LOCAL_MEMORY_UNROLL_INCH; z4++)
155
tmp_v[lx][z4][ly] = sfp2lfpvec4(bottom_blob_data[v_offset + z4 * psc(cstep) + ly]);
161
for (int z4 = 0; z4 < LOCAL_MEMORY_UNROLL_INCH; z4++)
163
tmp_k[ly][z4][lx] = sfp2lfpvec4(weight_data[w_offset + z4 * 4 + lx]);
169
for (int z4 = 0; z4 < LOCAL_MEMORY_UNROLL_INCH; z4++)
171
afpvec4 v0 = lfp2afpvec4(tmp_v[lx][z4][0]);
172
afpvec4 v1 = lfp2afpvec4(tmp_v[lx][z4][1]);
173
afpvec4 v2 = lfp2afpvec4(tmp_v[lx][z4][2]);
174
afpvec4 v3 = lfp2afpvec4(tmp_v[lx][z4][3]);
176
afpvec4 k0 = lfp2afpvec4(tmp_k[ly][z4][0]);
177
afpvec4 k1 = lfp2afpvec4(tmp_k[ly][z4][1]);
178
afpvec4 k2 = lfp2afpvec4(tmp_k[ly][z4][2]);
179
afpvec4 k3 = lfp2afpvec4(tmp_k[ly][z4][3]);
181
afpmat4 k = afpmat4(k0, k1, k2, k3);
189
v_offset += LOCAL_MEMORY_UNROLL_INCH * psc(cstep);
190
w_offset += LOCAL_MEMORY_UNROLL_INCH * 4;
197
const int remain = psc(c) - z;
201
for (int z4 = 0; z4 < remain; z4++)
203
tmp_v[lx][z4][ly] = sfp2lfpvec4(bottom_blob_data[v_offset + z4 * psc(cstep) + ly]);
209
for (int z4 = 0; z4 < remain; z4++)
211
tmp_k[ly][z4][lx] = sfp2lfpvec4(weight_data[w_offset + z4 * 4 + lx]);
217
for (int z4 = 0; z4 < remain; z4++)
219
afpvec4 v0 = lfp2afpvec4(tmp_v[lx][z4][0]);
220
afpvec4 v1 = lfp2afpvec4(tmp_v[lx][z4][1]);
221
afpvec4 v2 = lfp2afpvec4(tmp_v[lx][z4][2]);
222
afpvec4 v3 = lfp2afpvec4(tmp_v[lx][z4][3]);
224
afpvec4 k0 = lfp2afpvec4(tmp_k[ly][z4][0]);
225
afpvec4 k1 = lfp2afpvec4(tmp_k[ly][z4][1]);
226
afpvec4 k2 = lfp2afpvec4(tmp_k[ly][z4][2]);
227
afpvec4 k3 = lfp2afpvec4(tmp_k[ly][z4][3]);
229
afpmat4 k = afpmat4(k0, k1, k2, k3);
238
for (int z = 0; z < psc(c); z++)
240
afpvec4 v0 = buffer_ld4(bottom_blob_data, v_offset + 0);
241
afpvec4 v1 = buffer_ld4(bottom_blob_data, v_offset + 1);
242
afpvec4 v2 = buffer_ld4(bottom_blob_data, v_offset + 2);
243
afpvec4 v3 = buffer_ld4(bottom_blob_data, v_offset + 3);
246
buffer_ld4(weight_data, w_offset + 0),
247
buffer_ld4(weight_data, w_offset + 1),
248
buffer_ld4(weight_data, w_offset + 2),
249
buffer_ld4(weight_data, w_offset + 3)
258
v_offset += psc(cstep);
263
sum0 = activation_afpvec4(sum0, activation_type, activation_param_0, activation_param_1);
264
sum1 = activation_afpvec4(sum1, activation_type, activation_param_0, activation_param_1);
265
sum2 = activation_afpvec4(sum2, activation_type, activation_param_0, activation_param_1);
266
sum3 = activation_afpvec4(sum3, activation_type, activation_param_0, activation_param_1);
269
image3d_st4(top_blob, ivec3(sx4.r, sy4.r, gy), sum0);
270
image3d_st4(top_blob, ivec3(sx4.g, sy4.g, gy), sum1);
271
image3d_st4(top_blob, ivec3(sx4.b, sy4.b, gy), sum2);
272
image3d_st4(top_blob, ivec3(sx4.a, sy4.a, gy), sum3);
274
#if NCNN_shader_local_memory
275
if (gx >= psc(outcstep) || gy >= psc(outc))
279
int gi = gy * psc(outcstep) + gx;
281
buffer_st4(top_blob_data, gi + 0, sum0);
282
if (gx + 1 < psc(outcstep)) buffer_st4(top_blob_data, gi + 1, sum1);
283
if (gx + 2 < psc(outcstep)) buffer_st4(top_blob_data, gi + 2, sum2);
284
if (gx + 3 < psc(outcstep)) buffer_st4(top_blob_data, gi + 3, sum3);