Bug Summary

File:nnc/ccv_cnnp_model_addons.c
Warning:line 400, column 3
Declared variable-length array (VLA) has negative size

Annotated Source Code

Press '?' to see keyboard shortcuts

clang -cc1 -cc1 -triple x86_64-unknown-linux-gnu -analyze -disable-free -clear-ast-before-backend -disable-llvm-verifier -discard-value-names -main-file-name ccv_cnnp_model_addons.c -analyzer-checker=core -analyzer-checker=apiModeling -analyzer-checker=unix -analyzer-checker=deadcode -analyzer-checker=security.insecureAPI.UncheckedReturn -analyzer-checker=security.insecureAPI.getpw -analyzer-checker=security.insecureAPI.gets -analyzer-checker=security.insecureAPI.mktemp -analyzer-checker=security.insecureAPI.mkstemp -analyzer-checker=security.insecureAPI.vfork -analyzer-checker=nullability.NullPassedToNonnull -analyzer-checker=nullability.NullReturnedFromNonnull -analyzer-output plist -w -setup-static-analyzer -mrelocation-model pic -pic-level 2 -pic-is-pie -mframe-pointer=none -fmath-errno -ffp-contract=on -fno-rounding-math -mconstructor-aliases -funwind-tables=2 -target-cpu x86-64 -target-feature +sse2 -tune-cpu generic -debugger-tuning=gdb -fdebug-compilation-dir=/home/liu/actions-runner/_work/ccv/ccv/lib/nnc -fcoverage-compilation-dir=/home/liu/actions-runner/_work/ccv/ccv/lib/nnc -resource-dir /usr/local/lib/clang/19 -I ../ -I /usr/local/cuda/include -D HAVE_CBLAS -D HAVE_LIBPNG -D HAVE_LIBJPEG -D HAVE_FFTW3 -D HAVE_PTHREAD -D HAVE_LIBLINEAR -D HAVE_TESSERACT -D HAVE_AVCODEC -D HAVE_AVFORMAT -D HAVE_AVUTIL -D HAVE_SWSCALE -D HAVE_SSE2 -D HAVE_GSL -D HAVE_CUDA -D HAVE_CUDNN -D HAVE_NCCL -D USE_SYSTEM_CUB -I /usr/local/include -internal-isystem /usr/local/lib/clang/19/include -internal-isystem /usr/local/include -internal-isystem /usr/lib/gcc/x86_64-linux-gnu/12/../../../../x86_64-linux-gnu/include -internal-externc-isystem /usr/include/x86_64-linux-gnu -internal-externc-isystem /include -internal-externc-isystem /usr/include -O3 -ferror-limit 19 -fgnuc-version=4.2.1 -fskip-odr-check-in-gmf -vectorize-loops -vectorize-slp -analyzer-output=html -faddrsig -D__GCC_HAVE_DWARF2_CFI_ASM=1 -o /home/liu/actions-runner/_work/ccv/ccv/_analyze/2026-04-22-160924-948331-1 -x c ccv_cnnp_model_addons.c
1#include "ccv_nnc.h"
2#include "ccv_nnc_easy.h"
3#include "ccv_nnc_internal.h"
4#include "ccv_internal.h"
5#include "_ccv_cnnp_model.h"
6
7// MARK - Add-on Functions
8
9static int _ccv_cnnp_model_clip_grad_norm_reduce_norm2(const ccv_nnc_cmd_t cmd, const ccv_nnc_hint_t hint, const int flags, ccv_nnc_tensor_t* const* const inputs, const int input_size, ccv_nnc_tensor_t* const* const outputs, const int output_size, ccv_nnc_stream_context_t* const stream_context)
10{
11 const int device_id = CCV_TENSOR_GET_DEVICE_ID(inputs[0]->info.type)(((inputs[0]->info.type) & 0xfff00) >> 8);
12 ccv_nnc_tensor_t* const old_norm2 = outputs[1 + device_id * 2];
13 ccv_nnc_tensor_t* const norm2 = outputs[1 + device_id * 2 + 1];
14 const int tensor_count = ccv_nnc_tensor_count(inputs[0]->info);
15 if (tensor_count == 1)
16 ccv_nnc_cmd_exec(CMD_MUL_FORWARD(1)ccv_nnc_cmd(CCV_NNC_MUL_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={1,}}}, 0)
, hint, flags, TENSOR_LIST(inputs[0], inputs[0])(ccv_nnc_tensor_t* []){inputs[0], inputs[0]}, (1 +1 +1 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_LIST(norm2)(ccv_nnc_tensor_t* []){norm2}, (1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_context);
17 else {
18 ccv_nnc_cmd_exec(CMD_REDUCE_NORM2_FORWARD()ccv_nnc_cmd(CCV_NNC_REDUCE_NORM2_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.reduce={.count=(1 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1),.axis={}}}
), 0)
, hint, flags, TENSOR_LIST(inputs[0])(ccv_nnc_tensor_t* []){inputs[0]}, (1 +1 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_LIST(norm2)(ccv_nnc_tensor_t* []){norm2}, (1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_context);
19 ccv_nnc_cmd_exec(CMD_MUL_FORWARD(1)ccv_nnc_cmd(CCV_NNC_MUL_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={1,}}}, 0)
, hint, flags, TENSOR_LIST(norm2, norm2)(ccv_nnc_tensor_t* []){norm2, norm2}, (1 +1 +1 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_LIST(norm2)(ccv_nnc_tensor_t* []){norm2}, (1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_context);
20 }
21 ccv_nnc_cmd_exec(CMD_ADD_FORWARD(1, 1)ccv_nnc_cmd(CCV_NNC_ADD_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={1, 1}}}, 0)
, hint, flags, TENSOR_LIST(old_norm2, norm2)(ccv_nnc_tensor_t* []){old_norm2, norm2}, (1 +1 +1 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_LIST(old_norm2)(ccv_nnc_tensor_t* []){old_norm2}, (1 +1 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_context);
22 return CCV_NNC_EXEC_SUCCESS;
23}
24
25static ccv_nnc_cmd_vtab_t clip_grad_norm_reduce_norm2_vtab = {
26 .exec = _ccv_cnnp_model_clip_grad_norm_reduce_norm2
27};
28
29static int _ccv_cnnp_model_clip_grad_norm_scatter_norm2(const ccv_nnc_cmd_t cmd, const ccv_nnc_hint_t hint, const int flags, ccv_nnc_tensor_t* const* const inputs, const int input_size, ccv_nnc_tensor_t* const* const outputs, const int output_size, ccv_nnc_stream_context_t* const stream_context)
30{
31 const int device_id = CCV_TENSOR_GET_DEVICE_ID(inputs[0]->info.type)(((inputs[0]->info.type) & 0xfff00) >> 8);
32 ccv_nnc_tensor_t* const norm2 = inputs[1 + device_id * 2];
33 ccv_nnc_cmd_exec(CMD_MUL_FORWARD(1)ccv_nnc_cmd(CCV_NNC_MUL_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={1,}}}, 0)
, hint, flags, TENSOR_LIST(inputs[0], norm2)(ccv_nnc_tensor_t* []){inputs[0], norm2}, (1 +1 +1 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_LIST(outputs[0])(ccv_nnc_tensor_t* []){outputs[0]}, (1 +1 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_context);
34 return CCV_NNC_EXEC_SUCCESS;
35}
36
37static ccv_nnc_cmd_vtab_t clip_grad_norm_scatter_norm2_vtab = {
38 .exec = _ccv_cnnp_model_clip_grad_norm_scatter_norm2
39};
40
41void ccv_cnnp_model_parameters_clip_grad_norm(ccv_cnnp_model_t* const model, const ccv_cnnp_model_io_t parameters, int norm_type, float max_norm, ccv_nnc_stream_context_t* const stream_context)
42{
43 assert(norm_type == 2)((void) sizeof ((norm_type == 2) ? 1 : 0), __extension__ ({ if
(norm_type == 2) ; else __assert_fail ("norm_type == 2", "ccv_cnnp_model_addons.c"
, 43, __extension__ __PRETTY_FUNCTION__); }))
;
44 ccv_cnnp_compiled_data_t* const compiled_data = model->compiled_data;
45 assert(compiled_data)((void) sizeof ((compiled_data) ? 1 : 0), __extension__ ({ if
(compiled_data) ; else __assert_fail ("compiled_data", "ccv_cnnp_model_addons.c"
, 45, __extension__ __PRETTY_FUNCTION__); }))
;
46 const int parallel_count = ccv_max(model->parallel_count, 1)({ typeof (model->parallel_count) _a = (model->parallel_count
); typeof (1) _b = (1); (_a > _b) ? _a : _b; })
;
47 ccv_nnc_tensor_t* norm2[parallel_count * 2];
48 ccv_nnc_tensor_t* max_normt[parallel_count];
49 const int stream_type = model->compiled_data->stream_type;
50 int i;
51 if (stream_type == CCV_STREAM_CONTEXT_GPU)
52 {
53 for (i = 0; i < parallel_count; i++)
54 {
55 ccv_nnc_tensor_param_t info = {
56 .type = CCV_TENSOR_GPU_MEMORY,
57 .format = CCV_TENSOR_FORMAT_NHWC,
58 .datatype = CCV_32F,
59 .dim = {1},
60 };
61 CCV_TENSOR_SET_DEVICE_ID(info.type, i)(info.type) = (((info.type) & ~0xfff00) | (((i) & 0xfff
) << 8))
;
62 norm2[i * 2] = ccv_nnc_tensor_new(ccv_nnc_xpu_alloc(&compiled_data->xpu_alloc, i, stream_context, ccv_nnc_tensor_data_size(info)), info, 0);
63 norm2[i * 2 + 1] = ccv_nnc_tensor_new(ccv_nnc_xpu_alloc(&compiled_data->xpu_alloc, i, stream_context, ccv_nnc_tensor_data_size(info)), info, 0);
64 max_normt[i] = ccv_nnc_tensor_new(ccv_nnc_xpu_alloc(&compiled_data->xpu_alloc, i, stream_context, ccv_nnc_tensor_data_size(info)), info, 0);
65 }
66 } else {
67 for (i = 0; i < parallel_count; i++)
68 {
69 ccv_nnc_tensor_param_t info = {
70 .type = CCV_TENSOR_CPU_MEMORY,
71 .format = CCV_TENSOR_FORMAT_NHWC,
72 .datatype = CCV_32F,
73 .dim = {1},
74 };
75 norm2[i * 2] = ccv_nnc_tensor_new(0, info, 0);
76 norm2[i * 2 + 1] = ccv_nnc_tensor_new(0, info, 0);
77 max_normt[i] = ccv_nnc_tensor_new(0, info, 0);
78 }
79 }
80 // zero out old norm2.
81 if (parallel_count > 1)
82 {
83 ccv_nnc_stream_context_t* streams[parallel_count];
84 ccv_nnc_stream_signal_t* signal;
85 if (stream_context)
86 signal = ccv_nnc_stream_context_emit_signal_new(stream_context);
87 for (i = 0; i < parallel_count; i++)
88 {
89 const int stream_type = CCV_TENSOR_GET_MEMORY(norm2[i * 2]->info.type)((norm2[i * 2]->info.type) & 0x3) == CCV_TENSOR_GPU_MEMORY ? CCV_STREAM_CONTEXT_GPU : CCV_STREAM_CONTEXT_CPU;
90 const int device_id = CCV_TENSOR_GET_DEVICE_ID(norm2[i * 2]->info.type)(((norm2[i * 2]->info.type) & 0xfff00) >> 8);
91 int type = stream_type;
92 CCV_STREAM_SET_DEVICE_ID(type, device_id)(type) = (((type) & ~0xfff00) | (((device_id) & 0xfff
) << 8))
;
93 ccv_nnc_stream_context_t* const stream_0 = ccv_cnnp_compiled_data_get_stream(compiled_data, type);
94 // Wait signal to finish.
95 if (stream_context)
96 ccv_nnc_stream_context_wait_signal(stream_0, signal);
97 ccv_nnc_cmd_exec(CMD_SET_FORWARD(0)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={0,}}}, 0)
, ccv_nnc_no_hint, 0, 0, 0, TENSOR_LIST(norm2[i * 2])(ccv_nnc_tensor_t* []){norm2[i * 2]}, (1 +1 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_0);
98 if (stream_context)
99 {
100 ccv_nnc_stream_signal_t* const signal = ccv_nnc_stream_context_emit_signal_new(stream_0);
101 ccv_nnc_stream_context_wait_signal(stream_context, signal);
102 }
103 streams[i] = stream_0;
104 }
105 // If this should be blocking, blocking it.
106 if (!stream_context)
107 for (i = 0; i < parallel_count; i++)
108 if (streams[i])
109 ccv_nnc_stream_context_wait(streams[i]);
110 } else {
111 ccv_nnc_cmd_exec(CMD_SET_FORWARD(0)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={0,}}}, 0)
, ccv_nnc_no_hint, 0, 0, 0, TENSOR_LIST(norm2[0])(ccv_nnc_tensor_t* []){norm2[0]}, (1 +1 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_context);
112 }
113 // Gather norm2.
114 ccv_nnc_cmd_t reduce_cmd = {
115 .cmd = CCV_NNC_CUSTOM_FORWARD,
116 .isa = &clip_grad_norm_reduce_norm2_vtab,
117 };
118 ccv_cnnp_model_parameter_gradients_map(model, parameters, reduce_cmd, ccv_nnc_no_hint, 0, 0, 0, norm2, parallel_count * 2, stream_context);
119 // Now compute max(max_norm / norm2, 1.0).
120 if (parallel_count > 1)
121 {
122 ccv_nnc_stream_context_t* streams[parallel_count];
123 ccv_nnc_stream_signal_t* signal;
124 if (stream_context)
125 signal = ccv_nnc_stream_context_emit_signal_new(stream_context);
126 for (i = 0; i < parallel_count; i++)
127 {
128 const int stream_type = CCV_TENSOR_GET_MEMORY(norm2[i * 2]->info.type)((norm2[i * 2]->info.type) & 0x3) == CCV_TENSOR_GPU_MEMORY ? CCV_STREAM_CONTEXT_GPU : CCV_STREAM_CONTEXT_CPU;
129 const int device_id = CCV_TENSOR_GET_DEVICE_ID(norm2[i * 2]->info.type)(((norm2[i * 2]->info.type) & 0xfff00) >> 8);
130 int type = stream_type;
131 CCV_STREAM_SET_DEVICE_ID(type, device_id)(type) = (((type) & ~0xfff00) | (((device_id) & 0xfff
) << 8))
;
132 ccv_nnc_stream_context_t* const stream_0 = ccv_cnnp_compiled_data_get_stream(compiled_data, type);
133 // Wait signal to finish.
134 if (stream_context)
135 ccv_nnc_stream_context_wait_signal(stream_0, signal);
136 ccv_nnc_cmd_exec(CMD_EWSQRT_FORWARD()ccv_nnc_cmd(CCV_NNC_EWSQRT_FORWARD, 0, ccv_nnc_cmd_auto, 0), ccv_nnc_no_hint, 0, TENSOR_LIST(norm2[i * 2])(ccv_nnc_tensor_t* []){norm2[i * 2]}, (1 +1 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_LIST(norm2[i * 2])(ccv_nnc_tensor_t* []){norm2[i * 2]}, (1 +1 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_0);
137 ccv_nnc_cmd_exec(CMD_SET_FORWARD(max_norm)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={max_norm,}}}, 0)
, ccv_nnc_no_hint, 0, 0, 0, TENSOR_LIST(max_normt[i])(ccv_nnc_tensor_t* []){max_normt[i]}, (1 +1 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_0);
138 ccv_nnc_cmd_exec(CMD_EWDIV_FORWARD()ccv_nnc_cmd(CCV_NNC_EWDIV_FORWARD, 0, ccv_nnc_cmd_auto, 0), ccv_nnc_no_hint, 0, TENSOR_LIST(max_normt[i], norm2[i * 2])(ccv_nnc_tensor_t* []){max_normt[i], norm2[i * 2]}, (1 +1 +1 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1
)
, TENSOR_LIST(norm2[i * 2])(ccv_nnc_tensor_t* []){norm2[i * 2]}, (1 +1 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_0);
139 ccv_nnc_cmd_exec(CMD_CLAMP_FORWARD(NAN, 1)ccv_nnc_cmd(CCV_NNC_CLAMP_FORWARD, 0, (ccv_nnc_cmd_param_t){.
size={.dim={1,1,1}},.clamp={.min=(__builtin_nanf ("")),.max=1
}}, 0)
, ccv_nnc_no_hint, 0, TENSOR_LIST(norm2[i * 2])(ccv_nnc_tensor_t* []){norm2[i * 2]}, (1 +1 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_LIST(norm2[i * 2])(ccv_nnc_tensor_t* []){norm2[i * 2]}, (1 +1 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_0);
140 if (stream_context)
141 {
142 ccv_nnc_stream_signal_t* const signal = ccv_nnc_stream_context_emit_signal_new(stream_0);
143 ccv_nnc_stream_context_wait_signal(stream_context, signal);
144 }
145 streams[i] = stream_0;
146 }
147 // If this should be blocking, blocking it.
148 if (!stream_context)
149 for (i = 0; i < parallel_count; i++)
150 if (streams[i])
151 ccv_nnc_stream_context_wait(streams[i]);
152 } else {
153 ccv_nnc_cmd_exec(CMD_EWSQRT_FORWARD()ccv_nnc_cmd(CCV_NNC_EWSQRT_FORWARD, 0, ccv_nnc_cmd_auto, 0), ccv_nnc_no_hint, 0, TENSOR_LIST(norm2[0])(ccv_nnc_tensor_t* []){norm2[0]}, (1 +1 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_LIST(norm2[0])(ccv_nnc_tensor_t* []){norm2[0]}, (1 +1 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_context);
154 ccv_nnc_cmd_exec(CMD_SET_FORWARD(max_norm)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={max_norm,}}}, 0)
, ccv_nnc_no_hint, 0, 0, 0, TENSOR_LIST(max_normt[0])(ccv_nnc_tensor_t* []){max_normt[0]}, (1 +1 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_context);
155 ccv_nnc_cmd_exec(CMD_EWDIV_FORWARD()ccv_nnc_cmd(CCV_NNC_EWDIV_FORWARD, 0, ccv_nnc_cmd_auto, 0), ccv_nnc_no_hint, 0, TENSOR_LIST(max_normt[0], norm2[0])(ccv_nnc_tensor_t* []){max_normt[0], norm2[0]}, (1 +1 +1 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_LIST(norm2[0])(ccv_nnc_tensor_t* []){norm2[0]}, (1 +1 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_context);
156 ccv_nnc_cmd_exec(CMD_CLAMP_FORWARD(NAN, 1)ccv_nnc_cmd(CCV_NNC_CLAMP_FORWARD, 0, (ccv_nnc_cmd_param_t){.
size={.dim={1,1,1}},.clamp={.min=(__builtin_nanf ("")),.max=1
}}, 0)
, ccv_nnc_no_hint, 0, TENSOR_LIST(norm2[0])(ccv_nnc_tensor_t* []){norm2[0]}, (1 +1 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_LIST(norm2[0])(ccv_nnc_tensor_t* []){norm2[0]}, (1 +1 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_context);
157 }
158 ccv_nnc_cmd_t scatter_cmd = {
159 .cmd = CCV_NNC_CUSTOM_FORWARD,
160 .isa = &clip_grad_norm_scatter_norm2_vtab,
161 };
162 ccv_cnnp_model_parameter_gradients_map(model, parameters, scatter_cmd, ccv_nnc_no_hint, 0, norm2, parallel_count * 2, 0, 0, stream_context);
163 if (stream_type == CCV_STREAM_CONTEXT_GPU)
164 for (i = 0; i < parallel_count; i++)
165 {
166 ccv_nnc_xpu_free(&compiled_data->xpu_alloc, norm2[i * 2]->data.u8);
167 ccv_nnc_xpu_free(&compiled_data->xpu_alloc, norm2[i * 2 + 1]->data.u8);
168 ccv_nnc_xpu_free(&compiled_data->xpu_alloc, max_normt[i]->data.u8);
169 }
170 for (i = 0; i < parallel_count; i++)
171 {
172 ccv_nnc_tensor_free(norm2[i * 2]);
173 ccv_nnc_tensor_free(norm2[i * 2 + 1]);
174 ccv_nnc_tensor_free(max_normt[i]);
175 }
176}
177
178// MARK - Add-on Functions
179
180static int _ccv_cnnp_model_isnan(const ccv_nnc_cmd_t cmd, const ccv_nnc_hint_t hint, const int flags, ccv_nnc_tensor_t* const* const inputs, const int input_size, ccv_nnc_tensor_t* const* const outputs, const int output_size, ccv_nnc_stream_context_t* const stream_context)
181{
182 const int device_id = CCV_TENSOR_GET_DEVICE_ID(inputs[0]->info.type)(((inputs[0]->info.type) & 0xfff00) >> 8);
183 ccv_nnc_tensor_t* const old_isnanr = outputs[1 + device_id * 2];
184 ccv_nnc_tensor_t* const isnanr = outputs[1 + device_id * 2 + 1];
185 ccv_nnc_cmd_t reduce_cmd = CMD_REDUCE_ISNAN_FORWARD()ccv_nnc_cmd(CCV_NNC_REDUCE_ISNAN_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.reduce={.count=(1 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1),.axis={}}}
), 0)
;
186 reduce_cmd.info.reduce.count = ccv_nnc_tensor_nd(inputs[0]->info.dim);
187 int i;
188 for (i = 0; i < cmd.info.reduce.count; i++)
189 reduce_cmd.info.reduce.axis[i] = i;
190 ccv_nnc_cmd_exec(reduce_cmd, hint, flags, TENSOR_LIST(inputs[0])(ccv_nnc_tensor_t* []){inputs[0]}, (1 +1 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_LIST(isnanr)(ccv_nnc_tensor_t* []){isnanr}, (1 +1 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_context);
191 ccv_nnc_cmd_exec(CMD_EWSUM_FORWARD()ccv_nnc_cmd(CCV_NNC_EWSUM_FORWARD, 0, ccv_nnc_cmd_auto, 0), hint, flags, TENSOR_LIST(old_isnanr, isnanr)(ccv_nnc_tensor_t* []){old_isnanr, isnanr}, (1 +1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_LIST(old_isnanr)(ccv_nnc_tensor_t* []){old_isnanr}, (1 +1 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_context);
192 return CCV_NNC_EXEC_SUCCESS;
193}
194
195static ccv_nnc_cmd_vtab_t reduce_isnan_vtab = {
196 .exec = _ccv_cnnp_model_isnan
197};
198
199int ccv_cnnp_model_parameter_gradients_isnan(ccv_cnnp_model_t* const model, const ccv_cnnp_model_io_t parameters, ccv_nnc_stream_context_t* const stream_context)
200{
201 ccv_cnnp_compiled_data_t* const compiled_data = model->compiled_data;
202 assert(compiled_data)((void) sizeof ((compiled_data) ? 1 : 0), __extension__ ({ if
(compiled_data) ; else __assert_fail ("compiled_data", "ccv_cnnp_model_addons.c"
, 202, __extension__ __PRETTY_FUNCTION__); }))
;
203 const int parallel_count = ccv_max(model->parallel_count, 1)({ typeof (model->parallel_count) _a = (model->parallel_count
); typeof (1) _b = (1); (_a > _b) ? _a : _b; })
;
204 ccv_nnc_tensor_t* isnanr[parallel_count * 2];
205 const int stream_type = model->compiled_data->stream_type;
206 int i;
207 if (stream_type == CCV_STREAM_CONTEXT_GPU)
208 {
209 for (i = 0; i < parallel_count; i++)
210 {
211 ccv_nnc_tensor_param_t info = {
212 .type = CCV_TENSOR_GPU_MEMORY,
213 .format = CCV_TENSOR_FORMAT_NHWC,
214 .datatype = CCV_32S,
215 .dim = {1},
216 };
217 CCV_TENSOR_SET_DEVICE_ID(info.type, i)(info.type) = (((info.type) & ~0xfff00) | (((i) & 0xfff
) << 8))
;
218 isnanr[i * 2] = ccv_nnc_tensor_new(ccv_nnc_xpu_alloc(&compiled_data->xpu_alloc, i, stream_context, ccv_nnc_tensor_data_size(info)), info, 0);
219 isnanr[i * 2 + 1] = ccv_nnc_tensor_new(ccv_nnc_xpu_alloc(&compiled_data->xpu_alloc, i, stream_context, ccv_nnc_tensor_data_size(info)), info, 0);
220 }
221 } else {
222 for (i = 0; i < parallel_count; i++)
223 {
224 ccv_nnc_tensor_param_t info = {
225 .type = CCV_TENSOR_CPU_MEMORY,
226 .format = CCV_TENSOR_FORMAT_NHWC,
227 .datatype = CCV_32S,
228 .dim = {1},
229 };
230 isnanr[i * 2] = ccv_nnc_tensor_new(0, info, 0);
231 isnanr[i * 2 + 1] = ccv_nnc_tensor_new(0, info, 0);
232 }
233 }
234 // zero out old isnanr.
235 if (parallel_count > 1)
236 {
237 ccv_nnc_stream_context_t* streams[parallel_count];
238 ccv_nnc_stream_signal_t* signal;
239 if (stream_context)
240 signal = ccv_nnc_stream_context_emit_signal_new(stream_context);
241 for (i = 0; i < parallel_count; i++)
242 {
243 const int stream_type = CCV_TENSOR_GET_MEMORY(isnanr[i * 2]->info.type)((isnanr[i * 2]->info.type) & 0x3) == CCV_TENSOR_GPU_MEMORY ? CCV_STREAM_CONTEXT_GPU : CCV_STREAM_CONTEXT_CPU;
244 const int device_id = CCV_TENSOR_GET_DEVICE_ID(isnanr[i * 2]->info.type)(((isnanr[i * 2]->info.type) & 0xfff00) >> 8);
245 int type = stream_type;
246 CCV_STREAM_SET_DEVICE_ID(type, device_id)(type) = (((type) & ~0xfff00) | (((device_id) & 0xfff
) << 8))
;
247 ccv_nnc_stream_context_t* const stream_0 = ccv_cnnp_compiled_data_get_stream(compiled_data, type);
248 // Wait signal to finish.
249 if (stream_context)
250 ccv_nnc_stream_context_wait_signal(stream_0, signal);
251 ccv_nnc_cmd_exec(CMD_SET_FORWARD(0)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={0,}}}, 0)
, ccv_nnc_no_hint, 0, 0, 0, TENSOR_LIST(isnanr[i * 2])(ccv_nnc_tensor_t* []){isnanr[i * 2]}, (1 +1 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_0);
252 if (stream_context)
253 {
254 ccv_nnc_stream_signal_t* const signal = ccv_nnc_stream_context_emit_signal_new(stream_0);
255 ccv_nnc_stream_context_wait_signal(stream_context, signal);
256 }
257 streams[i] = stream_0;
258 }
259 // If this should be blocking, blocking it.
260 if (!stream_context)
261 for (i = 0; i < parallel_count; i++)
262 if (streams[i])
263 ccv_nnc_stream_context_wait(streams[i]);
264 } else
265 ccv_nnc_cmd_exec(CMD_SET_FORWARD(0)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={0,}}}, 0)
, ccv_nnc_no_hint, 0, 0, 0, TENSOR_LIST(isnanr[0])(ccv_nnc_tensor_t* []){isnanr[0]}, (1 +1 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, stream_context);
266 // Gather isnanr.
267 ccv_nnc_cmd_t reduce_cmd = {
268 .cmd = CCV_NNC_CUSTOM_FORWARD,
269 .isa = &reduce_isnan_vtab,
270 };
271 ccv_cnnp_model_parameter_gradients_map(model, parameters, reduce_cmd, ccv_nnc_no_hint, 0, 0, 0, isnanr, parallel_count * 2, stream_context);
272 for (i = 0; i < parallel_count; i++)
273 ccv_nnc_tensor_free(isnanr[i * 2 + 1]);
274 int retval = 0;
275 if (stream_type == CCV_TENSOR_GPU_MEMORY)
276 {
277 ccv_nnc_tensor_param_t info = {
278 .type = CCV_TENSOR_CPU_MEMORY,
279 .format = CCV_TENSOR_FORMAT_NHWC,
280 .datatype = CCV_32S,
281 .dim = {1},
282 };
283 ccv_nnc_tensor_t* checknan = ccv_nnc_tensor_new(0, info, 0);
284 for (i = 0; i < parallel_count; i++)
285 {
286 ccv_nnc_cmd_exec(CMD_DATA_TRANSFER_FORWARD()ccv_nnc_cmd(CCV_NNC_DATA_TRANSFER_FORWARD, 0, ccv_nnc_cmd_auto
, 0)
, ccv_nnc_no_hint, 0, TENSOR_LIST(isnanr[i * 2])(ccv_nnc_tensor_t* []){isnanr[i * 2]}, (1 +1 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_LIST(checknan)(ccv_nnc_tensor_t* []){checknan}, (1 +1 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, 0);
287 if (checknan->data.i32[0] > 0)
288 {
289 retval = 1;
290 break;
291 }
292 }
293 ccv_nnc_tensor_free(checknan);
294 } else {
295 for (i = 0; i < parallel_count; i++)
296 if (isnanr[i * 2]->data.i32[0] > 0)
297 {
298 retval = 1;
299 break;
300 }
301 }
302 for (i = 0; i < parallel_count; i++)
303 ccv_nnc_tensor_free(isnanr[i * 2]);
304 return retval;
305}
306
307// MARK - Core Layers
308
309static void _ccv_cnnp_sum_build(ccv_cnnp_model_t* const self, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
310{
311 PRINT(CCV_CLI_VERBOSE, "[cnnp_sum_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_sum_build] -\n"); fflush(stdout); } } while (
0)
;
312 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 312, __extension__ __PRETTY_FUNCTION__
); }))
;
313 outputs[0] = ccv_nnc_tensor_symbol_new(graph, ccv_nnc_tensor_symbol_params(graph, inputs[0]), 0);
314 ccv_nnc_graph_exec_symbol_new(graph, CMD_EWSUM_FORWARD()ccv_nnc_cmd(CCV_NNC_EWSUM_FORWARD, 0, ccv_nnc_cmd_auto, 0), inputs, input_size, outputs, output_size, 0);
315}
316
317static ccv_cnnp_model_t* _ccv_cnnp_sum_copy(const ccv_cnnp_model_t* const self, void* const context);
318
319static const ccv_cnnp_model_vtab_t ccv_cnnp_sum_isa = {
320 .build = _ccv_cnnp_sum_build,
321 .copy = _ccv_cnnp_sum_copy,
322};
323
324typedef struct {
325 ccv_cnnp_model_t super;
326 ccv_nnc_tensor_symbol_t output;
327} ccv_cnnp_model_sum_t;
328
329ccv_cnnp_model_t* ccv_cnnp_sum(const char* const name)
330{
331 ccv_cnnp_model_sum_t* const model_sum = (ccv_cnnp_model_sum_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_sum_t));
332 model_sum->super.isa = &ccv_cnnp_sum_isa;
333 model_sum->super.input_size = 0;
334 model_sum->super.outputs = &model_sum->output;
335 model_sum->super.output_size = 1;
336 ccv_cnnp_model_copy_name(&model_sum->super, name);
337 return (ccv_cnnp_model_t*)model_sum;
338}
339
340static ccv_cnnp_model_t* _ccv_cnnp_sum_copy(const ccv_cnnp_model_t* const self, void* const context)
341{
342 return ccv_cnnp_sum(self->name);
343}
344
345typedef struct {
346 ccv_cnnp_model_t super;
347 int axis;
348 ccv_nnc_tensor_symbol_t output;
349} ccv_cnnp_model_concat_t;
350
351static void _ccv_cnnp_concat_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
352{
353 const ccv_cnnp_model_concat_t* const self = (const ccv_cnnp_model_concat_t*)super;
354 PRINT(CCV_CLI_VERBOSE, "[cnnp_concat_build] 1. -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_concat_build] 1. -\n"); fflush(stdout); } } while
(0)
;
1
Assuming the condition is false
2
Taking false branch
3
Loop condition is false. Exiting loop
355 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 355, __extension__ __PRETTY_FUNCTION__
); }))
;
4
Assuming 'output_size' is equal to 1
5
Taking true branch
356 ccv_nnc_tensor_param_t output_params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
357 int i, j;
358 if (output_params.dim[0] == 0)
6
Assuming the condition is false
7
Taking false branch
359 for (i = 1; i < input_size; i++)
360 {
361 output_params = ccv_nnc_tensor_symbol_params(graph, inputs[i]);
362 if (output_params.dim[0] != 0)
363 break;
364 }
365 const int nd = ccv_nnc_tensor_nd(output_params.dim);
366 const int axis = self->axis;
367 assert(axis < nd)((void) sizeof ((axis < nd) ? 1 : 0), __extension__ ({ if (
axis < nd) ; else __assert_fail ("axis < nd", "ccv_cnnp_model_addons.c"
, 367, __extension__ __PRETTY_FUNCTION__); }))
;
8
Assuming 'axis' is < 'nd'
9
Taking true branch
368 output_params.dim[axis] = 0;
369 int input_is_contiguous = 1;
370 for (i = 0; i < input_size; i++)
10
Assuming 'i' is >= 'input_size'
11
Loop condition is false. Execution continues on line 394
371 {
372 const ccv_nnc_tensor_param_t input_params = ccv_nnc_tensor_symbol_params(graph, inputs[i]);
373 const int input_nd = ccv_nnc_tensor_nd(input_params.dim);
374 if (input_nd == 0)
375 {
376 PRINT(CCV_CLI_VERBOSE, "[cnnp_concat_build] %d. input[%d]: -\n", i + 2, i)do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_concat_build] %d. input[%d]: -\n", i + 2, i)
; fflush(stdout); } } while (0)
;
377 input_is_contiguous = 0;
378 continue;
379 }
380 if (CCV_CLI_OUTPUT_LEVEL_IS(CCV_CLI_VERBOSE)(CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
381 {
382 PRINT(CCV_CLI_VERBOSE, "[cnnp_concat_build] %d. input[%d]: (%d", i + 2, i, input_params.dim[0])do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_concat_build] %d. input[%d]: (%d", i + 2, i,
input_params.dim[0]); fflush(stdout); } } while (0)
;
383 int i;
384 for (i = 1; i < CCV_NNC_MAX_DIM_ALLOC(12) && input_params.dim[i] > 0; i++)
385 PRINT(CCV_CLI_VERBOSE, ", %d", input_params.dim[i])do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf(", %d", input_params.dim[i]); fflush(stdout); } } while
(0)
;
386 PRINT(CCV_CLI_VERBOSE, ")\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf(")\n"); fflush(stdout); } } while (0)
;
387 }
388 assert(input_nd == nd)((void) sizeof ((input_nd == nd) ? 1 : 0), __extension__ ({ if
(input_nd == nd) ; else __assert_fail ("input_nd == nd", "ccv_cnnp_model_addons.c"
, 388, __extension__ __PRETTY_FUNCTION__); }))
;
389 for (j = 0; j < nd; j++)
390 if (j != axis)
391 { assert(input_params.dim[j] == output_params.dim[j])((void) sizeof ((input_params.dim[j] == output_params.dim[j])
? 1 : 0), __extension__ ({ if (input_params.dim[j] == output_params
.dim[j]) ; else __assert_fail ("input_params.dim[j] == output_params.dim[j]"
, "ccv_cnnp_model_addons.c", 391, __extension__ __PRETTY_FUNCTION__
); }))
; }
392 output_params.dim[axis] += input_params.dim[axis];
393 }
394 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
395 int ofs[CCV_NNC_MAX_DIM_ALLOC(12)] = {};
396 int stride[CCV_NNC_MAX_DIM_ALLOC(12)] = {};
397 ccv_nnc_tensor_get_stride(output_params.dim, stride);
398 if (input_is_contiguous
11.1
'input_is_contiguous' is 1
)
12
Taking true branch
399 {
400 ccv_nnc_tensor_symbol_t aliases[input_size];
13
Declared variable-length array (VLA) has negative size
401 for (i = 0; i < input_size; i++)
402 {
403 const ccv_nnc_tensor_param_t input_params = ccv_nnc_tensor_symbol_params(graph, inputs[i]);
404 aliases[i] = ccv_nnc_tensor_symbol_alias_new(graph, outputs[0], ofs, stride, input_params, 0);
405 ofs[axis] += input_params.dim[axis];
406 }
407 // Format transform is more flexible.
408 ccv_nnc_graph_exec_symbol_new(graph, CMD_FORMAT_TRANSFORM_FORWARD()ccv_nnc_cmd(CCV_NNC_FORMAT_TRANSFORM_FORWARD, 0, ccv_nnc_cmd_auto
, 0)
, inputs, input_size, aliases, input_size, "concat");
409 } else {
410 ccv_nnc_tensor_symbol_t aliases[input_size];
411 for (i = 0; i < input_size; i++)
412 {
413 const ccv_nnc_tensor_param_t input_params = ccv_nnc_tensor_symbol_params(graph, inputs[i]);
414 if (input_params.dim[0] == 0)
415 {
416 // Create a new alias anyway, but not going to use it, in this way, the alias count will match during absorb.
417 aliases[i] = ccv_nnc_tensor_symbol_alias_new(graph, outputs[0], ofs, stride, input_params, 0);
418 continue;
419 }
420 aliases[i] = ccv_nnc_tensor_symbol_alias_new(graph, outputs[0], ofs, stride, input_params, 0);
421 ofs[axis] += input_params.dim[axis];
422 }
423 // Format transform is more flexible.
424 ccv_nnc_graph_exec_symbol_new(graph, CMD_FORMAT_TRANSFORM_FORWARD()ccv_nnc_cmd(CCV_NNC_FORMAT_TRANSFORM_FORWARD, 0, ccv_nnc_cmd_auto
, 0)
, inputs, input_size, aliases, input_size, "concat");
425 }
426}
427
428static ccv_cnnp_model_t* _ccv_cnnp_concat_copy(const ccv_cnnp_model_t* const self, void* const context);
429
430static const ccv_cnnp_model_vtab_t ccv_cnnp_concat_isa = {
431 .build = _ccv_cnnp_concat_build,
432 .copy = _ccv_cnnp_concat_copy,
433};
434
435ccv_cnnp_model_t* ccv_cnnp_concat(const int axis, const char* const name)
436{
437 ccv_cnnp_model_concat_t* const model_concat = (ccv_cnnp_model_concat_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_concat_t));
438 model_concat->super.isa = &ccv_cnnp_concat_isa;
439 model_concat->super.input_size = 0;
440 model_concat->super.outputs = &model_concat->output;
441 model_concat->super.output_size = 1;
442 model_concat->axis = axis;
443 ccv_cnnp_model_copy_name(&model_concat->super, name);
444 return (ccv_cnnp_model_t*)model_concat;
445}
446
447static ccv_cnnp_model_t* _ccv_cnnp_concat_copy(const ccv_cnnp_model_t* const super, void* const context)
448{
449 const ccv_cnnp_model_concat_t* const self = (const ccv_cnnp_model_concat_t*)super;
450 return ccv_cnnp_concat(self->axis, self->super.name);
451}
452
453typedef struct {
454 ccv_cnnp_model_t super;
455 int axis;
456 ccv_nnc_tensor_symbol_t outputs[1];
457} ccv_cnnp_model_chunk_t;
458
459static void _ccv_cnnp_chunk_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
460{
461 const ccv_cnnp_model_concat_t* const self = (const ccv_cnnp_model_concat_t*)super;
462 PRINT(CCV_CLI_VERBOSE, "[cnnp_chunk_build] 1. axis: %d\n", self->axis)do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_chunk_build] 1. axis: %d\n", self->axis);
fflush(stdout); } } while (0)
;
463 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 463, __extension__ __PRETTY_FUNCTION__); }))
;
464 const ccv_nnc_tensor_param_t input_params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
465 if (CCV_CLI_OUTPUT_LEVEL_IS(CCV_CLI_VERBOSE)(CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
466 {
467 PRINT(CCV_CLI_VERBOSE, "[cnnp_chunk_build] 2. input: (%d", input_params.dim[0])do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_chunk_build] 2. input: (%d", input_params.dim
[0]); fflush(stdout); } } while (0)
;
468 int i;
469 for (i = 1; i < CCV_NNC_MAX_DIM_ALLOC(12) && input_params.dim[i] > 0; i++)
470 PRINT(CCV_CLI_VERBOSE, ", %d", input_params.dim[i])do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf(", %d", input_params.dim[i]); fflush(stdout); } } while
(0)
;
471 PRINT(CCV_CLI_VERBOSE, ")\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf(")\n"); fflush(stdout); } } while (0)
;
472 }
473 ccv_nnc_tensor_param_t output_params = input_params;
474 int i;
475 const int nd = ccv_nnc_tensor_nd(output_params.dim);
476 const int axis = self->axis;
477 assert(axis < nd)((void) sizeof ((axis < nd) ? 1 : 0), __extension__ ({ if (
axis < nd) ; else __assert_fail ("axis < nd", "ccv_cnnp_model_addons.c"
, 477, __extension__ __PRETTY_FUNCTION__); }))
;
478 const int n = self->super.output_size;
479 assert(n == output_size)((void) sizeof ((n == output_size) ? 1 : 0), __extension__ ({
if (n == output_size) ; else __assert_fail ("n == output_size"
, "ccv_cnnp_model_addons.c", 479, __extension__ __PRETTY_FUNCTION__
); }))
;
480 assert(output_params.dim[axis] % n == 0)((void) sizeof ((output_params.dim[axis] % n == 0) ? 1 : 0), __extension__
({ if (output_params.dim[axis] % n == 0) ; else __assert_fail
("output_params.dim[axis] % n == 0", "ccv_cnnp_model_addons.c"
, 480, __extension__ __PRETTY_FUNCTION__); }))
;
481 output_params.dim[axis] = output_params.dim[axis] / n;
482 int ofs[CCV_NNC_MAX_DIM_ALLOC(12)] = {};
483 int stride[CCV_NNC_MAX_DIM_ALLOC(12)] = {};
484 ccv_nnc_tensor_get_stride(input_params.dim, stride);
485 ccv_nnc_tensor_symbol_t to = ccv_nnc_tensor_symbol_alias_to(graph, inputs[0]);
486 if (to.d == CCV_NNC_NO_TENSOR_SYMBOL) // If we are not reshape an alias, it is straightforward.
487 {
488 for (i = 0; i < output_size; i++)
489 {
490 outputs[i] = ccv_nnc_tensor_symbol_alias_new(graph, inputs[0], ofs, stride, output_params, 0);
491 ofs[axis] += output_params.dim[axis];
492 }
493 } else {
494 // Otherwise, we need to check if it is permute. For permute, we cannot do alias directly.
495 // We need to first materialize the permute and then run reshape on top of it, otherwise it will be wrong.
496 int old_stride[CCV_NNC_MAX_DIM_ALLOC(12)];
497 ccv_nnc_tensor_symbol_alias_params(graph, inputs[0], 0, old_stride);
498 // We identify permute by checking if the stride is not in descending order.
499 // This also covered "permute" through reshape, rather than using ccv_cnnp_permute directly.
500 int i, no_permute = 1;
501 for (i = 1; no_permute && i < nd; i++)
502 if (old_stride[i - 1] < old_stride[i])
503 no_permute = 0;
504 if (no_permute)
505 { // Just straightforward reshape if there is no no permute.
506 for (i = 0; i < output_size; i++)
507 {
508 outputs[i] = ccv_nnc_tensor_symbol_alias_new(graph, inputs[0], ofs, old_stride, output_params, 0);
509 ofs[axis] += output_params.dim[axis];
510 }
511 } else {
512 // Otherwise, we first do format transform to plain tensor and then do reshape.
513 ccv_nnc_tensor_symbol_t permuted = ccv_nnc_tensor_symbol_new(graph, input_params, 0);
514 ccv_nnc_graph_exec_symbol_new(graph, CMD_FORMAT_TRANSFORM_FORWARD()ccv_nnc_cmd(CCV_NNC_FORMAT_TRANSFORM_FORWARD, 0, ccv_nnc_cmd_auto
, 0)
, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(permuted)(const ccv_nnc_tensor_symbol_t []){permuted}, (1 +1 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "reshape");
515 for (i = 0; i < output_size; i++)
516 {
517 outputs[i] = ccv_nnc_tensor_symbol_alias_new(graph, permuted, ofs, stride, output_params, 0);
518 ofs[axis] += output_params.dim[axis];
519 }
520 }
521 }
522}
523
524static ccv_cnnp_model_t* _ccv_cnnp_chunk_copy(const ccv_cnnp_model_t* const self, void* const context);
525
526static const ccv_cnnp_model_vtab_t ccv_cnnp_chunk_isa = {
527 .build = _ccv_cnnp_chunk_build,
528 .copy = _ccv_cnnp_chunk_copy,
529};
530
531ccv_cnnp_model_t* ccv_cnnp_chunk(const int n, const int axis, const char* const name)
532{
533 assert(n >= 1)((void) sizeof ((n >= 1) ? 1 : 0), __extension__ ({ if (n >=
1) ; else __assert_fail ("n >= 1", "ccv_cnnp_model_addons.c"
, 533, __extension__ __PRETTY_FUNCTION__); }))
;
534 ccv_cnnp_model_chunk_t* const model_chunk = (ccv_cnnp_model_chunk_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_chunk_t) + sizeof(ccv_nnc_tensor_symbol_t) * (n - 1));
535 model_chunk->super.isa = &ccv_cnnp_chunk_isa;
536 model_chunk->super.input_size = 1;
537 model_chunk->super.outputs = model_chunk->outputs;
538 model_chunk->super.output_size = n;
539 model_chunk->axis = axis;
540 ccv_cnnp_model_copy_name(&model_chunk->super, name);
541 return (ccv_cnnp_model_t*)model_chunk;
542}
543
544static ccv_cnnp_model_t* _ccv_cnnp_chunk_copy(const ccv_cnnp_model_t* const super, void* const context)
545{
546 const ccv_cnnp_model_chunk_t* const self = (const ccv_cnnp_model_chunk_t*)super;
547 return ccv_cnnp_chunk(self->super.output_size, self->axis, self->super.name);
548}
549
550typedef struct {
551 ccv_cnnp_model_t super;
552 ccv_nnc_tensor_symbol_t output;
553 int format;
554 int dim[CCV_NNC_MAX_DIM_ALLOC(12)];
555 int ofs[CCV_NNC_MAX_DIM_ALLOC(12)];
556 int stride[CCV_NNC_MAX_DIM_ALLOC(12)];
557} ccv_cnnp_model_reshape_t;
558
559static void _ccv_cnnp_reshape_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
560{
561 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 561, __extension__ __PRETTY_FUNCTION__); }))
;
562 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 562, __extension__ __PRETTY_FUNCTION__
); }))
;
563 ccv_cnnp_model_reshape_t* const self = (ccv_cnnp_model_reshape_t*)super;
564 if (CCV_CLI_OUTPUT_LEVEL_IS(CCV_CLI_VERBOSE)(CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
565 {
566 PRINT(CCV_CLI_VERBOSE, "[cnnp_reshape_build] 1. dim: (%d", self->dim[0])do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_reshape_build] 1. dim: (%d", self->dim[0]
); fflush(stdout); } } while (0)
;
567 int i;
568 for (i = 1; i < CCV_NNC_MAX_DIM_ALLOC(12) && self->dim[i] > 0; i++)
569 PRINT(CCV_CLI_VERBOSE, ", %d", self->dim[i])do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf(", %d", self->dim[i]); fflush(stdout); } } while
(0)
;
570 const int count = i;
571 PRINT(CCV_CLI_VERBOSE, "), ofs: (%d", self->ofs[0])do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("), ofs: (%d", self->ofs[0]); fflush(stdout); } }
while (0)
;
572 for (i = 1; i < count; i++)
573 PRINT(CCV_CLI_VERBOSE, ", %d", self->ofs[i])do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf(", %d", self->ofs[i]); fflush(stdout); } } while
(0)
;
574 PRINT(CCV_CLI_VERBOSE, "), stride: (%d", self->stride[0])do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("), stride: (%d", self->stride[0]); fflush(stdout
); } } while (0)
;
575 for (i = 1; i < count; i++)
576 PRINT(CCV_CLI_VERBOSE, ", %d", self->stride[i])do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf(", %d", self->stride[i]); fflush(stdout); } } while
(0)
;
577 PRINT(CCV_CLI_VERBOSE, ")\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf(")\n"); fflush(stdout); } } while (0)
;
578 }
579 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
580 int dim[CCV_NNC_MAX_DIM_ALLOC(12)];
581 memcpy(dim, self->dim, sizeof(dim));
582 int i, auto_idx = -1;
583 size_t known = 1;
584 const size_t tensor_count = ccv_nnc_tensor_count(params);
585 for (i = 0; i < CCV_NNC_MAX_DIM_ALLOC(12) && dim[i]; i++)
586 if (dim[i] == -1)
587 auto_idx = i;
588 else
589 known *= dim[i];
590 if (auto_idx >= 0)
591 {
592 assert(known > 0 && tensor_count % known == 0)((void) sizeof ((known > 0 && tensor_count % known
== 0) ? 1 : 0), __extension__ ({ if (known > 0 &&
tensor_count % known == 0) ; else __assert_fail ("known > 0 && tensor_count % known == 0"
, "ccv_cnnp_model_addons.c", 592, __extension__ __PRETTY_FUNCTION__
); }))
;
593 dim[auto_idx] = tensor_count / known;
594 }
595 if (CCV_CLI_OUTPUT_LEVEL_IS(CCV_CLI_VERBOSE)(CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
596 {
597 PRINT(CCV_CLI_VERBOSE, "[cnnp_reshape_build] 2. input: (%d", params.dim[0])do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_reshape_build] 2. input: (%d", params.dim[0]
); fflush(stdout); } } while (0)
;
598 int i;
599 for (i = 1; i < CCV_NNC_MAX_DIM_ALLOC(12) && params.dim[i] > 0; i++)
600 PRINT(CCV_CLI_VERBOSE, ", %d", params.dim[i])do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf(", %d", params.dim[i]); fflush(stdout); } } while (
0)
;
601 PRINT(CCV_CLI_VERBOSE, ")\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf(")\n"); fflush(stdout); } } while (0)
;
602 }
603 assert(ccv_nnc_dimension_count(dim) <= ccv_nnc_tensor_count(params))((void) sizeof ((ccv_nnc_dimension_count(dim) <= ccv_nnc_tensor_count
(params)) ? 1 : 0), __extension__ ({ if (ccv_nnc_dimension_count
(dim) <= ccv_nnc_tensor_count(params)) ; else __assert_fail
("ccv_nnc_dimension_count(dim) <= ccv_nnc_tensor_count(params)"
, "ccv_cnnp_model_addons.c", 603, __extension__ __PRETTY_FUNCTION__
); }))
;
604 ccv_nnc_tensor_symbol_t to = ccv_nnc_tensor_symbol_alias_to(graph, inputs[0]);
605 int stride_from_dim[CCV_NNC_MAX_DIM_ALLOC(12)];
606 if (to.d == CCV_NNC_NO_TENSOR_SYMBOL) // If we are not reshape an alias, it is straightforward.
607 {
608 memcpy(params.dim, dim, sizeof(params.dim));
609 int* stride;
610 if (self->stride[0] == 0)
611 {
612 ccv_nnc_tensor_get_stride(dim, stride_from_dim);
613 stride = stride_from_dim;
614 } else
615 stride = self->stride;
616 if (self->format > 0)
617 params.format = self->format;
618 outputs[0] = ccv_nnc_tensor_symbol_alias_new(graph, inputs[0], self->ofs, stride, params, 0);
619 } else {
620 // Otherwise, we need to check if it is permute. For permute, we cannot do alias directly.
621 // We need to first materialize the permute and then run reshape on top of it, otherwise it will be wrong.
622 int old_stride[CCV_NNC_MAX_DIM_ALLOC(12)];
623 ccv_nnc_tensor_symbol_alias_params(graph, inputs[0], 0, old_stride);
624 // We identify permute by checking if the stride is not in descending order.
625 // This also covered "permute" through reshape, rather than using ccv_cnnp_permute directly.
626 const int nd = ccv_nnc_tensor_nd(params.dim);
627 const int new_nd = ccv_nnc_tensor_nd(dim);
628 int i, no_permute = 1;
629 // If the new dim has different nd, or we actually have a stride, we need to check if it is no permute or not.
630 if (new_nd != nd || (self->stride[0] != 0 && memcmp(self->stride, old_stride, sizeof(self->stride))))
631 for (i = 1; no_permute && i < nd; i++)
632 if (old_stride[i - 1] < old_stride[i])
633 no_permute = 0;
634 if (no_permute)
635 { // Just straightforward reshape if there is no no permute.
636 memcpy(params.dim, dim, sizeof(params.dim));
637 int* stride;
638 if (self->stride[0] == 0)
639 {
640 if (new_nd != nd) // Cannot use old stride.
641 {
642 ccv_nnc_tensor_get_stride(dim, stride_from_dim);
643 stride = stride_from_dim;
644 } else
645 stride = old_stride;
646 } else
647 stride = self->stride;
648 if (self->format > 0)
649 params.format = self->format;
650 outputs[0] = ccv_nnc_tensor_symbol_alias_new(graph, inputs[0], self->ofs, stride, params, 0);
651 } else {
652 // Otherwise, we first do format transform to plain tensor and then do reshape.
653 ccv_nnc_tensor_symbol_t permuted = ccv_nnc_tensor_symbol_new(graph, params, 0);
654 ccv_nnc_graph_exec_symbol_new(graph, CMD_FORMAT_TRANSFORM_FORWARD()ccv_nnc_cmd(CCV_NNC_FORMAT_TRANSFORM_FORWARD, 0, ccv_nnc_cmd_auto
, 0)
, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(permuted)(const ccv_nnc_tensor_symbol_t []){permuted}, (1 +1 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "reshape");
655 memcpy(params.dim, dim, sizeof(params.dim));
656 int* stride;
657 if (self->stride[0] == 0)
658 {
659 ccv_nnc_tensor_get_stride(dim, stride_from_dim);
660 stride = stride_from_dim;
661 } else
662 stride = self->stride;
663 if (self->format > 0)
664 params.format = self->format;
665 // And then we create alias against the permuted one.
666 outputs[0] = ccv_nnc_tensor_symbol_alias_new(graph, permuted, self->ofs, stride, params, 0);
667 }
668 }
669}
670
671static ccv_cnnp_model_t* _ccv_cnnp_reshape_copy(const ccv_cnnp_model_t* const super, void* const context);
672
673static const ccv_cnnp_model_vtab_t ccv_cnnp_reshape_isa = {
674 .build = _ccv_cnnp_reshape_build,
675 .copy = _ccv_cnnp_reshape_copy,
676};
677
678ccv_cnnp_model_t* ccv_cnnp_reshape(const int format, const int dim[CCV_NNC_MAX_DIM_ALLOC(12)], const int ofs[CCV_NNC_MAX_DIM_ALLOC(12)], const int stride[CCV_NNC_MAX_DIM_ALLOC(12)], const char* const name)
679{
680 ccv_cnnp_model_reshape_t* const model_reshape = (ccv_cnnp_model_reshape_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_reshape_t));
681 model_reshape->super.isa = &ccv_cnnp_reshape_isa;
682 model_reshape->super.input_size = 1;
683 model_reshape->super.outputs = &model_reshape->output;
684 model_reshape->super.output_size = 1;
685 ccv_cnnp_model_copy_name(&model_reshape->super, name);
686 model_reshape->format = format;
687 memcpy(model_reshape->dim, dim, sizeof(model_reshape->dim));
688 memcpy(model_reshape->ofs, ofs, sizeof(model_reshape->ofs));
689 if (stride[0] != 0)
690 memcpy(model_reshape->stride, stride, sizeof(model_reshape->stride));
691 return (ccv_cnnp_model_t*)model_reshape;
692}
693
694static ccv_cnnp_model_t* _ccv_cnnp_reshape_copy(const ccv_cnnp_model_t* const super, void* const context)
695{
696 const ccv_cnnp_model_reshape_t* const self = (const ccv_cnnp_model_reshape_t*)super;
697 return ccv_cnnp_reshape(self->format, self->dim, self->ofs, self->stride, self->super.name);
698}
699
700typedef struct {
701 ccv_cnnp_model_t super;
702 ccv_nnc_tensor_symbol_t output;
703 int type;
704 int begin[CCV_NNC_MAX_DIM_ALLOC(12)];
705 int end[CCV_NNC_MAX_DIM_ALLOC(12)];
706} ccv_cnnp_model_pad_t;
707
708static void _ccv_cnnp_pad_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
709{
710 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 710, __extension__ __PRETTY_FUNCTION__); }))
;
711 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 711, __extension__ __PRETTY_FUNCTION__
); }))
;
712 ccv_cnnp_model_pad_t* const self = (ccv_cnnp_model_pad_t*)super;
713 PRINT(CCV_CLI_VERBOSE, "[cnnp_pad_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_pad_build] -\n"); fflush(stdout); } } while (
0)
;
714 const ccv_nnc_tensor_param_t input_params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
715 const int nd = ccv_nnc_tensor_nd(input_params.dim);
716 ccv_nnc_tensor_param_t params = input_params;
717 int i;
718 for (i = 0 ; i < nd; i++)
719 params.dim[i] += self->begin[i] + self->end[i];
720 const ccv_nnc_tensor_symbol_t padded = ccv_nnc_tensor_symbol_new(graph, params, 0);
721 ccv_nnc_cmd_t pad = CMD_PAD_FORWARD(self->type, (), ())ccv_nnc_cmd(CCV_NNC_PAD_FORWARD, 0, ((ccv_nnc_cmd_param_t){.size
={.dim={}},.pad={.type=self->type,.end={}}}), 0)
;
722 memcpy(pad.info.size.dim, self->begin, sizeof(pad.info.size.dim));
723 memcpy(pad.info.pad.end, self->end, sizeof(pad.info.pad.end));
724 ccv_nnc_graph_exec_symbol_new(graph, pad, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(padded)(const ccv_nnc_tensor_symbol_t []){padded}, (1 +1 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "pad");
725 outputs[0] = padded;
726}
727
728static ccv_cnnp_model_t* _ccv_cnnp_pad_copy(const ccv_cnnp_model_t* const super, void* const context);
729
730static const ccv_cnnp_model_vtab_t ccv_cnnp_pad_isa = {
731 .build = _ccv_cnnp_pad_build,
732 .copy = _ccv_cnnp_pad_copy,
733};
734
735ccv_cnnp_model_t* ccv_cnnp_pad(const int type, const int begin[CCV_NNC_MAX_DIM_ALLOC(12)], const int end[CCV_NNC_MAX_DIM_ALLOC(12)], const char* const name)
736{
737 ccv_cnnp_model_pad_t* const model_pad = (ccv_cnnp_model_pad_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_pad_t));
738 model_pad->super.isa = &ccv_cnnp_pad_isa;
739 model_pad->super.input_size = 1;
740 model_pad->super.outputs = &model_pad->output;
741 model_pad->super.output_size = 1;
742 ccv_cnnp_model_copy_name(&model_pad->super, name);
743 model_pad->type = type;
744 memcpy(model_pad->begin, begin, sizeof(model_pad->begin));
745 memcpy(model_pad->end, end, sizeof(model_pad->end));
746 return (ccv_cnnp_model_t*)model_pad;
747}
748
749static ccv_cnnp_model_t* _ccv_cnnp_pad_copy(const ccv_cnnp_model_t* const super, void* const context)
750{
751 const ccv_cnnp_model_pad_t* const self = (const ccv_cnnp_model_pad_t*)super;
752 return ccv_cnnp_pad(self->type, self->begin, self->end, self->super.name);
753}
754
755typedef struct {
756 ccv_cnnp_model_t super;
757 ccv_nnc_tensor_symbol_t output;
758} ccv_cnnp_model_identity_t;
759
760static void _ccv_cnnp_identity_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
761{
762 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 762, __extension__ __PRETTY_FUNCTION__); }))
;
763 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 763, __extension__ __PRETTY_FUNCTION__
); }))
;
764 PRINT(CCV_CLI_VERBOSE, "[cnnp_identity_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_identity_build] -\n"); fflush(stdout); } } while
(0)
;
765 outputs[0] = inputs[0];
766}
767
768static ccv_cnnp_model_t* _ccv_cnnp_identity_copy(const ccv_cnnp_model_t* const super, void* const context);
769
770static const ccv_cnnp_model_vtab_t ccv_cnnp_identity_isa = {
771 .build = _ccv_cnnp_identity_build,
772 .copy = _ccv_cnnp_identity_copy,
773};
774
775ccv_cnnp_model_t* ccv_cnnp_identity(const char* const name)
776{
777 ccv_cnnp_model_identity_t* const model_identity = (ccv_cnnp_model_identity_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_identity_t));
778 model_identity->super.isa = &ccv_cnnp_identity_isa;
779 model_identity->super.input_size = 1;
780 model_identity->super.outputs = &model_identity->output;
781 model_identity->super.output_size = 1;
782 ccv_cnnp_model_copy_name(&model_identity->super, name);
783 return (ccv_cnnp_model_t*)model_identity;
784}
785
786static ccv_cnnp_model_t* _ccv_cnnp_identity_copy(const ccv_cnnp_model_t* const super, void* const context)
787{
788 const ccv_cnnp_model_identity_t* const self = (const ccv_cnnp_model_identity_t*)super;
789 return ccv_cnnp_identity(self->super.name);
790}
791
792typedef struct {
793 ccv_cnnp_model_t super;
794 ccv_nnc_tensor_symbol_t output;
795 int index[CCV_NNC_MAX_DIM_ALLOC(12)];
796} ccv_cnnp_model_permute_t;
797
798static void _ccv_cnnp_permute_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
799{
800 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 800, __extension__ __PRETTY_FUNCTION__); }))
;
801 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 801, __extension__ __PRETTY_FUNCTION__
); }))
;
802 ccv_cnnp_model_permute_t* const self = (ccv_cnnp_model_permute_t*)super;
803 PRINT(CCV_CLI_VERBOSE, "[cnnp_permute_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_permute_build] -\n"); fflush(stdout); } } while
(0)
;
804 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
805 ccv_nnc_tensor_symbol_t to = ccv_nnc_tensor_symbol_alias_to(graph, inputs[0]);
806 const int nd = ccv_nnc_tensor_nd(params.dim);
807 int input_dim[CCV_NNC_MAX_DIM_ALLOC(12)];
808 memcpy(input_dim, params.dim, sizeof(params.dim));
809 int input_stride[CCV_NNC_MAX_DIM_ALLOC(12)] = {};
810 int output_stride[CCV_NNC_MAX_DIM_ALLOC(12)] = {};
811 if (to.d == CCV_NNC_NO_TENSOR_SYMBOL) // If it is not an alias. Find stride and permute.
812 {
813 ccv_nnc_tensor_get_stride(input_dim, input_stride);
814 int i;
815 for (i = 0; i < nd; i++)
816 {
817 const int idx = self->index[i];
818 assert(idx >= 0 && idx < nd)((void) sizeof ((idx >= 0 && idx < nd) ? 1 : 0)
, __extension__ ({ if (idx >= 0 && idx < nd) ; else
__assert_fail ("idx >= 0 && idx < nd", "ccv_cnnp_model_addons.c"
, 818, __extension__ __PRETTY_FUNCTION__); }))
;
819 params.dim[i] = input_dim[idx];
820 output_stride[i] = input_stride[idx];
821 }
822 outputs[0] = ccv_nnc_tensor_symbol_alias_new(graph, inputs[0], ccv_nnc_no_ofs, output_stride, params, 0);
823 } else {
824 // if it is an alias, we can get the stride from it and use that.
825 int input_ofs[CCV_NNC_MAX_DIM_ALLOC(12)];
826 ccv_nnc_tensor_symbol_alias_params(graph, inputs[0], input_ofs, input_stride);
827 assert(input_stride[0] != 0)((void) sizeof ((input_stride[0] != 0) ? 1 : 0), __extension__
({ if (input_stride[0] != 0) ; else __assert_fail ("input_stride[0] != 0"
, "ccv_cnnp_model_addons.c", 827, __extension__ __PRETTY_FUNCTION__
); }))
;
828 int output_ofs[CCV_NNC_MAX_DIM_ALLOC(12)] = {};
829 int i;
830 for (i = 0; i < nd; i++)
831 {
832 const int idx = self->index[i];
833 assert(idx >= 0 && idx < nd)((void) sizeof ((idx >= 0 && idx < nd) ? 1 : 0)
, __extension__ ({ if (idx >= 0 && idx < nd) ; else
__assert_fail ("idx >= 0 && idx < nd", "ccv_cnnp_model_addons.c"
, 833, __extension__ __PRETTY_FUNCTION__); }))
;
834 params.dim[i] = input_dim[idx];
835 output_stride[i] = input_stride[idx];
836 output_ofs[i] = input_ofs[idx];
837 }
838 outputs[0] = ccv_nnc_tensor_symbol_alias_new(graph, inputs[0], output_ofs, output_stride, params, 0);
839 }
840}
841
842static ccv_cnnp_model_t* _ccv_cnnp_permute_copy(const ccv_cnnp_model_t* const super, void* const context);
843
844static const ccv_cnnp_model_vtab_t ccv_cnnp_permute_isa = {
845 .build = _ccv_cnnp_permute_build,
846 .copy = _ccv_cnnp_permute_copy,
847};
848
849ccv_cnnp_model_t* ccv_cnnp_permute(const int index[CCV_NNC_MAX_DIM_ALLOC(12)], const char* const name)
850{
851 ccv_cnnp_model_permute_t* const model_permute = (ccv_cnnp_model_permute_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_permute_t));
852 model_permute->super.isa = &ccv_cnnp_permute_isa;
853 model_permute->super.input_size = 1;
854 model_permute->super.outputs = &model_permute->output;
855 model_permute->super.output_size = 1;
856 ccv_cnnp_model_copy_name(&model_permute->super, name);
857 memcpy(model_permute->index, index, sizeof(model_permute->index));
858 return (ccv_cnnp_model_t*)model_permute;
859}
860
861static ccv_cnnp_model_t* _ccv_cnnp_permute_copy(const ccv_cnnp_model_t* const super, void* const context)
862{
863 const ccv_cnnp_model_permute_t* const self = (const ccv_cnnp_model_permute_t*)super;
864 return ccv_cnnp_permute(self->index, self->super.name);
865}
866
867typedef struct {
868 ccv_cnnp_model_t super;
869 int index;
870 ccv_nnc_tensor_symbol_t output;
871} ccv_cnnp_model_extract_t;
872
873static void _ccv_cnnp_extract_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
874{
875 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 875, __extension__ __PRETTY_FUNCTION__
); }))
;
876 ccv_cnnp_model_extract_t* const self = (ccv_cnnp_model_extract_t*)super;
877 PRINT(CCV_CLI_VERBOSE, "[cnnp_extract_build] index: %d\n", self->index)do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_extract_build] index: %d\n", self->index)
; fflush(stdout); } } while (0)
;
878 outputs[0] = inputs[self->index];
879}
880
881static ccv_cnnp_model_t* _ccv_cnnp_extract_copy(const ccv_cnnp_model_t* const self, void* const context);
882
883static const ccv_cnnp_model_vtab_t ccv_cnnp_extract_isa = {
884 .build = _ccv_cnnp_extract_build,
885 .copy = _ccv_cnnp_extract_copy,
886};
887
888ccv_cnnp_model_t* ccv_cnnp_extract(const int index, const char* const name)
889{
890 ccv_cnnp_model_extract_t* const model_extract = (ccv_cnnp_model_extract_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_extract_t));
891 model_extract->index = index;
892 model_extract->super.isa = &ccv_cnnp_extract_isa;
893 model_extract->super.input_size = 0;
894 model_extract->super.outputs = &model_extract->output;
895 model_extract->super.output_size = 1;
896 ccv_cnnp_model_copy_name(&model_extract->super, name);
897 return (ccv_cnnp_model_t*)model_extract;
898}
899
900static ccv_cnnp_model_t* _ccv_cnnp_extract_copy(const ccv_cnnp_model_t* const super, void* const context)
901{
902 ccv_cnnp_model_extract_t* const self = (ccv_cnnp_model_extract_t*)super;
903 return ccv_cnnp_extract(self->index, self->super.name);
904}
905
906typedef struct {
907 ccv_cnnp_model_t super;
908 ccv_nnc_tensor_symbol_t output;
909} ccv_cnnp_model_flatten_t;
910
911static void _ccv_cnnp_flatten_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
912{
913 PRINT(CCV_CLI_VERBOSE, "[cnnp_flatten_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_flatten_build] -\n"); fflush(stdout); } } while
(0)
;
914 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 914, __extension__ __PRETTY_FUNCTION__); }))
;
915 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 915, __extension__ __PRETTY_FUNCTION__
); }))
;
916 const ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
917 ccv_nnc_tensor_param_t output_params = params;
918 memset(output_params.dim, 0, sizeof(output_params.dim));
919 output_params.dim[0] = ccv_nnc_tensor_get_n(params);
920 assert(output_params.dim[0] > 0)((void) sizeof ((output_params.dim[0] > 0) ? 1 : 0), __extension__
({ if (output_params.dim[0] > 0) ; else __assert_fail ("output_params.dim[0] > 0"
, "ccv_cnnp_model_addons.c", 920, __extension__ __PRETTY_FUNCTION__
); }))
;
921 output_params.dim[1] = ccv_nnc_tensor_count(params) / output_params.dim[0];
922 int stride[CCV_NNC_MAX_DIM_ALLOC(12)] = {};
923 ccv_nnc_tensor_get_stride(output_params.dim, stride);
924 outputs[0] = ccv_nnc_tensor_symbol_alias_new(graph, inputs[0], DIM_ALLOC()(int [(12)]){}, stride, output_params, 0);
925}
926
927static ccv_cnnp_model_t* _ccv_cnnp_flatten_copy(const ccv_cnnp_model_t* const self, void* const context);
928
929static const ccv_cnnp_model_vtab_t ccv_cnnp_flatten_isa = {
930 .build = _ccv_cnnp_flatten_build,
931 .copy = _ccv_cnnp_flatten_copy,
932};
933
934ccv_cnnp_model_t* ccv_cnnp_flatten(const char* const name)
935{
936 ccv_cnnp_model_flatten_t* const model_flatten = (ccv_cnnp_model_flatten_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_flatten_t));
937 model_flatten->super.isa = &ccv_cnnp_flatten_isa;
938 model_flatten->super.input_size = 1;
939 model_flatten->super.outputs = &model_flatten->output;
940 model_flatten->super.output_size = 1;
941 ccv_cnnp_model_copy_name(&model_flatten->super, name);
942 return (ccv_cnnp_model_t*)model_flatten;
943}
944
945static ccv_cnnp_model_t* _ccv_cnnp_flatten_copy(const ccv_cnnp_model_t* const self, void* const context)
946{
947 return ccv_cnnp_flatten(self->name);
948}
949
950// MARK - Batch Norm Layer
951
952typedef struct {
953 ccv_cnnp_model_t super;
954 ccv_nnc_tensor_symbol_t output;
955 ccv_nnc_tensor_symbol_t bias;
956 ccv_nnc_tensor_symbol_t scale;
957 ccv_nnc_graph_exec_symbol_t batch_norm;
958 ccv_nnc_cmd_param_t params;
959 ccv_array_t* zero_inits;
960 ccv_array_t* retainables;
961} ccv_cnnp_model_batch_norm_t;
962
963static void _ccv_cnnp_batch_norm_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
964{
965 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 965, __extension__ __PRETTY_FUNCTION__); }))
;
966 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 966, __extension__ __PRETTY_FUNCTION__
); }))
;
967 ccv_cnnp_model_batch_norm_t* const self = (ccv_cnnp_model_batch_norm_t*)super;
968 PRINT(CCV_CLI_VERBOSE, "[cnnp_batch_norm_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_batch_norm_build] -\n"); fflush(stdout); } }
while (0)
;
969 const ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
970 const int nd = ccv_nnc_tensor_nd(params.dim);
971 ccv_nnc_tensor_param_t bias_params = params;
972 memset(bias_params.dim, 0, sizeof(bias_params.dim));
973 // If the accuracy is not enough, bump it to 32-bit floating point.
974 if (bias_params.datatype != CCV_32F && bias_params.datatype != CCV_64F)
975 bias_params.datatype = CCV_32F;
976 bias_params.dim[0] = nd > 1 ? ccv_nnc_tensor_get_c(params) : params.dim[0];
977 const ccv_nnc_tensor_symbol_t output = ccv_nnc_tensor_symbol_new(graph, params, 0);
978 // Both scale and bias are shared between if this model is reused.
979 if (!self->scale.graph)
980 self->scale = ccv_nnc_tensor_symbol_new(graph, bias_params, "scale");
981 if (!self->bias.graph)
982 self->bias = ccv_nnc_tensor_symbol_new(graph, bias_params, "bias");
983 const ccv_nnc_tensor_symbol_t mean = ccv_nnc_tensor_symbol_new(graph, bias_params, "mean");
984 const ccv_nnc_tensor_symbol_t var = ccv_nnc_tensor_symbol_new(graph, bias_params, "var");
985 // Otherwise, notice mean, var, saved_mean, saved_inv_std are not reused.
986 if (!self->zero_inits)
987 self->zero_inits = ccv_array_new(sizeof(ccv_nnc_tensor_symbol_t), 0, 0);
988 ccv_array_push(self->zero_inits, &mean);
989 ccv_array_push(self->zero_inits, &var);
990 const ccv_nnc_tensor_symbol_t out_mean = ccv_nnc_tensor_symbol_new(graph, bias_params, "out_mean");
991 const ccv_nnc_tensor_symbol_t out_var = ccv_nnc_tensor_symbol_new(graph, bias_params, "out_var");
992 if (!self->retainables)
993 self->retainables = ccv_array_new(sizeof(ccv_nnc_tensor_symbol_t), 0, 0);
994 ccv_array_push(self->retainables, &out_mean);
995 ccv_array_push(self->retainables, &out_var);
996 const ccv_nnc_tensor_symbol_t saved_mean = ccv_nnc_tensor_symbol_new(graph, bias_params, "saved_mean");
997 const ccv_nnc_tensor_symbol_t saved_inv_std = ccv_nnc_tensor_symbol_new(graph, bias_params, "saved_inv_std");
998 const int hw = ccv_nnc_tensor_hw(params, ccv_nnc_tensor_nd(params.dim), CCV_NNC_MAX_DIM(2));
999 ccv_nnc_cmd_param_t batch_norm = self->params;
1000 batch_norm.bnorm.count = hw >= 0 ? CCV_NNC_MAX_DIM(2) + 1 : 1;
1001 int i;
1002 batch_norm.bnorm.axis[0] = (params.format == CCV_TENSOR_FORMAT_CHWN) ? 3 : 0;
1003 if (hw >= 0)
1004 for (i = 0; i < CCV_NNC_MAX_DIM(2); i++)
1005 batch_norm.bnorm.axis[i + 1] = i + hw;
1006 self->params = batch_norm;
1007 self->batch_norm = ccv_nnc_graph_exec_symbol_new(graph, ccv_nnc_cmd(CCV_NNC_BATCH_NORM_FORWARD, 0, batch_norm, 0), TENSOR_SYMBOL_LIST(inputs[0], self->scale, self->bias, mean, var)(const ccv_nnc_tensor_symbol_t []){inputs[0], self->scale,
self->bias, mean, var}, (1 +1 +1 +1 +1 +1 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output, out_mean, out_var, saved_mean, saved_inv_std)(const ccv_nnc_tensor_symbol_t []){output, out_mean, out_var,
saved_mean, saved_inv_std}, (1 +1 +1 +1 +1 +1 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "batch_norm");
1008 outputs[0] = output;
1009}
1010
1011static void _ccv_cnnp_batch_norm_init_states(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_cnnp_state_initializer_f initializer, void* const context)
1012{
1013 ccv_cnnp_model_batch_norm_t* const self = (ccv_cnnp_model_batch_norm_t*)super;
1014 if (self->scale.graph)
1015 initializer(context, CMD_RANDOM_UNIFORM_FORWARD(0, 1)ccv_nnc_cmd(CCV_NNC_RANDOM_UNIFORM_FORWARD, 0, (ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.blas={.a={0, 1}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->scale);
1016 if (self->bias.graph)
1017 initializer(context, CMD_SET_FORWARD(0)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={0,}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->bias);
1018 int i;
1019 if (self->zero_inits)
1020 for (i = 0; i < self->zero_inits->rnum; i++)
1021 initializer(context, CMD_SET_FORWARD(0)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={0,}}}, 0)
, ccv_nnc_no_hint, 0, 0, *(ccv_nnc_tensor_symbol_t*)ccv_array_get(self->zero_inits, i)((void*)(((char*)((self->zero_inits)->data)) + (size_t)
(self->zero_inits)->rsize * (size_t)(i)))
);
1022}
1023
1024static void _ccv_cnnp_batch_norm_add_to_parameter(ccv_cnnp_model_t* const super, const ccv_cnnp_add_to_array_f add_to_array, void* const parameters, const int is_trainable)
1025{
1026 ccv_cnnp_model_batch_norm_t* const self = (ccv_cnnp_model_batch_norm_t*)super;
1027 if (self->scale.graph)
1028 add_to_array(parameters, self->scale, is_trainable);
1029 if (self->bias.graph)
1030 add_to_array(parameters, self->bias, is_trainable);
1031}
1032
1033static void _ccv_cnnp_batch_norm_add_to_output(ccv_cnnp_model_t* const super, const ccv_cnnp_add_to_array_f add_to_array, void* const outputs)
1034{
1035 ccv_cnnp_model_batch_norm_t* const self = (ccv_cnnp_model_batch_norm_t*)super;
1036 int i;
1037 if (self->retainables)
1038 for (i = 0; i < self->retainables->rnum; i++)
1039 {
1040 const ccv_nnc_tensor_symbol_t symbol = *(ccv_nnc_tensor_symbol_t*)ccv_array_get(self->retainables, i)((void*)(((char*)((self->retainables)->data)) + (size_t
)(self->retainables)->rsize * (size_t)(i)))
;
1041 add_to_array(outputs, symbol, 0);
1042 }
1043}
1044
1045static void _ccv_cnnp_batch_norm_set_is_test(ccv_cnnp_model_t* const super, const int is_test, const ccv_cnnp_cmd_updater_f updater, void* const context)
1046{
1047 ccv_cnnp_model_batch_norm_t* const self = (ccv_cnnp_model_batch_norm_t*)super;
1048 if (self->batch_norm.graph)
1049 {
1050 self->params.bnorm.is_test = is_test;
1051 updater(context, self->batch_norm, ccv_nnc_cmd(CCV_NNC_BATCH_NORM_FORWARD, 0, self->params, 0), ccv_nnc_no_hint);
1052 }
1053}
1054
1055static void _ccv_cnnp_batch_norm_deinit(ccv_cnnp_model_t* const super)
1056{
1057 ccv_cnnp_model_batch_norm_t* const self = (ccv_cnnp_model_batch_norm_t*)super;
1058 if (self->zero_inits)
1059 ccv_array_free(self->zero_inits);
1060 if (self->retainables)
1061 ccv_array_free(self->retainables);
1062}
1063
1064static ccv_cnnp_model_t* _ccv_cnnp_batch_norm_copy(const ccv_cnnp_model_t* const super, void* const context);
1065
1066static const ccv_cnnp_model_vtab_t ccv_cnnp_batch_norm_isa = {
1067 .build = _ccv_cnnp_batch_norm_build,
1068 .init_states = _ccv_cnnp_batch_norm_init_states,
1069 .add_to_parameter = _ccv_cnnp_batch_norm_add_to_parameter,
1070 .add_to_output = _ccv_cnnp_batch_norm_add_to_output,
1071 .copy = _ccv_cnnp_batch_norm_copy,
1072 .set_is_test = _ccv_cnnp_batch_norm_set_is_test,
1073 .deinit = _ccv_cnnp_batch_norm_deinit,
1074};
1075
1076ccv_cnnp_model_t* ccv_cnnp_batch_norm(const float momentum, const float epsilon, const int is_trainable, const char* const name)
1077{
1078 ccv_cnnp_model_batch_norm_t* const model_batch_norm = (ccv_cnnp_model_batch_norm_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_batch_norm_t));
1079 model_batch_norm->super.isa = &ccv_cnnp_batch_norm_isa;
1080 model_batch_norm->super.input_size = 1;
1081 model_batch_norm->super.outputs = &model_batch_norm->output;
1082 model_batch_norm->super.output_size = 1;
1083 model_batch_norm->super.is_trainable = is_trainable;
1084 ccv_cnnp_model_copy_name(&model_batch_norm->super, name);
1085 model_batch_norm->scale.d = CCV_NNC_NO_TENSOR_SYMBOL;
1086 model_batch_norm->scale.graph = 0;
1087 model_batch_norm->bias.d = CCV_NNC_NO_TENSOR_SYMBOL;
1088 model_batch_norm->bias.graph = 0;
1089 model_batch_norm->params.bnorm.momentum = momentum;
1090 model_batch_norm->params.bnorm.epsilon = epsilon;
1091 return (ccv_cnnp_model_t*)model_batch_norm;
1092}
1093
1094static ccv_cnnp_model_t* _ccv_cnnp_batch_norm_copy(const ccv_cnnp_model_t* const super, void* const context)
1095{
1096 const ccv_cnnp_model_batch_norm_t* const self = (const ccv_cnnp_model_batch_norm_t*)super;
1097 return ccv_cnnp_batch_norm(self->params.bnorm.momentum, self->params.bnorm.epsilon, self->super.is_trainable, self->super.name);
1098}
1099
1100// MARK - Convolution Layer
1101
1102typedef struct {
1103 ccv_cnnp_model_t super;
1104 ccv_nnc_tensor_symbol_t output;
1105 ccv_nnc_tensor_symbol_t weights;
1106 ccv_nnc_tensor_symbol_t bias;
1107 int groups;
1108 int filters;
1109 int kdim[CCV_NNC_MAX_DIM_ALLOC(12)];
1110 int dilation[CCV_NNC_MAX_DIM_ALLOC(12)];
1111 int no_bias;
1112 int format;
1113 ccv_nnc_hint_t hint;
1114} ccv_cnnp_model_convolution_t;
1115
1116static void _ccv_cnnp_convolution_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
1117{
1118 ccv_cnnp_model_convolution_t* const self = (ccv_cnnp_model_convolution_t*)super;
1119 PRINT(CCV_CLI_VERBOSE, "[cnnp_convolution_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_convolution_build] -\n"); fflush(stdout); } }
while (0)
;
1120 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 1120, __extension__ __PRETTY_FUNCTION__); }))
;
1121 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 1121, __extension__ __PRETTY_FUNCTION__
); }))
;
1122 const ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
1123 int i;
1124 const int k_nd = ccv_nnc_tensor_nd(self->kdim);
1125 const int nd = k_nd + 2;
1126 ccv_nnc_tensor_param_t weights_params = params;
1127 if (self->format)
1128 weights_params.format = self->format;
1129 ccv_nnc_tensor_set_n(&weights_params, self->filters);
1130 const int a_nd = ccv_nnc_tensor_nd(params.dim);
1131 int c;
1132 switch (params.format)
1133 {
1134 case CCV_TENSOR_FORMAT_NHWC:
1135 c = params.dim[a_nd - 1];
1136 break;
1137 case CCV_TENSOR_FORMAT_NCHW:
1138 if (a_nd == k_nd + 1)
1139 c = params.dim[0];
1140 else
1141 c = params.dim[a_nd <= 1 ? 0 : 1];
1142 break;
1143 case CCV_TENSOR_FORMAT_CHWN:
1144 c = params.dim[0];
1145 break;
1146 }
1147 assert(c % self->groups == 0)((void) sizeof ((c % self->groups == 0) ? 1 : 0), __extension__
({ if (c % self->groups == 0) ; else __assert_fail ("c % self->groups == 0"
, "ccv_cnnp_model_addons.c", 1147, __extension__ __PRETTY_FUNCTION__
); }))
;
1148 ccv_nnc_tensor_set_c(&weights_params, nd, c / self->groups);
1149 int hw = -1;
1150 if (weights_params.format == CCV_TENSOR_FORMAT_NHWC || weights_params.format == CCV_TENSOR_FORMAT_CHWN)
1151 hw = 1;
1152 else if (weights_params.format == CCV_TENSOR_FORMAT_NCHW)
1153 hw = 2;
1154 assert(hw >= 0)((void) sizeof ((hw >= 0) ? 1 : 0), __extension__ ({ if (hw
>= 0) ; else __assert_fail ("hw >= 0", "ccv_cnnp_model_addons.c"
, 1154, __extension__ __PRETTY_FUNCTION__); }))
;
1155 for (i = 0; i < k_nd; i++)
1156 weights_params.dim[i + hw] = self->kdim[i];
1157 if (!self->weights.graph)
1158 self->weights = ccv_nnc_tensor_symbol_new(graph, weights_params, "weights");
1159 assert(self->weights.graph == graph)((void) sizeof ((self->weights.graph == graph) ? 1 : 0), __extension__
({ if (self->weights.graph == graph) ; else __assert_fail
("self->weights.graph == graph", "ccv_cnnp_model_addons.c"
, 1159, __extension__ __PRETTY_FUNCTION__); }))
;
1160 ccv_nnc_tensor_param_t bias_params = params;
1161 if (self->format)
1162 bias_params.format = self->format;
1163 memset(bias_params.dim, 0, sizeof(bias_params.dim));
1164 bias_params.dim[0] = self->filters;
1165 ccv_nnc_cmd_t cmd = CMD_CONVOLUTION_FORWARD(self->groups, self->filters)ccv_nnc_cmd(CCV_NNC_CONVOLUTION_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={}},.convolution={.count=self->filters,.groups
=self->groups}}), 0)
;
1166 for (i = 0; i < k_nd; i++)
1167 cmd.info.size.dim[i] = self->kdim[i];
1168 cmd.info.size.dim[k_nd] = c;
1169 memcpy(cmd.info.convolution.dilation, self->dilation, sizeof(self->dilation));
1170 ccv_nnc_tensor_param_t output_params;
1171 // Dilate weight size based on the dilation factor.
1172 for (i = 0; i < k_nd; i++)
1173 weights_params.dim[i + hw] = (self->kdim[i] - 1) * ccv_max(self->dilation[i], 1)({ typeof (self->dilation[i]) _a = (self->dilation[i]);
typeof (1) _b = (1); (_a > _b) ? _a : _b; })
+ 1;
1174 ccv_nnc_hint_tensor_auto(cmd, (ccv_nnc_tensor_param_t []){
1175 params,
1176 weights_params,
1177 bias_params,
1178 }, 3, self->hint, &output_params, 1);
1179 const ccv_nnc_tensor_symbol_t output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
1180 ccv_nnc_graph_exec_symbol_t convolution;
1181 if (self->no_bias)
1182 convolution = ccv_nnc_graph_exec_symbol_new(graph, cmd, TENSOR_SYMBOL_LIST(inputs[0], self->weights)(const ccv_nnc_tensor_symbol_t []){inputs[0], self->weights
}, (1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output)(const ccv_nnc_tensor_symbol_t []){output}, (1 +1 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "convolution");
1183 else {
1184 if (!self->bias.graph)
1185 self->bias = ccv_nnc_tensor_symbol_new(graph, bias_params, "bias");
1186 convolution = ccv_nnc_graph_exec_symbol_new(graph, cmd, TENSOR_SYMBOL_LIST(inputs[0], self->weights, self->bias)(const ccv_nnc_tensor_symbol_t []){inputs[0], self->weights
, self->bias}, (1 +1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output)(const ccv_nnc_tensor_symbol_t []){output}, (1 +1 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "convolution");
1187 }
1188 ccv_nnc_graph_exec_symbol_set_hint(graph, convolution, self->hint);
1189 outputs[0] = output;
1190}
1191
1192static void _ccv_cnnp_convolution_init_states(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_cnnp_state_initializer_f initializer, void* const context)
1193{
1194 ccv_cnnp_model_convolution_t* const self = (ccv_cnnp_model_convolution_t*)super;
1195 const ccv_nnc_tensor_param_t weight_params = ccv_nnc_tensor_symbol_params(graph, self->weights);
1196 const int n = ccv_max(ccv_nnc_tensor_get_n(weight_params), 1)({ typeof (ccv_nnc_tensor_get_n(weight_params)) _a = (ccv_nnc_tensor_get_n
(weight_params)); typeof (1) _b = (1); (_a > _b) ? _a : _b
; })
;
1197 const int count = ccv_nnc_tensor_count(weight_params);
1198 const float std = sqrtf(2) / sqrtf(count / n);
1199 const float bound = sqrtf(3) * std;
1200 initializer(context, CMD_RANDOM_UNIFORM_FORWARD(-bound, bound)ccv_nnc_cmd(CCV_NNC_RANDOM_UNIFORM_FORWARD, 0, (ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.blas={.a={-bound, bound}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->weights);
1201 if (self->bias.graph)
1202 initializer(context, CMD_SET_FORWARD(0)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={0,}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->bias);
1203}
1204
1205static void _ccv_cnnp_convolution_add_to_parameter(ccv_cnnp_model_t* const super, const ccv_cnnp_add_to_array_f add_to_array, void* const parameters, const int is_trainable)
1206{
1207 ccv_cnnp_model_convolution_t* const self = (ccv_cnnp_model_convolution_t*)super;
1208 add_to_array(parameters, self->weights, is_trainable);
1209 if (self->bias.graph)
1210 add_to_array(parameters, self->bias, is_trainable);
1211}
1212
1213static ccv_cnnp_model_t* _ccv_cnnp_convolution_copy(const ccv_cnnp_model_t* const super, void* const context);
1214
1215static const ccv_cnnp_model_vtab_t ccv_cnnp_convolution_isa = {
1216 .build = _ccv_cnnp_convolution_build,
1217 .init_states = _ccv_cnnp_convolution_init_states,
1218 .add_to_parameter = _ccv_cnnp_convolution_add_to_parameter,
1219 .copy = _ccv_cnnp_convolution_copy,
1220};
1221
1222ccv_cnnp_model_t* ccv_cnnp_convolution(const int groups, const int filters, const int kdim[CCV_NNC_MAX_DIM_ALLOC(12)], const int dilation[CCV_NNC_MAX_DIM_ALLOC(12)], const int no_bias, ccv_nnc_hint_t hint, const int format, const int is_trainable, const char* const name)
1223{
1224 ccv_cnnp_model_convolution_t* const model_convolution = (ccv_cnnp_model_convolution_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_convolution_t));
1225 model_convolution->super.isa = &ccv_cnnp_convolution_isa;
1226 model_convolution->super.input_size = 1;
1227 model_convolution->super.outputs = &model_convolution->output;
1228 model_convolution->super.output_size = 1;
1229 model_convolution->super.is_trainable = is_trainable;
1230 ccv_cnnp_model_copy_name(&model_convolution->super, name);
1231 model_convolution->weights.d = CCV_NNC_NO_TENSOR_SYMBOL;
1232 model_convolution->weights.graph = 0;
1233 model_convolution->bias.d = CCV_NNC_NO_TENSOR_SYMBOL;
1234 model_convolution->bias.graph = 0;
1235 model_convolution->groups = groups;
1236 model_convolution->filters = filters;
1237 memcpy(model_convolution->kdim, kdim, sizeof(model_convolution->kdim));
1238 memcpy(model_convolution->dilation, dilation, sizeof(model_convolution->dilation));
1239 model_convolution->no_bias = no_bias;
1240 model_convolution->hint = hint;
1241 model_convolution->format = format;
1242 return (ccv_cnnp_model_t*)model_convolution;
1243}
1244
1245static ccv_cnnp_model_t* _ccv_cnnp_convolution_copy(const ccv_cnnp_model_t* const super, void* const context)
1246{
1247 ccv_cnnp_model_convolution_t* const self = (ccv_cnnp_model_convolution_t*)super;
1248 return ccv_cnnp_convolution(self->groups, self->filters, self->kdim, self->dilation, self->no_bias, self->hint, self->format, self->super.is_trainable, self->super.name);
1249}
1250
1251// MARK - Convolution Transpose Layer
1252
1253typedef struct {
1254 ccv_cnnp_model_t super;
1255 ccv_nnc_tensor_symbol_t output;
1256 ccv_nnc_tensor_symbol_t weights;
1257 ccv_nnc_tensor_symbol_t bias;
1258 int groups;
1259 int filters;
1260 int kdim[CCV_NNC_MAX_DIM_ALLOC(12)];
1261 int dilation[CCV_NNC_MAX_DIM_ALLOC(12)];
1262 int output_padding;
1263 int no_bias;
1264 int format;
1265 ccv_nnc_hint_t hint;
1266} ccv_cnnp_model_convolution_transpose_t;
1267
1268static void _ccv_cnnp_convolution_transpose_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
1269{
1270 ccv_cnnp_model_convolution_transpose_t* const self = (ccv_cnnp_model_convolution_transpose_t*)super;
1271 PRINT(CCV_CLI_VERBOSE, "[cnnp_convolution_transpose_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_convolution_transpose_build] -\n"); fflush(stdout
); } } while (0)
;
1272 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 1272, __extension__ __PRETTY_FUNCTION__); }))
;
1273 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 1273, __extension__ __PRETTY_FUNCTION__
); }))
;
1274 const ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
1275 int i;
1276 const int nd = CCV_NNC_MAX_DIM(2) + 2;
1277 ccv_nnc_tensor_param_t weights_params = params;
1278 if (self->format)
1279 weights_params.format = self->format;
1280 const int c = ccv_nnc_tensor_get_c(params);
1281 ccv_nnc_tensor_set_n(&weights_params, c);
1282 assert(c % self->groups == 0)((void) sizeof ((c % self->groups == 0) ? 1 : 0), __extension__
({ if (c % self->groups == 0) ; else __assert_fail ("c % self->groups == 0"
, "ccv_cnnp_model_addons.c", 1282, __extension__ __PRETTY_FUNCTION__
); }))
;
1283 ccv_nnc_tensor_set_c(&weights_params, nd, self->filters / self->groups);
1284 const int hw = ccv_nnc_tensor_hw(weights_params, nd, CCV_NNC_MAX_DIM(2));
1285 assert(hw >= 0)((void) sizeof ((hw >= 0) ? 1 : 0), __extension__ ({ if (hw
>= 0) ; else __assert_fail ("hw >= 0", "ccv_cnnp_model_addons.c"
, 1285, __extension__ __PRETTY_FUNCTION__); }))
;
1286 for (i = 0; i < CCV_NNC_MAX_DIM(2); i++)
1287 weights_params.dim[i + hw] = self->kdim[i];
1288 if (!self->weights.graph)
1289 self->weights = ccv_nnc_tensor_symbol_new(graph, weights_params, "weights");
1290 assert(self->weights.graph == graph)((void) sizeof ((self->weights.graph == graph) ? 1 : 0), __extension__
({ if (self->weights.graph == graph) ; else __assert_fail
("self->weights.graph == graph", "ccv_cnnp_model_addons.c"
, 1290, __extension__ __PRETTY_FUNCTION__); }))
;
1291 ccv_nnc_tensor_param_t bias_params = params;
1292 if (self->format)
1293 bias_params.format = self->format;
1294 memset(bias_params.dim, 0, sizeof(bias_params.dim));
1295 bias_params.dim[0] = self->filters;
1296 ccv_nnc_cmd_t cmd = CMD_CONVOLUTION_TRANSPOSE_FORWARD(self->groups, self->filters, self->output_padding)ccv_nnc_cmd(CCV_NNC_CONVOLUTION_TRANSPOSE_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={}},.convolution_transpose={.count=self->filters
,.groups=self->groups,.output_padding=self->output_padding
}}), 0)
;
1297 for (i = 0; i < CCV_NNC_MAX_DIM(2); i++)
1298 cmd.info.size.dim[i] = self->kdim[i];
1299 cmd.info.size.dim[CCV_NNC_MAX_DIM(2)] = c;
1300 memcpy(cmd.info.convolution_transpose.dilation, self->dilation, sizeof(self->dilation));
1301 ccv_nnc_tensor_param_t output_params;
1302 // Dilate weight size based on the dilation factor.
1303 for (i = 0; i < CCV_NNC_MAX_DIM(2); i++)
1304 weights_params.dim[i + hw] = (self->kdim[i] - 1) * ccv_max(self->dilation[i], 1)({ typeof (self->dilation[i]) _a = (self->dilation[i]);
typeof (1) _b = (1); (_a > _b) ? _a : _b; })
+ 1;
1305 ccv_nnc_hint_tensor_auto(cmd, (ccv_nnc_tensor_param_t []){
1306 params,
1307 weights_params,
1308 bias_params,
1309 }, 3, self->hint, &output_params, 1);
1310 const ccv_nnc_tensor_symbol_t output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
1311 ccv_nnc_graph_exec_symbol_t convolution_transpose;
1312 if (self->no_bias)
1313 convolution_transpose = ccv_nnc_graph_exec_symbol_new(graph, cmd, TENSOR_SYMBOL_LIST(inputs[0], self->weights)(const ccv_nnc_tensor_symbol_t []){inputs[0], self->weights
}, (1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output)(const ccv_nnc_tensor_symbol_t []){output}, (1 +1 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "convolution_transpose");
1314 else {
1315 if (!self->bias.graph)
1316 self->bias = ccv_nnc_tensor_symbol_new(graph, bias_params, "bias");
1317 convolution_transpose = ccv_nnc_graph_exec_symbol_new(graph, cmd, TENSOR_SYMBOL_LIST(inputs[0], self->weights, self->bias)(const ccv_nnc_tensor_symbol_t []){inputs[0], self->weights
, self->bias}, (1 +1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output)(const ccv_nnc_tensor_symbol_t []){output}, (1 +1 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "convolution_transpose");
1318 }
1319 ccv_nnc_graph_exec_symbol_set_hint(graph, convolution_transpose, self->hint);
1320 outputs[0] = output;
1321}
1322
1323static void _ccv_cnnp_convolution_transpose_init_states(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_cnnp_state_initializer_f initializer, void* const context)
1324{
1325 ccv_cnnp_model_convolution_transpose_t* const self = (ccv_cnnp_model_convolution_transpose_t*)super;
1326 const ccv_nnc_tensor_param_t weight_params = ccv_nnc_tensor_symbol_params(graph, self->weights);
1327 const int n = ccv_max(ccv_nnc_tensor_get_n(weight_params), 1)({ typeof (ccv_nnc_tensor_get_n(weight_params)) _a = (ccv_nnc_tensor_get_n
(weight_params)); typeof (1) _b = (1); (_a > _b) ? _a : _b
; })
;
1328 const int count = ccv_nnc_tensor_count(weight_params);
1329 const float std = sqrtf(2) / sqrtf(count / n);
1330 const float bound = sqrtf(3) * std;
1331 initializer(context, CMD_RANDOM_UNIFORM_FORWARD(-bound, bound)ccv_nnc_cmd(CCV_NNC_RANDOM_UNIFORM_FORWARD, 0, (ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.blas={.a={-bound, bound}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->weights);
1332 if (self->bias.graph)
1333 initializer(context, CMD_SET_FORWARD(0)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={0,}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->bias);
1334}
1335
1336static void _ccv_cnnp_convolution_transpose_add_to_parameter(ccv_cnnp_model_t* const super, const ccv_cnnp_add_to_array_f add_to_array, void* const parameters, const int is_trainable)
1337{
1338 ccv_cnnp_model_convolution_transpose_t* const self = (ccv_cnnp_model_convolution_transpose_t*)super;
1339 add_to_array(parameters, self->weights, is_trainable);
1340 if (self->bias.graph)
1341 add_to_array(parameters, self->bias, is_trainable);
1342}
1343
1344static ccv_cnnp_model_t* _ccv_cnnp_convolution_transpose_copy(const ccv_cnnp_model_t* const super, void* const context);
1345
1346static const ccv_cnnp_model_vtab_t ccv_cnnp_convolution_transpose_isa = {
1347 .build = _ccv_cnnp_convolution_transpose_build,
1348 .init_states = _ccv_cnnp_convolution_transpose_init_states,
1349 .add_to_parameter = _ccv_cnnp_convolution_transpose_add_to_parameter,
1350 .copy = _ccv_cnnp_convolution_transpose_copy,
1351};
1352
1353ccv_cnnp_model_t* ccv_cnnp_convolution_transpose(const int groups, const int filters, const int kdim[CCV_NNC_MAX_DIM_ALLOC(12)], const int dilation[CCV_NNC_MAX_DIM_ALLOC(12)], const int output_padding, const int no_bias, ccv_nnc_hint_t hint, const int format, const int is_trainable, const char* const name)
1354{
1355 ccv_cnnp_model_convolution_transpose_t* const model_convolution_transpose = (ccv_cnnp_model_convolution_transpose_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_convolution_transpose_t));
1356 model_convolution_transpose->super.isa = &ccv_cnnp_convolution_transpose_isa;
1357 model_convolution_transpose->super.input_size = 1;
1358 model_convolution_transpose->super.outputs = &model_convolution_transpose->output;
1359 model_convolution_transpose->super.output_size = 1;
1360 model_convolution_transpose->super.is_trainable = is_trainable;
1361 ccv_cnnp_model_copy_name(&model_convolution_transpose->super, name);
1362 model_convolution_transpose->weights.d = CCV_NNC_NO_TENSOR_SYMBOL;
1363 model_convolution_transpose->weights.graph = 0;
1364 model_convolution_transpose->bias.d = CCV_NNC_NO_TENSOR_SYMBOL;
1365 model_convolution_transpose->bias.graph = 0;
1366 model_convolution_transpose->groups = groups;
1367 model_convolution_transpose->filters = filters;
1368 memcpy(model_convolution_transpose->kdim, kdim, sizeof(model_convolution_transpose->kdim));
1369 memcpy(model_convolution_transpose->dilation, dilation, sizeof(model_convolution_transpose->dilation));
1370 model_convolution_transpose->output_padding = output_padding;
1371 model_convolution_transpose->no_bias = no_bias;
1372 model_convolution_transpose->hint = hint;
1373 model_convolution_transpose->format = format;
1374 return (ccv_cnnp_model_t*)model_convolution_transpose;
1375}
1376
1377static ccv_cnnp_model_t* _ccv_cnnp_convolution_transpose_copy(const ccv_cnnp_model_t* const super, void* const context)
1378{
1379 ccv_cnnp_model_convolution_transpose_t* const self = (ccv_cnnp_model_convolution_transpose_t*)super;
1380 return ccv_cnnp_convolution_transpose(self->groups, self->filters, self->kdim, self->dilation, self->output_padding, self->no_bias, self->hint, self->format, self->super.is_trainable, self->super.name);
1381}
1382
1383// MARK - Dense Layer
1384
1385typedef struct {
1386 ccv_cnnp_model_t super;
1387 ccv_nnc_tensor_symbol_t output;
1388 ccv_nnc_tensor_symbol_t weights;
1389 ccv_nnc_tensor_symbol_t bias;
1390 int count;
1391 int no_bias;
1392 int flags;
1393} ccv_cnnp_model_dense_t;
1394
1395static void _ccv_cnnp_dense_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
1396{
1397 ccv_cnnp_model_dense_t* const self = (ccv_cnnp_model_dense_t*)super;
1398 PRINT(CCV_CLI_VERBOSE, "[cnnp_dense_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_dense_build] -\n"); fflush(stdout); } } while
(0)
;
1399 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 1399, __extension__ __PRETTY_FUNCTION__); }))
;
1400 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 1400, __extension__ __PRETTY_FUNCTION__
); }))
;
1401 const ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
1402 ccv_nnc_tensor_param_t weights_params = params;
1403 memset(weights_params.dim, 0, sizeof(weights_params.dim));
1404 weights_params.dim[0] = self->count;
1405 weights_params.dim[1] = params.dim[ccv_nnc_tensor_nd(params.dim) - 1];
1406 if (!self->weights.graph)
1407 self->weights = ccv_nnc_tensor_symbol_new(graph, weights_params, "weights");
1408 assert(self->weights.graph == graph)((void) sizeof ((self->weights.graph == graph) ? 1 : 0), __extension__
({ if (self->weights.graph == graph) ; else __assert_fail
("self->weights.graph == graph", "ccv_cnnp_model_addons.c"
, 1408, __extension__ __PRETTY_FUNCTION__); }))
;
1409 ccv_nnc_tensor_param_t bias_params = params;
1410 memset(bias_params.dim, 0, sizeof(bias_params.dim));
1411 bias_params.dim[0] = self->count;
1412 ccv_nnc_cmd_t cmd = {0};
1413 cmd.cmd = CCV_NNC_GEMM_FORWARD;
1414 cmd.info.blas.a[0] = 1;
1415 cmd.info.blas.a[1] = 1;
1416 cmd.info.blas.transpose_b[0] = 0;
1417 cmd.info.blas.transpose_b[1] = 1;
1418 cmd.info.blas.flags = self->flags;
1419 ccv_nnc_tensor_param_t output_params;
1420 ccv_nnc_hint_tensor_auto(cmd, (ccv_nnc_tensor_param_t []){
1421 params,
1422 weights_params,
1423 bias_params,
1424 }, 3, ccv_nnc_no_hint, &output_params, 1);
1425 const ccv_nnc_tensor_symbol_t output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
1426 if (self->no_bias)
1427 ccv_nnc_graph_exec_symbol_new(graph, cmd, TENSOR_SYMBOL_LIST(inputs[0], self->weights)(const ccv_nnc_tensor_symbol_t []){inputs[0], self->weights
}, (1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output)(const ccv_nnc_tensor_symbol_t []){output}, (1 +1 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "dense");
1428 else {
1429 if (!self->bias.graph)
1430 self->bias = ccv_nnc_tensor_symbol_new(graph, bias_params, "bias");
1431 ccv_nnc_graph_exec_symbol_new(graph, cmd, TENSOR_SYMBOL_LIST(inputs[0], self->weights, self->bias)(const ccv_nnc_tensor_symbol_t []){inputs[0], self->weights
, self->bias}, (1 +1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output)(const ccv_nnc_tensor_symbol_t []){output}, (1 +1 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "dense");
1432 }
1433 outputs[0] = output;
1434}
1435
1436static void _ccv_cnnp_dense_init_states(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_cnnp_state_initializer_f initializer, void* const context)
1437{
1438 ccv_cnnp_model_dense_t* const self = (ccv_cnnp_model_dense_t*)super;
1439 const ccv_nnc_tensor_param_t weight_params = ccv_nnc_tensor_symbol_params(graph, self->weights);
1440 const int c = weight_params.dim[1];
1441 const float std = sqrtf(2) / sqrtf(c);
1442 const float bound = sqrtf(3) * std;
1443 initializer(context, CMD_RANDOM_UNIFORM_FORWARD(-bound, bound)ccv_nnc_cmd(CCV_NNC_RANDOM_UNIFORM_FORWARD, 0, (ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.blas={.a={-bound, bound}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->weights);
1444 if (self->bias.graph)
1445 initializer(context, CMD_SET_FORWARD(0)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={0,}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->bias);
1446}
1447
1448static void _ccv_cnnp_dense_add_to_parameter(ccv_cnnp_model_t* const super, const ccv_cnnp_add_to_array_f add_to_array, void* const parameters, const int is_trainable)
1449{
1450 ccv_cnnp_model_dense_t* const self = (ccv_cnnp_model_dense_t*)super;
1451 add_to_array(parameters, self->weights, is_trainable);
1452 if (self->bias.graph)
1453 add_to_array(parameters, self->bias, is_trainable);
1454}
1455
1456static ccv_cnnp_model_t* _ccv_cnnp_dense_copy(const ccv_cnnp_model_t* const super, void* const context);
1457
1458static const ccv_cnnp_model_vtab_t ccv_cnnp_dense_isa = {
1459 .build = _ccv_cnnp_dense_build,
1460 .init_states = _ccv_cnnp_dense_init_states,
1461 .add_to_parameter = _ccv_cnnp_dense_add_to_parameter,
1462 .copy = _ccv_cnnp_dense_copy,
1463};
1464
1465ccv_cnnp_model_t* ccv_cnnp_dense(const int count, const int no_bias, const int flags, const int is_trainable, const char* const name)
1466{
1467 ccv_cnnp_model_dense_t* const model_dense = (ccv_cnnp_model_dense_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_dense_t));
1468 model_dense->super.isa = &ccv_cnnp_dense_isa;
1469 model_dense->super.input_size = 1;
1470 model_dense->super.outputs = &model_dense->output;
1471 model_dense->super.output_size = 1;
1472 model_dense->super.is_trainable = is_trainable;
1473 ccv_cnnp_model_copy_name(&model_dense->super, name);
1474 model_dense->weights.d = CCV_NNC_NO_TENSOR_SYMBOL;
1475 model_dense->weights.graph = 0;
1476 model_dense->bias.d = CCV_NNC_NO_TENSOR_SYMBOL;
1477 model_dense->bias.graph = 0;
1478 model_dense->count = count;
1479 model_dense->no_bias = no_bias;
1480 model_dense->flags = flags;
1481 return (ccv_cnnp_model_t*)model_dense;
1482}
1483
1484static ccv_cnnp_model_t* _ccv_cnnp_dense_copy(const ccv_cnnp_model_t* const super, void* const context)
1485{
1486 const ccv_cnnp_model_dense_t* const self = (const ccv_cnnp_model_dense_t*)super;
1487 return ccv_cnnp_dense(self->count, self->no_bias, self->flags, self->super.is_trainable, self->super.name);
1488}
1489
1490// MARK - Pool Layers
1491
1492typedef struct {
1493 ccv_cnnp_model_t super;
1494 ccv_nnc_tensor_symbol_t output;
1495 int kdim[CCV_NNC_MAX_DIM_ALLOC(12)];
1496 ccv_nnc_hint_t hint;
1497} ccv_cnnp_model_pool_t;
1498
1499static void _ccv_cnnp_max_pool_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
1500{
1501 ccv_cnnp_model_pool_t* const self = (ccv_cnnp_model_pool_t*)super;
1502 PRINT(CCV_CLI_VERBOSE, "[cnnp_max_pool_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_max_pool_build] -\n"); fflush(stdout); } } while
(0)
;
1503 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 1503, __extension__ __PRETTY_FUNCTION__); }))
;
1504 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 1504, __extension__ __PRETTY_FUNCTION__
); }))
;
1505 const ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
1506 const int hw = ccv_nnc_tensor_hw(params, ccv_nnc_tensor_nd(params.dim), CCV_NNC_MAX_DIM(2));
1507 ccv_nnc_cmd_t cmd;
1508 if (hw >= 0 && self->kdim[0] == 0 && self->kdim[1] == 0)
1509 cmd = CMD_MAX_POOL_FORWARD(params.dim[hw], params.dim[hw + 1])ccv_nnc_cmd(CCV_NNC_MAX_POOL_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={params.dim[hw], params.dim[hw + 1],1}}}), 0)
;
1510 else
1511 cmd = CMD_MAX_POOL_FORWARD(self->kdim[0], self->kdim[1])ccv_nnc_cmd(CCV_NNC_MAX_POOL_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={self->kdim[0], self->kdim[1],1}}}), 0)
;
1512 ccv_nnc_tensor_param_t output_params;
1513 ccv_nnc_hint_tensor_auto(cmd, &params, 1, self->hint, &output_params, 1);
1514 const ccv_nnc_tensor_symbol_t pool_output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
1515 const ccv_nnc_graph_exec_symbol_t exec = ccv_nnc_graph_exec_symbol_new(graph, cmd, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(pool_output)(const ccv_nnc_tensor_symbol_t []){pool_output}, (1 +1 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "max_pool");
1516 ccv_nnc_graph_exec_symbol_set_hint(graph, exec, self->hint);
1517 outputs[0] = pool_output;
1518}
1519
1520static ccv_cnnp_model_t* _ccv_cnnp_max_pool_copy(const ccv_cnnp_model_t* const super, void* const context);
1521
1522static const ccv_cnnp_model_vtab_t ccv_cnnp_max_pool_isa = {
1523 .build = _ccv_cnnp_max_pool_build,
1524 .copy = _ccv_cnnp_max_pool_copy,
1525};
1526
1527ccv_cnnp_model_t* ccv_cnnp_max_pool(const int kdim[CCV_NNC_MAX_DIM_ALLOC(12)], const ccv_nnc_hint_t hint, const char* const name)
1528{
1529 ccv_cnnp_model_pool_t* const model_pool = (ccv_cnnp_model_pool_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_pool_t));
1530 model_pool->super.isa = &ccv_cnnp_max_pool_isa;
1531 model_pool->super.input_size = 1;
1532 model_pool->super.outputs = &model_pool->output;
1533 model_pool->super.output_size = 1;
1534 ccv_cnnp_model_copy_name(&model_pool->super, name);
1535 memcpy(model_pool->kdim, kdim, sizeof(model_pool->kdim));
1536 model_pool->hint = hint;
1537 return (ccv_cnnp_model_t*)model_pool;
1538}
1539
1540static ccv_cnnp_model_t* _ccv_cnnp_max_pool_copy(const ccv_cnnp_model_t* const super, void* const context)
1541{
1542 const ccv_cnnp_model_pool_t* const self = (const ccv_cnnp_model_pool_t*)super;
1543 return ccv_cnnp_max_pool(self->kdim, self->hint, self->super.name);
1544}
1545
1546static void _ccv_cnnp_average_pool_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
1547{
1548 ccv_cnnp_model_pool_t* const self = (ccv_cnnp_model_pool_t*)super;
1549 PRINT(CCV_CLI_VERBOSE, "[cnnp_average_pool_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_average_pool_build] -\n"); fflush(stdout); }
} while (0)
;
1550 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 1550, __extension__ __PRETTY_FUNCTION__); }))
;
1551 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 1551, __extension__ __PRETTY_FUNCTION__
); }))
;
1552 const ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
1553 const int hw = ccv_nnc_tensor_hw(params, ccv_nnc_tensor_nd(params.dim), CCV_NNC_MAX_DIM(2));
1554 ccv_nnc_cmd_t cmd;
1555 if (hw >= 0 && self->kdim[0] == 0 && self->kdim[1] == 0)
1556 cmd = CMD_AVERAGE_POOL_FORWARD(params.dim[hw], params.dim[hw + 1])ccv_nnc_cmd(CCV_NNC_AVERAGE_POOL_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={params.dim[hw], params.dim[hw + 1],1}}}), 0)
;
1557 else
1558 cmd = CMD_AVERAGE_POOL_FORWARD(self->kdim[0], self->kdim[1])ccv_nnc_cmd(CCV_NNC_AVERAGE_POOL_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={self->kdim[0], self->kdim[1],1}}}), 0)
;
1559 ccv_nnc_tensor_param_t output_params;
1560 ccv_nnc_hint_tensor_auto(cmd, &params, 1, self->hint, &output_params, 1);
1561 const ccv_nnc_tensor_symbol_t pool_output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
1562 const ccv_nnc_graph_exec_symbol_t exec = ccv_nnc_graph_exec_symbol_new(graph, cmd, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(pool_output)(const ccv_nnc_tensor_symbol_t []){pool_output}, (1 +1 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "average_pool");
1563 ccv_nnc_graph_exec_symbol_set_hint(graph, exec, self->hint);
1564 outputs[0] = pool_output;
1565}
1566
1567static ccv_cnnp_model_t* _ccv_cnnp_average_pool_copy(const ccv_cnnp_model_t* const super, void* const context);
1568
1569static const ccv_cnnp_model_vtab_t ccv_cnnp_average_pool_isa = {
1570 .build = _ccv_cnnp_average_pool_build,
1571 .copy = _ccv_cnnp_average_pool_copy,
1572};
1573
1574ccv_cnnp_model_t* ccv_cnnp_average_pool(const int kdim[CCV_NNC_MAX_DIM_ALLOC(12)], const ccv_nnc_hint_t hint, const char* const name)
1575{
1576 ccv_cnnp_model_pool_t* const model_pool = (ccv_cnnp_model_pool_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_pool_t));
1577 model_pool->super.isa = &ccv_cnnp_average_pool_isa;
1578 model_pool->super.input_size = 1;
1579 model_pool->super.outputs = &model_pool->output;
1580 model_pool->super.output_size = 1;
1581 ccv_cnnp_model_copy_name(&model_pool->super, name);
1582 memcpy(model_pool->kdim, kdim, sizeof(model_pool->kdim));
1583 model_pool->hint = hint;
1584 return (ccv_cnnp_model_t*)model_pool;
1585}
1586
1587static ccv_cnnp_model_t* _ccv_cnnp_average_pool_copy(const ccv_cnnp_model_t* const super, void* const context)
1588{
1589 const ccv_cnnp_model_pool_t* const self = (const ccv_cnnp_model_pool_t*)super;
1590 return ccv_cnnp_average_pool(self->kdim, self->hint, self->super.name);
1591}
1592
1593// MARK - RELU Layer
1594
1595typedef struct {
1596 ccv_cnnp_model_t super;
1597 ccv_nnc_tensor_symbol_t output;
1598} ccv_cnnp_model_relu_t;
1599
1600static void _ccv_cnnp_relu_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
1601{
1602 PRINT(CCV_CLI_VERBOSE, "[cnnp_relu_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_relu_build] -\n"); fflush(stdout); } } while
(0)
;
1603 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 1603, __extension__ __PRETTY_FUNCTION__); }))
;
1604 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 1604, __extension__ __PRETTY_FUNCTION__
); }))
;
1605 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
1606 ccv_nnc_tensor_param_t output_params;
1607 const ccv_nnc_cmd_t relu = CMD_RELU_FORWARD()ccv_nnc_cmd(CCV_NNC_RELU_FORWARD, 0, ccv_nnc_cmd_auto, 0);
1608 ccv_nnc_hint_tensor_auto(relu, (ccv_nnc_tensor_param_t []){
1609 params,
1610 }, 1, ccv_nnc_no_hint, &output_params, 1);
1611 const ccv_nnc_tensor_symbol_t relu_output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
1612 ccv_nnc_graph_exec_symbol_new(graph, relu, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(relu_output)(const ccv_nnc_tensor_symbol_t []){relu_output}, (1 +1 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "relu");
1613 outputs[0] = relu_output;
1614}
1615
1616static ccv_cnnp_model_t* _ccv_cnnp_relu_copy(const ccv_cnnp_model_t* const self, void* const context);
1617
1618static const ccv_cnnp_model_vtab_t ccv_cnnp_relu_isa = {
1619 .build = _ccv_cnnp_relu_build,
1620 .copy = _ccv_cnnp_relu_copy,
1621};
1622
1623ccv_cnnp_model_t* ccv_cnnp_relu(const char* const name)
1624{
1625 ccv_cnnp_model_relu_t* const model_relu = (ccv_cnnp_model_relu_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_relu_t));
1626 model_relu->super.isa = &ccv_cnnp_relu_isa;
1627 model_relu->super.input_size = 1;
1628 model_relu->super.outputs = &model_relu->output;
1629 model_relu->super.output_size = 1;
1630 ccv_cnnp_model_copy_name(&model_relu->super, name);
1631 return (ccv_cnnp_model_t*)model_relu;
1632}
1633
1634static ccv_cnnp_model_t* _ccv_cnnp_relu_copy(const ccv_cnnp_model_t* const self, void* const context)
1635{
1636 return ccv_cnnp_relu(self->name);
1637}
1638
1639// MARK - Sigmoid Layer
1640
1641typedef struct {
1642 ccv_cnnp_model_t super;
1643 ccv_nnc_tensor_symbol_t output;
1644} ccv_cnnp_model_sigmoid_t;
1645
1646static void _ccv_cnnp_sigmoid_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
1647{
1648 PRINT(CCV_CLI_VERBOSE, "[cnnp_sigmoid_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_sigmoid_build] -\n"); fflush(stdout); } } while
(0)
;
1649 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 1649, __extension__ __PRETTY_FUNCTION__); }))
;
1650 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 1650, __extension__ __PRETTY_FUNCTION__
); }))
;
1651 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
1652 ccv_nnc_tensor_param_t output_params;
1653 const ccv_nnc_cmd_t sigmoid = CMD_SIGMOID_FORWARD()ccv_nnc_cmd(CCV_NNC_SIGMOID_FORWARD, 0, ccv_nnc_cmd_auto, 0);
1654 ccv_nnc_hint_tensor_auto(sigmoid, (ccv_nnc_tensor_param_t []){
1655 params,
1656 }, 1, ccv_nnc_no_hint, &output_params, 1);
1657 const ccv_nnc_tensor_symbol_t sigmoid_output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
1658 ccv_nnc_graph_exec_symbol_new(graph, sigmoid, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(sigmoid_output)(const ccv_nnc_tensor_symbol_t []){sigmoid_output}, (1 +1 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1
)
, "sigmoid");
1659 outputs[0] = sigmoid_output;
1660}
1661
1662static ccv_cnnp_model_t* _ccv_cnnp_sigmoid_copy(const ccv_cnnp_model_t* const self, void* const context);
1663
1664static const ccv_cnnp_model_vtab_t ccv_cnnp_sigmoid_isa = {
1665 .build = _ccv_cnnp_sigmoid_build,
1666 .copy = _ccv_cnnp_sigmoid_copy,
1667};
1668
1669ccv_cnnp_model_t* ccv_cnnp_sigmoid(const char* const name)
1670{
1671 ccv_cnnp_model_sigmoid_t* const model_sigmoid = (ccv_cnnp_model_sigmoid_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_sigmoid_t));
1672 model_sigmoid->super.isa = &ccv_cnnp_sigmoid_isa;
1673 model_sigmoid->super.input_size = 1;
1674 model_sigmoid->super.outputs = &model_sigmoid->output;
1675 model_sigmoid->super.output_size = 1;
1676 ccv_cnnp_model_copy_name(&model_sigmoid->super, name);
1677 return (ccv_cnnp_model_t*)model_sigmoid;
1678}
1679
1680static ccv_cnnp_model_t* _ccv_cnnp_sigmoid_copy(const ccv_cnnp_model_t* const self, void* const context)
1681{
1682 return ccv_cnnp_sigmoid(self->name);
1683}
1684
1685// MARK - Tanh Layer
1686
1687typedef struct {
1688 ccv_cnnp_model_t super;
1689 ccv_nnc_tensor_symbol_t output;
1690} ccv_cnnp_model_tanh_t;
1691
1692static void _ccv_cnnp_tanh_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
1693{
1694 PRINT(CCV_CLI_VERBOSE, "[cnnp_tanh_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_tanh_build] -\n"); fflush(stdout); } } while
(0)
;
1695 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 1695, __extension__ __PRETTY_FUNCTION__); }))
;
1696 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 1696, __extension__ __PRETTY_FUNCTION__
); }))
;
1697 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
1698 ccv_nnc_tensor_param_t output_params;
1699 const ccv_nnc_cmd_t tanh = CMD_TANH_FORWARD()ccv_nnc_cmd(CCV_NNC_TANH_FORWARD, 0, ccv_nnc_cmd_auto, 0);
1700 ccv_nnc_hint_tensor_auto(tanh, (ccv_nnc_tensor_param_t []){
1701 params,
1702 }, 1, ccv_nnc_no_hint, &output_params, 1);
1703 const ccv_nnc_tensor_symbol_t tanh_output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
1704 ccv_nnc_graph_exec_symbol_new(graph, tanh, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(tanh_output)(const ccv_nnc_tensor_symbol_t []){tanh_output}, (1 +1 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "tanh");
1705 outputs[0] = tanh_output;
1706}
1707
1708static ccv_cnnp_model_t* _ccv_cnnp_tanh_copy(const ccv_cnnp_model_t* const self, void* const context);
1709
1710static const ccv_cnnp_model_vtab_t ccv_cnnp_tanh_isa = {
1711 .build = _ccv_cnnp_tanh_build,
1712 .copy = _ccv_cnnp_tanh_copy,
1713};
1714
1715ccv_cnnp_model_t* ccv_cnnp_tanh(const char* const name)
1716{
1717 ccv_cnnp_model_tanh_t* const model_tanh = (ccv_cnnp_model_tanh_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_tanh_t));
1718 model_tanh->super.isa = &ccv_cnnp_tanh_isa;
1719 model_tanh->super.input_size = 1;
1720 model_tanh->super.outputs = &model_tanh->output;
1721 model_tanh->super.output_size = 1;
1722 ccv_cnnp_model_copy_name(&model_tanh->super, name);
1723 return (ccv_cnnp_model_t*)model_tanh;
1724}
1725
1726static ccv_cnnp_model_t* _ccv_cnnp_tanh_copy(const ccv_cnnp_model_t* const self, void* const context)
1727{
1728 return ccv_cnnp_tanh(self->name);
1729}
1730
1731// MARK - Swish Layer
1732
1733typedef struct {
1734 ccv_cnnp_model_t super;
1735 ccv_nnc_tensor_symbol_t output;
1736 float beta;
1737} ccv_cnnp_model_swish_t;
1738
1739static void _ccv_cnnp_swish_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
1740{
1741 PRINT(CCV_CLI_VERBOSE, "[cnnp_swish_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_swish_build] -\n"); fflush(stdout); } } while
(0)
;
1742 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 1742, __extension__ __PRETTY_FUNCTION__); }))
;
1743 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 1743, __extension__ __PRETTY_FUNCTION__
); }))
;
1744 ccv_cnnp_model_swish_t* const self = (ccv_cnnp_model_swish_t*)super;
1745 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
1746 ccv_nnc_tensor_param_t output_params;
1747 const ccv_nnc_cmd_t swish = CMD_SWISH_FORWARD(self->beta)ccv_nnc_cmd(CCV_NNC_SWISH_FORWARD, 0, ((ccv_nnc_cmd_param_t){
.size={.dim={1,1,1}},.swish={.beta=self->beta}}), 0)
;
1748 ccv_nnc_hint_tensor_auto(swish, (ccv_nnc_tensor_param_t []){
1749 params,
1750 }, 1, ccv_nnc_no_hint, &output_params, 1);
1751 const ccv_nnc_tensor_symbol_t swish_output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
1752 ccv_nnc_graph_exec_symbol_new(graph, swish, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(swish_output)(const ccv_nnc_tensor_symbol_t []){swish_output}, (1 +1 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "swish");
1753 outputs[0] = swish_output;
1754}
1755
1756static ccv_cnnp_model_t* _ccv_cnnp_swish_copy(const ccv_cnnp_model_t* const self, void* const context);
1757
1758static const ccv_cnnp_model_vtab_t ccv_cnnp_swish_isa = {
1759 .build = _ccv_cnnp_swish_build,
1760 .copy = _ccv_cnnp_swish_copy,
1761};
1762
1763ccv_cnnp_model_t* ccv_cnnp_swish(const float beta, const char* const name)
1764{
1765 ccv_cnnp_model_swish_t* const model_swish = (ccv_cnnp_model_swish_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_swish_t));
1766 model_swish->super.isa = &ccv_cnnp_swish_isa;
1767 model_swish->super.input_size = 1;
1768 model_swish->super.outputs = &model_swish->output;
1769 model_swish->super.output_size = 1;
1770 model_swish->beta = beta;
1771 ccv_cnnp_model_copy_name(&model_swish->super, name);
1772 return (ccv_cnnp_model_t*)model_swish;
1773}
1774
1775static ccv_cnnp_model_t* _ccv_cnnp_swish_copy(const ccv_cnnp_model_t* const self, void* const context)
1776{
1777 const ccv_cnnp_model_swish_t* const swish = (const ccv_cnnp_model_swish_t*)self;
1778 return ccv_cnnp_swish(swish->beta, self->name);
1779}
1780
1781// MARK - GELU Layer
1782
1783typedef struct {
1784 ccv_cnnp_model_t super;
1785 ccv_nnc_tensor_symbol_t output;
1786 int tanh;
1787} ccv_cnnp_model_gelu_t;
1788
1789static void _ccv_cnnp_gelu_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
1790{
1791 PRINT(CCV_CLI_VERBOSE, "[cnnp_gelu_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_gelu_build] -\n"); fflush(stdout); } } while
(0)
;
1792 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 1792, __extension__ __PRETTY_FUNCTION__); }))
;
1793 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 1793, __extension__ __PRETTY_FUNCTION__
); }))
;
1794 ccv_cnnp_model_gelu_t* const self = (ccv_cnnp_model_gelu_t*)super;
1795 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
1796 ccv_nnc_tensor_param_t output_params;
1797 const ccv_nnc_cmd_t gelu = CMD_GELU_FORWARD(self->tanh)ccv_nnc_cmd(CCV_NNC_GELU_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.gelu={.tanh=self->tanh}}, 0)
;
1798 ccv_nnc_hint_tensor_auto(gelu, (ccv_nnc_tensor_param_t []){
1799 params,
1800 }, 1, ccv_nnc_no_hint, &output_params, 1);
1801 const ccv_nnc_tensor_symbol_t gelu_output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
1802 ccv_nnc_graph_exec_symbol_new(graph, gelu, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(gelu_output)(const ccv_nnc_tensor_symbol_t []){gelu_output}, (1 +1 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "gelu");
1803 outputs[0] = gelu_output;
1804}
1805
1806static ccv_cnnp_model_t* _ccv_cnnp_gelu_copy(const ccv_cnnp_model_t* const self, void* const context);
1807
1808static const ccv_cnnp_model_vtab_t ccv_cnnp_gelu_isa = {
1809 .build = _ccv_cnnp_gelu_build,
1810 .copy = _ccv_cnnp_gelu_copy,
1811};
1812
1813ccv_cnnp_model_t* ccv_cnnp_gelu(const int tanh, const char* const name)
1814{
1815 ccv_cnnp_model_gelu_t* const model_gelu = (ccv_cnnp_model_gelu_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_gelu_t));
1816 model_gelu->super.isa = &ccv_cnnp_gelu_isa;
1817 model_gelu->super.input_size = 1;
1818 model_gelu->super.outputs = &model_gelu->output;
1819 model_gelu->super.output_size = 1;
1820 model_gelu->tanh = tanh;
1821 ccv_cnnp_model_copy_name(&model_gelu->super, name);
1822 return (ccv_cnnp_model_t*)model_gelu;
1823}
1824
1825static ccv_cnnp_model_t* _ccv_cnnp_gelu_copy(const ccv_cnnp_model_t* const super, void* const context)
1826{
1827 ccv_cnnp_model_gelu_t* const self = (ccv_cnnp_model_gelu_t*)super;
1828 return ccv_cnnp_gelu(self->tanh, self->super.name);
1829}
1830
1831// MARK - Leaky ReLU Layer
1832
1833typedef struct {
1834 ccv_cnnp_model_t super;
1835 ccv_nnc_tensor_symbol_t output;
1836 float negative_slope;
1837} ccv_cnnp_model_leaky_relu_t;
1838
1839static void _ccv_cnnp_leaky_relu_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
1840{
1841 PRINT(CCV_CLI_VERBOSE, "[cnnp_leaky_relu_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_leaky_relu_build] -\n"); fflush(stdout); } }
while (0)
;
1842 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 1842, __extension__ __PRETTY_FUNCTION__); }))
;
1843 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 1843, __extension__ __PRETTY_FUNCTION__
); }))
;
1844 ccv_cnnp_model_leaky_relu_t* const self = (ccv_cnnp_model_leaky_relu_t*)super;
1845 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
1846 ccv_nnc_tensor_param_t output_params;
1847 const ccv_nnc_cmd_t leaky_relu = CMD_LEAKY_RELU_FORWARD(self->negative_slope)ccv_nnc_cmd(CCV_NNC_LEAKY_RELU_FORWARD, 0, (ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.leaky_relu={.negative_slope=self->
negative_slope}}, 0)
;
1848 ccv_nnc_hint_tensor_auto(leaky_relu, (ccv_nnc_tensor_param_t []){
1849 params,
1850 }, 1, ccv_nnc_no_hint, &output_params, 1);
1851 const ccv_nnc_tensor_symbol_t leaky_relu_output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
1852 ccv_nnc_graph_exec_symbol_new(graph, leaky_relu, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(leaky_relu_output)(const ccv_nnc_tensor_symbol_t []){leaky_relu_output}, (1 +1 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
-1)
, "leaky_relu");
1853 outputs[0] = leaky_relu_output;
1854}
1855
1856static ccv_cnnp_model_t* _ccv_cnnp_leaky_relu_copy(const ccv_cnnp_model_t* const self, void* const context);
1857
1858static const ccv_cnnp_model_vtab_t ccv_cnnp_leaky_relu_isa = {
1859 .build = _ccv_cnnp_leaky_relu_build,
1860 .copy = _ccv_cnnp_leaky_relu_copy,
1861};
1862
1863ccv_cnnp_model_t* ccv_cnnp_leaky_relu(const float negative_slope, const char* const name)
1864{
1865 ccv_cnnp_model_leaky_relu_t* const model_leaky_relu = (ccv_cnnp_model_leaky_relu_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_leaky_relu_t));
1866 model_leaky_relu->super.isa = &ccv_cnnp_leaky_relu_isa;
1867 model_leaky_relu->super.input_size = 1;
1868 model_leaky_relu->super.outputs = &model_leaky_relu->output;
1869 model_leaky_relu->super.output_size = 1;
1870 model_leaky_relu->negative_slope = negative_slope;
1871 ccv_cnnp_model_copy_name(&model_leaky_relu->super, name);
1872 return (ccv_cnnp_model_t*)model_leaky_relu;
1873}
1874
1875static ccv_cnnp_model_t* _ccv_cnnp_leaky_relu_copy(const ccv_cnnp_model_t* const super, void* const context)
1876{
1877 ccv_cnnp_model_leaky_relu_t* const self = (ccv_cnnp_model_leaky_relu_t*)super;
1878 return ccv_cnnp_leaky_relu(self->negative_slope, self->super.name);
1879}
1880
1881// MARK - Softmax Layer
1882
1883typedef struct {
1884 ccv_cnnp_model_t super;
1885 ccv_nnc_tensor_symbol_t output;
1886} ccv_cnnp_model_softmax_t;
1887
1888static void _ccv_cnnp_softmax_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
1889{
1890 PRINT(CCV_CLI_VERBOSE, "[cnnp_softmax_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_softmax_build] -\n"); fflush(stdout); } } while
(0)
;
1891 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 1891, __extension__ __PRETTY_FUNCTION__); }))
;
1892 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 1892, __extension__ __PRETTY_FUNCTION__
); }))
;
1893 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
1894 ccv_nnc_tensor_param_t output_params;
1895 const ccv_nnc_cmd_t softmax = CMD_SOFTMAX_FORWARD()ccv_nnc_cmd(CCV_NNC_SOFTMAX_FORWARD, 0, ccv_nnc_cmd_auto, 0);
1896 ccv_nnc_hint_tensor_auto(softmax, (ccv_nnc_tensor_param_t []){
1897 params,
1898 }, 1, ccv_nnc_no_hint, &output_params, 1);
1899 const ccv_nnc_tensor_symbol_t softmax_output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
1900 ccv_nnc_graph_exec_symbol_new(graph, softmax, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(softmax_output)(const ccv_nnc_tensor_symbol_t []){softmax_output}, (1 +1 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1
)
, "softmax");
1901 outputs[0] = softmax_output;
1902}
1903
1904static ccv_cnnp_model_t* _ccv_cnnp_softmax_copy(const ccv_cnnp_model_t* const self, void* const context);
1905
1906static const ccv_cnnp_model_vtab_t ccv_cnnp_softmax_isa = {
1907 .build = _ccv_cnnp_softmax_build,
1908 .copy = _ccv_cnnp_softmax_copy,
1909};
1910
1911ccv_cnnp_model_t* ccv_cnnp_softmax(const char* const name)
1912{
1913 ccv_cnnp_model_softmax_t* const model_softmax = (ccv_cnnp_model_softmax_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_softmax_t));
1914 model_softmax->super.isa = &ccv_cnnp_softmax_isa;
1915 model_softmax->super.input_size = 1;
1916 model_softmax->super.outputs = &model_softmax->output;
1917 model_softmax->super.output_size = 1;
1918 ccv_cnnp_model_copy_name(&model_softmax->super, name);
1919 return (ccv_cnnp_model_t*)model_softmax;
1920}
1921
1922static ccv_cnnp_model_t* _ccv_cnnp_softmax_copy(const ccv_cnnp_model_t* const self, void* const context)
1923{
1924 return ccv_cnnp_softmax(self->name);
1925}
1926
1927// MARK - Add Layer
1928
1929typedef struct {
1930 ccv_cnnp_model_t super;
1931 float p;
1932 float q;
1933 ccv_nnc_tensor_symbol_t output;
1934} ccv_cnnp_model_add_t;
1935
1936static void _ccv_cnnp_add_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
1937{
1938 PRINT(CCV_CLI_VERBOSE, "[cnnp_add_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_add_build] -\n"); fflush(stdout); } } while (
0)
;
1939 const ccv_cnnp_model_add_t* const self = (const ccv_cnnp_model_add_t*)super;
1940 assert(input_size == 2)((void) sizeof ((input_size == 2) ? 1 : 0), __extension__ ({ if
(input_size == 2) ; else __assert_fail ("input_size == 2", "ccv_cnnp_model_addons.c"
, 1940, __extension__ __PRETTY_FUNCTION__); }))
;
1941 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 1941, __extension__ __PRETTY_FUNCTION__
); }))
;
1942 ccv_nnc_tensor_param_t input_params[2];
1943 int i;
1944 for (i = 0; i < 2; i++)
1945 input_params[i] = ccv_nnc_tensor_symbol_params(graph, inputs[i]);
1946 ccv_nnc_tensor_param_t output_params;
1947 const ccv_nnc_cmd_t add = CMD_ADD_FORWARD(self->p, self->q)ccv_nnc_cmd(CCV_NNC_ADD_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={self->p, self->q}}}, 0)
;
1948 ccv_nnc_hint_tensor_auto(add, input_params, 2, ccv_nnc_no_hint, &output_params, 1);
1949 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
1950 ccv_nnc_graph_exec_symbol_new(graph, add, inputs, input_size, outputs, output_size, "add");
1951}
1952
1953static ccv_cnnp_model_t* _ccv_cnnp_add_copy(const ccv_cnnp_model_t* const self, void* const context);
1954
1955static const ccv_cnnp_model_vtab_t ccv_cnnp_add_isa = {
1956 .build = _ccv_cnnp_add_build,
1957 .copy = _ccv_cnnp_add_copy,
1958};
1959
1960ccv_cnnp_model_t* ccv_cnnp_add(const float p, const float q, const char* const name)
1961{
1962 ccv_cnnp_model_add_t* const model_add = (ccv_cnnp_model_add_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_add_t));
1963 model_add->super.isa = &ccv_cnnp_add_isa;
1964 model_add->super.input_size = 2;
1965 model_add->super.outputs = &model_add->output;
1966 model_add->super.output_size = 1;
1967 model_add->p = p;
1968 model_add->q = q;
1969 ccv_cnnp_model_copy_name(&model_add->super, name);
1970 return (ccv_cnnp_model_t*)model_add;
1971}
1972
1973static ccv_cnnp_model_t* _ccv_cnnp_add_copy(const ccv_cnnp_model_t* const super, void* const context)
1974{
1975 const ccv_cnnp_model_add_t* const self = (const ccv_cnnp_model_add_t*)super;
1976 return ccv_cnnp_add(self->p, self->q, self->super.name);
1977}
1978
1979// MARK - Mul Layer
1980
1981typedef struct {
1982 ccv_cnnp_model_t super;
1983 ccv_nnc_tensor_symbol_t output;
1984 float p;
1985} ccv_cnnp_model_mul_t;
1986
1987static void _ccv_cnnp_mul_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
1988{
1989 PRINT(CCV_CLI_VERBOSE, "[cnnp_mul_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_mul_build] -\n"); fflush(stdout); } } while (
0)
;
1990 const ccv_cnnp_model_mul_t* const self = (const ccv_cnnp_model_mul_t*)super;
1991 assert(input_size == 2)((void) sizeof ((input_size == 2) ? 1 : 0), __extension__ ({ if
(input_size == 2) ; else __assert_fail ("input_size == 2", "ccv_cnnp_model_addons.c"
, 1991, __extension__ __PRETTY_FUNCTION__); }))
;
1992 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 1992, __extension__ __PRETTY_FUNCTION__
); }))
;
1993 ccv_nnc_tensor_param_t input_params[2];
1994 int i;
1995 for (i = 0; i < 2; i++)
1996 input_params[i] = ccv_nnc_tensor_symbol_params(graph, inputs[i]);
1997 ccv_nnc_tensor_param_t output_params;
1998 const ccv_nnc_cmd_t mul = CMD_MUL_FORWARD(self->p)ccv_nnc_cmd(CCV_NNC_MUL_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={self->p,}}}, 0)
;
1999 ccv_nnc_hint_tensor_auto(mul, input_params, 2, ccv_nnc_no_hint, &output_params, 1);
2000 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
2001 ccv_nnc_graph_exec_symbol_new(graph, mul, inputs, input_size, outputs, output_size, "mul");
2002}
2003
2004static ccv_cnnp_model_t* _ccv_cnnp_mul_copy(const ccv_cnnp_model_t* const self, void* const context);
2005
2006static const ccv_cnnp_model_vtab_t ccv_cnnp_mul_isa = {
2007 .build = _ccv_cnnp_mul_build,
2008 .copy = _ccv_cnnp_mul_copy,
2009};
2010
2011ccv_cnnp_model_t* ccv_cnnp_mul(const float p, const char* const name)
2012{
2013 ccv_cnnp_model_mul_t* const model_mul = (ccv_cnnp_model_mul_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_mul_t));
2014 model_mul->super.isa = &ccv_cnnp_mul_isa;
2015 model_mul->super.input_size = 2;
2016 model_mul->super.outputs = &model_mul->output;
2017 model_mul->super.output_size = 1;
2018 model_mul->p = p;
2019 ccv_cnnp_model_copy_name(&model_mul->super, name);
2020 return (ccv_cnnp_model_t*)model_mul;
2021}
2022
2023static ccv_cnnp_model_t* _ccv_cnnp_mul_copy(const ccv_cnnp_model_t* const super, void* const context)
2024{
2025 const ccv_cnnp_model_mul_t* const self = (const ccv_cnnp_model_mul_t*)super;
2026 return ccv_cnnp_mul(self->p, self->super.name);
2027}
2028
2029// MARK - Scalar Mul Layer
2030
2031typedef struct {
2032 ccv_cnnp_model_t super;
2033 ccv_nnc_tensor_symbol_t output;
2034 float a;
2035} ccv_cnnp_model_scalar_mul_t;
2036
2037static void _ccv_cnnp_scalar_mul_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
2038{
2039 PRINT(CCV_CLI_VERBOSE, "[cnnp_scalar_mul_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_scalar_mul_build] -\n"); fflush(stdout); } }
while (0)
;
2040 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 2040, __extension__ __PRETTY_FUNCTION__); }))
;
2041 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 2041, __extension__ __PRETTY_FUNCTION__
); }))
;
2042 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
2043 ccv_nnc_tensor_param_t output_params;
2044 ccv_cnnp_model_scalar_mul_t* const self = (ccv_cnnp_model_scalar_mul_t*)super;
2045 const ccv_nnc_cmd_t scalar_mul = CMD_SCALAR_MUL_FORWARD(self->a)ccv_nnc_cmd(CCV_NNC_SCALAR_MUL_FORWARD, 0, (ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.blas={.a={self->a,}}}, 0)
;
2046 ccv_nnc_hint_tensor_auto(scalar_mul, (ccv_nnc_tensor_param_t []){
2047 params,
2048 }, 1, ccv_nnc_no_hint, &output_params, 1);
2049 const ccv_nnc_tensor_symbol_t scalar_mul_output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
2050 ccv_nnc_graph_exec_symbol_new(graph, scalar_mul, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(scalar_mul_output)(const ccv_nnc_tensor_symbol_t []){scalar_mul_output}, (1 +1 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
-1)
, "scalar_mul");
2051 outputs[0] = scalar_mul_output;
2052}
2053
2054static ccv_cnnp_model_t* _ccv_cnnp_scalar_mul_copy(const ccv_cnnp_model_t* const super, void* const context);
2055
2056static const ccv_cnnp_model_vtab_t ccv_cnnp_scalar_mul_isa = {
2057 .build = _ccv_cnnp_scalar_mul_build,
2058 .copy = _ccv_cnnp_scalar_mul_copy,
2059};
2060
2061ccv_cnnp_model_t* ccv_cnnp_scalar_mul(const float a, const char* const name)
2062{
2063 ccv_cnnp_model_scalar_mul_t* const model_scalar_mul = (ccv_cnnp_model_scalar_mul_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_scalar_mul_t));
2064 model_scalar_mul->super.isa = &ccv_cnnp_scalar_mul_isa;
2065 model_scalar_mul->super.input_size = 1;
2066 model_scalar_mul->super.outputs = &model_scalar_mul->output;
2067 model_scalar_mul->super.output_size = 1;
2068 model_scalar_mul->a = a;
2069 ccv_cnnp_model_copy_name(&model_scalar_mul->super, name);
2070 return (ccv_cnnp_model_t*)model_scalar_mul;
2071}
2072
2073static ccv_cnnp_model_t* _ccv_cnnp_scalar_mul_copy(const ccv_cnnp_model_t* const super, void* const context)
2074{
2075 const ccv_cnnp_model_scalar_mul_t* const self = (const ccv_cnnp_model_scalar_mul_t*)super;
2076 return ccv_cnnp_scalar_mul(self->a, self->super.name);
2077}
2078
2079// MARK - Div Layer
2080
2081typedef struct {
2082 ccv_cnnp_model_t super;
2083 ccv_nnc_tensor_symbol_t output;
2084 int reciprocal;
2085} ccv_cnnp_model_div_t;
2086
2087static void _ccv_cnnp_div_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
2088{
2089 const ccv_cnnp_model_div_t* const self = (const ccv_cnnp_model_div_t*)super;
2090 PRINT(CCV_CLI_VERBOSE, "[cnnp_div_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_div_build] -\n"); fflush(stdout); } } while (
0)
;
2091 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 2091, __extension__ __PRETTY_FUNCTION__
); }))
;
2092 ccv_nnc_tensor_param_t input_params[2];
2093 int i;
2094 ccv_nnc_tensor_param_t output_params;
2095 const ccv_nnc_cmd_t div = CMD_EWDIV_FORWARD()ccv_nnc_cmd(CCV_NNC_EWDIV_FORWARD, 0, ccv_nnc_cmd_auto, 0);
2096 if (self->reciprocal)
2097 {
2098 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 2098, __extension__ __PRETTY_FUNCTION__); }))
;
2099 input_params[0] = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
2100 input_params[1] = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
2101 ccv_nnc_hint_tensor_auto(div, input_params, 2, ccv_nnc_no_hint, &output_params, 1);
2102 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
2103 ccv_nnc_graph_exec_symbol_new(graph, div, TENSOR_SYMBOL_LIST(NO_TENSOR_SYMBOL, inputs[0])(const ccv_nnc_tensor_symbol_t []){(const ccv_nnc_tensor_symbol_t
){.d = CCV_NNC_NO_TENSOR_SYMBOL}, inputs[0]}, (1 +1 +1 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, outputs, output_size, "div");
2104 } else {
2105 assert(input_size == 2)((void) sizeof ((input_size == 2) ? 1 : 0), __extension__ ({ if
(input_size == 2) ; else __assert_fail ("input_size == 2", "ccv_cnnp_model_addons.c"
, 2105, __extension__ __PRETTY_FUNCTION__); }))
;
2106 for (i = 0; i < 2; i++)
2107 input_params[i] = ccv_nnc_tensor_symbol_params(graph, inputs[i]);
2108 ccv_nnc_hint_tensor_auto(div, input_params, input_size, ccv_nnc_no_hint, &output_params, 1);
2109 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
2110 ccv_nnc_graph_exec_symbol_new(graph, div, inputs, input_size, outputs, output_size, "div");
2111 }
2112}
2113
2114static ccv_cnnp_model_t* _ccv_cnnp_div_copy(const ccv_cnnp_model_t* const self, void* const context);
2115
2116static const ccv_cnnp_model_vtab_t ccv_cnnp_div_isa = {
2117 .build = _ccv_cnnp_div_build,
2118 .copy = _ccv_cnnp_div_copy,
2119};
2120
2121ccv_cnnp_model_t* ccv_cnnp_div(const int reciprocal, const char* const name)
2122{
2123 ccv_cnnp_model_div_t* const model_div = (ccv_cnnp_model_div_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_div_t));
2124 model_div->super.isa = &ccv_cnnp_div_isa;
2125 model_div->super.input_size = reciprocal ? 1 : 2;
2126 model_div->super.outputs = &model_div->output;
2127 model_div->super.output_size = 1;
2128 model_div->reciprocal = reciprocal;
2129 ccv_cnnp_model_copy_name(&model_div->super, name);
2130 return (ccv_cnnp_model_t*)model_div;
2131}
2132
2133static ccv_cnnp_model_t* _ccv_cnnp_div_copy(const ccv_cnnp_model_t* const super, void* const context)
2134{
2135 const ccv_cnnp_model_div_t* const self = (const ccv_cnnp_model_div_t*)super;
2136 return ccv_cnnp_div(self->reciprocal, self->super.name);
2137}
2138
2139// MARK - Sqrt Layer
2140
2141typedef struct {
2142 ccv_cnnp_model_t super;
2143 ccv_nnc_tensor_symbol_t output;
2144} ccv_cnnp_model_sqrt_t;
2145
2146static void _ccv_cnnp_sqrt_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
2147{
2148 PRINT(CCV_CLI_VERBOSE, "[cnnp_sqrt_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_sqrt_build] -\n"); fflush(stdout); } } while
(0)
;
2149 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 2149, __extension__ __PRETTY_FUNCTION__
); }))
;
2150 ccv_nnc_tensor_param_t input_params[1];
2151 ccv_nnc_tensor_param_t output_params;
2152 const ccv_nnc_cmd_t sqrt = CMD_EWSQRT_FORWARD()ccv_nnc_cmd(CCV_NNC_EWSQRT_FORWARD, 0, ccv_nnc_cmd_auto, 0);
2153 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 2153, __extension__ __PRETTY_FUNCTION__); }))
;
2154 input_params[0] = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
2155 ccv_nnc_hint_tensor_auto(sqrt, input_params, 1, ccv_nnc_no_hint, &output_params, 1);
2156 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
2157 ccv_nnc_graph_exec_symbol_new(graph, sqrt, inputs, 1, outputs, output_size, "sqrt");
2158}
2159
2160static ccv_cnnp_model_t* _ccv_cnnp_sqrt_copy(const ccv_cnnp_model_t* const self, void* const context);
2161
2162static const ccv_cnnp_model_vtab_t ccv_cnnp_sqrt_isa = {
2163 .build = _ccv_cnnp_sqrt_build,
2164 .copy = _ccv_cnnp_sqrt_copy,
2165};
2166
2167ccv_cnnp_model_t* ccv_cnnp_sqrt(const char* const name)
2168{
2169 ccv_cnnp_model_sqrt_t* const model_sqrt = (ccv_cnnp_model_sqrt_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_sqrt_t));
2170 model_sqrt->super.isa = &ccv_cnnp_sqrt_isa;
2171 model_sqrt->super.input_size = 1;
2172 model_sqrt->super.outputs = &model_sqrt->output;
2173 model_sqrt->super.output_size = 1;
2174 ccv_cnnp_model_copy_name(&model_sqrt->super, name);
2175 return (ccv_cnnp_model_t*)model_sqrt;
2176}
2177
2178static ccv_cnnp_model_t* _ccv_cnnp_sqrt_copy(const ccv_cnnp_model_t* const super, void* const context)
2179{
2180 const ccv_cnnp_model_sqrt_t* const self = (const ccv_cnnp_model_sqrt_t*)super;
2181 return ccv_cnnp_sqrt(self->super.name);
2182}
2183
2184// MARK - Log Layer
2185
2186typedef struct {
2187 ccv_cnnp_model_t super;
2188 ccv_nnc_tensor_symbol_t output;
2189} ccv_cnnp_model_log_t;
2190
2191static void _ccv_cnnp_log_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
2192{
2193 PRINT(CCV_CLI_VERBOSE, "[cnnp_log_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_log_build] -\n"); fflush(stdout); } } while (
0)
;
2194 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 2194, __extension__ __PRETTY_FUNCTION__
); }))
;
2195 ccv_nnc_tensor_param_t input_params[1];
2196 ccv_nnc_tensor_param_t output_params;
2197 const ccv_nnc_cmd_t log = CMD_EWLOG_FORWARD()ccv_nnc_cmd(CCV_NNC_EWLOG_FORWARD, 0, ccv_nnc_cmd_auto, 0);
2198 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 2198, __extension__ __PRETTY_FUNCTION__); }))
;
2199 input_params[0] = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
2200 ccv_nnc_hint_tensor_auto(log, input_params, 1, ccv_nnc_no_hint, &output_params, 1);
2201 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
2202 ccv_nnc_graph_exec_symbol_new(graph, log, inputs, 1, outputs, output_size, "log");
2203}
2204
2205static ccv_cnnp_model_t* _ccv_cnnp_log_copy(const ccv_cnnp_model_t* const self, void* const context);
2206
2207static const ccv_cnnp_model_vtab_t ccv_cnnp_log_isa = {
2208 .build = _ccv_cnnp_log_build,
2209 .copy = _ccv_cnnp_log_copy,
2210};
2211
2212ccv_cnnp_model_t* ccv_cnnp_log(const char* const name)
2213{
2214 ccv_cnnp_model_log_t* const model_log = (ccv_cnnp_model_log_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_log_t));
2215 model_log->super.isa = &ccv_cnnp_log_isa;
2216 model_log->super.input_size = 1;
2217 model_log->super.outputs = &model_log->output;
2218 model_log->super.output_size = 1;
2219 ccv_cnnp_model_copy_name(&model_log->super, name);
2220 return (ccv_cnnp_model_t*)model_log;
2221}
2222
2223static ccv_cnnp_model_t* _ccv_cnnp_log_copy(const ccv_cnnp_model_t* const super, void* const context)
2224{
2225 return ccv_cnnp_log(super->name);
2226}
2227
2228// MARK - Pow Layer
2229
2230typedef struct {
2231 ccv_cnnp_model_t super;
2232 ccv_nnc_tensor_symbol_t output;
2233 ccv_nnc_cmd_param_t params;
2234} ccv_cnnp_model_pow_t;
2235
2236static void _ccv_cnnp_pow_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
2237{
2238 ccv_cnnp_model_pow_t* const self = (ccv_cnnp_model_pow_t*)super;
2239 PRINT(CCV_CLI_VERBOSE, "[cnnp_pow_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_pow_build] -\n"); fflush(stdout); } } while (
0)
;
2240 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 2240, __extension__ __PRETTY_FUNCTION__); }))
;
2241 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 2241, __extension__ __PRETTY_FUNCTION__
); }))
;
2242 ccv_nnc_tensor_param_t input_params[1];
2243 input_params[0] = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
2244 ccv_nnc_tensor_param_t output_params;
2245 const ccv_nnc_cmd_t pow = ccv_nnc_cmd(CCV_NNC_EWPOW_FORWARD, 0, self->params, 0);
2246 ccv_nnc_hint_tensor_auto(pow, input_params, 1, ccv_nnc_no_hint, &output_params, 1);
2247 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
2248 ccv_nnc_graph_exec_symbol_new(graph, pow, inputs, input_size, outputs, output_size, "pow");
2249}
2250
2251static ccv_cnnp_model_t* _ccv_cnnp_pow_copy(const ccv_cnnp_model_t* const self, void* const context);
2252
2253static const ccv_cnnp_model_vtab_t ccv_cnnp_pow_isa = {
2254 .build = _ccv_cnnp_pow_build,
2255 .copy = _ccv_cnnp_pow_copy,
2256};
2257
2258ccv_cnnp_model_t* ccv_cnnp_pow(const float exponent, const char* const name)
2259{
2260 ccv_cnnp_model_pow_t* const model_pow = (ccv_cnnp_model_pow_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_pow_t));
2261 model_pow->super.isa = &ccv_cnnp_pow_isa;
2262 model_pow->super.input_size = 1;
2263 model_pow->super.outputs = &model_pow->output;
2264 model_pow->super.output_size = 1;
2265 model_pow->params = (ccv_nnc_cmd_param_t){
2266 .size = {
2267 .dim = { 1, 1, 1 }
2268 },
2269 .pow = {
2270 .exponent = exponent,
2271 },
2272 };
2273 ccv_cnnp_model_copy_name(&model_pow->super, name);
2274 return (ccv_cnnp_model_t*)model_pow;
2275}
2276
2277static ccv_cnnp_model_t* _ccv_cnnp_pow_copy(const ccv_cnnp_model_t* const super, void* const context)
2278{
2279 const ccv_cnnp_model_pow_t* const self = (const ccv_cnnp_model_pow_t*)super;
2280 return ccv_cnnp_pow(self->params.pow.exponent, super->name);
2281}
2282
2283// MARK - Sin Layer
2284
2285typedef struct {
2286 ccv_cnnp_model_t super;
2287 ccv_nnc_tensor_symbol_t output;
2288} ccv_cnnp_model_sin_t;
2289
2290static void _ccv_cnnp_sin_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
2291{
2292 PRINT(CCV_CLI_VERBOSE, "[cnnp_sin_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_sin_build] -\n"); fflush(stdout); } } while (
0)
;
2293 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 2293, __extension__ __PRETTY_FUNCTION__
); }))
;
2294 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 2294, __extension__ __PRETTY_FUNCTION__); }))
;
2295 ccv_nnc_tensor_param_t input_params[1];
2296 ccv_nnc_tensor_param_t output_params;
2297 const ccv_nnc_cmd_t sin = CMD_EWSIN_FORWARD()ccv_nnc_cmd(CCV_NNC_EWSIN_FORWARD, 0, ccv_nnc_cmd_auto, 0);
2298 input_params[0] = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
2299 ccv_nnc_hint_tensor_auto(sin, input_params, 1, ccv_nnc_no_hint, &output_params, 1);
2300 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
2301 ccv_nnc_graph_exec_symbol_new(graph, sin, inputs, 1, outputs, output_size, "sin");
2302}
2303
2304static ccv_cnnp_model_t* _ccv_cnnp_sin_copy(const ccv_cnnp_model_t* const self, void* const context);
2305
2306static const ccv_cnnp_model_vtab_t ccv_cnnp_sin_isa = {
2307 .build = _ccv_cnnp_sin_build,
2308 .copy = _ccv_cnnp_sin_copy,
2309};
2310
2311ccv_cnnp_model_t* ccv_cnnp_sin(const char* const name)
2312{
2313 ccv_cnnp_model_sin_t* const model_sin = (ccv_cnnp_model_sin_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_sin_t));
2314 model_sin->super.isa = &ccv_cnnp_sin_isa;
2315 model_sin->super.input_size = 1;
2316 model_sin->super.outputs = &model_sin->output;
2317 model_sin->super.output_size = 1;
2318 ccv_cnnp_model_copy_name(&model_sin->super, name);
2319 return (ccv_cnnp_model_t*)model_sin;
2320}
2321
2322static ccv_cnnp_model_t* _ccv_cnnp_sin_copy(const ccv_cnnp_model_t* const super, void* const context)
2323{
2324 return ccv_cnnp_sin(super->name);
2325}
2326
2327// MARK - Cos Layer
2328
2329typedef struct {
2330 ccv_cnnp_model_t super;
2331 ccv_nnc_tensor_symbol_t output;
2332} ccv_cnnp_model_cos_t;
2333
2334static void _ccv_cnnp_cos_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
2335{
2336 PRINT(CCV_CLI_VERBOSE, "[cnnp_cos_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_cos_build] -\n"); fflush(stdout); } } while (
0)
;
2337 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 2337, __extension__ __PRETTY_FUNCTION__
); }))
;
2338 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 2338, __extension__ __PRETTY_FUNCTION__); }))
;
2339 ccv_nnc_tensor_param_t input_params[1];
2340 ccv_nnc_tensor_param_t output_params;
2341 const ccv_nnc_cmd_t cos = CMD_EWCOS_FORWARD()ccv_nnc_cmd(CCV_NNC_EWCOS_FORWARD, 0, ccv_nnc_cmd_auto, 0);
2342 input_params[0] = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
2343 ccv_nnc_hint_tensor_auto(cos, input_params, 1, ccv_nnc_no_hint, &output_params, 1);
2344 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
2345 ccv_nnc_graph_exec_symbol_new(graph, cos, inputs, 1, outputs, output_size, "cos");
2346}
2347
2348static ccv_cnnp_model_t* _ccv_cnnp_cos_copy(const ccv_cnnp_model_t* const self, void* const context);
2349
2350static const ccv_cnnp_model_vtab_t ccv_cnnp_cos_isa = {
2351 .build = _ccv_cnnp_cos_build,
2352 .copy = _ccv_cnnp_cos_copy,
2353};
2354
2355ccv_cnnp_model_t* ccv_cnnp_cos(const char* const name)
2356{
2357 ccv_cnnp_model_cos_t* const model_cos = (ccv_cnnp_model_cos_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_cos_t));
2358 model_cos->super.isa = &ccv_cnnp_cos_isa;
2359 model_cos->super.input_size = 1;
2360 model_cos->super.outputs = &model_cos->output;
2361 model_cos->super.output_size = 1;
2362 ccv_cnnp_model_copy_name(&model_cos->super, name);
2363 return (ccv_cnnp_model_t*)model_cos;
2364}
2365
2366static ccv_cnnp_model_t* _ccv_cnnp_cos_copy(const ccv_cnnp_model_t* const super, void* const context)
2367{
2368 return ccv_cnnp_cos(super->name);
2369}
2370
2371// MARK - Rotate Half Layer
2372
2373typedef struct {
2374 ccv_cnnp_model_t super;
2375 ccv_nnc_tensor_symbol_t output;
2376} ccv_cnnp_model_rotate_half_t;
2377
2378static void _ccv_cnnp_rotate_half_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
2379{
2380 PRINT(CCV_CLI_VERBOSE, "[cnnp_rotate_half_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_rotate_half_build] -\n"); fflush(stdout); } }
while (0)
;
2381 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 2381, __extension__ __PRETTY_FUNCTION__); }))
;
2382 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 2382, __extension__ __PRETTY_FUNCTION__
); }))
;
2383 ccv_nnc_tensor_param_t input_params[1];
2384 ccv_nnc_tensor_param_t output_params;
2385 const ccv_nnc_cmd_t rotate_half = CMD_ROTATE_HALF_FORWARD()ccv_nnc_cmd(CCV_NNC_ROTATE_HALF_FORWARD, 0, (ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}}}, 0)
;
2386 input_params[0] = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
2387 ccv_nnc_hint_tensor_auto(rotate_half, input_params, 1, ccv_nnc_no_hint, &output_params, 1);
2388 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
2389 ccv_nnc_graph_exec_symbol_new(graph, rotate_half, inputs, 1, outputs, output_size, "rotate_half");
2390}
2391
2392static ccv_cnnp_model_t* _ccv_cnnp_rotate_half_copy(const ccv_cnnp_model_t* const self, void* const context);
2393
2394static const ccv_cnnp_model_vtab_t ccv_cnnp_rotate_half_isa = {
2395 .build = _ccv_cnnp_rotate_half_build,
2396 .copy = _ccv_cnnp_rotate_half_copy,
2397};
2398
2399ccv_cnnp_model_t* ccv_cnnp_rotate_half(const char* const name)
2400{
2401 ccv_cnnp_model_rotate_half_t* const model_rotate_half = (ccv_cnnp_model_rotate_half_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_rotate_half_t));
2402 model_rotate_half->super.isa = &ccv_cnnp_rotate_half_isa;
2403 model_rotate_half->super.input_size = 1;
2404 model_rotate_half->super.outputs = &model_rotate_half->output;
2405 model_rotate_half->super.output_size = 1;
2406 ccv_cnnp_model_copy_name(&model_rotate_half->super, name);
2407 return (ccv_cnnp_model_t*)model_rotate_half;
2408}
2409
2410static ccv_cnnp_model_t* _ccv_cnnp_rotate_half_copy(const ccv_cnnp_model_t* const super, void* const context)
2411{
2412 return ccv_cnnp_rotate_half(super->name);
2413}
2414
2415// MARK - Cmul Layer
2416
2417typedef struct {
2418 ccv_cnnp_model_t super;
2419 ccv_nnc_tensor_symbol_t output;
2420} ccv_cnnp_model_cmul_t;
2421
2422static void _ccv_cnnp_cmul_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
2423{
2424 PRINT(CCV_CLI_VERBOSE, "[cnnp_cmul_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_cmul_build] -\n"); fflush(stdout); } } while
(0)
;
2425 assert(input_size == 2)((void) sizeof ((input_size == 2) ? 1 : 0), __extension__ ({ if
(input_size == 2) ; else __assert_fail ("input_size == 2", "ccv_cnnp_model_addons.c"
, 2425, __extension__ __PRETTY_FUNCTION__); }))
;
2426 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 2426, __extension__ __PRETTY_FUNCTION__
); }))
;
2427 ccv_nnc_tensor_param_t input_params[2];
2428 int i;
2429 for (i = 0; i < 2; i++)
2430 input_params[i] = ccv_nnc_tensor_symbol_params(graph, inputs[i]);
2431 ccv_nnc_tensor_param_t output_params;
2432 const ccv_nnc_cmd_t mul = CMD_CMUL_FORWARD()ccv_nnc_cmd(CCV_NNC_CMUL_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}}}, 0)
;
2433 ccv_nnc_hint_tensor_auto(mul, input_params, 2, ccv_nnc_no_hint, &output_params, 1);
2434 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
2435 ccv_nnc_graph_exec_symbol_new(graph, mul, inputs, input_size, outputs, output_size, "cmul");
2436}
2437
2438static ccv_cnnp_model_t* _ccv_cnnp_cmul_copy(const ccv_cnnp_model_t* const self, void* const context);
2439
2440static const ccv_cnnp_model_vtab_t ccv_cnnp_cmul_isa = {
2441 .build = _ccv_cnnp_cmul_build,
2442 .copy = _ccv_cnnp_cmul_copy,
2443};
2444
2445ccv_cnnp_model_t* ccv_cnnp_cmul(const char* const name)
2446{
2447 ccv_cnnp_model_cmul_t* const model_cmul = (ccv_cnnp_model_cmul_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_cmul_t));
2448 model_cmul->super.isa = &ccv_cnnp_cmul_isa;
2449 model_cmul->super.input_size = 2;
2450 model_cmul->super.outputs = &model_cmul->output;
2451 model_cmul->super.output_size = 1;
2452 ccv_cnnp_model_copy_name(&model_cmul->super, name);
2453 return (ccv_cnnp_model_t*)model_cmul;
2454}
2455
2456static ccv_cnnp_model_t* _ccv_cnnp_cmul_copy(const ccv_cnnp_model_t* const super, void* const context)
2457{
2458 return ccv_cnnp_cmul(super->name);
2459}
2460
2461// MARK - Transpose Layer
2462
2463typedef struct {
2464 ccv_cnnp_model_t super;
2465 ccv_nnc_tensor_symbol_t output;
2466 int transpose[2];
2467} ccv_cnnp_model_transpose_t;
2468
2469static void _ccv_cnnp_transpose_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
2470{
2471 ccv_cnnp_model_transpose_t* const self = (ccv_cnnp_model_transpose_t*)super;
2472 PRINT(CCV_CLI_VERBOSE, "[cnnp_transpose_build] (%d, %d)\n", self->transpose[0], self->transpose[1])do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_transpose_build] (%d, %d)\n", self->transpose
[0], self->transpose[1]); fflush(stdout); } } while (0)
;
2473 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 2473, __extension__ __PRETTY_FUNCTION__); }))
;
2474 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 2474, __extension__ __PRETTY_FUNCTION__
); }))
;
2475 if (self->transpose[0] == self->transpose[1])
2476 {
2477 outputs[0] = inputs[0];
2478 return;
2479 }
2480 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
2481 ccv_nnc_tensor_param_t output_params;
2482 const ccv_nnc_cmd_t transpose = CMD_TRANSPOSE_FORWARD(self->transpose[0], self->transpose[1])ccv_nnc_cmd(CCV_NNC_TRANSPOSE_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.transpose={.axis={self->transpose[
0], self->transpose[1]}}}), 0)
;
2483 ccv_nnc_hint_tensor_auto(transpose, (ccv_nnc_tensor_param_t []){
2484 params,
2485 }, 1, ccv_nnc_no_hint, &output_params, 1);
2486 const ccv_nnc_tensor_symbol_t transpose_output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
2487 ccv_nnc_graph_exec_symbol_new(graph, transpose, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(transpose_output)(const ccv_nnc_tensor_symbol_t []){transpose_output}, (1 +1 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
-1)
, "transpose");
2488 outputs[0] = transpose_output;
2489}
2490
2491static ccv_cnnp_model_t* _ccv_cnnp_transpose_copy(const ccv_cnnp_model_t* const super, void* const context);
2492
2493static const ccv_cnnp_model_vtab_t ccv_cnnp_transpose_isa = {
2494 .build = _ccv_cnnp_transpose_build,
2495 .copy = _ccv_cnnp_transpose_copy,
2496};
2497
2498ccv_cnnp_model_t* ccv_cnnp_transpose(const int axis_a, const int axis_b, const char* const name)
2499{
2500 ccv_cnnp_model_transpose_t* const model_transpose = (ccv_cnnp_model_transpose_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_transpose_t));
2501 model_transpose->super.isa = &ccv_cnnp_transpose_isa;
2502 model_transpose->super.input_size = 1;
2503 model_transpose->super.outputs = &model_transpose->output;
2504 model_transpose->super.output_size = 1;
2505 model_transpose->transpose[0] = axis_a;
2506 model_transpose->transpose[1] = axis_b;
2507 ccv_cnnp_model_copy_name(&model_transpose->super, name);
2508 return (ccv_cnnp_model_t*)model_transpose;
2509}
2510
2511static ccv_cnnp_model_t* _ccv_cnnp_transpose_copy(const ccv_cnnp_model_t* const super, void* const context)
2512{
2513 const ccv_cnnp_model_transpose_t* const self = (const ccv_cnnp_model_transpose_t*)super;
2514 return ccv_cnnp_transpose(self->transpose[0], self->transpose[1], self->super.name);
2515}
2516
2517// MARK - Layer Norm Layer
2518
2519typedef struct {
2520 ccv_cnnp_model_t super;
2521 ccv_nnc_tensor_symbol_t output;
2522 ccv_nnc_tensor_symbol_t bias;
2523 ccv_nnc_tensor_symbol_t scale;
2524 ccv_nnc_cmd_param_t params;
2525} ccv_cnnp_model_layer_norm_t;
2526
2527static void _ccv_cnnp_layer_norm_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
2528{
2529 PRINT(CCV_CLI_VERBOSE, "[cnnp_layer_norm_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_layer_norm_build] -\n"); fflush(stdout); } }
while (0)
;
2530 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 2530, __extension__ __PRETTY_FUNCTION__); }))
;
2531 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 2531, __extension__ __PRETTY_FUNCTION__
); }))
;
2532 ccv_cnnp_model_layer_norm_t* const self = (ccv_cnnp_model_layer_norm_t*)super;
2533 const ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
2534 ccv_nnc_tensor_param_t bias_params = params;
2535 const int nd = ccv_nnc_tensor_nd(params.dim);
2536 int i;
2537 for (i = 0; i < nd; i++)
2538 bias_params.dim[i] = 1;
2539 for (i = 0; i < self->params.lnorm.count; i++)
2540 bias_params.dim[self->params.lnorm.axis[i]] = params.dim[self->params.lnorm.axis[i]];
2541 if (self->params.lnorm.elementwise_affine)
2542 {
2543 // Both scale and bias are shared between if this model is reused.
2544 if (!self->scale.graph)
2545 self->scale = ccv_nnc_tensor_symbol_new(graph, bias_params, "scale");
2546 if (!self->bias.graph)
2547 self->bias = ccv_nnc_tensor_symbol_new(graph, bias_params, "bias");
2548 }
2549 const ccv_nnc_cmd_t layer_norm = ccv_nnc_cmd(CCV_NNC_LAYER_NORM_FORWARD, 0, self->params, 0);
2550 ccv_nnc_tensor_param_t output_params[3];
2551 if (self->params.lnorm.elementwise_affine)
2552 ccv_nnc_hint_tensor_auto(layer_norm, (ccv_nnc_tensor_param_t []){
2553 params,
2554 bias_params,
2555 bias_params,
2556 }, 3, ccv_nnc_no_hint, output_params, 3);
2557 else
2558 ccv_nnc_hint_tensor_auto(layer_norm, (ccv_nnc_tensor_param_t []){
2559 params,
2560 }, 1, ccv_nnc_no_hint, output_params, 3);
2561 const ccv_nnc_tensor_symbol_t output = ccv_nnc_tensor_symbol_new(graph, output_params[0], 0);
2562 const ccv_nnc_tensor_symbol_t saved_mean = ccv_nnc_tensor_symbol_new(graph, output_params[1], "saved_mean");
2563 const ccv_nnc_tensor_symbol_t saved_inv_std = ccv_nnc_tensor_symbol_new(graph, output_params[2], "saved_inv_std");
2564 if (self->params.lnorm.elementwise_affine)
2565 ccv_nnc_graph_exec_symbol_new(graph, layer_norm, TENSOR_SYMBOL_LIST(inputs[0], self->scale, self->bias)(const ccv_nnc_tensor_symbol_t []){inputs[0], self->scale,
self->bias}, (1 +1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output, saved_mean, saved_inv_std)(const ccv_nnc_tensor_symbol_t []){output, saved_mean, saved_inv_std
}, (1 +1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 -1)
, "layer_norm");
2566 else
2567 ccv_nnc_graph_exec_symbol_new(graph, layer_norm, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output, saved_mean, saved_inv_std)(const ccv_nnc_tensor_symbol_t []){output, saved_mean, saved_inv_std
}, (1 +1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 -1)
, "layer_norm");
2568 outputs[0] = output;
2569}
2570
2571static void _ccv_cnnp_layer_norm_init_states(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_cnnp_state_initializer_f initializer, void* const context)
2572{
2573 ccv_cnnp_model_layer_norm_t* const self = (ccv_cnnp_model_layer_norm_t*)super;
2574 if (self->scale.graph)
2575 initializer(context, CMD_SET_FORWARD(1)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={1,}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->scale);
2576 if (self->bias.graph)
2577 initializer(context, CMD_SET_FORWARD(0)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={0,}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->bias);
2578}
2579
2580static void _ccv_cnnp_layer_norm_add_to_parameter(ccv_cnnp_model_t* const super, const ccv_cnnp_add_to_array_f add_to_array, void* const parameters, const int is_trainable)
2581{
2582 ccv_cnnp_model_layer_norm_t* const self = (ccv_cnnp_model_layer_norm_t*)super;
2583 if (self->scale.graph)
2584 add_to_array(parameters, self->scale, is_trainable);
2585 if (self->bias.graph)
2586 add_to_array(parameters, self->bias, is_trainable);
2587}
2588
2589static ccv_cnnp_model_t* _ccv_cnnp_layer_norm_copy(const ccv_cnnp_model_t* const super, void* const context);
2590
2591static const ccv_cnnp_model_vtab_t ccv_cnnp_layer_norm_isa = {
2592 .build = _ccv_cnnp_layer_norm_build,
2593 .init_states = _ccv_cnnp_layer_norm_init_states,
2594 .add_to_parameter = _ccv_cnnp_layer_norm_add_to_parameter,
2595 .copy = _ccv_cnnp_layer_norm_copy,
2596};
2597
2598ccv_cnnp_model_t* ccv_cnnp_layer_norm(const float epsilon, const int axis[CCV_NNC_MAX_DIM_ALLOC(12)], const int axis_count, const int elementwise_affine, const int is_trainable, const char* const name)
2599{
2600 ccv_cnnp_model_layer_norm_t* const model_layer_norm = (ccv_cnnp_model_layer_norm_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_layer_norm_t));
2601 model_layer_norm->super.isa = &ccv_cnnp_layer_norm_isa;
2602 model_layer_norm->super.input_size = 1;
2603 model_layer_norm->super.outputs = &model_layer_norm->output;
2604 model_layer_norm->super.output_size = 1;
2605 model_layer_norm->super.is_trainable = is_trainable;
2606 ccv_cnnp_model_copy_name(&model_layer_norm->super, name);
2607 model_layer_norm->scale.d = CCV_NNC_NO_TENSOR_SYMBOL;
2608 model_layer_norm->scale.graph = 0;
2609 model_layer_norm->bias.d = CCV_NNC_NO_TENSOR_SYMBOL;
2610 model_layer_norm->bias.graph = 0;
2611 model_layer_norm->params.lnorm.epsilon = epsilon;
2612 model_layer_norm->params.lnorm.count = axis_count;
2613 model_layer_norm->params.lnorm.elementwise_affine = elementwise_affine;
2614 memcpy(model_layer_norm->params.lnorm.axis, axis, sizeof(int) * axis_count);
2615 return (ccv_cnnp_model_t*)model_layer_norm;
2616}
2617
2618static ccv_cnnp_model_t* _ccv_cnnp_layer_norm_copy(const ccv_cnnp_model_t* const super, void* const context)
2619{
2620 const ccv_cnnp_model_layer_norm_t* const self = (const ccv_cnnp_model_layer_norm_t*)super;
2621 return ccv_cnnp_layer_norm(self->params.lnorm.epsilon, self->params.lnorm.axis, self->params.lnorm.count, self->params.lnorm.elementwise_affine, self->super.is_trainable, self->super.name);
2622}
2623
2624// MARK - Group Norm Layer
2625
2626typedef struct {
2627 ccv_cnnp_model_t super;
2628 ccv_nnc_tensor_symbol_t output;
2629 ccv_nnc_tensor_symbol_t bias;
2630 ccv_nnc_tensor_symbol_t scale;
2631 ccv_nnc_cmd_param_t params;
2632} ccv_cnnp_model_group_norm_t;
2633
2634static void _ccv_cnnp_group_norm_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
2635{
2636 PRINT(CCV_CLI_VERBOSE, "[cnnp_group_norm_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_group_norm_build] -\n"); fflush(stdout); } }
while (0)
;
2637 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 2637, __extension__ __PRETTY_FUNCTION__); }))
;
2638 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 2638, __extension__ __PRETTY_FUNCTION__
); }))
;
2639 ccv_cnnp_model_group_norm_t* const self = (ccv_cnnp_model_group_norm_t*)super;
2640 const ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
2641 ccv_nnc_tensor_param_t bias_params = params;
2642 const int nd = ccv_nnc_tensor_nd(params.dim);
2643 int i;
2644 for (i = 0; i < nd; i++)
2645 bias_params.dim[i] = 1;
2646 bias_params.dim[self->params.gnorm.group_axis] = params.dim[self->params.gnorm.group_axis];
2647 if (self->params.gnorm.elementwise_affine)
2648 {
2649 // Both scale and bias are shared between if this model is reused.
2650 if (!self->scale.graph)
2651 self->scale = ccv_nnc_tensor_symbol_new(graph, bias_params, "scale");
2652 if (!self->bias.graph)
2653 self->bias = ccv_nnc_tensor_symbol_new(graph, bias_params, "bias");
2654 }
2655 const ccv_nnc_cmd_t group_norm = ccv_nnc_cmd(CCV_NNC_GROUP_NORM_FORWARD, 0, self->params, 0);
2656 ccv_nnc_tensor_param_t output_params[3];
2657 if (self->params.gnorm.elementwise_affine)
2658 ccv_nnc_hint_tensor_auto(group_norm, (ccv_nnc_tensor_param_t []){
2659 params,
2660 bias_params,
2661 bias_params,
2662 }, 3, ccv_nnc_no_hint, output_params, 3);
2663 else
2664 ccv_nnc_hint_tensor_auto(group_norm, (ccv_nnc_tensor_param_t []){
2665 params,
2666 }, 1, ccv_nnc_no_hint, output_params, 3);
2667 const ccv_nnc_tensor_symbol_t output = ccv_nnc_tensor_symbol_new(graph, output_params[0], 0);
2668 const ccv_nnc_tensor_symbol_t saved_mean = ccv_nnc_tensor_symbol_new(graph, output_params[1], "saved_mean");
2669 const ccv_nnc_tensor_symbol_t saved_inv_std = ccv_nnc_tensor_symbol_new(graph, output_params[2], "saved_inv_std");
2670 if (self->params.gnorm.elementwise_affine)
2671 ccv_nnc_graph_exec_symbol_new(graph, group_norm, TENSOR_SYMBOL_LIST(inputs[0], self->scale, self->bias)(const ccv_nnc_tensor_symbol_t []){inputs[0], self->scale,
self->bias}, (1 +1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output, saved_mean, saved_inv_std)(const ccv_nnc_tensor_symbol_t []){output, saved_mean, saved_inv_std
}, (1 +1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 -1)
, "group_norm");
2672 else
2673 ccv_nnc_graph_exec_symbol_new(graph, group_norm, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output, saved_mean, saved_inv_std)(const ccv_nnc_tensor_symbol_t []){output, saved_mean, saved_inv_std
}, (1 +1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 -1)
, "group_norm");
2674 outputs[0] = output;
2675}
2676
2677static void _ccv_cnnp_group_norm_init_states(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_cnnp_state_initializer_f initializer, void* const context)
2678{
2679 ccv_cnnp_model_group_norm_t* const self = (ccv_cnnp_model_group_norm_t*)super;
2680 if (self->scale.graph)
2681 initializer(context, CMD_SET_FORWARD(1)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={1,}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->scale);
2682 if (self->bias.graph)
2683 initializer(context, CMD_SET_FORWARD(0)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={0,}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->bias);
2684}
2685
2686static void _ccv_cnnp_group_norm_add_to_parameter(ccv_cnnp_model_t* const super, const ccv_cnnp_add_to_array_f add_to_array, void* const parameters, const int is_trainable)
2687{
2688 ccv_cnnp_model_group_norm_t* const self = (ccv_cnnp_model_group_norm_t*)super;
2689 if (self->scale.graph)
2690 add_to_array(parameters, self->scale, is_trainable);
2691 if (self->bias.graph)
2692 add_to_array(parameters, self->bias, is_trainable);
2693}
2694
2695static ccv_cnnp_model_t* _ccv_cnnp_group_norm_copy(const ccv_cnnp_model_t* const super, void* const context);
2696
2697static const ccv_cnnp_model_vtab_t ccv_cnnp_group_norm_isa = {
2698 .build = _ccv_cnnp_group_norm_build,
2699 .init_states = _ccv_cnnp_group_norm_init_states,
2700 .add_to_parameter = _ccv_cnnp_group_norm_add_to_parameter,
2701 .copy = _ccv_cnnp_group_norm_copy,
2702};
2703
2704ccv_cnnp_model_t* ccv_cnnp_group_norm(const int group_axis, const int groups, const float epsilon, const int reduce_axis[CCV_NNC_MAX_DIM_ALLOC(12)], const int axis_count, const int elementwise_affine, const int is_trainable, const char* const name)
2705{
2706 ccv_cnnp_model_group_norm_t* const model_group_norm = (ccv_cnnp_model_group_norm_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_group_norm_t));
2707 model_group_norm->super.isa = &ccv_cnnp_group_norm_isa;
2708 model_group_norm->super.input_size = 1;
2709 model_group_norm->super.outputs = &model_group_norm->output;
2710 model_group_norm->super.output_size = 1;
2711 model_group_norm->super.is_trainable = is_trainable;
2712 ccv_cnnp_model_copy_name(&model_group_norm->super, name);
2713 model_group_norm->scale.d = CCV_NNC_NO_TENSOR_SYMBOL;
2714 model_group_norm->scale.graph = 0;
2715 model_group_norm->bias.d = CCV_NNC_NO_TENSOR_SYMBOL;
2716 model_group_norm->bias.graph = 0;
2717 model_group_norm->params.gnorm.group_axis = group_axis;
2718 model_group_norm->params.gnorm.groups = groups;
2719 model_group_norm->params.gnorm.epsilon = epsilon;
2720 model_group_norm->params.gnorm.reduce_count = axis_count;
2721 model_group_norm->params.gnorm.elementwise_affine = elementwise_affine;
2722 memcpy(model_group_norm->params.gnorm.reduce_axis, reduce_axis, sizeof(int) * axis_count);
2723 return (ccv_cnnp_model_t*)model_group_norm;
2724}
2725
2726static ccv_cnnp_model_t* _ccv_cnnp_group_norm_copy(const ccv_cnnp_model_t* const super, void* const context)
2727{
2728 const ccv_cnnp_model_group_norm_t* const self = (const ccv_cnnp_model_group_norm_t*)super;
2729 return ccv_cnnp_group_norm(self->params.gnorm.group_axis, self->params.gnorm.groups, self->params.gnorm.epsilon, self->params.gnorm.reduce_axis, self->params.gnorm.reduce_count, self->params.gnorm.elementwise_affine, self->super.is_trainable, self->super.name);
2730}
2731
2732// MARK - RMSNorm Layer
2733
2734typedef struct {
2735 ccv_cnnp_model_t super;
2736 ccv_nnc_tensor_symbol_t output;
2737 ccv_nnc_tensor_symbol_t scale;
2738 ccv_nnc_cmd_param_t params;
2739} ccv_cnnp_model_rmsnorm_t;
2740
2741static void _ccv_cnnp_rmsnorm_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
2742{
2743 PRINT(CCV_CLI_VERBOSE, "[cnnp_rmsnorm_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_rmsnorm_build] -\n"); fflush(stdout); } } while
(0)
;
2744 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 2744, __extension__ __PRETTY_FUNCTION__); }))
;
2745 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 2745, __extension__ __PRETTY_FUNCTION__
); }))
;
2746 ccv_cnnp_model_rmsnorm_t* const self = (ccv_cnnp_model_rmsnorm_t*)super;
2747 const ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
2748 ccv_nnc_tensor_param_t scale_params = params;
2749 const int nd = ccv_nnc_tensor_nd(params.dim);
2750 int i;
2751 for (i = 0; i < nd; i++)
2752 scale_params.dim[i] = 1;
2753 for (i = 0; i < self->params.rmsnorm.count; i++)
2754 scale_params.dim[self->params.rmsnorm.axis[i]] = params.dim[self->params.rmsnorm.axis[i]];
2755 // Both scale and bias are shared between if this model is reused.
2756 if (self->params.rmsnorm.elementwise_affine)
2757 {
2758 if (!self->scale.graph)
2759 self->scale = ccv_nnc_tensor_symbol_new(graph, scale_params, "scale");
2760 }
2761 const ccv_nnc_cmd_t rmsnorm = ccv_nnc_cmd(CCV_NNC_RMSNORM_FORWARD, 0, self->params, 0);
2762 ccv_nnc_tensor_param_t output_params[2];
2763 if (self->params.rmsnorm.elementwise_affine)
2764 ccv_nnc_hint_tensor_auto(rmsnorm, (ccv_nnc_tensor_param_t []){
2765 params,
2766 scale_params,
2767 }, 2, ccv_nnc_no_hint, output_params, 2);
2768 else
2769 ccv_nnc_hint_tensor_auto(rmsnorm, (ccv_nnc_tensor_param_t []){
2770 params,
2771 }, 1, ccv_nnc_no_hint, output_params, 2);
2772 const ccv_nnc_tensor_symbol_t output = ccv_nnc_tensor_symbol_new(graph, output_params[0], 0);
2773 const ccv_nnc_tensor_symbol_t saved_inv_std = ccv_nnc_tensor_symbol_new(graph, output_params[1], "saved_inv_std");
2774 if (self->params.rmsnorm.elementwise_affine)
2775 ccv_nnc_graph_exec_symbol_new(graph, rmsnorm, TENSOR_SYMBOL_LIST(inputs[0], self->scale)(const ccv_nnc_tensor_symbol_t []){inputs[0], self->scale}
, (1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output, saved_inv_std)(const ccv_nnc_tensor_symbol_t []){output, saved_inv_std}, (1
+1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 -1)
, "rmsnorm");
2776 else
2777 ccv_nnc_graph_exec_symbol_new(graph, rmsnorm, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output, saved_inv_std)(const ccv_nnc_tensor_symbol_t []){output, saved_inv_std}, (1
+1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 -1)
, "rmsnorm");
2778 outputs[0] = output;
2779}
2780
2781static void _ccv_cnnp_rmsnorm_init_states(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_cnnp_state_initializer_f initializer, void* const context)
2782{
2783 ccv_cnnp_model_rmsnorm_t* const self = (ccv_cnnp_model_rmsnorm_t*)super;
2784 if (self->scale.graph)
2785 initializer(context, CMD_SET_FORWARD(1)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={1,}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->scale);
2786}
2787
2788static void _ccv_cnnp_rmsnorm_add_to_parameter(ccv_cnnp_model_t* const super, const ccv_cnnp_add_to_array_f add_to_array, void* const parameters, const int is_trainable)
2789{
2790 ccv_cnnp_model_rmsnorm_t* const self = (ccv_cnnp_model_rmsnorm_t*)super;
2791 if (self->scale.graph)
2792 add_to_array(parameters, self->scale, is_trainable);
2793}
2794
2795static ccv_cnnp_model_t* _ccv_cnnp_rmsnorm_copy(const ccv_cnnp_model_t* const super, void* const context);
2796
2797static const ccv_cnnp_model_vtab_t ccv_cnnp_rmsnorm_isa = {
2798 .build = _ccv_cnnp_rmsnorm_build,
2799 .init_states = _ccv_cnnp_rmsnorm_init_states,
2800 .add_to_parameter = _ccv_cnnp_rmsnorm_add_to_parameter,
2801 .copy = _ccv_cnnp_rmsnorm_copy,
2802};
2803
2804ccv_cnnp_model_t* ccv_cnnp_rmsnorm(const float epsilon, const int axis[CCV_NNC_MAX_DIM_ALLOC(12)], const int axis_count, const int elementwise_affine, const int is_trainable, const char* const name)
2805{
2806 ccv_cnnp_model_rmsnorm_t* const model_rmsnorm = (ccv_cnnp_model_rmsnorm_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_rmsnorm_t));
2807 model_rmsnorm->super.isa = &ccv_cnnp_rmsnorm_isa;
2808 model_rmsnorm->super.input_size = 1;
2809 model_rmsnorm->super.outputs = &model_rmsnorm->output;
2810 model_rmsnorm->super.output_size = 1;
2811 model_rmsnorm->super.is_trainable = is_trainable;
2812 ccv_cnnp_model_copy_name(&model_rmsnorm->super, name);
2813 model_rmsnorm->scale.d = CCV_NNC_NO_TENSOR_SYMBOL;
2814 model_rmsnorm->scale.graph = 0;
2815 model_rmsnorm->params.rmsnorm.epsilon = epsilon;
2816 model_rmsnorm->params.rmsnorm.count = axis_count;
2817 model_rmsnorm->params.rmsnorm.elementwise_affine = elementwise_affine;
2818 memcpy(model_rmsnorm->params.lnorm.axis, axis, sizeof(int) * axis_count);
2819 return (ccv_cnnp_model_t*)model_rmsnorm;
2820}
2821
2822static ccv_cnnp_model_t* _ccv_cnnp_rmsnorm_copy(const ccv_cnnp_model_t* const super, void* const context)
2823{
2824 const ccv_cnnp_model_rmsnorm_t* const self = (const ccv_cnnp_model_rmsnorm_t*)super;
2825 return ccv_cnnp_rmsnorm(self->params.rmsnorm.epsilon, self->params.rmsnorm.axis, self->params.rmsnorm.count, self->params.rmsnorm.elementwise_affine, self->super.is_trainable, self->super.name);
2826}
2827
2828// MARK - Batched Matrix Mul Layer
2829
2830typedef struct {
2831 ccv_cnnp_model_t super;
2832 ccv_nnc_tensor_symbol_t output;
2833 int transpose_a[2];
2834 int transpose_b[2];
2835 int flags;
2836} ccv_cnnp_model_matmul_t;
2837
2838static void _ccv_cnnp_matmul_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
2839{
2840 PRINT(CCV_CLI_VERBOSE, "[cnnp_matmul_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_matmul_build] -\n"); fflush(stdout); } } while
(0)
;
2841 assert(input_size == 2)((void) sizeof ((input_size == 2) ? 1 : 0), __extension__ ({ if
(input_size == 2) ; else __assert_fail ("input_size == 2", "ccv_cnnp_model_addons.c"
, 2841, __extension__ __PRETTY_FUNCTION__); }))
;
2842 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 2842, __extension__ __PRETTY_FUNCTION__
); }))
;
2843 ccv_cnnp_model_matmul_t* const self = (ccv_cnnp_model_matmul_t*)super;
2844 ccv_nnc_tensor_param_t a_params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
2845 ccv_nnc_tensor_param_t b_params = ccv_nnc_tensor_symbol_params(graph, inputs[1]);
2846 ccv_nnc_tensor_param_t output_params;
2847 ccv_nnc_cmd_t matmul = CMD_GEMM_FORWARD(self->transpose_a, self->transpose_b)ccv_nnc_cmd(CCV_NNC_GEMM_FORWARD, 0, ((ccv_nnc_cmd_param_t){.
size={.dim={1,1,1}},.blas={.a={1,1},.transpose_a={self->transpose_a
[0],self->transpose_a[1]},.transpose_b={self->transpose_b
[0],self->transpose_b[1]},}}), 0)
;
2848 matmul.info.blas.flags = self->flags;
2849 ccv_nnc_hint_tensor_auto(matmul, (ccv_nnc_tensor_param_t []){
2850 a_params,
2851 b_params,
2852 }, 2, ccv_nnc_no_hint, &output_params, 1);
2853 const ccv_nnc_tensor_symbol_t matmul_output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
2854 ccv_nnc_graph_exec_symbol_new(graph, matmul, inputs, input_size, TENSOR_SYMBOL_LIST(matmul_output)(const ccv_nnc_tensor_symbol_t []){matmul_output}, (1 +1 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1
)
, "matmul");
2855 outputs[0] = matmul_output;
2856}
2857
2858static ccv_cnnp_model_t* _ccv_cnnp_matmul_copy(const ccv_cnnp_model_t* const super, void* const context);
2859
2860static const ccv_cnnp_model_vtab_t ccv_cnnp_matmul_isa = {
2861 .build = _ccv_cnnp_matmul_build,
2862 .copy = _ccv_cnnp_matmul_copy,
2863};
2864
2865ccv_cnnp_model_t* ccv_cnnp_matmul(const int transpose_a[2], const int transpose_b[2], const int flags, const char* const name)
2866{
2867 ccv_cnnp_model_matmul_t* const model_matmul = (ccv_cnnp_model_matmul_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_matmul_t));
2868 model_matmul->super.isa = &ccv_cnnp_matmul_isa;
2869 model_matmul->super.input_size = 2;
2870 model_matmul->super.outputs = &model_matmul->output;
2871 model_matmul->super.output_size = 1;
2872 model_matmul->transpose_a[0] = transpose_a[0];
2873 model_matmul->transpose_a[1] = transpose_a[1];
2874 model_matmul->transpose_b[0] = transpose_b[0];
2875 model_matmul->transpose_b[1] = transpose_b[1];
2876 model_matmul->flags = flags;
2877 ccv_cnnp_model_copy_name(&model_matmul->super, name);
2878 return (ccv_cnnp_model_t*)model_matmul;
2879}
2880
2881static ccv_cnnp_model_t* _ccv_cnnp_matmul_copy(const ccv_cnnp_model_t* const super, void* const context)
2882{
2883 const ccv_cnnp_model_matmul_t* const self = (const ccv_cnnp_model_matmul_t*)super;
2884 return ccv_cnnp_matmul(self->transpose_a, self->transpose_b, self->flags, self->super.name);
2885}
2886
2887// MARK - Dropout Layer
2888
2889typedef struct {
2890 ccv_cnnp_model_t super;
2891 ccv_nnc_tensor_symbol_t output;
2892 ccv_nnc_graph_exec_symbol_t dropout;
2893 float p;
2894 int entirety;
2895} ccv_cnnp_model_dropout_t;
2896
2897static void _ccv_cnnp_dropout_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
2898{
2899 PRINT(CCV_CLI_VERBOSE, "[cnnp_dropout_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_dropout_build] -\n"); fflush(stdout); } } while
(0)
;
2900 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 2900, __extension__ __PRETTY_FUNCTION__); }))
;
2901 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 2901, __extension__ __PRETTY_FUNCTION__
); }))
;
2902 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
2903 ccv_nnc_tensor_param_t output_params[2];
2904 ccv_cnnp_model_dropout_t* const self = (ccv_cnnp_model_dropout_t*)super;
2905 const ccv_nnc_cmd_t dropout = CMD_DROPOUT_FORWARD(self->p, self->entirety)ccv_nnc_cmd(CCV_NNC_DROPOUT_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.dropout={.p=self->p,.entirety=self
->entirety}}), 0)
;
2906 ccv_nnc_hint_tensor_auto(dropout, (ccv_nnc_tensor_param_t []){
2907 params,
2908 }, 1, ccv_nnc_no_hint, output_params, 2);
2909 const ccv_nnc_tensor_symbol_t dropout_output = ccv_nnc_tensor_symbol_new(graph, output_params[0], 0);
2910 const ccv_nnc_tensor_symbol_t mask = ccv_nnc_tensor_symbol_new(graph, output_params[1], "mask");
2911 self->dropout = ccv_nnc_graph_exec_symbol_new(graph, dropout, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(dropout_output, mask)(const ccv_nnc_tensor_symbol_t []){dropout_output, mask}, (1 +
1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 -1)
, "dropout");
2912 outputs[0] = dropout_output;
2913}
2914
2915static void _ccv_cnnp_dropout_set_is_test(ccv_cnnp_model_t* const super, const int is_test, const ccv_cnnp_cmd_updater_f updater, void* const context)
2916{
2917 ccv_cnnp_model_dropout_t* const self = (ccv_cnnp_model_dropout_t*)super;
2918 if (self->dropout.graph)
2919 {
2920 if (is_test)
2921 // During test, the dropout is not applied. Data transfer is perfect because if these are the same tensor, it will skip.
2922 updater(context, self->dropout, CMD_DATA_TRANSFER_FORWARD()ccv_nnc_cmd(CCV_NNC_DATA_TRANSFER_FORWARD, 0, ccv_nnc_cmd_auto
, 0)
, ccv_nnc_no_hint);
2923 else
2924 updater(context, self->dropout, CMD_DROPOUT_FORWARD(self->p, self->entirety)ccv_nnc_cmd(CCV_NNC_DROPOUT_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.dropout={.p=self->p,.entirety=self
->entirety}}), 0)
, ccv_nnc_no_hint);
2925 }
2926}
2927
2928static ccv_cnnp_model_t* _ccv_cnnp_dropout_copy(const ccv_cnnp_model_t* const super, void* const context);
2929
2930static const ccv_cnnp_model_vtab_t ccv_cnnp_dropout_isa = {
2931 .build = _ccv_cnnp_dropout_build,
2932 .set_is_test = _ccv_cnnp_dropout_set_is_test,
2933 .copy = _ccv_cnnp_dropout_copy,
2934};
2935
2936ccv_cnnp_model_t* ccv_cnnp_dropout(const float p, const int entirety, const char* const name)
2937{
2938 ccv_cnnp_model_dropout_t* const model_dropout = (ccv_cnnp_model_dropout_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_dropout_t));
2939 model_dropout->super.isa = &ccv_cnnp_dropout_isa;
2940 model_dropout->super.input_size = 1;
2941 model_dropout->super.outputs = &model_dropout->output;
2942 model_dropout->super.output_size = 1;
2943 model_dropout->p = p;
2944 model_dropout->entirety = entirety;
2945 ccv_cnnp_model_copy_name(&model_dropout->super, name);
2946 return (ccv_cnnp_model_t*)model_dropout;
2947}
2948
2949static ccv_cnnp_model_t* _ccv_cnnp_dropout_copy(const ccv_cnnp_model_t* const super, void* const context)
2950{
2951 const ccv_cnnp_model_dropout_t* const self = (const ccv_cnnp_model_dropout_t*)super;
2952 return ccv_cnnp_dropout(self->p, self->entirety, self->super.name);
2953}
2954
2955// MARK - Masked Fill Layer
2956
2957typedef struct {
2958 ccv_cnnp_model_t super;
2959 ccv_nnc_tensor_symbol_t output;
2960 float eq;
2961 float fill;
2962} ccv_cnnp_model_masked_fill_t;
2963
2964static void _ccv_cnnp_masked_fill_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
2965{
2966 PRINT(CCV_CLI_VERBOSE, "[cnnp_masked_fill_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_masked_fill_build] -\n"); fflush(stdout); } }
while (0)
;
2967 assert(input_size == 2)((void) sizeof ((input_size == 2) ? 1 : 0), __extension__ ({ if
(input_size == 2) ; else __assert_fail ("input_size == 2", "ccv_cnnp_model_addons.c"
, 2967, __extension__ __PRETTY_FUNCTION__); }))
;
2968 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 2968, __extension__ __PRETTY_FUNCTION__
); }))
;
2969 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
2970 ccv_cnnp_model_masked_fill_t* const self = (ccv_cnnp_model_masked_fill_t*)super;
2971 const ccv_nnc_tensor_symbol_t masked_fill_output = ccv_nnc_tensor_symbol_new(graph, params, 0);
2972 ccv_nnc_graph_exec_symbol_new(graph, CMD_MASKED_FILL_FORWARD(self->eq, self->fill)ccv_nnc_cmd(CCV_NNC_MASKED_FILL_FORWARD, 0, (ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.blas={.a={self->eq, self->fill}
}}, 0)
, TENSOR_SYMBOL_LIST(inputs[0], inputs[1])(const ccv_nnc_tensor_symbol_t []){inputs[0], inputs[1]}, (1 +
1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 -1)
, TENSOR_SYMBOL_LIST(masked_fill_output)(const ccv_nnc_tensor_symbol_t []){masked_fill_output}, (1 +1
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 -1)
, "masked_fill");
2973 outputs[0] = masked_fill_output;
2974}
2975
2976static ccv_cnnp_model_t* _ccv_cnnp_masked_fill_copy(const ccv_cnnp_model_t* const super, void* const context);
2977
2978static const ccv_cnnp_model_vtab_t ccv_cnnp_masked_fill_isa = {
2979 .build = _ccv_cnnp_masked_fill_build,
2980 .copy = _ccv_cnnp_masked_fill_copy,
2981};
2982
2983ccv_cnnp_model_t* ccv_cnnp_masked_fill(const float eq, const float fill, const char* const name)
2984{
2985 ccv_cnnp_model_masked_fill_t* const model_masked_fill = (ccv_cnnp_model_masked_fill_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_masked_fill_t));
2986 model_masked_fill->super.isa = &ccv_cnnp_masked_fill_isa;
2987 model_masked_fill->super.input_size = 2;
2988 model_masked_fill->super.outputs = &model_masked_fill->output;
2989 model_masked_fill->super.output_size = 1;
2990 model_masked_fill->eq = eq;
2991 model_masked_fill->fill = fill;
2992 ccv_cnnp_model_copy_name(&model_masked_fill->super, name);
2993 return (ccv_cnnp_model_t*)model_masked_fill;
2994}
2995
2996static ccv_cnnp_model_t* _ccv_cnnp_masked_fill_copy(const ccv_cnnp_model_t* const super, void* const context)
2997{
2998 const ccv_cnnp_model_masked_fill_t* const self = (const ccv_cnnp_model_masked_fill_t*)super;
2999 return ccv_cnnp_masked_fill(self->eq, self->fill, self->super.name);
3000}
3001
3002// MARK - Index Select Layer
3003
3004typedef struct {
3005 ccv_cnnp_model_t super;
3006 ccv_nnc_tensor_symbol_t output;
3007} ccv_cnnp_model_index_select_t;
3008
3009static void _ccv_cnnp_index_select_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
3010{
3011 PRINT(CCV_CLI_VERBOSE, "[cnnp_index_select_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_index_select_build] -\n"); fflush(stdout); }
} while (0)
;
3012 assert(input_size == 2)((void) sizeof ((input_size == 2) ? 1 : 0), __extension__ ({ if
(input_size == 2) ; else __assert_fail ("input_size == 2", "ccv_cnnp_model_addons.c"
, 3012, __extension__ __PRETTY_FUNCTION__); }))
;
3013 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 3013, __extension__ __PRETTY_FUNCTION__
); }))
;
3014 const ccv_nnc_tensor_param_t vocab_params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
3015 const ccv_nnc_tensor_param_t index_params = ccv_nnc_tensor_symbol_params(graph, inputs[1]);
3016 ccv_nnc_tensor_param_t output_params;
3017 const ccv_nnc_cmd_t index_select = CMD_INDEX_SELECT_FORWARD()ccv_nnc_cmd(CCV_NNC_INDEX_SELECT_FORWARD, 0, ccv_nnc_cmd_auto
, 0)
;
3018 ccv_nnc_hint_tensor_auto(index_select, (ccv_nnc_tensor_param_t []){
3019 vocab_params,
3020 index_params,
3021 }, 2, ccv_nnc_no_hint, &output_params, 1);
3022 const ccv_nnc_tensor_symbol_t output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
3023 ccv_nnc_graph_exec_symbol_new(graph, index_select, TENSOR_SYMBOL_LIST(inputs[0], inputs[1])(const ccv_nnc_tensor_symbol_t []){inputs[0], inputs[1]}, (1 +
1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 -1)
, TENSOR_SYMBOL_LIST(output)(const ccv_nnc_tensor_symbol_t []){output}, (1 +1 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "index_select");
3024 outputs[0] = output;
3025}
3026
3027static ccv_cnnp_model_t* _ccv_cnnp_index_select_copy(const ccv_cnnp_model_t* const super, void* const context);
3028
3029static const ccv_cnnp_model_vtab_t ccv_cnnp_index_select_isa = {
3030 .build = _ccv_cnnp_index_select_build,
3031 .copy = _ccv_cnnp_index_select_copy,
3032};
3033
3034ccv_cnnp_model_t* ccv_cnnp_index_select(const char* const name)
3035{
3036 ccv_cnnp_model_index_select_t* const model_index_select = (ccv_cnnp_model_index_select_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_index_select_t));
3037 model_index_select->super.isa = &ccv_cnnp_index_select_isa;
3038 model_index_select->super.input_size = 2;
3039 model_index_select->super.outputs = &model_index_select->output;
3040 model_index_select->super.output_size = 1;
3041 ccv_cnnp_model_copy_name(&model_index_select->super, name);
3042 return (ccv_cnnp_model_t*)model_index_select;
3043}
3044
3045static ccv_cnnp_model_t* _ccv_cnnp_index_select_copy(const ccv_cnnp_model_t* const super, void* const context)
3046{
3047 ccv_cnnp_model_index_select_t* const self = (ccv_cnnp_model_index_select_t*)super;
3048 return ccv_cnnp_index_select(self->super.name);
3049}
3050
3051// MARK - Embedding Layer
3052
3053typedef struct {
3054 ccv_cnnp_model_t super;
3055 ccv_nnc_tensor_symbol_t output;
3056 ccv_nnc_tensor_symbol_t vocab;
3057 int datatype;
3058 int vocab_size;
3059 int embed_size;
3060} ccv_cnnp_model_embedding_t;
3061
3062static void _ccv_cnnp_embedding_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
3063{
3064 ccv_cnnp_model_embedding_t* const self = (ccv_cnnp_model_embedding_t*)super;
3065 PRINT(CCV_CLI_VERBOSE, "[cnnp_embedding_build] vocab_size: %d, embed_size: %d\n", self->vocab_size, self->embed_size)do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_embedding_build] vocab_size: %d, embed_size: %d\n"
, self->vocab_size, self->embed_size); fflush(stdout); }
} while (0)
;
3066 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 3066, __extension__ __PRETTY_FUNCTION__); }))
;
3067 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 3067, __extension__ __PRETTY_FUNCTION__
); }))
;
3068 const ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
3069 ccv_nnc_tensor_param_t vocab_params = params;
3070 memset(vocab_params.dim, 0, sizeof(vocab_params.dim));
3071 vocab_params.datatype = self->datatype;
3072 vocab_params.dim[0] = self->vocab_size;
3073 vocab_params.dim[1] = self->embed_size;
3074 if (!self->vocab.graph)
3075 self->vocab = ccv_nnc_tensor_symbol_new(graph, vocab_params, "vocab");
3076 assert(self->vocab.graph == graph)((void) sizeof ((self->vocab.graph == graph) ? 1 : 0), __extension__
({ if (self->vocab.graph == graph) ; else __assert_fail (
"self->vocab.graph == graph", "ccv_cnnp_model_addons.c", 3076
, __extension__ __PRETTY_FUNCTION__); }))
;
3077 ccv_nnc_tensor_param_t output_params;
3078 const ccv_nnc_cmd_t embedding = CMD_INDEX_SELECT_FORWARD()ccv_nnc_cmd(CCV_NNC_INDEX_SELECT_FORWARD, 0, ccv_nnc_cmd_auto
, 0)
;
3079 ccv_nnc_hint_tensor_auto(embedding, (ccv_nnc_tensor_param_t []){
3080 vocab_params,
3081 params,
3082 }, 2, ccv_nnc_no_hint, &output_params, 1);
3083 const ccv_nnc_tensor_symbol_t output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
3084 ccv_nnc_graph_exec_symbol_new(graph, embedding, TENSOR_SYMBOL_LIST(self->vocab, inputs[0])(const ccv_nnc_tensor_symbol_t []){self->vocab, inputs[0]}
, (1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output)(const ccv_nnc_tensor_symbol_t []){output}, (1 +1 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "embedding");
3085 outputs[0] = output;
3086}
3087
3088static void _ccv_cnnp_embedding_init_states(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_cnnp_state_initializer_f initializer, void* const context)
3089{
3090 ccv_cnnp_model_embedding_t* const self = (ccv_cnnp_model_embedding_t*)super;
3091 const float std = sqrtf(2) / sqrtf(self->vocab_size + self->embed_size);
3092 const float bound = sqrtf(3) * std;
3093 initializer(context, CMD_RANDOM_UNIFORM_FORWARD(-bound, bound)ccv_nnc_cmd(CCV_NNC_RANDOM_UNIFORM_FORWARD, 0, (ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.blas={.a={-bound, bound}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->vocab);
3094}
3095
3096static void _ccv_cnnp_embedding_add_to_parameter(ccv_cnnp_model_t* const super, const ccv_cnnp_add_to_array_f add_to_array, void* const parameters, const int is_trainable)
3097{
3098 ccv_cnnp_model_embedding_t* const self = (ccv_cnnp_model_embedding_t*)super;
3099 add_to_array(parameters, self->vocab, is_trainable);
3100}
3101
3102static ccv_cnnp_model_t* _ccv_cnnp_embedding_copy(const ccv_cnnp_model_t* const super, void* const context);
3103
3104static const ccv_cnnp_model_vtab_t ccv_cnnp_embedding_isa = {
3105 .build = _ccv_cnnp_embedding_build,
3106 .init_states = _ccv_cnnp_embedding_init_states,
3107 .add_to_parameter = _ccv_cnnp_embedding_add_to_parameter,
3108 .copy = _ccv_cnnp_embedding_copy,
3109};
3110
3111ccv_cnnp_model_t* ccv_cnnp_embedding(const int datatype, const int vocab_size, const int embed_size, const int is_trainable, const char* const name)
3112{
3113 ccv_cnnp_model_embedding_t* const model_embedding = (ccv_cnnp_model_embedding_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_embedding_t));
3114 model_embedding->super.isa = &ccv_cnnp_embedding_isa;
3115 model_embedding->super.input_size = 1;
3116 model_embedding->super.outputs = &model_embedding->output;
3117 model_embedding->super.output_size = 1;
3118 model_embedding->super.is_trainable = is_trainable;
3119 ccv_cnnp_model_copy_name(&model_embedding->super, name);
3120 model_embedding->vocab.d = CCV_NNC_NO_TENSOR_SYMBOL;
3121 model_embedding->vocab.graph = 0;
3122 assert(datatype == CCV_32F || datatype == CCV_16F || datatype == CCV_16BF)((void) sizeof ((datatype == CCV_32F || datatype == CCV_16F ||
datatype == CCV_16BF) ? 1 : 0), __extension__ ({ if (datatype
== CCV_32F || datatype == CCV_16F || datatype == CCV_16BF) ;
else __assert_fail ("datatype == CCV_32F || datatype == CCV_16F || datatype == CCV_16BF"
, "ccv_cnnp_model_addons.c", 3122, __extension__ __PRETTY_FUNCTION__
); }))
;
3123 model_embedding->datatype = datatype;
3124 assert(vocab_size > 0)((void) sizeof ((vocab_size > 0) ? 1 : 0), __extension__ (
{ if (vocab_size > 0) ; else __assert_fail ("vocab_size > 0"
, "ccv_cnnp_model_addons.c", 3124, __extension__ __PRETTY_FUNCTION__
); }))
;
3125 model_embedding->vocab_size = vocab_size;
3126 assert(embed_size > 0)((void) sizeof ((embed_size > 0) ? 1 : 0), __extension__ (
{ if (embed_size > 0) ; else __assert_fail ("embed_size > 0"
, "ccv_cnnp_model_addons.c", 3126, __extension__ __PRETTY_FUNCTION__
); }))
;
3127 model_embedding->embed_size = embed_size;
3128 return (ccv_cnnp_model_t*)model_embedding;
3129}
3130
3131static ccv_cnnp_model_t* _ccv_cnnp_embedding_copy(const ccv_cnnp_model_t* const super, void* const context)
3132{
3133 ccv_cnnp_model_embedding_t* const self = (ccv_cnnp_model_embedding_t*)super;
3134 return ccv_cnnp_embedding(self->datatype, self->vocab_size, self->embed_size, self->super.is_trainable, self->super.name);
3135}
3136
3137// MARK - Pool Layers
3138
3139typedef struct {
3140 ccv_cnnp_model_t super;
3141 ccv_nnc_tensor_symbol_t output;
3142 int type;
3143 float width_scale;
3144 float height_scale;
3145 int align_corners;
3146} ccv_cnnp_model_upsample_t;
3147
3148static void _ccv_cnnp_upsample_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
3149{
3150 PRINT(CCV_CLI_VERBOSE, "[cnnp_upsample_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_upsample_build] -\n"); fflush(stdout); } } while
(0)
;
3151 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 3151, __extension__ __PRETTY_FUNCTION__); }))
;
3152 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 3152, __extension__ __PRETTY_FUNCTION__
); }))
;
3153 ccv_cnnp_model_upsample_t* const self = (ccv_cnnp_model_upsample_t*)super;
3154 const ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
3155 ccv_nnc_cmd_t cmd = CMD_UPSAMPLE_FORWARD(self->type, self->width_scale, self->height_scale, self->align_corners)ccv_nnc_cmd(CCV_NNC_UPSAMPLE_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.upsample={.type=self->type,.width_scale
=self->width_scale,.height_scale=self->height_scale,.align_corners
=self->align_corners}}), 0)
;
3156 ccv_nnc_tensor_param_t output_params;
3157 ccv_nnc_hint_tensor_auto(cmd, &params, 1, ccv_nnc_no_hint, &output_params, 1);
3158 const ccv_nnc_tensor_symbol_t output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
3159 ccv_nnc_graph_exec_symbol_new(graph, cmd, TENSOR_SYMBOL_LIST(inputs[0])(const ccv_nnc_tensor_symbol_t []){inputs[0]}, (1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output)(const ccv_nnc_tensor_symbol_t []){output}, (1 +1 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "upsample");
3160 outputs[0] = output;
3161}
3162
3163static ccv_cnnp_model_t* _ccv_cnnp_upsample_copy(const ccv_cnnp_model_t* const super, void* const context);
3164
3165static const ccv_cnnp_model_vtab_t ccv_cnnp_upsample_isa = {
3166 .build = _ccv_cnnp_upsample_build,
3167 .copy = _ccv_cnnp_upsample_copy,
3168};
3169
3170ccv_cnnp_model_t* ccv_cnnp_upsample(const int type, const float width_scale, const float height_scale, const int align_corners, const char* const name)
3171{
3172 ccv_cnnp_model_upsample_t* const model_upsample = (ccv_cnnp_model_upsample_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_upsample_t));
3173 model_upsample->super.isa = &ccv_cnnp_upsample_isa;
3174 model_upsample->super.input_size = 1;
3175 model_upsample->super.outputs = &model_upsample->output;
3176 model_upsample->super.output_size = 1;
3177 ccv_cnnp_model_copy_name(&model_upsample->super, name);
3178 assert(type == CCV_NNC_UPSAMPLE_NEAREST || type == CCV_NNC_UPSAMPLE_BILINEAR)((void) sizeof ((type == CCV_NNC_UPSAMPLE_NEAREST || type == CCV_NNC_UPSAMPLE_BILINEAR
) ? 1 : 0), __extension__ ({ if (type == CCV_NNC_UPSAMPLE_NEAREST
|| type == CCV_NNC_UPSAMPLE_BILINEAR) ; else __assert_fail (
"type == CCV_NNC_UPSAMPLE_NEAREST || type == CCV_NNC_UPSAMPLE_BILINEAR"
, "ccv_cnnp_model_addons.c", 3178, __extension__ __PRETTY_FUNCTION__
); }))
;
3179 model_upsample->type = type;
3180 model_upsample->width_scale = width_scale;
3181 model_upsample->height_scale = height_scale;
3182 model_upsample->align_corners = align_corners;
3183 return (ccv_cnnp_model_t*)model_upsample;
3184}
3185
3186static ccv_cnnp_model_t* _ccv_cnnp_upsample_copy(const ccv_cnnp_model_t* const super, void* const context)
3187{
3188 const ccv_cnnp_model_upsample_t* const self = (const ccv_cnnp_model_upsample_t*)super;
3189 return ccv_cnnp_upsample(self->type, self->width_scale, self->height_scale, self->align_corners, self->super.name);
3190}
3191
3192// MARK - Reduce Sum Layer
3193
3194typedef struct {
3195 ccv_cnnp_model_t super;
3196 int axis[CCV_NNC_MAX_DIM_ALLOC(12)];
3197 int count;
3198 ccv_nnc_tensor_symbol_t output;
3199} ccv_cnnp_model_reduce_sum_t;
3200
3201static void _ccv_cnnp_reduce_sum_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
3202{
3203 PRINT(CCV_CLI_VERBOSE, "[cnnp_reduce_sum_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_reduce_sum_build] -\n"); fflush(stdout); } }
while (0)
;
3204 const ccv_cnnp_model_reduce_sum_t* const self = (const ccv_cnnp_model_reduce_sum_t*)super;
3205 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 3205, __extension__ __PRETTY_FUNCTION__); }))
;
3206 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 3206, __extension__ __PRETTY_FUNCTION__
); }))
;
3207 ccv_nnc_tensor_param_t input_params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
3208 ccv_nnc_tensor_param_t output_params;
3209 ccv_nnc_cmd_t reduce_sum = CMD_REDUCE_SUM_FORWARD()ccv_nnc_cmd(CCV_NNC_REDUCE_SUM_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.reduce={.count=(1 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1),.axis={}}}
), 0)
;
3210 int i;
3211 for (i = 0; i < self->count; i++)
3212 reduce_sum.info.reduce.axis[i] = self->axis[i];
3213 reduce_sum.info.reduce.count = self->count;
3214 ccv_nnc_hint_tensor_auto(reduce_sum, &input_params, 1, ccv_nnc_no_hint, &output_params, 1);
3215 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
3216 ccv_nnc_graph_exec_symbol_new(graph, reduce_sum, inputs, input_size, outputs, output_size, "reduce_sum");
3217}
3218
3219static ccv_cnnp_model_t* _ccv_cnnp_reduce_sum_copy(const ccv_cnnp_model_t* const self, void* const context);
3220
3221static const ccv_cnnp_model_vtab_t ccv_cnnp_reduce_sum_isa = {
3222 .build = _ccv_cnnp_reduce_sum_build,
3223 .copy = _ccv_cnnp_reduce_sum_copy,
3224};
3225
3226ccv_cnnp_model_t* ccv_cnnp_reduce_sum(const int* const axis, const int axis_count, const char* const name)
3227{
3228 ccv_cnnp_model_reduce_sum_t* const model_reduce_sum = (ccv_cnnp_model_reduce_sum_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_reduce_sum_t));
3229 model_reduce_sum->super.isa = &ccv_cnnp_reduce_sum_isa;
3230 model_reduce_sum->super.input_size = 1;
3231 model_reduce_sum->super.outputs = &model_reduce_sum->output;
3232 model_reduce_sum->super.output_size = 1;
3233 ccv_cnnp_model_copy_name(&model_reduce_sum->super, name);
3234 assert(axis_count <= CCV_NNC_MAX_DIM_ALLOC)((void) sizeof ((axis_count <= (12)) ? 1 : 0), __extension__
({ if (axis_count <= (12)) ; else __assert_fail ("axis_count <= CCV_NNC_MAX_DIM_ALLOC"
, "ccv_cnnp_model_addons.c", 3234, __extension__ __PRETTY_FUNCTION__
); }))
;
3235 int i;
3236 for (i = 0; i < axis_count; i++)
3237 model_reduce_sum->axis[i] = axis[i];
3238 model_reduce_sum->count = axis_count;
3239 return (ccv_cnnp_model_t*)model_reduce_sum;
3240}
3241
3242static ccv_cnnp_model_t* _ccv_cnnp_reduce_sum_copy(const ccv_cnnp_model_t* const super, void* const context)
3243{
3244 const ccv_cnnp_model_reduce_sum_t* const self = (const ccv_cnnp_model_reduce_sum_t*)super;
3245 return ccv_cnnp_reduce_sum(self->axis, self->count, self->super.name);
3246}
3247
3248// MARK - Reduce Mean Layer
3249
3250typedef struct {
3251 ccv_cnnp_model_t super;
3252 int axis[CCV_NNC_MAX_DIM_ALLOC(12)];
3253 int count;
3254 ccv_nnc_tensor_symbol_t output;
3255} ccv_cnnp_model_reduce_mean_t;
3256
3257static void _ccv_cnnp_reduce_mean_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
3258{
3259 PRINT(CCV_CLI_VERBOSE, "[cnnp_reduce_mean_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_reduce_mean_build] -\n"); fflush(stdout); } }
while (0)
;
3260 const ccv_cnnp_model_reduce_mean_t* const self = (const ccv_cnnp_model_reduce_mean_t*)super;
3261 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 3261, __extension__ __PRETTY_FUNCTION__); }))
;
3262 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 3262, __extension__ __PRETTY_FUNCTION__
); }))
;
3263 ccv_nnc_tensor_param_t input_params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
3264 ccv_nnc_tensor_param_t output_params;
3265 ccv_nnc_cmd_t reduce_mean = CMD_REDUCE_MEAN_FORWARD()ccv_nnc_cmd(CCV_NNC_REDUCE_MEAN_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.reduce={.count=(1 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1),.axis={}}}
), 0)
;
3266 int i;
3267 for (i = 0; i < self->count; i++)
3268 reduce_mean.info.reduce.axis[i] = self->axis[i];
3269 reduce_mean.info.reduce.count = self->count;
3270 ccv_nnc_hint_tensor_auto(reduce_mean, &input_params, 1, ccv_nnc_no_hint, &output_params, 1);
3271 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
3272 ccv_nnc_graph_exec_symbol_new(graph, reduce_mean, inputs, input_size, outputs, output_size, "reduce_mean");
3273}
3274
3275static ccv_cnnp_model_t* _ccv_cnnp_reduce_mean_copy(const ccv_cnnp_model_t* const self, void* const context);
3276
3277static const ccv_cnnp_model_vtab_t ccv_cnnp_reduce_mean_isa = {
3278 .build = _ccv_cnnp_reduce_mean_build,
3279 .copy = _ccv_cnnp_reduce_mean_copy,
3280};
3281
3282ccv_cnnp_model_t* ccv_cnnp_reduce_mean(const int* const axis, const int axis_count, const char* const name)
3283{
3284 ccv_cnnp_model_reduce_mean_t* const model_reduce_mean = (ccv_cnnp_model_reduce_mean_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_reduce_mean_t));
3285 model_reduce_mean->super.isa = &ccv_cnnp_reduce_mean_isa;
3286 model_reduce_mean->super.input_size = 1;
3287 model_reduce_mean->super.outputs = &model_reduce_mean->output;
3288 model_reduce_mean->super.output_size = 1;
3289 ccv_cnnp_model_copy_name(&model_reduce_mean->super, name);
3290 assert(axis_count <= CCV_NNC_MAX_DIM_ALLOC)((void) sizeof ((axis_count <= (12)) ? 1 : 0), __extension__
({ if (axis_count <= (12)) ; else __assert_fail ("axis_count <= CCV_NNC_MAX_DIM_ALLOC"
, "ccv_cnnp_model_addons.c", 3290, __extension__ __PRETTY_FUNCTION__
); }))
;
3291 int i;
3292 for (i = 0; i < axis_count; i++)
3293 model_reduce_mean->axis[i] = axis[i];
3294 model_reduce_mean->count = axis_count;
3295 return (ccv_cnnp_model_t*)model_reduce_mean;
3296}
3297
3298static ccv_cnnp_model_t* _ccv_cnnp_reduce_mean_copy(const ccv_cnnp_model_t* const super, void* const context)
3299{
3300 const ccv_cnnp_model_reduce_mean_t* const self = (const ccv_cnnp_model_reduce_mean_t*)super;
3301 return ccv_cnnp_reduce_mean(self->axis, self->count, self->super.name);
3302}
3303
3304// MARK - Reduce Max Layer
3305
3306typedef struct {
3307 ccv_cnnp_model_t super;
3308 int axis[CCV_NNC_MAX_DIM_ALLOC(12)];
3309 int count;
3310 ccv_nnc_tensor_symbol_t output;
3311} ccv_cnnp_model_reduce_max_t;
3312
3313static void _ccv_cnnp_reduce_max_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
3314{
3315 PRINT(CCV_CLI_VERBOSE, "[cnnp_reduce_max_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_reduce_max_build] -\n"); fflush(stdout); } }
while (0)
;
3316 const ccv_cnnp_model_reduce_max_t* const self = (const ccv_cnnp_model_reduce_max_t*)super;
3317 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 3317, __extension__ __PRETTY_FUNCTION__); }))
;
3318 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 3318, __extension__ __PRETTY_FUNCTION__
); }))
;
3319 ccv_nnc_tensor_param_t input_params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
3320 ccv_nnc_tensor_param_t output_params;
3321 ccv_nnc_cmd_t reduce_max = CMD_REDUCE_MAX_FORWARD()ccv_nnc_cmd(CCV_NNC_REDUCE_MAX_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.reduce={.count=(1 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1),.axis={}}}
), 0)
;
3322 int i;
3323 for (i = 0; i < self->count; i++)
3324 reduce_max.info.reduce.axis[i] = self->axis[i];
3325 reduce_max.info.reduce.count = self->count;
3326 ccv_nnc_hint_tensor_auto(reduce_max, &input_params, 1, ccv_nnc_no_hint, &output_params, 1);
3327 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
3328 ccv_nnc_graph_exec_symbol_new(graph, reduce_max, inputs, input_size, outputs, output_size, "reduce_max");
3329}
3330
3331static ccv_cnnp_model_t* _ccv_cnnp_reduce_max_copy(const ccv_cnnp_model_t* const self, void* const context);
3332
3333static const ccv_cnnp_model_vtab_t ccv_cnnp_reduce_max_isa = {
3334 .build = _ccv_cnnp_reduce_max_build,
3335 .copy = _ccv_cnnp_reduce_max_copy,
3336};
3337
3338ccv_cnnp_model_t* ccv_cnnp_reduce_max(const int* const axis, const int axis_count, const char* const name)
3339{
3340 ccv_cnnp_model_reduce_max_t* const model_reduce_max = (ccv_cnnp_model_reduce_max_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_reduce_max_t));
3341 model_reduce_max->super.isa = &ccv_cnnp_reduce_max_isa;
3342 model_reduce_max->super.input_size = 1;
3343 model_reduce_max->super.outputs = &model_reduce_max->output;
3344 model_reduce_max->super.output_size = 1;
3345 ccv_cnnp_model_copy_name(&model_reduce_max->super, name);
3346 assert(axis_count <= CCV_NNC_MAX_DIM_ALLOC)((void) sizeof ((axis_count <= (12)) ? 1 : 0), __extension__
({ if (axis_count <= (12)) ; else __assert_fail ("axis_count <= CCV_NNC_MAX_DIM_ALLOC"
, "ccv_cnnp_model_addons.c", 3346, __extension__ __PRETTY_FUNCTION__
); }))
;
3347 int i;
3348 for (i = 0; i < axis_count; i++)
3349 model_reduce_max->axis[i] = axis[i];
3350 model_reduce_max->count = axis_count;
3351 return (ccv_cnnp_model_t*)model_reduce_max;
3352}
3353
3354static ccv_cnnp_model_t* _ccv_cnnp_reduce_max_copy(const ccv_cnnp_model_t* const super, void* const context)
3355{
3356 const ccv_cnnp_model_reduce_max_t* const self = (const ccv_cnnp_model_reduce_max_t*)super;
3357 return ccv_cnnp_reduce_max(self->axis, self->count, self->super.name);
3358}
3359
3360// MARK - Reduce Min Layer
3361
3362typedef struct {
3363 ccv_cnnp_model_t super;
3364 int axis[CCV_NNC_MAX_DIM_ALLOC(12)];
3365 int count;
3366 ccv_nnc_tensor_symbol_t output;
3367} ccv_cnnp_model_reduce_min_t;
3368
3369static void _ccv_cnnp_reduce_min_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
3370{
3371 PRINT(CCV_CLI_VERBOSE, "[cnnp_reduce_min_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_reduce_min_build] -\n"); fflush(stdout); } }
while (0)
;
3372 const ccv_cnnp_model_reduce_min_t* const self = (const ccv_cnnp_model_reduce_min_t*)super;
3373 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 3373, __extension__ __PRETTY_FUNCTION__); }))
;
3374 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 3374, __extension__ __PRETTY_FUNCTION__
); }))
;
3375 ccv_nnc_tensor_param_t input_params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
3376 ccv_nnc_tensor_param_t output_params;
3377 ccv_nnc_cmd_t reduce_min = CMD_REDUCE_MIN_FORWARD()ccv_nnc_cmd(CCV_NNC_REDUCE_MIN_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.reduce={.count=(1 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1),.axis={}}}
), 0)
;
3378 int i;
3379 for (i = 0; i < self->count; i++)
3380 reduce_min.info.reduce.axis[i] = self->axis[i];
3381 reduce_min.info.reduce.count = self->count;
3382 ccv_nnc_hint_tensor_auto(reduce_min, &input_params, 1, ccv_nnc_no_hint, &output_params, 1);
3383 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
3384 ccv_nnc_graph_exec_symbol_new(graph, reduce_min, inputs, input_size, outputs, output_size, "reduce_min");
3385}
3386
3387static ccv_cnnp_model_t* _ccv_cnnp_reduce_min_copy(const ccv_cnnp_model_t* const self, void* const context);
3388
3389static const ccv_cnnp_model_vtab_t ccv_cnnp_reduce_min_isa = {
3390 .build = _ccv_cnnp_reduce_min_build,
3391 .copy = _ccv_cnnp_reduce_min_copy,
3392};
3393
3394ccv_cnnp_model_t* ccv_cnnp_reduce_min(const int* const axis, const int axis_count, const char* const name)
3395{
3396 ccv_cnnp_model_reduce_min_t* const model_reduce_min = (ccv_cnnp_model_reduce_min_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_reduce_min_t));
3397 model_reduce_min->super.isa = &ccv_cnnp_reduce_min_isa;
3398 model_reduce_min->super.input_size = 1;
3399 model_reduce_min->super.outputs = &model_reduce_min->output;
3400 model_reduce_min->super.output_size = 1;
3401 ccv_cnnp_model_copy_name(&model_reduce_min->super, name);
3402 assert(axis_count <= CCV_NNC_MAX_DIM_ALLOC)((void) sizeof ((axis_count <= (12)) ? 1 : 0), __extension__
({ if (axis_count <= (12)) ; else __assert_fail ("axis_count <= CCV_NNC_MAX_DIM_ALLOC"
, "ccv_cnnp_model_addons.c", 3402, __extension__ __PRETTY_FUNCTION__
); }))
;
3403 int i;
3404 for (i = 0; i < axis_count; i++)
3405 model_reduce_min->axis[i] = axis[i];
3406 model_reduce_min->count = axis_count;
3407 return (ccv_cnnp_model_t*)model_reduce_min;
3408}
3409
3410static ccv_cnnp_model_t* _ccv_cnnp_reduce_min_copy(const ccv_cnnp_model_t* const super, void* const context)
3411{
3412 const ccv_cnnp_model_reduce_min_t* const self = (const ccv_cnnp_model_reduce_min_t*)super;
3413 return ccv_cnnp_reduce_min(self->axis, self->count, self->super.name);
3414}
3415
3416// MARK - Reduce Norm2 Layer
3417
3418typedef struct {
3419 ccv_cnnp_model_t super;
3420 int axis[CCV_NNC_MAX_DIM_ALLOC(12)];
3421 int count;
3422 ccv_nnc_tensor_symbol_t output;
3423} ccv_cnnp_model_reduce_norm2_t;
3424
3425static void _ccv_cnnp_reduce_norm2_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
3426{
3427 const ccv_cnnp_model_reduce_norm2_t* const self = (const ccv_cnnp_model_reduce_norm2_t*)super;
3428 PRINT(CCV_CLI_VERBOSE, "[cnnp_reduce_norm2_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_reduce_norm2_build] -\n"); fflush(stdout); }
} while (0)
;
3429 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 3429, __extension__ __PRETTY_FUNCTION__); }))
;
3430 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 3430, __extension__ __PRETTY_FUNCTION__
); }))
;
3431 ccv_nnc_tensor_param_t input_params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
3432 ccv_nnc_tensor_param_t output_params;
3433 ccv_nnc_cmd_t reduce_norm2 = CMD_REDUCE_NORM2_FORWARD()ccv_nnc_cmd(CCV_NNC_REDUCE_NORM2_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.reduce={.count=(1 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1),.axis={}}}
), 0)
;
3434 int i;
3435 for (i = 0; i < self->count; i++)
3436 reduce_norm2.info.reduce.axis[i] = self->axis[i];
3437 reduce_norm2.info.reduce.count = self->count;
3438 ccv_nnc_hint_tensor_auto(reduce_norm2, &input_params, 1, ccv_nnc_no_hint, &output_params, 1);
3439 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
3440 ccv_nnc_graph_exec_symbol_new(graph, reduce_norm2, inputs, input_size, outputs, output_size, "reduce_norm2");
3441}
3442
3443static ccv_cnnp_model_t* _ccv_cnnp_reduce_norm2_copy(const ccv_cnnp_model_t* const self, void* const context);
3444
3445static const ccv_cnnp_model_vtab_t ccv_cnnp_reduce_norm2_isa = {
3446 .build = _ccv_cnnp_reduce_norm2_build,
3447 .copy = _ccv_cnnp_reduce_norm2_copy,
3448};
3449
3450ccv_cnnp_model_t* ccv_cnnp_reduce_norm2(const int* const axis, const int axis_count, const char* const name)
3451{
3452 ccv_cnnp_model_reduce_norm2_t* const model_reduce_norm2 = (ccv_cnnp_model_reduce_norm2_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_reduce_norm2_t));
3453 model_reduce_norm2->super.isa = &ccv_cnnp_reduce_norm2_isa;
3454 model_reduce_norm2->super.input_size = 1;
3455 model_reduce_norm2->super.outputs = &model_reduce_norm2->output;
3456 model_reduce_norm2->super.output_size = 1;
3457 ccv_cnnp_model_copy_name(&model_reduce_norm2->super, name);
3458 assert(axis_count <= CCV_NNC_MAX_DIM_ALLOC)((void) sizeof ((axis_count <= (12)) ? 1 : 0), __extension__
({ if (axis_count <= (12)) ; else __assert_fail ("axis_count <= CCV_NNC_MAX_DIM_ALLOC"
, "ccv_cnnp_model_addons.c", 3458, __extension__ __PRETTY_FUNCTION__
); }))
;
3459 int i;
3460 for (i = 0; i < axis_count; i++)
3461 model_reduce_norm2->axis[i] = axis[i];
3462 model_reduce_norm2->count = axis_count;
3463 return (ccv_cnnp_model_t*)model_reduce_norm2;
3464}
3465
3466static ccv_cnnp_model_t* _ccv_cnnp_reduce_norm2_copy(const ccv_cnnp_model_t* const super, void* const context)
3467{
3468 const ccv_cnnp_model_reduce_norm2_t* const self = (const ccv_cnnp_model_reduce_norm2_t*)super;
3469 return ccv_cnnp_reduce_norm2(self->axis, self->count, self->super.name);
3470}
3471
3472// MARK - Argmax Layer
3473
3474typedef struct {
3475 ccv_cnnp_model_t super;
3476 int axis;
3477 ccv_nnc_tensor_symbol_t output;
3478} ccv_cnnp_model_argmax_t;
3479
3480static void _ccv_cnnp_argmax_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
3481{
3482 const ccv_cnnp_model_argmax_t* const self = (const ccv_cnnp_model_argmax_t*)super;
3483 PRINT(CCV_CLI_VERBOSE, "[cnnp_argmax_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_argmax_build] -\n"); fflush(stdout); } } while
(0)
;
3484 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 3484, __extension__ __PRETTY_FUNCTION__); }))
;
3485 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 3485, __extension__ __PRETTY_FUNCTION__
); }))
;
3486 ccv_nnc_tensor_param_t input_params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
3487 ccv_nnc_tensor_param_t output_params;
3488 ccv_nnc_cmd_t argmax = CMD_ARGMAX_FORWARD()ccv_nnc_cmd(CCV_NNC_ARGMAX_FORWARD, 0, ((ccv_nnc_cmd_param_t)
{.size={.dim={1,1,1}},.reduce={.count=(1 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1),.axis={}}})
, 0)
;
3489 argmax.info.reduce.axis[0] = self->axis;
3490 argmax.info.reduce.count = 1;
3491 ccv_nnc_hint_tensor_auto(argmax, &input_params, 1, ccv_nnc_no_hint, &output_params, 1);
3492 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
3493 ccv_nnc_graph_exec_symbol_new(graph, argmax, inputs, input_size, outputs, output_size, "argmax");
3494}
3495
3496static ccv_cnnp_model_t* _ccv_cnnp_argmax_copy(const ccv_cnnp_model_t* const self, void* const context);
3497
3498static const ccv_cnnp_model_vtab_t ccv_cnnp_argmax_isa = {
3499 .build = _ccv_cnnp_argmax_build,
3500 .copy = _ccv_cnnp_argmax_copy,
3501};
3502
3503ccv_cnnp_model_t* ccv_cnnp_argmax(const int axis, const char* const name)
3504{
3505 ccv_cnnp_model_argmax_t* const model_argmax = (ccv_cnnp_model_argmax_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_argmax_t));
3506 model_argmax->super.isa = &ccv_cnnp_argmax_isa;
3507 model_argmax->super.input_size = 1;
3508 model_argmax->super.outputs = &model_argmax->output;
3509 model_argmax->super.output_size = 1;
3510 ccv_cnnp_model_copy_name(&model_argmax->super, name);
3511 model_argmax->axis = axis;
3512 return (ccv_cnnp_model_t*)model_argmax;
3513}
3514
3515static ccv_cnnp_model_t* _ccv_cnnp_argmax_copy(const ccv_cnnp_model_t* const super, void* const context)
3516{
3517 const ccv_cnnp_model_argmax_t* const self = (const ccv_cnnp_model_argmax_t*)super;
3518 return ccv_cnnp_argmax(self->axis, self->super.name);
3519}
3520
3521// MARK - Argmin Layer
3522
3523typedef struct {
3524 ccv_cnnp_model_t super;
3525 int axis;
3526 ccv_nnc_tensor_symbol_t output;
3527} ccv_cnnp_model_argmin_t;
3528
3529static void _ccv_cnnp_argmin_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
3530{
3531 const ccv_cnnp_model_argmin_t* const self = (const ccv_cnnp_model_argmin_t*)super;
3532 PRINT(CCV_CLI_VERBOSE, "[cnnp_argmin_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_argmin_build] -\n"); fflush(stdout); } } while
(0)
;
3533 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 3533, __extension__ __PRETTY_FUNCTION__); }))
;
3534 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 3534, __extension__ __PRETTY_FUNCTION__
); }))
;
3535 ccv_nnc_tensor_param_t input_params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
3536 ccv_nnc_tensor_param_t output_params;
3537 ccv_nnc_cmd_t argmin = CMD_ARGMIN_FORWARD()ccv_nnc_cmd(CCV_NNC_ARGMIN_FORWARD, 0, ((ccv_nnc_cmd_param_t)
{.size={.dim={1,1,1}},.reduce={.count=(1 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1),.axis={}}})
, 0)
;
3538 argmin.info.reduce.axis[0] = self->axis;
3539 argmin.info.reduce.count = 1;
3540 ccv_nnc_hint_tensor_auto(argmin, &input_params, 1, ccv_nnc_no_hint, &output_params, 1);
3541 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
3542 ccv_nnc_graph_exec_symbol_new(graph, argmin, inputs, input_size, outputs, output_size, "argmin");
3543}
3544
3545static ccv_cnnp_model_t* _ccv_cnnp_argmin_copy(const ccv_cnnp_model_t* const self, void* const context);
3546
3547static const ccv_cnnp_model_vtab_t ccv_cnnp_argmin_isa = {
3548 .build = _ccv_cnnp_argmin_build,
3549 .copy = _ccv_cnnp_argmin_copy,
3550};
3551
3552ccv_cnnp_model_t* ccv_cnnp_argmin(const int axis, const char* const name)
3553{
3554 ccv_cnnp_model_argmin_t* const model_argmin = (ccv_cnnp_model_argmin_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_argmin_t));
3555 model_argmin->super.isa = &ccv_cnnp_argmin_isa;
3556 model_argmin->super.input_size = 1;
3557 model_argmin->super.outputs = &model_argmin->output;
3558 model_argmin->super.output_size = 1;
3559 ccv_cnnp_model_copy_name(&model_argmin->super, name);
3560 model_argmin->axis = axis;
3561 return (ccv_cnnp_model_t*)model_argmin;
3562}
3563
3564static ccv_cnnp_model_t* _ccv_cnnp_argmin_copy(const ccv_cnnp_model_t* const super, void* const context)
3565{
3566 const ccv_cnnp_model_argmin_t* const self = (const ccv_cnnp_model_argmin_t*)super;
3567 return ccv_cnnp_argmin(self->axis, self->super.name);
3568}
3569
3570// MARK - Min Layer
3571
3572typedef struct {
3573 ccv_cnnp_model_t super;
3574 ccv_nnc_tensor_symbol_t output;
3575} ccv_cnnp_model_min_t;
3576
3577static void _ccv_cnnp_min_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
3578{
3579 PRINT(CCV_CLI_VERBOSE, "[cnnp_min_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_min_build] -\n"); fflush(stdout); } } while (
0)
;
3580 assert(input_size == 2)((void) sizeof ((input_size == 2) ? 1 : 0), __extension__ ({ if
(input_size == 2) ; else __assert_fail ("input_size == 2", "ccv_cnnp_model_addons.c"
, 3580, __extension__ __PRETTY_FUNCTION__); }))
;
3581 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 3581, __extension__ __PRETTY_FUNCTION__
); }))
;
3582 ccv_nnc_tensor_param_t input_params[2];
3583 int i;
3584 for (i = 0; i < 2; i++)
3585 input_params[i] = ccv_nnc_tensor_symbol_params(graph, inputs[i]);
3586 ccv_nnc_tensor_param_t output_params;
3587 const ccv_nnc_cmd_t min = CMD_MIN_FORWARD()ccv_nnc_cmd(CCV_NNC_MIN_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}}}, 0)
;
3588 ccv_nnc_hint_tensor_auto(min, input_params, 2, ccv_nnc_no_hint, &output_params, 1);
3589 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
3590 ccv_nnc_graph_exec_symbol_new(graph, min, inputs, input_size, outputs, output_size, "min");
3591}
3592
3593static ccv_cnnp_model_t* _ccv_cnnp_min_copy(const ccv_cnnp_model_t* const self, void* const context);
3594
3595static const ccv_cnnp_model_vtab_t ccv_cnnp_min_isa = {
3596 .build = _ccv_cnnp_min_build,
3597 .copy = _ccv_cnnp_min_copy,
3598};
3599
3600ccv_cnnp_model_t* ccv_cnnp_min(const char* const name)
3601{
3602 ccv_cnnp_model_min_t* const model_min = (ccv_cnnp_model_min_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_min_t));
3603 model_min->super.isa = &ccv_cnnp_min_isa;
3604 model_min->super.input_size = 2;
3605 model_min->super.outputs = &model_min->output;
3606 model_min->super.output_size = 1;
3607 ccv_cnnp_model_copy_name(&model_min->super, name);
3608 return (ccv_cnnp_model_t*)model_min;
3609}
3610
3611static ccv_cnnp_model_t* _ccv_cnnp_min_copy(const ccv_cnnp_model_t* const super, void* const context)
3612{
3613 const ccv_cnnp_model_min_t* const self = (const ccv_cnnp_model_min_t*)super;
3614 return ccv_cnnp_min(self->super.name);
3615}
3616
3617// MARK - Max Layer
3618
3619typedef struct {
3620 ccv_cnnp_model_t super;
3621 ccv_nnc_tensor_symbol_t output;
3622} ccv_cnnp_model_max_t;
3623
3624static void _ccv_cnnp_max_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
3625{
3626 PRINT(CCV_CLI_VERBOSE, "[cnnp_max_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_max_build] -\n"); fflush(stdout); } } while (
0)
;
3627 assert(input_size == 2)((void) sizeof ((input_size == 2) ? 1 : 0), __extension__ ({ if
(input_size == 2) ; else __assert_fail ("input_size == 2", "ccv_cnnp_model_addons.c"
, 3627, __extension__ __PRETTY_FUNCTION__); }))
;
3628 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 3628, __extension__ __PRETTY_FUNCTION__
); }))
;
3629 ccv_nnc_tensor_param_t input_params[2];
3630 int i;
3631 for (i = 0; i < 2; i++)
3632 input_params[i] = ccv_nnc_tensor_symbol_params(graph, inputs[i]);
3633 ccv_nnc_tensor_param_t output_params;
3634 const ccv_nnc_cmd_t max = CMD_MAX_FORWARD()ccv_nnc_cmd(CCV_NNC_MAX_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}}}, 0)
;
3635 ccv_nnc_hint_tensor_auto(max, input_params, 2, ccv_nnc_no_hint, &output_params, 1);
3636 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
3637 ccv_nnc_graph_exec_symbol_new(graph, max, inputs, input_size, outputs, output_size, "max");
3638}
3639
3640static ccv_cnnp_model_t* _ccv_cnnp_max_copy(const ccv_cnnp_model_t* const self, void* const context);
3641
3642static const ccv_cnnp_model_vtab_t ccv_cnnp_max_isa = {
3643 .build = _ccv_cnnp_max_build,
3644 .copy = _ccv_cnnp_max_copy,
3645};
3646
3647ccv_cnnp_model_t* ccv_cnnp_max(const char* const name)
3648{
3649 ccv_cnnp_model_max_t* const model_max = (ccv_cnnp_model_max_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_max_t));
3650 model_max->super.isa = &ccv_cnnp_max_isa;
3651 model_max->super.input_size = 2;
3652 model_max->super.outputs = &model_max->output;
3653 model_max->super.output_size = 1;
3654 ccv_cnnp_model_copy_name(&model_max->super, name);
3655 return (ccv_cnnp_model_t*)model_max;
3656}
3657
3658static ccv_cnnp_model_t* _ccv_cnnp_max_copy(const ccv_cnnp_model_t* const super, void* const context)
3659{
3660 const ccv_cnnp_model_max_t* const self = (const ccv_cnnp_model_max_t*)super;
3661 return ccv_cnnp_max(self->super.name);
3662}
3663
3664// MARK - LSTM Layer
3665
3666typedef struct {
3667 ccv_cnnp_model_t super;
3668 int masked;
3669 ccv_nnc_tensor_symbol_t output;
3670 ccv_nnc_tensor_symbol_t weights;
3671 ccv_nnc_tensor_symbol_t reserves;
3672 ccv_nnc_cmd_param_t params;
3673 ccv_nnc_graph_exec_symbol_t lstm;
3674} ccv_cnnp_model_lstm_t;
3675
3676static int _ccv_cnnp_lstm_weight_dim(int bidirectional, int num_layers, int input_size, int hidden_size, int proj_size, int bias)
3677{
3678 const int D = !!bidirectional + 1;
3679 if (hidden_size == proj_size)
3680 return (num_layers * (bias ? 8 : 0) + (num_layers - 1) * (hidden_size * 4 * D + hidden_size * 4) + input_size * 4 + hidden_size * 4) * D;
3681 else
3682 return (num_layers * (bias ? 8 : 0) + (num_layers - 1) * (proj_size * 4 * D + proj_size * 4) + (proj_size * 4 + input_size * 4) + num_layers * proj_size) * D;
3683}
3684
3685static void _ccv_cnnp_lstm_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
3686{
3687 ccv_cnnp_model_lstm_t* const self = (ccv_cnnp_model_lstm_t*)super;
3688 PRINT(CCV_CLI_VERBOSE, "[cnnp_lstm_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_lstm_build] -\n"); fflush(stdout); } } while
(0)
;
3689 assert(input_size == self->super.input_size)((void) sizeof ((input_size == self->super.input_size) ? 1
: 0), __extension__ ({ if (input_size == self->super.input_size
) ; else __assert_fail ("input_size == self->super.input_size"
, "ccv_cnnp_model_addons.c", 3689, __extension__ __PRETTY_FUNCTION__
); }))
;
3690 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 3690, __extension__ __PRETTY_FUNCTION__
); }))
;
3691 const int proj_size = self->params.rnn.proj_size == 0 ? self->params.rnn.hidden_size : self->params.rnn.proj_size;
3692 ccv_nnc_tensor_param_t input_params[5];
3693 input_params[0]= ccv_nnc_tensor_symbol_params(graph, inputs[0]);
3694 if (input_size == 2)
3695 input_params[1] = ccv_nnc_tensor_symbol_params(graph, inputs[1]);
3696 input_params[4] = input_params[0];
3697 memset(input_params[4].dim, 0, sizeof(input_params[4].dim));
3698 const int x_nd = ccv_nnc_tensor_nd(input_params[0].dim);
3699 const int feature_count = input_params[0].dim[x_nd - 1];
3700 input_params[4].dim[0] = _ccv_cnnp_lstm_weight_dim(self->params.rnn.bidirectional, self->params.rnn.num_layers, feature_count, self->params.rnn.hidden_size, proj_size, self->params.rnn.bias);
3701 input_params[4].dim[1] = self->params.rnn.hidden_size;
3702 const ccv_nnc_cmd_t lstm = ccv_nnc_cmd(CCV_NNC_LSTM_FORWARD, 0, self->params, 0);
3703 ccv_nnc_tensor_param_t output_params[4];
3704 ccv_nnc_hint_tensor_auto(lstm, input_params, 5, ccv_nnc_no_hint, output_params, 4);
3705 outputs[0] = ccv_nnc_tensor_symbol_new(graph, output_params[0], 0);
3706 if (!self->weights.graph)
3707 self->weights = ccv_nnc_tensor_symbol_new(graph, input_params[4], "weights");
3708 if (!self->reserves.graph)
3709 self->reserves = ccv_nnc_tensor_symbol_new(graph, output_params[3], "reserves");
3710 const ccv_nnc_tensor_symbol_t mask = input_size == 2 ? inputs[1] : NO_TENSOR_SYMBOL(const ccv_nnc_tensor_symbol_t){.d = CCV_NNC_NO_TENSOR_SYMBOL
}
;
3711 self->lstm = ccv_nnc_graph_exec_symbol_new(graph, lstm, TENSOR_SYMBOL_LIST(inputs[0], mask, NO_TENSOR_SYMBOL, NO_TENSOR_SYMBOL, self->weights)(const ccv_nnc_tensor_symbol_t []){inputs[0], mask, (const ccv_nnc_tensor_symbol_t
){.d = CCV_NNC_NO_TENSOR_SYMBOL}, (const ccv_nnc_tensor_symbol_t
){.d = CCV_NNC_NO_TENSOR_SYMBOL}, self->weights}, (1 +1 +1
+1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -
1)
, TENSOR_SYMBOL_LIST(outputs[0], NO_TENSOR_SYMBOL, NO_TENSOR_SYMBOL, self->reserves)(const ccv_nnc_tensor_symbol_t []){outputs[0], (const ccv_nnc_tensor_symbol_t
){.d = CCV_NNC_NO_TENSOR_SYMBOL}, (const ccv_nnc_tensor_symbol_t
){.d = CCV_NNC_NO_TENSOR_SYMBOL}, self->reserves}, (1 +1 +
1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
-1)
, "lstm");
3712}
3713
3714static void _ccv_cnnp_lstm_init_states(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_cnnp_state_initializer_f initializer, void* const context)
3715{
3716 ccv_cnnp_model_lstm_t* const self = (ccv_cnnp_model_lstm_t*)super;
3717 if (self->weights.graph)
3718 {
3719 const float stdv = 1.0 / sqrt(self->params.rnn.hidden_size);
3720 initializer(context, CMD_RANDOM_UNIFORM_FORWARD(-stdv, stdv)ccv_nnc_cmd(CCV_NNC_RANDOM_UNIFORM_FORWARD, 0, (ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.blas={.a={-stdv, stdv}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->weights);
3721 }
3722}
3723
3724static void _ccv_cnnp_lstm_add_to_parameter(ccv_cnnp_model_t* const super, const ccv_cnnp_add_to_array_f add_to_array, void* const parameters, const int is_trainable)
3725{
3726 ccv_cnnp_model_lstm_t* const self = (ccv_cnnp_model_lstm_t*)super;
3727 if (self->weights.graph)
3728 add_to_array(parameters, self->weights, is_trainable);
3729}
3730
3731static void _ccv_cnnp_lstm_set_is_test(ccv_cnnp_model_t* const super, const int is_test, const ccv_cnnp_cmd_updater_f updater, void* const context)
3732{
3733 ccv_cnnp_model_lstm_t* const self = (ccv_cnnp_model_lstm_t*)super;
3734 if (self->lstm.graph)
3735 {
3736 self->params.rnn.is_test = is_test;
3737 updater(context, self->lstm, ccv_nnc_cmd(CCV_NNC_LSTM_FORWARD, 0, self->params, 0), ccv_nnc_no_hint);
3738 }
3739}
3740
3741static ccv_cnnp_model_t* _ccv_cnnp_lstm_copy(const ccv_cnnp_model_t* const self, void* const context);
3742
3743static const ccv_cnnp_model_vtab_t ccv_cnnp_lstm_isa = {
3744 .build = _ccv_cnnp_lstm_build,
3745 .init_states = _ccv_cnnp_lstm_init_states,
3746 .add_to_parameter = _ccv_cnnp_lstm_add_to_parameter,
3747 .copy = _ccv_cnnp_lstm_copy,
3748 .set_is_test = _ccv_cnnp_lstm_set_is_test,
3749};
3750
3751ccv_cnnp_model_t* ccv_cnnp_lstm(const int masked, const int hidden_size, const int proj_size, const int num_layers, const int bias, const int batch_first, const int bidirectional, const float dropout, const int is_trainable, const char* const name)
3752{
3753 ccv_cnnp_model_lstm_t* const model_lstm = (ccv_cnnp_model_lstm_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_lstm_t));
3754 model_lstm->super.isa = &ccv_cnnp_lstm_isa;
3755 model_lstm->super.input_size = masked ? 2 : 1;
3756 model_lstm->super.outputs = &model_lstm->output;
3757 model_lstm->super.output_size = 1;
3758 model_lstm->super.is_trainable = is_trainable;
3759 ccv_cnnp_model_copy_name(&model_lstm->super, name);
3760 model_lstm->masked = masked;
3761 model_lstm->weights.d = CCV_NNC_NO_TENSOR_SYMBOL;
3762 model_lstm->weights.graph = 0;
3763 model_lstm->params.rnn.hidden_size = hidden_size;
3764 model_lstm->params.rnn.proj_size = proj_size;
3765 model_lstm->params.rnn.num_layers = num_layers;
3766 model_lstm->params.rnn.bias = bias;
3767 model_lstm->params.rnn.batch_first = batch_first;
3768 model_lstm->params.rnn.bidirectional = bidirectional;
3769 model_lstm->params.rnn.dropout = dropout;
3770 return (ccv_cnnp_model_t*)model_lstm;
3771}
3772
3773static ccv_cnnp_model_t* _ccv_cnnp_lstm_copy(const ccv_cnnp_model_t* const super, void* const context)
3774{
3775 const ccv_cnnp_model_lstm_t* const self = (const ccv_cnnp_model_lstm_t*)super;
3776 return ccv_cnnp_lstm(self->masked, self->params.rnn.hidden_size, self->params.rnn.proj_size, self->params.rnn.num_layers, self->params.rnn.bias, self->params.rnn.batch_first, self->params.rnn.bidirectional, self->params.rnn.dropout, self->super.is_trainable, self->super.name);
3777}
3778
3779/// MARK - Datatype conversion layer.
3780
3781typedef struct {
3782 ccv_cnnp_model_t super;
3783 ccv_nnc_tensor_symbol_t output;
3784 int datatype;
3785 int ref_to_last;
3786} ccv_cnnp_model_datatype_conversion_t;
3787
3788static void _ccv_cnnp_datatype_conversion_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
3789{
3790 ccv_cnnp_model_datatype_conversion_t* const self = (ccv_cnnp_model_datatype_conversion_t*)super;
3791 PRINT(CCV_CLI_VERBOSE, "[cnnp_datatype_conversion_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_datatype_conversion_build] -\n"); fflush(stdout
); } } while (0)
;
3792 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
3793 if (self->ref_to_last)
3794 {
3795 assert(input_size > 1)((void) sizeof ((input_size > 1) ? 1 : 0), __extension__ (
{ if (input_size > 1) ; else __assert_fail ("input_size > 1"
, "ccv_cnnp_model_addons.c", 3795, __extension__ __PRETTY_FUNCTION__
); }))
;
3796 const ccv_nnc_tensor_param_t last_params = ccv_nnc_tensor_symbol_params(graph, inputs[input_size - 1]);
3797 params.datatype = last_params.datatype;
3798 } else
3799 params.datatype = self->datatype;
3800 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 3800, __extension__ __PRETTY_FUNCTION__
); }))
;
3801 outputs[0] = ccv_nnc_tensor_symbol_new(graph, params, 0);
3802 ccv_nnc_graph_exec_symbol_new(graph, CMD_DATATYPE_CONVERSION_FORWARD()ccv_nnc_cmd(CCV_NNC_DATATYPE_CONVERSION_FORWARD, 0, ccv_nnc_cmd_auto
, 0)
, inputs, output_size /* intentional */, outputs, output_size, 0);
3803}
3804
3805static ccv_cnnp_model_t* _ccv_cnnp_datatype_conversion_copy(const ccv_cnnp_model_t* const self, void* const context);
3806
3807static const ccv_cnnp_model_vtab_t ccv_cnnp_datatype_conversion_isa = {
3808 .build = _ccv_cnnp_datatype_conversion_build,
3809 .copy = _ccv_cnnp_datatype_conversion_copy,
3810};
3811
3812ccv_cnnp_model_t* ccv_cnnp_datatype_conversion(const int datatype, const int ref_to_last, const char* const name)
3813{
3814 ccv_cnnp_model_datatype_conversion_t* const model_datatype_conversion = (ccv_cnnp_model_datatype_conversion_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_datatype_conversion_t));
3815 model_datatype_conversion->super.isa = &ccv_cnnp_datatype_conversion_isa;
3816 model_datatype_conversion->super.input_size = 0;
3817 model_datatype_conversion->super.outputs = &model_datatype_conversion->output;
3818 model_datatype_conversion->super.output_size = 1;
3819 model_datatype_conversion->datatype = datatype;
3820 model_datatype_conversion->ref_to_last = ref_to_last;
3821 ccv_cnnp_model_copy_name(&model_datatype_conversion->super, name);
3822 return (ccv_cnnp_model_t*)model_datatype_conversion;
3823}
3824
3825static ccv_cnnp_model_t* _ccv_cnnp_datatype_conversion_copy(const ccv_cnnp_model_t* const super, void* const context)
3826{
3827 ccv_cnnp_model_datatype_conversion_t* const self = (ccv_cnnp_model_datatype_conversion_t*)super;
3828 return ccv_cnnp_datatype_conversion(self->datatype, self->ref_to_last, self->super.name);
3829}
3830
3831/// MARK - Clamp layer.
3832
3833typedef struct {
3834 ccv_cnnp_model_t super;
3835 ccv_nnc_tensor_symbol_t output;
3836 float min;
3837 float max;
3838} ccv_cnnp_model_clamp_t;
3839
3840static void _ccv_cnnp_clamp_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
3841{
3842 ccv_cnnp_model_clamp_t* const self = (ccv_cnnp_model_clamp_t*)super;
3843 PRINT(CCV_CLI_VERBOSE, "[cnnp_clamp_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_clamp_build] -\n"); fflush(stdout); } } while
(0)
;
3844 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
3845 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 3845, __extension__ __PRETTY_FUNCTION__
); }))
;
3846 outputs[0] = ccv_nnc_tensor_symbol_new(graph, params, 0);
3847 ccv_nnc_graph_exec_symbol_new(graph, CMD_CLAMP_FORWARD(self->min, self->max)ccv_nnc_cmd(CCV_NNC_CLAMP_FORWARD, 0, (ccv_nnc_cmd_param_t){.
size={.dim={1,1,1}},.clamp={.min=self->min,.max=self->max
}}, 0)
, inputs, output_size /* intentional */, outputs, output_size, 0);
3848}
3849
3850static ccv_cnnp_model_t* _ccv_cnnp_clamp_copy(const ccv_cnnp_model_t* const self, void* const context);
3851
3852static const ccv_cnnp_model_vtab_t ccv_cnnp_clamp_isa = {
3853 .build = _ccv_cnnp_clamp_build,
3854 .copy = _ccv_cnnp_clamp_copy,
3855};
3856
3857ccv_cnnp_model_t* ccv_cnnp_clamp(const float min, const float max, const char* const name)
3858{
3859 ccv_cnnp_model_clamp_t* const model_clamp = (ccv_cnnp_model_clamp_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_clamp_t));
3860 model_clamp->super.isa = &ccv_cnnp_clamp_isa;
3861 model_clamp->super.input_size = 0;
3862 model_clamp->super.outputs = &model_clamp->output;
3863 model_clamp->super.output_size = 1;
3864 model_clamp->min = min;
3865 model_clamp->max = max;
3866 ccv_cnnp_model_copy_name(&model_clamp->super, name);
3867 return (ccv_cnnp_model_t*)model_clamp;
3868}
3869
3870static ccv_cnnp_model_t* _ccv_cnnp_clamp_copy(const ccv_cnnp_model_t* const super, void* const context)
3871{
3872 ccv_cnnp_model_clamp_t* const self = (ccv_cnnp_model_clamp_t*)super;
3873 return ccv_cnnp_clamp(self->min, self->max, self->super.name);
3874}
3875
3876// MARK - Parameter Layer
3877
3878typedef struct {
3879 ccv_cnnp_model_t super;
3880 float init_bound;
3881 ccv_nnc_tensor_symbol_t weights;
3882 ccv_nnc_tensor_param_t weights_params;
3883 ccv_nnc_tensor_symbol_t output;
3884} ccv_cnnp_model_parameter_t;
3885
3886static void _ccv_cnnp_parameter_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
3887{
3888 PRINT(CCV_CLI_VERBOSE, "[cnnp_parameter_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_parameter_build] -\n"); fflush(stdout); } } while
(0)
;
3889 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 3889, __extension__ __PRETTY_FUNCTION__
); }))
;
3890 ccv_cnnp_model_parameter_t* const self = (ccv_cnnp_model_parameter_t*)super;
3891 if (!self->weights.graph)
3892 self->weights = ccv_nnc_tensor_symbol_new(graph, self->weights_params, "weights");
3893 assert(self->weights.graph == graph)((void) sizeof ((self->weights.graph == graph) ? 1 : 0), __extension__
({ if (self->weights.graph == graph) ; else __assert_fail
("self->weights.graph == graph", "ccv_cnnp_model_addons.c"
, 3893, __extension__ __PRETTY_FUNCTION__); }))
;
3894 outputs[0] = self->weights;
3895}
3896
3897static void _ccv_cnnp_parameter_init_states(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_cnnp_state_initializer_f initializer, void* const context)
3898{
3899 ccv_cnnp_model_parameter_t* const self = (ccv_cnnp_model_parameter_t*)super;
3900 if (self->init_bound > 0)
3901 initializer(context, CMD_RANDOM_UNIFORM_FORWARD(-self->init_bound, self->init_bound)ccv_nnc_cmd(CCV_NNC_RANDOM_UNIFORM_FORWARD, 0, (ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.blas={.a={-self->init_bound, self->
init_bound}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->weights);
3902 else
3903 initializer(context, CMD_SET_FORWARD(0)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={0,}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->weights);
3904}
3905
3906static void _ccv_cnnp_parameter_add_to_parameter(ccv_cnnp_model_t* const super, const ccv_cnnp_add_to_array_f add_to_array, void* const parameters, const int is_trainable)
3907{
3908 ccv_cnnp_model_parameter_t* const self = (ccv_cnnp_model_parameter_t*)super;
3909 add_to_array(parameters, self->weights, is_trainable);
3910}
3911
3912static ccv_cnnp_model_t* _ccv_cnnp_parameter_copy(const ccv_cnnp_model_t* const super, void* const context);
3913
3914static const ccv_cnnp_model_vtab_t ccv_cnnp_parameter_isa = {
3915 .build = _ccv_cnnp_parameter_build,
3916 .init_states = _ccv_cnnp_parameter_init_states,
3917 .add_to_parameter = _ccv_cnnp_parameter_add_to_parameter,
3918 .copy = _ccv_cnnp_parameter_copy,
3919};
3920
3921ccv_cnnp_model_t* ccv_cnnp_parameter(const ccv_nnc_tensor_param_t params, const float init_bound, const int is_trainable, const char* const name)
3922{
3923 ccv_cnnp_model_parameter_t* const model_parameter = (ccv_cnnp_model_parameter_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_parameter_t));
3924 model_parameter->super.isa = &ccv_cnnp_parameter_isa;
3925 model_parameter->super.input_size = 0;
3926 model_parameter->super.outputs = &model_parameter->output;
3927 model_parameter->super.output_size = 1;
3928 model_parameter->super.is_trainable = is_trainable;
3929 ccv_cnnp_model_copy_name(&model_parameter->super, name);
3930 model_parameter->weights.d = CCV_NNC_NO_TENSOR_SYMBOL;
3931 model_parameter->weights.graph = 0;
3932 model_parameter->weights_params = params;
3933 return (ccv_cnnp_model_t*)model_parameter;
3934}
3935
3936static ccv_cnnp_model_t* _ccv_cnnp_parameter_copy(const ccv_cnnp_model_t* const super, void* const context)
3937{
3938 const ccv_cnnp_model_parameter_t* const self = (const ccv_cnnp_model_parameter_t*)super;
3939 return ccv_cnnp_parameter(self->weights_params, self->init_bound, self->super.is_trainable, self->super.name);
3940}
3941
3942// MARK - Scalar Layer
3943
3944typedef struct {
3945 ccv_cnnp_model_t super;
3946 int type;
3947 int format;
3948 int datatype;
3949 float value;
3950 ccv_nnc_tensor_symbol_t output;
3951} ccv_cnnp_model_scalar_t;
3952
3953static void _ccv_cnnp_scalar_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
3954{
3955 PRINT(CCV_CLI_VERBOSE, "[cnnp_scalar_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_scalar_build] -\n"); fflush(stdout); } } while
(0)
;
3956 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 3956, __extension__ __PRETTY_FUNCTION__
); }))
;
3957 ccv_cnnp_model_scalar_t* const self = (ccv_cnnp_model_scalar_t*)super;
3958 ccv_nnc_tensor_param_t params = {
3959 .type = self->type,
3960 .format = self->format,
3961 .datatype = self->datatype,
3962 .dim = {
3963 1
3964 }
3965 };
3966 if (input_size > 0)
3967 {
3968 ccv_nnc_tensor_param_t input_params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
3969 params.type = input_params.type;
3970 params.format = input_params.format;
3971 params.datatype = input_params.datatype;
3972 }
3973 outputs[0] = ccv_nnc_tensor_symbol_new(graph, params, 0);
3974 ccv_nnc_graph_exec_symbol_new(graph, CMD_SET_FORWARD(self->value)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={self->value,}}}, 0)
, 0, 0, outputs, 1, 0);
3975}
3976
3977static ccv_cnnp_model_t* _ccv_cnnp_scalar_copy(const ccv_cnnp_model_t* const super, void* const context);
3978
3979static const ccv_cnnp_model_vtab_t ccv_cnnp_scalar_isa = {
3980 .build = _ccv_cnnp_scalar_build,
3981 .copy = _ccv_cnnp_scalar_copy,
3982};
3983
3984ccv_cnnp_model_t* ccv_cnnp_scalar(const int type, const int format, const int datatype, const float value, const char* const name)
3985{
3986 ccv_cnnp_model_scalar_t* const model_scalar = (ccv_cnnp_model_scalar_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_scalar_t));
3987 model_scalar->super.isa = &ccv_cnnp_scalar_isa;
3988 model_scalar->super.input_size = 0;
3989 model_scalar->super.outputs = &model_scalar->output;
3990 model_scalar->super.output_size = 1;
3991 ccv_cnnp_model_copy_name(&model_scalar->super, name);
3992 model_scalar->type = type;
3993 model_scalar->format = format;
3994 model_scalar->datatype = datatype;
3995 model_scalar->value = value;
3996 return (ccv_cnnp_model_t*)model_scalar;
3997}
3998
3999static ccv_cnnp_model_t* _ccv_cnnp_scalar_copy(const ccv_cnnp_model_t* const super, void* const context)
4000{
4001 const ccv_cnnp_model_scalar_t* const self = (const ccv_cnnp_model_scalar_t*)super;
4002 return ccv_cnnp_scalar(self->type, self->format, self->datatype, self->value, self->super.name);
4003}
4004
4005// MARK - Variable Layer
4006
4007typedef struct {
4008 ccv_cnnp_model_t super;
4009 ccv_nnc_tensor_param_t params;
4010 ccv_nnc_tensor_symbol_t output;
4011} ccv_cnnp_model_variable_t;
4012
4013static void _ccv_cnnp_variable_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
4014{
4015 PRINT(CCV_CLI_VERBOSE, "[cnnp_variable_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_variable_build] -\n"); fflush(stdout); } } while
(0)
;
4016 assert(input_size == 0)((void) sizeof ((input_size == 0) ? 1 : 0), __extension__ ({ if
(input_size == 0) ; else __assert_fail ("input_size == 0", "ccv_cnnp_model_addons.c"
, 4016, __extension__ __PRETTY_FUNCTION__); }))
;
4017 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 4017, __extension__ __PRETTY_FUNCTION__
); }))
;
4018 ccv_cnnp_model_variable_t* const self = (ccv_cnnp_model_variable_t*)super;
4019 outputs[0] = ccv_nnc_tensor_symbol_new(graph, self->params, 0);
4020}
4021
4022static ccv_cnnp_model_t* _ccv_cnnp_variable_copy(const ccv_cnnp_model_t* const super, void* const context);
4023
4024static const ccv_cnnp_model_vtab_t ccv_cnnp_variable_isa = {
4025 .build = _ccv_cnnp_variable_build,
4026 .copy = _ccv_cnnp_variable_copy,
4027};
4028
4029ccv_cnnp_model_t* ccv_cnnp_variable(const ccv_nnc_tensor_param_t params, const char* const name)
4030{
4031 ccv_cnnp_model_variable_t* const model_variable = (ccv_cnnp_model_variable_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_variable_t));
4032 model_variable->super.isa = &ccv_cnnp_variable_isa;
4033 model_variable->super.input_size = 0;
4034 model_variable->super.outputs = &model_variable->output;
4035 model_variable->super.output_size = 1;
4036 ccv_cnnp_model_copy_name(&model_variable->super, name);
4037 model_variable->params = params;
4038 return (ccv_cnnp_model_t*)model_variable;
4039}
4040
4041static ccv_cnnp_model_t* _ccv_cnnp_variable_copy(const ccv_cnnp_model_t* const super, void* const context)
4042{
4043 const ccv_cnnp_model_variable_t* const self = (const ccv_cnnp_model_variable_t*)super;
4044 return ccv_cnnp_variable(self->params, self->super.name);
4045}
4046
4047// MARK - Move Layer
4048
4049typedef struct {
4050 ccv_cnnp_model_t super;
4051 ccv_nnc_tensor_symbol_t output;
4052} ccv_cnnp_model_move_t;
4053
4054static void _ccv_cnnp_move_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
4055{
4056 PRINT(CCV_CLI_VERBOSE, "[cnnp_move_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_move_build] -\n"); fflush(stdout); } } while
(0)
;
4057 assert(input_size == 2)((void) sizeof ((input_size == 2) ? 1 : 0), __extension__ ({ if
(input_size == 2) ; else __assert_fail ("input_size == 2", "ccv_cnnp_model_addons.c"
, 4057, __extension__ __PRETTY_FUNCTION__); }))
;
4058 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 4058, __extension__ __PRETTY_FUNCTION__
); }))
;
4059 outputs[0] = inputs[1];
4060 ccv_nnc_graph_exec_symbol_new(graph, CMD_FORMAT_TRANSFORM_FORWARD()ccv_nnc_cmd(CCV_NNC_FORMAT_TRANSFORM_FORWARD, 0, ccv_nnc_cmd_auto
, 0)
, inputs, 1, outputs, 1, "move");
4061}
4062
4063static ccv_cnnp_model_t* _ccv_cnnp_move_copy(const ccv_cnnp_model_t* const super, void* const context);
4064
4065static const ccv_cnnp_model_vtab_t ccv_cnnp_move_isa = {
4066 .build = _ccv_cnnp_move_build,
4067 .copy = _ccv_cnnp_move_copy,
4068};
4069
4070ccv_cnnp_model_t* ccv_cnnp_move(const char* const name)
4071{
4072 ccv_cnnp_model_move_t* const model_move = (ccv_cnnp_model_move_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_move_t));
4073 model_move->super.isa = &ccv_cnnp_move_isa;
4074 model_move->super.input_size = 2;
4075 model_move->super.outputs = &model_move->output;
4076 model_move->super.output_size = 1;
4077 ccv_cnnp_model_copy_name(&model_move->super, name);
4078 return (ccv_cnnp_model_t*)model_move;
4079}
4080
4081static ccv_cnnp_model_t* _ccv_cnnp_move_copy(const ccv_cnnp_model_t* const super, void* const context)
4082{
4083 const ccv_cnnp_model_move_t* const self = (const ccv_cnnp_model_move_t*)super;
4084 return ccv_cnnp_move(self->super.name);
4085}
4086
4087// MARK - "Making" Contiguous Layer
4088
4089typedef struct {
4090 ccv_cnnp_model_t super;
4091 ccv_nnc_tensor_symbol_t output;
4092} ccv_cnnp_model_contiguous_t;
4093
4094static void _ccv_cnnp_contiguous_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
4095{
4096 PRINT(CCV_CLI_VERBOSE, "[cnnp_contiguous_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_contiguous_build] -\n"); fflush(stdout); } }
while (0)
;
4097 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 4097, __extension__ __PRETTY_FUNCTION__); }))
;
4098 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 4098, __extension__ __PRETTY_FUNCTION__
); }))
;
4099 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
4100 ccv_nnc_tensor_symbol_t to = ccv_nnc_tensor_symbol_alias_to(graph, inputs[0]);
4101 if (to.d == CCV_NNC_NO_TENSOR_SYMBOL) // If we are not reshape an alias, it is straightforward.
4102 {
4103 outputs[0] = inputs[0];
4104 return;
4105 }
4106 // Otherwise, we need to check its stride to know if it is contiguous.
4107 int old_stride[CCV_NNC_MAX_DIM_ALLOC(12)];
4108 ccv_nnc_tensor_symbol_alias_params(graph, inputs[0], 0, old_stride);
4109 // We identify permute by checking if the stride is not in descending order.
4110 // This also covered "permute" through reshape, rather than using ccv_cnnp_permute directly.
4111 if (ccv_nnc_is_tensor_stride_packed(old_stride, params.dim))
4112 {
4113 outputs[0] = inputs[0];
4114 return;
4115 }
4116 outputs[0] = ccv_nnc_tensor_symbol_new(graph, params, 0);
4117 ccv_nnc_graph_exec_symbol_t make_contiguous = ccv_nnc_graph_exec_symbol_new(graph, CMD_FORMAT_TRANSFORM_FORWARD()ccv_nnc_cmd(CCV_NNC_FORMAT_TRANSFORM_FORWARD, 0, ccv_nnc_cmd_auto
, 0)
, inputs, 1, outputs, 1, "contiguous");
4118 ccv_nnc_graph_exec_symbol_set_flags(graph, make_contiguous, CCV_NNC_GRAPH_EXEC_DISABLE_OPT);
4119}
4120
4121static ccv_cnnp_model_t* _ccv_cnnp_contiguous_copy(const ccv_cnnp_model_t* const super, void* const context);
4122
4123static const ccv_cnnp_model_vtab_t ccv_cnnp_contiguous_isa = {
4124 .build = _ccv_cnnp_contiguous_build,
4125 .copy = _ccv_cnnp_contiguous_copy,
4126};
4127
4128ccv_cnnp_model_t* ccv_cnnp_contiguous(const char* const name)
4129{
4130 ccv_cnnp_model_contiguous_t* const model_contiguous = (ccv_cnnp_model_contiguous_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_contiguous_t));
4131 model_contiguous->super.isa = &ccv_cnnp_contiguous_isa;
4132 model_contiguous->super.input_size = 1;
4133 model_contiguous->super.outputs = &model_contiguous->output;
4134 model_contiguous->super.output_size = 1;
4135 ccv_cnnp_model_copy_name(&model_contiguous->super, name);
4136 return (ccv_cnnp_model_t*)model_contiguous;
4137}
4138
4139static ccv_cnnp_model_t* _ccv_cnnp_contiguous_copy(const ccv_cnnp_model_t* const super, void* const context)
4140{
4141 const ccv_cnnp_model_contiguous_t* const self = (const ccv_cnnp_model_contiguous_t*)super;
4142 return ccv_cnnp_contiguous(self->super.name);
4143}
4144
4145// MARK - "Making" Copy Layer
4146
4147typedef struct {
4148 ccv_cnnp_model_t super;
4149 ccv_nnc_tensor_symbol_t output;
4150} ccv_cnnp_model_copy_t;
4151
4152static void _ccv_cnnp_copy_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
4153{
4154 PRINT(CCV_CLI_VERBOSE, "[cnnp_copy_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_copy_build] -\n"); fflush(stdout); } } while
(0)
;
4155 assert(input_size == 1)((void) sizeof ((input_size == 1) ? 1 : 0), __extension__ ({ if
(input_size == 1) ; else __assert_fail ("input_size == 1", "ccv_cnnp_model_addons.c"
, 4155, __extension__ __PRETTY_FUNCTION__); }))
;
4156 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 4156, __extension__ __PRETTY_FUNCTION__
); }))
;
4157 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
4158 ccv_nnc_tensor_symbol_t to = ccv_nnc_tensor_symbol_alias_to(graph, inputs[0]);
4159 if (to.d == CCV_NNC_NO_TENSOR_SYMBOL) // If we are not reshape an alias, it is straightforward.
4160 {
4161 outputs[0] = inputs[0];
4162 return;
4163 }
4164 outputs[0] = ccv_nnc_tensor_symbol_new(graph, params, 0);
4165 ccv_nnc_graph_exec_symbol_t make_contiguous = ccv_nnc_graph_exec_symbol_new(graph, CMD_FORMAT_TRANSFORM_FORWARD()ccv_nnc_cmd(CCV_NNC_FORMAT_TRANSFORM_FORWARD, 0, ccv_nnc_cmd_auto
, 0)
, inputs, 1, outputs, 1, "contiguous");
4166 ccv_nnc_graph_exec_symbol_set_flags(graph, make_contiguous, CCV_NNC_GRAPH_EXEC_DISABLE_OPT);
4167}
4168
4169static ccv_cnnp_model_t* _ccv_cnnp_copy_copy(const ccv_cnnp_model_t* const super, void* const context);
4170
4171static const ccv_cnnp_model_vtab_t ccv_cnnp_copy_isa = {
4172 .build = _ccv_cnnp_copy_build,
4173 .copy = _ccv_cnnp_copy_copy,
4174};
4175
4176ccv_cnnp_model_t* ccv_cnnp_copy(const char* const name)
4177{
4178 ccv_cnnp_model_copy_t* const model_copy = (ccv_cnnp_model_copy_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_copy_t));
4179 model_copy->super.isa = &ccv_cnnp_copy_isa;
4180 model_copy->super.input_size = 1;
4181 model_copy->super.outputs = &model_copy->output;
4182 model_copy->super.output_size = 1;
4183 ccv_cnnp_model_copy_name(&model_copy->super, name);
4184 return (ccv_cnnp_model_t*)model_copy;
4185}
4186
4187static ccv_cnnp_model_t* _ccv_cnnp_copy_copy(const ccv_cnnp_model_t* const super, void* const context)
4188{
4189 const ccv_cnnp_model_copy_t* const self = (const ccv_cnnp_model_copy_t*)super;
4190 return ccv_cnnp_copy(self->super.name);
4191}
4192
4193// MARK - Scaled-Dot Product Attention Layer
4194
4195typedef struct {
4196 ccv_cnnp_model_t super;
4197 ccv_nnc_tensor_symbol_t output;
4198 ccv_nnc_tensor_symbol_t weights;
4199 ccv_nnc_tensor_symbol_t bias;
4200 float scale;
4201 int is_causal;
4202 int has_attn_mask;
4203 int flags;
4204 int fused_unify_head_weights;
4205 int no_bias;
4206} ccv_cnnp_model_scaled_dot_product_attention_t;
4207
4208static void _ccv_cnnp_scaled_dot_product_attention_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
4209{
4210 PRINT(CCV_CLI_VERBOSE, "[cnnp_scaled_dot_product_attention_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_scaled_dot_product_attention_build] -\n"); fflush
(stdout); } } while (0)
;
4211 assert(input_size == 3 || input_size == 4)((void) sizeof ((input_size == 3 || input_size == 4) ? 1 : 0)
, __extension__ ({ if (input_size == 3 || input_size == 4) ; else
__assert_fail ("input_size == 3 || input_size == 4", "ccv_cnnp_model_addons.c"
, 4211, __extension__ __PRETTY_FUNCTION__); }))
;
4212 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 4212, __extension__ __PRETTY_FUNCTION__
); }))
;
4213 ccv_cnnp_model_scaled_dot_product_attention_t* const self = (ccv_cnnp_model_scaled_dot_product_attention_t*)super;
4214 const ccv_nnc_tensor_param_t q_params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
4215 const ccv_nnc_tensor_param_t k_params = ccv_nnc_tensor_symbol_params(graph, inputs[1]);
4216 const ccv_nnc_tensor_param_t v_params = ccv_nnc_tensor_symbol_params(graph, inputs[2]);
4217 const int v_nd = ccv_nnc_tensor_nd(v_params.dim);
4218 assert(v_nd == 3 || v_nd == 4)((void) sizeof ((v_nd == 3 || v_nd == 4) ? 1 : 0), __extension__
({ if (v_nd == 3 || v_nd == 4) ; else __assert_fail ("v_nd == 3 || v_nd == 4"
, "ccv_cnnp_model_addons.c", 4218, __extension__ __PRETTY_FUNCTION__
); }))
;
4219 const int hEv = (v_nd == 3 ? 1 : v_params.dim[2]) * v_params.dim[v_nd - 1];
4220 ccv_nnc_tensor_param_t weights_params = q_params;
4221 memset(weights_params.dim, 0, sizeof(weights_params.dim));
4222 weights_params.dim[0] = hEv;
4223 weights_params.dim[1] = hEv;
4224 ccv_nnc_tensor_param_t bias_params = q_params;
4225 memset(bias_params.dim, 0, sizeof(bias_params.dim));
4226 bias_params.dim[0] = hEv;
4227 ccv_nnc_cmd_t cmd = {0};
4228 cmd.cmd = CCV_NNC_SCALED_DOT_PRODUCT_ATTENTION_FORWARD;
4229 cmd.info.scaled_dot_product_attention.scale = self->scale;
4230 cmd.info.scaled_dot_product_attention.is_causal = self->is_causal;
4231 cmd.info.scaled_dot_product_attention.flags = self->flags;
4232 ccv_nnc_tensor_param_t output_params[3];
4233 ccv_nnc_tensor_symbol_t output;
4234 ccv_nnc_tensor_symbol_t saved_softmax_lse;
4235 ccv_nnc_tensor_symbol_t saved_v_proj = NO_TENSOR_SYMBOL(const ccv_nnc_tensor_symbol_t){.d = CCV_NNC_NO_TENSOR_SYMBOL
}
;
4236 ccv_nnc_tensor_symbol_t attn_mask = NO_TENSOR_SYMBOL(const ccv_nnc_tensor_symbol_t){.d = CCV_NNC_NO_TENSOR_SYMBOL
}
;
4237 ccv_nnc_tensor_symbol_t weights = NO_TENSOR_SYMBOL(const ccv_nnc_tensor_symbol_t){.d = CCV_NNC_NO_TENSOR_SYMBOL
}
;
4238 ccv_nnc_tensor_symbol_t bias = NO_TENSOR_SYMBOL(const ccv_nnc_tensor_symbol_t){.d = CCV_NNC_NO_TENSOR_SYMBOL
}
;
4239 if (self->has_attn_mask)
4240 attn_mask = inputs[3];
4241 if (self->fused_unify_head_weights)
4242 {
4243 if (!self->weights.graph)
4244 self->weights = ccv_nnc_tensor_symbol_new(graph, weights_params, "weights");
4245 weights = self->weights;
4246 if (!self->no_bias)
4247 {
4248 if (!self->bias.graph)
4249 self->bias = ccv_nnc_tensor_symbol_new(graph, bias_params, "bias");
4250 bias = self->bias;
4251 }
4252 ccv_nnc_hint_tensor_auto(cmd, (ccv_nnc_tensor_param_t []){
4253 q_params,
4254 k_params,
4255 v_params,
4256 (ccv_nnc_tensor_param_t){},
4257 weights_params,
4258 bias_params,
4259 }, 6, ccv_nnc_no_hint, output_params, 3);
4260 output = ccv_nnc_tensor_symbol_new(graph, output_params[0], 0);
4261 saved_softmax_lse = ccv_nnc_tensor_symbol_new(graph, output_params[1], 0);
4262 saved_v_proj = ccv_nnc_tensor_symbol_new(graph, output_params[2], 0);
4263 } else {
4264 ccv_nnc_hint_tensor_auto(cmd, (ccv_nnc_tensor_param_t []){
4265 q_params,
4266 k_params,
4267 v_params,
4268 }, 3, ccv_nnc_no_hint, output_params, 2);
4269 output = ccv_nnc_tensor_symbol_new(graph, output_params[0], 0);
4270 saved_softmax_lse = ccv_nnc_tensor_symbol_new(graph, output_params[1], 0);
4271 }
4272 ccv_nnc_graph_exec_symbol_new(graph, cmd, TENSOR_SYMBOL_LIST(inputs[0], inputs[1], inputs[2], attn_mask, weights, bias)(const ccv_nnc_tensor_symbol_t []){inputs[0], inputs[1], inputs
[2], attn_mask, weights, bias}, (1 +1 +1 +1 +1 +1 +1 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output, saved_softmax_lse, saved_v_proj)(const ccv_nnc_tensor_symbol_t []){output, saved_softmax_lse,
saved_v_proj}, (1 +1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 -1)
, "scaled_dot_product_attention");
4273 outputs[0] = output;
4274}
4275
4276static void _ccv_cnnp_scaled_dot_product_attention_init_states(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_cnnp_state_initializer_f initializer, void* const context)
4277{
4278 ccv_cnnp_model_scaled_dot_product_attention_t* const self = (ccv_cnnp_model_scaled_dot_product_attention_t*)super;
4279 if (self->weights.graph)
4280 {
4281 assert(self->fused_unify_head_weights)((void) sizeof ((self->fused_unify_head_weights) ? 1 : 0),
__extension__ ({ if (self->fused_unify_head_weights) ; else
__assert_fail ("self->fused_unify_head_weights", "ccv_cnnp_model_addons.c"
, 4281, __extension__ __PRETTY_FUNCTION__); }))
;
4282 const ccv_nnc_tensor_param_t weight_params = ccv_nnc_tensor_symbol_params(graph, self->weights);
4283 const int c = weight_params.dim[1];
4284 const float std = sqrtf(2) / sqrtf(c);
4285 const float bound = sqrtf(3) * std;
4286 initializer(context, CMD_RANDOM_UNIFORM_FORWARD(-bound, bound)ccv_nnc_cmd(CCV_NNC_RANDOM_UNIFORM_FORWARD, 0, (ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.blas={.a={-bound, bound}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->weights);
4287 if (self->bias.graph)
4288 initializer(context, CMD_SET_FORWARD(0)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={0,}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->bias);
4289 }
4290}
4291
4292static void _ccv_cnnp_scaled_dot_product_attention_add_to_parameter(ccv_cnnp_model_t* const super, const ccv_cnnp_add_to_array_f add_to_array, void* const parameters, const int is_trainable)
4293{
4294 ccv_cnnp_model_scaled_dot_product_attention_t* const self = (ccv_cnnp_model_scaled_dot_product_attention_t*)super;
4295 if (self->weights.graph)
4296 {
4297 assert(self->fused_unify_head_weights)((void) sizeof ((self->fused_unify_head_weights) ? 1 : 0),
__extension__ ({ if (self->fused_unify_head_weights) ; else
__assert_fail ("self->fused_unify_head_weights", "ccv_cnnp_model_addons.c"
, 4297, __extension__ __PRETTY_FUNCTION__); }))
;
4298 add_to_array(parameters, self->weights, is_trainable);
4299 if (self->bias.graph)
4300 add_to_array(parameters, self->bias, is_trainable);
4301 }
4302}
4303
4304static ccv_cnnp_model_t* _ccv_cnnp_scaled_dot_product_attention_copy(const ccv_cnnp_model_t* const super, void* const context);
4305
4306static const ccv_cnnp_model_vtab_t ccv_cnnp_scaled_dot_product_attention_isa = {
4307 .build = _ccv_cnnp_scaled_dot_product_attention_build,
4308 .copy = _ccv_cnnp_scaled_dot_product_attention_copy,
4309};
4310
4311static const ccv_cnnp_model_vtab_t ccv_cnnp_scaled_dot_product_attention_fused_isa = {
4312 .build = _ccv_cnnp_scaled_dot_product_attention_build,
4313 .init_states = _ccv_cnnp_scaled_dot_product_attention_init_states,
4314 .add_to_parameter = _ccv_cnnp_scaled_dot_product_attention_add_to_parameter,
4315 .copy = _ccv_cnnp_scaled_dot_product_attention_copy,
4316};
4317
4318ccv_cnnp_model_t* ccv_cnnp_scaled_dot_product_attention(const float scale, const int is_causal, const int has_attn_mask, const int flags, const int fused_unify_head_weights, const int no_bias, const int is_trainable, const char* const name)
4319{
4320 ccv_cnnp_model_scaled_dot_product_attention_t* const model_scaled_dot_product_attention = (ccv_cnnp_model_scaled_dot_product_attention_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_scaled_dot_product_attention_t));
4321 model_scaled_dot_product_attention->super.isa = fused_unify_head_weights ? &ccv_cnnp_scaled_dot_product_attention_fused_isa : &ccv_cnnp_scaled_dot_product_attention_isa;
4322 model_scaled_dot_product_attention->super.input_size = has_attn_mask ? 4 : 3;
4323 model_scaled_dot_product_attention->super.outputs = &model_scaled_dot_product_attention->output;
4324 model_scaled_dot_product_attention->super.output_size = 1;
4325 model_scaled_dot_product_attention->super.is_trainable = is_trainable;
4326 ccv_cnnp_model_copy_name(&model_scaled_dot_product_attention->super, name);
4327 model_scaled_dot_product_attention->weights.d = CCV_NNC_NO_TENSOR_SYMBOL;
4328 model_scaled_dot_product_attention->weights.graph = 0;
4329 model_scaled_dot_product_attention->bias.d = CCV_NNC_NO_TENSOR_SYMBOL;
4330 model_scaled_dot_product_attention->bias.graph = 0;
4331 model_scaled_dot_product_attention->scale = scale;
4332 model_scaled_dot_product_attention->is_causal = is_causal;
4333 model_scaled_dot_product_attention->has_attn_mask = has_attn_mask;
4334 model_scaled_dot_product_attention->flags = flags;
4335 model_scaled_dot_product_attention->fused_unify_head_weights = fused_unify_head_weights;
4336 model_scaled_dot_product_attention->no_bias = no_bias;
4337 return (ccv_cnnp_model_t*)model_scaled_dot_product_attention;
4338}
4339
4340static ccv_cnnp_model_t* _ccv_cnnp_scaled_dot_product_attention_copy(const ccv_cnnp_model_t* const super, void* const context)
4341{
4342 const ccv_cnnp_model_scaled_dot_product_attention_t* const self = (const ccv_cnnp_model_scaled_dot_product_attention_t*)super;
4343 return ccv_cnnp_scaled_dot_product_attention(self->scale, self->is_causal, self->has_attn_mask, self->flags, self->fused_unify_head_weights, self->no_bias, self->super.is_trainable, self->super.name);
4344}
4345
4346// MARK - Debug Layer
4347
4348typedef struct {
4349 ccv_cnnp_model_t super;
4350 ccv_nnc_tensor_symbol_t output;
4351 ccv_cnnp_model_debug_f debugger;
4352 ccv_cnnp_model_debug_context_deinit_f debug_deinit;
4353 ccv_cnnp_model_debug_context_copy_f debug_copy;
4354 void* debug_context;
4355} ccv_cnnp_model_debug_t;
4356
4357static int _ccv_cnnp_debug_exec(const ccv_nnc_cmd_t cmd, const ccv_nnc_hint_t hint, const int flags, ccv_nnc_tensor_t* const* const inputs, const int input_size, ccv_nnc_tensor_t* const* const outputs, const int output_size, ccv_nnc_stream_context_t* const stream_context)
4358{
4359 if (cmd.cmd == CCV_NNC_CUSTOM_BACKWARD)
4360 {
4361 assert(0 && "don't support debug backward pass yet")((void) sizeof ((0 && "don't support debug backward pass yet"
) ? 1 : 0), __extension__ ({ if (0 && "don't support debug backward pass yet"
) ; else __assert_fail ("0 && \"don't support debug backward pass yet\""
, "ccv_cnnp_model_addons.c", 4361, __extension__ __PRETTY_FUNCTION__
); }))
;
4362 }
4363 ccv_cnnp_model_debug_t* const self = (ccv_cnnp_model_debug_t*)cmd.data;
4364 self->debugger(inputs, input_size, stream_context, self->debug_context);
4365 return CCV_NNC_EXEC_SUCCESS;
4366}
4367
4368static ccv_nnc_cmd_vtab_t ccv_cnnp_debug_exec_isa = {
4369 .exec = _ccv_cnnp_debug_exec
4370};
4371
4372static void _ccv_cnnp_debug_build(ccv_cnnp_model_t* const self, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
4373{
4374 PRINT(CCV_CLI_VERBOSE, "[cnnp_debug_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_debug_build] -\n"); fflush(stdout); } } while
(0)
;
4375 assert(input_size >= 1)((void) sizeof ((input_size >= 1) ? 1 : 0), __extension__ (
{ if (input_size >= 1) ; else __assert_fail ("input_size >= 1"
, "ccv_cnnp_model_addons.c", 4375, __extension__ __PRETTY_FUNCTION__
); }))
;
4376 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 4376, __extension__ __PRETTY_FUNCTION__
); }))
;
4377 ccv_nnc_tensor_symbol_t to = ccv_nnc_tensor_symbol_alias_to(graph, inputs[0]);
4378 ccv_nnc_tensor_param_t output_params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
4379 if (to.d == CCV_NNC_NO_TENSOR_SYMBOL) // If we are not reshape an alias, it is straightforward.
4380 {
4381 int ofs[CCV_NNC_MAX_DIM_ALLOC(12)] = {0};
4382 int stride[CCV_NNC_MAX_DIM_ALLOC(12)];
4383 ccv_nnc_tensor_get_stride(output_params.dim, stride);
4384 outputs[0] = ccv_nnc_tensor_symbol_alias_new(graph, inputs[0], ofs, stride, output_params, 0);
4385 } else {
4386 int old_ofs[CCV_NNC_MAX_DIM_ALLOC(12)];
4387 int old_stride[CCV_NNC_MAX_DIM_ALLOC(12)];
4388 ccv_nnc_tensor_symbol_alias_params(graph, inputs[0], old_ofs, old_stride);
4389 outputs[0] = ccv_nnc_tensor_symbol_alias_new(graph, to, old_ofs, old_stride, output_params, 0);
4390 }
4391 ccv_nnc_cmd_t cmd = ccv_nnc_cmd(CCV_NNC_CUSTOM_FORWARD, (ccv_nnc_cmd_vtab_t*)&ccv_cnnp_debug_exec_isa, (ccv_nnc_cmd_param_t){}, 0);
4392 cmd.data = self;
4393 ccv_nnc_graph_exec_symbol_t make_debug = ccv_nnc_graph_exec_symbol_new(graph, cmd, inputs, input_size, outputs, 1, "debug");
4394 // Disable any optimizations.
4395 ccv_nnc_graph_exec_symbol_set_flags(graph, make_debug, CCV_NNC_GRAPH_EXEC_DISABLE_OPT);
4396}
4397
4398static void _ccv_cnnp_debug_deinit(ccv_cnnp_model_t* const super)
4399{
4400 const ccv_cnnp_model_debug_t* const self = (const ccv_cnnp_model_debug_t*)super;
4401 if (self->debug_deinit && self->debug_context)
4402 self->debug_deinit(self->debug_context);
4403}
4404
4405static ccv_cnnp_model_t* _ccv_cnnp_debug_copy(const ccv_cnnp_model_t* const super, void* const context);
4406
4407static const ccv_cnnp_model_vtab_t ccv_cnnp_debug_isa = {
4408 .build = _ccv_cnnp_debug_build,
4409 .deinit = _ccv_cnnp_debug_deinit,
4410 .copy = _ccv_cnnp_debug_copy,
4411};
4412
4413ccv_cnnp_model_t* ccv_cnnp_debug(ccv_cnnp_model_debug_f func, void* const context, ccv_cnnp_model_debug_context_deinit_f deinit, ccv_cnnp_model_debug_context_copy_f copy, const char* const name)
4414{
4415 ccv_cnnp_model_debug_t* const model_debug = (ccv_cnnp_model_debug_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_debug_t));
4416 model_debug->super.isa = &ccv_cnnp_debug_isa;
4417 model_debug->super.input_size = 0;
4418 model_debug->super.outputs = &model_debug->output;
4419 model_debug->super.output_size = 1;
4420 model_debug->debugger = func;
4421 model_debug->debug_context = context;
4422 model_debug->debug_deinit = deinit;
4423 model_debug->debug_copy = copy;
4424 ccv_cnnp_model_copy_name(&model_debug->super, name);
4425 return (ccv_cnnp_model_t*)model_debug;
4426}
4427
4428static ccv_cnnp_model_t* _ccv_cnnp_debug_copy(const ccv_cnnp_model_t* const super, void* const context)
4429{
4430 const ccv_cnnp_model_debug_t* const self = (const ccv_cnnp_model_debug_t*)super;
4431 void* debug_context = self->debug_context;
4432 if (self->debug_copy && self->debug_context)
4433 debug_context = self->debug_copy(self->debug_context);
4434 return ccv_cnnp_debug(self->debugger, debug_context, self->debug_deinit, self->debug_copy, self->super.name);
4435}
4436
4437/// MARK - Sort layer.
4438
4439typedef struct {
4440 ccv_cnnp_model_t super;
4441 ccv_nnc_tensor_symbol_t outputs[2];
4442 int along_axis;
4443 int descending;
4444} ccv_cnnp_model_sort_t;
4445
4446static void _ccv_cnnp_sort_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
4447{
4448 ccv_cnnp_model_sort_t* const self = (ccv_cnnp_model_sort_t*)super;
4449 PRINT(CCV_CLI_VERBOSE, "[cnnp_sort_build] - along_axis: %d, descending: %d\n", self->along_axis, self->descending)do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_sort_build] - along_axis: %d, descending: %d\n"
, self->along_axis, self->descending); fflush(stdout); }
} while (0)
;
4450 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
4451 assert(output_size == 2)((void) sizeof ((output_size == 2) ? 1 : 0), __extension__ ({
if (output_size == 2) ; else __assert_fail ("output_size == 2"
, "ccv_cnnp_model_addons.c", 4451, __extension__ __PRETTY_FUNCTION__
); }))
;
4452 outputs[0] = ccv_nnc_tensor_symbol_new(graph, params, 0);
4453 params.datatype = CCV_32S;
4454 outputs[1] = ccv_nnc_tensor_symbol_new(graph, params, 0);
4455 ccv_nnc_graph_exec_symbol_new(graph, CMD_SORT_FORWARD(self->along_axis, self->descending)ccv_nnc_cmd(CCV_NNC_SORT_FORWARD, 0, ((ccv_nnc_cmd_param_t){.
size={.dim={1,1,1}},.sort={.along_axis=self->along_axis,.descending
=self->descending}}), 0)
, inputs, input_size, outputs, output_size, "sort");
4456}
4457
4458static ccv_cnnp_model_t* _ccv_cnnp_sort_copy(const ccv_cnnp_model_t* const self, void* const context);
4459
4460static const ccv_cnnp_model_vtab_t ccv_cnnp_sort_isa = {
4461 .build = _ccv_cnnp_sort_build,
4462 .copy = _ccv_cnnp_sort_copy,
4463};
4464
4465ccv_cnnp_model_t* ccv_cnnp_sort(const int along_axis, const int descending, const char* const name)
4466{
4467 ccv_cnnp_model_sort_t* const model_sort = (ccv_cnnp_model_sort_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_sort_t));
4468 model_sort->super.isa = &ccv_cnnp_sort_isa;
4469 model_sort->super.input_size = 0;
4470 model_sort->super.outputs = model_sort->outputs;
4471 model_sort->super.output_size = 2;
4472 model_sort->along_axis = along_axis;
4473 model_sort->descending = descending;
4474 ccv_cnnp_model_copy_name(&model_sort->super, name);
4475 return (ccv_cnnp_model_t*)model_sort;
4476}
4477
4478static ccv_cnnp_model_t* _ccv_cnnp_sort_copy(const ccv_cnnp_model_t* const super, void* const context)
4479{
4480 ccv_cnnp_model_sort_t* const self = (ccv_cnnp_model_sort_t*)super;
4481 return ccv_cnnp_sort(self->along_axis, self->descending, self->super.name);
4482}
4483
4484/// MARK - Partition layer.
4485
4486typedef struct {
4487 ccv_cnnp_model_t super;
4488 ccv_nnc_tensor_symbol_t outputs[2];
4489 int kth;
4490 int along_axis;
4491 int descending;
4492} ccv_cnnp_model_partition_t;
4493
4494static void _ccv_cnnp_partition_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
4495{
4496 ccv_cnnp_model_partition_t* const self = (ccv_cnnp_model_partition_t*)super;
4497 PRINT(CCV_CLI_VERBOSE, "[cnnp_partition_build] - kth: %d, along_axis: %d, descending: %d\n", self->kth, self->along_axis, self->descending)do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_partition_build] - kth: %d, along_axis: %d, descending: %d\n"
, self->kth, self->along_axis, self->descending); fflush
(stdout); } } while (0)
;
4498 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
4499 assert(output_size == 2)((void) sizeof ((output_size == 2) ? 1 : 0), __extension__ ({
if (output_size == 2) ; else __assert_fail ("output_size == 2"
, "ccv_cnnp_model_addons.c", 4499, __extension__ __PRETTY_FUNCTION__
); }))
;
4500 if (self->kth > 0)
4501 params.dim[self->along_axis] = self->kth;
4502 outputs[0] = ccv_nnc_tensor_symbol_new(graph, params, 0);
4503 params.datatype = CCV_32S;
4504 outputs[1] = ccv_nnc_tensor_symbol_new(graph, params, 0);
4505 ccv_nnc_graph_exec_symbol_new(graph, CMD_PARTITION_FORWARD(self->kth, self->along_axis, self->descending)ccv_nnc_cmd(CCV_NNC_PARTITION_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.partition={.kth=self->kth,.along_axis
=self->along_axis,.descending=self->descending}}), 0)
, inputs, input_size, outputs, output_size, "partition");
4506}
4507
4508static ccv_cnnp_model_t* _ccv_cnnp_partition_copy(const ccv_cnnp_model_t* const self, void* const context);
4509
4510static const ccv_cnnp_model_vtab_t ccv_cnnp_partition_isa = {
4511 .build = _ccv_cnnp_partition_build,
4512 .copy = _ccv_cnnp_partition_copy,
4513};
4514
4515ccv_cnnp_model_t* ccv_cnnp_partition(const int kth, const int along_axis, const int descending, const char* const name)
4516{
4517 ccv_cnnp_model_partition_t* const model_partition = (ccv_cnnp_model_partition_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_partition_t));
4518 model_partition->super.isa = &ccv_cnnp_partition_isa;
4519 model_partition->super.input_size = 0;
4520 model_partition->super.outputs = model_partition->outputs;
4521 model_partition->super.output_size = 2;
4522 model_partition->kth = kth;
4523 model_partition->along_axis = along_axis;
4524 model_partition->descending = descending;
4525 ccv_cnnp_model_copy_name(&model_partition->super, name);
4526 return (ccv_cnnp_model_t*)model_partition;
4527}
4528
4529static ccv_cnnp_model_t* _ccv_cnnp_partition_copy(const ccv_cnnp_model_t* const super, void* const context)
4530{
4531 ccv_cnnp_model_partition_t* const self = (ccv_cnnp_model_partition_t*)super;
4532 return ccv_cnnp_partition(self->kth, self->along_axis, self->descending, self->super.name);
4533}
4534
4535/// MARK - Unique consecutive layer.
4536
4537typedef struct {
4538 ccv_cnnp_model_t super;
4539 ccv_nnc_tensor_symbol_t outputs[2];
4540 int bincount;
4541} ccv_cnnp_model_unique_consecutive_t;
4542
4543static void _ccv_cnnp_unique_consecutive_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
4544{
4545 ccv_cnnp_model_unique_consecutive_t* const self = (ccv_cnnp_model_unique_consecutive_t*)super;
4546 PRINT(CCV_CLI_VERBOSE, "[cnnp_unique_consecutive_build] - bincount: %d\n", self->bincount)do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_unique_consecutive_build] - bincount: %d\n",
self->bincount); fflush(stdout); } } while (0)
;
4547 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
4548 assert(output_size == 2)((void) sizeof ((output_size == 2) ? 1 : 0), __extension__ ({
if (output_size == 2) ; else __assert_fail ("output_size == 2"
, "ccv_cnnp_model_addons.c", 4548, __extension__ __PRETTY_FUNCTION__
); }))
;
4549 if (self->bincount > 0)
4550 params.dim[0] = ccv_min(params.dim[0], self->bincount)({ typeof (params.dim[0]) _a = (params.dim[0]); typeof (self->
bincount) _b = (self->bincount); (_a < _b) ? _a : _b; }
)
;
4551 outputs[0] = ccv_nnc_tensor_symbol_new(graph, params, 0);
4552 params.datatype = CCV_32S;
4553 outputs[1] = ccv_nnc_tensor_symbol_new(graph, params, 0);
4554 ccv_nnc_graph_exec_symbol_new(graph, CMD_UNIQUE_CONSECUTIVE_FORWARD(self->bincount)ccv_nnc_cmd(CCV_NNC_UNIQUE_CONSECUTIVE_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.unique_consecutive={.bincount=self->
bincount}}), 0)
, inputs, input_size, outputs, output_size, "unique_consecutive");
4555}
4556
4557static ccv_cnnp_model_t* _ccv_cnnp_unique_consecutive_copy(const ccv_cnnp_model_t* const self, void* const context);
4558
4559static const ccv_cnnp_model_vtab_t ccv_cnnp_unique_consecutive_isa = {
4560 .build = _ccv_cnnp_unique_consecutive_build,
4561 .copy = _ccv_cnnp_unique_consecutive_copy,
4562};
4563
4564ccv_cnnp_model_t* ccv_cnnp_unique_consecutive(const int bincount, const char* const name)
4565{
4566 ccv_cnnp_model_unique_consecutive_t* const model_unique_consecutive = (ccv_cnnp_model_unique_consecutive_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_unique_consecutive_t));
4567 model_unique_consecutive->super.isa = &ccv_cnnp_unique_consecutive_isa;
4568 model_unique_consecutive->super.input_size = 0;
4569 model_unique_consecutive->super.outputs = model_unique_consecutive->outputs;
4570 model_unique_consecutive->super.output_size = 2;
4571 model_unique_consecutive->bincount = bincount;
4572 ccv_cnnp_model_copy_name(&model_unique_consecutive->super, name);
4573 return (ccv_cnnp_model_t*)model_unique_consecutive;
4574}
4575
4576static ccv_cnnp_model_t* _ccv_cnnp_unique_consecutive_copy(const ccv_cnnp_model_t* const super, void* const context)
4577{
4578 ccv_cnnp_model_unique_consecutive_t* const self = (ccv_cnnp_model_unique_consecutive_t*)super;
4579 return ccv_cnnp_unique_consecutive(self->bincount, self->super.name);
4580}
4581
4582/// MARK - Scatter add layer.
4583
4584typedef struct {
4585 ccv_cnnp_model_t super;
4586 ccv_nnc_tensor_symbol_t output;
4587 int bincount;
4588} ccv_cnnp_model_scatter_add_t;
4589
4590static void _ccv_cnnp_scatter_add_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
4591{
4592 ccv_cnnp_model_scatter_add_t* const self = (ccv_cnnp_model_scatter_add_t*)super;
4593 PRINT(CCV_CLI_VERBOSE, "[cnnp_scatter_add_build] - bincount: %d\n", self->bincount)do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_scatter_add_build] - bincount: %d\n", self->
bincount); fflush(stdout); } } while (0)
;
4594 ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
4595 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 4595, __extension__ __PRETTY_FUNCTION__
); }))
;
4596 assert(self->bincount > 0)((void) sizeof ((self->bincount > 0) ? 1 : 0), __extension__
({ if (self->bincount > 0) ; else __assert_fail ("self->bincount > 0"
, "ccv_cnnp_model_addons.c", 4596, __extension__ __PRETTY_FUNCTION__
); }))
;
4597 params.dim[0] = self->bincount;
4598 outputs[0] = ccv_nnc_tensor_symbol_new(graph, params, 0);
4599 ccv_nnc_graph_exec_symbol_new(graph, CMD_SCATTER_ADD_FORWARD(self->bincount)ccv_nnc_cmd(CCV_NNC_SCATTER_ADD_FORWARD, 0, ((ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.scatter_add={.bincount=self->bincount
}}), 0)
, inputs, input_size, outputs, output_size, "scatter_add");
4600}
4601
4602static ccv_cnnp_model_t* _ccv_cnnp_scatter_add_copy(const ccv_cnnp_model_t* const self, void* const context);
4603
4604static const ccv_cnnp_model_vtab_t ccv_cnnp_scatter_add_isa = {
4605 .build = _ccv_cnnp_scatter_add_build,
4606 .copy = _ccv_cnnp_scatter_add_copy,
4607};
4608
4609ccv_cnnp_model_t* ccv_cnnp_scatter_add(const int bincount, const char* const name)
4610{
4611 assert(bincount > 0)((void) sizeof ((bincount > 0) ? 1 : 0), __extension__ ({ if
(bincount > 0) ; else __assert_fail ("bincount > 0", "ccv_cnnp_model_addons.c"
, 4611, __extension__ __PRETTY_FUNCTION__); }))
;
4612 ccv_cnnp_model_scatter_add_t* const model_scatter_add = (ccv_cnnp_model_scatter_add_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_scatter_add_t));
4613 model_scatter_add->super.isa = &ccv_cnnp_scatter_add_isa;
4614 model_scatter_add->super.input_size = 0;
4615 model_scatter_add->super.outputs = &model_scatter_add->output;
4616 model_scatter_add->super.output_size = 1;
4617 model_scatter_add->bincount = bincount;
4618 ccv_cnnp_model_copy_name(&model_scatter_add->super, name);
4619 return (ccv_cnnp_model_t*)model_scatter_add;
4620}
4621
4622static ccv_cnnp_model_t* _ccv_cnnp_scatter_add_copy(const ccv_cnnp_model_t* const super, void* const context)
4623{
4624 ccv_cnnp_model_scatter_add_t* const self = (ccv_cnnp_model_scatter_add_t*)super;
4625 return ccv_cnnp_scatter_add(self->bincount, self->super.name);
4626}
4627
4628// MARK - Segmented Dense Layer
4629
4630typedef struct {
4631 ccv_cnnp_model_t super;
4632 ccv_nnc_tensor_symbol_t output;
4633 ccv_nnc_tensor_symbol_t weights;
4634 ccv_nnc_tensor_symbol_t bias;
4635 int segments;
4636 int count;
4637 int no_bias;
4638 int flags;
4639} ccv_cnnp_model_segmented_dense_t;
4640
4641static void _ccv_cnnp_segmented_dense_build(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_nnc_tensor_symbol_t* const inputs, const int input_size, ccv_nnc_tensor_symbol_t* const outputs, const int output_size)
4642{
4643 ccv_cnnp_model_segmented_dense_t* const self = (ccv_cnnp_model_segmented_dense_t*)super;
4644 PRINT(CCV_CLI_VERBOSE, "[cnnp_segmented_dense_build] -\n")do { if ((CCV_CLI_VERBOSE & ccv_cli_get_output_levels()))
{ printf("[cnnp_segmented_dense_build] -\n"); fflush(stdout)
; } } while (0)
;
4645 assert(input_size == 3)((void) sizeof ((input_size == 3) ? 1 : 0), __extension__ ({ if
(input_size == 3) ; else __assert_fail ("input_size == 3", "ccv_cnnp_model_addons.c"
, 4645, __extension__ __PRETTY_FUNCTION__); }))
;
4646 assert(output_size == 1)((void) sizeof ((output_size == 1) ? 1 : 0), __extension__ ({
if (output_size == 1) ; else __assert_fail ("output_size == 1"
, "ccv_cnnp_model_addons.c", 4646, __extension__ __PRETTY_FUNCTION__
); }))
;
4647 const ccv_nnc_tensor_param_t params = ccv_nnc_tensor_symbol_params(graph, inputs[0]);
4648 const ccv_nnc_tensor_param_t indices_params = ccv_nnc_tensor_symbol_params(graph, inputs[1]);
4649 const ccv_nnc_tensor_param_t counts_params = ccv_nnc_tensor_symbol_params(graph, inputs[2]);
4650 ccv_nnc_tensor_param_t weights_params = params;
4651 memset(weights_params.dim, 0, sizeof(weights_params.dim));
4652 weights_params.dim[0] = self->segments;
4653 weights_params.dim[1] = self->count;
4654 weights_params.dim[2] = params.dim[ccv_nnc_tensor_nd(params.dim) - 1];
4655 if (!self->weights.graph)
4656 self->weights = ccv_nnc_tensor_symbol_new(graph, weights_params, "weights");
4657 assert(self->weights.graph == graph)((void) sizeof ((self->weights.graph == graph) ? 1 : 0), __extension__
({ if (self->weights.graph == graph) ; else __assert_fail
("self->weights.graph == graph", "ccv_cnnp_model_addons.c"
, 4657, __extension__ __PRETTY_FUNCTION__); }))
;
4658 ccv_nnc_tensor_param_t bias_params = params;
4659 memset(bias_params.dim, 0, sizeof(bias_params.dim));
4660 bias_params.dim[0] = self->segments;
4661 bias_params.dim[1] = self->count;
4662 ccv_nnc_cmd_t cmd = {0};
4663 cmd.cmd = CCV_NNC_SEGMENTED_GEMM_FORWARD;
4664 cmd.info.blas.a[0] = 1;
4665 cmd.info.blas.a[1] = 1;
4666 cmd.info.blas.transpose_b[0] = 1;
4667 cmd.info.blas.transpose_b[1] = 2;
4668 cmd.info.blas.flags = self->flags;
4669 ccv_nnc_tensor_param_t output_params;
4670 ccv_nnc_hint_tensor_auto(cmd, (ccv_nnc_tensor_param_t []){
4671 params, indices_params, counts_params,
4672 weights_params,
4673 bias_params,
4674 }, 5, ccv_nnc_no_hint, &output_params, 1);
4675 const ccv_nnc_tensor_symbol_t output = ccv_nnc_tensor_symbol_new(graph, output_params, 0);
4676 if (self->no_bias)
4677 ccv_nnc_graph_exec_symbol_new(graph, cmd, TENSOR_SYMBOL_LIST(inputs[0], inputs[1], inputs[2], self->weights)(const ccv_nnc_tensor_symbol_t []){inputs[0], inputs[1], inputs
[2], self->weights}, (1 +1 +1 +1 +1 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output)(const ccv_nnc_tensor_symbol_t []){output}, (1 +1 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "segmented_dense");
4678 else {
4679 if (!self->bias.graph)
4680 self->bias = ccv_nnc_tensor_symbol_new(graph, bias_params, "bias");
4681 ccv_nnc_graph_exec_symbol_new(graph, cmd, TENSOR_SYMBOL_LIST(inputs[0], inputs[1], inputs[2], self->weights, self->bias)(const ccv_nnc_tensor_symbol_t []){inputs[0], inputs[1], inputs
[2], self->weights, self->bias}, (1 +1 +1 +1 +1 +1 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, TENSOR_SYMBOL_LIST(output)(const ccv_nnc_tensor_symbol_t []){output}, (1 +1 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +
0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0
+0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 +0 -1)
, "segmented_dense");
4682 }
4683 outputs[0] = output;
4684}
4685
4686static void _ccv_cnnp_segmented_dense_init_states(ccv_cnnp_model_t* const super, ccv_nnc_symbolic_graph_t* const graph, const ccv_cnnp_state_initializer_f initializer, void* const context)
4687{
4688 ccv_cnnp_model_segmented_dense_t* const self = (ccv_cnnp_model_segmented_dense_t*)super;
4689 const ccv_nnc_tensor_param_t weight_params = ccv_nnc_tensor_symbol_params(graph, self->weights);
4690 const int c = weight_params.dim[1];
4691 const float std = sqrtf(2) / sqrtf(c);
4692 const float bound = sqrtf(3) * std;
4693 initializer(context, CMD_RANDOM_UNIFORM_FORWARD(-bound, bound)ccv_nnc_cmd(CCV_NNC_RANDOM_UNIFORM_FORWARD, 0, (ccv_nnc_cmd_param_t
){.size={.dim={1,1,1}},.blas={.a={-bound, bound}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->weights);
4694 if (self->bias.graph)
4695 initializer(context, CMD_SET_FORWARD(0)ccv_nnc_cmd(CCV_NNC_SET_FORWARD, 0, (ccv_nnc_cmd_param_t){.size
={.dim={1,1,1}},.blas={.a={0,}}}, 0)
, ccv_nnc_no_hint, 0, 0, self->bias);
4696}
4697
4698static void _ccv_cnnp_segmented_dense_add_to_parameter(ccv_cnnp_model_t* const super, const ccv_cnnp_add_to_array_f add_to_array, void* const parameters, const int is_trainable)
4699{
4700 ccv_cnnp_model_segmented_dense_t* const self = (ccv_cnnp_model_segmented_dense_t*)super;
4701 add_to_array(parameters, self->weights, is_trainable);
4702 if (self->bias.graph)
4703 add_to_array(parameters, self->bias, is_trainable);
4704}
4705
4706static ccv_cnnp_model_t* _ccv_cnnp_segmented_dense_copy(const ccv_cnnp_model_t* const super, void* const context);
4707
4708static const ccv_cnnp_model_vtab_t ccv_cnnp_segmented_dense_isa = {
4709 .build = _ccv_cnnp_segmented_dense_build,
4710 .init_states = _ccv_cnnp_segmented_dense_init_states,
4711 .add_to_parameter = _ccv_cnnp_segmented_dense_add_to_parameter,
4712 .copy = _ccv_cnnp_segmented_dense_copy,
4713};
4714
4715ccv_cnnp_model_t* ccv_cnnp_segmented_dense(const int segments, const int count, const int no_bias, const int flags, const int is_trainable, const char* const name)
4716{
4717 ccv_cnnp_model_segmented_dense_t* const model_segmented_dense = (ccv_cnnp_model_segmented_dense_t*)cccalloccalloc(1, sizeof(ccv_cnnp_model_segmented_dense_t));
4718 model_segmented_dense->super.isa = &ccv_cnnp_segmented_dense_isa;
4719 model_segmented_dense->super.input_size = 3;
4720 model_segmented_dense->super.outputs = &model_segmented_dense->output;
4721 model_segmented_dense->super.output_size = 1;
4722 model_segmented_dense->super.is_trainable = is_trainable;
4723 ccv_cnnp_model_copy_name(&model_segmented_dense->super, name);
4724 model_segmented_dense->weights.d = CCV_NNC_NO_TENSOR_SYMBOL;
4725 model_segmented_dense->weights.graph = 0;
4726 model_segmented_dense->bias.d = CCV_NNC_NO_TENSOR_SYMBOL;
4727 model_segmented_dense->bias.graph = 0;
4728 model_segmented_dense->segments = segments;
4729 model_segmented_dense->count = count;
4730 model_segmented_dense->no_bias = no_bias;
4731 model_segmented_dense->flags = flags;
4732 return (ccv_cnnp_model_t*)model_segmented_dense;
4733}
4734
4735static ccv_cnnp_model_t* _ccv_cnnp_segmented_dense_copy(const ccv_cnnp_model_t* const super, void* const context)
4736{
4737 const ccv_cnnp_model_segmented_dense_t* const self = (const ccv_cnnp_model_segmented_dense_t*)super;
4738 return ccv_cnnp_segmented_dense(self->segments, self->count, self->no_bias, self->flags, self->super.is_trainable, self->super.name);
4739}