blob: 0c8d3532aac2bc534838f187ede7916d7737cb05 [file] [log] [blame]
Anthony Barbier871448e2017-03-24 14:54:29 +00001/*
Jenkins6a7771e2020-05-28 11:28:36 +01002 * Copyright (c) 2017-2020 ARM Limited.
Anthony Barbier871448e2017-03-24 14:54:29 +00003 *
4 * SPDX-License-Identifier: MIT
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to
8 * deal in the Software without restriction, including without limitation the
9 * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
10 * sell copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in all
14 * copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
19 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
22 * SOFTWARE.
23 */
24#include "arm_compute/runtime/CL/functions/CLCannyEdge.h"
25
26#include "arm_compute/core/CL/ICLTensor.h"
27#include "arm_compute/core/CL/OpenCL.h"
28#include "arm_compute/core/Error.h"
Anthony Barbier871448e2017-03-24 14:54:29 +000029#include "arm_compute/core/Validate.h"
30#include "arm_compute/runtime/CL/CLScheduler.h"
31#include "arm_compute/runtime/CL/functions/CLSobel3x3.h"
32#include "arm_compute/runtime/CL/functions/CLSobel5x5.h"
33#include "arm_compute/runtime/CL/functions/CLSobel7x7.h"
Jenkins6a7771e2020-05-28 11:28:36 +010034#include "support/MemorySupport.h"
Anthony Barbier871448e2017-03-24 14:54:29 +000035
36using namespace arm_compute;
37
Kaizen8938bd32017-09-28 14:38:23 +010038CLCannyEdge::CLCannyEdge(std::shared_ptr<IMemoryManager> memory_manager) // NOLINT
39 : _memory_group(std::move(memory_manager)),
40 _sobel(),
41 _gradient(),
42 _border_mag_gradient(),
43 _non_max_suppr(),
44 _edge_trace(),
45 _gx(),
46 _gy(),
47 _mag(),
48 _phase(),
49 _nonmax(),
50 _visited(),
51 _recorded(),
52 _l1_list_counter(),
Jenkins52ba29e2018-08-29 15:32:11 +000053 _l1_stack(),
54 _output(nullptr)
Anthony Barbier871448e2017-03-24 14:54:29 +000055{
56}
57
Jenkins52ba29e2018-08-29 15:32:11 +000058void CLCannyEdge::configure(ICLTensor *input, ICLTensor *output, int32_t upper_thr, int32_t lower_thr, int32_t gradient_size, int32_t norm_type, BorderMode border_mode,
59 uint8_t constant_border_value)
Anthony Barbier871448e2017-03-24 14:54:29 +000060{
Jenkins6a7771e2020-05-28 11:28:36 +010061 configure(CLKernelLibrary::get().get_compile_context(), input, output, upper_thr, lower_thr, gradient_size, norm_type, border_mode, constant_border_value);
62}
63
64void CLCannyEdge::configure(const CLCompileContext &compile_context, ICLTensor *input, ICLTensor *output, int32_t upper_thr, int32_t lower_thr, int32_t gradient_size, int32_t norm_type,
65 BorderMode border_mode,
66 uint8_t constant_border_value)
67{
Jenkins52ba29e2018-08-29 15:32:11 +000068 ARM_COMPUTE_ERROR_ON_NULLPTR(input, output);
Anthony Barbier871448e2017-03-24 14:54:29 +000069 ARM_COMPUTE_ERROR_ON_DATA_TYPE_CHANNEL_NOT_IN(input, 1, DataType::U8);
70 ARM_COMPUTE_ERROR_ON_DATA_TYPE_CHANNEL_NOT_IN(output, 1, DataType::U8);
71 ARM_COMPUTE_ERROR_ON((1 != norm_type) && (2 != norm_type));
Jenkins52ba29e2018-08-29 15:32:11 +000072 ARM_COMPUTE_ERROR_ON((gradient_size != 3) && (gradient_size != 5) && (gradient_size != 7));
73 ARM_COMPUTE_ERROR_ON((lower_thr < 0) || (lower_thr >= upper_thr));
74
75 _output = output;
Anthony Barbier871448e2017-03-24 14:54:29 +000076
77 const unsigned int L1_hysteresis_stack_size = 8;
Anthony Barbier871448e2017-03-24 14:54:29 +000078 const TensorShape shape = input->info()->tensor_shape();
79
80 TensorInfo gradient_info;
81 TensorInfo info;
82
83 // Initialize images
84 if(gradient_size < 7)
85 {
Anthony Barbiera4376382017-04-12 15:12:46 +010086 gradient_info.init(shape, 1, arm_compute::DataType::S16);
87 info.init(shape, 1, arm_compute::DataType::U16);
Anthony Barbier871448e2017-03-24 14:54:29 +000088 }
89 else
90 {
Anthony Barbiera4376382017-04-12 15:12:46 +010091 gradient_info.init(shape, 1, arm_compute::DataType::S32);
92 info.init(shape, 1, arm_compute::DataType::U32);
Anthony Barbier871448e2017-03-24 14:54:29 +000093 }
94
95 _gx.allocator()->init(gradient_info);
Anthony Barbier871448e2017-03-24 14:54:29 +000096 _gy.allocator()->init(gradient_info);
Anthony Barbier871448e2017-03-24 14:54:29 +000097 _mag.allocator()->init(info);
Anthony Barbier871448e2017-03-24 14:54:29 +000098 _nonmax.allocator()->init(info);
Anthony Barbier871448e2017-03-24 14:54:29 +000099
100 TensorInfo info_u8(shape, 1, arm_compute::DataType::U8);
Anthony Barbier871448e2017-03-24 14:54:29 +0000101 _phase.allocator()->init(info_u8);
Anthony Barbier871448e2017-03-24 14:54:29 +0000102 _l1_list_counter.allocator()->init(info_u8);
Anthony Barbier871448e2017-03-24 14:54:29 +0000103
104 TensorInfo info_u32(shape, 1, arm_compute::DataType::U32);
Anthony Barbier871448e2017-03-24 14:54:29 +0000105 _visited.allocator()->init(info_u32);
Anthony Barbier871448e2017-03-24 14:54:29 +0000106 _recorded.allocator()->init(info_u32);
Anthony Barbier871448e2017-03-24 14:54:29 +0000107
108 TensorShape shape_l1_stack = input->info()->tensor_shape();
109 shape_l1_stack.set(0, input->info()->dimension(0) * L1_hysteresis_stack_size);
110 TensorInfo info_s32(shape_l1_stack, 1, arm_compute::DataType::S32);
Anthony Barbier871448e2017-03-24 14:54:29 +0000111 _l1_stack.allocator()->init(info_s32);
Anthony Barbier871448e2017-03-24 14:54:29 +0000112
Kaizen8938bd32017-09-28 14:38:23 +0100113 // Manage intermediate buffers
114 _memory_group.manage(&_gx);
115 _memory_group.manage(&_gy);
116
Anthony Barbier871448e2017-03-24 14:54:29 +0000117 // Configure/Init sobelNxN
118 if(gradient_size == 3)
119 {
Kaizen8938bd32017-09-28 14:38:23 +0100120 auto k = arm_compute::support::cpp14::make_unique<CLSobel3x3>();
Jenkins6a7771e2020-05-28 11:28:36 +0100121 k->configure(compile_context, input, &_gx, &_gy, border_mode, constant_border_value);
Anthony Barbier871448e2017-03-24 14:54:29 +0000122 _sobel = std::move(k);
123 }
124 else if(gradient_size == 5)
125 {
Kaizen8938bd32017-09-28 14:38:23 +0100126 auto k = arm_compute::support::cpp14::make_unique<CLSobel5x5>();
Jenkins6a7771e2020-05-28 11:28:36 +0100127 k->configure(compile_context, input, &_gx, &_gy, border_mode, constant_border_value);
Anthony Barbier871448e2017-03-24 14:54:29 +0000128 _sobel = std::move(k);
129 }
130 else if(gradient_size == 7)
131 {
Kaizen8938bd32017-09-28 14:38:23 +0100132 auto k = arm_compute::support::cpp14::make_unique<CLSobel7x7>();
Jenkins6a7771e2020-05-28 11:28:36 +0100133 k->configure(compile_context, input, &_gx, &_gy, border_mode, constant_border_value);
Anthony Barbier871448e2017-03-24 14:54:29 +0000134 _sobel = std::move(k);
135 }
136 else
137 {
Jenkins0e205f72019-11-28 16:53:35 +0000138 ARM_COMPUTE_ERROR_VAR("Gradient size %d not supported", gradient_size);
Anthony Barbier871448e2017-03-24 14:54:29 +0000139 }
140
Kaizen8938bd32017-09-28 14:38:23 +0100141 // Manage intermediate buffers
142 _memory_group.manage(&_mag);
143 _memory_group.manage(&_phase);
144
Anthony Barbier871448e2017-03-24 14:54:29 +0000145 // Configure gradient
Jenkins6a7771e2020-05-28 11:28:36 +0100146 _gradient.configure(compile_context, &_gx, &_gy, &_mag, &_phase, norm_type);
Anthony Barbier871448e2017-03-24 14:54:29 +0000147
Kaizen8938bd32017-09-28 14:38:23 +0100148 // Allocate intermediate buffers
149 _gx.allocator()->allocate();
150 _gy.allocator()->allocate();
151
152 // Manage intermediate buffers
153 _memory_group.manage(&_nonmax);
154
Anthony Barbier871448e2017-03-24 14:54:29 +0000155 // Configure non-maxima suppression
Jenkins6a7771e2020-05-28 11:28:36 +0100156 _non_max_suppr.configure(compile_context, &_mag, &_phase, &_nonmax, lower_thr, border_mode == BorderMode::UNDEFINED);
Anthony Barbiera4376382017-04-12 15:12:46 +0100157
Kaizen8938bd32017-09-28 14:38:23 +0100158 // Allocate intermediate buffers
159 _phase.allocator()->allocate();
160
Anthony Barbiera4376382017-04-12 15:12:46 +0100161 // Fill border around magnitude image as non-maxima suppression will access
162 // it. If border mode is undefined filling the border is a nop.
Jenkins6a7771e2020-05-28 11:28:36 +0100163 _border_mag_gradient.configure(compile_context, &_mag, _non_max_suppr.border_size(), border_mode, constant_border_value);
Anthony Barbier871448e2017-03-24 14:54:29 +0000164
Kaizen8938bd32017-09-28 14:38:23 +0100165 // Allocate intermediate buffers
166 _mag.allocator()->allocate();
167
168 // Manage intermediate buffers
169 _memory_group.manage(&_visited);
170 _memory_group.manage(&_recorded);
171 _memory_group.manage(&_l1_stack);
172 _memory_group.manage(&_l1_list_counter);
173
Anthony Barbier871448e2017-03-24 14:54:29 +0000174 // Configure edge tracing
Jenkins6a7771e2020-05-28 11:28:36 +0100175 _edge_trace.configure(compile_context, &_nonmax, output, upper_thr, lower_thr, &_visited, &_recorded, &_l1_stack, &_l1_list_counter);
Anthony Barbiera4376382017-04-12 15:12:46 +0100176
Kaizen8938bd32017-09-28 14:38:23 +0100177 // Allocate intermediate buffers
Anthony Barbiera4376382017-04-12 15:12:46 +0100178 _visited.allocator()->allocate();
179 _recorded.allocator()->allocate();
180 _l1_stack.allocator()->allocate();
181 _l1_list_counter.allocator()->allocate();
182 _nonmax.allocator()->allocate();
Anthony Barbier871448e2017-03-24 14:54:29 +0000183}
184
185void CLCannyEdge::run()
186{
Jenkins4ba87db2019-05-23 17:11:51 +0100187 MemoryGroupResourceScope scope_mg(_memory_group);
Kaizen8938bd32017-09-28 14:38:23 +0100188
Anthony Barbier871448e2017-03-24 14:54:29 +0000189 // Run sobel
190 _sobel->run();
191
192 // Run phase and magnitude calculation
193 CLScheduler::get().enqueue(_gradient, false);
194
Anthony Barbiera4376382017-04-12 15:12:46 +0100195 // Fill border before non-maxima suppression. Nop for border mode undefined.
196 CLScheduler::get().enqueue(_border_mag_gradient, false);
197
Anthony Barbier871448e2017-03-24 14:54:29 +0000198 // Run non max suppresion
199 _nonmax.clear(CLScheduler::get().queue());
200 CLScheduler::get().enqueue(_non_max_suppr, false);
201
202 // Clear temporary structures and run edge trace
Jenkins52ba29e2018-08-29 15:32:11 +0000203 _output->clear(CLScheduler::get().queue());
Anthony Barbier871448e2017-03-24 14:54:29 +0000204 _visited.clear(CLScheduler::get().queue());
205 _recorded.clear(CLScheduler::get().queue());
206 _l1_list_counter.clear(CLScheduler::get().queue());
207 _l1_stack.clear(CLScheduler::get().queue());
208 CLScheduler::get().enqueue(_edge_trace, true);
Anthony Barbier871448e2017-03-24 14:54:29 +0000209}