Skip to main content

METHODS article

Front. Neurorobot., 12 January 2022
Volume 15 - 2021 | https://doi.org/10.3389/fnbot.2021.815144

Event-Based Circular Detection for AUV Docking Based on Spiking Neural Network

Feihu Zhang* Yaohui Zhong Liyuan Chen Zhiliang Wang
  • School of Marine Science and Technology, Northwestern Polytechnical University, Xi'an, China

In this paper, a circular objects detection method for Autonomous Underwater Vehicle (AUV) docking is proposed, based on the Dynamic Vision Sensor (DVS) and the Spiking Neural Network (SNN) framework. In contrast to the related work, the proposed method not only avoids motion blur caused by frame-based recognition during docking procedure but also reduces data redundancy with limited on-chip resources. First, four coplanar and rectangular constrained circular light sources are constructed as the docking landmark. By combining asynchronous Hough circle transform with the SNN model, the coordinates of landmarks in the image are detected. Second, a Perspective-4-Point (P4P) algorithm is utilized to calculate the relative pose between AUV and the landmark. In addition, a spatiotemporal filter is also used to eliminate noises generated by the background. Finally, experimental results are demonstrated from both software simulation and experimental pool, respectively, to verify the proposed method. It is concluded that the proposed method achieves better performance in accuracy and efficiency in underwater docking scenarios.

1. Introduction

Although the exploitation of ocean resources has attracted significant interests from both industrial and societal, the development of marine science and technology still suffers limited activities (Saeki, 1985). Exploring underwater environments presents many problems, such as water pressure changing and oxygen supplying (Stachiw, 2004). Autonomous Underwater Vehicles (AUV) (Figure 1), often referred to the Unmanned Underwater Vehicles (UUV), have been developed along with the rapid exploitation of the ocean, and leading to a reduction in operational costs.

FIGURE 1
www.frontiersin.org

Figure 1. Autonomous underwater vehicle.

However, due to the volume and mass issues, AUV carries limited energy (Chiche et al., 2018). It is challenging to perform better in large-scale environments, and the AUV is often required to replenish energy and transmit information frequently. Therefore, the underwater docking technique is developed to provide powerful energy supply, information processing and communication support for AUVs (Benton et al., 2004). To the best knowledge of authors, almost all underwater docking tasks rely on optical cameras for short-range pose estimation between AUV and docking station (Wang et al., 2016). The docking system developed by Woods Hole Oceanographic Institute for Remus series AUV (Stokey et al., 2001) and the docking system designed by MBARI for bluefine AUV (McEwen et al., 2008) are two typical inclusive docking systems. In docking process, the guidance system plays a vital role concerning the whole system, while visual perception contributes a lot in short-range docking (Zhao et al., 2013).

In Zhong et al. (2019), a binocular localization method for AUV docking is presented, and an adaptively weighted OTSU method is developed for feature extraction, the operation frequency of which is about 10 Hz. Jointly, in Wang et al. (2016), gray-scale feature analysis, edge detection and morphology methods are used to improve the algorithm of calculating the centers of target lights. Furthermore, in Yan et al. (2019), a visual positioning algorithm based on the L-shaped light array is proposed. Previous studies generally used the frame-based camera to carry out detection, which contains redundant background data, lacks solutions to high exposure as well as high-speed response capability such as more than 1kHz.

As frame-based camera also suffers the over-exposure issue while close to the docking light, it is quite challenging to determine the surrounding environment and its own pose. Considering the instability of underwater motion, it is also difficult to keep relatively stationary. As a result, motion blur could not be directly eliminated. In contrast to frame-based cameras, the event-based camera is sensitive to dynamic information and suitable for moving target recognition. In Piatkowska et al. (2012), an algorithm for spatiotemporal tracking to detect moving persons that is suitable for DVS was proposed. In Chen (2018), discriminative knowledge was transferred from a frame-based convolutional neural network (CNN) to the event-based modality via intermediate pseudo-labels, and then supervised learning was combined to detect cars. In Seifozzakerini et al. (2016), a method using Hough transform and event-based clustering algorithm to track multiple lines was proposed. So far, most research in underwater applications still relies on frame-based cameras, whereas the event-based visual perception method has not been well-explored.

In this paper, an event-based detection of multiple circles for AUV Docking based on the spiking neural network method is proposed. The main contributions of this work are concluded as follows:

First, the proposed approach significantly eliminates the redundant information during the docking task. The frame-based camera produces information on the whole image. However, in underwater scenarios, most of the image backgrounds are adaptively filtered by the event-based camera. Thus, the computation performance is guaranteed with respect to the on-chip resource in the AUV.

Second, based on the Spiking Neural Network, the relative position information is acquired between the AUV and the docking ring. Furthermore, the PnP algorithm and a spatiotemporal filter are simultaneously utilized to estimate the relative depth and reduce the noise interference caused by vibration and background activities, respectively.

Third, the proposed approach keeps robust in complex underwater environments. The event-based camera could effectively eliminate motion blur and over-exposure, which is a natural advantage in underwater docking applications.

The structure of this paper is organized as follows: section 2 briefly introduces the backgrounds. Section 3 investigates the preprocessing work with respect to spatiotemporal filter. Section 4 presents the SNN detection framework and section 5 exhibits experimental results. Finally, this paper is concluded in section 6.

2. Background

In this paper, the overview of the proposed detection method could be briefly expressed in Figure 2, which is based on asynchronous Hough circle transform and the theory of Spiking Neural Network.

FIGURE 2
www.frontiersin.org

Figure 2. Overview of the detection method.

2.1. Dynamic Vision Sensor

The Dynamic Vision Sensor (DVS, also called event-based camera) is a neuromorphic camera that behaves similar to the human visual system by modeling the human retina (Lichtsteiner Patrick and Tobi, 2008). In contrast to the frame-based camera, which captures and transmits frames synchronously at a fixed frame rate, DVS keeps super performances by asynchronously transmitting events as soon as each occurs in a pixel. Once the logarithmic intensity change of a pixel is larger than a predefined threshold, an event of the corresponding polarity will asynchronously generate depending on the direction of the change of brightness (Seifozzakerini et al., 2016). Hence, it is sensitive to intensity logarithmic change. However, the information of the magnitude change is not transmitted. Every event consists of four parameters (t, x, y, p) including the timestamp t in μs, position (x, y) in pixels and polarity p which is binary (+/−), where parameters t,x, and y are integer values. Each pixel sensor is independent of the other pixel sensors so that its own intensity change can be adapted, and a very high dynamic range of DVS is formed. DVS outputs compressed digital data as a tuple, avoiding redundancy and latency caused by conventional cameras.

2.2. Spiking Neuron Model and Spiking Neural Network

Spiking neural network (SNN) is the third generation of neural network models, improving the realistic level of neural simulation (Maass, 1997). Each Spiking Neuron receives some spike inputs and generates one spike output (Burkitt, 2006). The input is a sequence of spikes that happen at different times, increasing or decreasing the neurons' Membrane Potential (MP). Moreover, MP is constantly decaying linearly till zero. Whenever the MP exceeds the positive or negative threshold (only a positive threshold considered to simplify the calculation), a spike is produced as output. Later MP of the neuron and its local lateral neighbors are inhibited to zero and then enter a short no response period.

In this paper, the leaky integrate-and-fire (LIF) (Brunel and Sergi, 1998) spiking neurons are considered to establish SNN. An SNN representing the parameter space is also utilized to detect circle objects based on the asynchronous Hough transform. The neurons' inputs are the pulses generated by Hough transform mappings, and the output neurons' coordinates are the pixel coordinates of the object centers. Algorithm 1 shows this neuron model, where ti is the current timestamp, ti−1 is the last time the neuron is stimulated, td is the attenuation duration, ui is the MP at time ti, sign(ui−1) is the positive or negative signal of the potential at the previous time and si is the input spike at time ti.

ALGORITHM 1
www.frontiersin.org

Algorithm 1. Updating procedure of a spiking neuron when receiving an input spike.

3. Spatiotemporal Filter for Reducing Noise

In fact, small changes in the lower intensity of a pixel often lead to an apparent change which may generate an event afterward. Hence, in darker places, more and more noisy events will be produced (Seifozzakerini et al., 2016). However, the frame-based filtering algorithm is not suitable for DVS. As Hough Transform (Illingworth and Kittler, 1987) is sensitive to noisy measurements, a spatiotemporal filter is utilized to process raw events before the detection phase.

Background Activity (BA) noise is expected in the event stream, which is produced by thermal noise and junction leakage currents (Lichtsteiner Patrick and Tobi, 2008). However, unlike actual events, BA events lack time correlation with other events in their spatial neighborhood. Besides, the BA events are proved to correspond to Poisson distribution (Khodamoradi and Kastner, 2021), the probability of a known number n of events if all those events are independent and happen at a given average rate λ:

P{n}=λn·e-λn!    (1)

In order to recover interested events from raw data, the spatiotemporal filter also records the early timestamps. Once an event is processed, the filter searches the corresponding spatial neighborhood. If the timestamp difference between two adjacent events in the very near spatial coordinate is found less than a threshold dt, it is regarded as an actual event. Otherwise, it is discarded. The principle could be briefly expressed as follows:

e(ti,xi,yi,si)BAnoise               |ttmn|dt,s.t.|mx|1|ny|1    (2)

Where e is the new event to be processed, tmn is the timestamp of the last event at the position (m, n) not including the new event, dt is the time threshold.

The Spatiotemporal filter proposed by Alireza Khodamoradi (Khodamoradi and Kastner, 2021) is suitable for embedded applications and moving cameras, which is applied in this paper. It uses only two memory cells for recursively filtering the image, which can significantly eliminate memory requirements. The computational performance is thus reduced from O(N2) to O(N). Meanwhile, this filter increases the data density of real events by 180%.

4. Event-Based Multiple Circle Detection and Pose Estimation

As mentioned above, optical guidance plays an essential role in close range, and the AUV is usually guided by lights mounted around the docking station. It is observed that using single lights makes the task challenging, as the 3D pose information is always missing. Therefore, at least three light sources are required for AUV docking. In this paper, four circle shape LED lights are utilized to overcome the aforementioned issue. From an underwater perspective, each circular light looks like a dot from long distances and a circle from nearby places. Furthermore, the halo often appears sparse compared to the dense light source, and the corresponding position includes bias regarding the light sources. In order to improve the localization accuracy, the position of the light source should be considered instead of the halo. Hence, the spatiotemporal noise filter is utilized to eliminate the halo around.

In this paper, the detection of multiple lights is performed using Hough transform (Hough, 1959). Note that the sparse events data has already been acquired with redundancy, while the temporal asynchrony process is implemented afterward. According to the difference between frame-based cameras and DVS that gradient information is challenging to obtain from frame-free events, traditional Hough circle transform is adapted to proceed asynchronous events. A simple way to solve this problem is to accumulate all events in a period into a pseudo frame, and then the conventional frame-based gradient Hough transform (Chen et al., 2012) can be used. However, ignoring the time information of every event like that will reduce the sensitivity and dynamic characteristics of the algorithm. To achieve this goal, an asynchronous Hough circle transform based on Spiking Neural Network is thus proposed to accurately and effectively detect underwater lights.

4.1. The Proposed SNN for Asynchronous Hough Circle Transform

Pixels are simultaneously generated on frame-based cameras. However, the stream from the event-based camera is asynchronous; that is, events are generated with time sequence. In order to take advantage of event streams, every single event must be processed asynchronously. In this paper, new events are processed immediately without being accumulated as a frame. Therefore, the asynchronous Hough transform algorithm in SNN is proposed to effectively identify the sparse events in the time scale of microseconds.

In this paper, we propose the SNN model with Hough Transform to detect multiple circle objects. As shown in Seifozzakerini et al. (2016), a straight line is detected based on a 2-dimensional SNN model, which only contains two parameters: distance ρ and angle θ. However, for circular feature detection, three parameters (x-coordinates, y-coordinates, and radius r) should be taken into account. Once the radii are unknown, the event-based Hough transform should be performed in 3D space (xc, yc, r). Ni et al. (2012) extracted the microspheres with known radius, but without utilizing the characteristics of SNN. To extend both the number and radius of potential objects, Hough circle transform and SNN are jointly utilized. Noting that constructing a 3D SNN leads to huge computation resources, only two parameters xc and yc are selected to avoid the computational cost. Here, the range of parameter r is manually selected outside the SNN. The process is as follows:

First, conduct a spatiotemporal filter (in section 3) with raw events to eliminate the noises and halos around the underwater lights.

Second, the Hough circle transform algorithm based on asynchronous events is utilized to map events from Cartesian coordinate space to 2D parameter space.

By continuously fetching the latest event ei = (ti, xi, yi, si) from the flow queue, Ptn is defined as a collection of points generated at timestamp tn:

Ptn={(yi,xi)|ei(ti,xi,yi,si)}    (3)

For each acquired event, the coordinates (yi, xi) are extracted, and the mapping from the Cartesian coordinate space to the Hough parameter space is performed. Note that the Hough transformation for each event is asynchronously processed. Especially every time an event produces mappings to the circle centered on itself with radius r and central angle θi from 0 to 360 degrees. The radius range is set from rmin to rmax aiming to detect different sizes. Therefore, 360 · (rmaxrmin) mappings are generated for one event at its timestamp ti. Mappings from every event which occurs at a circle's edge would include one mapping at its center (yc, xc). Thus the calculation formulas of the Hough circle transform are calculated as follows:

     xc=xi+r·cosθi     yc=yi+r·sinθis.t.r[rmin,rmax],θi[0,360)    (4)

Where xc and yc are the horizontal and vertical coordinates of the center, respectively. r is radius,θi is the central angle from (yi, xi) to (yc, xc), and (yi, xi) ∈ Ptn. Mappings outside the range scope are not considered.

Third, improve Hough transformation with SNN for circular detection.

It is observed from event streams that neurons at the object location output spikes. Hence laterally suppressing adjacent neurons make the detection task possible. Considering that the original Hough parameter space does not contain time information, a time-sensitive SNN is constructed as an intermediate layer for asynchronous processing. The input spikes of SNN are all mappings generated by asynchronous Hough transform, and the output spike is the pixel coordinates of targets.

A M × N SNN is established by using the LIF spiking neuron model (section 2.2), where M and N represent the parameter range of Y-coordinate and X-coordinate of the circle's center, respectively. The height and width of SNN can be selected according to the resolution of the camera. The membrane potentials of all neurons are initialized as 0 at the beginning. Meanwhile, a matrix is utilized to update the timestamp ti of each event, which is initialized by the first incoming event's timestamp. Noticed that each event in the receptor layer inputs spikes si to 360 · (rmaxrmin) neurons (yc, xc) in the intermediate layer, and each spike input increases the absolute value of MP regards to the neuron, whereas the MP always decreases with a fixed linear rate λ. The residual MP is calculated by MP at the last time ti−1 minus the decay value and adds the current spike value. td is the time duration between the current spike input time ti and the last spike input time ti−1. MP will not attenuate after decay to 0. The events can therefore be mapped to SNN by using Equation (4) and defined as follow:

Hough(Ptn):PtnSNNtn    (5)

where SNNtn is a matrix that changes over time:

SNNtn(yc,xc)={SNNtn1(yc,xc)+|si|ifHough(Ptn)SNNtn1(yc,xc)λ·tdwhenever               SNNtn1(yc,xc)>0                               foryc[0,rows),xc[0,columns),si=±1.    (6)

The intermediate layer mapped by all incoming events can be expressed as Hough(Ptnk+1, tn+1), where tn+1 is the current time, si is the input spike, λ is the rate of decay, and td is decay time. The following recursive formula for continuous conversion with time is utilized, which is also called continuous SNN based Hough mapping:

Hough(Ptn-k+1,tn+1)=Hough(Ptn-k,tn)          +Hough(Ptn+1)-Hough(Ptn-k)  Hough(Ptn-k,tn)=i=tn-ktnHough(Pti)    (7)

The continuous Hough mapping is performed in the intermediate layer of SNN. Once the value of MP exceeds the positive threshold, the neuron is activated and outputs a positive pulse at (xc, yc) at timestamp ti to the output layer. Then MP is reset to zero. Afterward, the neuron enters a short period of no response, during which the input pulse is ignored to avoid multiple outputs originating from the same target. Meanwhile, the activated neurons inhibit the MP of other neurons within the local lateral margin of m × m to reduce duplicate detections of one object. The MP of non-activated neurons and neurons not near activated neurons are not inhibited and still decay naturally. Once a neuron receives a pulse, it updates the timestamp of the corresponding coordinates within the timestamp matrix. The updating procedure of a neuron when receiving an input spike is shown in Algorithm 1 (section 2.2).

As the original polarity si of the event is divided into positive and negative, the two symmetrical edges of the target generate spiking inputs with opposite polarities, which leads to false alarms. Therefore, the absolute value of spikes is used in calculation, while all inputs are considered positive spikes.

Once a neuron (yc, xc) outputs a spike to the output layer at ti, the position of the center is detected. All spikes output in a period tnktn are counted, that is, all targets detected during that time. The process of SNN could be briefly shown in Figure 3. Firstly, the receptor layer's yellow, blue, and green events generated input spikes with time sequence. Secondly, the MP of corresponding neurons in the intermediate layer changed in turn. Finally, a red spike outputs afterward.

FIGURE 3
www.frontiersin.org

Figure 3. The running process of SNN based asynchronous hough circle transform. (A) Receptor layer (target in Cartesian space). (B) Output layer (results of detection). (C) Intermediate layer. (D) Change of membrane potential (with time and spike order).

Fourth, matching feature to the rectangular target with the following rules:

- The number of different spikes after ignoring duplicate is 4.

- The included angle between the two diagonal circles' connecting lines is within a specific pixel range. k is the slope of a line and α is the angle.

tan(α)=|(k2-k1)/(1+k1·k2)|s.t.tan(α1)<tan(α)<tan(α2)    (8)

- The length difference between two diagonal circles' connecting lines is within a specific pixel range, where l means the length.

l1<|(x1-x4)2+(y1-y4)2-(x2-x3)2+(y2-y3)2|<l2    (9)

Thus, the whole procedure of the proposed SNN for asynchronous Hough circle transform is presented as Algorithm 2.

ALGORITHM 2
www.frontiersin.org

Algorithm 2. Event-based multi-circle detecting asynchronously in the SNN.

4.2. Perspective-4-Point Algorithm for Pose Estimation

PnP algorithm is a method to estimate the pose of the camera relative to the world coordinate system by knowing the 3D coordinates of n points in space, the corresponding 2D point coordinates, and the internal parameter matrix of the camera. Considering the feasibility, four LED light sources are used to constitute the perspective-4-point (P4P) problem (Horaud et al., 1989). The problem is cast into solving an unknown biquadratic polynomial equation. It was developed as part of a monocular object recognition system (Horaud, 1987).

In the condition that 3D coordinates of 4 points (P1, P2, P3, P4) in the world coordinate system were known, the 2D coordinates of 4 points(p1, p2, p3, p4) in pixel coordinate system were calculated by Algorithm 2 and the internal parameter matrix K of the camera was calibrated, the camera pose relative to the world coordinate system can be calculated as follow.

[XcYcZc]=[Rt][XwYwZw]    (10)

In the formula, (Xc, Yc, Zc) and (Xw, Yw, Zw) represent both the camera and world coordinate system. Besides, R is the rotation matrix and t is the translation vector, which describe the transformation relationship between the two coordinate systems.

In order to solve the P4P problem faster on-chip, the algorithm of Gao et al. (2003) combined with the projection method was used. Firstly, four groups of solutions are calculated with three points to obtain four rotation matrices and translation matrices. Then the result [R t] can be calculated according to the formula:

(uv1)~[fx0cx0fycy001]·[r11r12r13t1r21r22r23t2r31r32r33t3]·[XYZ1]    (11)

Where fx, fy, cx, cy are parameters of the DVS, u, v are coordinates in image, R is the rotation matrix, t is the translation vector and (X, Y, Z) is the world coordinates of the 4th points.

After substituting (X, Y, Z) into the formula, four projections (u, v) in the image are obtained. The matrix [R t]with the slightest projection error is the right solution. Therefore, the pose is obtained as well as the relative depth between AUV and docking ring, which guides the AUV during docking task.

5. Experiments and Results

The proposed multiple circle detection method was evaluated in both the simulator and real scenario, while the effectiveness of estimating the position of AUV was evaluated.

In the simulation, the docking of AUV is carried out in V-REP to verify the practicability of the proposed positioning algorithm in SNN. The scene and the events are displayed respectively as Figures 4A,B. In general, the visual guidance system consists of four parts: 4 circular LED lights constrained by rectangle, a dynamic vision sensor, a computer and AUV. To achieve docking, AUV's three degrees-of-freedom (DOF) in transverse X, longitudinal Y, and vertical Z direction could be artificially controlled. Besides, the DVS was fixed on the head of the AUV, with the visual field being limited to 65 degrees. In addition, the targets remained stationary in the scene while AUV approached them vertically according to the relative position deviation. Moreover, the computer detected targets and estimated AUV's position by using SNN and P4P algorithms. In this case, X-axis and Y-axis were parallel to the target plane, while Z-axis was perpendicular. Note that Z-axis is less than 0 according to the right-handed coordinate system. The initial position (X, Y, Z) in millimeters of the four targets are (1000, 2000, 0), (1000, 4000, 0), (3000, 2000, 0), (3000, 4000, 0), and AUV's initial placement point is (2000, 3000, −3000). In order to compare the estimation difference among 3 DOF, AUV's position in the X direction is limited to 2000.

FIGURE 4
www.frontiersin.org

Figure 4. (A) AUV and targets in V-REP. (B) Events of targets in V-REP.

During the docking process, the motion command was sent to AUV to approach the targets gradually. When AUV was approaching, the position deviation of its Z-direction would gradually increase up to 0, which means the relative depth is getting closer. In Figure 5, the blue lines represent the actual spatial trajectory of the AUV relative to targets in 3 DOF, which is recorded by a graph fixed on AUV. The yellow lines represent the estimated position of AUV by using the SNN and P4P algorithm. Meanwhile, red lines represent the deviation of estimation during the AUV's movement, which is obtained by comparing the difference between the ground truth and estimation. As a result, Figure 5 verifies the feasibility of the visual positioning method.

FIGURE 5
www.frontiersin.org

Figure 5. Estimation of the AUV's position in V-REP. (A) X-axis position of AUV in V-REP. (B) Y-axis position of AUV in V-REP. (C) Z-axis position of AUV in V-REP.

In the pool experiments, the detection effects of the frame-based Gradient Hough Transform (GHT) (Chen et al., 2012) and the proposed SNN based method were compared.

The visual guidance system in the experimental pool consists of three parts: 4 LED light sources, a dynamic vision sensor and a computer. In the pool, X-axis and Y-axis were parallel to the target plane, while Z-axis was perpendicular. Four underwater circular lights were bound vertically to the corners of a rectangular plate and kept stationary underwater. In addition, the DVS used in experiment is produced by IniVation, which has a 320 × 240 spatial resolution, 1μs temporal accuracy and the internal parameters matrix (257.3, 164.1, 255.4, 130.4). Meanwhile, a waterproof shell coated with a metal oxide layer is processed. The experimental equipment is displayed as Figure 6. And to simulate the docking motion of AUV, the DVS was made close to or away from the stationary targets. During the process, the computer parsed DVS's data, detected targets and estimated DVS's position by using SNN and P4P algorithms.

FIGURE 6
www.frontiersin.org

Figure 6. The experimental equipments. (A) DVS with waterproof shell. (B) Targets with lights and steel plate.

In order to verify the detection effect of the SNN algorithm, it was compared with a baseline method. The frame-based Gradient Hough Transform(GHT) is a classical method for detecting circles and utilized as baseline. During the experiment for GHT, raw event stream in a period was first accumulated as a pseudo-frame. Then it was changed to a grayscale image, filtered by the median, operated by morphological open, and extracted the edges. Later the OpenCV library was utilized to realize the GHT method, and the parameters of GHT were set as Table 1. Detection effects of GHT were as Figure 7.

TABLE 1
www.frontiersin.org

Table 1. The parameters of pseudo-frame gradient Hough circle transform.

FIGURE 7
www.frontiersin.org

Figure 7. Results of frame-based GHT. (A) Raw event frame. (B) Effect of detection.

In the proposed method, the event stream was first put into the spatiotemporal filter compared to GHT. The time interval (dt) regarding the spatiotemporal neighbors and the number (n) of supporting pixels were configured as 1 ms and 76,800, respectively. A 240 × 320 SNN was established according to the DVS's resolution. Then the proposed Asynchronous Hough Circle Transform in SNN (section 4.1) was implemented in Python. The parameters in Table 2 were set to SNN, and the detected objects were exhibited in Figure 8 with the camera approaching targets. In Figure 8, the green points represent positive events which generate positive spikes to SNN, while the red points represent negative events.

TABLE 2
www.frontiersin.org

Table 2. The parameters of SNN based asynchronous Hough transform.

FIGURE 8
www.frontiersin.org

Figure 8. The detection effect of SNN. (A) Raw events of far targets. (B) Detection of far targets. (C) Raw events of near targets. (D) Detection of near targets.

Table 3 reports the statistics of detection results by using SNN. With time increasing, massive input events from the receptor layer generated fewer output spikes to the output layer, which indicated the changes of targets in the image. Therefore, the generated time, end time and the number of spikes were recorded for quantitative analysis. As we can see, the positions (X, Y) of all targets can be accurately detected and tracked.

TABLE 3
www.frontiersin.org

Table 3. Quantitative analysis of multiple circle detection by SNN.

Figure 9 demonstrates the detection performance between two methods when the pixel position of multiple targets changes with event sequence, where red and blue points respectively present the results of GHT and SNN. As illustrated in the figure, the accuracy of multi-target tracking of the two algorithms is close.

FIGURE 9
www.frontiersin.org

Figure 9. Multi-target tracking effect in image.

As presented in Figure 10A, the X-coordinate trajectories of the upper and the lower circle are basically the same, and so on in Y-coordinate of the left and right circle, which shows the accuracy of target detection. Besides, Figure 10B shows the detected X-axis trajectories of right and bottom objects by two different methods, Figure 10C shows the Y-axis trajectories of left and top objects, and Figure 10D shows the Y-axis trajectories of right and bottom objects. In addition, the detection results of the GHT and SNN methods are almost equal. However, the jitter of the curve of SNN method is smaller than GHT method. After obtaining image coordinates at the detection stage, relative position in the real scenario was estimated by the P4P algorithm. The estimation performance from two methods is displayed in Figure 11, where X and Y axes are positions parallel to the target plane, and Z-axis represents relative depth. Note that Z-axis position is always less than 0 because the camera only appears on one side of the target plane. It can be seen from the figure that two methods have similar results for position estimation.

FIGURE 10
www.frontiersin.org

Figure 10. Multi-target tracking effects in X and Y direction. (A) X-axis of left and top objects. (B) X-axis of right and bottom objects. (C) Y-axis of left and top objects. (D) Y-axis of right and bottom objects.

FIGURE 11
www.frontiersin.org

Figure 11. Position estimation in real world. (A) X-axis position of DVS in real world. (B) Y-axis position of DVS in real world. (C) Z-axis position of DVS in real world.

6. Conclusion

In this paper, a multiple circles detection and relative pose estimation method was proposed, which combined monocular DVS and SNN for AUV docking. The proposed method not only avoids motion blur caused by frame-based recognition but also adapts to high exposure at close range and reduces data redundancy effectively, which utilizes the biological characteristics of SNN and hardware features of DVS. We focus on calculating asynchronous Hough mappings and constructing an SNN model. The accuracy of our method is compared with a frame-based method. Pool experiments and simulations are carried out to verify the effectiveness of the method. In future work, stereo event-based cameras are considered to combine with SNN to improve the recognition accuracy and speed.

Data Availability Statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author/s.

Author Contributions

FZ and YZ present the idea in this article and write the paper together. LC and ZW provide the technical support and help to complete the experimental verification. All authors contributed to the article and approved the submitted version.

Funding

This study was supported by the National Natural Science Foundation of China (52171322), the National Key Research and Development Program (2020YFB1313200), and the Fundamental Research Funds for the Central Universities (D5000210944).

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Benton, C., Kenney, J., Nitzel, R., Blidberg, R., Chappell, S., and Mupparapu, S. (2004). “Autonomous undersea systems network (AUSNet) - protocols to support ad-hoc AUV communications,” in 2004 IEEE/OES Autonomous Underwater Vehicles (Wiscasset, ME), 83–87. doi: 10.1109/AUV.2004.1431197

PubMed Abstract | CrossRef Full Text | Google Scholar

Brunel, N., and Sergi, S. (1998). Firing frequency of leaky intergrate-and-fire neurons with synaptic current dynamics. J. Theoret. Biol. 195, 87–95. doi: 10.1006/jtbi.1998.0782

PubMed Abstract | CrossRef Full Text | Google Scholar

Burkitt, A. N.. (2006). A review of the integrate-and-fire neuron model: I. Homogeneous synaptic input. Biol. Cybernet. 95, 1–19. doi: 10.1007/s00422-006-0068-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, N. F. Y.. (2018). “Pseudo-labels for supervised learning on dynamic vision sensor data, applied to object detection under ego-motion,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops (Singapore), 644–653. doi: 10.1109/CVPRW.2018.00107

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, X., Lu, L., and Gao, Y. (2012). “A new concentric circle detection method based on Hough transform,” in 2012 7th International Conference on Computer Science Education (ICCSE) (Zhejiang), 753–758. doi: 10.1109/ICCSE.2012.6295182

PubMed Abstract | CrossRef Full Text | Google Scholar

Chiche, A., Lagergren, C., Lindbergh, G., and Stenius, I. (2018). “Sizing the energy system on long-range AUV,” in 2018 IEEE/OES Autonomous Underwater Vehicle Workshop (AUV), 1–6. doi: 10.1109/AUV.2018.8729812

PubMed Abstract | CrossRef Full Text | Google Scholar

Gao, X.-S., Hou, X.-R., Tang, J., and Cheng, H.-F. (2003). Complete solution classification for the perspective-three-point problem. IEEE Trans. Pattern Anal. Mach. Intell. 25, 930–943. doi: 10.1109/TPAMI.2003.1217599

PubMed Abstract | CrossRef Full Text | Google Scholar

Horaud, R.. (1987). New methods for matching 3-d objects with single perspective views. IEEE Trans. Pattern Anal. Mach. Intell. 9, 401–412. doi: 10.1109/TPAMI.1987.4767922

PubMed Abstract | CrossRef Full Text | Google Scholar

Horaud, R., Conio, B., Leboulleux, O., and Lacolle, B. (1989). An analytic solution for the perspective 4-point problem. Comput. Vision Graph. Image Process. 47, 33–44. doi: 10.1016/0734-189X(89)90052-2

CrossRef Full Text | Google Scholar

Hough, P. V. C.. (1959). “Machine analysis of bubble chamber pictures,” in Proceedings of the International Conference on High Energy Accelerators and Instrumentation, 554–556. Available online at: https://ci.nii.ac.jp/naid/10025522532/en/

Google Scholar

Illingworth, J., and Kittler, J. (1987). The adaptive hough transform. IEEE Trans. Pattern Anal. Mach. Intell. 9, 690–698. doi: 10.1109/TPAMI.1987.4767964

PubMed Abstract | CrossRef Full Text | Google Scholar

Khodamoradi, A., and Kastner, R. (2021). O(N)O(N)-space spatiotemporal filter for reducing noise in neuromorphic vision sensors. IEEE Trans. Emerg. Top. Comput. 9, 15–23. doi: 10.1109/TETC.2017.2788865

PubMed Abstract | CrossRef Full Text | Google Scholar

Lichtsteiner Patrick, P. C., and Tobi, D. (2008). A 128 × 128 120 db 15 μs latency asynchronous temporal contrast vision sensor. IEEE J. Solid State Circ. 43, 566–576. doi: 10.1109/JSSC.2007.914337

PubMed Abstract | CrossRef Full Text

Maass, W.. (1997). Networks of spiking neurons: the third generation of neural network models. Neural Netw. 10, 1659–1671. doi: 10.1016/S0893-6080(97)00011-7

CrossRef Full Text | Google Scholar

McEwen, R. S., Hobson, B. W., McBride, L., and Bellingham, J. G. (2008). Docking control system for a 54-cm-diameter (21-in) AUV. IEEE J. Ocean. Eng. 33, 550–562. doi: 10.1109/JOE.2008.2005348

PubMed Abstract | CrossRef Full Text | Google Scholar

Ni, Z., Pacoret, C., Benosman, R., Ieng, S., and Regnier, S. (2012). Asynchronous event-based high speed vision for microparticle tracking. J. Microsc. 245, 236–244. doi: 10.1111/j.1365-2818.2011.03565.x

CrossRef Full Text | Google Scholar

Piatkowska, E., Belbachir, A. N., Schraml, S., and Gelautz, M. (2012). “Spatiotemporal multiple persons tracking using dynamic vision sensor,” in 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, 35–40.

Google Scholar

Saeki, M.. (1985). Current research and development at the Japan Marine Science & Technology Center (JAMSTEC). IEEE J. Ocean. Eng. 10, 182–182. doi: 10.1109/JOE.1985.1145087

PubMed Abstract | CrossRef Full Text | Google Scholar

Seifozzakerini, S., Yau, W. Y., Zhao, B., and Mao, K. (2016). Event-based hough transform in a spiking neural network for multiple line detection and tracking using a dynamic vision sensor. BMVC 94, 1–12. doi: 10.5244/C.30.94

CrossRef Full Text | Google Scholar

Stachiw, J.. (2004). “Acrylic plastic as structural material for underwater vehicles,” in Proceedings of the 2004 International Symposium on Underwater Technology, 289–296. doi: 10.1109/UT.2004.1405581

PubMed Abstract | CrossRef Full Text | Google Scholar

Stokey, R., Allen, B., Austin, T., Goldsborough, R., Forrester, N., Purcell, M., et al. (2001). Enabling technologies for remus docking: an integral component of an autonomous ocean-sampling network. IEEE J. Ocean. Eng. 26, 487–497. doi: 10.1109/48.972082

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, G., Han, J., Wang, X., and Jing, D. (2016). “Improvement on vision guidance in AUV docking,” in OCEANS 2016 (Shanghai), 1–6. doi: 10.1109/OCEANSAP.2016.7485653

PubMed Abstract | CrossRef Full Text | Google Scholar

Yan, Z., Gong, P., Zhang, W., Li, Z., and Teng, Y. (2019). Autonomous underwater vehicle vision guided docking experiments based on l-shaped light array. IEEE Access 7, 72567–72576. doi: 10.1109/ACCESS.2019.2917791

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhao, C., Wang, X.-Y., Zhuang, G.-J., Zhao, M., and Ge, T. (2013). Motion of an underwater self-reconfigurable robot with tree-like configurations. J. Shanghai Jiaotong Univ. 18, 598–605. doi: 10.1007/s12204-013-1433-y

CrossRef Full Text | Google Scholar

Zhong, L., Li, D., Lin, M., Lin, R., and Yang, C. (2019). A fast binocular localisation method for AUV docking. Sensors 19:1735. doi: 10.3390/s19071735

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: DVS, SNN, AUV, Hough transform, P4P, docking

Citation: Zhang F, Zhong Y, Chen L and Wang Z (2022) Event-Based Circular Detection for AUV Docking Based on Spiking Neural Network. Front. Neurorobot. 15:815144. doi: 10.3389/fnbot.2021.815144

Received: 15 November 2021; Accepted: 08 December 2021;
Published: 12 January 2022.

Edited by:

Rui Li, Chongqing University, China

Reviewed by:

Xieyuanli Chen, University of Bonn, Germany
Yanyan Li, Technical University of Munich, Germany

Copyright © 2022 Zhang, Zhong, Chen and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Feihu Zhang, feihu.zhang@nwpu.edu.cn

Download