Abstract
Insect behaviour is an important research topic in plant protection. To study insect behaviour accurately, it is necessary to observe and record their flight trajectory quantitatively and precisely in three dimensions (3D). The goal of this research was to analyse frames extracted from videos using Kernelized Correlation Filters (KCF) and Background Subtraction (BS) (KCF-BS) to plot the 3D trajectory of cabbage butterfly (P. rapae). Considering the experimental environment with a wind tunnel, a quadrature binocular vision insect video capture system was designed and applied in this study. The KCF-BS algorithm was used to track the butterfly in video frames and obtain coordinates of the target centroid in two videos. Finally the 3D trajectory was calculated according to the matching relationship in the corresponding frames of two angles in the video. To verify the validity of the KCF-BS algorithm, Compressive Tracking (CT) and Spatio-Temporal Context Learning (STC) algorithms were performed. The results revealed that the KCF-BS tracking algorithm performed more favourably than CT and STC in terms of accuracy and robustness.
Similar content being viewed by others
Introduction
Insect behaviour has become an important research direction in the field of plant protection1. Behavioural research may inform methods for biological control2,3, biological model construction4,5 and plant-insect interactions6,7. To study the behaviour of flying insects accurately, it is necessary to observe and record their flight trajectory quantitatively and precisely in three dimensions (3D).
Traditional detection methods for insect behaviour depend mainly on direct and manual observation, which are complicated by arbitrary qualification, wasting of human resources and low effectiveness8. Recent developments in computer vision have stimulated the application of these techniques to insect tracking9,10,11. Straw et al.12 used three cameras to three-dimensionally track flying animals and can also track flies and birds, but the use of three cameras increased the difficulty of matching and the amount of 3D coordinate calculations. Okubo et al.13 obtained images of a group of mosquitoes with a single camera and constructed the 3D trajectory of mosquitoes based on the geometric relationship between the mosquito group and its shadow on a white background. Stowers et al.14 used the FreemoVR platform to establish a height-aversion assay in mice and studied visuomotor effects in Drosophila and zebrafish. However, this method was not directly suitable for the investigation of behaviours for which stereopsis is important because it rendered visual stimuli in a perspective-correct manner for a single viewpoint. Jantzen and Eisner15 implemented Lepidoptera’s 3D trajectory tracking, and Lihoreau et al.16 obtained the three dimensional foraging flights of bumblebees. However, in these studies, the experimental environment was relatively simple, and the target was obvious.
Automated image-based tracking has been applied for outdoor research, and the imaging method includes thermal infrared17, sonar18, 3D19, and harmonic radar20 methods. Xu et al.21 proposed a method for the 3D observation of fish based on a single camera. A waterproof mirror was installed above an experimental fish tank to simulate a camera shooting from top to bottom. Although monocular vision was able to determine a 3D trajectory, the shooting video was largely influenced by environmental factors, and the calculation process was complex. Hardie and Powell22 investigated the use of two or more parallel cameras to obtain an image sequence of aphid-landing outdoors, and drew the 3D landing trajectory, but this method involved a complex 3D matching process due to the complicated external environment, which increased the difficulty of matching. Risse et al.23 used a freely moving camera to track small animals in cluttered natural environments. Animals were represented by a single point, and the animal had to move in the majority of the frames to be tracked correctly. Although outdoor research has scored some achievements, complex natural environments (e.g., light, temperature, spatial structure of habitats, leaves, animals) increases the difficulty of outdoor research. Indoor research can weaken the impact of environmental factors and be conducive to the scholarly understanding of insect behaviour and also provide a technical and theoretical basis for outdoor experiments.
The cabbage butterfly is one of the most harmful insects to agriculture24, and thus it is necessary to track them and obtain their flight trajectories. In this study, we chose the cabbage butterfly adult as the research object, and performed research in a wind tunnel to which plants and wind were added to introduce a more natural environment than an empty tunnel. We then designed a orthogonal binocular vision measurement system with top-view and side-view cameras aimed at the wind-tunnel experimental environment, reducing the difficulty of camera matching. The insect can be obscured by plants. We propose a method for tracking and obtaining 3D trajectories based on a combination of the kernelized correlation filter (KCF) algorithm and background subtraction (BS) (KCF-BS). This method can occlusion and re-track the target after the target was lost. The proposed method will provide a theoretical foundation and valuable reference for further behavioural research on this insect and a technical reference for target tracking based on orthogonal binocular vision.
Results
Kernel function and characteristic parameter optimization
To optimize the kernel functions and the characteristics of tracking of cabbage butterfly targets, a top-view video of flight in the wind tunnel was employed as a test. Linear, Polynomial and Gaussian kernels were used successively in tracking tests under Histogram of Oriented Gradient (HOG)25 and Gray characteristics, while the actual flight curve of the cabbage butterfly target centroid coordinates in the test video frames was manually marked. Comparisons of the trajectory results from Linear, Polynomial and Gaussian kernels under the two characteristics in the top view video with the actual flight path are shown in Fig. 1a, and the error distance (Euclidean distance) vs. time (frame) for all algorithm and parameters are shown in Fig. 1b. The Euclidean distance in pixels was used to express the relative error between tracking and the actual trajectories, as shown in Table 1.
Figure 1a shows that the tracking trajectory and the actual trajectory curves were quite close to each other, but there was a relatively large deviation between the tracking and actual trajectory curves using HOG/Gaussian, and HOG/Polynomials (Fig. 1b) and a relatively smaller tracking error using the Gray characteristic (Fig. 1b and Table 1). Although the average relative error of Gray/Gaussian was smallest, the maximum and minimum relative errors of Gray/Gaussian were larger and might lead to large fluctuations in estimates. Considering this synthetically using multiple factors, Gray/Linear with smaller maximum, minimum and average relative errors was selected as optimal for tracking the 3D trajectory in this paper.
Target tracking and centroid trajectory acquisition
In the preliminary tracking test, the threshold T was 3. The test was conducted in an open area without plant shelter, when the insect was occluded by the plant leaves or was located at the edge of the wind tunnel, and when the background was similar to the target colour. Figure 2 shows tracking data for the insect target in the top-view video based on the KCF algorithm (first column), and the KCF- BS method (second column).
Figure 2a shows the tracking results when the insect was located in the open area in the centre of the wind tunnel, and both algorithms could be used to completely track and plot the trajectory of the target centroid. Figure 2b–d shows the tracking results when the insect was occluded by plant leaves, when located at the edge of wind tunnel, and when the target colour was similar to the background. Comparing the tracking results of the two methods indicates that, in those circumstances, the KCF algorithm alone failed to track continuously but the method of KCF-BS accurately tracked the actual flying trajectory of the cabbage butterfly, but missed information when the background colour was similar to the target. The success rates of the two algorithms in the four cases are shown in Table 2.
Comparison experiment
To verify the effectiveness of the tracking method described in this paper, Compressive Tracking (CT)26 and the Spatio-Temporal Context Learning (STC)27 were performed in the top-view video. Figure 3 shows comparisons of the object using the KCF-BS algorithm (first column), the CT algorithm (second column) and the STC algorithm (third column) in the wind tunnel centre case. The CT algorithm dropped the target at the 45th frame (Fig. 3c2) and could not effectively track the target again (Fig. 3d2). The target was lost in the 59th frame using the STC algorithm (Fig. 3d3). The KCF-BS algorithm could still track the target in the 45th and 59th frames.
Figure 4a shows the actual trajectory and tracking trajectory using the KCF-BS algorithm, the CT algorithm and the STC algorithm in the wind tunnel centre case, and Fig. 4b shows the error distance (Euclidean distance) vs. time (frame) for all algorithms. CT and STC algorithms could not continuously track the target (Fig. 4a) and showed a large deviation in Fig. 4b. When the target was not lost, the average errors of the CT and STC algorithms were 8.3 pixels and 10.9 pixels, respectively. The KCF-BS algorithm tracked the target continuously with an average error was 7.1 pixels. The success rates of the CT, STC and KCF-BS algorithms in the four cases are shown in Table 3.
Acquisition of the 3D trajectory
A target-tracking test was performed on the acquired cabbage butterfly video, and a representative result for the side-view and top-view video target tracking is shown in Fig. 5. The 3D spatial coordinate sequence P(x n , y n , z n ) during cabbage butterfly flying was generated from the coordinate matrices P(x n , z n ) and P(x n , y n ) obtained by side-view and top-view video tracking, respectively. Thus, the 3D trajectory of the cabbage butterfly in flight could be drawn (Fig. 6).
Figure 6 shows that the 3D trajectory obtained in the wind tunnel centre case was better and smoother. The abnormality in the 3D trajectory (Fig. 6b) was caused by the occlusion of the plant, as seen from Fig. 5b1,b2, but the target could still be effectively tracked afterwards. In the case of the target at the edge of wind tunnel, the poor 3D trajectory (Fig. 6c) was caused by the similar colour of the background area and the target, which as seen from Fig. 5c1,c2.
Accuracy evaluation of target three-dimensional coordinates
To evaluate the accuracy of the 3D coordinates of the insect target in this paper, five white ping-pong balls with a diameter of 40 mm were numbered and suspended from fine lines in different locations in the wind tunnel (the fan was stopped), and the actual distances between the ping-pong balls were measured using a steel tape. The 3D spatial coordinates of the centre of each sphere were also obtained using the methods described in this paper, and the Euclidean distance between the balls was calculated. The comparison of the calculated and actual distance between balls is shown in Table 4. The minimum, maximum and average errors were 4 mm, 20 mm, and 8.8 mm, respectively. The standard deviation was 4.9 mm.
The 3D trajectory error of the actual distance of KCF-BS, CT and STC algorithms in the wind tunnel centre case are shown in Fig. 7. The minimum, maximum and average errors of the KCF-BS algorithm were 0.11 cm, 1.65 cm, 0.76 cm, respectively. As seen from Fig. 7b,c, the error of CT and STC algorithms is large because the CT and STC algorithms could not redetect the target when the target was lost, and the position of the tracking box was basically unchanged.
Discussion
The experiment was performed in the wind tunnel, and cabbage butterfly occupied a small proportion in the scene with deformation, making video capture and tracking detection difficult. To meet the tracking test at 60fps, the present study proposed improved upon KCF and compared these results to KCF.
Tracking experiments using the KCF algorithm found that the algorithm missed the target when the butterfly was occluded by a plant, or when it was near the edge of wind tunnel, or when it was similar in colour to the background due to deformation during flight (Fig. 2b1–d1). For example, when the wings were in the closed state during flying, a partial lack of information led to tracking failure. The grey-black body and white wings of the adult cabbage butterfly resembled the colour of the floor in the top view as seen through the partial area of the glass frame, and thus the target could not be effectively detected. After missing the tracking target, the target box remained in its current position and was unable to re-detect and recognize the target. However, in the above situation, the proposed KCF-BS method could still obtain better test results (Fig. 2b2–d2). The success rates of KFC-BS detection in the cases of wind tunnel centre, obscured target, wind tunnel edge and similar colours were 100%, 81.0%, 96.4% and 52.6%, respectively (Table 2). In the obscured target case, the frames missed by the KCF-BS were the target occlusion frames. In the wind tunnel edge case, the frames lost by KCF-BS were due to the similar colour of the target and the background, as proved in the similar colours case (Fig. 2d2), and KCF-BS missed more target frames, which was a further consideration.
To verify the effectiveness of the tracking method described in this paper, the CT algorithm and STC algorithm were used to assess the performance of the KCF-BS method. As seen from Figs 3 and 4, the KCF-BS method can more precisely track the target in this case.
CT and STC algorithms are commonly used in tracking algorithms28,29,30,31. The CT algorithm is a discriminant tracking algorithm that uses a rectangular window to represent the tracking result. Most of the pixels in the tracking window come from the target and a small part comes from the background. When sparse local Haar-like features are extracted in the window, the pixels in the background affect the robustness of the algorithm. The cabbage butterfly possess a grey-black body and white wings with fine scales, and has deformation in motion. When the wings of the cabbage butterfly were closed, the background pixels in the tracking window occupy a large proportion, and thus, the CT algorithm could not effectively detect the target. As seen from the second column of Figs 3c2,d2 and 4a, the CT algorithm dropped the target and could not effectively track the target again. This was reflected in Table 3, with a lower accuracy rate. The STC algorithm was based on the generative appearance model and was integrated with tracking, learning and detection. In STC, Zhang et al.27 added background information including spatial location in the generative appearance model and adopted Fast Fourier Transform (FFT) to reduce calculations, so this algorithm could achieve real-time, robust and efficient tracking results. However, the method was less effective in terms of target distortion, background colour similarity and motion blur. In tracking the cabbage butterfly, the STC algorithm also dropped the target (Figs 3d3 and 4a). This was reflected in Table 3 lower accuracy. When the target was not lost, the average errors of the CT and STC algorithms were 8.3 pixels and 10.9 pixels, respectively. The KCF-BS algorithm could track the target continuously, and the average error was 7.1 pixels. When the target disappeared or the tracking effect was poor, the BS algorithm was used to retrieve the target position. However, when the target was similar in colour to the background, the phenomenon of missing frames was still observed (Fig. 2d2).
The two-dimensional coordinates of the target centroid were obtained by tracking cabbage butterflies in top-view and side-view videos using the KCF-BS algorithm. The 3D coordinates of the target centroid were obtained by matching the corresponding frames in the videos from the two cameras, from which the 3D trajectories was obtained. Comparing the measured distance and actual distance between the balls (40 mm diameter) of the test ball with the Euclidean distance obtained by the algorithm proposed in this paper, the results showed that the average error was 8.8 mm and the standard deviation was 4.9 mm. The distance error was mainly caused by camera calibration error, orthogonal binocular visual matching error and manual measurement error. The calibration pixel errors for the top and side view cameras were (0.51; 0.49) and (0.42; 0.52), respectively. The orthogonal binocular visual matching error, caused by hardware performance, was ignored. The manual measurement errors were inevitable, and we invited experienced researchers to make the measurements to reduce the errors. In Fig. 7, the minimum, maximum and average errors of KCF-BS algorithm were 0.11 cm, 1.65 cm, and 0.76 cm, respectively. The centroid positioning deviation caused by deformation of the cabbage butterfly during flying affected the measurement accuracy of the target. The problem of how to effectively acquire the target centroid must still be solved.
The method in this paper achieves target tracking and obtains three-dimensional motion trajectory, and this study of the cabbage butterfly provides a technical basis for the study of insect behaviour and a reference for tracking research. The proposed method was only suitable for a single moving insect, and the tracking trajectory may be incomplete when the target and background are similar. Further study is needed to apply this method detection and tracking to the high-speed flying of multiple targets and to improved tracking performance when the colour of the target and background are similar.
Methods
Insects
The cabbage butterfly has a wide distribution worldwide, and cabbage butterfly adults are employed as the experimental subjects (Fig. 8); these adults have a grey-black body and white wings with fine scales. They prefer to fly under strong light in the daytime. A total of 26 cabbage butterfly adults were acquired from the Chinese Arid Area Research Institute of Water-Saving Agriculture, Northwest A & F University, Yangling, Shannxi, China from May 25 to June 8, 2016. Their body lengths were 12–20 mm and wing expanses were 45–55 mm. The captured cabbage butterflies were fed honey-water for 1–2 days before use in experiments.
Experimental scene
Wind-tunnel technology is an important experimental setup to study the behaviour of insects32. This experiment was conducted in the wind tunnel of the Insect Behavioral Laboratory of Northwest A&F University. The dimensions of the wind tunnel are 1.8 m × 0.58 m × 0.58 m (L × W × H). Potted cabbage seedlings ~20 cm high and potted green pepper seedlings ~18 cm high were placed on the left side of the wind tunnel to serve as an experimental chemical source for the butterfly. An axial fan was installed on the right side of the tunnel to produce a weak airflow from right to left. In this paper, wind tunnel airflow velocity was set to 0.4 m/s to simulate the breeze in a natural environment.
Design of the video capture system
To obtain behavioural information from the butterfly, we designed an orthogonal binocular vision video acquisition system appropriate to the location and size of the wind tunnel (Fig. 9). The system consisted of two CCD cameras, a trigger, a computer, a camera-mounting bracket and ancillary equipment. To detect flight motion parameters for the fast-flying insects, we selected a gigabit network industrial camera (MICROVIEW, RS-A2300-GM60, BeiJing, China) with 1600*1200 pixels, a sensor of 7.2 × 5.4 mm, a lens (KOWA, LM3NCF, BeiJing, China) with a focal length of 3.5 mm, and a field of view of 89.0° × 73.8°. The camera was connected to a gigabit network card via an RJ45 interface to transmit grey video information. The optical axes of two cameras with same configuration were mounted orthogonally on brackets to obtain side-view and top-view image sequences, respectively. To effectively obtain 3D information for the targets, a MATLAB camera calibration toolbox was employed for internal parameter calibration of the two cameras33,34. To achieve two-camera synchronized shooting, an external trigger (MVM-B-0008-S00-V0001) was used to achieve synchronous acquisition. A 6-pin IO connector at the rear of the camera powered the camera and provided a trigger I/O. The trigger was connected to the two cameras via network cables connected to the computer. The unified trigger input pulse signals were set on the computer with control software to directly control the synchronized acquisition of two cameras.
Data processing platform
The video data were obtained using the above system, and valid video segments were randomly selected for experimentation. Then, a computer with an Intel Core I5-2400 processor with 3.2 GHz dominant frequency, 8 GB memory and a 500 G byte hard drive, was used to process video data. All tests were conducted using the MATLAB 2014a environment.
Cabbage butterfly target tracking
To obtain a 3D trajectory, it was first necessary to track the cabbage butterfly targets. In this study, the frame rates of the two cameras reached 60 fps to prevent the target displacement between frames from becoming too large when cabbage butterfly flying at maximum speed, and this improved the tracking accuracy. Henriques et al.35 proposed the KCF tracking algorithm in 2015, which is a new type of high-speed tracking algorithm. By constructing a classifier between the target and the background to determine the target, the achieved algorithm works at high speed with fast training and fast detection. Therefore, in applications to fast-moving target tracking, the KCF tracking algorithm has broad prospects. Tracking experiments using the KCF algorithm found that when the target was lost, the method was unable to re-detect and recognize the target. The background subtraction (BS) method is an effective motion detection algorithm, for which the motion area was detected by comparing the current frame with the background image difference36, and BS was used to regain the goal in this paper. Therefore, the KCF algorithm and BS method were combined in this paper to track cabbage butterfly targets. The principle of this method was to manually select the first frame target and to use the KCF-BS algorithm to achieve automatic target tracking. The main steps of the tracking algorithm for the cabbage butterfly target were as follows35:
-
(1)
Obtaining the cabbage butterfly base sample a and building a circulant matrix.
In the original frames of the video, the rectangular area containing the butterfly and some surrounding environmental information were manually selected as the base sample a, which represents the characteristic information of the cabbage butterfly target. For the grey characteristic, with a as the two-dimensional matrix, the sample a was cyclically shifted to obtain the circulant matrix:
$$A=[\begin{array}{ccccc}a & {R}^{1}a & {R}^{2}a & \cdots & {R}^{n}a\\ {D}^{1}a & {D}^{1}{R}^{1}a & {D}^{1}{R}^{2}a & \cdots & {D}^{1}{R}^{n}a\\ {D}^{2}a & {D}^{2}{R}^{1}a & {D}^{2}{R}^{2}a & \cdots & {D}^{2}{R}^{n}a\\ \vdots & \vdots & \vdots & \ddots & \vdots \\ {D}^{n}a & {D}^{n}{R}^{1}a & {D}^{n}{R}^{2}a & \cdots & {D}^{n}{R}^{n}a\end{array}]$$(1)where Ri is the bias matrix circularly moving i places to the right; and Dj is the bias matrix circularly moving j places downward.
-
(2)
Diagonalization of the Fourier space of the circulant matrix and the search for a simplified linear regression function.
The sample training process for butterfly target tracking was a ridge regression, which aimed to create a linear regression function f(c) = wTc, to make f(a i ) = wTa i = b i , and to find a w that minimizes errors in sample a i and the desired output b i 37:
$$\mathop{{\rm{\min }}}\limits_{w}{\sum _{i}(f({a}_{i})-{b}_{i})}^{2}+\lambda {\Vert w\Vert }^{2}$$(2)$$w={({A}^{H}A+\lambda I)}^{-1}{A}^{H}b$$(3)where w is a regression coefficient; λ is a regularization parameter; AH is the conjugate transpose of A; I is the unit matrix; and b i is the desired output.
The essential feature of a circulant matrix is the diagonalization with DFT, irrespective of the values generated37:
$$A={F}^{H}diag(\hat{a})F$$(4)where F is a discrete Fourier transform matrix; FH is the conjugate transpose of F; \(\hat{a}\) is the Fourier transform of a; and diag is the vector diagonalization.
Substituting eq. (4) into eq. (3):
$$\widehat{w}=\frac{\hat{a}\odot \widehat{b}}{{\hat{a}}^{\ast }\odot \hat{a}+\lambda }$$(5)where \(\hat{w}\) is the Fourier transform of w; and \(\hat{b}\) is the Fourier transform of b.
-
(3)
Building the kernel ridge regression classifier
Segmentation of the cabbage butterfly target from the background is a nonlinear problem, so it was transformed into a linear problem, and the solution was then determined. Here, a kernel function was used for the nonlinear transformation of the target tracking algorithm. A nonlinear mapping function ψ(c) transformed the sample characteristic space into a higher dimensional space so that the nonlinear regression became linearly separable38. Thus, the regression coefficient w and the regression function change to
$$w=\sum _{i}{\alpha }_{i}\psi ({a}_{i})$$(6)$$f(c)={w}^{T}\psi (c)={\alpha }^{T}\kappa (c)$$(7)where κ(c) is a kernel function, κ(c) = ψ(c) · ψ(a i ) mainly including Gaussian, polynomial and linear kernels; and α is a new regression parameter (the solution parameter is changed from w to α).
From eq. (7), we can see that f(c) is a nonlinear function of c, but the linear function on κ(c) uses the optimization method of the linear function to solve α, with the result that:
$$\alpha ={(K+\lambda I)}^{-1}b$$(8)$$\hat{\alpha }=\frac{\hat{b}}{\hat{k}+\lambda }$$(9)where K is the kernel function matrix (circulant matrix), K ij = κ(c i , a j ); and k is the first row of the constituent elements of K.
-
(4)
Fast-tracking of cabbage butterfly targets
After designing a classifier, the next frame image was sampled in the predicted area and the sample was cyclically shifted to obtain a new circulant matrix as the sample set for detection. The sample with the largest response f(c i ) was selected as the detected new butterfly target area, and the target displacement was judged from the relative displacement from the predicted area.
-
(5)
Re-detecting the target when it was lost When a target was lost, in the tracking process, the position of the tracking box was substantially unchanged (Fig. 2b1–d1). Thus, threshold T and distance S between the target centroid of the current frame and the target centroid of the previous frame were set for comparison to determine whether the target was lost. If S was less than T, the target was considered lost, and the new target centroid was obtained using the background subtraction method to serve as the target centroid of the current frame. The schematic diagram for reacquiring the target is shown in Fig. 10.
By repeating these steps frame-by-frame, we were able to achieve reliable target tracking. The time from frame acquisition to insect position was approximately 1.3 s.
Acquisition of 3D trajectory parameters
The key to obtaining a 3D trajectory is to detect the coordinates of the target centroid of the butterfly. In this paper, two cameras were arranged orthogonally. Therefore, the 3-dimensional coordinates can be obtained only using the calibrated camera internal parameters and the orthogonal relationship between the two cameras. The coordinates x, y and z corresponded to the directions of the wind-tunnel length, depth and height, respectively. The top left of the wind tunnel was the coordinate origin. The parameters of the 3D trajectory were obtained as follows:
-
(1)
Side-view and top-view videos were tracked frame-by-frame using the algorithms described above. The centre coordinates of the rectangular box obtained were used as the two-dimensional centroid coordinates (x n , z n ) and (x n , y n ) of the butterfly in the corresponding side-view and top-view videos. The horizontal axis of the two image sequences was the same. There may be a deviation, but the average can be used instead of the abscissa, y n corresponds to the y coordinate of 3D, and z n corresponds to the z coordinate of 3D, so the 3D spatial coordinates are (x n , y n , z n ).
-
(2)
The internal parameters of the cameras obtained during calibration were used to convert the 3D coordinates in pixel units to spatial coordinates in physical units, and the 3D coordinates of the target in each image frame were saved as.txt files for subsequent processing.
References
Mo, S. S., Zhao, D. X. & Chen, Q. Advances on Relationships Between Plant Volatiles and Insect Behavior. Chinese Journal of Tropical Agriculture. 26(6), 84–89 (2006).
Giunti, G. et al. Parasitoid learning: Current knowledge and implications for biological control. Biological Control. 90, 208–219 (2015).
Zeng, C. et al. Effect of Infochemicals on Insect Behavior and Application inBiological Control. Jiangxi Plant Protection 32(2), 63–67 (2009).
Wu, W. et al. Butterfly catastrophe model for wheat aphid population dynamics: Construction, analysis and application. Ecological Modelling. 288(5), 55–61 (2014).
Senior, A. M., Lihoreau, M., Buhl, J., Raubenheimer, D. & Simpson, S. J. Social network analysis and nutritional behavior: an integrated modeling approach. Frontiers in Psychology. 7, 18 (2016).
Nandi, A. K., Sumana, A. & Bhattacharya, K. Social insect colony as a biological regulatory system: modelling information flow in dominance networks. Journal of the Royal Society Interface. 11(101), 20140951 (2014).
Raguso, R. A., Boland, W., Hartmann, T., Pickett, J. A. & Strack, D. Plant-insect interactions. Phytochemistry. 72(13), 1495 (2011).
Wu, B., Feng, H., Zhao, K. & Wu, K. Application of video trajectory analysis technology in insect behavior research. Plant Protection. 35(2), 1–6 (2009).
Robie, A. A., Seagraves, K. M., Egnor, S. E. & Branson, K. Machine Vision Methods for Analyzing Social Interactions. J Exp Biol. 220(Pt 1), 25–34 (2017).
Risse, B. et al. Comparison of two 3d tracking paradigms for freely flying insects. Eurasip Journal on Image & Video Processing. 2013(1), 57 (2013).
Cheng, X. E., Wang, S. H. & Chen, Y. Q. Estimating Orientation in Tracking Individuals of Flying Swarms. in ICASSP. IEEE 2016, 1496–1500 (2016).
Straw, A. D., Branson, K., Neumann, T. R. & Dickinson, M. H. Multi-camera real-time three-dimensional tracking of multiple flying animals. Journal of the Royal Society Interface. 8(56), 395–409 (2011).
Okubo, A., Bray, D. J. & Chiang, H. C. Use of Shadows for Studying the Three-Dimensional Structure of Insect Swarms. Annals of the Entomological Society of America. 74(1), 48–50 (1981).
Stowers, J. R. et al. Virtual reality for freely moving animals. Nature Methods 14(10), 995 (2017).
Jantzen, B. & Eisner, T. Hindwings are unnecessary for flight but essential for execution of normal evasive flight in lepidoptera. Proc Natl Acad Sci USA 105(43), 16636–16640 (2008).
Lihoreau, M., Ings, T. C., Chittka, L. & Reynolds, A. M. Signatures of a globally optimal searching strategy in the three-dimensional foraging flights of bumblebees. Sci Rep. 6, 30401 (2016).
Hristov, N. I., Betke, M. & Kunz, T. H. Applications of thermal infrared imaging for research in aeroecology. Integrative & Comparative Biology. 48(1), 50 (2008).
Handegard, N. O. et al. The dynamics of coordinated group hunting and collective information transfer among schooling prey. Current Biology Cb. 22(13), 1213 (2012).
Butail, S. & Paley, D. A. Three-dimensional reconstruction of the fast-start swimming kinematics of densely schooling fish. Journal of the Royal Society Interface. 9(66), 77 (2012).
O’Neal, M. E., Landis, D. A., Rothwell, E., Kempel, L. & Reinhard, D. Tracking insects with harmonic radar: a case study. American Entomologist. 50(4), 212–218 (2004).
Xu, P. L., Han, J. & Tong, J. F. Preliminary studies on an automated 3D fish tracking method based on a single video camera. Journal of Fisheries of China 36(4), 623–628 (2012).
Hardie, J. & Powell, G. Video Analysis of Aphid Flight Behaviour. Computers & Electronics in Agriculture 35(2), 229–242 (2002).
Risse, B., Mangan, M., Webb, B. & Pero, L. D. Visual Tracking of Small Animals in Cluttered Natural Environments Using a Freely Moving Camera. IEEE International Conference on Computer Vision Workshop IEEE Computer Society 2017, 2840–2849 (2017).
Yin, J. L. Occurrence law of cabbage butterfly in China and its identification and prevention. Plant Diseases and Pests. 1(2), 21–25 (2010).
Dalal, N. & Triggs, B. Histograms of Oriented Gradients for Human Detection. IEEE Computer Society Conference on Computer Vision & Pattern Recognition IEEE Computer Society 2005, 886–893 (2005).
Zhang, K., Zhang, L. & Yang, M. H. Real-Time Compressive Tracking. European Conference on Computer Vision 7574, 864–877 (2012).
Zhang, K., Zhang, L., Yang, M. H. & Zhang, D. Fast tracking via spatio-temporal context learning. Computer Science (2013).
Zhang, W. Z., Ji, J. G., Jing, Z. Z., Jing, W. F. & Zhang, Y. Adaptive Real-Time Compressive Tracking. International Conference on Network and Information Systems for Computers, 236–240 (2016).
Li, G. et al. Nano-Manipulation Based on Real-Time Compressive Tracking. IEEE Transactions on Nanotechnology. 14(5), 837–846 (2015).
Pang, X., Ma, D., Dong, M., Chen, Z. & Yu, Z. Design and implementation of the face tracking algorithm based on spatio-temporal context learning. Journal of Information & Computational Science. 11(17), 6319–6326 (2014).
Liu, W., Li, J., Shi, Z., Chen, X. & Chen, X. Oversaturated part-based visual tracking via spatio-temporal context learning. Applied Optics. 55(25), 6960 (2016).
Li, J. J., Huang, C. H., Zhou, L. & Yan, F. M. The windtunnel technique. Chinese Journal of Applied Entomology 51(2), 591–596 (2014).
Zhang, Z. Flexible Camera Calibration by Viewing a Plane From Unknown Orientations. The Proceedings of the Seventh IEEE International Conference on Computer Vision 1, 666–673 (1999).
Maimone, A. & Fuchs, H. Encumbrance-Free Tele-Presence System with Real-Time 3D Capture and Display Using Commodity Depth Cameras. IEEE International Symposium on Mixed and Augmented Reality 1416(1), 137–146 (2012).
Henriques, J. F., Rui, C., Martins, P. & Batista, J. High-Speed Tracking with Kernelized Correlation Filters. IEEE Transactions on Pattern Analysis & Machine Intelligence. 37(3), 583–596 (2015).
Piccardi, M. Background subtraction techniques: a review. IEEE International Conference on Systems, Man and Cybernetics. IEEE. 4(2), 3099–3104 (2005).
Rifkin, R., Yeo, G. & Poggio, T. Regularized Least-Squares Classification. Acta Electronica Sinica. 190(1), 93–104 (2003).
Gray, R. M. Toeplitz and circulant Matrices: A Review. Foundations & Trends® in Communications & Information Theory. 2(3), 155–239 (2006).
Acknowledgements
This work was supported by the general program from the Principal Fund of Northwest A&F University: Research and Development of Intelligent Agricultural Information System, grant number: A213020901. The authors thank all team members and staff on the Plant Protection College for their help in this study. The authors also thank the anonymous referees for their helpful comments and suggestions.
Author information
Authors and Affiliations
Contributions
Y.Y. Guo, D.J. He and C. Liu conceived the idea, and performed the numerical calculations and theoretical analysis. Y.Y. Guo and C. Liu performed the experiments described in this manuscript. Y.Y. Guo and C. Liu wrote the manuscript text. All authors reviewed the manuscript for important intellectual content and approved its final version prior to submission.
Corresponding author
Ethics declarations
Competing Interests
The authors declare no competing interests.
Additional information
Publisher's note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Guo, Yy., He, Dj. & Liu, C. Target tracking and 3D trajectory acquisition of cabbage butterfly (P. rapae) based on the KCF-BS algorithm. Sci Rep 8, 9622 (2018). https://doi.org/10.1038/s41598-018-27520-z
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-018-27520-z
This article is cited by
-
Moving object tracking in clinical scenarios: application to cardiac surgery and cerebral aneurysm clipping
International Journal of Computer Assisted Radiology and Surgery (2019)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.