Next Article in Journal
Experimental Comparison between Ethanol and Hexane as Solvents for Oil Extraction from Peanut Press Cake
Previous Article in Journal
Recovery of Protein from Industrial Hemp Waste (Cannabis sativa, L.) Using High-Pressure Processing and Ultrasound Technologies
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Tiny Object Detection Approach for Maize Cleaning Operations

1
Beijing Advanced Innovation Center for Food Nutrition and Human Health, College of Engineering, China Agricultural University, 17 Qinghua Donglu, P.O. Box 50, Beijing 100083, China
2
School of Electronic and Computer Engineering, Peking University, Shenzhen 518055, China
3
Beijing Key Laboratory of Functional Food from Plant Resources, College of Food Science and Nutritional Engineering, China Agricultural University, Beijing 100083, China
4
School of Chemical Engineering, University of New South Wales, Sydney, NSW 2052, Australia
*
Authors to whom correspondence should be addressed.
Foods 2023, 12(15), 2885; https://doi.org/10.3390/foods12152885
Submission received: 25 June 2023 / Revised: 10 July 2023 / Accepted: 27 July 2023 / Published: 29 July 2023
(This article belongs to the Section Food Engineering and Technology)

Abstract

:
Real-time and accurate awareness of the grain situation proves beneficial for making targeted and dynamic adjustments to cleaning parameters and strategies, leading to efficient and effective removal of impurities with minimal losses. In this study, harvested maize was employed as the raw material, and a specialized object detection network focused on impurity-containing maize images was developed to determine the types and distribution of impurities during the cleaning operations. On the basis of the classic contribution Faster Region Convolutional Neural Network, EfficientNetB7 was introduced as the backbone of the feature learning network and a cross-stage feature integration mechanism was embedded to obtain the global features that contained multi-scale mappings. The spatial information and semantic descriptions of feature matrices from different hierarchies could be fused through continuous convolution and upsampling operations. At the same time, taking into account the geometric properties of the objects to be detected and combining the images’ resolution, the adaptive region proposal network (ARPN) was designed and utilized to generate candidate boxes with appropriate sizes for the detectors, which was beneficial to the capture and localization of tiny objects. The effectiveness of the proposed tiny object detection model and each improved component were validated through ablation experiments on the constructed RGB impurity-containing image datasets.

1. Introduction

The performance of the cleaning system is of paramount importance, as it is a critical step in combined harvesting. It exerts a direct influence on the loss rate and impurity content of grain kernels, while also playing a vital role in ensuring efficient drying, quality-guaranteed transportation and safe storage of the harvested grains [1,2]. The cleaning principles are often based on the significant differences in shape, specific gravity, volume and density, etc. among normal kernels, damaged ones, rotten ones and impurities. The principle involves the actions of throwing, blowing, transporting and screening the mixture, through the coupling of multiphysics [3,4]. For this purpose, Krzysiak et al. proposed a rotary cleaning device suitable for wheat grains and analyzed the influences of the inclination angle of the sieve drum on the quality of the processes. Three metrics were used to assess the cleaning results: the coefficients of plump grain mass separation and fine impurity separation and the overall coefficient of cleaning effectiveness [5]. On the premise of sufficiently considering the air-flow uniformity in the technological processes of grain air-screen cleaning, Aldoshin et al. installed an additional fine-mesh sieve between the lower sieve and inclined bottom of the cleaning system to isolate the small impurities [6]. The countersunk screen designed by Wang et al. was utilized in the cleaning device so that the maize particles could move towards the screen holes, which increased the penetration possibility of maize kernels [7]. These contributions optimized the mechanical structure of the equipment based on the physical characteristics of different varieties of crops, which could improve the cleaning efficiency to a certain extent. However, the parameter setting process still relied on operators to manually track and supervise the entire cleaning processes, relying on their experience. This approach was evidently time-consuming and expensive. Therefore, the implementation of real-time grain situation awareness can offer valuable guidance and reference for the adaptive and dynamic adjustments of cleaning strategies, addressing these limitations.
As a kind of information carrier, images can provide research foundation and data resources for numerous fields [8]. Based on the hyperspectral, a rapid and cost-effective way was proposed to generate records of sediment properties and composition at the micrometer-scale [9]. Yuan et al. designed a compact proxy-based deep learning framework to perform highly accurate hyperspectral image classification with superb efficiency and robustness [10]. In addition, the maize kernel images supplied information support to the classification tasks of planted cultivars [11]. Object detection, through the integration of object localization and recognition techniques, enables accurate regression of bounding box coordinates and identification of object categories. This approach was widely applied in the domains of face recognition, medical image processing and agricultural product processing, etc. [12,13]. The traditional object detection algorithms devised the corresponding feature extraction modules for different kinds of objects to be detected, so they were more pertinent and interpretable [14]. Nevertheless, these methods exhibited limitations in terms of robustness and scalability. This was primarily attributed to their heavy dependence on manually crafted features and the need for extensive parameter adjustments [15,16]. Relying on the powerful feature extraction capabilities, deep learning-based object detection technologies could adaptively capture the deep semantic information of images through the multi-structured network models, thus significantly improving the efficiency and accuracy of detection tasks [17,18]. Wang et al. constructed Pest24, which was a large-scale multi-target standardized dataset of agricultural pests. On this basis, they utilized a variety of deep learning-based object detection models to detect the pests in the datasets, which achieved encouraging results in the real-time monitoring of field crop pests [19]. Based on the deep neural network frameworks, Bazame et al. proposed a computer vision system with the object detection algorithms as the core to measure the ripeness of Coffea Arabica fruits on the branches, thereby demonstrating the potential in objectively guiding the decision-making of the coffee farmers [20]. As one of the classic two-stage detectors, Faster Region Convolutional Neural Network (Faster R-CNN) could be used to identify the weeds in cropping areas and detect the cracks in bridges [21,22]. Compared with the representative contributions of the one-stage algorithms You Only Look Once (YOLO) [23,24] and Single-Shot Multi-Box Detector (SSD) [25,26], due to the initial generation of the candidate box and the further adjustment of the bounding box, the detection accuracy of the two-stage models was relatively higher, while the one-stage models had faster detection speed. In order to comprehensively detect all kinds of objects with different geometric characteristics in the images, the multi-feature fusion based on convolution, the setting of residual module and the introduction of attention mechanism were exerted to the basic framework of the backbones, which gradually complicated the structure of the feature learning networks. A feature pyramid architecture AugFPN was designed by Guo et al. to realize the fusion of multi-scale image features; the ResNet50 and MobileNet-v2 were employed as the backbone respectively to demonstrate its effectiveness on the MS COCO detection datasets [27]. For the purpose of capturing the rich context features of the image to be detected, Zhao et al. proposed a context-aware pyramid feature extraction module (CPFE) for the high-level feature maps. At the same time, the enhancement of contextual features and the refinement of boundary information (contained in the low-level feature maps) were realized with the aids of the channel-wise attention and spatial attention, and the final matrix was generated through feature fusion [28].
Limited by the lack of visual feature information caused by fewer pixels, the detection accuracy of tiny objects was relatively low [29,30]. In addition, the information loss during the forward propagation of the networks, the uneven distribution of the sample quantities and the setting of anchor boxes, etc. could all affect the final object classification and the coordinate regression results [31]. Therefore, tiny object detection has become one of the most challenging tasks in computer vision [32]. In allusion to the smaller size and higher density of the objects in the aerial images, Wei et al. proposed an efficacious calibrated-guidance (CG) scheme to intensify the channel communication through the feature transformer fashion, which could adaptively determine the calibration weights for each channel based on the global feature affinity correlations [33]. The concept of fusion factor was proposed by Gong et al. to control the information that delivered from deep layers to the shallow ones, which adapted the feather pyramid network (FPN) to tiny object detection, and its effective value was estimated based on a statistical method [34]. By use of the improved K-means clustering algorithm, Wang et al. generated suitable anchors for the traffic sign datasets and then promoted the detection recall rate and target positioning accuracy of the proposed lightweight recognition algorithm, which was improved on the basis of YOLOv4–Tiny [35]. Similarly, Cheng et al. adjusted the sizes and aspect ratios of the anchors and label frames according to the dimensioning of the tiny objects in the capacitance samples, thereupon achieving effective training of the network in the candidate areas [36]. In addition, different data augmentation strategies had been testified to expand and enrich the scale and diversity of the datasets, thus enhancing the robustness and generalization ability of the detection models [37].
Maize (Zea mays L.) is a traditional global grain crop known for its strong environmental adaptability, high nutritional value and diverse applications. It serves as a crucial feed source in the animal husbandry and breeding industry [38,39]. As a consequence, the rational utilization of maize production capacity had momentous strategic significance for the development of national economy and the promotion of agricultural technology [40]. However, during the harvesting process, maize kernels often become contaminated with a variety of impurities, including rotten and damaged kernels, cobs, husks, gravel and clods. These result in resource waste and pose safety hazards during subsequent processing and storage [41]. Hence, this study focused on harvested maize as the primary material and introduced a tiny object detection network specifically designed for impurity-containing maize images. This network enabled real-time identification and analysis of impurity categories and their distribution during cyclic cleaning operations. By utilizing the feedback on grain conditions during impurity removal, targeted and dynamic adjustments of parameters and strategies could be made to enhance the efficiency and minimize losses in the maize cleaning process. The major contribution points are summarized as follows:
(1)
The EfficientNetB7 was introduced as the backbone of the feature learning network and a tiny object detection network was proposed for analyzing the categories and distribution of impurities in the harvested maize based on the classic contribution Faster R-CNN;
(2)
The designed cross-stage feature integration mechanism was able to fuse the semantic descriptions and spatial information of feature matrices from different hierarchies through continuous convolution and upsampling operations;
(3)
Based on the geometric properties of the objects to be detected and the resolution of images, the adaptive region proposal network was able to generate appropriate anchor boxes for the detectors;
(4)
The impurity-containing maize datasets was constructed to measure the comprehensive performance of the end-to-end tiny object detection network.

2. Materials and Methods

The variety of maize in this research was Wannuo 2000, which was purchased from Shangzhuang experimental station of China Agricultural University (Beijing, China). The moisture content was about 25% and the samples were stored in the refrigerator at 4 °C. Figure 1 revealed the overall framework of tiny object detection for the impurity-containing maize images and it could be divided into three parts, which were image feature learning network, adaptive region proposal network and classification and regression layers of candidate box, according to the propagation sequence. The image feature learning network was used to extract the global features that contained multi-scale mappings. The adaptive region proposal network performed coordinate adjustment and classification of generated anchor boxes through continuous convolution. Eventually, the obtained high-quality candidate boxes were subjected to specific classification and location regression.

2.1. Image Feature Learning Network

EfficientNet has marked a significant milestone in compound model scaling research by effectively balancing network width, depth and resolution. This balance enables the models to sufficiently capture the feature of images, while simultaneously making them more effortless to be trained [42]. Therefore, based on the fine-grained object detection task, EfficientNetB7 was introduced as the backbone of the image feature learning network. In the feed-forward processes of the model, compared with the feature matrices from the deep hierarchies, those from shallow hierarchies contain abundant spatial information but exhibit relatively ambiguous semantic descriptions [43]. Therefore, the cross-stage integration mechanism shown in Figure 2 was embedded in the basic framework of EfficientNetB7. By performing convolution and upsampling operations on the feature matrices from deep hierarchies and fusing them with those from shallow hierarchies, a cross-stage integrated feature with multi-scale mappings was acquired [44]. Among them, the convolution operations with different receptive fields could simultaneously improve the expression ability of the model and adjust the dimension of the feature matrices.
The feature learning of the impurity-containing maize images was conducted through eight convolution stages; as shown in Table 1, the width and depth of each stage were closely related to the dimension of the original images, which were obtained by multiplying the magnification factor corresponding to the resolution with the parameters of the baseline (EfficientNetB0) [45,46] (where H i W i C i are the dimensions of the feature matrix before operation O i in Figure 2). L i denotes the quantity of repetitions of the operation O i , i.e., the depth of stage i . The rightmost column lists the kernel sliding strides of the first convolutions in the repeated operations for each stage. Compared with the subsequent stages, the operations in the first stage adapted a traditional convolution with a kernel size of 3*3. Furthermore, the incorporation of BN (Batch Normalization) layers and Swish activation functions effectively addressed gradient vanishing and exploding issues during back-propagation, thereby enhancing the model’s generalization capability [47,48].
The detailed structure of MBConv in Stages 2–8 is exhibited in Figure 3, which shows the close layouts with the MobileNetV3 blocks [49]. The first convolution operation, with a kernel size of 1*1, was utilized to increase the dimension of the input feature matrix. MBConv6 in Table 1 signified that the scale of convolution kernels was 6 times that of the input feature channels, while MBConv1 indicated that there was no 1*1 convolution operation of dimensionality enhancement in the current stage. Similarly, k3*3 and k5*5 were the convolution kernel sizes for the depthwise convolution in the corresponding stage [50]. The utilization of depthwise convolution effectively reduced the quantity of network parameters, which meant less memory consumption and faster computing speed. The padding of the 3*3 and 5*5 kernels were 1 and 2 respectively, which meant that the matrix size and channel quantity of the feature did not change after the planar convolution with a stride of 1. Furthermore, it was a necessary and sufficient condition for the existence of shortcut connections and dropout layers that the input and output feature matrices in Figure 3 had the same dimensionality.
The SE block, depicted in Figure 4, serves as a lightweight plug-and-play channel attention mechanism. It compresses features in the spatial dimension using squeeze, excitation and reweight processes. Consequently, based on the correlation among channels, new weights were generated for them and exerted to the input matrices in turn [51]. By virtue of its cross-channel interaction capability, the SE block was able to selectively enhance the more significant features through learning global information [52]. In this case, global average pooling was applied to each channel of the input matrices and Swish and Sigmoid activation functions were utilized for the two one-dimension fully connected layers. The global average pooling downsampled the matrices to the specified size and the activation functions were able to improve the nonlinearity of the network. Different from the SE block in the image classification tasks, the quantity of neurons in the channel-reduced FC Layer 1 was a quarter of the feature width (the quantity of channel) input to the current MBConv. The scale of FC Layer 2 was the same as the feature width after depthwise convolution. With regard to the cross-stage integration mechanism, the convolution operation with a kernel size of 3*3 was exploited to improve the local perception competence of the model and the quantity of 1*1 convolution kernel could flexibly adjust the stacking of channels. Moreover, the double upsampling processes after feature integration were implemented through bilinear interpolation [53].

2.2. Adaptive Region Proposal Network (ARPN)

ARPN (Adaptive Region Proposal Network) leverages the distribution characteristics and geometric properties of impurities and maize kernels to classify and adjust the coordinates of generated anchors through continuous convolution. Specifically, the convolution kernel and sliding window with the size of 3*3 were employed to sequentially traverse each position of the cross-stage integrated feature, thereby obtaining the intermediate layer (in the same size and dimension as the cross-stage integrated feature) and generating initial anchor boxes in the meantime [54]. In order to more completely and accurately cover the various objects in the impurity-containing maize images, the aspect ratios were set to 1:1, 1:2 and 2:1, as shown in Figure 5, and the area scales were 642, 1282 and 2562, which could correspondingly generate about 50K (75*75*9) anchor boxes on each original image [55]. These were determined by conducting experiments on different categories of target contours in the impurity-containing maize images. Eventually, the classification and coordinate regression parameters of each anchor box were attained by concatenating two convolution operations with a kernel size of 1*1. The classification information included the probability of foreground (with object) and background (without object), and the regression parameters were oriented towards the center coordinates, width and height of the anchor boxes, so the quantities of convolution kernels were 2 n and 4 n , respectively.
In the end-to-end training processes based on the back propagation and stochastic gradient descent, the positive anchor samples were defined as (i) anchors that had IoU (intersection-over-union) overlaps higher than 0.7 with any ground-truth box, or (ii) anchors with the highest IoU ratio with the ground-truth boxes. In contrast, the anchor was regarded as a negative sample when the IoU ratios were lower than 0.3 for all ground-truth boxes [56]. Anchors that were neither positive nor negative did not participate in the updates of the networks. In order to avoid the degradation and poor generalization of the model caused by excessive negative samples, the loss of mini-batch was counted by randomly sampling the equal quantity of positive and negative samples [57]. The loss function of ARPN is shown in Equation (1), which was measured through division of the sum of classification loss and regression loss by the quantity of mini-batch. Among them, N m = 256 was the capacity of each mini-batch. If the quantity of positive samples was fewer than 128, then the mini-batch was supplemented with negative samples. i represents the index of an anchor in the current mini-batch, c i denotes the probability that the i th anchor was predicted to be the real label. The ground-truth of c i * is 1 if the current anchor box is a positive sample and 0 for a negative sample [58]. r i * = { r x * , r y * , r w * , r h * } indicates the coordinate regression parameters of the i th anchor corresponding to the ground-truth box and r i = { r x , r y , r w , r h } is the predicted value.
L o s s ( { c i } , { r i } ) = i L c l s ( c i , c i * ) + i c i * L r e g ( r i , r i * ) N m
L c l s = ln ( c i )
L r e g ( r i , r i * ) = i s m o o t h L 1 ( r i r i * )
The classification loss L c l s and regression loss L r e g were separately defined by the logarithmic and cumulative operations of Equations (2) and (3), and the s m o o t h L 1 revealed in Equation (4), was introduced as a robust loss function [31]. Furthermore, Equation (5) describes the relationships among A t t r b e f = { x b e f , y b e f , w b e f , h b e f } , r i and r i * . A t t r b e f and A t t r a f t = { x a f t , y a f t , w a f t , h a f t } are the attribute information of the anchor and the coordinate-adjusted candidate box, respectively. A t t r g t = { x * , y * , w * , h * } is the attribute information of the ground-truth box corresponding to the current anchor [59]. The attribute information included the centre coordinates, width and height. The network parameters were randomly initialized through drawing weights from the zero-mean Gaussian distribution with standard deviation of 0.01. Meanwhile, since the cross-boundary anchor boxes brought about a large number of error terms that were difficult to correct, the anchor boxes with boundary-crossing outliers were ignored in the training processes. Finally, based on the classification information of the generated proposal regions, a non-maximum suppression (NMS) approach was adapted to deal with the highly overlapping candidate boxes; the IoU threshold for NMS was fixed at 0.7 [60].
s m o o t h L 1 ( x ) = 0.5 x 2           i f x < 1 x 0.5       o t h e r w i s e
r x = x a f t x b e f w b e f ,   r y = y a f t y b e f h b e f , r w = ln ( w a f t w b e f ) ,   r h = ln ( h a f t h b e f ) , r x * = x * x b e f w b e f ,   r y * = y * y b e f h b e f , r w * = ln ( w * w b e f ) ,   r h * = ln ( h * h b e f )

2.3. Classification and Regression Layers of Candidate Box

The candidate boxes generated by ARPN served as the regions of interest (ROI) for the follow-up specific classification and location regression. These regions were projected to the cross-stage integrated matrix obtained through the feature learning network [61]. After ROI pooling, the feature matrices were regularized to a consistent size and flattened. Both of the two following fully connection layers had 1024 neurons and were exploited as the inputs of the classifier and regressor. The outputs of the classification layer with softmax included k + 1 outcomes, which, respectively, represented the probability of objects in different varieties. Among them, k was the quantity of object categories and the circumstances of background were also taken into consideration [62]. Similar to the regression layer in the ARPN, the candidate box regressor contained 4 ( k + 1 ) neurons, which could adjust each location through 4 parameters. As shown in Equation (6), P = { P x , P y , P w , P h } represents the center coordinates, width and height of the candidate box, { U x , U y , U w , U h } are the attribute information of the final bounding box output by the tiny object detection network and { f x , f y , f w , f h } are the coordinate regression parameters of k + 1 object categories exported by the regressor.
U x = P w f x ( P ) + P x U y = P h f y ( P ) + P y U w = P w e f w ( P ) U h = P h e f h ( P )
The loss of each candidate box in the tiny object detection network was composed of category loss L c a t and regression loss L l o c , as shown in Equation (7) [63]. q = { q 0 , q 1 , . . . , q k } is the softmax probability distribution predicted by the classifier, v denotes the real category label corresponding to the object in the candidate box and the category loss L c a t is measured through Equation (8). b g = { b x g , b y g , b w g , b h g } is the coordinate regression parameters predicted by the regressor for the corresponding category g and s = { s x , s y , s w , s h } is that of the candidate box for the corresponding ground-truth object. The regression loss L l o c is obtained through Equation (9) and α is the hyper-parameter utilized to balance the two losses [64]. Additionally, the values of the Iverson bracket indicator function [ v > 0 ] are 1 when v > 0 ; otherwise, it is 0. Compared with the basic Faster R-CNN network, in order to capture the multi-hierarchy features in the fine-grained impurity-containing maize images, the proposed model replaced the original backbone ZFNet with EfficientNetB7 and embedded a cross-stage feature integration mechanism. At the same time, the area scale of the anchor box in the adaptive region proposal network was also adjusted accordingly for the tiny object detection tasks.
L o s s ( q , v , b g , s ) = L c a t ( q , v ) + α [ v > 0 ] L l o c ( b g , s )
L c a t ( q , v ) = ln q v
L l o c ( b g , s ) = i { x , y , w , h } s m o o t h L 1 ( b i g s i )

3. Results and Discussion

The image acquisition modules, as illustrated in Figure 6C, were positioned at the feed port and discharge port of the cleaning equipment. Their purpose was to capture images and provide the necessary data for the end-to-end tiny object detection network. The multiphysics-coupled cleaning equipment removed impurities with a certain mass through two screens with different sizes and shapes, while relatively light impurities were removed by means of the air separation unit. The industrial cameras (BFS-U3-51S5C-C, LUSTER LightTech Co., LTD., Beijing, China) with global shutters were designed and manufactured by FLIR and the supporting development tool was Spinnaker 2.6.0.160 (FLIR Systems, Wilsonville, OR, USA). The ring lights (RI15045-W) developed by OPT-Machine Vision were utilized to ensure the uniformity of imaging brightness. The resolution of RGB impurity-containing maize images were standardized to 600*600, which was beneficial to the feature learning network. In order to avoid the uncertain convergence direction and over-fitting conditions of the entire models caused by insufficient quantity of samples, data augmentation approaches were exerted to expand the datasets [65]. Specifically, we performed rotations, vertical mirror symmetry, horizontal mirror symmetry, adjustments of contrast and brightness, insertions of Gaussian noise and salt and pepper noise on the 1000 original images, as shown in Figure 7, and divided the impurity-containing maize datasets into a training set and test set, according to the ratio of 3:1 [66]. The adjustment and addition of brightness, contrast and noise enabled the model to have better robustness and greater adaptability to the image acquisition conditions.
The proposed model was regarded as the adaptive region proposal network (ARPN) and remaining detector network, which were trained through alternating optimization [67]. To be specific, the ImageNet-pre-trained models were used to initialize the feature learning network and the end-to-end training was performed on ARPN. Afterwards, the feature learning network was initialized again through the ImageNet-pre-trained models and the detector network was trained based on the proposals generated by ARPN. Eventually, both of the components shared the same convolutional layers and, sequentially, fine-tuned the layers unique to ARPN and the detector network, thereby forming a unified network [68]. The utilized deep learning framework was Pytorch 1.10, the version of Python was 3.7, the vision toolkit was Torchvision 0.11.1 and the strategy of stochastic gradient descent (SGD) was adopted to optimize the processes of parameter updating.
The comprehensive performance of the proposed tiny object detection network was measured through the evaluation indicators applied to the COCO datasets [69,70]. The AP in Figure 8 was the mean value of all mAPs (mean average precisions) when the IoU threshold was between 0.5 and 0.95 (with a value interval of 0.05), which indicated the localization capability of the model [71]. Among them, mAP was the average of the areas under the curves in the PR Graphs that corresponded to each object category. AP50 was the mAP (IoU threshold was 0.5) for all kinds of objects and APs could be defined as the AP for objects with sizes less than 642 [72]. AR100 and AR10 (the range and value interval of IoU threshold were the same as those of AP) separately denoted the average of all mARs (mean average recall) for the n top-scoring detections after NMS (Non-Maximum Suppression) [73]. The mAR was twice the mean value of the areas under the curves in the Recall–IoU Graph corresponding to each object category. Similar to APs, ARs could be defined as the AR for objects with sizes less than 642. Basic-ResNet101 and Basic-EfficientNetB7 represented replacing the backbone of the classic work Faster R-CNN with ResNet101 and EfficientNetB7, respectively. Basic + ARPN and Basic + Cross-stage integration mechanism individually signified the introduction of ARPN and Cross-stage integration on the basis of EfficientNetB7 as the feature learning network. The ablation experiments sequentially demonstrated the effectiveness of each improved component on the basic model, thereby revealing the superiority of the proposed model (ours), which exhibited stronger performance in various evaluation indicators (Figure 8A) [74,75]. The selection of EfficientNetB7 could significantly improve the tiny object detection capability, while the cross-stage integration mechanism and ARPN also had strong adaptability. Since the quantity of objects in each image was mostly no more than 10, the results of AR100 and AR10 were comparable. In addition, Figure 8B shows the average detection precision of various objects for different models when the IoU threshold was 0.5, reflecting the better equilibrium of the proposed model. Among them, the relatively lower average precision of the category Damaged might be caused by the similar appearance of partially damaged maize to normal kernels, and this also explained the higher average precision of category Weeds due to their more prominent profiles. Figure 9 exhibits the object detection outcomes on part of the images in the test datasets, including the predicted category and confidence score of the objects. The overall performance was consistent with the data pattern in Figure 8, which could reflect the object distribution in the maize cleaning processes.

4. Conclusions

In this study, we proposed a tiny object detection network specifically designed for harvested maize, to accurately identify and analyze the categories and distribution of impurities during the cleaning process. Firstly, on the basis of EfficientNetB7, a cross-stage integration mechanism was introduced to obtain feature matrices that contained spatial information and semantic descriptions. Then, the appropriate candidate boxes were generated through ARPN. Eventually, the classification and regression layers output the final detection results after adjusting the attribute information. The superiority of the proposed approach over the basic model was demonstrated through the ablation experiments on the constructed impurity-containing maize datasets and the effectiveness of each introduced component was illustrated as well. The introduction of the components individually or simultaneously enabled the model to have a stronger detection capability, which proved the compatibility between them. In addition, the proposed tiny object detection network also had better performance in actual continuous maize cleaning operations.

5. Future Direction

By virtue of the distribution information of various objects derived in the maize cleaning operations, the current study could provide significant references for the qualitative production. In the future, the structural design of the detection network will be optimized according to the comprehensive characteristics of more types of crops, so that it can be applied to more scenarios of cleaning operations.

Author Contributions

Conceptualization, Methodology, Software, Data curation, Formal analysis, Writing—original draft, Investigation, H.Y.; Visualization, Methodology, Z.L.; Data curation, Software, W.L.; Methodology, Investigation, W.G.; Writing—review & editing, Resources, Validation, Funding acquisition, Project administration, D.L.; Software, Visualization, Supervision, L.W.; Methodology, Conceptualization, M.W.; Visualization, Supervision, Writing—review & editing, Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program of China (2021YFD2100600).

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wu, J.; Tang, Q.; Mu, S.L.; Yang, X.X.; Jiang, L.; Hu, Z.C. Design and Test of Self-Leveling System for Cleaning Screen of Grain Combine Harvester. Agriculture 2023, 13, 377. [Google Scholar] [CrossRef]
  2. Liang, Y.Q.; Tang, Z.; Zhang, H.; Li, Y.M.; Ding, Z.; Su, Z. Cross-flow fan on multi-dimensional airflow field of air screen cleaning system for rice grain. Int. J. Agric. Biol. Eng. 2022, 15, 223–235. [Google Scholar] [CrossRef]
  3. Badretdinov, I.; Mudarisov, S.; Lukmanov, R.; Permyakov, V.; Ibragimov, R.; Nasyrov, R. Mathematical modeling and research of the work of the grain combine harvester cleaning system. Comput. Electron. Agric. 2019, 165, 104966. [Google Scholar] [CrossRef]
  4. Tang, H.; Xu, C.S.; Zhao, J.L.; Wang, Y.J. Screening and impurity removal device to improve the accuracy of moisture content detection device for rice. Int. J. Agric. Biol. Eng. 2022, 15, 113–123. [Google Scholar] [CrossRef]
  5. Krzysiak, Z.; Samociuk, W.; Skic, A.; Bartnik, G.; Zarajczyk, J.; Szmigielski, M.; Dziki, D.; Wierzbicki, S.; Krzywonos, L. Effect of sieve drum inclination angle on wheat grain cleaning in a novel rotary cleaning device. Trans. Asabe 2017, 60, 1751–1758. [Google Scholar] [CrossRef]
  6. Aldoshin, N.; Didmanidze, O.; Lylin, N.; Mosyakov, M. Work improvement of air-and-screen cleaner of combine Harvester. In Proceedings of the 18th International Scientific Conference Engineering for Rural Development, Jelgava, Latvia, 22–24 May 2019; pp. 100–104. [Google Scholar]
  7. Wang, L.J.; Chai, J.; Wang, H.S.; Wang, Y.S. Design and performance of a countersunk screen in a maize cleaning device. Biosyst. Eng. 2021, 209, 300–314. [Google Scholar] [CrossRef]
  8. Vasefi, F.; MacKinnon, N.; Farkas, D.L. Chapter 16—Hyperspectral and Multispectral Imaging in Dermatology. In Imaging in Dermatology; Hamblin, M.R., Avci, P., Gupta, G.K., Eds.; Academic Press: Boston, MA, USA, 2016; pp. 187–201. [Google Scholar]
  9. Amann, B.; Butz, C.; Rein, B.; Tylmann, W. Hyperspectral imaging: A novel, non-destructive method for investigating sub-annual sediment structures and composition. Past Glob. Changes Mag. 2014, 22, 10–11. [Google Scholar] [CrossRef] [Green Version]
  10. Yuan, Y.; Wang, C.; Jiang, Z. Proxy-Based Deep Learning Framework for Spectral–Spatial Hyperspectral Image Classification: Efficient and Robust. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  11. Alimohammadi, F.; Rasekh, M.; Sayyah, A.H.A.; Abbaspour-Gilandeh, Y.; Karami, H.; Sharabiani, V.R.; Fioravanti, A.; Gancarz, M.; Findura, P.; Kwasniewski, D. Hyperspectral imaging coupled with multivariate analysis and artificial intelligence to the classification of maize kernels. Int. Agrophysics 2022, 36, 83–91. [Google Scholar] [CrossRef]
  12. Tong, K.; Wu, Y.Q. Deep learning-based detection from the perspective of small or tiny objects: A survey. Image Vis. Comput. 2022, 123, 104471. [Google Scholar] [CrossRef]
  13. Shuang, K.; Lyu, Z.H.; Loo, J.; Zhang, W.T. Scale-balanced loss for object detection. Pattern Recognit. 2021, 117, 107997. [Google Scholar] [CrossRef]
  14. Felzenszwalb, P.F.; Girshick, R.B.; McAllester, D.; Ramanan, D. Object Detection with Discriminatively Trained Part-Based Models. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1627–1645. [Google Scholar] [CrossRef] [Green Version]
  15. Shi, Y.Y.; Li, J.Y.; Yu, Z.Y.; Li, Y.; Hu, Y.P.Q.; Wu, L.S. Multi-Barley Seed Detection Using iPhone Images and YOLOv5 Model. Foods 2022, 11, 3531. [Google Scholar] [CrossRef]
  16. Shirpour, M.; Khairdoost, N.; Bauer, M.A.; Beauchemin, S.S. Traffic Object Detection and Recognition Based on the Attentional Visual Field of Drivers. IEEE Trans. Intell. Veh. 2023, 8, 594–604. [Google Scholar] [CrossRef]
  17. Wang, J.; Zhang, T.J.; Cheng, Y.; Al-Nabhan, N. Deep Learning for Object Detection: A Survey. Comput. Syst. Sci. Eng. 2021, 38, 165–182. [Google Scholar] [CrossRef]
  18. Nalepa, J. Recent Advances in Multi- and Hyperspectral Image Analysis. Sensors 2021, 21, 6002. [Google Scholar] [CrossRef]
  19. Wang, Q.J.; Zhang, S.Y.; Dong, S.F.; Zhang, G.C.; Yang, J.; Li, R.; Wang, H.Q. Pest24: A large-scale very small object data set of agricultural pests for multi-target detection. Comput. Electron. Agric. 2020, 175, 105585. [Google Scholar] [CrossRef]
  20. Bazame, H.C.; Molin, J.P.; Althoff, D.; Martello, M. Detection of coffee fruits on tree branches using computer vision. Sci. Agric. 2023, 80, e20220064. [Google Scholar] [CrossRef]
  21. Mu, Y.; Feng, R.L.; Ni, R.W.; Li, J.; Luo, T.Y.; Liu, T.H.; Li, X.; Gong, H.; Guo, Y.; Sun, Y.; et al. A Faster R-CNN-Based Model for the Identification of Weed Seedling. Agronomy 2022, 12, 2867. [Google Scholar] [CrossRef]
  22. Li, R.X.; Yu, J.Y.; Li, F.; Yang, R.T.; Wang, Y.D.; Peng, Z.H. Automatic bridge crack detection using Unmanned aerial vehicle and Faster R-CNN. Constr. Build. Mater. 2023, 362, 129659. [Google Scholar] [CrossRef]
  23. Kim, M.; Jeong, J.; Kim, S. ECAP-YOLO: Efficient Channel Attention Pyramid YOLO for Small Object Detection in Aerial Image. Remote Sens. 2021, 13, 4851. [Google Scholar] [CrossRef]
  24. Chen, Y.L.; Xu, H.L.; Zhang, X.J.; Gao, P.; Xu, Z.G.; Huang, X.B. An object detection method for bayberry trees based on an improved YOLO algorithm. Int. J. Digit. Earth 2023, 16, 781–805. [Google Scholar] [CrossRef]
  25. Xu, X.L.; Zhao, J.H.; Li, Y.; Gao, H.H.; Wang, X.H. BANet: A Balanced Atrous Net Improved from SSD for Autonomous Driving in Smart Transportation. IEEE Sens. J. 2021, 21, 25018–25026. [Google Scholar] [CrossRef]
  26. Gu, G.S.; Gan, S.W.; Deng, J.H.; Du, Y.K.; Qiu, Z.W.; Liu, J.J.; Liu, C.; Zhao, J. Automated diatom detection in forensic drowning diagnosis using a single shot multibox detector with plump receptive field. Appl. Soft Comput. 2022, 122, 108885. [Google Scholar] [CrossRef]
  27. Guo, C.; Fan, B.; Zhang, Q.; Xiang, S.; Pan, C. AugFPN: Improving Multi-Scale Feature Learning for Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  28. Zhao, T.; Wu, X.Q.; Soc, I.C. Pyramid Feature Attention Network for Saliency detection. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2019), Long Beach, CA, USA, 15–20 June 2019; pp. 3080–3089. [Google Scholar]
  29. Liu, J.W.; Gu, Y.; Han, S.M.; Zhang, Z.B.; Guo, J.F.; Cheng, X.Q. Feature Rescaling and Fusion for Tiny Object Detection. IEEE ACCESS 2021, 9, 62946–62955. [Google Scholar] [CrossRef]
  30. Chen, G.; Wang, H.T.; Chen, K.; Li, Z.J.; Song, Z.D.; Liu, Y.L.; Chen, W.K.; Knoll, A. A Survey of the Four Pillars for Small Object Detection: Multiscale Representation, Contextual Information, Super-Resolution, and Region Proposal. IEEE Trans. Syst. Man Cybern. Syst. 2022, 52, 936–953. [Google Scholar] [CrossRef]
  31. Liu, Y.; Sun, P.; Wergeles, N.; Shang, Y. A survey and performance evaluation of deep learning methods for small object detection. Expert Syst. Appl. 2021, 172, 114602. [Google Scholar] [CrossRef]
  32. Xiao, J.S.; Guo, H.W.; Zhou, J.; Zhao, T.; Yu, Q.Z.; Chen, Y.H.; Wang, Z.Y. Tiny object detection with context enhancement and feature purification. Expert Syst. Appl. 2023, 211, 118665. [Google Scholar] [CrossRef]
  33. Wei, Z.Q.; Liang, D.; Zhang, D.; Zhang, L.Y.; Geng, Q.X.; Wei, M.Q.; Zhou, H.Y. Learning Calibrated-Guidance for Object Detection in Aerial Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 2721–2733. [Google Scholar] [CrossRef]
  34. Gong, Y.Q.; Yu, X.H.; Ding, Y.; Peng, X.K.; Zhao, J.; Han, Z.J. Effective Fusion Factor in FPN for Tiny Object Detection. In Proceedings of the 2021 IEEE Winter Conference on Applications of Computer Vision (WACV 2021), Waikoloa, HI, USA, 3–8 January 2021; pp. 1159–1167. [Google Scholar]
  35. Wang, L.M.; Zhou, K.; Chu, A.L.; Wang, G.B.; Wang, L.Z. An Improved Light-Weight Traffic Sign Recognition Algorithm Based on YOLOv4-Tiny. IEEE Access 2021, 9, 124963–124971. [Google Scholar] [CrossRef]
  36. Cheng, C.; Dai, N.; Huang, J.; Zhuang, Y.H.; Tang, T.; Liu, L.L. Capacitance pin defect detection based on deep learning. J. Comb. Optim. 2022, 44, 3477–3494. [Google Scholar] [CrossRef]
  37. Zhang, S.F.; Xie, Y.L.; Wan, J.; Xia, H.S.; Li, S.Z.; Guo, G.D. WiderPerson: A Diverse Dataset for Dense Pedestrian Detection in the Wild. Ieee Trans. Multimed. 2020, 22, 380–393. [Google Scholar] [CrossRef] [Green Version]
  38. Orzech, K.; Wanic, M.; Zaluski, D. Gas Exchanges in the Leaves of Silage Maize Depending on the Forecrop and Maize Development Stage. Agronomy 2022, 12, 396. [Google Scholar] [CrossRef]
  39. Wang, L.J.; Zhang, S.; Gao, Y.P.; Cui, T.; Ma, Z.; Wang, B. Investigation of maize grains penetrating holes on a novel screen based on CFD-DEM simulation. Powder Technol. 2023, 419, 118332. [Google Scholar] [CrossRef]
  40. Yu, H.-z.; Li, Z.-z.; Guo, W.-b.; Li, D.; Wang, L.-j.; Wang, Y. An estimation method of maize impurity rate based on the deep residual networks. Ind. Crops Prod. 2023, 196, 116455. [Google Scholar] [CrossRef]
  41. Zhang, N.; Fu, J.; Wang, R.X.; Chen, Z.; Fu, Q.K.; Chen, X.G. Experimental Study on the Particle Size and Weight Distribution of the Threshed Mixture in Corn Combine Harvester. Agriculture 2022, 12, 1214. [Google Scholar] [CrossRef]
  42. Tan, M.X.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; Volume 97. [Google Scholar]
  43. Wang, J.K.; Shao, F.M.; He, X.H.; Lu, G.L. A Novel Method of Small Object Detection in UAV Remote Sensing Images Based on Feature Alignment of Candidate Regions. Drones 2022, 6, 292. [Google Scholar] [CrossRef]
  44. Sun, Z.Q.; Meng, C.N.; Cheng, J.R.; Zhang, Z.Q.; Chang, S.J. A Multi-Scale Feature Pyramid Network for Detection and Instance Segmentation of Marine Ships in SAR Images. Remote Sens. 2022, 14, 6312. [Google Scholar] [CrossRef]
  45. Devi, N.; Sarma, K.K.; Laskar, S. Design of an intelligent bean cultivation approach using computer vision, IoT and spatio-temporal deep learning structures. Ecol. Inform. 2023, 75, 102044. [Google Scholar] [CrossRef]
  46. Hanh, B.T.; Manh, H.V.; Nguyen, N.V. Enhancing the performance of transferred efficientnet models in leaf image-based plant disease classification. J. Plant Dis. Prot. 2022, 129, 623–634. [Google Scholar] [CrossRef]
  47. Li, S.W.; Yang, Y.C. A deep generative framework for data-driven surrogate modeling and visualization of parameterized nonlinear dynamical systems. Nonlinear Dyn. 2023, 111, 10287–10307. [Google Scholar] [CrossRef]
  48. Liu, J.Y.; Song, S.N.; Wang, J.Y.; Balaiti, M.; Song, N.A.; Li, S. Flatness Prediction of Cold Rolled Strip Based on Deep Neural Network with Improved Activation Function. Sensors 2022, 22, 656. [Google Scholar] [CrossRef] [PubMed]
  49. Howard, A.; Sandler, M.; Chu, G.; Chen, L.C.; Chen, B.; Tan, M.X.; Wang, W.J.; Zhu, Y.K.; Pang, R.M.; Vasudevan, V.; et al. Searching for MobileNetV3. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV 2019), Seoul, Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar]
  50. Hu, M.J.; Lin, H.Z.; Fan, Z.M.; Gao, W.J.; Yang, L.; Liu, C.; Song, Q. Learning to Recognize Chest-Xray Images Faster and More Efficiently Based on Multi-Kernel Depthwise Convolution. IEEE Access 2020, 8, 37265–37274. [Google Scholar] [CrossRef]
  51. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E.H. Squeeze-and-Excitation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2011–2023. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Uzen, H.; Turkoglu, M.; Aslan, M.; Hanbay, D. Depth-wise Squeeze and Excitation Block-based Efficient-Unet model for surface defect detection. Vis. Comput. 2023, 39, 1745–1764. [Google Scholar] [CrossRef]
  53. Huang, H.H.; Ge, P. Depth extraction in computational integral imaging based on bilinear interpolation. Opt. Appl. 2020, 50, 497–509. [Google Scholar] [CrossRef]
  54. Sun, K.L.; Wen, Q.F.; Zhou, H.P. Ganster R-CNN: Occluded Object Detection Network Based on Generative Adversarial Nets and Faster R-CNN. IEEE Access 2022, 10, 105022–105030. [Google Scholar] [CrossRef]
  55. Yuan, H.F.; Shao, Y.J.; Liu, Z.H.; Wang, H.Q. An Improved Faster R-CNN for Pulmonary Embolism Detection from CTPA Images. IEEE Access 2021, 9, 105382–105392. [Google Scholar] [CrossRef]
  56. Yi, Z.R.; Yao, D.Y.; Li, G.J.; Ai, J.Y.; Xie, W. Detection and localization for lake floating objects based on CA-faster R-CNN. Multimed. Tools Appl. 2022, 81, 17263–17281. [Google Scholar] [CrossRef]
  57. Peng, C.; Zhao, K.; Lovell, B.C. Faster ILOD: Incremental learning for object detectors based on faster RCNN. Pattern Recognit. Lett. 2020, 140, 109–115. [Google Scholar] [CrossRef]
  58. Kim, E.J.; Park, H.C.; Ham, S.W.; Kho, S.Y.; Kim, D.K. Extracting Vehicle Trajectories Using Unmanned Aerial Vehicles in Congested Traffic Conditions. J. Adv. Transp. 2019, 2019, 9060797. [Google Scholar] [CrossRef] [Green Version]
  59. Yan, J.Q.; Wang, H.Q.; Yan, M.L.; Diao, W.H.; Sun, X.; Li, H. IoU-Adaptive Deformable R-CNN: Make Full Use of IoU for Multi-Class Object Detection in Remote Sensing Imagery. Remote Sens. 2019, 11, 286. [Google Scholar] [CrossRef] [Green Version]
  60. Alganci, U.; Soydas, M.; Sertel, E. Comparative Research on Deep Learning Approaches for Airplane Detection from Very High-Resolution Satellite Images. Remote Sens. 2020, 12, 458. [Google Scholar] [CrossRef] [Green Version]
  61. Guo, Z.M.; Tian, Y.Y.; Mao, W.D. A Robust Faster R-CNN Model with Feature Enhancement for Rust Detection of Transmission Line Fitting. Sensors 2022, 22, 7961. [Google Scholar] [CrossRef] [PubMed]
  62. Huang, L.; Chen, C.; Yun, J.T.; Sun, Y.; Tian, J.R.; Hao, Z.Q.; Yu, H.; Ma, H.J. Multi-Scale Feature Fusion Convolutional Neural Network for Indoor Small Target Detection. Front. Neurorobotics 2022, 16, 881021. [Google Scholar] [CrossRef]
  63. Chaoxia, C.Y.; Shang, W.W.; Zhang, F. Information-Guided Flame Detection Based on Faster R-CNN. IEEE Access 2020, 8, 58923–58932. [Google Scholar] [CrossRef]
  64. Liu, Y.; Wang, S.N. A quantitative detection algorithm based on improved faster R-CNN for marine benthos. Ecol. Inform. 2021, 61, 101228. [Google Scholar] [CrossRef]
  65. Shorten, C.; Khoshgoftaar, T.M.; Furht, B. Text Data Augmentation for Deep Learning. J. Big Data 2021, 8, 101. [Google Scholar] [CrossRef]
  66. Shorten, C.; Khoshgoftaar, T.M. A survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef] [Green Version]
  67. Lee, H.; Eum, S.; Kwon, H. ME R-CNN: Multi-Expert R-CNN for Object Detection. IEEE Trans. Image Process. 2020, 29, 1030–1044. [Google Scholar] [CrossRef] [Green Version]
  68. Jiang, J.; Xu, H.; Zhang, S.C.; Fang, Y.J.; Kang, L. FSNet: A Target Detection Algorithm Based on a Fusion Shared Network. IEEE Access 2019, 7, 169417–169425. [Google Scholar] [CrossRef]
  69. Zhang, Y.Q.; Bai, Y.H.; Ding, M.L.; Ghanem, B. Multi-task Generative Adversarial Network for Detecting Small Objects in the Wild. Int. J. Comput. Vis. 2020, 128, 1810–1828. [Google Scholar] [CrossRef]
  70. Law, H.; Deng, J. CornerNet: Detecting Objects as Paired Keypoints. Int. J. Comput. Vis. 2020, 128, 642–656. [Google Scholar] [CrossRef] [Green Version]
  71. Zhang, Y.Q.; Chu, J.; Leng, L.; Miao, J. Mask-Refined R-CNN: A Network for Refining Object Details in Instance Segmentation. Sensors 2020, 20, 1010. [Google Scholar] [CrossRef] [Green Version]
  72. Cai, Z.W.; Vasconcelos, N. Cascade R-CNN: High Quality Object Detection and Instance Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 1483–1498. [Google Scholar] [CrossRef] [Green Version]
  73. Mohan, H.M.; Rao, P.V.; Kumara, H.C.S.; Manasa, S. Non-invasive technique for real-time myocardial infarction detection using faster R-CNN. Multimed. Tools Appl. 2021, 80, 26939–26967. [Google Scholar] [CrossRef]
  74. Dai, Y.M.; Wu, Y.Q.; Zhou, F.; Barnard, K. Attentional Local Contrast Networks for Infrared Small Target Detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9813–9824. [Google Scholar] [CrossRef]
  75. Zhang, Z.; Lin, Z.; Xu, J.; Jin, W.D.; Lu, S.P.; Fan, D.P. Bilateral Attention Network for RGB-D Salient Object Detection. IEEE Trans. Image Process. 2021, 30, 1949–1961. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The tiny object detection network for impurity-containing maize images.
Figure 1. The tiny object detection network for impurity-containing maize images.
Foods 12 02885 g001
Figure 2. The cross-stage feature integration mechanism.
Figure 2. The cross-stage feature integration mechanism.
Foods 12 02885 g002
Figure 3. The detailed structure of MBConv operation.
Figure 3. The detailed structure of MBConv operation.
Foods 12 02885 g003
Figure 4. The module arrangement of the SE block.
Figure 4. The module arrangement of the SE block.
Foods 12 02885 g004
Figure 5. The area scale and aspect ratio of the anchor box.
Figure 5. The area scale and aspect ratio of the anchor box.
Foods 12 02885 g005
Figure 6. The multiphysics-coupled cleaning equipment (processed by Zhengzhou Wangu Machinery Co., Ltd., Zhengzhou, China) and its components. (A) The overall structure. (B,C) The image acquisition modules (pink frames) and the feed port (yellow frame). (D) The air separation unit (green frame). (E) The main part of the vibrating screen unit (purple frame). (F) The discharge port (cyan frame).
Figure 6. The multiphysics-coupled cleaning equipment (processed by Zhengzhou Wangu Machinery Co., Ltd., Zhengzhou, China) and its components. (A) The overall structure. (B,C) The image acquisition modules (pink frames) and the feed port (yellow frame). (D) The air separation unit (green frame). (E) The main part of the vibrating screen unit (purple frame). (F) The discharge port (cyan frame).
Foods 12 02885 g006
Figure 7. The impurity-containing maize images after data augmentation. (A) The inserted Gaussian noise. (B) The inserted salt and pepper noise.
Figure 7. The impurity-containing maize images after data augmentation. (A) The inserted Gaussian noise. (B) The inserted salt and pepper noise.
Foods 12 02885 g007
Figure 8. (A) Comparison of comprehensive detection performances among the proposed model and baselines. (B) The average detection precision of different models for various objects under the condition that the IoU threshold was 0.5 (corresponding to AP50).
Figure 8. (A) Comparison of comprehensive detection performances among the proposed model and baselines. (B) The average detection precision of different models for various objects under the condition that the IoU threshold was 0.5 (corresponding to AP50).
Foods 12 02885 g008
Figure 9. The object detection results on part of the images in the test datasets.
Figure 9. The object detection results on part of the images in the test datasets.
Foods 12 02885 g009
Table 1. The specific components of the feature learning network.
Table 1. The specific components of the feature learning network.
Stage
i
Operation
Oi
Resolution (Input) Hi*Wi*CiLayers
Li
Strides
(First Layer)
13*3 Conv600*600*312
2MBConv1, k3*3300*300*6441
3MBConv6, k3*3300*300*3272
4MBConv6, k5*5150*150*4872
5MBConv6, k3*375*75*80102
6MBConv6, k5*538*38*160101
7MBConv6, k5*538*38*224132
8MBConv6, k3*319*19*38441
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, H.; Li, Z.; Li, W.; Guo, W.; Li, D.; Wang, L.; Wu, M.; Wang, Y. A Tiny Object Detection Approach for Maize Cleaning Operations. Foods 2023, 12, 2885. https://doi.org/10.3390/foods12152885

AMA Style

Yu H, Li Z, Li W, Guo W, Li D, Wang L, Wu M, Wang Y. A Tiny Object Detection Approach for Maize Cleaning Operations. Foods. 2023; 12(15):2885. https://doi.org/10.3390/foods12152885

Chicago/Turabian Style

Yu, Haoze, Zhuangzi Li, Wei Li, Wenbo Guo, Dong Li, Lijun Wang, Min Wu, and Yong Wang. 2023. "A Tiny Object Detection Approach for Maize Cleaning Operations" Foods 12, no. 15: 2885. https://doi.org/10.3390/foods12152885

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop