Svoboda | Graniru | BBC Russia | Golosameriki | Facebook
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (3,122)

Search Parameters:
Keywords = YOLOv7

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
16 pages, 76531 KiB  
Article
KSL-POSE: A Real-Time 2D Human Pose Estimation Method Based on Modified YOLOv8-Pose Framework
by Tianyi Lu, Ke Cheng, Xuecheng Hua and Suning Qin
Sensors 2024, 24(19), 6249; https://doi.org/10.3390/s24196249 (registering DOI) - 26 Sep 2024
Abstract
Two-dimensional human pose estimation aims to equip computers with the ability to accurately recognize human keypoints and comprehend their spatial contexts within media content. However, the accuracy of real-time human pose estimation diminishes when processing images with occluded body parts or overlapped individuals. [...] Read more.
Two-dimensional human pose estimation aims to equip computers with the ability to accurately recognize human keypoints and comprehend their spatial contexts within media content. However, the accuracy of real-time human pose estimation diminishes when processing images with occluded body parts or overlapped individuals. To address these issues, we propose a method based on the YOLO framework. We integrate the convolutional concepts of Kolmogorov–Arnold Networks (KANs) through introducing non-linear activation functions to enhance the feature extraction capabilities of the convolutional kernels. Moreover, to improve the detection of small target keypoints, we integrate the cross-stage partial (CSP) approach and utilize the small object enhance pyramid (SOEP) module for feature integration. We also innovatively incorporate a layered shared convolution with batch normalization detection head (LSCB), consisting of multiple shared convolutional layers and batch normalization layers, to enable cross-stage feature fusion and address the low utilization of model parameters. Given the structure and purpose of the proposed model, we name it KSL-POSE. Compared to the baseline model YOLOv8l-POSE, KSL-POSE achieves significant improvements, increasing the average detection accuracy by 1.5% on the public MS COCO 2017 data set. Furthermore, the model also demonstrates competitive performance on the CrowdPOSE data set, thus validating its generalization ability. Full article
(This article belongs to the Section Intelligent Sensors)
22 pages, 7527 KiB  
Article
EAAnet: Efficient Attention and Aggregation Network for Crowd Person Detection
by Wenzhuo Chen, Wen Wu, Wantao Dai and Feng Huang
Appl. Sci. 2024, 14(19), 8692; https://doi.org/10.3390/app14198692 (registering DOI) - 26 Sep 2024
Abstract
With the frequent occurrence of natural disasters and the acceleration of urbanization, it is necessary to carry out efficient evacuation, especially when earthquakes, fires, terrorist attacks, and other serious threats occur. However, due to factors such as small targets, complex posture, occlusion, and [...] Read more.
With the frequent occurrence of natural disasters and the acceleration of urbanization, it is necessary to carry out efficient evacuation, especially when earthquakes, fires, terrorist attacks, and other serious threats occur. However, due to factors such as small targets, complex posture, occlusion, and dense distribution, the current mainstream algorithms still have problems such as low precision and poor real-time performance in crowd person detection. Therefore, this paper proposes EAAnet, a crowd person detection algorithm. It is based on YOLOv5, with CBAM (Convolutional Block Attention Module) introduced into the backbone, BiFPN (Bidirectional Feature Pyramid Network) introduced into the neck, and combined with a loss function of CIoU_Loss to better predict the person number. The experimental results show that compared with other mainstream detection algorithms, EAAnet has achieved significant improvement in precision and real-time performance. The precision value of all categories was 78.6%, which was increased by 1.8. Among these, the categories of riders and partially visible person were increased by 4.6 and 0.8, respectively. At the same time, the parameter number of EAAnet is only 7.1M, with a calculation amount of 16.0G FLOPs. Therefore, it is proved that EAAnet has the ability of the efficient real-time detection of the crowd person and is feasible in the field of emergency management. Full article
(This article belongs to the Special Issue Deep Learning for Object Detection)
Show Figures

Figure 1

26 pages, 11965 KiB  
Article
AMFEF-DETR: An End-to-End Adaptive Multi-Scale Feature Extraction and Fusion Object Detection Network Based on UAV Aerial Images
by Sen Wang, Huiping Jiang, Jixiang Yang, Xuan Ma and Jiamin Chen
Drones 2024, 8(10), 523; https://doi.org/10.3390/drones8100523 - 26 Sep 2024
Abstract
To address the challenge of low detection accuracy and slow detection speed in unmanned aerial vehicle (UAV) aerial images target detection tasks, caused by factors such as complex ground environments, varying UAV flight altitudes and angles, and changes in lighting conditions, this study [...] Read more.
To address the challenge of low detection accuracy and slow detection speed in unmanned aerial vehicle (UAV) aerial images target detection tasks, caused by factors such as complex ground environments, varying UAV flight altitudes and angles, and changes in lighting conditions, this study proposes an end-to-end adaptive multi-scale feature extraction and fusion detection network, named AMFEF-DETR. Specifically, to extract target features from complex backgrounds more accurately, we propose an adaptive backbone network, FADC-ResNet, which dynamically adjusts dilation rates and performs adaptive frequency awareness. This enables the convolutional kernels to effectively adapt to varying scales of ground targets, capturing more details while expanding the receptive field. We also propose a HiLo attention-based intra-scale feature interaction (HLIFI) module to handle high-level features from the backbone. This module uses dual-pathway encoding of high and low frequencies to enhance the focus on the details of dense small targets while reducing noise interference. Additionally, the bidirectional adaptive feature pyramid network (BAFPN) is proposed for cross-scale feature fusion, integrating semantic information and enhancing adaptability. The Inner-Shape-IoU loss function, designed to focus on bounding box shapes and incorporate auxiliary boxes, is introduced to accelerate convergence and improve regression accuracy. When evaluated on the VisDrone dataset, the AMFEF-DETR demonstrated improvements of 4.02% and 16.71% in mAP50 and FPS, respectively, compared to the RT-DETR. Additionally, the AMFEF-DETR model exhibited strong robustness, achieving mAP50 values 2.68% and 3.75% higher than the RT-DETR and YOLOv10, respectively, on the HIT-UAV dataset. Full article
Show Figures

Figure 1

18 pages, 6006 KiB  
Article
Lightweight Insulator and Defect Detection Method Based on Improved YOLOv8
by Yanxing Liu, Xudong Li, Ruyu Qiao, Yu Chen, Xueliang Han, Agyemang Paul and Zhefu Wu
Appl. Sci. 2024, 14(19), 8691; https://doi.org/10.3390/app14198691 (registering DOI) - 26 Sep 2024
Abstract
Insulator and defect detection is a critical technology for the automated inspection of transmission and distribution lines within smart grids. However, the development of a lightweight, real-time detection platform suitable for deployment on drones faces significant challenges. These include the high complexity of [...] Read more.
Insulator and defect detection is a critical technology for the automated inspection of transmission and distribution lines within smart grids. However, the development of a lightweight, real-time detection platform suitable for deployment on drones faces significant challenges. These include the high complexity of existing algorithms, limited availability of UAV images, and persistent issues with false positives and missed detections. To address this issue, this paper proposed a lightweight drone-based insulator defect detection method (LDIDD) that integrates data augmentation and attention mechanisms based on YOLOv8. Firstly, to address the limitations of the existing insulator dataset, data augmentation techniques are developed to enhance the diversity and quantity of samples in the dataset. Secondly, to address the issue of the network model’s complexity hindering its application on UAV equipment, depthwise separable convolution is incorporated for lightweight enhancement within the YOLOv8 algorithm framework. Thirdly, a convolutional block attention mechanism is integrated into the feature extraction module to enhance the detection of small insulator targets in aerial images. The experimental results show that the improved network reduces the computational volume by 46.6% and the mAP stably maintains at 98.3% compared to YOLOv8, which enables the implementation of a lightweight insulator defect network suitable for the UAV equipment side without affecting the detection performance. Full article
Show Figures

Figure 1

22 pages, 5586 KiB  
Article
Vegetation Type Preferences in Red Deer (Cervus elaphus) Determined by Object Detection Models
by Annika Fugl, Lasse Lange Jensen, Andreas Hein Korsgaard, Cino Pertoldi and Sussie Pagh
Drones 2024, 8(10), 522; https://doi.org/10.3390/drones8100522 - 26 Sep 2024
Abstract
This study investigates the possibility of utilising a drone equipped with a thermal camera to monitor the spatial distribution of red deer (Cervus elaphus) and to determine their behavioural patterns, as well as preferences for vegetation types in a moor in [...] Read more.
This study investigates the possibility of utilising a drone equipped with a thermal camera to monitor the spatial distribution of red deer (Cervus elaphus) and to determine their behavioural patterns, as well as preferences for vegetation types in a moor in Denmark. The spatial distribution of red deer was mapped according to time of day and vegetation types. Reed deer were separated manually from fallow deer (Dama dama) due to varying footage quality. Automated object detection from thermal camera footage was used to identification of two behaviours, “Eating” and “Lying”, enabling insights into the behavioural patterns of red deer in different vegetation types. The results showed a migration of red deer from the moors to agricultural fields during the night. The higher proportion of time spent eating in agricultural grass fields compared to two natural vegetation types, “Grey dune” and “Decalcified fixed dune”, indicates that fields are important foraging habitats for red deer. The red deer populations were observed significantly later on grass fields compared to the natural vegetation types. This may be due to human disturbance or lack of randomisation of the flight time with the drone. Further studies are suggested across different seasons as well as the time of day for a better understanding of the annual and diurnal foraging patterns of red deer. Full article
(This article belongs to the Special Issue Drone Advances in Wildlife Research: 2nd Edition)
Show Figures

Figure 1

11 pages, 4657 KiB  
Communication
Assessment of the Performance of a Field Weeding Location-Based Robot Using YOLOv8
by Reetta Palva, Eerikki Kaila, Borja García Pascual and Victor Bloch
Agronomy 2024, 14(10), 2215; https://doi.org/10.3390/agronomy14102215 - 26 Sep 2024
Abstract
Field robots are an important tool when improving the efficiency and decreasing the climatic impact of food production. Although several commercial field robots are available, the advantages, limitations, and optimal utilization methods of this technology are still not well understood due to its [...] Read more.
Field robots are an important tool when improving the efficiency and decreasing the climatic impact of food production. Although several commercial field robots are available, the advantages, limitations, and optimal utilization methods of this technology are still not well understood due to its novelty. This study aims to evaluate the performance of a commercial field robot for seeding and weeding tasks. The evaluation was carried out in a 2-hectare sugar beet field. The robot’s performance was assessed by counting plants and weeds using image processing. The YOLOv8 model was trained to detect sugar beets and weeds. The plant and weed densities were compared on a robotically weeded area of the field, a chemically weeded control area, and an untreated control area. The average weed density on the robotically treated area was about two times lower than that on the untreated area and about three times higher than on the chemically treated area. The testing robot in the specific testing environment and mode showed intermediate results, weeding a majority of the weeds between the rows; however, it left the most harmful weeds close to the plants. Software for robot performance assessment can be used for monitoring robot performance and plant conditions several times during plant growth according to the weeding frequency. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

24 pages, 25641 KiB  
Article
SA-SRYOLOv8: A Research on Star Anise Variety Recognition Based on a Lightweight Cascaded Neural Network and Diversified Fusion Dataset
by Haosong Chen, Fujie Zhang, Chaofan Guo, Junjie Yi and Xiangkai Ma
Agronomy 2024, 14(10), 2211; https://doi.org/10.3390/agronomy14102211 - 25 Sep 2024
Viewed by 185
Abstract
Star anise, a widely popular spice, benefits from classification that enhances its economic value. In response to the low identification efficiency and accuracy of star anise varieties in the market, as well as the scarcity of related research, this study proposes an efficient [...] Read more.
Star anise, a widely popular spice, benefits from classification that enhances its economic value. In response to the low identification efficiency and accuracy of star anise varieties in the market, as well as the scarcity of related research, this study proposes an efficient identification method based on non-similarity augmentation and a lightweight cascaded neural network. Specifically, this approach utilizes a Siamese enhanced data network and a front-end SRGAN network to address sample imbalance and the challenge of identifying blurred images. The YOLOv8 model is further lightweight to reduce memory usage and increase detection speed, followed by optimization of the weight parameters through an extended training strategy. Additionally, a diversified fusion dataset of star anise, incorporating open data, was constructed to further validate the feasibility and effectiveness of this method. Testing showed that the SA-SRYOLOv8 detection model achieved an average detection precision (mAP) of 96.37%, with a detection speed of 146 FPS. Ablation experiment results showed that compared to the original YOLOv8 and the improved YOLOv8, the cascade model’s mAP increased by 0.09 to 0.81 percentage points. Additionally, when compared to mainstream detection models such as SSD, Fast R-CNN, YOLOv3, YOLOv5, YOLOX, and YOLOv7, the cascade model’s mAP increased by 1.81 to 19.7 percentage points. Furthermore, the model was significantly lighter, at only about 7.4% of the weight of YOLOv3, and operated at twice the speed of YOLOv7. Visualization results demonstrated that the cascade model accurately detected multiple star anise varieties across different scenarios, achieving high-precision detection targets. The model proposed in this study can provide new theoretical frameworks and ideas for constructing real-time star anise detection systems, offering new technological applications for smart agriculture. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

15 pages, 5543 KiB  
Article
Reflective Adversarial Attacks against Pedestrian Detection Systems for Vehicles at Night
by Yuanwan Chen, Yalun Wu, Xiaoshu Cui, Qiong Li, Jiqiang Liu and Wenjia Niu
Symmetry 2024, 16(10), 1262; https://doi.org/10.3390/sym16101262 - 25 Sep 2024
Viewed by 195
Abstract
The advancements in deep learning have significantly enhanced the accuracy and robustness of pedestrian detection. However, recent studies reveal that adversarial attacks can exploit the vulnerabilities of deep learning models to mislead detection systems. These attacks are effective not only in digital environments [...] Read more.
The advancements in deep learning have significantly enhanced the accuracy and robustness of pedestrian detection. However, recent studies reveal that adversarial attacks can exploit the vulnerabilities of deep learning models to mislead detection systems. These attacks are effective not only in digital environments but also pose significant threats to the reliability of pedestrian detection systems in the physical world. Existing adversarial attacks targeting pedestrian detection primarily focus on daytime scenarios and are easily noticeable by road observers. In this paper, we propose a novel adversarial attack method against vehicle–pedestrian detection systems at night. Our approach utilizes reflective optical materials that can effectively reflect light back to its source. We optimize the placement of these reflective patches using the particle swarm optimization (PSO) algorithm and deploy patches that blend with the color of pedestrian clothing in real-world scenarios. These patches remain inconspicuous during the day or under low-light conditions, but at night, the reflected light from vehicle headlights effectively disrupts the vehicle’s pedestrian detection systems. Considering that real-world detection models are often black-box systems, we propose a “symmetry” strategy, which involves using the behavior of an alternative model to simulate the response of the target model to adversarial patches. We generate adversarial examples using YOLOv5 and apply our attack to various types of pedestrian detection models. Experiments demonstrate that our approach is both effective and broadly applicable. Full article
(This article belongs to the Special Issue Advanced Studies of Symmetry/Asymmetry in Cybersecurity)
Show Figures

Figure 1

24 pages, 6889 KiB  
Article
SOD-YOLOv8—Enhancing YOLOv8 for Small Object Detection in Aerial Imagery and Traffic Scenes
by Boshra Khalili and Andrew W. Smyth
Sensors 2024, 24(19), 6209; https://doi.org/10.3390/s24196209 - 25 Sep 2024
Viewed by 268
Abstract
Object detection, as a crucial aspect of computer vision, plays a vital role in traffic management, emergency response, autonomous vehicles, and smart cities. Despite the significant advancements in object detection, detecting small objects in images captured by high-altitude cameras remains challenging, due to [...] Read more.
Object detection, as a crucial aspect of computer vision, plays a vital role in traffic management, emergency response, autonomous vehicles, and smart cities. Despite the significant advancements in object detection, detecting small objects in images captured by high-altitude cameras remains challenging, due to factors such as object size, distance from the camera, varied shapes, and cluttered backgrounds. To address these challenges, we propose small object detection YOLOv8 (SOD-YOLOv8), a novel model specifically designed for scenarios involving numerous small objects. Inspired by efficient generalized feature pyramid networks (GFPNs), we enhance multi-path fusion within YOLOv8 to integrate features across different levels, preserving details from shallower layers and improving small object detection accuracy. Additionally, we introduce a fourth detection layer to effectively utilize high-resolution spatial information. The efficient multi-scale attention module (EMA) in the C2f-EMA module further enhances feature extraction by redistributing weights and prioritizing relevant features. We introduce powerful-IoU (PIoU) as a replacement for CIoU, focusing on moderate quality anchor boxes and adding a penalty based on differences between predicted and ground truth bounding box corners. This approach simplifies calculations, speeds up convergence, and enhances detection accuracy. SOD-YOLOv8 significantly improves small object detection, surpassing widely used models across various metrics, without substantially increasing the computational cost or latency compared to YOLOv8s. Specifically, it increased recall from 40.1% to 43.9%, precision from 51.2% to 53.9%, mAP0.5 from 40.6% to 45.1%, and mAP0.5:0.95 from 24% to 26.6%. Furthermore, experiments conducted in dynamic real-world traffic scenes illustrated SOD-YOLOv8’s significant enhancements across diverse environmental conditions, highlighting its reliability and effective object detection capabilities in challenging scenarios. Full article
Show Figures

Figure 1

20 pages, 6830 KiB  
Article
Deep Learning-Based Intelligent Detection Algorithm for Surface Disease in Concrete Buildings
by Jing Gu, Yijuan Pan and Jingjing Zhang
Buildings 2024, 14(10), 3058; https://doi.org/10.3390/buildings14103058 - 25 Sep 2024
Viewed by 246
Abstract
In this study, the extent of concrete building distress is used to determine whether a building needs to be demolished and maintained, and the study focuses on accurately identifying target distress in different complex contexts and accurately distinguishing between their categories. To solve [...] Read more.
In this study, the extent of concrete building distress is used to determine whether a building needs to be demolished and maintained, and the study focuses on accurately identifying target distress in different complex contexts and accurately distinguishing between their categories. To solve the problem of insufficient feature extraction of small targets in bridge disease images under complex backgrounds and noise, we propose the YOLOv8 Dynamic Plus model. First, we enhanced attention on multi-scale disease features by implementing structural reparameterization with parallel small-kernel expansion convolution. Next, we reconstructed the relationship between localization and classification tasks in the detection head and implemented dynamic selection of interactive features using a feature extractor to improve the accuracy of classification and recognition. Finally, to address problems of missed detection, such as inadequate extraction of small targets, we extended the original YOLOv8 architecture by adding a layer in the feature extraction phase dedicated to small-target detection. This modification integrated the neck part more effectively with the shallow features of the original three-layer YOLOv8 feature extraction stage. The improved YOLOv8 Dynamic Plus model demonstrated a 7.4 percentage-point increase in performance compared to the original model, validating the feasibility of our approach and enhancing its capability for building disease detection. In practice, this improvement has led to more accurate maintenance and safety assessments of concrete buildings and earlier detection of potential structural problems, resulting in lower maintenance costs and longer building life. This not only improves the safety of buildings but also brings significant economic benefits and social value to the industries involved. Full article
Show Figures

Figure 1

9 pages, 2128 KiB  
Proceeding Paper
GEB-YOLO: Optimized YOLOv7 Model for Surface Defect Detection on Aluminum Profiles
by Zihao Xu, Jinran Hu, Xingyi Xiao and Yujian Xu
Eng. Proc. 2024, 75(1), 28; https://doi.org/10.3390/engproc2024075028 - 25 Sep 2024
Viewed by 65
Abstract
In recent years, achieving high-precision and high-speed target detection of surface defects on aluminum profiles to meet the requirements of industrial applications has been challenging. In this paper, the GEB-YOLO is proposed based on the YOLOv7 algorithm. First, the global attention mechanism (GAM) [...] Read more.
In recent years, achieving high-precision and high-speed target detection of surface defects on aluminum profiles to meet the requirements of industrial applications has been challenging. In this paper, the GEB-YOLO is proposed based on the YOLOv7 algorithm. First, the global attention mechanism (GAM) is introduced, highlighting defect features. Second, the Explicit Visual Center Block (EVCBlock) is integrated into the network for key information extraction. Meanwhile, the BiFPN network structure is adopted to enhance feature fusion. The ablation experiments have demonstrated that the defect detection accuracy of the GEB-YOLO model is improved by 6.3%, and the speed is increased by 15% compared to the YOLOv7 model. Full article
Show Figures

Figure 1

15 pages, 11412 KiB  
Article
Night Lighting Fault Detection Based on Improved YOLOv5
by Feng Zhang, Congqi Dai, Wanlu Zhang, Shu Liu and Ruiqian Guo
Buildings 2024, 14(10), 3051; https://doi.org/10.3390/buildings14103051 - 25 Sep 2024
Viewed by 155
Abstract
Night lighting is essential for urban life, and the occurrence of faults can significantly affect the presentation of lighting effects. Many reasons can cause lighting faults, including the damage of lamps and circuits, and the typical manifestation of the faults is that the [...] Read more.
Night lighting is essential for urban life, and the occurrence of faults can significantly affect the presentation of lighting effects. Many reasons can cause lighting faults, including the damage of lamps and circuits, and the typical manifestation of the faults is that the lights do not light up. The current troubleshooting mainly relies on artificial visual inspection, making detecting faults difficult and time-consuming. Therefore, it is necessary to introduce technical means to detect lighting faults. However, current research on lighting fault detection mainly focuses on using non-visual methods such as sensor data analysis, which has the disadvantages of having a high cost and difficulty adapting to large-scale fault detection. Therefore, this study mainly focuses on solving the problem of the automatic detection of night lighting faults using machine vision methods, especially object detection methods. Based on the YOLOv5 model, two data fusion models have been developed based on the characteristics of lighting fault detection inverse problems: YOLOv5 Channel Concatenation and YOLOv5 Image Fusion. Based on the dataset obtained from the developed automatic image collection and annotation system, the training and evaluation of these three models, including the original YOLOv5, YOLOv5 Channel Concatenation, and YOLOv5 Image Fusion, have been completed. Research has found that applying complete lighting images is essential for the problem of lighting fault detection. The developed Image Fusion model can effectively fuse information and accurately detect the occurrence and area of faults, with a mAP value of 0.984. This study is expected to play an essential role in the intelligent development of urban night lighting. Full article
(This article belongs to the Section Building Materials, and Repair & Renovation)
Show Figures

Figure 1

22 pages, 6602 KiB  
Article
Performance Analysis of YOLO and Detectron2 Models for Detecting Corn and Soybean Pests Employing Customized Dataset
by Guilherme Pires Silva de Almeida, Leonardo Nazário Silva dos Santos, Leandro Rodrigues da Silva Souza, Pablo da Costa Gontijo, Ruy de Oliveira, Matheus Cândido Teixeira, Mario De Oliveira, Marconi Batista Teixeira and Heyde Francielle do Carmo França
Agronomy 2024, 14(10), 2194; https://doi.org/10.3390/agronomy14102194 - 24 Sep 2024
Viewed by 390
Abstract
One of the most challenging aspects of agricultural pest control is accurate detection of insects in crops. Inadequate control measures for insect pests can seriously impact the production of corn and soybean plantations. In recent years, artificial intelligence (AI) algorithms have been extensively [...] Read more.
One of the most challenging aspects of agricultural pest control is accurate detection of insects in crops. Inadequate control measures for insect pests can seriously impact the production of corn and soybean plantations. In recent years, artificial intelligence (AI) algorithms have been extensively used for detecting insect pests in the field. In this line of research, this paper introduces a method to detect four key insect species that are predominant in Brazilian agriculture. Our model relies on computer vision techniques, including You Only Look Once (YOLO) and Detectron2, and adapts them to lightweight formats—TensorFlow Lite (TFLite) and Open Neural Network Exchange (ONNX)—for resource-constrained devices. Our method leverages two datasets: a comprehensive one and a smaller sample for comparison purposes. With this setup, the authors aimed at using these two datasets to evaluate the performance of the computer vision models and subsequently convert the best-performing models into TFLite and ONNX formats, facilitating their deployment on edge devices. The results are promising. Even in the worst-case scenario, where the ONNX model with the reduced dataset was compared to the YOLOv9-gelan model with the full dataset, the precision reached 87.3%, and the accuracy achieved was 95.0%. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

15 pages, 3657 KiB  
Article
Multi-Domain Joint Synthetic Aperture Radar Ship Detection Method Integrating Complex Information with Deep Learning
by Chaoyang Tian, Zongsen Lv, Fengli Xue, Xiayi Wu and Dacheng Liu
Remote Sens. 2024, 16(19), 3555; https://doi.org/10.3390/rs16193555 - 24 Sep 2024
Viewed by 257
Abstract
With the flourishing development of deep learning, synthetic aperture radar (SAR) ship detection based on this method has been widely applied across various domains. However, most deep-learning-based detection methods currently only use the amplitude information from SAR images. In fact, phase information and [...] Read more.
With the flourishing development of deep learning, synthetic aperture radar (SAR) ship detection based on this method has been widely applied across various domains. However, most deep-learning-based detection methods currently only use the amplitude information from SAR images. In fact, phase information and time-frequency features can also play a role in ship detection. Additionally, the background noise and the small size of ships also pose challenges to detection. Finally, satellite-based detection requires the model to be lightweight and capable of real-time processing. To address these difficulties, we propose a multi-domain joint SAR ship detection method that integrates complex information with deep learning. Based on the imaging mechanism of line-by-line scanning, we can first confirm the presence of ships within echo returns in the eigen-subspace domain, which can reduce detection time. Benefiting from the complex information of single-look complex (SLC) SAR images, we transform the echo returns containing ships into the time-frequency domain. In the time-frequency domain, ships exhibit distinctive features that are different from noise, without the limitation of size, which is highly advantageous for detection. Therefore, we constructed a time-frequency SAR image dataset (TFSID) using the images in the time-frequency domain, and utilizing the advantages of this dataset, we combined space-to-depth convolution (SPDConv) and Inception depthwise convolution (InceptionDWConv) to propose Efficient SPD-InceptionDWConv (ESIDConv). Using this module as the core, we proposed a lightweight SAR ship detector (LSDet) based on YOLOv5n. The detector achieves a detection accuracy of 99.5 with only 0.3 M parameters and 1.2 G operations on the dataset. Extensive experiments on different datasets demonstrated the superiority and effectiveness of our proposed method. Full article
Show Figures

Figure 1

23 pages, 5682 KiB  
Article
IV-YOLO: A Lightweight Dual-Branch Object Detection Network
by Dan Tian, Xin Yan, Dong Zhou, Chen Wang and Wenshuai Zhang
Sensors 2024, 24(19), 6181; https://doi.org/10.3390/s24196181 - 24 Sep 2024
Viewed by 306
Abstract
With the rapid growth in demand for security surveillance, assisted driving, and remote sensing, object detection networks with robust environmental perception and high detection accuracy have become a research focus. However, single-modality image detection technologies face limitations in environmental adaptability, often affected by [...] Read more.
With the rapid growth in demand for security surveillance, assisted driving, and remote sensing, object detection networks with robust environmental perception and high detection accuracy have become a research focus. However, single-modality image detection technologies face limitations in environmental adaptability, often affected by factors such as lighting conditions, fog, rain, and obstacles like vegetation, leading to information loss and reduced detection accuracy. We propose an object detection network that integrates features from visible light and infrared images—IV-YOLO—to address these challenges. This network is based on YOLOv8 (You Only Look Once v8) and employs a dual-branch fusion structure that leverages the complementary features of infrared and visible light images for target detection. We designed a Bidirectional Pyramid Feature Fusion structure (Bi-Fusion) to effectively integrate multimodal features, reducing errors from feature redundancy and extracting fine-grained features for small object detection. Additionally, we developed a Shuffle-SPP structure that combines channel and spatial attention to enhance the focus on deep features and extract richer information through upsampling. Regarding model optimization, we designed a loss function tailored for multi-scale object detection, accelerating the convergence speed of the network during training. Compared with the current state-of-the-art Dual-YOLO model, IV-YOLO achieves mAP improvements of 2.8%, 1.1%, and 2.2% on the Drone Vehicle, FLIR, and KAIST datasets, respectively. On the Drone Vehicle and FLIR datasets, IV-YOLO has a parameter count of 4.31 M and achieves a frame rate of 203.2 fps, significantly outperforming YOLOv8n (5.92 M parameters, 188.6 fps on the Drone Vehicle dataset) and YOLO-FIR (7.1 M parameters, 83.3 fps on the FLIR dataset), which had previously achieved the best performance on these datasets. This demonstrates that IV-YOLO achieves higher real-time detection performance while maintaining lower parameter complexity, making it highly promising for applications in autonomous driving, public safety, and beyond. Full article
(This article belongs to the Section Sensor Networks)
Show Figures

Figure 1

Back to TopTop