Svoboda | Graniru | BBC Russia | Golosameriki | Facebook
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (36,766)

Search Parameters:
Keywords = visualization

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
15 pages, 10353 KiB  
Article
Development of a Diagnostic Microfluidic Chip for SARS-CoV-2 Detection in Saliva and Nasopharyngeal Samples
by Sandhya Sharma, Massimo Caputi and Waseem Asghar
Viruses 2024, 16(8), 1190; https://doi.org/10.3390/v16081190 (registering DOI) - 25 Jul 2024
Abstract
The novel coronavirus SARS-CoV-2 was first isolated in late 2019; it has spread to all continents, infected over 700 million people, and caused over 7 million deaths worldwide to date. The high transmissibility of the virus and the emergence of novel strains with [...] Read more.
The novel coronavirus SARS-CoV-2 was first isolated in late 2019; it has spread to all continents, infected over 700 million people, and caused over 7 million deaths worldwide to date. The high transmissibility of the virus and the emergence of novel strains with altered pathogenicity and potential resistance to therapeutics and vaccines are major challenges in the study and treatment of the virus. Ongoing screening efforts aim to identify new cases to monitor the spread of the virus and help determine the danger connected to the emergence of new variants. Given its sensitivity and specificity, nucleic acid amplification tests (NAATs) such as RT-qPCR are the gold standard for SARS-CoV-2 detection. However, due to high costs, complexity, and unavailability in low-resource and point-of-care (POC) settings, the available RT-qPCR assays cannot match global testing demands. An alternative NAAT, RT-LAMP-based SARS-CoV-2 detection offers scalable, low-cost, and rapid testing capabilities. We have developed an automated RT-LAMP-based microfluidic chip that combines the RNA isolation, purification, and amplification steps on the same device and enables the visual detection of SARS-CoV-2 within 40 min from saliva and nasopharyngeal samples. The entire assay is executed inside a uniquely designed, inexpensive disposable microfluidic chip, where assay components and reagents have been optimized to provide precise and qualitative results and can be effectively deployed in POC settings. Furthermore, this technology could be easily adapted for other novel emerging viruses. Full article
(This article belongs to the Section Coronaviruses)
Show Figures

Figure 1

13 pages, 4679 KiB  
Article
Combination of Evidence from Bibliometrics and Bioinformatics Analysis Identifies miR-21 as a Potential Therapeutical Target for Diabetes
by Yiqing Chen, Xuan Ye, Xiao Zhang, Zilin Guo, Wei Chen, Zihan Pan, Zengjie Zhang, Bing Li, Hongyun Wang and Jianhua Yao
Metabolites 2024, 14(8), 403; https://doi.org/10.3390/metabo14080403 (registering DOI) - 25 Jul 2024
Abstract
Many microRNAs (miRNAs) have been identified as being involved in diabetes; however, the question of which ones may be the most promising therapeutical targets still needs more investigation. This study aims to understand the overall development tendency and identify a specific miRNA molecule [...] Read more.
Many microRNAs (miRNAs) have been identified as being involved in diabetes; however, the question of which ones may be the most promising therapeutical targets still needs more investigation. This study aims to understand the overall development tendency and identify a specific miRNA molecule to attenuate diabetes. We developed a combined analysis method based on bibliometrics and bioinformatics to visualize research institutions, authors, cited references, and keywords to identify a promising target for diabetes. Our data showed that diabetes-related miRNA is receiving continuously increasing attention, with a large number of publications, indicating that this is still a hot topic in diabetes research. Scientists from different institutions are collaborating closely in this field. miR-21, miR-146a, miR-155, and miR-34a are frequently mentioned as high-frequency keywords in the related references. Moreover, among all the above miRNAs, bioinformatics analysis further strengthens the argument that miR-21 is the top significantly upregulated molecule in diabetes patients and plays an important role in the pathogenesis of diabetes. Our study may provide a way to identify targets and promote the clinical translation of miRNA-related therapeutical strategies for diabetes, which could also indicate present and future directions for research in this area. Full article
Show Figures

Figure 1

12 pages, 3638 KiB  
Article
Exploring Edge Computing for Sustainable CV-Based Worker Detection in Construction Site Monitoring: Performance and Feasibility Analysis
by Xue Xiao, Chen Chen, Martin Skitmore, Heng Li and Yue Deng
Buildings 2024, 14(8), 2299; https://doi.org/10.3390/buildings14082299 (registering DOI) - 25 Jul 2024
Abstract
This research explores edge computing for construction site monitoring using computer vision (CV)-based worker detection methods. The feasibility of using edge computing is validated by testing worker detection models (yolov5 and yolov8) on local computers and three edge computing devices (Jetson Nano, Raspberry [...] Read more.
This research explores edge computing for construction site monitoring using computer vision (CV)-based worker detection methods. The feasibility of using edge computing is validated by testing worker detection models (yolov5 and yolov8) on local computers and three edge computing devices (Jetson Nano, Raspberry Pi 4B, and Jetson Xavier NX). The results show comparable mAP values for all devices, with the local computer processing frames six times faster than the Jetson Xavier NX. This study contributes by proposing an edge computing solution to address data security, installation complexity, and time delay issues in CV-based construction site monitoring. This approach also enhances data sustainability by mitigating potential risks associated with data loss, privacy breaches, and network connectivity issues. Additionally, it illustrates the practicality of employing edge computing devices for automated visual monitoring and provides valuable information for construction managers to select the appropriate device. Full article
Show Figures

Figure 1

15 pages, 936 KiB  
Article
Facial Expression Recognition Based on Vision Transformer with Hybrid Local Attention
by Yuan Tian, Jingxuan Zhu, Huang Yao and Di Chen
Appl. Sci. 2024, 14(15), 6471; https://doi.org/10.3390/app14156471 (registering DOI) - 24 Jul 2024
Abstract
Facial expression recognition has wide application prospects in many occasions. Due to the complexity and variability of facial expressions, facial expression recognition has become a very challenging research topic. This paper proposes a Vision Transformer expression recognition method based on hybrid local attention [...] Read more.
Facial expression recognition has wide application prospects in many occasions. Due to the complexity and variability of facial expressions, facial expression recognition has become a very challenging research topic. This paper proposes a Vision Transformer expression recognition method based on hybrid local attention (HLA-ViT). The network adopts a dual-stream structure. One stream extracts the hybrid local features and the other stream extracts the global contextual features. These two streams constitute a global–local fusion attention. The hybrid local attention module is proposed to enhance the network’s robustness to face occlusion and head pose variations. The convolutional neural network is combined with the hybrid local attention module to obtain feature maps with local prominent information. Robust features are then captured by the ViT from the global perspective of the visual sequence context. Finally, the decision-level fusion mechanism fuses the expression features with local prominent information, adding complementary information to enhance the network’s recognition performance and robustness against interference factors such as occlusion and head posture changes in natural scenes. Extensive experiments demonstrate that our HLA-ViT network achieves an excellent performance with 90.45% on RAF-DB, 90.13% on FERPlus, and 65.07% on AffectNet. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
20 pages, 2789 KiB  
Article
Performance Investigations of VSLAM and Google Street View Integration in Outdoor Location-Based Augmented Reality Under Various Lighting Conditions
by Komang Candra Brata, Nobuo Funabiki, Prismahardi Aji Riyantoko, Yohanes Yohanie Fridelin Panduman and Mustika Mentari
Electronics 2024, 13(15), 2930; https://doi.org/10.3390/electronics13152930 (registering DOI) - 24 Jul 2024
Abstract
The growing demand for Location-based Augmented Reality (LAR) experiences has driven the integration of Visual Simultaneous Localization And Mapping (VSLAM) with Google Street View (GSV) to enhance the accuracy. However, the impact of the ambient light intensity on the accuracy and reliability is [...] Read more.
The growing demand for Location-based Augmented Reality (LAR) experiences has driven the integration of Visual Simultaneous Localization And Mapping (VSLAM) with Google Street View (GSV) to enhance the accuracy. However, the impact of the ambient light intensity on the accuracy and reliability is underexplored, posing significant challenges in outdoor LAR implementations. This paper investigates the impact of light conditions on the accuracy and reliability of the VSLAM/GSV integration approach in outdoor LAR implementations. This study fills a gap in the current literature and offers valuable insights into vision-based approach implementation under different light conditions. Extensive experiments were conducted at five Point of Interest (POI) locations under various light conditions with a total of 100 datasets. Descriptive statistic methods were employed to analyze the data and assess the performance variation. Additionally, the Analysis of Variance (ANOVA) analysis was utilized to assess the impact of different light conditions on the accuracy metric and horizontal tracking time, determining whether there are significant differences in performance across varying levels of light intensity. The experimental results revealed that a significant correlation (p < 0.05) exists between the ambient light intensity and the accuracy of the VSLAM/GSV integration approach. Through the confidence interval estimation, the minimum illuminance 434lx is needed to provide a feasible and consistent accuracy. Variations in visual references, such as wet surfaces in the rainy season, also impact the horizontal tracking time and accuracy. Full article
(This article belongs to the Special Issue Perception and Interaction in Mixed, Augmented, and Virtual Reality)
21 pages, 8557 KiB  
Article
An Improved Weighted Cross-Entropy-Based Convolutional Neural Network for Auxiliary Diagnosis of Pneumonia
by Zhenyu Song, Zhanling Shi, Xuemei Yan, Bin Zhang, Shuangbao Song and Cheng Tang
Electronics 2024, 13(15), 2929; https://doi.org/10.3390/electronics13152929 (registering DOI) - 24 Jul 2024
Abstract
Pneumonia has long been a significant concern in global public health. With the advancement of convolutional neural networks (CNNs), new technological methods have emerged to address this challenge. However, the application of CNNs to pneumonia diagnosis still faces several critical issues. First, the [...] Read more.
Pneumonia has long been a significant concern in global public health. With the advancement of convolutional neural networks (CNNs), new technological methods have emerged to address this challenge. However, the application of CNNs to pneumonia diagnosis still faces several critical issues. First, the datasets used for training models often suffer from insufficient sample sizes and imbalanced class distributions, leading to reduced classification performance. Second, although CNNs can automatically extract features and make decisions from complex image data, their interpretability is relatively poor, limiting their widespread use in clinical diagnosis to some extent. To address these issues, a novel weighted cross-entropy loss function is proposed, which calculates weights via an inverse proportion exponential function to handle data imbalance more efficiently. Additionally, we employ a transfer learning approach that combines pretrained CNN model parameter fine-tuning to improve classification performance. Finally, we introduce the gradient-weighted class activation mapping method to enhance the interpretability of the model’s decisions by visualizing the image regions of focus. The experimental results indicate that our proposed approach significantly enhances CNN performance in pneumonia diagnosis tasks. Among the four selected models, the accuracy rates improved to over 90%, and visualized results were provided. Full article
(This article belongs to the Section Networks)
23 pages, 4443 KiB  
Article
Visualizing a Sustainable Future in Rural Romania: Agrotourism and Vernacular Architecture
by Raul-Cătălin Oltean, Carl T. Dahlman and Felix-Horatiu Arion
Agriculture 2024, 14(8), 1219; https://doi.org/10.3390/agriculture14081219 - 24 Jul 2024
Abstract
In Romania, rural communities grapple with decades of depopulation and economic decline, endangering the natural and cultural richness of their landscapes. The implementation of Romania’s 2030 sustainable development plan presents an opportunity to reverse these trends by merging economic and community development with [...] Read more.
In Romania, rural communities grapple with decades of depopulation and economic decline, endangering the natural and cultural richness of their landscapes. The implementation of Romania’s 2030 sustainable development plan presents an opportunity to reverse these trends by merging economic and community development with cultural preservation. This paper examines the potential for creating new livelihood opportunities through a program that integrates sustainable agrotourism with culturally appropriate vernacular architecture in Romania’s distinct rural regions. Focusing on two such regions characterized by significant rural population decline yet endowed with ecological services capable of supporting a diverse rural economy, we collaborated with an architect and landscape engineer to devise three specific and detailed agritourist housing scenarios. These scenarios draw upon local architectural forms harmonious with the vernacular landscape, providing accommodations for agrotourism guests and facilitating craft workshops for visitors interested in rural crafts and traditions. We evaluated the cultural appropriateness of the architectural designs through a social survey and assessed the broader social utility of the development plan via an expansive cost–benefit analysis, treating the project’s sustainability features as quasi-public goods. Such interdisciplinary endeavours are essential for effectively bridging conceptually driven social analysis with pragmatic design and planning strategies, essential for achieving sustainable futures for rural communities and landscapes, as exemplified by rural Romania. Full article
(This article belongs to the Special Issue Leveraging Agritourism for Rural Development)
Show Figures

Figure 1

13 pages, 1729 KiB  
Article
Prediction of Compressive Strength of Concrete Specimens Based on Interpretable Machine Learning
by Wenhu Wang, Yihui Zhong, Gang Liao, Qing Ding, Tuan Zhang and Xiangyang Li
Materials 2024, 17(15), 3661; https://doi.org/10.3390/ma17153661 - 24 Jul 2024
Abstract
The aim of this paper is to explore an effective model for predicting the compressive strength of concrete using machine learning technology, as well as to interpret the model using an interpretable method, which overcomes the limitation of the unknowable prediction processes of [...] Read more.
The aim of this paper is to explore an effective model for predicting the compressive strength of concrete using machine learning technology, as well as to interpret the model using an interpretable method, which overcomes the limitation of the unknowable prediction processes of previous machine learning models. An experimental database containing 228 samples of the compressive strength of standard cubic specimens was built in this study, and six algorithms were applied to build the predictive model. The results show that the XGBoost model has the highest prediction accuracy among all models, as the R2 of the training set and testing set are 0.982 and 0.966, respectively. Further analysis was conducted on the XGBoost model to discuss its applicability. The main steps include the following: (i) obtaining key features, (ii) obtaining trends in the evolution of features, (iii) single-sample analysis, and (iv) conducting a correlation analysis to explore methods of visualizing the variations in the factors that exert influence. The interpretability analyses on the XGBoost model show that the contribution to the compressive strength by each factor is highly in line with the conventional theory. In summary, the XGBoost model proved to be effective in predicting concrete’s compressive strength. Full article
Show Figures

Figure 1

23 pages, 71663 KiB  
Article
Design, Technical Development, and Evaluation of an Autonomous Compost Turner: An Approach towards Smart Composting
by Max Cichocki, Eva Buchmayer, Fabian Theurl and Christoph Schmied
Sustainability 2024, 16(15), 6347; https://doi.org/10.3390/su16156347 - 24 Jul 2024
Abstract
In a sustainable circular economy, the composting of organic waste plays an essential role. This paper presents the design and technical development of a smart and self-driving compost turner. The architecture of the hardware, including the sensor setup, navigation module, and control module, [...] Read more.
In a sustainable circular economy, the composting of organic waste plays an essential role. This paper presents the design and technical development of a smart and self-driving compost turner. The architecture of the hardware, including the sensor setup, navigation module, and control module, is presented. Furthermore, the methodological development using model-based systems engineering of the architecture of concepts, models, and their subsequent software integration in ROS is discussed. The validation and verification of the overall system are carried out in an industrial environment using three scenarios. The capabilities of the compost turner are demonstrated by requiring it to autonomously follow pre-defined trajectories at the composting plant and perform required composting tasks. The results prove that the autonomous compost turner can perform the required activities. In addition to autonomous driving, the compost turner is capable of intelligent processing of the compost data and of transferring, visualizing, and storing them in a cloud server. The overall system of the intelligent, autonomous compost turner can provide essential leverage for improving sustainability efforts, thus contributing substantially to an environmentally friendly and sustainable future. Full article
(This article belongs to the Special Issue Smart Manufacturing and Supply Chain Management in Industry 4.0)
27 pages, 1458 KiB  
Article
How the Degree of Anthropomorphism of Human-like Robots Affects Users’ Perceptual and Emotional Processing: Evidence from an EEG Study
by Jinchun Wu, Xiaoxi Du, Yixuan Liu, Wenzhe Tang and Chengqi Xue
Sensors 2024, 24(15), 4809; https://doi.org/10.3390/s24154809 - 24 Jul 2024
Abstract
Anthropomorphized robots are increasingly integrated into human social life, playing vital roles across various fields. This study aimed to elucidate the neural dynamics underlying users’ perceptual and emotional responses to robots with varying levels of anthropomorphism. We investigated event-related potentials (ERPs) and event-related [...] Read more.
Anthropomorphized robots are increasingly integrated into human social life, playing vital roles across various fields. This study aimed to elucidate the neural dynamics underlying users’ perceptual and emotional responses to robots with varying levels of anthropomorphism. We investigated event-related potentials (ERPs) and event-related spectral perturbations (ERSPs) elicited while participants viewed, perceived, and rated the affection of robots with low (L-AR), medium (M-AR), and high (H-AR) levels of anthropomorphism. EEG data were recorded from 42 participants. Results revealed that H-AR induced a more negative N1 and increased frontal theta power, but decreased P2 in early time windows. Conversely, M-AR and L-AR elicited larger P2 compared to H-AR. In later time windows, M-AR generated greater late positive potential (LPP) and enhanced parietal-occipital theta oscillations than H-AR and L-AR. These findings suggest distinct neural processing phases: early feature detection and selective attention allocation, followed by later affective appraisal. Early detection of facial form and animacy, with P2 reflecting higher-order visual processing, appeared to correlate with anthropomorphism levels. This research advances the understanding of emotional processing in anthropomorphic robot design and provides valuable insights for robot designers and manufacturers regarding emotional and feature design, evaluation, and promotion of anthropomorphic robots. Full article
(This article belongs to the Section Biomedical Sensors)
18 pages, 2164 KiB  
Article
Light-YOLO: A Study of a Lightweight YOLOv8n-Based Method for Underwater Fishing Net Detection
by Nuo Chen, Jin Zhu and Linhan Zheng
Appl. Sci. 2024, 14(15), 6461; https://doi.org/10.3390/app14156461 - 24 Jul 2024
Viewed by 2
Abstract
Detecting small dark targets underwater, such as fishing nets, is critical to the operation of underwater robots. Existing techniques often require more computational resources and operate under harsh underwater imaging conditions when handling such tasks. This study aims to develop a model with [...] Read more.
Detecting small dark targets underwater, such as fishing nets, is critical to the operation of underwater robots. Existing techniques often require more computational resources and operate under harsh underwater imaging conditions when handling such tasks. This study aims to develop a model with low computational resource consumption and high efficiency to improve the detection accuracy of fishing nets for safe and efficient underwater operations. The Light-YOLO model proposed in this paper introduces an attention mechanism based on sparse connectivity and deformable convolution optimized for complex underwater lighting and visual conditions. This novel attention mechanism enhances the detection performance by focusing on the key visual features of fishing nets, while the introduced CoTAttention and SEAM modules further improve the model’s recognition accuracy of fishing nets through deeper feature interactions. The results demonstrate that the proposed Light-YOLO model achieves a precision of 89.3%, a recall of 80.7%, and an [email protected] of 86.7%. Compared to other models, our model has the highest precision for its computational size and is the lightest while maintaining similar accuracy, providing an effective solution for fishing net detection and identification. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
21 pages, 16757 KiB  
Article
Flow-Field Inference for Turbulent Exhale Flow Measurement
by Shane Transue, Do-kyeong Lee, Jae-Sung Choi, Seongjun Choi, Min Hong and Min-Hyung Choi
Diagnostics 2024, 14(15), 1596; https://doi.org/10.3390/diagnostics14151596 - 24 Jul 2024
Viewed by 74
Abstract
Background: Vision-based pulmonary diagnostics present a unique approach for tracking and measuring natural breathing behaviors through remote imaging. While many existing methods correlate chest and diaphragm movements to respiratory behavior, we look at how the direct visualization of thermal CO2 exhale flow [...] Read more.
Background: Vision-based pulmonary diagnostics present a unique approach for tracking and measuring natural breathing behaviors through remote imaging. While many existing methods correlate chest and diaphragm movements to respiratory behavior, we look at how the direct visualization of thermal CO2 exhale flow patterns can be tracked to directly measure expiratory flow. Methods: In this work, we present a novel method for isolating and extracting turbulent exhale flow signals from thermal image sequences through flow-field prediction and optical flow measurement. The objective of this work is to introduce a respiratory diagnostic tool that can be used to capture and quantify natural breathing, to identify and measure respiratory metrics such as breathing rate, flow, and volume. One of the primary contributions of this work is a method for capturing and measuring natural exhale behaviors that describe individualized pulmonary traits. By monitoring subtle individualized respiratory traits, we can perform secondary analysis to identify unique personalized signatures and abnormalities to gain insight into pulmonary function. In our study, we perform data acquisition within a clinical setting to train an inference model (FieldNet) that predicts flow-fields to quantify observed exhale behaviors over time. Results: Expiratory flow measurements capturing individualized flow signatures from our initial cohort demonstrate how the proposed flow field model can be used to isolate and analyze turbulent exhale behaviors and measure anomalous behavior. Conclusions: Our results illustrate that detailed spatial flow analysis can contribute to unique signatures for identifying patient specific natural breathing behaviors and abnormality detection. This provides the first-step towards a non-contact respiratory technology that directly captures effort-independent behaviors based on the direct measurement of imaged CO2 exhaled airflow patterns. Full article
(This article belongs to the Special Issue Diagnosis, Classification, and Monitoring of Pulmonary Diseases)
Show Figures

Figure 1

18 pages, 20456 KiB  
Article
RCRFNet: Enhancing Object Detection with Self-Supervised Radar–Camera Fusion and Open-Set Recognition
by Minwei Chen, Yajun Liu, Zenghui Zhang and Weiwei Guo
Sensors 2024, 24(15), 4803; https://doi.org/10.3390/s24154803 - 24 Jul 2024
Viewed by 82
Abstract
Robust object detection in complex environments, poor visual conditions, and open scenarios presents significant technical challenges in autonomous driving. These challenges necessitate the development of advanced fusion methods for millimeter-wave (mmWave) radar point cloud data and visual images. To address these issues, this [...] Read more.
Robust object detection in complex environments, poor visual conditions, and open scenarios presents significant technical challenges in autonomous driving. These challenges necessitate the development of advanced fusion methods for millimeter-wave (mmWave) radar point cloud data and visual images. To address these issues, this paper proposes a radar–camera robust fusion network (RCRFNet), which leverages self-supervised learning and open-set recognition to effectively utilise the complementary information from both sensors. Specifically, the network uses matched radar–camera data through a frustum association approach to generate self-supervised signals, enhancing network training. The integration of global and local depth consistencies between radar point clouds and visual images, along with image features, helps construct object class confidence levels for detecting unknown targets. Additionally, these techniques are combined with a multi-layer feature extraction backbone and a multimodal feature detection head to achieve robust object detection. Experiments on the nuScenes public dataset demonstrate that RCRFNet outperforms state-of-the-art (SOTA) methods, particularly in conditions of low visual visibility and when detecting unknown class objects. Full article
24 pages, 8432 KiB  
Article
Lane Attribute Classification Based on Fine-Grained Description
by Zhonghe He, Pengfei Gong, Hongcheng Ye and Zizheng Gan
Sensors 2024, 24(15), 4800; https://doi.org/10.3390/s24154800 - 24 Jul 2024
Viewed by 109
Abstract
As an indispensable part of the vehicle environment perception task, road traffic marking detection plays a vital role in correctly understanding the current traffic situation. However, the existing traffic marking detection algorithms still have some limitations. Taking lane detection as an example, the [...] Read more.
As an indispensable part of the vehicle environment perception task, road traffic marking detection plays a vital role in correctly understanding the current traffic situation. However, the existing traffic marking detection algorithms still have some limitations. Taking lane detection as an example, the current detection methods mainly focus on the location information detection of lane lines, and they only judge the overall attribute of each detected lane line instance, thus lacking more fine-grained dynamic detection of lane line attributes. In order to meet the needs of intelligent vehicles for the dynamic attribute detection of lane lines and more perfect road environment information in urban road environment, this paper constructs a fine-grained attribute detection method for lane lines, which uses pixel-level attribute sequence points to describe the complete attribute distribution of lane lines and then matches the detection results of the lane lines. Realizing the attribute judgment of different segment positions of lane instances is called the fine-grained attribute detection of lane lines (Lane-FGA). In addition, in view of the lack of annotation information in the current open-source lane data set, this paper constructs a lane data set with both lane instance information and fine-grained attribute information by combining manual annotation and intelligent annotation. At the same time, a cyclic iterative attribute inference algorithm is designed to solve the difficult problem of lane attribute labeling in areas without visual cues such as occlusion and damage. In the end, the average accuracy of the proposed algorithm reaches 97% on various types of lane attribute detection. Full article
(This article belongs to the Topic Applications in Image Analysis and Pattern Recognition)
Show Figures

Figure 1

16 pages, 483 KiB  
Article
Query-Based Object Visual Tracking with Parallel Sequence Generation
by Chang Liu, Bin Zhang, Chunjuan Bo and Dong Wang
Sensors 2024, 24(15), 4802; https://doi.org/10.3390/s24154802 - 24 Jul 2024
Viewed by 92
Abstract
Query decoders have been shown to achieve good performance in object detection. However, they suffer from insufficient object tracking performance. Sequence-to-sequence learning in this context has recently been explored, with the idea of describing a target as a sequence of discrete tokens. In [...] Read more.
Query decoders have been shown to achieve good performance in object detection. However, they suffer from insufficient object tracking performance. Sequence-to-sequence learning in this context has recently been explored, with the idea of describing a target as a sequence of discrete tokens. In this study, we experimentally determine that, with appropriate representation, a parallel approach for predicting a target coordinate sequence with a query decoder can achieve good performance and speed. We propose a concise query-based tracking framework for predicting a target coordinate sequence in a parallel manner, named QPSTrack. A set of queries are designed to be responsible for different coordinates of the tracked target. All the queries jointly represent a target rather than a traditional one-to-one matching pattern between the query and target. Moreover, we adopt an adaptive decoding scheme including a one-layer adaptive decoder and learnable adaptive inputs for the decoder. This decoding scheme assists the queries in decoding the template-guided search features better. Furthermore, we explore the use of the plain ViT-Base, ViT-Large, and lightweight hierarchical LeViT architectures as the encoder backbone, providing a family of three variants in total. All the trackers are found to obtain a good trade-off between speed and performance; for instance, our tracker QPSTrack-B256 with the ViT-Base encoder achieves a 69.1% AUC on the LaSOT benchmark at 104.8 FPS. Full article
Show Figures

Figure 1

Back to TopTop