Svoboda | Graniru | BBC Russia | Golosameriki | Facebook
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (2,353)

Search Parameters:
Keywords = multi-task learning

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 7653 KiB  
Article
TransPVP: A Transformer-Based Method for Ultra-Short-Term Photovoltaic Power Forecasting
by Jinfeng Wang, Wenshan Hu, Lingfeng Xuan, Feiwu He, Chaojie Zhong and Guowei Guo
Energies 2024, 17(17), 4426; https://doi.org/10.3390/en17174426 (registering DOI) - 4 Sep 2024
Abstract
The increasing adoption of renewable energy, particularly photovoltaic (PV) power, has highlighted the importance of accurate PV power forecasting. Despite advances driven by deep learning (DL), significant challenges remain, particularly in capturing the long-term dependencies essential for accurate forecasting. This study presents TransPVP, [...] Read more.
The increasing adoption of renewable energy, particularly photovoltaic (PV) power, has highlighted the importance of accurate PV power forecasting. Despite advances driven by deep learning (DL), significant challenges remain, particularly in capturing the long-term dependencies essential for accurate forecasting. This study presents TransPVP, a novel transformer-based methodology that addresses these challenges and advances PV power forecasting. TransPVP employs a deep fusion technique alongside a multi-task joint learning framework, effectively integrating heterogeneous data sources and capturing long-term dependencies. This innovative approach enhances the model’s ability to detect patterns of PV power variation, surpassing the capabilities of traditional models. The effectiveness of TransPVP was rigorously evaluated using real data from a PV power plant. Experimental results showed that TransPVP significantly outperformed established baseline models on key performance metrics including RMSE, R2, and CC, underscoring its accuracy, predictive power, and reliability in practical forecasting scenarios. Full article
(This article belongs to the Section A2: Solar Energy and Photovoltaic Systems)
Show Figures

Figure 1

17 pages, 6059 KiB  
Article
ECF-Net: Enhanced, Channel-Based, Multi-Scale Feature Fusion Network for COVID-19 Image Segmentation
by Zhengjie Ji, Junhao Zhou, Linjing Wei, Shudi Bao, Meng Chen, Hongxing Yuan and Jianjun Zheng
Electronics 2024, 13(17), 3501; https://doi.org/10.3390/electronics13173501 - 3 Sep 2024
Abstract
Accurate segmentation of COVID-19 lesion regions in lung CT images aids physicians in analyzing and diagnosing patients’ conditions. However, the varying morphology and blurred contours of these regions make this task complex and challenging. Existing methods utilizing Transformer architecture lack attention to local [...] Read more.
Accurate segmentation of COVID-19 lesion regions in lung CT images aids physicians in analyzing and diagnosing patients’ conditions. However, the varying morphology and blurred contours of these regions make this task complex and challenging. Existing methods utilizing Transformer architecture lack attention to local features, leading to the loss of detailed information in tiny lesion regions. To address these issues, we propose a multi-scale feature fusion network, ECF-Net, based on channel enhancement. Specifically, we leverage the learning capabilities of both CNN and Transformer architectures to design parallel channel extraction blocks in three different ways, effectively capturing diverse lesion features. Additionally, to minimize irrelevant information in the high-dimensional feature space and focus the network on useful and critical information, we develop adaptive feature generation blocks. Lastly, a bidirectional pyramid-structured feature fusion approach is introduced to integrate features at different levels, enhancing the diversity of feature representations and improving segmentation accuracy for lesions of various scales. The proposed method is tested on four COVID-19 datasets, demonstrating mIoU values of 84.36%, 87.15%, 83.73%, and 75.58%, respectively, outperforming several current state-of-the-art methods and exhibiting excellent segmentation performance. These findings provide robust technical support for medical image segmentation in clinical practice. Full article
(This article belongs to the Special Issue Biomedical Image Processing and Classification, 2nd Edition)
Show Figures

Figure 1

19 pages, 1200 KiB  
Article
Optimized Intrusion Detection for IoMT Networks with Tree-Based Machine Learning and Filter-Based Feature Selection
by Ghaida Balhareth and Mohammad Ilyas
Sensors 2024, 24(17), 5712; https://doi.org/10.3390/s24175712 - 2 Sep 2024
Viewed by 253
Abstract
The Internet of Medical Things (IoMTs) is a network of connected medical equipment such as pacemakers, prosthetics, and smartwatches. Utilizing the IoMT-based system, a huge amount of data is generated, offering experts a valuable resource for tasks such as prediction, real-time monitoring, and [...] Read more.
The Internet of Medical Things (IoMTs) is a network of connected medical equipment such as pacemakers, prosthetics, and smartwatches. Utilizing the IoMT-based system, a huge amount of data is generated, offering experts a valuable resource for tasks such as prediction, real-time monitoring, and diagnosis. To do so, the patient’s health data must be transferred to database storage for processing because of the limitations of the storage and computation capabilities of IoMT devices. Consequently, concerns regarding security and privacy can arise due to the limited control over the transmitted information and reliance on wireless transmission, which leaves the network vulnerable to several kinds of attacks. Motivated by this, in this study, we aim to build and improve an efficient intrusion detection system (IDS) for IoMT networks. The proposed IDS leverages tree-based machine learning classifiers combined with filter-based feature selection techniques to enhance detection accuracy and efficiency. The proposed model is used for monitoring and identifying unauthorized or malicious activities within medical devices and networks. To optimize performance and minimize computation costs, we utilize Mutual Information (MI) and XGBoost as filter-based feature selection methods. Then, to reduce the number of the chosen features selected, we apply a mathematical set (intersection) to extract the common features. The proposed method can detect intruders while data are being transferred, allowing for the accurate and efficient analysis of healthcare data at the network’s edge. The system’s performance is assessed using the CICIDS2017 dataset. We evaluate the proposed model in terms of accuracy, F1 score, recall, precision, true positive rate, and false positive rate. The proposed model achieves 98.79% accuracy and a low false alarm rate 0.007 FAR on the CICIDS2017 dataset according to the experimental results. While this study focuses on binary classification for intrusion detection, we are planning to build a multi-classification approach for future work which will be able to not only detect the attacks but also categorize them. Additionally, we will consider using our proposed feature selection technique for different ML classifiers and evaluate the model’s performance empirically in real-world IoMT scenarios. Full article
(This article belongs to the Section Internet of Things)
Show Figures

Figure 1

22 pages, 13240 KiB  
Article
SpemNet: A Cotton Disease and Pest Identification Method Based on Efficient Multi-Scale Attention and Stacking Patch Embedding
by Keyuan Qiu, Yingjie Zhang, Zekai Ren, Meng Li, Qian Wang, Yiqiang Feng and Feng Chen
Insects 2024, 15(9), 667; https://doi.org/10.3390/insects15090667 - 2 Sep 2024
Viewed by 247
Abstract
We propose a cotton pest and disease recognition method, SpemNet, based on efficient multi-scale attention and stacking patch embedding. By introducing the SPE module and the EMA module, we successfully solve the problems of local feature learning difficulty and insufficient multi-scale feature integration [...] Read more.
We propose a cotton pest and disease recognition method, SpemNet, based on efficient multi-scale attention and stacking patch embedding. By introducing the SPE module and the EMA module, we successfully solve the problems of local feature learning difficulty and insufficient multi-scale feature integration in the traditional Vision Transformer model, which significantly improve the performance and efficiency of the model. In our experiments, we comprehensively validate the SpemNet model on the CottonInsect dataset, and the results show that SpemNet performs well in the cotton pest recognition task, with significant effectiveness and superiority. The SpemNet model excels in key metrics such as precision and F1 score, demonstrating significant potential and superiority in the cotton pest and disease recognition task. This study provides an efficient and reliable solution in the field of cotton pest and disease identification, which is of great theoretical and applied significance. Full article
Show Figures

Figure 1

11 pages, 1023 KiB  
Article
Research on the Migration and Settlement Laws of Backflow Proppants after Fracturing Tight Sandstone
by Hanlie Cheng and Qiang Qin
Appl. Sci. 2024, 14(17), 7746; https://doi.org/10.3390/app14177746 - 2 Sep 2024
Viewed by 201
Abstract
This article studies the migration and settlement laws of backflow proppants after fracturing tight sandstone. This paper proposes a fitting method based on a multi-task learning network to address the issue of interference from multiple physical parameters during the transport and settlement processes [...] Read more.
This article studies the migration and settlement laws of backflow proppants after fracturing tight sandstone. This paper proposes a fitting method based on a multi-task learning network to address the issue of interference from multiple physical parameters during the transport and settlement processes of proppants. This method can effectively handle multi-dimensional interference factors and fit the mapping logic of multiple engineering parameters to transport patterns through the continuous correction of multi-layer networks. We first introduce the characteristics of tight sandstone reservoirs and their important value in mining, as well as the status of current research on the migration and settlement laws of proppants at home and abroad. Based on this, we then deeply analyze the sedimentation rate model of proppants in tight sandstone backflow and the equilibrium height of proppants under multiple factors of interference while considering the distribution characteristics of proppants. In order to more accurately simulate the transport and settlement laws of proppants, this paper introduces a multi-task learning network. This network can comprehensively consider multi-dimensional parameters, learn the inherent laws of data through training, and achieve accurate fitting of the transport and settlement laws of proppants. This study trained and tested the model using actual production data, and the results showed that the proposed model can fit the input–output relationship well, thus effectively supporting the study of proppant transport and settlement laws. Full article
Show Figures

Figure 1

17 pages, 3620 KiB  
Article
Image Registration Algorithm for Stamping Process Monitoring Based on Improved Unsupervised Homography Estimation
by Yujie Zhang and Yinuo Du
Appl. Sci. 2024, 14(17), 7721; https://doi.org/10.3390/app14177721 - 2 Sep 2024
Viewed by 307
Abstract
Homography estimation is a crucial task in aligning template images with target images in stamping monitoring systems. To enhance the robustness and accuracy of homography estimation against random vibrations and lighting variations in stamping environments, this paper proposes an improved unsupervised homography estimation [...] Read more.
Homography estimation is a crucial task in aligning template images with target images in stamping monitoring systems. To enhance the robustness and accuracy of homography estimation against random vibrations and lighting variations in stamping environments, this paper proposes an improved unsupervised homography estimation model. The model takes as input the channel-stacked template and target images and outputs the estimated homography matrix. First, a specialized deformable convolution module and Group Normalization (GN) layer are introduced to expand the receptive field and enhance the model’s ability to learn rotational invariance when processing large, high-resolution images. Next, a multi-scale, multi-stage unsupervised homography estimation network structure is constructed to improve the accuracy of homography estimation by refining the estimation through multiple stages, thereby enhancing the model’s resistance to scale variations. Finally, stamping monitoring image data is incorporated into the training through data fusion, with data augmentation techniques applied to randomly introduce various levels of perturbation, brightness, contrast, and filtering to improve the model’s robustness to complex changes in the stamping environment, making it more suitable for monitoring applications in this specific industrial context. Compared to traditional methods, this approach provides better homography matrix estimation when handling images with low texture, significant lighting variations, or large viewpoint changes. Compared to other deep-learning-based homography estimation methods, it reduces estimation errors and performs better on stamping monitoring images, while also offering broader applicability. Full article
Show Figures

Figure 1

44 pages, 4286 KiB  
Article
Multitask Learning for Crash Analysis: A Fine-Tuned LLM Framework Using Twitter Data
by Shadi Jaradat, Richi Nayak, Alexander Paz, Huthaifa I. Ashqar and Mohammad Elhenawy
Smart Cities 2024, 7(5), 2422-2465; https://doi.org/10.3390/smartcities7050095 (registering DOI) - 1 Sep 2024
Viewed by 473
Abstract
Road traffic crashes (RTCs) are a global public health issue, with traditional analysis methods often hindered by delays and incomplete data. Leveraging social media for real-time traffic safety analysis offers a promising alternative, yet effective frameworks for this integration are scarce. This study [...] Read more.
Road traffic crashes (RTCs) are a global public health issue, with traditional analysis methods often hindered by delays and incomplete data. Leveraging social media for real-time traffic safety analysis offers a promising alternative, yet effective frameworks for this integration are scarce. This study introduces a novel multitask learning (MTL) framework utilizing large language models (LLMs) to analyze RTC-related tweets from Australia. We collected 26,226 traffic-related tweets from May 2022 to May 2023. Using GPT-3.5, we extracted fifteen distinct features categorized into six classification tasks and nine information retrieval tasks. These features were then used to fine-tune GPT-2 for language modeling, which outperformed baseline models, including GPT-4o mini in zero-shot mode and XGBoost, across most tasks. Unlike traditional single-task classifiers that may miss critical details, our MTL approach simultaneously classifies RTC-related tweets and extracts detailed information in natural language. Our fine-tunedGPT-2 model achieved an average accuracy of 85% across the six classification tasks, surpassing the baseline GPT-4o mini model’s 64% and XGBoost’s 83.5%. In information retrieval tasks, our fine-tuned GPT-2 model achieved a BLEU-4 score of 0.22, a ROUGE-I score of 0.78, and a WER of 0.30, significantly outperforming the baseline GPT-4 mini model’s BLEU-4 score of 0.0674, ROUGE-I score of 0.2992, and WER of 2.0715. These results demonstrate the efficacy of our fine-tuned GPT-2 model in enhancing both classification and information retrieval, offering valuable insights for data-driven decision-making to improve road safety. This study is the first to explicitly apply social media data and LLMs within an MTL framework to enhance traffic safety. Full article
(This article belongs to the Section Smart Transportation)
Show Figures

Figure 1

15 pages, 2857 KiB  
Article
Assessing the Capabilities of UV-NIR Spectroscopy for Predicting Macronutrients in Hydroponic Solutions with Single-Task and Multi-Task Learning
by Haijun Qi, Bin Li, Jun Nie, Yizhi Luo, Yu Yuan and Xingxing Zhou
Agronomy 2024, 14(9), 1974; https://doi.org/10.3390/agronomy14091974 - 1 Sep 2024
Viewed by 249
Abstract
Macronutrients, including nitrogen (N), phosphorus (P), potassium (K), calcium (Ca), magnesium (Mg), and sulfur (S), are the most basic nutrient elements in the solution for the hydroponic system. However, the current management of hydroponic nutrient solutions usually depends on EC and pH sensors [...] Read more.
Macronutrients, including nitrogen (N), phosphorus (P), potassium (K), calcium (Ca), magnesium (Mg), and sulfur (S), are the most basic nutrient elements in the solution for the hydroponic system. However, the current management of hydroponic nutrient solutions usually depends on EC and pH sensors due to the lack of accurate specific macronutrient sensing equipment, which easily leads to nutritional imbalance for the cultivated plant. In this study, the UV-NIR absorption spectroscopy (200–1100 nm) was used to predict six macronutrients in hydroponic solutions; two kinds of single-task learning algorithms, including partial least squares (PLS) and least absolute shrinkage and selection operator (LASSO), and two kinds of multi-task learning algorithms, including dirty multi-task learning (DMTL) and robust multi-task learning (RMTL), were investigated to develop prediction models and assess capabilities of UV-NIR. The results showed that N and Ca could be quantitatively predicted by UV-NIR with the ratio of performance to deviation (RPD) more than 2, K could be qualitatively predicted (1.4 < RPD < 2), and P, Mg, and S could not be successfully predicted (RPD < 1.4); the RMTL algorithm outperformed others for predicting K and Ca benefit from the underlying task relationships with N; and predicting P, Mg, and S were identified as irrelevant (outlier) tasks. Our study provides a potential approach for predicting several macronutrients in hydroponic solutions with UV-NIR, especially using RMTL to improve model prediction ability. Full article
(This article belongs to the Special Issue The Application of Near-Infrared Spectroscopy in Agriculture)
Show Figures

Figure 1

23 pages, 2299 KiB  
Article
CA-ViT: Contour-Guided and Augmented Vision Transformers to Enhance Glaucoma Classification Using Fundus Images
by Tewodros Gizaw Tohye, Zhiguang Qin, Mugahed A. Al-antari, Chiagoziem C. Ukwuoma, Zenbe Markos Lonseko and Yeong Hyeon Gu
Bioengineering 2024, 11(9), 887; https://doi.org/10.3390/bioengineering11090887 - 31 Aug 2024
Viewed by 345
Abstract
Glaucoma, a predominant cause of visual impairment on a global scale, poses notable challenges in diagnosis owing to its initially asymptomatic presentation. Early identification is vital to prevent irreversible vision impairment. Cutting-edge deep learning techniques, such as vision transformers (ViTs), have been employed [...] Read more.
Glaucoma, a predominant cause of visual impairment on a global scale, poses notable challenges in diagnosis owing to its initially asymptomatic presentation. Early identification is vital to prevent irreversible vision impairment. Cutting-edge deep learning techniques, such as vision transformers (ViTs), have been employed to tackle the challenge of early glaucoma detection. Nevertheless, limited approaches have been suggested to improve glaucoma classification due to issues like inadequate training data, variations in feature distribution, and the overall quality of samples. Furthermore, fundus images display significant similarities and slight discrepancies in lesion sizes, complicating glaucoma classification when utilizing ViTs. To address these obstacles, we introduce the contour-guided and augmented vision transformer (CA-ViT) for enhanced glaucoma classification using fundus images. We employ a Conditional Variational Generative Adversarial Network (CVGAN) to enhance and diversify the training dataset by incorporating conditional sample generation and reconstruction. Subsequently, a contour-guided approach is integrated to offer crucial insights into the disease, particularly concerning the optic disc and optic cup regions. Both the original images and extracted contours are given to the ViT backbone; then, feature alignment is performed with a weighted cross-entropy loss. Finally, in the inference phase, the ViT backbone, trained on the original fundus images and augmented data, is used for multi-class glaucoma categorization. By utilizing the Standardized Multi-Channel Dataset for Glaucoma (SMDG), which encompasses various datasets (e.g., EYEPACS, DRISHTI-GS, RIM-ONE, REFUGE), we conducted thorough testing. The results indicate that the proposed CA-ViT model significantly outperforms current methods, achieving a precision of 93.0%, a recall of 93.08%, an F1 score of 92.9%, and an accuracy of 93.0%. Therefore, the integration of augmentation with the CVGAN and contour guidance can effectively enhance glaucoma classification tasks. Full article
17 pages, 2023 KiB  
Article
Deep Siamese Neural Network-Driven Model for Robotic Multiple Peg-in-Hole Assembly System
by Jinlong Chen, Wei Tang and Minghao Yang
Electronics 2024, 13(17), 3453; https://doi.org/10.3390/electronics13173453 - 30 Aug 2024
Viewed by 355
Abstract
Robots are now widely used in assembly tasks. However, when robots perform the automatic assembly of Multi-Pin Circular Connectors (MPCCs), the small diameter of the pins and the narrow gaps between them present significant challenges. During the assembly process, the robot’s end effector [...] Read more.
Robots are now widely used in assembly tasks. However, when robots perform the automatic assembly of Multi-Pin Circular Connectors (MPCCs), the small diameter of the pins and the narrow gaps between them present significant challenges. During the assembly process, the robot’s end effector can obstruct the view, and the contact between the pins and the corresponding holes is completely blocked, making this task more precise and challenging than the common peg-in-hole assembly. Therefore, this paper proposes a robotic assembly strategy for MPCCs that includes two main aspects: (1) we employ a vision-based Deep Siamese Neural Network (DSNN) model to address the most challenging peg-in-hole alignment problem in MPCC assembly. This method avoids the difficulties of modeling in traditional control strategies, the high training costs, and the low sample efficiency in reinforcement learning. (2) This paper constructs a complete practical assembly system for MPCCs, covering everything from gripping to final screwing. The experimental results consistently demonstrate that the assembly system integrated with the DSNN can effectively accomplish the MPCC assembly task. Full article
Show Figures

Figure 1

26 pages, 5826 KiB  
Article
An Efficient Task Implementation Modeling Framework with Multi-Stage Feature Selection and AutoML: A Case Study in Forest Fire Risk Prediction
by Ye Su, Longlong Zhao, Hongzhong Li, Xiaoli Li, Jinsong Chen and Yuankai Ge
Remote Sens. 2024, 16(17), 3190; https://doi.org/10.3390/rs16173190 (registering DOI) - 29 Aug 2024
Viewed by 372
Abstract
As data science advances, automated machine learning (AutoML) gains attention for lowering barriers, saving time, and enhancing efficiency. However, with increasing data dimensionality, AutoML struggles with large-scale feature sets. Effective feature selection is crucial for efficient AutoML in multi-task applications. This study proposes [...] Read more.
As data science advances, automated machine learning (AutoML) gains attention for lowering barriers, saving time, and enhancing efficiency. However, with increasing data dimensionality, AutoML struggles with large-scale feature sets. Effective feature selection is crucial for efficient AutoML in multi-task applications. This study proposes an efficient modeling framework combining a multi-stage feature selection (MSFS) algorithm and AutoSklearn, a robust and efficient AutoML framework, to address high-dimensional data challenges. The MSFS algorithm includes three stages: mutual information gain (MIG), recursive feature elimination with cross-validation (RFECV), and a voting aggregation mechanism, ensuring comprehensive consideration of feature correlation, importance, and stability. Based on multi-source and time series remote sensing data, this study pioneers the application of AutoSklearn for forest fire risk prediction. Using this case study, we compare MSFS with five other feature selection (FS) algorithms, including three single FS algorithms and two hybrid FS algorithms. Results show that MSFS selects half of the original features (12/24), effectively handling collinearity (eliminating 11 out of 13 collinear feature groups) and increasing AutoSklearn’s success rate by 15%, outperforming two FS algorithms with the same number of features by 7% and 5%. Among the six FS algorithms and non-FS, MSFS demonstrates the highest prediction performance and stability with minimal variance (0.09%) across five evaluation metrics. MSFS efficiently filters redundant features, enhancing AutoSklearn’s operational efficiency and generalization ability in high-dimensional tasks. The MSFS–AutoSklearn framework significantly improves AutoML’s production efficiency and prediction accuracy, facilitating the efficient implementation of various real-world tasks and the wider application of AutoML. Full article
Show Figures

Figure 1

12 pages, 442 KiB  
Article
Exploring Autism Spectrum Disorder: A Comparative Study of Traditional Classifiers and Deep Learning Classifiers to Analyze Functional Connectivity Measures from a Multicenter Dataset
by Francesca Mainas, Bruno Golosio, Alessandra Retico and Piernicola Oliva
Appl. Sci. 2024, 14(17), 7632; https://doi.org/10.3390/app14177632 - 29 Aug 2024
Viewed by 312
Abstract
The investigation of functional magnetic resonance imaging (fMRI) data with traditional machine learning (ML) and deep learning (DL) classifiers has been widely used to study autism spectrum disorders (ASDs). This condition is characterized by symptoms that affect the individual’s behavioral aspects and social [...] Read more.
The investigation of functional magnetic resonance imaging (fMRI) data with traditional machine learning (ML) and deep learning (DL) classifiers has been widely used to study autism spectrum disorders (ASDs). This condition is characterized by symptoms that affect the individual’s behavioral aspects and social relationships. Early diagnosis is crucial for intervention, but the complexity of ASD poses challenges for the development of effective treatments. This study compares traditional ML and DL classifiers in the analysis of tabular data, in particular, functional connectivity measures obtained from the time series of a public multicenter dataset, and evaluates whether the features that contribute most to the classification task vary depending on the classifier used. Specifically, Support Vector Machine (SVM) classifiers, with both linear and radial basis function (RBF) kernels, and Extreme Gradient Boosting (XGBoost) classifiers are compared against the TabNet classifier (a DL architecture customized for tabular data analysis) and a Multi Layer Perceptron (MLP). The findings suggest that DL classifiers may not be optimal for the type of data analyzed, as their performance trails behind that of standard classifiers. Among the latter, SVMs outperform the other classifiers with an AUC of around 75%, whereas the best performances of TabNet and MLP reach 65% and 71% at most, respectively. Furthermore, the analysis of the feature importance showed that the brain regions that contribute the most to the classification task are those primarily responsible for sensory and spatial perception, as well as attention modulation, which is known to be altered in ASDs. Full article
Show Figures

Figure 1

23 pages, 11538 KiB  
Article
A Novel Deep Learning Model for Breast Tumor Ultrasound Image Classification with Lesion Region Perception
by Jinzhu Wei, Haoyang Zhang and Jiang Xie
Curr. Oncol. 2024, 31(9), 5057-5079; https://doi.org/10.3390/curroncol31090374 - 28 Aug 2024
Viewed by 247
Abstract
Multi-task learning (MTL) methods are widely applied in breast imaging for lesion area perception and classification to assist in breast cancer diagnosis and personalized treatment. A typical paradigm of MTL is the shared-backbone network architecture, which can lead to information sharing conflicts and [...] Read more.
Multi-task learning (MTL) methods are widely applied in breast imaging for lesion area perception and classification to assist in breast cancer diagnosis and personalized treatment. A typical paradigm of MTL is the shared-backbone network architecture, which can lead to information sharing conflicts and result in the decline or even failure of the main task’s performance. Therefore, extracting richer lesion features and alleviating information-sharing conflicts has become a significant challenge for breast cancer classification. This study proposes a novel Multi-Feature Fusion Multi-Task (MFFMT) model to effectively address this issue. Firstly, in order to better capture the local and global feature relationships of lesion areas, a Contextual Lesion Enhancement Perception (CLEP) module is designed, which integrates channel attention mechanisms with detailed spatial positional information to extract more comprehensive lesion feature information. Secondly, a novel Multi-Feature Fusion (MFF) module is presented. The MFF module effectively extracts differential features that distinguish between lesion-specific characteristics and the semantic features used for tumor classification, and enhances the common feature information of them as well. Experimental results on two public breast ultrasound imaging datasets validate the effectiveness of our proposed method. Additionally, a comprehensive study on the impact of various factors on the model’s performance is conducted to gain a deeper understanding of the working mechanism of the proposed framework. Full article
(This article belongs to the Topic Artificial Intelligence in Cancer Pathology and Prognosis)
Show Figures

Figure 1

14 pages, 1786 KiB  
Article
AI Services-Oriented Dynamic Computing Resource Scheduling Algorithm Based on Distributed Data Parallelism in Edge Computing Network of Smart Grid
by Jing Zou, Peizhe Xin, Chang Wang, Heli Zhang, Lei Wei and Ying Wang
Future Internet 2024, 16(9), 312; https://doi.org/10.3390/fi16090312 - 28 Aug 2024
Viewed by 336
Abstract
Massive computational resources are required by a booming number of artificial intelligence (AI) services in the communication network of the smart grid. To alleviate the computational pressure on data centers, edge computing first network (ECFN) can serve as an effective solution to realize [...] Read more.
Massive computational resources are required by a booming number of artificial intelligence (AI) services in the communication network of the smart grid. To alleviate the computational pressure on data centers, edge computing first network (ECFN) can serve as an effective solution to realize distributed model training based on data parallelism for AI services in smart grid. Due to AI services with diversified types, an edge data center has a changing workload in different time periods. Selfish edge data centers from different edge suppliers are reluctant to share their computing resources without a rule for fair competition. AI services-oriented dynamic computational resource scheduling of edge data centers affects both the economic profit of AI service providers and computational resource utilization. This letter mainly discusses the partition and distribution of AI data based on distributed model training and dynamic computational resource scheduling problems among multiple edge data centers for AI services. To this end, a mixed integer linear programming (MILP) model and a Deep Reinforcement Learning (DRL)-based algorithm are proposed. Simulation results show that the proposed DRL-based algorithm outperforms the benchmark in terms of profit of AI service provider, backlog of distributed model training tasks, running time and multi-objective optimization. Full article
Show Figures

Figure 1

16 pages, 4344 KiB  
Article
Multi-Scale Spatio-Temporal Attention Networks for Network-Scale Traffic Learning and Forecasting
by Cong Wu, Hui Ding, Zhongwang Fu and Ning Sun
Sensors 2024, 24(17), 5543; https://doi.org/10.3390/s24175543 (registering DOI) - 27 Aug 2024
Viewed by 410
Abstract
Accurate and timely forecasting of traffic on local road networks is crucial for deploying effective dynamic traffic control, advanced route planning, and navigation services. This task is particularly challenging due to complex spatio-temporal dependencies arising from non-Euclidean spatial relations in road networks and [...] Read more.
Accurate and timely forecasting of traffic on local road networks is crucial for deploying effective dynamic traffic control, advanced route planning, and navigation services. This task is particularly challenging due to complex spatio-temporal dependencies arising from non-Euclidean spatial relations in road networks and non-linear temporal dynamics influenced by changing road conditions. This paper introduces the spatio-temporal network embedding (STNE) model, a novel deep learning framework tailored for learning and forecasting graph-structured traffic data over extended input sequences. Unlike traditional convolutional neural networks (CNNs), the model employs graph convolutional networks (GCNs) to capture the spatial characteristics of local road network topologies. Moreover, the segmentation of very long input traffic data into multiple sub-sequences, based on significant temporal properties such as closeness, periodicity, and trend, is performed. Multi-dimensional long short-term memory neural networks (MDLSTM) are utilized to flexibly access multi-dimensional context. Experimental results demonstrate that the STNE model surpasses state-of-the-art traffic forecasting benchmarks on two large-scale real-world traffic datasets. Full article
Show Figures

Figure 1

Back to TopTop