Sign in to use this feature.

Years

Between: -

Search Results (574)

Search Parameters:
Keywords = AlexNet

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
16 pages, 4940 KiB  
Article
Potato Beetle Detection with Real-Time and Deep Learning
by Abdil Karakan
Processes 2024, 12(9), 2038; https://doi.org/10.3390/pr12092038 - 21 Sep 2024
Viewed by 280
Abstract
In this study, deep learning methods were used to detect potato beetles (Leptinotarsa decemlineata) on potato plants. High-resolution images were taken of fields with the help of a drone. Since these images were large in size, each one was divided into [...] Read more.
In this study, deep learning methods were used to detect potato beetles (Leptinotarsa decemlineata) on potato plants. High-resolution images were taken of fields with the help of a drone. Since these images were large in size, each one was divided into six equal parts. Then, according to the image, the potato beetles were divided into three classes: adult, late-stage potato beetle, and no beetles. A data set was created with 3000 images in each class, making 9000 in total. Different filters were applied to the images that made up the data set. In this way, problems that may have arisen from the camera in real-time detection were minimized. At the same time, the accuracy rate was increased. The created data set was used with six different deep learning models: MobileNet, InceptionV3, ResNet101, AlexNet, DenseNet121, and Xception. The deep learning models were tested with Sgd, Adam, and Rmsprop optimization methods and their performances were compared. In order to evaluate the success of the models more accurately, they were tested on a second data set created with images taken from a different field. As a result of this study, the highest accuracy of 99.81% was obtained. In the test results from a second field that did not exist in the data set, 92.95% accuracy was obtained. The average accuracy rate was 96.30%. Full article
(This article belongs to the Section Advanced Digital and Other Processes)
Show Figures

Figure 1

18 pages, 59323 KiB  
Article
Method for Augmenting Side-Scan Sonar Seafloor Sediment Image Dataset Based on BCEL1-CBAM-INGAN
by Haixing Xia, Yang Cui, Shaohua Jin, Gang Bian, Wei Zhang and Chengyang Peng
J. Imaging 2024, 10(9), 233; https://doi.org/10.3390/jimaging10090233 - 20 Sep 2024
Viewed by 308
Abstract
In this paper, a method for augmenting samples of side-scan sonar seafloor sediment images based on CBAM-BCEL1-INGAN is proposed, aiming to address the difficulties in acquiring and labeling datasets, as well as the insufficient diversity and quantity of data samples. Firstly, a Convolutional [...] Read more.
In this paper, a method for augmenting samples of side-scan sonar seafloor sediment images based on CBAM-BCEL1-INGAN is proposed, aiming to address the difficulties in acquiring and labeling datasets, as well as the insufficient diversity and quantity of data samples. Firstly, a Convolutional Block Attention Module (CBAM) is integrated into the residual blocks of the INGAN generator to enhance the learning of specific attributes and improve the quality of the generated images. Secondly, a BCEL1 loss function (combining binary cross-entropy and L1 loss functions) is introduced into the discriminator, enabling it to focus on both global image consistency and finer distinctions for better generation results. Finally, augmented samples are input into an AlexNet classifier to verify their authenticity. Experimental results demonstrate the excellent performance of the method in generating images of coarse sand, gravel, and bedrock, as evidenced by significant improvements in the Frechet Inception Distance (FID) and Inception Score (IS). The introduction of the CBAM and BCEL1 loss function notably enhances the quality and details of the generated images. Moreover, classification experiments using the AlexNet classifier show an increase in the recognition rate from 90.5% using only INGAN-generated images of bedrock to 97.3% using images augmented using our method, marking a 6.8% improvement. Additionally, the classification accuracy of bedrock-type matrices is improved by 5.2% when images enhanced using the method presented in this paper are added to the training set, which is 2.7% higher than that of the simple method amplification. This validates the effectiveness of our method in the task of generating seafloor sediment images, partially alleviating the scarcity of side-scan sonar seafloor sediment image data. Full article
(This article belongs to the Section Image and Video Processing)
Show Figures

Figure 1

20 pages, 4757 KiB  
Article
Combining Transfer Learning and Ensemble Algorithms for Improved Citrus Leaf Disease Classification
by Hongyan Zhu, Dani Wang, Yuzhen Wei, Xuran Zhang and Lin Li
Agriculture 2024, 14(9), 1549; https://doi.org/10.3390/agriculture14091549 - 7 Sep 2024
Viewed by 497
Abstract
Accurate categorization and timely control of leaf diseases are crucial for citrus growth. We proposed the Multi-Models Fusion Network (MMFN) for citrus leaf diseases detection based on model fusion and transfer learning. Compared to traditional methods, the algorithm (integrating transfer learning Alexnet, VGG, [...] Read more.
Accurate categorization and timely control of leaf diseases are crucial for citrus growth. We proposed the Multi-Models Fusion Network (MMFN) for citrus leaf diseases detection based on model fusion and transfer learning. Compared to traditional methods, the algorithm (integrating transfer learning Alexnet, VGG, and Resnet) we proposed can address the issues of limited categories, slow processing speed, and low recognition accuracy. By constructing efficient deep learning models and training and optimizing them with a large dataset of citrus leaf images, we ensured the broad applicability and accuracy of citrus leaf disease detection, achieving high-precision classification. Herein, various deep learning algorithms, including original Alexnet, VGG, Resnet, and transfer learning versions Resnet34 (Pre_Resnet34) and Resnet50 (Pre_Resnet50) were also discussed and compared. The results demonstrated that the MMFN model achieved an average accuracy of 99.72% in distinguishing between diseased and healthy leaves. Additionally, the model attained an average accuracy of 98.68% in the classification of multiple diseases (citrus huanglongbing (HLB), greasy spot disease and citrus canker), insect pests (citrus leaf miner), and deficiency disease (zinc deficiency). These findings conclusively illustrate that deep learning model fusion networks combining transfer learning and integration algorithms can automatically extract image features, enhance the automation and accuracy of disease recognition, demonstrate the significant potential and application value in citrus leaf disease classification, and potentially drive the development of smart agriculture. Full article
(This article belongs to the Special Issue Multi- and Hyper-Spectral Imaging Technologies for Crop Monitoring)
Show Figures

Figure 1

20 pages, 4755 KiB  
Article
Advanced Defect Detection in Wrap Film Products: A Hybrid Approach with Convolutional Neural Networks and One-Class Support Vector Machines with Variational Autoencoder-Derived Covariance Vectors
by Tatsuki Shimizu, Fusaomi Nagata, Maki K. Habib, Koki Arima, Akimasa Otsuka and Keigo Watanabe
Machines 2024, 12(9), 603; https://doi.org/10.3390/machines12090603 - 31 Aug 2024
Viewed by 427
Abstract
This study proposes a novel approach that utilizes Convolutional Neural Networks (CNNs) and Support Vector Machines (SVMs) to tackle a critical challenge: detecting defects in wrapped film products. With their delicate and reflective film wound around a core material, these products present formidable [...] Read more.
This study proposes a novel approach that utilizes Convolutional Neural Networks (CNNs) and Support Vector Machines (SVMs) to tackle a critical challenge: detecting defects in wrapped film products. With their delicate and reflective film wound around a core material, these products present formidable hurdles for conventional visual inspection systems. The complex task of identifying defects, such as unwound or protruding areas, remains a daunting endeavor. Despite the power of commercial image recognition systems, they struggle to capture anomalies within wrap film products. Our research methodology achieved a 90% defect detection accuracy, establishing its practical significance compared with existing methods. We introduce a pioneering methodology centered on covariance vectors extracted from latent variables, a product of a Variational Autoencoder (VAE). These covariance vectors serve as feature vectors for training a specialized One-Class SVM (OCSVM), a key component of our approach. Unlike conventional practices, our OCSVM does not require images containing defects for training; it uses defect-free images, thus circumventing the challenge of acquiring sufficient defect samples. We compare our methodology against feature vectors derived from the fully connected layers of established CNN models, AlexNet and VGG19, offering a comprehensive benchmarking perspective. Our research represents a significant advancement in defect detection technology. By harnessing the latent variable covariance vectors from a VAE encoder, our approach provides a unique solution to the challenges faced by commercial image recognition systems. These advancements in our study have the potential to revolutionize quality control mechanisms within manufacturing industries, offering a brighter future for product integrity and customer satisfaction. Full article
Show Figures

Figure 1

12 pages, 2079 KiB  
Article
Research on Default Classification of Unbalanced Credit Data Based on PixelCNN-WGAN
by Yutong Sun, Yanting Ji and Xiangxing Tao
Electronics 2024, 13(17), 3419; https://doi.org/10.3390/electronics13173419 - 28 Aug 2024
Viewed by 514
Abstract
Personal credit assessment plays a crucial role in the financial system, which not only relates to the financial activities of individuals but also affects the overall credit system and economic health of society. However, the current problem of data imbalance affecting classification results [...] Read more.
Personal credit assessment plays a crucial role in the financial system, which not only relates to the financial activities of individuals but also affects the overall credit system and economic health of society. However, the current problem of data imbalance affecting classification results in the field of personal credit assessment has not been fully solved. In order to solve this problem better, we propose a data-enhanced classification algorithm based on a Pixel Convolutional Neural Network (PixelCNN) and a Generative Adversarial Network (Wasserstein GAN, WGAN). Firstly, the historical data containing borrowers’ borrowing information are transformed into grayscale maps; then, data enhancement of default images is performed using the improved PixelCNN-WGAN model; and finally, the expanded image dataset is inputted into the CNN, AlexNet, SqueezeNet, and MobileNetV2 for classification. The results on the real dataset LendingClub show that the data enhancement algorithm designed in this paper improves the accuracy of the four algorithms by 1.548–3.568% compared with the original dataset, which can effectively improve the classification effect of the credit data, and to a certain extent, it provides a new idea for the classification task in the field of personal credit assessment. Full article
Show Figures

Figure 1

23 pages, 12258 KiB  
Article
Leakage Identification of Underground Structures Using Classification Deep Neural Networks and Transfer Learning
by Wenyang Wang, Qingwei Chen, Yongjiang Shen and Zhengliang Xiang
Sensors 2024, 24(17), 5569; https://doi.org/10.3390/s24175569 - 28 Aug 2024
Viewed by 402
Abstract
Water leakage defects often occur in underground structures, leading to accelerated structural aging and threatening structural safety. Leakage identification can detect early diseases of underground structures and provide important guidance for reinforcement and maintenance. Deep learning-based computer vision methods have been rapidly developed [...] Read more.
Water leakage defects often occur in underground structures, leading to accelerated structural aging and threatening structural safety. Leakage identification can detect early diseases of underground structures and provide important guidance for reinforcement and maintenance. Deep learning-based computer vision methods have been rapidly developed and widely used in many fields. However, establishing a deep learning model for underground structure leakage identification usually requires a lot of training data on leakage defects, which is very expensive. To overcome the data shortage, a deep neural network method for leakage identification is developed based on transfer learning in this paper. For comparison, four famous classification models, including VGG16, AlexNet, SqueezeNet, and ResNet18, are constructed. To train the classification models, a transfer learning strategy is developed, and a dataset of underground structure leakage is created. Finally, the classification performance on the leakage dataset of different deep learning models is comparatively studied under different sizes of training data. The results showed that the VGG16, AlexNet, and SqueezeNet models with transfer learning can overall provide higher and more stable classification performance on the leakage dataset than those without transfer learning. The ResNet18 model with transfer learning can overall provide a similar value of classification performance on the leakage dataset than that without transfer learning, but its classification performance is more stable than that without transfer learning. In addition, the SqueezeNet model obtains an overall higher and more stable performance than the comparative models on the leakage dataset for all classification metrics. Full article
Show Figures

Figure 1

15 pages, 1856 KiB  
Article
DaSAM: Disease and Spatial Attention Module-Based Explainable Model for Brain Tumor Detection
by Sara Tehsin, Inzamam Mashood Nasir, Robertas Damaševičius and Rytis Maskeliūnas
Big Data Cogn. Comput. 2024, 8(9), 97; https://doi.org/10.3390/bdcc8090097 - 25 Aug 2024
Viewed by 826
Abstract
Brain tumors are the result of irregular development of cells. It is a major cause of adult demise worldwide. Several deaths can be avoided with early brain tumor detection. Magnetic resonance imaging (MRI) for earlier brain tumor diagnosis may improve the chance of [...] Read more.
Brain tumors are the result of irregular development of cells. It is a major cause of adult demise worldwide. Several deaths can be avoided with early brain tumor detection. Magnetic resonance imaging (MRI) for earlier brain tumor diagnosis may improve the chance of survival for patients. The most common method of diagnosing brain tumors is MRI. The improved visibility of malignancies in MRI makes therapy easier. The diagnosis and treatment of brain cancers depend on their identification and treatment. Numerous deep learning models are proposed over the last decade including Alexnet, VGG, Inception, ResNet, DenseNet, etc. All these models are trained on a huge dataset, ImageNet. These general models have many parameters, which become irrelevant when implementing these models for a specific problem. This study uses a custom deep-learning model for the classification of brain MRIs. The proposed Disease and Spatial Attention Model (DaSAM) has two modules; (a) the Disease Attention Module (DAM), to distinguish between disease and non-disease regions of an image, and (b) the Spatial Attention Module (SAM), to extract important features. The experiments of the proposed model are conducted on two multi-class datasets that are publicly available, the Figshare and Kaggle datasets, where it achieves precision values of 99% and 96%, respectively. The proposed model is also tested using cross-dataset validation, where it achieved 85% accuracy when trained on the Figshare dataset and validated on the Kaggle dataset. The incorporation of DAM and SAM modules enabled the functionality of feature mapping, which proved to be useful for the highlighting of important features during the decision-making process of the model. Full article
Show Figures

Figure 1

24 pages, 2249 KiB  
Article
Enhancing Jujube Forest Growth Estimation and Disease Detection Using a Novel Diffusion-Transformer Architecture
by Xiangyi Hu, Zhihao Zhang, Liping Zheng, Tailai Chen, Chao Peng, Yilin Wang, Ruiheng Li, Xinyang Lv and Shuo Yan
Plants 2024, 13(17), 2348; https://doi.org/10.3390/plants13172348 - 23 Aug 2024
Viewed by 485
Abstract
This paper proposes an advanced deep learning model that integrates the Diffusion-Transformer structure and parallel attention mechanism for the tasks of growth estimation and disease detection in jujube forests. Existing methods in forestry monitoring often fall short in meeting the practical needs of [...] Read more.
This paper proposes an advanced deep learning model that integrates the Diffusion-Transformer structure and parallel attention mechanism for the tasks of growth estimation and disease detection in jujube forests. Existing methods in forestry monitoring often fall short in meeting the practical needs of large-scale and highly complex forest areas due to limitations in data processing capabilities and feature extraction precision. In response to this challenge, this paper designs and conducts a series of benchmark tests and ablation experiments to systematically evaluate and verify the performance of the proposed model across key performance metrics such as precision, recall, accuracy, and F1-score. Experimental results demonstrate that compared to traditional machine learning models like Support Vector Machines and Random Forests, as well as common deep learning models such as AlexNet and ResNet, the model proposed in this paper achieves a precision of 95%, a recall of 92%, an accuracy of 93%, and an F1-score of 94% in the task of disease detection in jujube forests, showing similarly superior performance in growth estimation tasks as well. Furthermore, ablation experiments with different attention mechanisms and loss functions further validate the effectiveness of parallel attention and parallel loss function in enhancing the overall performance of the model. These research findings not only provide a new technical path for forestry disease monitoring and health assessment but also contribute rich theoretical and experimental foundations for related fields. Full article
Show Figures

Figure 1

29 pages, 8768 KiB  
Article
HRIDM: Hybrid Residual/Inception-Based Deeper Model for Arrhythmia Detection from Large Sets of 12-Lead ECG Recordings
by Syed Atif Moqurrab, Hari Mohan Rai and Joon Yoo
Algorithms 2024, 17(8), 364; https://doi.org/10.3390/a17080364 - 19 Aug 2024
Cited by 2 | Viewed by 429
Abstract
Heart diseases such as cardiovascular and myocardial infarction are the foremost reasons of death in the world. The timely, accurate, and effective prediction of heart diseases is crucial for saving lives. Electrocardiography (ECG) is a primary non-invasive method to identify cardiac abnormalities. However, [...] Read more.
Heart diseases such as cardiovascular and myocardial infarction are the foremost reasons of death in the world. The timely, accurate, and effective prediction of heart diseases is crucial for saving lives. Electrocardiography (ECG) is a primary non-invasive method to identify cardiac abnormalities. However, manual interpretation of ECG recordings for heart disease diagnosis is a time-consuming and inaccurate process. For the accurate and efficient detection of heart diseases from the 12-lead ECG dataset, we have proposed a hybrid residual/inception-based deeper model (HRIDM). In this study, we have utilized ECG datasets from various sources, which are multi-institutional large ECG datasets. The proposed model is trained on 12-lead ECG data from over 10,000 patients. We have compared the proposed model with several state-of-the-art (SOTA) models, such as LeNet-5, AlexNet, VGG-16, ResNet-50, Inception, and LSTM, on the same training and test datasets. To show the effectiveness of the computational efficiency of the proposed model, we have only trained over 20 epochs without GPU support and we achieved an accuracy of 50.87% on the test dataset for 27 categories of heart abnormalities. We found that our proposed model outperformed the previous studies which participated in the official PhysioNet/CinC Challenge 2020 and achieved fourth place as compared with the 41 official ranking teams. The result of this study indicates that the proposed model is an implying new method for predicting heart diseases using 12-lead ECGs. Full article
Show Figures

Figure 1

10 pages, 1304 KiB  
Article
Age and Sex Estimation in Children and Young Adults Using Panoramic Radiographs with Convolutional Neural Networks
by Tuğçe Nur Şahin and Türkay Kölüş
Appl. Sci. 2024, 14(16), 7014; https://doi.org/10.3390/app14167014 - 9 Aug 2024
Viewed by 751
Abstract
Image processing with artificial intelligence has shown significant promise in various medical imaging applications. The present study aims to evaluate the performance of 16 different convolutional neural networks (CNNs) in predicting age and gender from panoramic radiographs in children and young adults. The [...] Read more.
Image processing with artificial intelligence has shown significant promise in various medical imaging applications. The present study aims to evaluate the performance of 16 different convolutional neural networks (CNNs) in predicting age and gender from panoramic radiographs in children and young adults. The networks tested included DarkNet-19, DarkNet-53, Inception-ResNet-v2, VGG-19, DenseNet-201, ResNet-50, GoogLeNet, VGG-16, SqueezeNet, ResNet-101, ResNet-18, ShuffleNet, MobileNet-v2, NasNet-Mobile, AlexNet, and Xception. These networks were trained on a dataset of 7336 radiographs from individuals aged between 5 and 21. Age and gender estimation accuracy and mean absolute age prediction errors were evaluated on 340 radiographs. Statistical analyses were conducted using Shapiro–Wilk, one-way ANOVA, and Tukey tests (p < 0.05). The gender prediction accuracy and the mean absolute age prediction error were, respectively, 87.94% and 0.582 for DarkNet-53, 86.18% and 0.427 for DarkNet-19, 84.71% and 0.703 for GoogLeNet, 81.76% and 0.756 for DenseNet-201, 81.76% and 1.115 for ResNet-18, 80.88% and 0.650 for VGG-19, 79.41% and 0.988 for SqueezeNet, 79.12% and 0.682 for Inception-Resnet-v2, 78.24% and 0.747 for ResNet-50, 77.35% and 1.047 for VGG-16, 76.47% and 1.109 for Xception, 75.88% and 0.977 for ResNet-101, 73.24% and 0.894 for ShuffleNet, 72.35% and 1.206 for AlexNet, 71.18% and 1.094 for NasNet-Mobile, and 62.94% and 1.327 for MobileNet-v2. No statistical difference in age prediction performance was found between DarkNet-19 and DarkNet-53, which demonstrated the most successful age estimation results. Despite these promising results, all tested CNNs performed below 90% accuracy and were not deemed suitable for clinical use. Future studies should continue with more-advanced networks and larger datasets. Full article
(This article belongs to the Special Issue Oral Diseases: Diagnosis and Therapy)
Show Figures

Figure 1

15 pages, 6283 KiB  
Article
Precision Detection of Salt Stress in Soybean Seedlings Based on Deep Learning and Chlorophyll Fluorescence Imaging
by Yixin Deng, Nan Xin, Longgang Zhao, Hongtao Shi, Limiao Deng, Zhongzhi Han and Guangxia Wu
Plants 2024, 13(15), 2089; https://doi.org/10.3390/plants13152089 - 27 Jul 2024
Viewed by 650
Abstract
Soil salinization poses a critical challenge to global food security, impacting plant growth, development, and crop yield. This study investigates the efficacy of deep learning techniques alongside chlorophyll fluorescence (ChlF) imaging technology for discerning varying levels of salt stress in soybean seedlings. Traditional [...] Read more.
Soil salinization poses a critical challenge to global food security, impacting plant growth, development, and crop yield. This study investigates the efficacy of deep learning techniques alongside chlorophyll fluorescence (ChlF) imaging technology for discerning varying levels of salt stress in soybean seedlings. Traditional methods for stress identification in plants are often laborious and time-intensive, prompting the exploration of more efficient approaches. A total of six classic convolutional neural network (CNN) models—AlexNet, GoogLeNet, ResNet50, ShuffleNet, SqueezeNet, and MobileNetv2—are evaluated for salt stress recognition based on three types of ChlF images. Results indicate that ResNet50 outperforms other models in classifying salt stress levels across three types of ChlF images. Furthermore, feature fusion after extracting three types of ChlF image features in the average pooling layer of ResNet50 significantly enhanced classification accuracy, achieving the highest accuracy of 98.61% in particular when fusing features from three types of ChlF images. UMAP dimensionality reduction analysis confirms the discriminative power of fused features in distinguishing salt stress levels. These findings underscore the efficacy of deep learning and ChlF imaging technologies in elucidating plant responses to salt stress, offering insights for precision agriculture and crop management. Overall, this study demonstrates the potential of integrating deep learning with ChlF imaging for precise and efficient crop stress detection, offering a robust tool for advancing precision agriculture. The findings contribute to enhancing agricultural sustainability and addressing global food security challenges by enabling more effective crop stress management. Full article
(This article belongs to the Special Issue Practical Applications of Chlorophyll Fluorescence Measurements)
Show Figures

Figure 1

20 pages, 5274 KiB  
Article
Innovative Research on Intelligent Recognition of Winter Jujube Defects by Applying Convolutional Neural Networks
by Jianjun Zhang, Weihui Wang and Qinglun Che
Electronics 2024, 13(15), 2941; https://doi.org/10.3390/electronics13152941 - 25 Jul 2024
Viewed by 545
Abstract
The current sorting process for winter jujubes relies heavily on manual labor, lacks uniform sorting standards, and is inefficient. Furthermore, existing devices have simple structures and can only be sorted based on size. This paper introduces a method for detecting surface defects on [...] Read more.
The current sorting process for winter jujubes relies heavily on manual labor, lacks uniform sorting standards, and is inefficient. Furthermore, existing devices have simple structures and can only be sorted based on size. This paper introduces a method for detecting surface defects on winter jujubes using convolutional neural networks (CNNs). According to the current situation in the winter jujube industry in Zhanhua District, Binzhou City, Shandong Province, China, we collected winter jujubes with different surface qualities in Zhanhua District; produced a winter jujube dataset containing 2000 winter jujube images; improved it based on the traditional AlexNet model; selected a total of four classical convolutional neural networks, AlexNet, VGG-16, Inception-V3, and ResNet-34, to conduct different learning rate comparison training experiments; and then took the accuracy rate, loss value, and F1-score of the validation set as evaluation indexes while analyzing and discussing the training results of each model. The experimental results show that the improved AlexNet model had the highest accuracy in the binary classification case, with an accuracy of 98% on the validation set; the accuracy of the Inception V3 model reached 97%. In the detailed classification case, the accuracy of the Inception V3 model was 95%. Different models have different performances and different hardware requirements, and different models can be used to build the system according to different needs. This study can provide a theoretical basis and technical reference for researching and developing winter jujube detection devices. Full article
(This article belongs to the Special Issue Convolutional Neural Networks and Vision Applications, 3rd Edition)
Show Figures

Figure 1

12 pages, 3623 KiB  
Article
Trajectory Classification and Recognition of Planar Mechanisms Based on ResNet18 Network
by Jianping Wang, Youchao Wang, Boyan Chen, Xiaoyue Jia and Dexi Pu
Algorithms 2024, 17(8), 324; https://doi.org/10.3390/a17080324 - 25 Jul 2024
Viewed by 549
Abstract
This study utilizes the ResNet18 network to classify and recognize trajectories of planar mechanisms. This research begins by deriving formulas for trajectory points in various typical planar mechanisms, and the resulting trajectory images are employed as samples for training and testing the network. [...] Read more.
This study utilizes the ResNet18 network to classify and recognize trajectories of planar mechanisms. This research begins by deriving formulas for trajectory points in various typical planar mechanisms, and the resulting trajectory images are employed as samples for training and testing the network. The classification of trajectory images for both upright and inverted configurations of a planar four-bar linkage is investigated. Compared with AlexNet and VGG16, the ResNet18 model demonstrates superior classification accuracy during testing, coupled with reduced training time and memory consumption. Furthermore, the ResNet18 model is applied to classify trajectory images for six different planar mechanisms in both upright and inverted configurations as well as to identify whether the trajectory images belong to the upright or inverted configuration for each mechanism. The test results affirm the feasibility and effectiveness of the ResNet18 network in the classification and recognition of planar mechanism trajectories. Full article
Show Figures

Figure 1

16 pages, 6945 KiB  
Article
Deep Learning for High-Speed Lightning Footage—A Semantic Segmentation Network Comparison
by Tyson Cross, Jason R. Smit, Carina Schumann, Tom A. Warner and Hugh G. P. Hunt
Atmosphere 2024, 15(8), 873; https://doi.org/10.3390/atmos15080873 - 23 Jul 2024
Viewed by 578
Abstract
We present a novel deep learning approach to a unique image processing application: high-speed (>1000 fps) video footage of lightning. High-speed cameras enable us to observe lightning with microsecond resolution, characterizing key processes previously analyzed manually. We evaluate different semantic segmentation networks (DeepLab3+, [...] Read more.
We present a novel deep learning approach to a unique image processing application: high-speed (>1000 fps) video footage of lightning. High-speed cameras enable us to observe lightning with microsecond resolution, characterizing key processes previously analyzed manually. We evaluate different semantic segmentation networks (DeepLab3+, SegNet, FCN8s, U-Net, and AlexNet) and provide a detailed explanation of the image processing methods for this unique imagery. Our system architecture includes an input image processing stage, a segmentation network stage, and a sequence classification stage. The ground-truth data consists of high-speed videos of lightning filmed in South Africa, totaling 48,381 labeled frames. DeepLab3+ performed the best (93–95% accuracy), followed by SegNet (92–95% accuracy) and FCN8s (89–90% accuracy). AlexNet and U-Net achieved below 80% accuracy. Full sequence classification was 48.1% and stroke classification was 74.1%, due to the linear dependence on the segmentation. We recommend utilizing exposure metadata to improve noise misclassifications and extending CNNs to use tapped gates with temporal memory. This work introduces a novel deep learning application to lightning imagery and is one of the first studies on high-speed video footage using deep learning. Full article
(This article belongs to the Special Issue Recent Advances in Lightning Research)
Show Figures

Figure 1

31 pages, 8378 KiB  
Article
A Transfer Learning Approach: Early Prediction of Alzheimer’s Disease on US Healthy Aging Dataset
by Kishor Kumar Reddy C, Aarti Rangarajan, Deepti Rangarajan, Mohammed Shuaib, Fathe Jeribi and Shadab Alam
Mathematics 2024, 12(14), 2204; https://doi.org/10.3390/math12142204 - 13 Jul 2024
Viewed by 768
Abstract
Alzheimer’s disease (AD) is a growing public health crisis, a very global health concern, and an irreversible progressive neurodegenerative disorder of the brain for which there is still no cure. Globally, it accounts for 60–80% of dementia cases, thereby raising the need for [...] Read more.
Alzheimer’s disease (AD) is a growing public health crisis, a very global health concern, and an irreversible progressive neurodegenerative disorder of the brain for which there is still no cure. Globally, it accounts for 60–80% of dementia cases, thereby raising the need for an accurate and effective early classification. The proposed work used a healthy aging dataset from the USA and focused on three transfer learning approaches: VGG16, VGG19, and Alex Net. This work leveraged how the convolutional model and pooling layers work to improve and reduce overfitting, despite challenges in training the numerical dataset. VGG was preferably chosen as a hidden layer as it has a more diverse, deeper, and simpler architecture with better performance when dealing with larger datasets. It consumes less memory and training time. A comparative analysis was performed using machine learning and neural network algorithm techniques. Performance metrics such as accuracy, error rate, precision, recall, F1 score, sensitivity, specificity, kappa statistics, ROC, and RMSE were experimented with and compared. The accuracy was 100% for VGG16 and VGG19 and 98.20% for Alex Net. The precision was 99.9% for VGG16, 96.6% for VGG19, and 100% for Alex Net; the recall values were 99.9% for all three cases of VGG16, VGG19, and Alex Net; and the sensitivity metric was 96.8% for VGG16, 97.9% for VGG19, and 98.7% for Alex Net, which has outperformed when compared with the existing approaches for the classification of Alzheimer’s disease. This research contributes to the advancement of predictive knowledge, leading to future empirical evaluation, experimentation, and testing in the biomedical field. Full article
Show Figures

Figure 1

Back to TopTop