Sign in to use this feature.

Years

Between: -

Search Results (1,402)

Search Parameters:
Keywords = BERT

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
15 pages, 803 KiB  
Article
Exploration of Deep-Learning-Based Approaches for False Fact Identification in Social Judicial Systems
by Yuzhuo Zou, Jiepin Chen, Jiebin Cai, Mengen Zhou and Yinghui Pan
Electronics 2024, 13(19), 3831; https://doi.org/10.3390/electronics13193831 - 27 Sep 2024
Abstract
With the many applications of artificial intelligence (AI) in social judicial systems, false fact identification becomes a challenging issue when the system is expected to be more autonomous and intelligent in assisting a judicial review. In particular, private lending disputes often involve false [...] Read more.
With the many applications of artificial intelligence (AI) in social judicial systems, false fact identification becomes a challenging issue when the system is expected to be more autonomous and intelligent in assisting a judicial review. In particular, private lending disputes often involve false facts that are intentionally concealed and manipulated due to unique and dynamic relationships and their nonconfrontational nature in the judicial system. In this article, we investigate deep learning techniques to identify false facts in loan cases for the purpose of reducing the judicial workload. Specifically, we adapt deep-learning-based natural language processing techniques to a dataset over 100 real-world judicial rules spanning four courts of different levels in China. The BERT (bidirectional encoder representations from transformers)-based classifier and T5 text generation models were trained to classify false litigation claims semantically. The experimental results demonstrate that T5 has a robust learning capability with a small number of legal text samples, outperforms BERT in identifying falsified facts, and provides explainable decisions to judges. This research shows that deep-learning-based false fact identification approaches provide promising solutions for addressing concealed information and manipulation in private lending lawsuits. This highlights the feasibility of deep learning to strengthen fact-finding and reduce labor costs in the judicial field. Full article
(This article belongs to the Special Issue Data-Driven Intelligence in Autonomous Systems)
15 pages, 8308 KiB  
Article
The Diatom Genus Navicula in Spring Ecosystems with the Description of Navicula aquaesuavis sp. nov.
by María Cid-Rodríguez, Marco Cantonati, Nicola Angeli, Olena Bilous, Maha Al-Harbi, Horst Lange-Bertalot, Zlatko Levkov, Lucia Piana, Daniel Spitale and Abdullah A. Saber
Water 2024, 16(19), 2751; https://doi.org/10.3390/w16192751 - 27 Sep 2024
Abstract
Given the limited understanding of species diversity and ecological preferences of diatoms of the genus Navicula in spring ecosystems, herein we present and describe as species new to science, Navicula aquaesuavis Lange-Bert., Levkov, Cid-Rodríguez, A.A.Saber and Cantonati sp. nov. This species was collected [...] Read more.
Given the limited understanding of species diversity and ecological preferences of diatoms of the genus Navicula in spring ecosystems, herein we present and describe as species new to science, Navicula aquaesuavis Lange-Bert., Levkov, Cid-Rodríguez, A.A.Saber and Cantonati sp. nov. This species was collected from a mountain spring located above the tree line at 1613 m a.s.l. in the Northern Apennines. The Fontana del Vescovo (Bishop’s spring), which is the locus classicus of the new species, has a low conductivity (60–70 µS cm−1), temperature of ca. 5 °C, circumneutral pH (7.3–7.5), relatively low nitrate (ca. 1 mg L−1), and also suffered from a discharge reduction from 1 to 0.1 L s−1 from 2011 to 2023. The putative new species was confirmed by a second finding in Northern Macedonia, and we thoroughly document this second population as well. We seized the opportunity to describe this new Navicula and review the global literature on the diatom genus Navicula in spring ecosystems. Using the results of this review and our own databases on springs and wells in central Europe and Egypt, we discuss the main Navicula species and their environmental preferences in spring habitats. Full article
(This article belongs to the Special Issue Biodiversity of Freshwater Ecosystems: Monitoring and Conservation)
Show Figures

Figure 1

14 pages, 1576 KiB  
Article
Language Model-Based Text Augmentation System for Cerebrovascular Disease Related Medical Report
by Yu-Hyeon Kim, Chulho Kim and Yu-Seop Kim
Appl. Sci. 2024, 14(19), 8652; https://doi.org/10.3390/app14198652 - 25 Sep 2024
Abstract
Texts in medical fields containing sensitive information pose challenges for AI research usability. However, there is increasing interest in generating synthetic text to make medical text data bigger for text-based medical AI research. Therefore, this paper suggests a text augmentation system for cerebrovascular [...] Read more.
Texts in medical fields containing sensitive information pose challenges for AI research usability. However, there is increasing interest in generating synthetic text to make medical text data bigger for text-based medical AI research. Therefore, this paper suggests a text augmentation system for cerebrovascular diseases, using a synthetic text generation model based on DistilGPT2 and a classification model based on BioBERT. The synthetic text generation model generates synthetic text using randomly extracted reports (5000, 10,000, 15,000, and 20,000) from 73,671 reports. The classification model is fine-tuned with the entire report to annotate synthetic text and build a new dataset. Subsequently, we fine-tuned a classification model by incrementally increasing the amount of augmented data added to each original dataset. Experimental results show that fine-tuning by adding augmented data improves model performance by up to 20%. Furthermore, we found that generating a large amount of synthetic text is not necessarily required to achieve better performance, and the appropriate amount of data augmentation depends on the size of the original data. Therefore, our proposed method reduces the time and resources needed for dataset construction, automating the annotation task and generating meaningful synthetic text for medical AI research. Full article
(This article belongs to the Topic Innovation, Communication and Engineering)
Show Figures

Figure 1

27 pages, 2051 KiB  
Article
A Transparent Pipeline for Identifying Sexism in Social Media: Combining Explainability with Model Prediction
by Hadi Mohammadi, Anastasia Giachanou and Ayoub Bagheri
Appl. Sci. 2024, 14(19), 8620; https://doi.org/10.3390/app14198620 - 24 Sep 2024
Abstract
In this study, we present a new approach that combines multiple Bidirectional Encoder Representations from Transformers (BERT) architectures with a Convolutional Neural Network (CNN) framework designed for sexism detection in text at a granular level. Our method relies on the analysis and identification [...] Read more.
In this study, we present a new approach that combines multiple Bidirectional Encoder Representations from Transformers (BERT) architectures with a Convolutional Neural Network (CNN) framework designed for sexism detection in text at a granular level. Our method relies on the analysis and identification of the most important terms contributing to sexist content using Shapley Additive Explanations (SHAP) values. This approach involves defining a range of Sexism Scores based on both model predictions and explainability, moving beyond binary classification to provide a deeper understanding of the sexism-detection process. Additionally, it enables us to identify specific parts of a sentence and their respective contributions to this range, which can be valuable for decision makers and future research. In conclusion, this study introduces an innovative method for enhancing the clarity of large language models (LLMs), which is particularly relevant in sensitive domains such as sexism detection. The incorporation of explainability into the model represents a significant advancement in this field. The objective of our study is to bridge the gap between advanced technology and human comprehension by providing a framework for creating AI models that are both efficient and transparent. This approach could serve as a pipeline for future studies to incorporate explainability into language models. Full article
(This article belongs to the Special Issue Data and Text Mining: New Approaches, Achievements and Applications)
Show Figures

Figure 1

18 pages, 2249 KiB  
Article
Fractal Self-Similarity in Semantic Convergence: Gradient of Embedding Similarity across Transformer Layers
by Minhyeok Lee
Fractal Fract. 2024, 8(10), 552; https://doi.org/10.3390/fractalfract8100552 - 24 Sep 2024
Abstract
This paper presents a mathematical analysis of semantic convergence in transformer-based language models, drawing inspiration from the concept of fractal self-similarity. We introduce and prove a novel theorem characterizing the gradient of embedding similarity across layers. Specifically, we establish that there exists a [...] Read more.
This paper presents a mathematical analysis of semantic convergence in transformer-based language models, drawing inspiration from the concept of fractal self-similarity. We introduce and prove a novel theorem characterizing the gradient of embedding similarity across layers. Specifically, we establish that there exists a monotonically increasing function that provides a lower bound on the rate at which the average cosine similarity between token embeddings at consecutive layers and the final layer increases. This establishes a fundamental property: semantic alignment of token representations consistently increases through the network, exhibiting a pattern of progressive refinement, analogous to fractal self-similarity. The key challenge addressed is the quantification and generalization of semantic convergence across diverse model architectures and input contexts. To validate our findings, we conduct experiments on BERT and DistilBERT models, analyzing embedding similarities for diverse input types. While our experiments are limited to these models, we empirically demonstrate consistent semantic convergence within these architectures. Quantitatively, we find that the average rates of semantic convergence are approximately 0.0826 for BERT and 0.1855 for DistilBERT. We observe that the rate of convergence varies based on token frequency and model depth, with rare words showing slightly higher similarities (differences of approximately 0.0167 for BERT and 0.0120 for DistilBERT). This work advances our understanding of transformer models’ internal mechanisms and provides a mathematical framework for comparing and optimizing model architectures. Full article
Show Figures

Figure 1

16 pages, 2463 KiB  
Article
Binning Metagenomic Contigs Using Contig Embedding and Decomposed Tetranucleotide Frequency
by Long Fu, Jiabin Shi and Baohua Huang
Biology 2024, 13(10), 755; https://doi.org/10.3390/biology13100755 - 24 Sep 2024
Abstract
Metagenomic binning is a crucial step in metagenomic research. It can aggregate the genome sequences belonging to the same microbial species into independent bins. Most existing methods ignore the semantic information of contigs and lack effective processing of tetranucleotide frequency, resulting in insufficient [...] Read more.
Metagenomic binning is a crucial step in metagenomic research. It can aggregate the genome sequences belonging to the same microbial species into independent bins. Most existing methods ignore the semantic information of contigs and lack effective processing of tetranucleotide frequency, resulting in insufficient and complex feature information extracted for binning and poor binning results. To address the above problems, we propose CedtBin, a metagenomic binning method based on contig embedding and decomposed tetranucleotide frequency. First, the improved BERT model is used to learn the contigs to obtain their embedding representation. Secondly, the tetranucleotide frequencies are decomposed using a non-negative matrix factorization (NMF) algorithm. After that, the two features are spliced and input into the clustering algorithm for binning. Considering the sensitivity of the DBSCAN clustering algorithm to input parameters, in order to solve the drawbacks of manual parameter input, we also propose an Annoy-DBSCAN algorithm that can adaptively determine the parameters of the DBSCAN algorithm. This algorithm uses Approximate Nearest Neighbors Oh Yeah (Annoy) and combines it with a grid search strategy to find the optimal parameters of the DBSCAN algorithm. On simulated and real datasets, CedtBin achieves better binning results than mainstream methods and can reconstruct more genomes, indicating that the proposed method is effective. Full article
(This article belongs to the Special Issue 2nd Edition of Computational Methods in Biology)
Show Figures

Figure 1

14 pages, 9755 KiB  
Article
Phoneme Recognition in Korean Singing Voices Using Self-Supervised English Speech Representations
by Wenqin Wu and Joonwhoan Lee
Appl. Sci. 2024, 14(18), 8532; https://doi.org/10.3390/app14188532 - 22 Sep 2024
Abstract
In general, it is difficult to obtain a huge, labeled dataset for deep learning-based phoneme recognition in singing voices. Studying singing voices also offers inherent challenges, compared to speech, because of the distinct variations in pitch, duration, and intensity. This paper proposes a [...] Read more.
In general, it is difficult to obtain a huge, labeled dataset for deep learning-based phoneme recognition in singing voices. Studying singing voices also offers inherent challenges, compared to speech, because of the distinct variations in pitch, duration, and intensity. This paper proposes a detouring method to overcome this insufficient dataset, and applies it to the recognition of Korean phonemes in singing voices. The method started with pre-training the HuBERT, a self-supervised speech representation model, on a large-scale English corpus. The model was then adapted to the Korean speech domain with a relatively small-scale Korean corpus, in which the Korean phonemes were interpreted as similar English ones. Finally, the speech-adapted model was again trained with a tiny-scale Korean singing voice corpus for speech–singing adaptation. In the final adaptation, melodic supervision was chosen, which utilizes pitch information to improve the performance. For evaluation, the performance on multi-level error rates based on Word Error Rate (WER) was taken. Using the HuBERT-based transfer learning for adaptation improved the phoneme-level error rate of Korean speech by as much as 31.19%. Again, on singing voices by melodic supervision, it improved the rate by 0.55%. The significant improvement in speech recognition underscores the considerable potential of a model equipped with general human voice representations captured from the English corpus that can improve phoneme recognition on less target speech data. Moreover, the musical variation in singing voices is beneficial for phoneme recognition in singing voices. The proposed method could be applied to the phoneme recognition of other languages that have less speech and singing voice corpora. Full article
Show Figures

Figure 1

17 pages, 791 KiB  
Article
Pre-Trained Language Model Ensemble for Arabic Fake News Detection
by Lama Al-Zahrani and Maha Al-Yahya
Mathematics 2024, 12(18), 2941; https://doi.org/10.3390/math12182941 - 21 Sep 2024
Abstract
Fake news detection (FND) remains a challenge due to its vast and varied sources, especially on social media platforms. While numerous attempts have been made by academia and the industry to develop fake news detection systems, research on Arabic content remains limited. This [...] Read more.
Fake news detection (FND) remains a challenge due to its vast and varied sources, especially on social media platforms. While numerous attempts have been made by academia and the industry to develop fake news detection systems, research on Arabic content remains limited. This study investigates transformer-based language models for Arabic FND. While transformer-based models have shown promising performance in various natural language processing tasks, they often struggle with tasks involving complex linguistic patterns and cultural contexts, resulting in unreliable performance and misclassification problems. To overcome these challenges, we investigated an ensemble of transformer-based models. We experimented with five Arabic transformer models: AraBERT, MARBERT, AraELECTRA, AraGPT2, and ARBERT. Various ensemble approaches, including a weighted-average ensemble, hard voting, and soft voting, were evaluated to determine the most effective techniques for boosting learning models and improving prediction accuracies. The results of this study demonstrate the effectiveness of ensemble models in significantly boosting the baseline model performance. An important finding is that ensemble models achieved excellent performance on the Arabic Multisource Fake News Detection (AMFND) dataset, reaching an F1 score of 94% using weighted averages. Moreover, changing the number of models in the ensemble has a slight effect on the performance. These key findings contribute to the advancement of fake news detection in Arabic, offering valuable insights for both academia and the industry Full article
Show Figures

Figure 1

25 pages, 6051 KiB  
Article
Cross-Task Rumor Detection: Model Optimization Based on Model Transfer Learning and Graph Convolutional Neural Networks (GCNs)
by Wen Jiang, Facheng Yan, Kelan Ren, Xiong Zhang, Bin Wei and Mingshu Zhang
Electronics 2024, 13(18), 3757; https://doi.org/10.3390/electronics13183757 - 21 Sep 2024
Abstract
With the widespread adoption of social media, the rapid dissemination of rumors poses a severe threat to public perception and social stability, emerging as a major challenge confronting society. Hence, the development of efficient and accurate rumor detection models has become an urgent [...] Read more.
With the widespread adoption of social media, the rapid dissemination of rumors poses a severe threat to public perception and social stability, emerging as a major challenge confronting society. Hence, the development of efficient and accurate rumor detection models has become an urgent need. Given the challenges of rumor detection tasks, including data scarcity, feature complexity, and difficulties in cross-task knowledge transfer, this paper proposes a BERT–GCN–Transfer Learning model, an integrated rumor detection model that combines BERT (Bidirectional Encoder Representations from Transformers), Graph Convolutional Networks (GCNs), and transfer learning techniques. By harnessing BERT’s robust text representation capabilities, the GCN’s feature extraction prowess on graph-structured data, and the advantage of transfer learning in cross-task knowledge sharing, the model achieves effective rumor detection on social media platforms. Experimental results indicate that this model achieves accuracies of 0.878 and 0.892 on the Twitter15 and Twitter16 datasets, respectively, significantly enhancing the accuracy of rumor detection compared to baseline models. Moreover, it greatly improves the efficiency of model training. Full article
Show Figures

Figure 1

13 pages, 787 KiB  
Article
Chinese Medical Named Entity Recognition Based on Context-Dependent Perception and Novel Memory Units
by Yufeng Kang, Yang Yan and Wenbo Huang
Appl. Sci. 2024, 14(18), 8471; https://doi.org/10.3390/app14188471 - 20 Sep 2024
Abstract
Medical named entity recognition (NER) focuses on extracting and classifying key entities from medical texts. Through automated medical information extraction, NER can effectively improve the efficiency of electronic medical record analysis, medical literature retrieval, and intelligent medical question–answering systems, enabling doctors and researchers [...] Read more.
Medical named entity recognition (NER) focuses on extracting and classifying key entities from medical texts. Through automated medical information extraction, NER can effectively improve the efficiency of electronic medical record analysis, medical literature retrieval, and intelligent medical question–answering systems, enabling doctors and researchers to obtain the required medical information more quickly and thereby helping to improve the accuracy of diagnosis and treatment decisions. The current methods have certain limitations in dealing with contextual dependencies and entity memory and fail to fully consider the contextual relevance and interactivity between entities. To address these issues, this paper proposes a Chinese medical named entity recognition model that combines contextual dependency perception and a new memory unit. The model combines the BERT pre-trained model with a new memory unit (GLMU) and a recall network (RMN). The GLMU can efficiently capture long-distance dependencies, while the RMN enhances multi-level semantic information processing. The model also incorporates fully connected layers (FC) and conditional random fields (CRF) to further optimize the performance of entity classification and sequence labeling. The experimental results show that the model achieved F1 values of 91.53% and 64.92% on the Chinese medical datasets MCSCSet and CMeEE, respectively, surpassing other related models and demonstrating significant advantages in the field of medical entity recognition. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

20 pages, 1391 KiB  
Article
A Hybrid Approach to Dimensional Aspect-Based Sentiment Analysis Using BERT and Large Language Models
by Yice Zhang, Hongling Xu, Delong Zhang and Ruifeng Xu
Electronics 2024, 13(18), 3724; https://doi.org/10.3390/electronics13183724 - 19 Sep 2024
Abstract
Dimensional aspect-based sentiment analysis (dimABSA) aims to recognize aspect-level quadruples from reviews, offering a fine-grained sentiment description for user opinions. A quadruple consists of aspect, category, opinion, and sentiment intensity, which is represented using continuous real-valued scores in the valence–arousal dimensions. To address [...] Read more.
Dimensional aspect-based sentiment analysis (dimABSA) aims to recognize aspect-level quadruples from reviews, offering a fine-grained sentiment description for user opinions. A quadruple consists of aspect, category, opinion, and sentiment intensity, which is represented using continuous real-valued scores in the valence–arousal dimensions. To address this task, we propose a hybrid approach that integrates the BERT model with a large language model (LLM). Firstly, we develop both the BERT-based and LLM-based methods for dimABSA. The BERT-based method employs a pipeline approach, while the LLM-based method transforms the dimABSA task into a text generation task. Secondly, we evaluate their performance in entity extraction, relation classification, and intensity prediction to determine their advantages. Finally, we devise a hybrid approach to fully utilize their advantages across different scenarios. Experiments demonstrate that the hybrid approach outperforms BERT-based and LLM-based methods, achieving state-of-the-art performance with an F1-score of 41.7% on the quadruple extraction. Full article
(This article belongs to the Special Issue New Advances in Affective Computing)
Show Figures

Figure 1

13 pages, 3143 KiB  
Article
Ensemble Techniques for Robust Fake News Detection: Integrating Transformers, Natural Language Processing, and Machine Learning
by Mohammed Al-alshaqi, Danda B. Rawat and Chunmei Liu
Sensors 2024, 24(18), 6062; https://doi.org/10.3390/s24186062 - 19 Sep 2024
Abstract
The proliferation of fake news across multiple modalities has emerged as a critical challenge in the modern information landscape, necessitating advanced detection methods. This study proposes a comprehensive framework for fake news detection integrating text, images, and videos using machine learning and deep [...] Read more.
The proliferation of fake news across multiple modalities has emerged as a critical challenge in the modern information landscape, necessitating advanced detection methods. This study proposes a comprehensive framework for fake news detection integrating text, images, and videos using machine learning and deep learning techniques. The research employs a dual-phased methodology, first analyzing textual data using various classifiers, then developing a multimodal approach combining BERT for text analysis and a modified CNN for visual data. Experiments on the ISOT fake news dataset and MediaEval 2016 image verification corpus demonstrate the effectiveness of the proposed models. For textual data, the Random Forest classifier achieved 99% accuracy, outperforming other algorithms. The multimodal approach showed superior performance compared to baseline models, with a 3.1% accuracy improvement over existing multimodal techniques. This research contributes to the ongoing efforts to combat misinformation by providing a robust, adaptable framework for detecting fake news across different media formats, addressing the complexities of modern information dissemination and manipulation. Full article
(This article belongs to the Section Internet of Things)
Show Figures

Figure 1

21 pages, 1323 KiB  
Article
Exploring Players’ Perspectives: A Comprehensive Topic Modeling Case Study on Elden Ring
by Fatemeh Dehghani and Loutfouz Zaman
Information 2024, 15(9), 573; https://doi.org/10.3390/info15090573 - 18 Sep 2024
Abstract
Game reviews heavily influence public perception. User feedback is crucial for developers, offering valuable insights to enhance game quality. In this research, Metacritic game reviews for Elden Ring were analyzed for topic modeling using Latent Dirichlet Allocation (LDA), Bidirectional Encoder Representations from Transformers [...] Read more.
Game reviews heavily influence public perception. User feedback is crucial for developers, offering valuable insights to enhance game quality. In this research, Metacritic game reviews for Elden Ring were analyzed for topic modeling using Latent Dirichlet Allocation (LDA), Bidirectional Encoder Representations from Transformers (BERT), and a hybrid model combining both to identify effective methods for extracting underlying themes in player feedback. We analyzed and interpreted these models’ outputs to learn the game reviews. We aimed to identify the differences, similarities, and variations between the three to determine which provided more valuable and instructive information. Our findings indicate that each method successfully identified keywords with some similarities in identified words. The LDA model had the highest silhouette score, indicating the most distinct clustering. The LDA-BERT model had a 1% higher coherence score than LDA, indicating more meaningful topics. Full article
(This article belongs to the Special Issue Artificial Intelligence Methods for Human-Computer Interaction)
Show Figures

Graphical abstract

21 pages, 6745 KiB  
Article
Multimodal Social Media Fake News Detection Based on 1D-CCNet Attention Mechanism
by Yuhan Yan, Haiyan Fu and Fan Wu
Electronics 2024, 13(18), 3700; https://doi.org/10.3390/electronics13183700 - 18 Sep 2024
Abstract
Due to the explosive rise of multimodal content in online social communities, cross-modal learning is crucial for accurate fake news detection. However, current multimodal fake news detection techniques face challenges in extracting features from multiple modalities and fusing cross-modal information, failing to fully [...] Read more.
Due to the explosive rise of multimodal content in online social communities, cross-modal learning is crucial for accurate fake news detection. However, current multimodal fake news detection techniques face challenges in extracting features from multiple modalities and fusing cross-modal information, failing to fully exploit the correlations and complementarities between different modalities. To address these issues, this paper proposes a fake news detection model based on a one-dimensional CCNet (1D-CCNet) attention mechanism, named BTCM. This method first utilizes BERT and BLIP-2 encoders to extract text and image features. Then, it employs the proposed 1D-CCNet attention mechanism module to process the input text and image sequences, enhancing the important aspects of the bimodal features. Meanwhile, this paper uses the pre-trained BLIP-2 model for object detection in images, generating image descriptions and augmenting text data to enhance the dataset. This operation aims to further strengthen the correlations between different modalities. Finally, this paper proposes a heterogeneous cross-feature fusion method (HCFFM) to integrate image and text features. Comparative experiments were conducted on three public datasets: Twitter, Weibo, and Gossipcop. The results show that the proposed model achieved excellent performance. Full article
(This article belongs to the Special Issue Application of Data Mining in Social Media)
Show Figures

Figure 1

26 pages, 6325 KiB  
Article
Improving the Accuracy and Effectiveness of Text Classification Based on the Integration of the Bert Model and a Recurrent Neural Network (RNN_Bert_Based)
by Chanthol Eang and Seungjae Lee
Appl. Sci. 2024, 14(18), 8388; https://doi.org/10.3390/app14188388 - 18 Sep 2024
Abstract
This paper proposes a new robust model for text classification on the Stanford Sentiment Treebank v2 (SST-2) dataset in terms of model accuracy. We developed a Recurrent Neural Network Bert based (RNN_Bert_based) model designed to improve classification accuracy on the SST-2 dataset. This [...] Read more.
This paper proposes a new robust model for text classification on the Stanford Sentiment Treebank v2 (SST-2) dataset in terms of model accuracy. We developed a Recurrent Neural Network Bert based (RNN_Bert_based) model designed to improve classification accuracy on the SST-2 dataset. This dataset consists of movie review sentences, each labeled with either positive or negative sentiment, making it a binary classification task. Recurrent Neural Networks (RNNs) are effective for text classification because they capture the sequential nature of language, which is crucial for understanding context and meaning. Bert excels in text classification by providing bidirectional context, generating contextual embeddings, and leveraging pre-training on large corpora. This allows Bert to capture nuanced meanings and relationships within the text effectively. Combining Bert with RNNs can be highly effective for text classification. Bert’s bidirectional context and rich embeddings provide a deep understanding of the text, while RNNs capture sequential patterns and long-range dependencies. Together, they leverage the strengths of both architectures, leading to improved performance on complex classification tasks. Next, we also developed an integration of the Bert model and a K-Nearest Neighbor based (KNN_Bert_based) method as a comparative scheme for our proposed work. Based on the results of experimentation, our proposed model outperforms traditional text classification models as well as existing models in terms of accuracy. Full article
(This article belongs to the Special Issue Natural Language Processing: Novel Methods and Applications)
Show Figures

Figure 1

Back to TopTop