[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (8)

Search Parameters:
Keywords = Portuguese BERT

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
29 pages, 6331 KiB  
Article
Multimodal Affective Communication Analysis: Fusing Speech Emotion and Text Sentiment Using Machine Learning
by Diego Resende Faria, Abraham Itzhak Weinberg and Pedro Paulo Ayrosa
Appl. Sci. 2024, 14(15), 6631; https://doi.org/10.3390/app14156631 - 29 Jul 2024
Viewed by 663
Abstract
Affective communication, encompassing verbal and non-verbal cues, is crucial for understanding human interactions. This study introduces a novel framework for enhancing emotional understanding by fusing speech emotion recognition (SER) and sentiment analysis (SA). We leverage diverse features and both classical and deep learning [...] Read more.
Affective communication, encompassing verbal and non-verbal cues, is crucial for understanding human interactions. This study introduces a novel framework for enhancing emotional understanding by fusing speech emotion recognition (SER) and sentiment analysis (SA). We leverage diverse features and both classical and deep learning models, including Gaussian naive Bayes (GNB), support vector machines (SVMs), random forests (RFs), multilayer perceptron (MLP), and a 1D convolutional neural network (1D-CNN), to accurately discern and categorize emotions in speech. We further extract text sentiment from speech-to-text conversion, analyzing it using pre-trained models like bidirectional encoder representations from transformers (BERT), generative pre-trained transformer 2 (GPT-2), and logistic regression (LR). To improve individual model performance for both SER and SA, we employ an extended dynamic Bayesian mixture model (DBMM) ensemble classifier. Our most significant contribution is the development of a novel two-layered DBMM (2L-DBMM) for multimodal fusion. This model effectively integrates speech emotion and text sentiment, enabling the classification of more nuanced, second-level emotional states. Evaluating our framework on the EmoUERJ (Portuguese) and ESD (English) datasets, the extended DBMM achieves accuracy rates of 96% and 98% for SER, 85% and 95% for SA, and 96% and 98% for combined emotion classification using the 2L-DBMM, respectively. Our findings demonstrate the superior performance of the extended DBMM for individual modalities compared to individual classifiers and the 2L-DBMM for merging different modalities, highlighting the value of ensemble methods and multimodal fusion in affective communication analysis. The results underscore the potential of our approach in enhancing emotional understanding with broad applications in fields like mental health assessment, human–robot interaction, and cross-cultural communication. Full article
Show Figures

Figure 1

Figure 1
<p>Overview of the proposed architecture for affective communication merging speech emotion and sentiment analysis.</p>
Full article ">Figure 2
<p>Frameworks for categorizing emotions: (<b>a</b>) Russell’s circumplex model of affect (adapted from [<a href="#B34-applsci-14-06631" class="html-bibr">34</a>]) and (<b>b</b>) Plutchik’s wheel of emotions (adapted from [<a href="#B35-applsci-14-06631" class="html-bibr">35</a>]).</p>
Full article ">Figure 3
<p>Overview of the proposed 2L-DBMM architecture for multimodality: SER and SA.</p>
Full article ">
20 pages, 5277 KiB  
Article
Sentiment Analysis in Portuguese Restaurant Reviews: Application of Transformer Models in Edge Computing
by Alexandre Branco, Daniel Parada, Marcos Silva, Fábio Mendonça, Sheikh Shanawaz Mostafa and Fernando Morgado-Dias
Electronics 2024, 13(3), 589; https://doi.org/10.3390/electronics13030589 - 31 Jan 2024
Cited by 1 | Viewed by 1368
Abstract
This study focuses on improving sentiment analysis in restaurant reviews by leveraging transfer learning and transformer-based pre-trained models. This work evaluates the suitability of pre-trained deep learning models for analyzing Natural Language Processing tasks in Portuguese. It also explores the viability of utilizing [...] Read more.
This study focuses on improving sentiment analysis in restaurant reviews by leveraging transfer learning and transformer-based pre-trained models. This work evaluates the suitability of pre-trained deep learning models for analyzing Natural Language Processing tasks in Portuguese. It also explores the viability of utilizing edge devices for Natural Language Processing tasks, considering their computational limitations and resource constraints. Specifically, we employ bidirectional encoder representations from transformers and robustly optimized BERT approach, two state-of-the-art models, to build a sentiment review classifier. The classifier’s performance is evaluated using accuracy and area under the receiver operating characteristic curve as the primary metrics. Our results demonstrate that the classifier developed using ensemble techniques outperforms the baseline model (from 0.80 to 0.84) in accurately classifying restaurant review sentiments when three classes are considered (negative, neutral, and positive), reaching an accuracy and area under the receiver operating characteristic curve higher than 0.8 when examining a Zomato restaurant review dataset, provided for this work. This study seeks to create a model for the precise classification of Portuguese reviews into positive, negative, or neutral categories. The flexibility of deploying our model on affordable hardware platforms suggests its potential to enable real-time solutions. The deployment of the model on edge computing platforms improves accessibility in resource-constrained environments. Full article
(This article belongs to the Special Issue Applications of Machine Learning in Real World)
Show Figures

Figure 1

Figure 1
<p>Representation of the methodology.</p>
Full article ">Figure 2
<p>Word count distribution.</p>
Full article ">Figure 3
<p>Raw dataset rating distribution.</p>
Full article ">Figure 4
<p>Train dataset weight rating distribution and weight assigned.</p>
Full article ">Figure 5
<p>Representation of a fully developed model.</p>
Full article ">Figure 6
<p>Architecture developed for SA.</p>
Full article ">Figure 7
<p>Architecture of the ensemble applied to each sentiment classifier.</p>
Full article ">Figure 8
<p>Full model architecture developed.</p>
Full article ">Figure 9
<p>Train history of the performance metrics ACC (<b>a</b>), and AUC (<b>b</b>), using SentAnalysisPt.</p>
Full article ">Figure 10
<p>Train history of the loss (<b>a</b>), and the resulting CM (<b>b</b>), using SentAnalysisPt.</p>
Full article ">Figure 11
<p>Comparison of the ACC per class of the three models.</p>
Full article ">Figure 12
<p>Train history of the performance metrics ACC (<b>a</b>), and AUC (<b>b</b>), using SentAnalysisPtRoBERTa.</p>
Full article ">Figure 13
<p>Train history of the loss (<b>a</b>), and the resulting CM (<b>b</b>), using SentAnalysisPtRoBERTa.</p>
Full article ">Figure 14
<p>Performance to complexity ratio in terms of parameters and time of the examined hardware platforms. The dash lines indicate the interpolation of the points.</p>
Full article ">
16 pages, 553 KiB  
Article
Towards Transfer Learning Techniques—BERT, DistilBERT, BERTimbau, and DistilBERTimbau for Automatic Text Classification from Different Languages: A Case Study
by Rafael Silva Barbon and Ademar Takeo Akabane
Sensors 2022, 22(21), 8184; https://doi.org/10.3390/s22218184 - 26 Oct 2022
Cited by 13 | Viewed by 3927
Abstract
The Internet of Things is a paradigm that interconnects several smart devices through the internet to provide ubiquitous services to users. This paradigm and Web 2.0 platforms generate countless amounts of textual data. Thus, a significant challenge in this context is automatically performing [...] Read more.
The Internet of Things is a paradigm that interconnects several smart devices through the internet to provide ubiquitous services to users. This paradigm and Web 2.0 platforms generate countless amounts of textual data. Thus, a significant challenge in this context is automatically performing text classification. State-of-the-art outcomes have recently been obtained by employing language models trained from scratch on corpora made up from news online to handle text classification better. A language model that we can highlight is BERT (Bidirectional Encoder Representations from Transformers) and also DistilBERT is a pre-trained smaller general-purpose language representation model. In this context, through a case study, we propose performing the text classification task with two previously mentioned models for two languages (English and Brazilian Portuguese) in different datasets. The results show that DistilBERT’s training time for English and Brazilian Portuguese was about 45% faster than its larger counterpart, it was also 40% smaller, and preserves about 96% of language comprehension skills for balanced datasets. Full article
(This article belongs to the Special Issue Sensors Data Processing Using Machine Learning)
Show Figures

Figure 1

Figure 1
<p>Transformer architecture [<a href="#B19-sensors-22-08184" class="html-bibr">19</a>].</p>
Full article ">Figure 2
<p>The BERT architecture [<a href="#B20-sensors-22-08184" class="html-bibr">20</a>] in a pre-training context (<b>left</b>) or fine-tuning for different tasks (<b>right</b>).</p>
Full article ">Figure 3
<p>Overfitting example.</p>
Full article ">Figure 4
<p>A 10-fold cross-validation example.</p>
Full article ">Figure 5
<p>Bar plot comparing BERT and DistilBERT’s model scores.</p>
Full article ">
23 pages, 3212 KiB  
Article
Survey of Text Mining Techniques Applied to Judicial Decisions Prediction
by Olga Alejandra Alcántara Francia, Miguel Nunez-del-Prado and Hugo Alatrista-Salas
Appl. Sci. 2022, 12(20), 10200; https://doi.org/10.3390/app122010200 - 11 Oct 2022
Cited by 10 | Viewed by 3737
Abstract
This paper reviews the most recent literature on experiments with different Machine Learning, Deep Learning and Natural Language Processing techniques applied to predict judicial and administrative decisions. Among the most outstanding findings, we have that the most used data mining techniques are Support [...] Read more.
This paper reviews the most recent literature on experiments with different Machine Learning, Deep Learning and Natural Language Processing techniques applied to predict judicial and administrative decisions. Among the most outstanding findings, we have that the most used data mining techniques are Support Vector Machine (SVM), K Nearest Neighbours (K-NN) and Random Forest (RF), and in terms of the most used deep learning techniques, we found Long-Term Memory (LSTM) and transformers such as BERT. An important finding in the papers reviewed was that the use of machine learning techniques has prevailed over those of deep learning. Regarding the place of origin of the research carried out, we found that 64% of the works belong to studies carried out in English-speaking countries, 8% in Portuguese and 28% in other languages (such as German, Chinese, Turkish, Spanish, etc.). Very few works of this type have been carried out in Spanish-speaking countries. The classification criteria of the works have been based, on the one hand, on the identification of the classifiers used to predict situations (or events with legal interference) or judicial decisions and, on the other hand, on the application of classifiers to the phenomena regulated by the different branches of law: criminal, constitutional, human rights, administrative, intellectual property, family law, tax law and others. The corpus size analyzed in the reviewed works reached 100,000 documents in 2020. Finally, another important finding lies in the accuracy of these predictive techniques, reaching predictions of over 60% in different branches of law. Full article
(This article belongs to the Special Issue Natural Language Processing (NLP) and Applications)
Show Figures

Figure 1

Figure 1
<p>Survey classification by types of law.</p>
Full article ">Figure 2
<p>Corpus size evolution.</p>
Full article ">Figure 3
<p>Languages distribution of analyzed works.</p>
Full article ">Figure 4
<p>Word cloud analysis.</p>
Full article ">Figure 5
<p>Historical performance of judicial decision prediction.</p>
Full article ">Figure 6
<p>Percentage of models used for ruling decision prediction.</p>
Full article ">Figure 7
<p>Scores using different metrics to assess the ruling decision prediction.</p>
Full article ">
17 pages, 2339 KiB  
Article
An Empirical Comparison of Portuguese and Multilingual BERT Models for Auto-Classification of NCM Codes in International Trade
by Roberta Rodrigues de Lima, Anita M. R. Fernandes, James Roberto Bombasar, Bruno Alves da Silva, Paul Crocker and Valderi Reis Quietinho Leithardt
Big Data Cogn. Comput. 2022, 6(1), 8; https://doi.org/10.3390/bdcc6010008 - 10 Jan 2022
Cited by 5 | Viewed by 4330
Abstract
Classification problems are common activities in many different domains and supervised learning algorithms have shown great promise in these areas. The classification of goods in international trade in Brazil represents a real challenge due to the complexity involved in assigning the correct category [...] Read more.
Classification problems are common activities in many different domains and supervised learning algorithms have shown great promise in these areas. The classification of goods in international trade in Brazil represents a real challenge due to the complexity involved in assigning the correct category codes to a good, especially considering the tax penalties and legal implications of a misclassification. This work focuses on the training process of a classifier based on bidirectional encoder representations from transformers (BERT) for tax classification of goods with MCN codes which are the official classification system for import and export products in Brazil. In particular, this article presents results from using a specific Portuguese-language-pretrained BERT model, as well as results from using a multilingual-pretrained BERT model. Experimental results show that Portuguese model had a slightly better performance than the multilingual model, achieving an MCC 0.8491, and confirms that the classifiers could be used to improve specialists’ performance in the classification of goods. Full article
Show Figures

Figure 1

Figure 1
<p>Composition of the NCM Code.</p>
Full article ">Figure 2
<p>Embedding composition.</p>
Full article ">Figure 3
<p>Fine-tuning process.</p>
Full article ">Figure 4
<p>Procedures.</p>
Full article ">Figure 5
<p>Chapter 90 import data sample [<a href="#B14-BDCC-06-00008" class="html-bibr">14</a>].</p>
Full article ">Figure 6
<p>Four-fold cross-validation.</p>
Full article ">Figure 7
<p>Multilingual BERT hyperparameter tuning.</p>
Full article ">Figure 8
<p>Portuguese BERT hyperparameter tuning.</p>
Full article ">Figure 9
<p>LIME applied to BERT Portuguese classifier.</p>
Full article ">
13 pages, 333 KiB  
Article
Compositional Distributional Semantics with Syntactic Dependencies and Selectional Preferences
by Pablo Gamallo
Appl. Sci. 2021, 11(12), 5743; https://doi.org/10.3390/app11125743 - 21 Jun 2021
Cited by 5 | Viewed by 2577
Abstract
This article describes a compositional model based on syntactic dependencies which has been designed to build contextualized word vectors, by following linguistic principles related to the concept of selectional preferences. The compositional strategy proposed in the current work has been evaluated on a [...] Read more.
This article describes a compositional model based on syntactic dependencies which has been designed to build contextualized word vectors, by following linguistic principles related to the concept of selectional preferences. The compositional strategy proposed in the current work has been evaluated on a syntactically controlled and multilingual dataset, and compared with Transformer BERT-like models, such as Sentence BERT, the state-of-the-art in sentence similarity. For this purpose, we created two new test datasets for Portuguese and Spanish on the basis of that defined for the English language, containing expressions with noun-verb-noun transitive constructions. The results we have obtained show that the linguistic-based compositional approach turns out to be competitive with Transformer models. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

Figure 1
<p>Dependency-based analysis of <span class="html-italic">the company fired the employee</span> and left-to-right interpretation process to build the contextualized word senses of the three lexical constituents.</p>
Full article ">Figure 2
<p>Dependency-based analysis of <span class="html-italic">the company fired the employee</span> and right-to-left interpretation process to build the contextualized word senses of the three lexical constituents.</p>
Full article ">Figure 3
<p>Bar plot with the Spearman scores of the best configurations for each model and language.</p>
Full article ">
24 pages, 912 KiB  
Article
Monolingual and Cross-Lingual Intent Detection without Training Data in Target Languages
by Jurgita Kapočiūtė-Dzikienė, Askars Salimbajevs and Raivis Skadiņš
Electronics 2021, 10(12), 1412; https://doi.org/10.3390/electronics10121412 - 11 Jun 2021
Cited by 7 | Viewed by 2884
Abstract
Due to recent DNN advancements, many NLP problems can be effectively solved using transformer-based models and supervised data. Unfortunately, such data is not available in some languages. This research is based on assumptions that (1) training data can be obtained by the machine [...] Read more.
Due to recent DNN advancements, many NLP problems can be effectively solved using transformer-based models and supervised data. Unfortunately, such data is not available in some languages. This research is based on assumptions that (1) training data can be obtained by the machine translating it from another language; (2) there are cross-lingual solutions that work without the training data in the target language. Consequently, in this research, we use the English dataset and solve the intent detection problem for five target languages (German, French, Lithuanian, Latvian, and Portuguese). When seeking the most accurate solutions, we investigate BERT-based word and sentence transformers together with eager learning classifiers (CNN, BERT fine-tuning, FFNN) and lazy learning approach (Cosine similarity as the memory-based method). We offer and evaluate several strategies to overcome the data scarcity problem with machine translation, cross-lingual models, and a combination of the previous two. The experimental investigation revealed the robustness of sentence transformers under various cross-lingual conditions. The accuracy equal to ~0.842 is achieved with the English dataset with completely monolingual models is considered our top-line. However, cross-lingual approaches demonstrate similar accuracy levels reaching ~0.831, ~0.829, ~0.853, ~0.831, and ~0.813 on German, French, Lithuanian, Latvian, and Portuguese languages. Full article
(This article belongs to the Special Issue Hybrid Methods for Natural Language Processing)
Show Figures

Figure 1

Figure 1
<p>The best accuracies + confidence intervals with BERT-w + CNN, BERT-w + BERT, BERT-s + FFNN, and BERT-s + COS approaches under the <span class="html-italic">MT-based</span> strategy. Dashed lines connect the best-achieved accuracy (within the same language) with those accuracies to which differences are not statistically significant. EN results are obtained on original data and represent the top-line.</p>
Full article ">Figure 2
<p>The best accuracies + confidence intervals with BERT-w + CNN, BERT-w + BERT, BERT-s + FFNN, and BERT-s + COS approaches under the <span class="html-italic">cross-lingual</span> strategy. For the notation, see <a href="#electronics-10-01412-f001" class="html-fig">Figure 1</a>.</p>
Full article ">Figure 3
<p>The best accuracies + confidence intervals of BERT-s + FFNN and BERT-s + COS models trained under the <span class="html-italic">combined</span> strategy. For the notation, see <a href="#electronics-10-01412-f001" class="html-fig">Figure 1</a>.</p>
Full article ">Figure 4
<p>The best accuracies + confidence intervals with BERT-s + FFNN and BERT-s + COS approaches under the <span class="html-italic">train all</span> strategy. For the notation see <a href="#electronics-10-01412-f001" class="html-fig">Figure 1</a>.</p>
Full article ">Figure 5
<p>The best accuracies + confidence intervals for different languages under different conditions. For the notation, see <a href="#electronics-10-01412-f001" class="html-fig">Figure 1</a>.</p>
Full article ">
19 pages, 383 KiB  
Article
Benchmarking Natural Language Inference and Semantic Textual Similarity for Portuguese
by Pedro Fialho, Luísa Coheur and Paulo Quaresma
Information 2020, 11(10), 484; https://doi.org/10.3390/info11100484 - 15 Oct 2020
Cited by 4 | Viewed by 2739
Abstract
Two sentences can be related in many different ways. Distinct tasks in natural language processing aim to identify different semantic relations between sentences. We developed several models for natural language inference and semantic textual similarity for the Portuguese language. We took advantage of [...] Read more.
Two sentences can be related in many different ways. Distinct tasks in natural language processing aim to identify different semantic relations between sentences. We developed several models for natural language inference and semantic textual similarity for the Portuguese language. We took advantage of pre-trained models (BERT); additionally, we studied the roles of lexical features. We tested our models in several datasets—ASSIN, SICK-BR and ASSIN2—and the best results were usually achieved with ptBERT-Large, trained in a Brazilian corpus and tuned in the latter datasets. Besides obtaining state-of-the-art results, this is, to the best of our knowledge, the most all-inclusive study about natural language inference and semantic textual similarity for the Portuguese language. Full article
(This article belongs to the Special Issue Selected Papers from PROPOR 2020)
Show Figures

Figure 1

Figure 1
<p>Top 100 examples of ASSIN2 with greater distance between predicted and true values of the STS task, where such distance is greater than 0.5.</p>
Full article ">Figure 2
<p>Top 100 examples of ASSIN-PTBR with greater distance between the prediction and true values of the STS task, where such distance is greater than 0.5</p>
Full article ">
Back to TopTop