Deep learning model for opinion mining
LSTM (Long Short-Term Memory) shows its performance in Sentiment Analysis, but it has a critical drawback in terms of how to do backpropagation, limiting the training time to more extended and the process slower. Attention mechanism more behavior like human understands the sentences by a focus on sp...
Saved in:
Main Author: | |
---|---|
Format: | Final Year Project / Dissertation / Thesis |
Published: |
2022
|
Subjects: | |
Online Access: | http://eprints.utar.edu.my/5098/1/fyp_CS_2022_LHJ.pdf http://eprints.utar.edu.my/5098/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
id |
my-utar-eprints.5098 |
---|---|
record_format |
eprints |
spelling |
my-utar-eprints.50982023-04-19T08:52:38Z Deep learning model for opinion mining Lee, Hao Jie T Technology (General) LSTM (Long Short-Term Memory) shows its performance in Sentiment Analysis, but it has a critical drawback in terms of how to do backpropagation, limiting the training time to more extended and the process slower. Attention mechanism more behavior like human understands the sentences by a focus on specific words to solve the issue from LSTM. The Bert (Bidirectional Encoder Representations from Transformers) use an attention mechanism and outperform other attention-based model such as GPT (Generative Pre-trained Transformer) and Elmo (Embeddings from Language Model) because it has learned the deep bidirectional presentations by the MLM (Masked Language Model) and NSP (Next Sentence Prediction). Ernie (Enhanced Language Representation with Informative Entities) model and Zen model modify how Bert model learns language and gains achievement in Chinese NLP (Natural Language Process). RoBERTa (A Robustly Optimized BERT Pretraining Approach) from Facebook proves the NSP is not helping the model, so we also modify the NSP task others for to learn the language. Sentiment Analysis is one of the NLP tasks Ernie and Zen successful in beating the Bert-Chinese, which uses Chinese characters as input and WordPiece embeddings to do word embeddings. Word level embeddings and input is needed to improve the Bert model works on Chinese Sentiment Analysis. With the motivation to improve the Chinese Sentiment Analysis, this project will combine experience from different models to propose a better version of the Bert model. This project will limit the scope to improve Sentiment Analysis among different NLP tasks. 2022-11-29 Final Year Project / Dissertation / Thesis NonPeerReviewed application/pdf http://eprints.utar.edu.my/5098/1/fyp_CS_2022_LHJ.pdf Lee, Hao Jie (2022) Deep learning model for opinion mining. Final Year Project, UTAR. http://eprints.utar.edu.my/5098/ |
institution |
Universiti Tunku Abdul Rahman |
building |
UTAR Library |
collection |
Institutional Repository |
continent |
Asia |
country |
Malaysia |
content_provider |
Universiti Tunku Abdul Rahman |
content_source |
UTAR Institutional Repository |
url_provider |
http://eprints.utar.edu.my |
topic |
T Technology (General) |
spellingShingle |
T Technology (General) Lee, Hao Jie Deep learning model for opinion mining |
description |
LSTM (Long Short-Term Memory) shows its performance in Sentiment Analysis, but it has a critical drawback in terms of how to do backpropagation, limiting the training time to more extended and the process slower. Attention mechanism more behavior like human understands the sentences by a focus on specific words to solve the issue from LSTM. The Bert (Bidirectional Encoder Representations from Transformers) use an attention mechanism and outperform other attention-based model such as GPT (Generative Pre-trained Transformer) and Elmo (Embeddings from Language Model) because it has learned the deep bidirectional presentations by the MLM (Masked Language Model) and NSP (Next Sentence Prediction).
Ernie (Enhanced Language Representation with Informative Entities) model and Zen model modify how Bert model learns language and gains achievement in Chinese NLP (Natural Language Process). RoBERTa (A Robustly Optimized BERT Pretraining Approach) from Facebook proves the NSP is not helping the model, so we also modify the NSP task others for to learn the language. Sentiment Analysis is one of the NLP tasks Ernie and Zen successful in beating the Bert-Chinese, which uses Chinese characters as input and WordPiece embeddings to do word embeddings. Word level embeddings and input is needed to improve the Bert model works on Chinese Sentiment Analysis.
With the motivation to improve the Chinese Sentiment Analysis, this project will combine experience from different models to propose a better version of the Bert model. This project will limit the scope to improve Sentiment Analysis among different NLP tasks. |
format |
Final Year Project / Dissertation / Thesis |
author |
Lee, Hao Jie |
author_facet |
Lee, Hao Jie |
author_sort |
Lee, Hao Jie |
title |
Deep learning model for opinion mining |
title_short |
Deep learning model for opinion mining |
title_full |
Deep learning model for opinion mining |
title_fullStr |
Deep learning model for opinion mining |
title_full_unstemmed |
Deep learning model for opinion mining |
title_sort |
deep learning model for opinion mining |
publishDate |
2022 |
url |
http://eprints.utar.edu.my/5098/1/fyp_CS_2022_LHJ.pdf http://eprints.utar.edu.my/5098/ |
_version_ |
1765301613417725952 |
score |
13.160551 |