Inference Algorithms in Latent Dirichlet Allocation for Semantic Classification

There are existing implementations of Latent Dirichlet Allocation (LDA) algorithm as a semantic classifier to arrange the data for efficient retrieval. However, the problem of learning or inferencing the posterior distribution of the algorithm is trivial. Inferencing directly the prior distribution...

Full description

Saved in:
Bibliographic Details
Main Authors: Mohammad Zubir, W.M.A., Abdul Aziz, I., Jaafar, J., Hasan, M.H.
Format: Article
Published: Springer Verlag 2018
Online Access:https://www.scopus.com/inward/record.uri?eid=2-s2.0-85029592113&doi=10.1007%2f978-3-319-67621-0_16&partnerID=40&md5=fcaad5d0774e2882d0438b21c3ada74a
http://eprints.utp.edu.my/21237/
Tags: Add Tag
No Tags, Be the first to tag this record!
id my.utp.eprints.21237
record_format eprints
spelling my.utp.eprints.212372019-02-26T03:19:49Z Inference Algorithms in Latent Dirichlet Allocation for Semantic Classification Mohammad Zubir, W.M.A. Abdul Aziz, I. Jaafar, J. Hasan, M.H. There are existing implementations of Latent Dirichlet Allocation (LDA) algorithm as a semantic classifier to arrange the data for efficient retrieval. However, the problem of learning or inferencing the posterior distribution of the algorithm is trivial. Inferencing directly the prior distribution could lead to time taken to increase exponentially. It is due to the coupling of the hyperparameters. Several inference algorithms have been implemented together with LDA to solve this issue. The inference algorithm used in this research work is Gibbs sampling. Research using Gibbs sampling shows promising results in comparison to other inference algorithms, especially in the performance of the algorithm. It still takes a long time to compute the topic distribution of the data. There are still room for improvement in the time taken for the algorithm to complete the topic distribution. Using two datasets, an evaluation of the performance of the algorithm has been conducted. Results show that Gibbs sampling as the inference algorithm provides a better prediction on the optimal number of topic of the data in comparison to Variational Expectation Maximization (VEM). © 2018, Springer International Publishing AG. Springer Verlag 2018 Article NonPeerReviewed https://www.scopus.com/inward/record.uri?eid=2-s2.0-85029592113&doi=10.1007%2f978-3-319-67621-0_16&partnerID=40&md5=fcaad5d0774e2882d0438b21c3ada74a Mohammad Zubir, W.M.A. and Abdul Aziz, I. and Jaafar, J. and Hasan, M.H. (2018) Inference Algorithms in Latent Dirichlet Allocation for Semantic Classification. Advances in Intelligent Systems and Computing, 662 . pp. 173-184. http://eprints.utp.edu.my/21237/
institution Universiti Teknologi Petronas
building UTP Resource Centre
collection Institutional Repository
continent Asia
country Malaysia
content_provider Universiti Teknologi Petronas
content_source UTP Institutional Repository
url_provider http://eprints.utp.edu.my/
description There are existing implementations of Latent Dirichlet Allocation (LDA) algorithm as a semantic classifier to arrange the data for efficient retrieval. However, the problem of learning or inferencing the posterior distribution of the algorithm is trivial. Inferencing directly the prior distribution could lead to time taken to increase exponentially. It is due to the coupling of the hyperparameters. Several inference algorithms have been implemented together with LDA to solve this issue. The inference algorithm used in this research work is Gibbs sampling. Research using Gibbs sampling shows promising results in comparison to other inference algorithms, especially in the performance of the algorithm. It still takes a long time to compute the topic distribution of the data. There are still room for improvement in the time taken for the algorithm to complete the topic distribution. Using two datasets, an evaluation of the performance of the algorithm has been conducted. Results show that Gibbs sampling as the inference algorithm provides a better prediction on the optimal number of topic of the data in comparison to Variational Expectation Maximization (VEM). © 2018, Springer International Publishing AG.
format Article
author Mohammad Zubir, W.M.A.
Abdul Aziz, I.
Jaafar, J.
Hasan, M.H.
spellingShingle Mohammad Zubir, W.M.A.
Abdul Aziz, I.
Jaafar, J.
Hasan, M.H.
Inference Algorithms in Latent Dirichlet Allocation for Semantic Classification
author_facet Mohammad Zubir, W.M.A.
Abdul Aziz, I.
Jaafar, J.
Hasan, M.H.
author_sort Mohammad Zubir, W.M.A.
title Inference Algorithms in Latent Dirichlet Allocation for Semantic Classification
title_short Inference Algorithms in Latent Dirichlet Allocation for Semantic Classification
title_full Inference Algorithms in Latent Dirichlet Allocation for Semantic Classification
title_fullStr Inference Algorithms in Latent Dirichlet Allocation for Semantic Classification
title_full_unstemmed Inference Algorithms in Latent Dirichlet Allocation for Semantic Classification
title_sort inference algorithms in latent dirichlet allocation for semantic classification
publisher Springer Verlag
publishDate 2018
url https://www.scopus.com/inward/record.uri?eid=2-s2.0-85029592113&doi=10.1007%2f978-3-319-67621-0_16&partnerID=40&md5=fcaad5d0774e2882d0438b21c3ada74a
http://eprints.utp.edu.my/21237/
_version_ 1738656261149294592
score 13.159267