Transformer in mRNA degradation prediction.

The unstable properties and the advantages of the mRNA vaccine have encouraged many experts worldwide in tackling the degradation problem. Machine learning models have been highly implemented in bioinformatics and the healthcare fieldstone insights from biological data. Thus, machine learning plays...

Full description

Saved in:
Bibliographic Details
Main Authors: Yit, Tan Wen, Hassan, Rohayanti, Zakaria, Noor Hidayah, Kasim, Shahreen, Moi, Sim Hiew, Khairuddin, Alif Ridzuan, Amnur, Hidra
Format: Article
Language:English
Published: Politeknik Negeri Padang 2023
Subjects:
Online Access:http://eprints.utm.my/105942/1/TanWenYit2023_TransformerinMRNADegradationPrediction.pdf
http://eprints.utm.my/105942/
http://dx.doi.org/10.30630/joiv.7.2.1165
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:The unstable properties and the advantages of the mRNA vaccine have encouraged many experts worldwide in tackling the degradation problem. Machine learning models have been highly implemented in bioinformatics and the healthcare fieldstone insights from biological data. Thus, machine learning plays an important role in predicting the degradation rate of mRNA vaccine candidates. Stanford University has held an OpenVaccine Challenge competition on Kaggle to gather top solutions in solving the mentioned problems, and a multi-column root means square error (MCRMSE) has been used as a main performance metric. The Nucleic Transformer has been proposed by different researchers as a deep learning solution that is able to utilize a self-attention mechanism and Convolutional Neural Network (CNN). Hence, this paper would like to enhance the existing Nucleic Transformer performance by utilizing the AdaBelief or RangerAdaBelief optimizer with a proposed decoder that consists of a normalization layer between two linear layers. Based on the experimental result, the performance of the enhanced Nucleic Transformer outperforms the existing solution. In this study, the AdaBelief optimizer performs better than the RangerAdaBelief optimizer, even though it possesses Ranger’s advantages. The advantages of the proposed decoder can only be shown when there is limited data. When the data is sufficient, the performance might be similar but still better than the linear decoder if and only if the AdaBelief optimizer is used. As a result, the combination of the AdaBelief optimizer with the proposed decoder performs the best with 2.79% and 1.38% performance boost in public and private MCRMSE, respectively.