Please use this identifier to cite or link to this item: https://scholarbank.nus.edu.sg/handle/10635/230686
Title: NEW METHODS FOR OMICS AND PHARMACEUTICAL DEEP LEARNING
Authors: SHEN WANXIANG
ORCID iD:   orcid.org/0000-0001-7114-3664
Keywords: Unsupervised learning, Machine learning, Deep learning, Convolutional neural network, Omics, Molecular property, Drug discovery
Issue Date: 11-Apr-2022
Citation: SHEN WANXIANG (2022-04-11). NEW METHODS FOR OMICS AND PHARMACEUTICAL DEEP LEARNING. ScholarBank@NUS Repository.
Abstract: Artificial intelligence (AI) has emerged as a solution recently and is transforming healthcare and therapeutics. However, there are still challenges for deep learning (DL)-based AI in non-image data learning of biomedical area. Biomedical investigations such as omics-based disease diagnosis and pharmaceutical drug discovery frequently rely on non-image data of high-dimensions, arbitrary-order but intercorrelated features, relatively low-sample sizes, diverse samples, and sparse features (Bio-HARDS), which pose challenges in exploring highly efficient DL algorithms for these tasks. There is thus a pressing need to develop more robust, high-performing, but explainable DL approaches for Bio-HARDS tasks. This thesis describes a novel methodology for omics and pharmaceutical DL using unsupervised and supervised learning to learn the omics and pharmaceutical data. The unsupervised data representation methods of AggMap/MolMap combined with supervised CNN-based AggMapNet/MolMapNet architectures establish the new paradigms for enhanced learning and interpretability of omics and pharmaceutical molecular data.
URI: https://scholarbank.nus.edu.sg/handle/10635/230686
Appears in Collections:Ph.D Theses (Restricted)

Show full item record
Files in This Item:
File Description SizeFormatAccess SettingsVersion 
PhD_Thesis_SWX_v12_final_submit.pdf16.78 MBAdobe PDF

RESTRICTED

NoneLog In

Page view(s)

20
checked on Dec 1, 2022

Download(s)

1
checked on Dec 1, 2022

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.