Please use this identifier to cite or link to this item:
https://scholarbank.nus.edu.sg/handle/10635/230686
Title: | NEW METHODS FOR OMICS AND PHARMACEUTICAL DEEP LEARNING | Authors: | SHEN WANXIANG | ORCID iD: | ![]() |
Keywords: | Unsupervised learning, Machine learning, Deep learning, Convolutional neural network, Omics, Molecular property, Drug discovery | Issue Date: | 11-Apr-2022 | Citation: | SHEN WANXIANG (2022-04-11). NEW METHODS FOR OMICS AND PHARMACEUTICAL DEEP LEARNING. ScholarBank@NUS Repository. | Abstract: | Artificial intelligence (AI) has emerged as a solution recently and is transforming healthcare and therapeutics. However, there are still challenges for deep learning (DL)-based AI in non-image data learning of biomedical area. Biomedical investigations such as omics-based disease diagnosis and pharmaceutical drug discovery frequently rely on non-image data of high-dimensions, arbitrary-order but intercorrelated features, relatively low-sample sizes, diverse samples, and sparse features (Bio-HARDS), which pose challenges in exploring highly efficient DL algorithms for these tasks. There is thus a pressing need to develop more robust, high-performing, but explainable DL approaches for Bio-HARDS tasks. This thesis describes a novel methodology for omics and pharmaceutical DL using unsupervised and supervised learning to learn the omics and pharmaceutical data. The unsupervised data representation methods of AggMap/MolMap combined with supervised CNN-based AggMapNet/MolMapNet architectures establish the new paradigms for enhanced learning and interpretability of omics and pharmaceutical molecular data. | URI: | https://scholarbank.nus.edu.sg/handle/10635/230686 |
Appears in Collections: | Ph.D Theses (Open) |
Show full item record
Files in This Item:
File | Description | Size | Format | Access Settings | Version | |
---|---|---|---|---|---|---|
PhD_Thesis_SWX_v12_final_submit.pdf | 16.78 MB | Adobe PDF | OPEN | None | View/Download |
Google ScholarTM
Check
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.