完整後設資料紀錄
DC 欄位語言
dc.contributor.authorHuang, CDen_US
dc.contributor.authorLiang, SFen_US
dc.contributor.authorLin, CTen_US
dc.contributor.authorWu, RCen_US
dc.date.accessioned2014-12-08T15:18:50Z-
dc.date.available2014-12-08T15:18:50Z-
dc.date.issued2005-07-01en_US
dc.identifier.issn1016-2364en_US
dc.identifier.urihttp://hdl.handle.net/11536/13539-
dc.description.abstractIn machine learning, both the properly used networks and the selected features are important factors which should be considered carefully. These two factors will influence the result, whether for better or worse. In bioinformatics, the amount of features may be very large to make machine learning possible. In this study we introduce the idea of feature selection in the problem of bioinformatics. We use neural networks to complete our task where each input node is associated with a gate. At the beginning of the training, all gates are almost closed, and, at this time, no features are allowed to enter the network. During the training phase, gates are either opened or closed, depending oil the requirements. After the selection training phase has completed, gates corresponding to the helpful features are completely opened while gates Corresponding to the useless features are closed more tightly. Some gates may be partially open, depending oil the importance of the corresponding features. So, the network can not only select features in an online manner during learning, but it also does some feature extraction. We combine feature selection with our novel hierarchical machine learning architecture and apply it to multi-class protein fold classification. At the first level the network classifies the data into four major folds: all alpha, all beta, alpha + beta and alpha/beta. In the next level, we have another set of networks which further classifies the data into twenty-seven folds. This approach helps achieve the following. The gating network is found to reduce the number of features drastically. It is interesting to observe that, for the first level using just 50 features selected by the gating network, we can get a test accuracy comparable to that using 125 features in neural classifiers. The process also helps us get a better insight into the folding process. For example, tracking the evolution of different gates, we call find which characteristics (features) of the data are more important for the folding process. Eventually, it reduces the computation time. The use of the hierarchical architecture helps LIS get a better performance also.en_US
dc.language.isoen_USen_US
dc.subjectmachine learningen_US
dc.subjecthierarchical architectureen_US
dc.subjectfeature selectionen_US
dc.subjectgateen_US
dc.subjectneural networken_US
dc.subjectprotein folden_US
dc.subjectbioinformaticsen_US
dc.titleMachine learning with automatic feature selection for multi-class protein fold classificationen_US
dc.typeArticleen_US
dc.identifier.journalJOURNAL OF INFORMATION SCIENCE AND ENGINEERINGen_US
dc.citation.volume21en_US
dc.citation.issue4en_US
dc.citation.spage711en_US
dc.citation.epage720en_US
dc.contributor.department電控工程研究所zh_TW
dc.contributor.departmentInstitute of Electrical and Control Engineeringen_US
dc.identifier.wosnumberWOS:000230500600003-
dc.citation.woscount1-
顯示於類別:期刊論文


文件中的檔案:

  1. 000230500600003.pdf

若為 zip 檔案,請下載檔案解壓縮後,用瀏覽器開啟資料夾中的 index.html 瀏覽全文。