簡易檢索 / 詳目顯示

研究生: 洪書帆
Hung, Shu-Fan
論文名稱: 以潛在樣本提升小樣本學習之正確性
Using potential Samples to improve learning accuracies of Small Sample sets
指導教授: 利德江
Li, Der-Chiang
學位類別: 碩士
Master
系所名稱: 管理學院 - 工業與資訊管理學系
Department of Industrial and Information Management
論文出版年: 2009
畢業學年度: 98
語文別: 中文
論文頁數: 73
中文關鍵詞: 小樣本學習類別屬性數值預測
外文關鍵詞: small data learning, virtual sample
相關次數: 點閱:107下載:3
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • 在真實生活中,欲進行資料分析時,常囿於資料取得困難、時間、或成本的考量,無法取得有效數量之樣本,因此如何從所獲得之少量資料中,發掘有意義的資訊以建立穩定的知識模式,在近年來已是重要的議題。目前小樣本學習方法,主要可從資料的轉換、虛擬資料量之擴充、以及新屬性的開發等方面著手。在虛擬資料量擴充部份,本研究提出一個兩階段的虛擬樣本產生方法,第一階段使用模式樹中的M5'演算法對小樣本進行學習,依其分類法則產生各葉部節點之可能值;第二階段則依此可能值於各葉部節點產生應存在於小樣本中之樣本,稱之潛在樣本。實驗結果顯示,透過此方法產生之潛在樣本,在加入原小樣本後所建構之預測模式,確實較原小樣本對未知母體的預測誤差有明顯的改善,並且於方法比較上,本研究比整體趨勢擴散方法有更佳之效果。此外,本研究之主要貢獻在於小樣本之名目屬性資料之學習,冀能成為後續相關研究之參考。

    In real world, while we try to learn information from data with machine learning algorithms, sometimes the size of acquirable data is quite small so that the learning results may not be reliable since the learning tools have their smallest required sample number respectively to ensure their robustness. The reasons of the small data learning problems occuring can be mainly owing to the consideration of time and cost. In this research, a two-phase procedure is proposed to solve such learning problems by generating more training samples for learning tools to improve their predictions. In the first phase, the M5’ model is employed to acqire classification rules which contain the corrleations between attributes and the possible values of each leaf-node can be thus obtained. In the second phase, samples store at a leaf-node can be permuted with correspondent possible values to generate the training samples which are very likely to happen are called potential samples. The experimental results reveal that the prediction accuracies can be improved while learning with potential samples, especially when the original sample size is small. Besides, compared with a virtual sample generation algorithm, Mega-Trend-Diffusion, the proposed procedure significantly outperforms.

    摘要 2 Abstract II 致謝 III 目錄 IV 圖目錄 VII 表目錄 IX _Toc263614454第一章 緒論 1 1.1 研究背景與動機 1 1.2 研究目的 3 1.3 研究架構與流程 4 第二章 文獻探討 6 2.1 虛擬樣本(virtual sample) 6 2.1.1 資訊擴展(information diffusion) 6 2.1.2 其他虛擬樣本產生方法 8 2.2 模式樹 9 2.3 整體趨勢擴展技術 11 2.4 預測模型 15 2.4.1 倒傳遞類神經網路(back-propagation network, BPN) 15 2.4.2 支撐向量迴歸(support vector regression, SVR) 16 2.4.3 多元線性迴歸(multiple linear regression, MLR) 17 2.4.4 決策樹(decision tree) 17 2.5 小結 17 第三章 研究方法 19 3.1 屬性轉換 20 3.1.1 名目屬性轉換 20 3.1.2 數值屬性離散化 21 3.2 模式樹M5'演算法學習 22 3.3 排列組合 24 3.3.1 輸出屬性(類別)值 24 3.3.2 輸入屬性值 24 3.3.3 排列法則 27 3.4 建構預測模型 28 3.4.1 M5'模式樹演算法 28 3.4.2 倒傳遞類神經網路 28 3.4.3 支撐向量迴歸 30 3.4.4 多元線性迴歸 32 3.4.5 決策樹 33 3.5 實驗方式 34 3.5.1 評估方式 35 3.5.2 預測誤差評估指標 35 3.5.3 假設檢定 36 第四章 實例驗證 37 4.1 軟體選擇 37 4.1.1 參數設定 37 4.2 資料蒐集與說明 38 4.2.1 UCI資料說明 38 4.2.2 文獻資料說明 39 4.3 樣本數敏感度分析與評估 39 4.3.1 方法驗證 39 4.3.2 方法比較 59 4.4 進階研究與探討 61 4.4.1 潛在和虛擬樣本之差異探討 61 4.4.2 潛在樣本之深入探討 63 第五章 結論與建議 66 參考文獻 67 附錄 71

    葉怡成,“類神經網路模式應用與實作”,儒林圖書公司,民國92年3月。
    Anthony, M., Biggs, N. (1997). Computational Learning Theory. Cambridge University Press.
    Breiman, L., Friedman, J.H., Olshen, R.A., and Stone, C.J. (1984). Classification and Regression Trees, Belmont, CA:Wadsworth International Group.
    Chaudhuri, P., Huang, M. C., Loh, W. Y.,& Rubin, R. (1994). Piecewise-polynomial regression tree, Statistica Sinica, Vol.4, 143-167.
    Dobra, A. and Gehrke. J.E., (2002). SECRET: A Scalable Linear Regression Tree Algorithm, Proceedings of the Eighth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 481-487.
    Drucker, H., Burges, C. J. C., Kaufman, L., Smola, A., and Vapnik, V. (1997). Support vector regression machines, in: M. Mozer, M. Jordan, and T. Petsche (Eds.), Advances in Neural Information Processing Systems, vol 9. Cambridge, MA: MIT Press, 1997, pp. 155-161.
    Frank E., Wang Y., Inglis S., Holmes G., and Witten IH (1998). Technical note: Using model trees for classification. Machine Learning, 32, 63-76.
    Huang, C. F., Moraga, C. (2004). A diffusion-neural-network for learning from small samples. International Journal of Approximate Reasoning, 35, 137-161.
    Huang, C. J., Wang, H. F. (2009). Prediction of the Period of Psychotic Episode in Individual Schizophrenics by Simulation-Data Construction Approach.
    Huang, C.F. (1997). Principle of information. Fuzzy Sets and Systems. 91,69-90.
    Jang, J.S.R. (1993). ANFIS: adaptive-network-based fuzzy inference systems. IEEE Transactions on Systems, Man and Cybernetics. 23, 665-685.
    Jennrich, R. I, Schluchter, M. D. (1986). Unbalanced repeated-measures models with structured covariance matrices. Biometrics, 42, 805-820.
    Karalic, A. (1992). Employing linear regression in regression tree leaves, Proceeding of the 10th European Conference on Artificial Intelligence, 440-441.
    Kerber, R. (1992). ChiMerge: discretization of numeric attributes, The Ninth ACM SIGKDDInternational Conference on Knowledge Discovery and Data Mining, Paris, France,123-128.
    Laird, N. M., Ware, J. H. (1982). Random-effects models for longitudinal data. Biometrics, 38, 963-974.
    Li, D. C., Chen, L. S., Lin, Y. S. (2003). Using functional virtual population as assistance to learn scheduling knowledge in dynamic manufacturing environments. International Journal of Production Research 41(17), 4011-4024.
    Li, D. C., Hsu, H. C., Tsai, T. I., Lu, T. J., Hu, S. C. (2007a). A new method to help diagnose cancers for small sample size. Expert Systems with Applications, 33 420-424.
    Li, D. C., Lin, Y. S. (2006). Using virtual sample generation to build up management knowledge in the early manufacturing stages. European Journal of Operational Research, 175 413-434.
    Li, D. C., Tsai, T. I., Shi, S. (2009). A prediction of the dielectric constant of multi-layer ceramic capacitors using the mega-trend-diffusion technique in powder pilot runs: Case study. International Journal of Production Research 19, 51-69.
    Li, D. C., Wu, C., Chen, F. M. (2005). Using Data-fuzzification Technology in Small Data Set Learning to Improve FMS Scheduling Accuracy. International Journal of Advanced Manufacturing Technology 27, 321-328.
    Li, D. C., Wu, C., Tsai, T. I., and Chang, F. M. (2006). Using Mega-Fuzzification and Data Trend Estimation in Small Data Set Learning for Early FMS Scheduling Knowledge. Computers & Operations Research 33, 1857-1869.
    Li, D. C., Wu, S. S., Tsai, T. I., Lina, Y. S. (2007b).Using mega-trend-diffusion and artificial samples in small data set learning for early flexible manufacturing system scheduling knowledge, Computers and Operations Research, 34, 966-982
    Loh, W.Y. (2002). Regression trees with unbiased variable selection and interaction detection. Statistica Sinica, 12, 361-386.
    Malerba, D., Esposito, F., Ceci, M.,& Appice, A. (2004). Top-Down Induction of Model Trees with Regression and Splitting Nodes. IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 26, No. 5, 612-625.
    Niyogi,P., Girosi, F., Tomaso, P. (1998). Incorporating prior information in machine learning by creating virtual examples. Proceeding of the IEEE 275-298.
    Quinlan, J. R. (1979). “Discovering rules from large collections of examples:a case study.” In Michie, D., editor , Expert System in the Microelectronic Age. Edinburgh Scotland: Edinburgh University Press.
    Quinlan, J. R. (1992). Learning with continuous classes. Proceedings of the Australian Joint Conference on Artificial Intelligence, 343-348.
    Quinlan, J. R. (1993). “C4.5:Programs for Machine Learning.” San Mateo CA :Morgan Kaufmann.
    Ribeiro, B. (2002). On Data Based Learning Using Support Vector Clustering. Proceeding of the 9th International Conference on Neural Information Processing (ICONIP’02), Vol. 5,2516-2521.
    Roiger, R. J.,& Geatz, M. W. (2003). Data Mining:A Tutorial-Based Primer. New York:Addison Wesley.
    Schervish, M. (1995). Probabilistic inference and inference diagrams. Operations Research 36, 589-604.
    Vapnik, V. N. (1995). The Nature of Statistical Learning Theory, Springer-Verlag, New York.
    Wang, H. F., Huang, C. J. (2008). Data construction method for the analysis of the spatial distribution of disastrous earthquakes in Taiwan. 189~212
    Wang, Y. (2003). On-Demand Forecasting of Stock Prices Using a Real-Time Predictor. IEEE Transactions on Knowledge and Data Engineering. 1033-1037 Wang, Y., and Witten, I.H. (1997). Inducing model trees for continuous classes, Proceedings of poster papers of the 9th European Conference on Machine Learning.
    Wang, Y., Song Q., MacDonell S., Shepperd M., Shen J. (2009). Integrate the GM(1,1) and Verhulst Models to Predict SoftWare Stage Effort. 647-658
    Witten, I. H.,& Frank, E. (2005). Data Mining:Practical Machine Learning Tools and Techniques (2nd ed.). San Francisco:Morgan Kaufmann.

    下載圖示 校內:2015-07-28公開
    校外:2015-07-28公開
    QR CODE