簡易檢索 / 詳目顯示

研究生: 陳慶鴻
Chen, Ching-Heng
論文名稱: 在樣本數不足的情況下降低SVM法的學習誤差
A Method to Reduce the Learning Error in Support Vector Machine under Insufficient Sample Cases
指導教授: 利德江
Li, Der-Chaing
學位類別: 碩士
Master
系所名稱: 管理學院 - 工業管理科學系
Department of Industrial Management Science
論文出版年: 2003
畢業學年度: 91
語文別: 英文
論文頁數: 39
中文關鍵詞:
外文關鍵詞: virtual examples, dimension, feature space, generalization theory, SVM, machine learning
相關次數: 點閱:61下載:3
分享至:
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報
  • In the field of machine learning, occasions of insufficient data are often encountered. Especially when time and cost are limited. Without exception, the emerging learning method SVM (support vector machine) also faces this situation. Since a small data set usually leads learning systems to a low learning accuracy, find a way to cope with the problem becoming meaningful in academics consequently.
    Theoretically, so called insufficient data don’t mean the absolute number of data is small, but mainly points on the inappropriate ratios between the number of data and the associated dimension. It is clean viewing this concept through the Generalization theory. The theory fully illustrates the relationship among the learning error, the data size, and the dimensions. Therefore, to reduce the learning error by determining a proper relationship between size and dimensions is the basic approach proposed in this study. Technically, two ways are the method to increase learning accuracies including (1) Virtual Samples generation, and (2) dimension reducing. The study will derivate an algorithm for generating a set of training samples for leaning and refer the reason of dimension reducing.

    ABSTRACT i LIST OF CONTENTS ii LIST OF TABLES iv LIST OF FIGURES v Chapter 1 Introduction 1 Chapter 2 Literature Review 4 2.1 Basic Model of the SVM 4 2.2 Generalization Theory 10 2.3 Regularization Theory 13 2.4 Virtual Examples and Features Selection 15 2.5 The experiment of the SVM 19 Chapter 3 Methodology 21 3.1 Machine learning 21 3.2 Error of learning with small training set in SVM 22 3.2 Generating training examples 23 Chapter 4 The Experiment of the study 28 4.1 The setting of the experiment 28 4.2 The results of the experiment 29 Chapter 5 Discussion and The Future 34 Reference 35 Appendix A 37

    Evgeniou, T., Poggio, T. Pontil, M., and Verri, A, Regularization and statistical learning theory for data analysis, Computational Statistics & Data Analysis, 38, pp. 421-431, 2002.
    Neumann, J, L earning the systematic transformation of holographic reduced representations, Cognitive Systems Research, pp. 227–235, 2002.
    Niyogi, P., Girosi, N., and Poggio, T., Incorporating Prior Information in Machine Learning by Creating Virtual Examples, Proceedings of the IEEE, Vol. 86,no. 11, 1998.
    Platt, J., Fast training of support vector machines using sequential minimal optimization. Advances in Kernel Methods – Support Vector Learning, pp. 185-208. MIT Press, 1999.
    Rosenblatt, F., The perceptron: a probabilistic model for information storage and organization in the brain. Psychological Reviews, 65, pp. 386-408, 1959.
    Smola, A.J., Bcholkopf, B., and Muller, K., The connection between regularization operators and support vector kernels, Neural Networks, 11, pp. 637-649, 1998.
    Tikhonov, A.N., Arsenin, V.Y., Solutions of Ill-posed Problems. Winston, W.H, Washington,D.C. , 1977.
    Vapnik, V. and Chervonenkis, A.Ja., On the uniform convergence of relative frequencies of events to their probabilities, Theory Probab. Apl.16, pp. 264-280, 1971.
    Vapnik, V., The Nature of Statistical Learning Theory, Springer-Verlag, NY, 1999.
    Vukovich,G., Pedrycz,w., Feature analysis through information granulation and fuzzy sets. Pattern Recognition, Vol. 35, pp.825-834, 2002.
    Wahba, G.., Splines Models for Observational Data. Series in Applied Mathematics, Philadelphia, Vol. 59, SIAM, 1990.

    下載圖示 校內:2004-07-08公開
    校外:2004-07-08公開
    QR CODE