[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/502512.502527acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
Article

Proximal support vector machine classifiers

Published: 26 August 2001 Publication History

Abstract

Instead of a standard support vector machine (SVM) that classifies points by assigning them to one of two disjoint half-spaces, points are classified by assigning them to the closest of two parallel planes (in input or feature space) that are pushed apart as far as possible. This formulation, which can also be interpreted as regularized least squares and considered in the much more general context of regularized networks [8, 9], leads to an extremely fast and simple algorithm for generating a linear or nonlinear classifier that merely requires the solution of a single system of linear equations. In contrast, standard SVMs solve a quadratic or a linear program that require considerably longer computational time. Computational results on publicly available datasets indicate that the proposed proximal SVM classifier has comparable test set correctness to that of standard SVM classifiers, but with considerably faster computational time that can be an order of magnitude faster. The linear proximal SVM can easily handle large datasets as indicated by the classification of a 2 million point 10-attribute set in 20.8 seconds. All computational results are based on 6 lines of MATLAB code.

References

[1]
E. Anderson, Z. Bai, C. Bischof, S. Blackford, J. Demmel, J. Dongarra, J. Du Croz, A. Greenbaum, S. Hammarling, A. McKenney, and D. Sorensen. LAPACK User's Guide. SIAM, Philadelphia, Pennsylvania, third edition, 1999. http://www.netlib.org/lapack/.
[2]
K. P. Bennett and O. L. Mangasarian. Robust linear programming discrimination of two linearly inseparable sets. Optimization Methods and Software, 1:23-34, 1992.
[3]
P. S. Bradley and O. L. Mangasarian. Massive data discrimination via linear support vector machines. Optimization Methods and Software, 13:1-10, 2000. ftp://ftp.cs.wisc.edu/math-prog/tech-reports/98- 03.ps.
[4]
US Census Bureau. Adult dataset. Publicly available from: www.sgi.com/Technology/mlc/db/.
[5]
C. J. C. Burges. A tutorial on support vector machines for pattern recognition. Data Mining and Knowledge Discovery, 2(2):121-167, 1998.
[6]
V. Cherkassky and F. Mulier. Learning from Data - Concepts, Theory and Methods. John Wiley & Sons, New York, 1998.
[7]
CPLEX Optimization Inc., Incline Village, Nevada. Using the CPLEX(TM) Linear Optimizer and CPLEX(TM) Mixed Integer Optimizer (Version 2.0), 1992.
[8]
T. Evgeniou, M. Pontil, and T. Poggio. Regularization networks and support vector machines. Advances in Computational Mathematics, 13:1-50, 2000.
[9]
T. Evgeniou, M. Pontil, and T. Poggio. Regularization networks and support vector machines. In A. Smola, P. Bartlett, B. Sch51kopf, and D. Schuurmans, editors, Advances in Large Margin Classifiers, pages 171-203, Cambridge, MA, 2000. MIT Press.
[10]
M. C. Ferris and T. S. Munson. Interior point methods for massive support vector machines. Technical Report 00-05, Computer Sciences Department, University of Wisconsin, Madison, Wisconsin, May 2000. ftp://ftp.cs.wisc.edu/pub/dmi/tech-reports/00-05.ps.
[11]
G. Fung and O. L. Mangasarian. Data selection for support vector machine classification. In R. Ramakrishnan and S. Stolfo, editors, Proceedings KDD2000: Knowledge Discovery and Data Mining, August 20-23, 2000, Boston, MA, pages 64-70, New York, 2000. Asscociation for Computing Machinery. ftp://ftp.cs.wisc.edu/pub/dmi/tech-reports/00-02.ps.
[12]
G. Fung, O. L. Mangasarian, and A. Smola. Minimal kernel classifiers. Technical Report 00-08, Data Mining Institute, Computer Sciences Department, University of Wisconsin, Madison, Wisconsin, November 2000. ftp: //ftp.cs, wisc.edu /pub/dmi/tech-reports/ O0-O8.ps.
[13]
D. Gale. The Theory of Linear Economic Models. McGraw-Hill Book Company, New York, 1960.
[14]
G. H. Golub and C. F. Van Loan. Matrix Computations. The John Hopkins University Press, Baltimore, Maryland, 3rd edition, 1996.
[15]
J. Gracke, M. Griebel, and M. Thess. Data mining with sparse grids. Technical report, Institut f/ir Angrwandte Mathematik, Universita~t Bonn, Bonn, Germany, 2000. http://wissrech.iam.unibonn.de/research/projects/garcke/sparsemining.html.
[16]
T. Joachims. Making large-scale support vector machine learning practical. In B. SchSlkopf, C. J. C. Burges, and A. J. Smola, editors, Advances in Kernel Methods - Support Vector Learning, pages 169-184. MIT Press, 1999.
[17]
Y.-J. Lee and O. L. Mangasarian. RSVM: Reduced support vector machines. Technical Report 00-07, Data Mining Institute, Computer Sciences Department, University of Wisconsin, Madison, Wisconsin, July 2000. Proceedings of the First SIAM International Conference on Data Mining, Chicago, April 5-7, 2001, CD-ROM. ftp://ftp.cs.wisc.edu/pub/dmi/tech-reports/00-07.ps.
[18]
Yuh-Jye Lee and O. L. Mangasarian. SSVM: A smooth support vector machine. Technical Report 99-03, Data Mining Institute, Computer Sciences Department, University of Wisconsin, Madison, Wisconsin, September 1999. Computational Optimization and Applications 20(1), October 2001, to appear. ftp://ftp.cs.wisc.edu/pub/dmi/tech-reports/99-03.ps.
[19]
O. L. Mangasarian. Nonlinear Programming. SIAM, Philadelphia, PA, 1994.
[20]
O. L. Mangasarian. Generalized support vector machines. In A. Smola, P. Bartlett, B. SchSlkopf, and D. Schuurmans, editors, Advances in Large Margin Classifiers, pages 135:146, Cambridge, MA, 2000. MIT Press. ftp://ftp.cs.wisc.edu/math-prog/techreports/98-14.ps.
[21]
O. L. Mangasarian and R. R. Meyer. Nonlinear perturbation of linear programs. SIAM Journal on Control and Optimization, 17(6):745-752, November 1979.
[22]
O. L. Mangasarian and D. R. Musicant. Successive overrelaxation for support vector machines. IEEE Transactions on Neural Networks, 10:1032-1037, 1999. ftp://ftp.cs.wisc.edu/math-prog/tech-reports/98- 18.ps.
[23]
O. L. Mangasarian and D. R. Musicant. Active support vector machine classification. Technical Report 00-04, Data Mining Institute, Computer Sciences Department, University of Wisconsin, Madison, Wisconsin, April 2000. Machine Learning, to appear. ftp://ftp.cs.wisc.edu/pub/dmi/tech-reports/OO-O4.ps.
[24]
O. L. Mangasarian and D. R. Musicant. Lagrangian support vector machines. Technical Report 00-06, Data Mining Institute, Computer Sciences Department, University of Wisconsin, Madison, Wisconsin, June 2000. Journal of Machine Learning Research, to appear. ftp://ftp.cs.wisc.edu/pub/dmi/tech-reports/00-06.ps.
[25]
O. L. Mangasarian and T.-H. Shiau. Lipschitz continuity of solutions of linear inequalities, programs and complementarity problems. SIAM Journal on Control and Optimization, 25(3):583-595, May 1987.
[26]
MATLAB. User's Guide. The MathWorks, Inc., Natick, MA 01760, 1994-2001. http://www.mathworks.com.
[27]
T. M. Mitchell. Machine Learning. McGraw-Hill, Boston, 1997.
[28]
P. M. Murphy and D. W. Aha. UCI repository of machine learning databases, 1992. www.ics.uci.edu/~mlearn/MLRepository.html.
[29]
D. R. Musicant. NDC: normally distributed clustered datasets, 1998. www.cs.wisc.edu/~-.musicant/data/ndc/.
[30]
S. Odewahn, E. Stockwell, R. Pennington, R. Humphreys, and W. Zumach. Automated star/galaxy discrimination with neural networks. Astronomical Journal, 103(1):318-331, 1992.
[31]
J. Platt. Sequential minimal optimization: A fast algorithm for training support vector machines. In B. SchSlkopf, C. J. C. Burges, and A. J. Smola, editors, Advances in Kernel Methods - Support Vector Learning, pages 185-208. MIT Press, 1999. http://www.research.microsoft.com/~jplatt/smo.html.
[32]
A. Smola, P. L. Bartlett, B. SchSlkopf, and J. Schiirmann (editors). Advances in Large Margin Classifiers. MIT Press, Cambridge, MA, 2000.
[33]
J. A. K. Suykens and J. Vandewalle. Least squares support vector machine classifiers. Neural Processing Letters, 9(3):293-300, 1999.
[34]
A. N. Tikhonov and V. Y. Arsenin. Solutions of Ill-Posed Problems. John Wiley ~: Sons, New York, 1977.
[35]
V. N. Vapnik. The Nature of Statistical Learning Theory. Springer, New York, 1995.
[36]
V. N. Vapnik. The Nature of Statistical Learning Theory. Springer, New York, second edition, 2000.
[37]
Alexis Wieland. Twin spiral dataset, http://wwwcgi.cs.cmu.edu/afs/cs.cmu.edu/project/airepository/ai/areas/neural/bench/cmu/0.html.

Cited By

View all
  • (2025)Support matrix machine: A reviewNeural Networks10.1016/j.neunet.2024.106767181(106767)Online publication date: Jan-2025
  • (2024)An Overview on the Advancements of Support Vector Machine Models in Healthcare Applications: A ReviewInformation10.3390/info1504023515:4(235)Online publication date: 19-Apr-2024
  • (2024)Robust control chart for nonlinear conditionally heteroscedastic time series based on Huber support vector regressionPLOS ONE10.1371/journal.pone.029912019:2(e0299120)Online publication date: 23-Feb-2024
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
KDD '01: Proceedings of the seventh ACM SIGKDD international conference on Knowledge discovery and data mining
August 2001
493 pages
ISBN:158113391X
DOI:10.1145/502512
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 26 August 2001

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. data classification
  2. linear equations
  3. support vector machines

Qualifiers

  • Article

Conference

KDD01
Sponsor:

Acceptance Rates

KDD '01 Paper Acceptance Rate 31 of 237 submissions, 13%;
Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)113
  • Downloads (Last 6 weeks)13
Reflects downloads up to 13 Dec 2024

Other Metrics

Citations

Cited By

View all
  • (2025)Support matrix machine: A reviewNeural Networks10.1016/j.neunet.2024.106767181(106767)Online publication date: Jan-2025
  • (2024)An Overview on the Advancements of Support Vector Machine Models in Healthcare Applications: A ReviewInformation10.3390/info1504023515:4(235)Online publication date: 19-Apr-2024
  • (2024)Robust control chart for nonlinear conditionally heteroscedastic time series based on Huber support vector regressionPLOS ONE10.1371/journal.pone.029912019:2(e0299120)Online publication date: 23-Feb-2024
  • (2024)Machine-learning-based models for the optimization of post-cervical spinal laminoplasty outpatient follow-up schedulesBMC Medical Informatics and Decision Making10.1186/s12911-024-02693-y24:1Online publication date: 30-Sep-2024
  • (2024)Twin Support Vector Machines Based on Chaotic Mapping Dung Beetle Optimization AlgorithmJournal of Computational Design and Engineering10.1093/jcde/qwae040Online publication date: 22-Apr-2024
  • (2024)Dual-modal measurements of suspended particles combining polarization and fluorescence analysisOptics & Laser Technology10.1016/j.optlastec.2024.111086177(111086)Online publication date: Oct-2024
  • (2024)Twin proximal support vector regression with heteroscedastic Gaussian noiseExpert Systems with Applications10.1016/j.eswa.2024.123840250(123840)Online publication date: Sep-2024
  • (2024)Intuitionistic fuzzy twin proximal SVM with fuzzy hyperplane and its application in EEG signal classificationApplied Soft Computing10.1016/j.asoc.2024.111816163(111816)Online publication date: Sep-2024
  • (2024)A comparative study of linear type multiple instance learning techniques for detecting COVID-19 by chest X-ray imagesProgress in Artificial Intelligence10.1007/s13748-024-00332-1Online publication date: 18-Sep-2024
  • (2024)Early Detection of Potato Disease Using an Enhanced Convolutional Neural Network-Long Short-Term Memory Deep Learning ModelPotato Research10.1007/s11540-024-09760-xOnline publication date: 8-Jul-2024
  • Show More Cited By

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media