Advertisement

Statistical Approach to the Identification of Separation Surface for Spatial Data

  • Yee LeungEmail author
Chapter
Part of the Advances in Spatial Science book series (ADVSPATIAL)

Abstract

In spatial clustering, spatial objects are grouped into clusters according to their similarities. In terms of learning or pattern recognition, it belongs to the identification of structures/classes through an unsupervised process. In terms of data mining, it is the discovery of intrinsic classes, particularly new classes, in spatial data. It formulates class structures and determines the number of classes. I have examined in Chap. 2 the importance of clustering as a means for unraveling interesting, useful and natural patterns in spatial data. The process generally does not involve how to separate predetermined classes, or how to determine whether classes are significantly different from each other, or how to assign new objects to given classes. Another fundamental issue of spatial knowledge discovery involves spatial classification. It essentially deals with the separation of pre-specified classes and the assignment of new spatial objects to these classes on the basis of some measurements (with respect to selected features) about them. In terms of learning or pattern recognition, it is actually a supervised learning process which searches for the decision surface separating appropriately various classes. In terms of data mining, it often involves the discovery of classification rules from the training/learning data set that can separate distinct/genuine classes of spatial objects and the assignment of new spatial objects to these labeled classes. Whether the pre-specified classes are significantly different is usually not the main concern in classification. It can be determined by procedures such as the analysis of variance in statistics.

Keywords

Support Vector Machine Feature Space Linear Discriminant Analysis Credit Card Decision Function 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

References

  1. Aha DW, Kibler D, Albert MK (1991) Instance-based learning algorithms. Mach Learn 6:37–66Google Scholar
  2. Allenby GM, Rossi PE (1994) Modeling household purchase behavior with logistic normal regression. J Am Stat Assoc 89:1218–1231CrossRefGoogle Scholar
  3. Anderson JA (1982) Logistic discrimination. In: Krishnaiah PR, Kanal L (eds) Hand book of statistics, vol 2. North-Holland, Amsterdam, pp 169–191Google Scholar
  4. Angulo C, Catala A (2000) K_SVCR. A Multi-class support vector machine. In: Lopez de Mantaras R, Plaza E (eds) ECML 2000, LNAI 1810. Springer, Berlin, pp. 31–38Google Scholar
  5. Berkson J (1944) Application of the logistic function to bio-assay. J Am Stat Assoc 39:357–365CrossRefGoogle Scholar
  6. Brown M, Lewis HG, Gunn SR (1999) Support vector machines for spectral unmixing. IGRASS’99 2:1363–1365Google Scholar
  7. Burges C, Scholkopf B (1997) Improving the accuracy and speed of support vector machines. In: Mozer M, Jordan M, Petsche T (eds) Neural information processing systems. MIT, CambridgeGoogle Scholar
  8. Burges CJC (1998) A tutorial on support vector machines for pattern recognition. Data Min Knowl Discov 2:121–167CrossRefGoogle Scholar
  9. Carpenter GA, Grossberg S, Reynolds JH (1991) ARTMAP: supervised real time learning and classification of nonstationary data by a self-organising neural network. Neural Networks 4:565–588CrossRefGoogle Scholar
  10. Collett D (1991) Modelling binary data. Chapman and Hall, LondonGoogle Scholar
  11. Das Gupta S (1980) Discriminant analysis. In: Krishnaiah PR, Kanal L, Fisher RA (eds) An appreciation. Springer, New York, pp 161–170Google Scholar
  12. Efron B (1975) The efficiency of logistic regression compared to normal discriminant analysis. J Am Stat Assoc 70(352):892–898CrossRefGoogle Scholar
  13. Fisher RA (1936) The use of multiple measurements in taxonomic problems. Ann Eugen 7:179–188Google Scholar
  14. Fitzpatrick DB (1976) An analysis of bank credit card profit. J Bank Res 7:199–205Google Scholar
  15. Fix E, Hodges JL (1951) Discriminatory analysis – nonparametric discrimination: consistency properties. Report No. 4, Project no. 21-29-004. USAF School of Aviation Medicine, Randolph Field, Texas. Reprinted in Int Stat Rev 57(1989):238–247Google Scholar
  16. Geisser S (1982) Beyesian discrimination. In: Krishnaiah PR, Kanal L (eds) Hand book of statistics, vol 2. North-Holland, Amsterdam, pp 101–120Google Scholar
  17. Guadagni P, Little J (1983) A logit model of brand choice calibrated on scanner data. Market Sci 2:203–238CrossRefGoogle Scholar
  18. Hand DJ (1982) Kernel discriminant analysis. Research Studies Press, LetchworthGoogle Scholar
  19. Hand DJ, Henley WE (1997) Statistical classification methods in consumer credit scoring. J Roy Stat Soc Series A 160:523–541Google Scholar
  20. Hart PE (1968) The condensed nearest neighbour rule. IEEE Trans Inform Theor 14:515–516CrossRefGoogle Scholar
  21. Hastie TJ, Tibshirani RJ (1996) Discriminant analysis by Gaussian mixtures. J R Stat Soc B 58:155–176Google Scholar
  22. Hearst MA, Scholkopf B, Dumais S (1998) Trends and controversies-support vector machines. IEEE Intell Syst 13(4):18–28CrossRefGoogle Scholar
  23. Hermes L, Frieauff D, Puzicha J, Buhmann JM (1999) Support vector machines for land usage classification in Landsat TM imagery. In: Proceeding of the IEEE international geoscience and remote sensing symposium, vol. 1. Hamburg, pp. 348–350Google Scholar
  24. Hosmer DW, Lemeshow S (1989) Applied logistic regression. Wiley, New YorkGoogle Scholar
  25. John GH, Langley P (1995) Estimating continuous distributions in Bayesian classifiers In Proceedings of the 11th conference on uncertainty in artificial intelligence. Morgan Kaufmann, San Mateo, CAGoogle Scholar
  26. Keerthi SS, Shevade SK, Bhattacharya U, Murthy KRK (2000) A fast iterative nearest point algorithm for support vector machine classifier design. IEEE Trans Neural Network 11(1):124–136CrossRefGoogle Scholar
  27. Kohavi R (1996) Scaling up the accuracy of naive-Bayes classifiers: a deceision-tree hybrid. In: Proceedings of the second international conference on knowledge discovery and data mining. Morgan Kaufman, San Mateo CAGoogle Scholar
  28. Langley P, Sage S (1994) Induction of selective Bayesian classifers. In: Proceedings of the 10th conference on uncertainty in artificial intelligence. Morgan Kaufmann, Seattle, WA,Google Scholar
  29. Leung Y, Leung KS, Ma JH (2003a) Data mining for bank databases (unpublished paper)Google Scholar
  30. Leung Y, Leung KS, Mei CL (2003b) Data mining for credit card promotion in the banking sector (unpublished paper)Google Scholar
  31. Leung Y, Luo JC, Zhou CH, Ma JH (2002b) Support vector machine for spatial feature extraction and classification of high resolution remote sensing images (unpublished paper)Google Scholar
  32. McLachlan GJ (1992) Discriminant analysis and statistical pattern recognition. Wiley, New YorkCrossRefGoogle Scholar
  33. Menard SW (1995) Applied logistic regression analysis. Sage Publication, Thousand Oaks, CAGoogle Scholar
  34. Pearl J (1988) Probabilistic reasoning in intelligent systems: networks of plausible inference. Morgan Kaufmann, San Mateo, CAGoogle Scholar
  35. Piramuthu S (1999) Feature selection for financial credit-risk evaluation. Inform J Comput 11(3):258–266CrossRefGoogle Scholar
  36. Powell MJD (1987) Radial basis functions for multivariable interpolation: a review. In: Mason JC, Cox MG (eds) Algorithms for Approximation of Functions and Data. Oxford University Press, Oxford, pp 143–167Google Scholar
  37. Ramoni M, Sebastiani P (1996) Robust learning with missing data. Technical Report Kmi-TR-28, Knowledge Media Institute, The Open UniversityGoogle Scholar
  38. Rubinstein YD, Hastie TJ (1997) Discriminative vs informative learning. In: Proceedings of 3rd international conference on knowledge discovery and data mining. AAAI Press, Menlo Park, CA, pp. 49–53Google Scholar
  39. Scholkopf B, Burges CJC, Smola AJ (1999) Advances in kernel methods: support vector learning. MIT, CambridgeGoogle Scholar
  40. Scholkopf B, Sung KK, Burges CJC, Girosi F, Niyogi P, Poggio T, Vapnik VN (1997) Comparing support vector machines with Gaussian Kernels to radial basis function classifiers. IEEE Trans Signal Process 45(11):2758–2765CrossRefGoogle Scholar
  41. Smith CAB (1947) Some examples of discrimination. Ann Eugen 13:272–282Google Scholar
  42. Stanfill C, Waltz D (1986) Toward memory-based reasoning. Commun ACM 29:1213–1228CrossRefGoogle Scholar
  43. Thomas LC, Crook JN, Edelman DB (eds) (1992) Credit scoring and credit control. Clarendon Press, OxfordGoogle Scholar
  44. Vapnik VN (1995) The nature of statistical learning theory. Springer, New YorkGoogle Scholar
  45. Wand MP, Jones MC (1995) Kernel smoothing. Chapman and Hall, LondonGoogle Scholar
  46. Xu ZB, Leung Y (2004) How neural networks can be made more effective and efficient: a view of learning theory (unpublished paper)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2010

Authors and Affiliations

  1. 1.Dept. of Geography & Resource Management ShatinThe Chinese University of Hong KongNew TerritoriesHong Kong SAR

Personalised recommendations