Abstract
With the advent of powerful desktop computers, organizations are recognizing that data can be more than that. By using the appropriate tools and techniques an experienced analyst can convert voluminous data into valuable information. This can be used to highlight the success (or failure) of marketing campaigns, display processes and be more responsive to customer needs. There are a wide variety of techniques that can be employed for data analysis and increasingly the term ‘data mining’ is used to describe these techniques.
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Anderberg M R: ‘Cluster analysis for applications’, Academic Press (1973).
Jain A K A and Dubes R C: ‘Algorithms for clustering data’, Prentice Hall (1988).
Taylor C, Michie D and Spiegelhalter D: ‘Machine learning, neural and statistical classification’, Ellis Horwood (1994).
Totton K A and Limb P R: ‘Electronic diagnosis using a multilayer perceptron’, BT Technol J, 10, No 3, pp 97–102 (1992).
Tattersall G D et al: ‘Feature extraction & visualization of decision support data’, BT Technol J, 10, No 3, pp 110–123 (1992).
Tattersall G D: ‘Investigation of data pre-processing for neural net classifiers’, University of East Anglia, DOC: 007, Issue 1 (Nov 1992).
Cox S J and Chichlowski K: ‘The application of standard statistical techniques to pattern classification’, University of East Anglia, DOC: 009, Issue 1 (March 1993).
Chatfield M G and Collins A J: ‘Introduction to multivariate analysis’, Chapman & Hall (1980).
McCulloch W C and Pitts W: ‘A logical calculus of the ideas imminent in nervous activity’, Bulletin of Mathematical Biophysics, 5, pp 115–133 (1943).
Rumelhart D E and McClelland J L: ‘Parallel distributed processing, Volume 1, Foundations’, The MIT Press (1987).
Caudhill M: ‘Neural networks primer’, Parts I-VIII, AI Expert (Dec 1987 to Aug 1989 ).
Lippman R P: ‘An introduction to computing with neural nets’, IEEE ASSP Magazine, 4–23 (Apr 1987).
Quinlan J R: ‘Cr.5! Programs for machine learning’, Morgan Kauffman, San Mateo, California (1993).
Hunt et al: ‘Experiments in induction’, Academic Press, New York (1966).
Quinlan J R: ‘Learning efficient classification procedures and their application to chess end games’ in Michalski R, Carbonnel J and Mitchell T (Eds): ‘Machine learning: an artificial intelligence approach’, Palo Alto! Tioga (1983).
Forsyth R and Rada R: ‘Machine intelligence: Applications in expert systems and information retrieval’, p 59–64 (1986).
Brieman L et al: ‘Classification and regression trees’, Wadsworth, Monterrey, CA (1984).
Boswell R A: ‘Manual for NewID version 2.0’, Technical Report TI/P2154/RAB54/, Turing Institute (Jan 1990).
Mueller W and Wysotzki F: ‘Automatic construction of decision trees for classification’, in Moser K and Schader M (Eds): ‘Annals of Operational Research 32’, J C Baltzer A G Science Publishers, Wijdenes, The Netherlands (1994).
Nadel B A: ‘Constraint satisfaction algorithms’, Computational Intelligence, 5, Part 4, pp 188–224 (Nov 1989).
Winston P H: ‘Learning structural descriptions from examples’, in Winston P H (Ed): ‘The psychology of computer vision’, McGraw-Hill (1975).
Michalski R and Larson J: ‘Incremental generation of VLI hypotheses: The underlying methodology and the description of program QA11’, Urbana: University of Illinois at Urbana-Champaign, Dept of Computer Science Report (ISG 83 - 5 ) (1973).
Thornton C J: ‘Techniques in computational learning’, Chapman & Hall (1992).
Clark P and Niblett T: ‘Induction in noisy domains’, in Brakto I and Lavrac N (Eds): ‘Progress i machine learning’, Sigma Press (1987).
Michaelski R, Mozetic I, Hong J and Lavrac N: ‘The multi-purpose incremental learning system AQ15 and its testing application to three medical domains’, in Proc AAAAI-86, California, Morgan Kauffman (1986).
Fisher R A: ‘The use of multiple measurements in taxonomic problems’, Annals of Eugenics, 7, pp 179–177 (1936).
Michalski R, Mozetic I, Hong J and Lavrac N: ‘The multi-purpose incremental learning system AQ15 and its testing application to three medical domains’, Proc AAI-86, California, Morgan Kauffman (1986).
Clarke W R et al: ‘How non-normality affects the quadratic discriminant function’, Comm Statistics — Theory and Methods IT-16, pp 41–46 (1979).
Cox D R: ‘Some procedures associated with the logistic qualitative response curve’ in Dvid F N (Ed): ‘Research papers on statistics: Festschrift for J. Neyman’, pp 57–77, John Wiley, New York (1966).
Mahalanobis P C: ‘Historical note on the D3-statistic’, Sankhya 9, p 237 (1948).
Fix E and Hodges J L: ‘Discriminatory analysis, nonparametric estimation: consistency properties’ Report 4, Project 21-49-004, USAF School of Aviation Medicine, Randolph Field, Texas (1951).
Freidman J H: ‘SMART’s user guide’ Technical Report No 1, Laboratory of Computational Statistics, Department of Statistics, Stanford University (1984).
Holland J H: ‘Adaptation in natural and artifical systems’, Ann Arbor: The University of Michigan Press (1975).
Kelly J and Davis L: ‘Hybridizing the GA and K-nearest neighbors classification algorithm’, in Proceedings of the Fourth International Conference on Genetic Algorithms, California (Jul 1991).
Dodd N: ‘Optimization of network structure using genetic techniques’, AIENG-91: Applications of artificial intelligence in engineering 6: Proceedings of the sixth international conference. Oxford UK, pp 939–944 (July 1991).
Goldberg D: ‘Genetic algorithms in search, optimization and machine learning’, (1989).
Davis L: ‘Genetic algorithms and simulated annealing’, Pitman, London (1987).
Stone M: ‘Cross-validatory choice and assessment of statistical predictions’, J Roy Statist Soc, 36, pp 111–33 (1974).
Efron B: ‘Estimating the error rate of a prediction rule: improvements on cross- validation’, J Amer Stat Ass, 78, pp 316–331 (1983).
Fukunaga K: ‘Introduction to tatistical pattern recognition’, Academic Press (1972).
Massy W F: ‘On methods: discriminant analysis of audience characteristics’, J of Advertising Research, 5, pp 39–48 (1965).
Wilks S S: ‘Mathematical statistics’, John Wiley & Sons (1963).
Weiss S M and Julikowski C A: ‘Computer systems that learn’, Morgan Kauffman (1991).
Gordon A D: ‘Classification’, Chapman & Hall (1981).
Piatesky-Shapiro G and Frawley W J: ‘Knowledge Discovery in Databases’, AAAI Press (1991).
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1996 British Telecommunications plc
About this chapter
Cite this chapter
Limb, P.R., Meggs, G.J. (1996). Data Mining — Tools and Techniques. In: Flavin, P.G., Totton, K.A.E. (eds) Computer Aided Decision Support in Telecommunications. BT Telecommunications Series, vol 8. Springer, Dordrecht. https://doi.org/10.1007/978-94-009-0081-3_3
Download citation
DOI: https://doi.org/10.1007/978-94-009-0081-3_3
Publisher Name: Springer, Dordrecht
Print ISBN: 978-94-010-6524-5
Online ISBN: 978-94-009-0081-3
eBook Packages: Springer Book Archive