Towards a parallel computationally efficient approach to scaling up data stream classificationTennant, M., Stahl, F. ORCID: https://orcid.org/0000-0002-4860-0203, Di Fatta, G. and Gomes, J. B. (2014) Towards a parallel computationally efficient approach to scaling up data stream classification. In: Thirty-fourth SGAI International Conference on Artificial Intelligence, 9-11 Dec 2014, Cambridge, England, pp. 51-65.
It is advisable to refer to the publisher's version if you intend to cite from this work. See Guidance on citing. Official URL: http://dx.doi.org/10.1007/978-3-319-12069-0_4 Abstract/SummaryAdvances in hardware technologies allow to capture and process data in real-time and the resulting high throughput data streams require novel data mining approaches. The research area of Data Stream Mining (DSM) is developing data mining algorithms that allow us to analyse these continuous streams of data in real-time. The creation and real-time adaption of classification models from data streams is one of the most challenging DSM tasks. Current classifiers for streaming data address this problem by using incremental learning algorithms. However, even so these algorithms are fast, they are challenged by high velocity data streams, where data instances are incoming at a fast rate. This is problematic if the applications desire that there is no or only a very little delay between changes in the patterns of the stream and absorption of these patterns by the classifier. Problems of scalability to Big Data of traditional data mining algorithms for static (non streaming) datasets have been addressed through the development of parallel classifiers. However, there is very little work on the parallelisation of data stream classification techniques. In this paper we investigate K-Nearest Neighbours (KNN) as the basis for a real-time adaptive and parallel methodology for scalable data stream classification tasks.
Download Statistics DownloadsDownloads per month over past year Deposit Details University Staff: Request a correction | Centaur Editors: Update this record |