About cookies on this site Our websites require some cookies to function properly (required). In addition, other cookies may be used with your consent to analyze site usage, improve the user experience and for advertising. For more information, please review your options. By visiting our website, you agree to our processing of information as described in IBM’sprivacy statement. To provide a smooth navigation, your cookie preferences will be shared across the IBM web domains listed here.
Publication
Statistical Analysis and Data Mining
Paper
Distance-based mixture modeling for classification via hypothetical local mapping
Abstract
We propose a new approach for mixture modeling based only upon pairwise distances via the concept of hypothetical local mapping (HLM). This work is motivated by the increasingly commonplace applications involving complex objects that cannot be effectively represented by tractable mathematical entities. The new modeling approach consists of two steps. A distance-based clustering algorithm is applied first. Then, HLM takes as input the distances between the training data and their corresponding cluster centroids to estimate the model parameters. In the special case where all the training data are taken as cluster centroids, we obtain a distance-based counterpart of the kernel density. The classification performance of the mixture models is compared with other state-of-the-art distance-based classification methods. Results demonstrate that HLM-based algorithms are highly competitive in terms of classification accuracy and are computationally efficient. Furthermore, the HLM-based modeling approach adapts readily to incremental learning. We have developed and tested two schemes of incremental learning scalable for dynamic data arriving at a high speed.