Show simple item record

dc.contributor.authorNelson, Joshua
dc.description.abstractA problem that arises quite frequently in statistics is that of identifying groups, or clusters, of data within a population or sample. The most widely used procedure to identify clusters in a set of observations is known as K-Means. The main limitation of this algorithm is that it uses the Euclidean distance metric to assign points to clusters. Hence, this algorithm operates well only if the covariance structures of the clusters are nearly spherical and homogeneous in nature. To remedy this shortfall in the K-Means algorithm the Mahalanobis distance metric was used to capture the variance structure of the clusters. The issue with using Mahalanobis distances is that the accuracy of the distance is sensitive to initialization. If this method serves as a signicant improvement over its competitors, then it will provide a useful tool for analyzing clusters.en_US
dc.publisherNorth Dakota State Universityen_US
dc.rightsNDSU Policy 190.6.2
dc.titleOn K-Means Clustering Using Mahalanobis Distanceen_US
dc.typeThesisen_US
dc.date.accessioned2017-11-05T23:06:44Z
dc.date.available2017-11-05T23:06:44Z
dc.date.issued2012
dc.identifier.urihttps://hdl.handle.net/10365/26766
dc.rights.urihttps://www.ndsu.edu/fileadmin/policy/190.pdf
ndsu.degreeMaster of Science (MS)en_US
ndsu.collegeScience and Mathematicsen_US
ndsu.departmentStatisticsen_US
ndsu.programStatisticsen_US
ndsu.advisorMelnykov, Volodymyr


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record