Academic literature on the topic 'K-means clustering'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the lists of relevant articles, books, theses, conference reports, and other scholarly sources on the topic 'K-means clustering.'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Journal articles on the topic "K-means clustering"
Hedar, Abdel-Rahman, Abdel-Monem Ibrahim, Alaa Abdel-Hakim, and Adel Sewisy. "K-Means Cloning: Adaptive Spherical K-Means Clustering." Algorithms 11, no. 10 (October 6, 2018): 151. http://dx.doi.org/10.3390/a11100151.
Full textJhun, Myoungshic. "BOOTSTRAPPING K-MEANS CLUSTERING." Journal of the Japanese Society of Computational Statistics 3, no. 1 (1990): 1–14. http://dx.doi.org/10.5183/jjscs1988.3.1.
Full textTimmerman, Marieke E., Eva Ceulemans, Kim De Roover, and Karla Van Leeuwen. "Subspace K-means clustering." Behavior Research Methods 45, no. 4 (March 23, 2013): 1011–23. http://dx.doi.org/10.3758/s13428-013-0329-y.
Full textXiao, Ethan. "Comprehensive K-Means Clustering." Journal of Computer and Communications 12, no. 03 (2024): 146–59. http://dx.doi.org/10.4236/jcc.2024.123009.
Full textYu, Hengjun, Kohei Inoue, Kenji Hara, and Kiichi Urahama. "A Robust K-Means for Document Clustering." Journal of the Institute of Industrial Applications Engineers 6, no. 2 (April 25, 2018): 60–65. http://dx.doi.org/10.12792/jiiae.6.60.
Full textMadhuri, K., and Mr K. Srinivasa Rao. "Social Media Analysis using Optimized K-Means Clustering." International Journal of Trend in Scientific Research and Development Volume-3, Issue-2 (February 28, 2019): 953–57. http://dx.doi.org/10.31142/ijtsrd21558.
Full textRavindran, R. Malathi, and Dr Antony Selvadoss Thanamani. "K-Means Document Clustering using Vector Space Model." Bonfring International Journal of Data Mining 5, no. 2 (July 31, 2015): 10–14. http://dx.doi.org/10.9756/bijdm.8076.
Full textHUA, C., Q. CHEN, H. WU, and T. WADA. "RK-Means Clustering: K-Means with Reliability." IEICE Transactions on Information and Systems E91-D, no. 1 (January 1, 2008): 96–104. http://dx.doi.org/10.1093/ietisy/e91-d.1.96.
Full textJain, Preeti, and Bala Buksh. "Accelerated K-means Clustering Algorithm." International Journal of Information Technology and Computer Science 8, no. 10 (October 8, 2016): 39–46. http://dx.doi.org/10.5815/ijitcs.2016.10.05.
Full textFarzizadeh, Mohammad, and Ali Abdolahi. "Clustering Students By K-means." International Journal of Computer Applications Technology and Research 5, no. 8 (July 26, 2016): 530–32. http://dx.doi.org/10.7753/ijcatr0508.1006.
Full textDissertations / Theses on the topic "K-means clustering"
Buchta, Christian, Martin Kober, Ingo Feinerer, and Kurt Hornik. "Spherical k-Means Clustering." American Statistical Association, 2012. http://epub.wu.ac.at/4000/1/paper.pdf.
Full textMusco, Cameron N. (Cameron Nicholas). "Dimensionality reduction for k-means clustering." Thesis, Massachusetts Institute of Technology, 2015. http://hdl.handle.net/1721.1/101473.
Full textThis electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.
Cataloged from student-submitted PDF version of thesis.
Includes bibliographical references (pages 123-131).
In this thesis we study dimensionality reduction techniques for approximate k-means clustering. Given a large dataset, we consider how to quickly compress to a smaller dataset (a sketch), such that solving the k-means clustering problem on the sketch will give an approximately optimal solution on the original dataset. First, we provide an exposition of technical results of [CEM+15], which show that provably accurate dimensionality reduction is possible using common techniques such as principal component analysis, random projection, and random sampling. We next present empirical evaluations of dimensionality reduction techniques to supplement our theoretical results. We show that our dimensionality reduction algorithms, along with heuristics based on these algorithms, indeed perform well in practice. Finally, we discuss possible extensions of our work to neurally plausible algorithms for clustering and dimensionality reduction. This thesis is based on joint work with Michael Cohen, Samuel Elder, Nancy Lynch, Christopher Musco, and Madalina Persu.
by Cameron N. Musco.
S.M.
Persu, Elena-Mădălina. "Approximate k-means clustering through random projections." Thesis, Massachusetts Institute of Technology, 2015. http://hdl.handle.net/1721.1/99847.
Full textCataloged from PDF version of thesis.
Includes bibliographical references (pages 39-41).
Using random row projections, we show how to approximate a data matrix A with a much smaller sketch à that can be used to solve a general class of constrained k-rank approximation problems to within (1 + [epsilon]) error. Importantly, this class of problems includes k-means clustering. By reducing data points to just O(k) dimensions, our methods generically accelerate any exact, approximate, or heuristic algorithm for these ubiquitous problems. For k-means dimensionality reduction, we provide (1+ [epsilon]) relative error results for random row projections which improve on the (2 + [epsilon]) prior known constant factor approximation associated with this sketching technique, while preserving the number of dimensions. For k-means clustering, we show how to achieve a (9 + [epsilon]) approximation by Johnson-Lindenstrauss projecting data points to just 0(log k/[epsilon]2 ) dimensions. This gives the first result that leverages the specific structure of k-means to achieve dimension independent of input size and sublinear in k.
by Elena-Mădălina Persu.
S.M. in Computer Science and Engineering
Xiang, Chongyuan. "Private k-means clustering : algorithms and applications." Thesis, Massachusetts Institute of Technology, 2016. http://hdl.handle.net/1721.1/106394.
Full textThis electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.
Cataloged from student-submitted PDF version of thesis.
Includes bibliographical references (pages 77-80).
Today is a new era of big data. We contribute our personal data for the common good simply by using our smart phones, searching the web and doing online transactions. Researchers, companies and governments use the collected data to learn various user behavior patterns and make impactful decisions based on that. Is it possible to publish and run queries on those databases without disclosing information about any specific individual? Differential privacy is a strong notion of privacy which guarantees that very little will be learned about individual records in the database, no matter what the attackers already know or wish to learn. Still, there is no practical system applying differential privacy algorithms for clustering points on real databases. This thesis describes the construction of small coresets for computing k-means clustering of a set of points while preserving differential privacy. As a result, it gives the first 𝑘-means clustering algorithm that is both differentially private, and has an approximation error that depends sub-linearly on the data’s dimension d. Previous results introduced errors that are exponential in d. This thesis implements this algorithm and uses it to create differentially private location data from GPS tracks. Specifically the algorithm allows clustering GPS databases generated from mobile nodes, while letting the user control the introduced noise due to privacy. This thesis also provides experimental results for the system and algorithms, and compares them to existing techniques. To the best of my knowledge, this is the first practical system that enables differentially private clustering on real data.
by Chongyuan Xiang.
M. Eng.
Nelson, Joshua. "On K-Means Clustering Using Mahalanobis Distance." Thesis, North Dakota State University, 2012. https://hdl.handle.net/10365/26766.
Full textLi, Yanjun. "High Performance Text Document Clustering." Wright State University / OhioLINK, 2007. http://rave.ohiolink.edu/etdc/view?acc_num=wright1181005422.
Full textELIASSON, PHILIP, and NIKLAS ROSÉN. "Efficient K-means clustering and the importanceof seeding." Thesis, KTH, Skolan för datavetenskap och kommunikation (CSC), 2013. http://urn.kb.se/resolve?urn=urn:nbn:se:kth:diva-134910.
Full textKlustring av data innebär att man grupperar dataelement baserat på någon typ a likhet mellan de grupperade elementen. Klustring har många olika användningsråden såsom datakompression, datautvinning, mönsterigenkänning, och maskininlärning och det finns många olika klustringsmetoder. Den här uppsatsen undersöker klustringsmetoden k-means och hur valet av startvärden för metoden påverkar resultatet. Lloyds algorithm används som utgångspunkt och den jämförs med en förbättrad algorithm som använder sig av kd-träd. Två olika metoder att välja startvärden jämförs, slumpmässigt val av startvärde och delklustring.
Kondo, Yumi. "Robustification of the sparse K-means clustering algorithm." Thesis, University of British Columbia, 2011. http://hdl.handle.net/2429/37093.
Full textChowuraya, Tawanda. "Online content clustering using variant K-Means Algorithms." Thesis, Cape Peninsula University of Technology, 2019. http://hdl.handle.net/20.500.11838/3089.
Full textWe live at a time when so much information is created. Unfortunately, much of the information is redundant. There is a huge amount of online information in the form of news articles that discuss similar stories. The number of articles is projected to grow. The growth makes it difficult for a person to process all that information in order to update themselves on a subject matter. There is an overwhelming amount of similar information on the internet. There is need for a solution that can organize this similar information into specific themes. The solution is a branch of Artificial intelligence (AI) called machine learning (ML) using clustering algorithms. This refers to clustering groups of information that is similar into containers. When the information is clustered people can be presented with information on their subject of interest, grouped together. The information in a group can be further processed into a summary. This research focuses on unsupervised learning. Literature has it that K-Means is one of the most widely used unsupervised clustering algorithm. K-Means is easy to learn, easy to implement and is also efficient. However, there is a horde of variations of K-Means. The research seeks to find a variant of K-Means that can be used with an acceptable performance, to cluster duplicate or similar news articles into correct semantic groups. The research is an experiment. News articles were collected from the internet using gocrawler. gocrawler is a program that takes Universal Resource Locators (URLs) as an argument and collects a story from a website pointed to by the URL. The URLs are read from a repository. The stories come riddled with adverts and images from the web page. This is referred to as a dirty text. The dirty text is sanitized. Sanitization is basically cleaning the collected news articles. This includes removing adverts and images from the web page. The clean text is stored in a repository, it is the input for the algorithm. The other input is the K value. All K-Means based variants take K value that defines the number of clusters to be produced. The stories are manually classified and labelled. The labelling is done to check the accuracy of machine clustering. Each story is labelled with a class to which it belongs. The data collection process itself was not unsupervised but the algorithms used to cluster are totally unsupervised. A total of 45 stories were collected and 9 manual clusters were identified. Under each manual cluster there are sub clusters of stories talking about one specific event. The performance of all the variants is compared to see the one with the best clustering results. Performance was checked by comparing the manual classification and the clustering results from the algorithm. Each K-Means variant is run on the same set of settings and same data set, that is 45 stories. The settings used are, • Dimensionality of the feature vectors, • Window size, • Maximum distance between the current and predicted word in a sentence, • Minimum word frequency, • Specified range of words to ignore, • Number of threads to train the model. • The training algorithm either distributed memory (PV-DM) or distributed bag of words (PV-DBOW), • The initial learning rate. The learning rate decreases to minimum alpha as training progresses, • Number of iterations per cycle, • Final learning rate, • Number of clusters to form, • The number of times the algorithm will be run, • The method used for initialization. The results obtained show that K-Means can perform better than K-Modes. The results are tabulated and presented in graphs in chapter six. Clustering can be improved by incorporating Named Entity (NER) recognition into the K-Means algorithms. Results can also be improved by implementing multi-stage clustering technique. Where initial clustering is done then you take the cluster group and further cluster it to achieve finer clustering results.
Li, Songzi. "K-groups: A Generalization of K-means by Energy Distance." Bowling Green State University / OhioLINK, 2015. http://rave.ohiolink.edu/etdc/view?acc_num=bgsu1428583805.
Full textBooks on the topic "K-means clustering"
Wu, Junjie. Advances in K-means Clustering. Berlin, Heidelberg: Springer Berlin Heidelberg, 2012. http://dx.doi.org/10.1007/978-3-642-29807-3.
Full textRoy, Falguni. Seismic signal detection using K-means clustering algorithm. Mumbai: Bhabha Atomic Research Centre, 2009.
Find full textservice), SpringerLink (Online, ed. Advances in K-means Clustering: A Data Mining Thinking. Berlin, Heidelberg: Springer Berlin Heidelberg, 2012.
Find full textSchreiber, Thomas. A voronoi diagram based adaptive k-means-type clustering algorithm for multidimensional weighted data. Kaiserslautern: Universität Kaiserslautern, 1991.
Find full textK, Kokula Krishna Hari, and K. Saravanan, eds. Identification of Brain Regions Related to Alzheimers’ Diseases using MRI Images Based on Eigenbrain and K-means Clustering. Tiruppur, Tamil Nadu, India: Association of Scientists, Developers and Faculties, 2016.
Find full textWu, Junjie. Advances in K-means Clustering: A Data Mining Thinking. Springer, 2012.
Find full textWu, JunJie. Advances in K-Means Clustering: A Data Mining Thinking. Springer Berlin / Heidelberg, 2014.
Find full textKaur, Arvind, and Nancy Nancy. Comparative Analysis of Hybrid Clustering Algorithm with K- Means. Independently Published, 2018.
Find full textWong, M. Anthony. Using the K-Means Clustering Method As a Density Estimation Procedure. Creative Media Partners, LLC, 2018.
Find full textTuckfield, Bradford, and Alok Malik. Applied Unsupervised Learning with R: Uncover Hidden Relationships and Patterns with K-Means Clustering, Hierarchical Clustering, and PCA. Packt Publishing, Limited, 2019.
Find full textBook chapters on the topic "K-means clustering"
Zhou, Hong. "K-Means Clustering." In Learn Data Mining Through Excel, 35–47. Berkeley, CA: Apress, 2020. http://dx.doi.org/10.1007/978-1-4842-5982-5_3.
Full textNg, Annalyn, and Kenneth Soo. "k-Means-Clustering." In Data Science – was ist das eigentlich?!, 19–28. Berlin, Heidelberg: Springer Berlin Heidelberg, 2018. http://dx.doi.org/10.1007/978-3-662-56776-0_2.
Full textDinov, Ivo D. "k-Means Clustering." In Data Science and Predictive Analytics, 443–73. Cham: Springer International Publishing, 2018. http://dx.doi.org/10.1007/978-3-319-72347-1_13.
Full textMannor, Shie, Xin Jin, Jiawei Han, Xin Jin, Jiawei Han, Xin Jin, Jiawei Han, and Xinhua Zhang. "K-Means Clustering." In Encyclopedia of Machine Learning, 563–64. Boston, MA: Springer US, 2011. http://dx.doi.org/10.1007/978-0-387-30164-8_425.
Full textStanberry, Larissa. "Clustering, k-Means." In Encyclopedia of Systems Biology, 430–31. New York, NY: Springer New York, 2013. http://dx.doi.org/10.1007/978-1-4419-9863-7_1189.
Full textSreevalsan-Nair, Jaya. "K-Means Clustering." In Encyclopedia of Mathematical Geosciences, 1–3. Cham: Springer International Publishing, 2021. http://dx.doi.org/10.1007/978-3-030-26050-7_171-1.
Full textJin, Xin, and Jiawei Han. "K-Means Clustering." In Encyclopedia of Machine Learning and Data Mining, 1–3. Boston, MA: Springer US, 2016. http://dx.doi.org/10.1007/978-1-4899-7502-7_431-1.
Full textJin, Xin, and Jiawei Han. "K-Means Clustering." In Encyclopedia of Machine Learning and Data Mining, 695–97. Boston, MA: Springer US, 2017. http://dx.doi.org/10.1007/978-1-4899-7687-1_431.
Full textZhou, Hong. "K-Means Clustering." In Learn Data Mining Through Excel, 37–52. Berkeley, CA: Apress, 2023. http://dx.doi.org/10.1007/978-1-4842-9771-1_3.
Full textSreevalsan-Nair, Jaya. "K-Means Clustering." In Encyclopedia of Mathematical Geosciences, 695–97. Cham: Springer International Publishing, 2023. http://dx.doi.org/10.1007/978-3-030-85040-1_171.
Full textConference papers on the topic "K-means clustering"
Di Fatta, Giuseppe, Francesco Blasa, Simone Cafiero, and Giancarlo Fortino. "Epidemic K-Means Clustering." In 2011 IEEE International Conference on Data Mining Workshops (ICDMW). IEEE, 2011. http://dx.doi.org/10.1109/icdmw.2011.76.
Full textAgarwal, Pankaj K., and Nabil H. Mustafa. "k-means projective clustering." In the twenty-third ACM SIGMOD-SIGACT-SIGART symposium. New York, New York, USA: ACM Press, 2004. http://dx.doi.org/10.1145/1055558.1055581.
Full textArandjelovic, Ognjen. "Discriminative k-means clustering." In 2013 International Joint Conference on Neural Networks (IJCNN 2013 - Dallas). IEEE, 2013. http://dx.doi.org/10.1109/ijcnn.2013.6707038.
Full textAsgharbeygi, Nima, and Arian Maleki. "Geodesic K-means clustering." In 2008 19th International Conference on Pattern Recognition (ICPR). IEEE, 2008. http://dx.doi.org/10.1109/icpr.2008.4761241.
Full textBorgelt, Christian, and Olha Yarikova. "Initializing k-means Clustering." In 9th International Conference on Data Science, Technology and Applications. SCITEPRESS - Science and Technology Publications, 2020. http://dx.doi.org/10.5220/0009872702600267.
Full textGoel, Anurag, and Angshul Majumdar. "Transformed K-means Clustering." In 2021 29th European Signal Processing Conference (EUSIPCO). IEEE, 2021. http://dx.doi.org/10.23919/eusipco54536.2021.9616177.
Full textNa, Shi, Liu Xumin, and Guan Yong. "Research on k-means Clustering Algorithm: An Improved k-means Clustering Algorithm." In 2010 Third International Symposium on Intelligent Information Technology and Security Informatics (IITSI). IEEE, 2010. http://dx.doi.org/10.1109/iitsi.2010.74.
Full textDashti, Hesam T., Tiago Simas, Rita A. Ribeiro, Amir Assadi, and Andre Moitinho. "MK-means - Modified K-means clustering algorithm." In 2010 International Joint Conference on Neural Networks (IJCNN). IEEE, 2010. http://dx.doi.org/10.1109/ijcnn.2010.5596300.
Full textQi, Jianpeng, Yanwei Yu, Lihong Wang, and Jinglei Liu. "K*-Means: An Effective and Efficient K-Means Clustering Algorithm." In 2016 IEEE International Conferences on Big Data and Cloud Computing (BDCloud), Social Computing and Networking (SocialCom), Sustainable Computing and Communications (SustainCom) (BDCloud-SocialCom-SustainCom). IEEE, 2016. http://dx.doi.org/10.1109/bdcloud-socialcom-sustaincom.2016.46.
Full textSingh, Vivek Kumar, Nisha Tiwari, and Shekhar Garg. "Document Clustering Using K-Means, Heuristic K-Means and Fuzzy C-Means." In 2011 International Conference on Computational Intelligence and Communication Networks (CICN). IEEE, 2011. http://dx.doi.org/10.1109/cicn.2011.62.
Full textReports on the topic "K-means clustering"
Kanungo, T., D. M. Mount, N. S. Netanyahu, C. Piatko, R. Silverman, and A. Y. Wu. The Analysis of a Simple k-Means Clustering Algorithm. Fort Belvoir, VA: Defense Technical Information Center, January 2000. http://dx.doi.org/10.21236/ada458738.
Full textCordeiro de Amorim, Renato. A survey on feature weighting based K-Means algorithms. Web of Open Science, December 2020. http://dx.doi.org/10.37686/ser.v1i2.79.
Full textKryzhanivs'kyi, Evstakhii, Liliana Horal, Iryna Perevozova, Vira Shyiko, Nataliia Mykytiuk, and Maria Berlous. Fuzzy cluster analysis of indicators for assessing the potential of recreational forest use. [б. в.], October 2020. http://dx.doi.org/10.31812/123456789/4470.
Full textHerrera, Allen, and Alexander Heifetz. Detection of Anomalies in Gamma Background Radiation Data with K-Means and Self-Organizing Map Clustering Algorithms - Consortium on Nuclear Security Technologies (CONNECT) Q1 Report. Office of Scientific and Technical Information (OSTI), December 2021. http://dx.doi.org/10.2172/1841591.
Full textEshed-Williams, Leor, and Daniel Zilberman. Genetic and cellular networks regulating cell fate at the shoot apical meristem. United States Department of Agriculture, January 2014. http://dx.doi.org/10.32747/2014.7699862.bard.
Full textMultiple Engine Faults Detection Using Variational Mode Decomposition and GA-K-means. SAE International, March 2022. http://dx.doi.org/10.4271/2022-01-0616.
Full text