top of page

CLUSTERING

1_o4PXxETRsZV_AjLhOxGk-g.png

OVERVIEW

 

Clustering is a technique in machine learning and data analysis that involves grouping similar data points together. It aims to discover inherent structures or patterns in the data, where items within the same group (cluster) are more similar to each other than to those in other groups. Clustering is often unsupervised, meaning that it does not rely on predefined labels. In the context of the text summarization project, clustering serves as a valuable technique for organizing and discovering patterns within the textual data. Clustering involves grouping together documents that share similarities, which can be instrumental in uncovering underlying structures in the dataset.

image-5.png
demo-clustering-biconnected-components.c7bcb893be.png

Partitional Clustering: 

Partitional clustering involves dividing the dataset into non-overlapping subsets or clusters. K-Means is a popular partitional clustering algorithm that partitions data points into 'k' clusters based on their similarity. Euclidean distance is commonly used in partitional clustering, but other metrics like cosine similarity can also be applied.

As the text dataset contains diverse articles, partitional clustering algorithms can be leveraged to partition the documents into distinct groups. K-means, for instance, can help identify clusters of articles with similar content.

Hierarchical Clustering:

Hierarchical clustering creates a tree-like structure (dendrogram) of nested clusters.

Agglomerative Hierarchical Clustering is a common hierarchical clustering algorithm that starts with individual data points as clusters and gradually merges them based on similarity. In hierarchical clustering, various distance metrics can be used, such as Euclidean distance or cosine similarity. The choice depends on the nature of the data. With the hierarchical nature of text data, hierarchical clustering allows the creation of a tree-like structure of clusters. This can be particularly beneficial when exploring relationships between different topics and subtopics within the dataset.

unnamed.png

Distance Metrics Used

Given that the dataset comprises textual information, choosing appropriate distance metrics is crucial. Commonly used metrics include:

Cosine Similarity: Measures the cosine of the angle between two vectors, providing a measure of similarity between documents.

Euclidean Distance: Euclidean distance is a measure of the straight-line distance between two points in a multidimensional space, commonly used as a distance metric in clustering algorithms.

.

Clustering plays a pivotal role in the discovery process for the this project:

Topic Discovery: Clustering helps identify articles that revolve around similar topics. This is beneficial for organizing large datasets and understanding the prevalent themes within the collection.

Content Relationships:By grouping documents with similar content, insights into relationships between articles can be uncovered, contributing to an understanding of how different topics are interconnected.

bottom of page