Skip to main content

Clustering and Visualising Documents using Word Embeddings

As corpora are increasingly ‘born digital’ on hard drives as well as web and email servers, we are moving from being able to select or group documents using keyword or manual searches to needing to be able to automate this task at scale. Moreover, large-ish, unlabelled corpora of thousands or tens-of-thousands of documents are not particularly well-suited to topic modelling or TF/IDF analysis either. Since we don’t have a sense of what kinds of groups might exist, what kinds of topics might be covered, or what level of distinctiveness in vocabulary might matter, we need different, more flexible ways to visualise and extract structure from texts.

This lesson shows one way to achieve this: uncovering meaningful structure in a large corpus of about 9,000 documents through the use of two techniques — dimensionality reduction and hierarchical clustering — to find and group similar documents with minimal human guidance. Our approach to document classification is unsupervised: we do not use either keywords or human expertise — except to validate the results and provide a measure of ‘quality’ — relying instead on the information contained in the text itself.

To do this we take advantage of word and document embeddings; these lie at the root of recent advances in text-mining and Natural Language Processing, and they provide us with a numerical representation of a text that extends what’s possible with counts or TF/IDF representations of text. We take these embeddings and then apply our selected techniques to extract a hierarchical structure of relationships from the corpus. In this lesson, we’ll explore why documents on similar topics tend be closer in the (numerical) ‘space’ of the word and document embeddings than those that are on very different topics.

Reviewed by:

  • Quinn Dombrowski
  • Barbara McGillivray

Learning outcomes

After completing this lesson, you will be able to:

  • Appreciate the ‘curse of dimensionality’ and understand why it is important to text mining
  • Use (nonlinear) dimensionality reduction to reveal structure in corpora
  • Use hierarchical clustering to group similar documents within a corpus

Cite as

Jonathan Reades and Jennie Williams (2023). Clustering and Visualising Documents using Word Embeddings. Version 1.0.0. Edited by Alex Wermer-Colan. ProgHist Ltd. [Training module]. https://doi.org/10.46430/phen0111

Reuse conditions

Resources hosted on DARIAH-Campus are subjects to the DARIAH-Campus Training Materials Reuse Charter

Full metadata

Title:
Clustering and Visualising Documents using Word Embeddings
Authors:
Jonathan Reades, Jennie Williams
Domain:
Social Sciences and Humanities
Language:
en
Published:
7/11/2024
Content type:
Training module
Licence:
CCBY 4.0
Sources:
Programming Historian
Topics:
DH, Open education, Open access, Data visualisation, Machine Learning, Python
Version:
1.0.0