Studying large language models as compression algorithms for human culture

Trends Cogn Sci. 2024 Mar;28(3):187-189. doi: 10.1016/j.tics.2024.01.001. Epub 2024 Jan 19.

Abstract

Large language models (LLMs) extract and reproduce the statistical regularities in their training data. Researchers can use these models to study the conceptual relationships encoded in this training data (i.e., the open internet), providing a remarkable opportunity to understand the cultural distinctions embedded within much of recorded human communication.

Keywords: compression; cultural psychology; large language models.

MeSH terms

  • Algorithms
  • Communication*
  • Humans
  • Language*