Abstract
In alignment with the distributional hypothesis of language, the work “Quantum Projections on Conceptual Subspaces” (Martínez-Mingo A, Jorge-Botana G, Martinez-Huertas JÁ, et al. Quantum projections on conceptual subspaces. Cogn Syst Res 2023; 82: 101154) proposed a methodology for generating conceptual subspaces from textual information based on previous work (Martinez-Mingo A, Jorge-Botana G and Olmos R. Quantum approach for similarity evaluation in LSA vector space models. 2020). These subspaces enable the utilization of the quantum model of similarity put forth by Pothos and Busemeyer (Pothos E, Busemeyer J. A quantum probability explanation for violations of symmetry in similarity judgments. In
Keywords
Introduction
Recently, we have witnessed a genuine revolution in computational language studies. Large language models (LLMs) such as the GPT family models have fundamentally altered the landscape, outperforming all prior expectations. Intriguingly, this leap did not result from a significant advance in academic research—although the merit of Transformer
7
architectures should not be understated. Instead, it has primarily been a matter of scaling up existing pre-trained language models (PLMs), supplemented by a suite of optimization and training methodologies tailored for commercial applications.
8
This scaling, in terms of both training data and model parameters, has spawned emergent behaviors in LLMs that have led many in the scientific community to reconsider the feasibility of general artificial intelligence in the not-so-distant future. Specifically,
With this overarching goal, we are keen to provide an academic commentary on the paper “Quantum Projections on Conceptual Subspaces,” 1 which implicitly raises a research question that challenges the very foundations of distributional language models. Our discussion leans on the seminal experiments conducted by Tversky about human cognitive biases like the violation of the assumptions of symmetry and triangle inequality in cognitive similarity assessments. 4 In his groundbreaking 1977 work, Tversky posed an exceptionally pertinent question: are spatial models adequately suited for assessing psychological similarity? Given Tversky's findings, the answer appears to be a resounding no. Building upon these insights, researchers like Pothos, Busemeyer, and Yearsley3,5,6 have proposed extending the conventional geometric model to a subspace-based model. This extension enables the incorporation of mathematical apparatus derived from the axiomatization of quantum mechanics into the study of similarity. 3 It is within this context that an important research question concerning computational language models emerges: Are we employing the correct geometry? This leads us to hypothesize that the classical geometric model will be insufficient if the goal is to emulate, in an ecologically valid manner, the cognitive processes inherent in human language. Studying perspectives that merge logic with conceptual subspaces could inspire better optimization methodologies over the architectures that allow LLMs to work. It could open ways to think new representation and updating mechanisms in order to overcome their sample-based style of learning. Considerations about computational and algorithmic human plausibility in some tasks as understanding metaphors, abstract reasoning, systematic compositionality and contextual adaptability could result in LLMs improvements.
Key points from the article
To the best of our knowledge, there are no existing methodologies that enable the creation of conceptual subspaces within a
Building upon an established semantic space, we introduce an approach for estimating the semantic contour of a target term. This contour serves as a reservoir of key information needed for constructing a subsequent conceptual subspace. The foundational idea for generating such a conceptual subspace is to create it based on terms that are semantically related to this target term, which we define as the semantic contour of the target term.
2
To ascertain this pertinent information, we suggest utilizing the similarity function between the target term and all terms in the corpus, which is then contrasted with the similarity function of an abstract, intermediary term within the semantic space, coined as the
Once the contour of the target concept is created, we delineate the corresponding subspace by undertaking dimensionality reduction. This aims to capture orthogonal components that account for the maximum variance within the contour, thereby ensuring that the semantic terms defining the contour are represented by a reduced number of dimensions that effectively summarize and simplify the semantic contents of the contour, and consequently, the target concept. The selection of the subspace dimensions is guided by the Parallel Analysis technique. Here, the deserved number of dimensions is chosen by comparing the variance explained by the contour's components against that explained by components in a random matrix of equivalent dimensions. Thus, the resulting representation subspace for a target concept will consist of
In the paper under discussion, the proposed methodology successfully accounts for asymmetry biases and diagnosticity effects. Regarding the violation of the triangle inequality, it holds in terms of similarity within both the classical geometric model and the quantum model. Importantly, no violations of this assumption in terms of distances were observed with the stimuli utilized across either model. 1 This commentary has a multifaceted aim: to provide methodological clarifications, delve into theoretical and practical implications, and offer informed speculation on the trajectory of future research in this domain. More specifically, the article advocates for employing various types of contours—either conceptual or contextual—as a basis for generating semantic subspaces. Once the contours are clearly defined, an analytical distinction is introduced between three types of subspaces: aggregate term subspaces (ATSs), aggregate context subspaces (ACSs), and aggregate feature subspaces (AFSs). Lastly, the commentary introduces new data that disrupts traditional geometric assumptions, specifically the triangular inequality assumption.
A deeper dive into methodological challenges and opportunities
Firstly, it is important to clarify that our choice of latent semantic analysis (LSA) as the technique for constructing the container semantic space is dictated by the resultant vector space's nature, which ensures orthogonality of the generated latent dimensions via singular value decomposition (SVD). This orthogonality constraint was initially adopted to ensure the container space's conformity to a Hilbert space—being this orthogonality a desirable but not essential feature for this type of vector spaces.
10
While this condition does confer interpretability to the system when it collapses into a basis vector, the latent dimensions in LSA are not inherently interpretable, thus adding no additional informative value and merely constraining our container space. Consequently, acknowledging the container space as merely a coordinate system that abstractly represents information (term vectors in the space can be considered
Regarding contour generation, the original paper outlines a methodology for composing contours based on the term vectors nearest to the target term (
Contour of China in a 176 dimensions BERT container space
In the final stage of the proposed methodology, 1 Principal Component Analysis (PCA) is employed for contour dimensionality reduction. However, we do not elaborate on the nature of the matrix subjected to decomposition. In the article under review, PCA is applied to the LSA latent dimensions in such a way that the retained eigenvectors, as determined by Parallel Analysis, have a direct representation in the container space, obviating the need for subsequent identification and orthogonalization. Yet, as evident from the findings of this study, these obtained dimensions have limited interpretability.
We want to clarify that, within the conceptual window provided by the conceptual contour of a target term, it will be feasible to perform dimensionality reduction either on the terms, or on the latent features. In this light, we suggest that for each conceptual contour, two types of subspaces could be defined: an

Descriptors WordClouds for China's estimated AFS and ATS.
Finally, we would like to augment this commentary with additional results from applying Aggregated Term Subspaces (ATSs) to investigate violations of the triangular inequality assumption. Although the commented article specifies that such violations should theoretically be possible within distance metrics, no instances were reported therein.
1
However, when we apply the conceptual distance measure proposed by Gabora and Aerts
14
to the ATS of
Conclusion and future work
In conclusion, we wish to highlight the novelty of the approach presented in the article under discussion, 1 which represents a promising attempt at the modeling of natural language using the mathematical tools provided by the axiomatization of the basic principles of quantum mechanics. Although this line of research is still in its early stages, it holds considerable potential, thereby creating numerous opportunities for future studies in this area. Moreover, the empirical testing of the effects outlined in the paper—and their potential relation to the results obtained from the computational application of this methodology—is of utmost importance, but it is essential to define metrics capable of evaluating the performance of the various versions of the proposed method, particularly in the context of standard tasks in natural language processing and semantic analysis. Specifically, one of the main challenges of the quantum similarity model proposed in the commented paper is its potential difficulty in evaluating hierarchical relationships between concepts (subspaces) and their features, which may also be concepts (other subspaces) with their own sets of features. This concept-feature duality can be easily integrated into the model in a recursive fashion, yet a method for representing semantic relationships that denote a certain hierarchy has not been established. The research conducted by Martínez-Huertas et al. 15 may provide insights into this issue by exploring possible higher- and lower-order dimensions in conceptual subspaces.
In a more targeted vein, we see considerable value in undertaking a thorough examination of the characteristics of the various subspaces that can be formed (ATS, ACS, or AFS) at different layers of abstraction in transformer-based models. This could potentially demystify the so-called
Footnotes
Declaration of conflicting interests
The authors declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The authors received no financial support for the research, authorship, and/or publication of this article.
Author biographies
Alejandro Martínez-Mingo holds a PhD in Cognitive Psychology and Language Models and serves as a professor of Natural Language Processing. His research area lies at the intersection of cognition and computational linguistics, exploring how models from quantum physics can be adapted to elucidate cognitive phenomena.
Jose Ángel Martínez-Huertas is a professor in Formal Models of Cognitive Processes. His research focuses on the development of computational language models and statistical models for explaining psychological processes.
Ricardo Olmos is a professor in Psychometrics and Data Analysis in Psychology. His research interests include computational linguistics, psychometrics in the realm of Classical Test Theory (CTT) and Item Response Theory (IRT), and statistics within linear models.
Guillermo Jorge-Botana is a professor in Supervised Learning. His research area concentrates on the development of computational models for the representation of linguistic and cognitive phenomena.
