Alert button
Picture for Jiaang Li

Jiaang Li

Alert button

Random Entity Quantization for Parameter-Efficient Compositional Knowledge Graph Representation

Oct 24, 2023
Jiaang Li, Quan Wang, Yi Liu, Licheng Zhang, Zhendong Mao

Figure 1 for Random Entity Quantization for Parameter-Efficient Compositional Knowledge Graph Representation
Figure 2 for Random Entity Quantization for Parameter-Efficient Compositional Knowledge Graph Representation
Figure 3 for Random Entity Quantization for Parameter-Efficient Compositional Knowledge Graph Representation
Figure 4 for Random Entity Quantization for Parameter-Efficient Compositional Knowledge Graph Representation

Representation Learning on Knowledge Graphs (KGs) is essential for downstream tasks. The dominant approach, KG Embedding (KGE), represents entities with independent vectors and faces the scalability challenge. Recent studies propose an alternative way for parameter efficiency, which represents entities by composing entity-corresponding codewords matched from predefined small-scale codebooks. We refer to the process of obtaining corresponding codewords of each entity as entity quantization, for which previous works have designed complicated strategies. Surprisingly, this paper shows that simple random entity quantization can achieve similar results to current strategies. We analyze this phenomenon and reveal that entity codes, the quantization outcomes for expressing entities, have higher entropy at the code level and Jaccard distance at the codeword level under random entity quantization. Therefore, different entities become more easily distinguished, facilitating effective KG representation. The above results show that current quantization strategies are not critical for KG representation, and there is still room for improvement in entity distinguishability beyond current strategies. The code to reproduce our results is available at https://github.com/JiaangL/RandomQuantization.

* Accepted to EMNLP 2023 
Viaarxiv icon

Copyright Violations and Large Language Models

Oct 20, 2023
Antonia Karamolegkou, Jiaang Li, Li Zhou, Anders Søgaard

Language models may memorize more than just facts, including entire chunks of texts seen during training. Fair use exemptions to copyright laws typically allow for limited use of copyrighted material without permission from the copyright holder, but typically for extraction of information from copyrighted materials, rather than {\em verbatim} reproduction. This work explores the issue of copyright violations and large language models through the lens of verbatim memorization, focusing on possible redistribution of copyrighted text. We present experiments with a range of language models over a collection of popular books and coding problems, providing a conservative characterization of the extent to which language models can redistribute these materials. Overall, this research highlights the need for further examination and the potential impact on future developments in natural language processing to ensure adherence to copyright regulations. Code is at \url{https://github.com/coastalcph/CopyrightLLMs}.

* EMNLP 2023 
Viaarxiv icon

PokemonChat: Auditing ChatGPT for Pokémon Universe Knowledge

Jun 05, 2023
Laura Cabello, Jiaang Li, Ilias Chalkidis

Figure 1 for PokemonChat: Auditing ChatGPT for Pokémon Universe Knowledge
Figure 2 for PokemonChat: Auditing ChatGPT for Pokémon Universe Knowledge
Figure 3 for PokemonChat: Auditing ChatGPT for Pokémon Universe Knowledge
Figure 4 for PokemonChat: Auditing ChatGPT for Pokémon Universe Knowledge

The recently released ChatGPT model demonstrates unprecedented capabilities in zero-shot question-answering. In this work, we probe ChatGPT for its conversational understanding and introduce a conversational framework (protocol) that can be adopted in future studies. The Pok\'emon universe serves as an ideal testing ground for auditing ChatGPT's reasoning capabilities due to its closed world assumption. After bringing ChatGPT's background knowledge (on the Pok\'emon universe) to light, we test its reasoning process when using these concepts in battle scenarios. We then evaluate its ability to acquire new knowledge and include it in its reasoning process. Our ultimate goal is to assess ChatGPT's ability to generalize, combine features, and to acquire and reason over newly introduced knowledge from human feedback. We find that ChatGPT has prior knowledge of the Pokemon universe, which can reason upon in battle scenarios to a great extent, even when new information is introduced. The model performs better with collaborative feedback and if there is an initial phase of information retrieval, but also hallucinates occasionally and is susceptible to adversarial attacks.

Viaarxiv icon

Large Language Models Converge on Brain-Like Word Representations

Jun 02, 2023
Jiaang Li, Antonia Karamolegkou, Yova Kementchedjhieva, Mostafa Abdou, Sune Lehmann, Anders Søgaard

Figure 1 for Large Language Models Converge on Brain-Like Word Representations
Figure 2 for Large Language Models Converge on Brain-Like Word Representations
Figure 3 for Large Language Models Converge on Brain-Like Word Representations
Figure 4 for Large Language Models Converge on Brain-Like Word Representations

One of the greatest puzzles of all time is how understanding arises from neural mechanics. Our brains are networks of billions of biological neurons transmitting chemical and electrical signals along their connections. Large language models are networks of millions or billions of digital neurons, implementing functions that read the output of other functions in complex networks. The failure to see how meaning would arise from such mechanics has led many cognitive scientists and philosophers to various forms of dualism -- and many artificial intelligence researchers to dismiss large language models as stochastic parrots or jpeg-like compressions of text corpora. We show that human-like representations arise in large language models. Specifically, the larger neural language models get, the more their representations are structurally similar to neural response measurements from brain imaging.

* Work in process 
Viaarxiv icon

Inductive Relation Prediction from Relational Paths and Context with Hierarchical Transformers

Apr 19, 2023
Jiaang Li, Quan Wang, Zhendong Mao

Figure 1 for Inductive Relation Prediction from Relational Paths and Context with Hierarchical Transformers
Figure 2 for Inductive Relation Prediction from Relational Paths and Context with Hierarchical Transformers
Figure 3 for Inductive Relation Prediction from Relational Paths and Context with Hierarchical Transformers
Figure 4 for Inductive Relation Prediction from Relational Paths and Context with Hierarchical Transformers

Relation prediction on knowledge graphs (KGs) is a key research topic. Dominant embedding-based methods mainly focus on the transductive setting and lack the inductive ability to generalize to new entities for inference. Existing methods for inductive reasoning mostly mine the connections between entities, i.e., relational paths, without considering the nature of head and tail entities contained in the relational context. This paper proposes a novel method that captures both connections between entities and the intrinsic nature of entities, by simultaneously aggregating RElational Paths and cOntext with a unified hieRarchical Transformer framework, namely REPORT. REPORT relies solely on relation semantics and can naturally generalize to the fully-inductive setting, where KGs for training and inference have no common entities. In the experiments, REPORT performs consistently better than all baselines on almost all the eight version subsets of two fully-inductive datasets. Moreover. REPORT is interpretable by providing each element's contribution to the prediction results.

* Accepted by ICASSP 2023 (Oral) 
Viaarxiv icon

Implications of the Convergence of Language and Vision Model Geometries

Feb 13, 2023
Jiaang Li, Yova Kementchedjhieva, Anders Søgaard

Figure 1 for Implications of the Convergence of Language and Vision Model Geometries
Figure 2 for Implications of the Convergence of Language and Vision Model Geometries
Figure 3 for Implications of the Convergence of Language and Vision Model Geometries
Figure 4 for Implications of the Convergence of Language and Vision Model Geometries

Large-scale pretrained language models (LMs) are said to ``lack the ability to connect [their] utterances to the world'' (Bender and Koller, 2020). If so, we would expect LM representations to be unrelated to representations in computer vision models. To investigate this, we present an empirical evaluation across three different LMs (BERT, GPT2, and OPT) and three computer vision models (VMs, including ResNet, SegFormer, and MAE). Our experiments show that LMs converge towards representations that are partially isomorphic to those of VMs, with dispersion, and polysemy both factoring into the alignability of vision and language spaces. We discuss the implications of this finding.

* work in progress 
Viaarxiv icon