Skip to content

Jianglin954/SFGL

Repository files navigation

Scale-Free Graph-Language Models (ICLR 2025)

Codes for paper Scale-Free Graph-Language Models

Contributions

  1. We identify two key challenges in existing GLMs: artificial structural assumptions in graph generation and unreliable LM finetuning for text embedding. We propose addressing these challenges simultaneously by exploring a well-grounded graph structural prior.

  2. We leverage the scale-free edge distribution in real networks as our graph structural prior. Our empirical validation and analysis reveal that a KNN graph, constructed using cosine similarity with an appropriately chosen k, effectively approximates a scale-free network.

  3. To the best of our knowledge, the proposed SFGL is the first work to unify graph generation and text embedding within a GLM framework, highlighting the synergistic potential of GNNs and LMs under a scale-free structural prior.

Datasets

Datasets can be download from here. Please place the downloaded files in the folder dataset.

Installation

conda env create -f SFGL_environment.yml

Usage

see bash files:

bash 23arxiv_gnn.sh
bash 23arxiv_gnn_lm.sh
bash 23arxiv_gnn_lm_gpt.sh
bash 23arxiv_gnn_lm_gnn.sh

The experimental results will be saved in "tmp_results" folder.

Reference

@inproceedings{
  lu2025scalefree,
  title={Scale-Free Graph-Language Models},
  author={Jianglin Lu and Yixuan Liu and Yitian Zhang and Yun Fu},
  booktitle={The Thirteenth International Conference on Learning Representations},
  year={2025},
  url={https://openreview.net/forum?id=nFcgay1Yo9}
}

Acknowledgement

Our code is mainly built on TAPE. We sincerely appreciate the authors for their valuable contributions.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published