GNN-as-Judge: Unleashing the Power of LLMs for Graph Learning with GNN Feedback

📰 ArXiv cs.AI

arXiv:2604.08553v1 Announce Type: cross Abstract: Large Language Models (LLMs) have shown strong performance on text-attributed graphs (TAGs) due to their superior semantic understanding ability on textual node features. However, their effectiveness as predictors in the low-resource setting, where labeled nodes are severely limited and scarce, remains constrained since fine-tuning LLMs usually requires sufficient labeled data, especially when the TAG shows complex structural patterns. In essence

Published 13 Apr 2026
Read full paper → ← Back to Reads