Advances in Graph Neural Networks and Multimodal Learning

The field of graph neural networks and multimodal learning is rapidly evolving, with a focus on developing more robust and generalizable models. Recent research has highlighted the importance of integrating structural information from graphs with semantic reasoning from large language models, leading to improved performance on a range of tasks such as node classification, link prediction, and graph regression. Additionally, there is a growing interest in developing models that can defend against various types of attacks, including backdoor and adversarial attacks. Noteworthy papers in this area include: UniGTE, which introduces a unified graph-text encoding framework for zero-shot generalization across graph tasks and domains. RELATE, which proposes a schema-agnostic perceiver encoder for multimodal relational graphs, achieving performance within 3% of schema-specific encoders while reducing parameter counts by up to 5x. Graph4MM, which presents a graph-based multimodal learning framework that integrates multi-hop structural information into self-attention, outperforming larger VLMs, LLMs, and multimodal graph baselines on both generative and discriminative tasks.

Sources

Backdoor or Manipulation? Graph Mixture of Experts Can Defend Against Various Graph Adversarial Attacks

Interpretable Graph-Language Modeling for Detecting Youth Illicit Drug Use

UniGTE: Unified Graph-Text Encoding for Zero-Shot Generalization across Graph Tasks and Domains

Robustness in Text-Attributed Graph Learning: Insights, Trade-offs, and New Defenses

Graph4MM: Weaving Multimodal Learning with Structural Information

Learning Noise-Resilient and Transferable Graph-Text Alignment via Dynamic Quality Assessment

RELATE: A Schema-Agnostic Perceiver Encoder for Multimodal Relational Graphs

The Temporal Graph of Bitcoin Transactions

BadGraph: A Backdoor Attack Against Latent Diffusion Model for Text-Guided Graph Generation

Built with on top of