Graph foundation models for relational data
GraphNews
Kumo’s ‘relational foundation model’ predicts the future your LLM can’t see
Forecasting is a fundamentally new capability that is missing from the current purview of generative AI. Here's how Kumo is changing that.
Use Graph Machine Learning to detect fraud with Amazon Neptune Analytics and GraphStorm | Amazon Web Services
Every year, businesses and consumers lose billions of dollars to fraud, with consumers reporting $12.5 billion lost to fraud in 2024, a 25% increase year over year. People who commit fraud often work together in organized fraud networks, running many different schemes that companies struggle to detect and stop. In this post, we discuss how to use Amazon Neptune Analytics, a memory-optimized graph database engine for analytics, and GraphStorm, a scalable open source graph machine learning (ML) library, to build a fraud analysis pipeline with AWS services.
Alice enters the magical, branchy world of Graphs and Graph Neural Networks
The first draft 'G' chapter of the geometric deep learning book is live! 🚀
Alice enters the magical, branchy world of Graphs and Graph Neural Networks 🕸️ (Large Language Models are there too!)
I've spent 7+ years studying, researching & talking about graphs -- This text is my best attempt at conveying everything i've learnt 💎
You may read this chapter in the usual place (link in comments!)
Any and all feedback / thoughts / questions on the content, and/or words of encouragement for finishing this book (pretty please! 😇) are warmly welcomed!
Michael Bronstein Joan Bruna Taco Cohen | 18 comments on LinkedIn
Alice enters the magical, branchy world of Graphs and Graph Neural Networks
GDL Book
Grids, Groups, Graphs, Geodesics, and Gauges
Towards Multi-modal Graph Large Language Model
Multi-modal graphs are everywhere in the digital world.
Yet the tools used to understand them haven't evolved as much as one would expect.
What if the same model could handle your social network analysis, molecular discovery, AND urban planning tasks?
A new paper from Tsinghua University proposes Multi-modal Graph Large Language Models (MG-LLM) - a paradigm shift in how we process complex interconnected data that combines text, images, audio, and structured relationships.
Think of it as ChatGPT for graphs, but, metaphorically speaking, with eyes, ears, and structural understanding.
Their key insight? Treating all graph tasks as generative problems.
Instead of training separate models for node classification, link prediction, or graph reasoning, MG-LLM frames everything as transforming one multi-modal graph into another.
This unified approach means the same model that predicts protein interactions could also analyze social media networks or urban traffic patterns.
What makes this particularly exciting is the vision for natural language interaction with graph data. Imagine querying complex molecular structures or editing knowledge graphs using plain English, without learning specialized query languages.
The challenges remain substantial - from handling the multi-granularity of data (pixels to full images) to managing multi-scale tasks (entire graph input, single node output).
But if successful, this could fundamentally change the level of graph-based insights across industries that have barely scratched the surface of AI adoption.
↓
𝐖𝐚𝐧𝐭 𝐭𝐨 𝐤𝐞𝐞𝐩 𝐮𝐩? Join my newsletter with 50k+ readers and be the first to learn about the latest AI research: llmwatch.com 💡
Towards Multi-modal Graph Large Language Model
GitHub - GraGODs/GraGOD: Anomaly detection with GNNs
Anomaly detection with GNNs. Contribute to GraGODs/GraGOD development by creating an account on GitHub.
Multi-modal Graph Large Language Models (MG-LLM)
Multi-modal graphs are everywhere in the digital world.
Yet the tools used to understand them haven't evolved as much as one would expect.
What if the same model could handle your social network analysis, molecular discovery, AND urban planning tasks?
A new paper from Tsinghua University proposes Multi-modal Graph Large Language Models (MG-LLM) - a paradigm shift in how we process complex interconnected data that combines text, images, audio, and structured relationships.
Think of it as ChatGPT for graphs, but, metaphorically speaking, with eyes, ears, and structural understanding.
Their key insight? Treating all graph tasks as generative problems.
Instead of training separate models for node classification, link prediction, or graph reasoning, MG-LLM frames everything as transforming one multi-modal graph into another.
This unified approach means the same model that predicts protein interactions could also analyze social media networks or urban traffic patterns.
What makes this particularly exciting is the vision for natural language interaction with graph data. Imagine querying complex molecular structures or editing knowledge graphs using plain English, without learning specialized query languages.
The challenges remain substantial - from handling the multi-granularity of data (pixels to full images) to managing multi-scale tasks (entire graph input, single node output).
But if successful, this could fundamentally change the level of graph-based insights across industries that have barely scratched the surface of AI adoption.
↓
𝐖𝐚𝐧𝐭 𝐭𝐨 𝐤𝐞𝐞𝐩 𝐮𝐩? Join my newsletter with 50k+ readers and be the first to learn about the latest AI research: llmwatch.com 💡
Multi-modal Graph Large Language Models (MG-LLM)
Graph Embeddings — The Summary - Towards Data Science
This article present what graph embeddings are, their use, and the comparison of the most commonly used graph embedding approaches.
Graph technologies landscape
6 years after, where are we?
Interested by your feedback concerning the evolution of the graph technologies landscape and about what the current landscape is.
https://lnkd.in/eEPkExH | 25 comments on LinkedIn
Nicolas Figay
Relational Graph Transformers: A New Frontier in AI for Relational Data - Kumo
Relational Graph Transformers represent the next evolution in Relational Deep Learning, allowing AI systems to seamlessly navigate and learn from data spread across multiple tables. By treating relational databases as the rich, interconnected graphs they inherently are, these models eliminate the need for extensive feature engineering and complex data pipelines that have traditionally slowed AI adoption.
In this post, we'll explore how Relational Graph Transformers work, why they're uniquely suited for enterprise data challenges, and how they're already revolutionizing applications from customer analytics and recommendation systems to fraud detection and demand forecasting.
Dynamic Graph Memory in Mem0- when facts needs relationships
Dynamic Graph Memory in Mem0- when facts needs relationships
Graph Learning Will Lose Relevance Due To Poor Benchmarks
📣 Our spicy ICML 2025 position paper: “Graph Learning Will Lose Relevance Due To Poor Benchmarks”.
Graph learning is less trendy in the ML world than it was in 2020-2022. We believe the problem is in poor benchmarks that hold the field back - and suggest ways to fix it!
We identified three problems:
#️⃣ P1: No transformative real-world applications - while LLMs and geometric generative models become more powerful and solve complex tasks every generation (from reasoning to protein folding), how transformative could a GNN on Cora or OGB be?
P1 Remedies: The community is overlooking many significant and transformative applications, including chip design and broader ML for systems, combinatorial optimization, and relational data (as highlighted by RelBench). Each of them offers $billions in potential outcomes.
#️⃣ P2: While everything can be modeled as a graph, often it should not be. We made a simple experiment and probed a vanilla DeepSet w/o edges and a GNN on Cayley graphs (fixed edges for a certain number of nodes) on molecular datasets and the performance is quite competitive.
#️⃣ P3: Bad benchmarking culture (this one hits hard) - it’s a mess :)
Small datasets (don’t use Cora and MUTAG in 2025), no standard splits, and in many cases recent models are clearly worse than GCN / Sage from 2020. It gets worse when evaluating generative models.
Remedies for P3: We need more holistic benchmarks which are harder to game and saturate - while it’s a common problem for all ML fields, standard graph learning benchmarks are egregiously old and rather irrelevant for the scale of problems doable in 2025.
💡 As a result, it’s hard to build a true foundation model for graphs. Instead of training each model on each dataset, we suggest using GNNs / GTs as processors in the “encoder-processor-decoder” blueprint, train them at scale, and only tune graph-specific encoders/decoders.
For example, we pre-trained several models on PCQM4M-v2, COCO-SP, and MalNet Tiny, and fine-tuned them on PascalVOC, Peptides-struct, and Stargazers to find that graph transformers benefit from pre-training.
---
The project started around NeurIPS 2024 when Christopher Morris gathered us to discuss the peeve points of graph learning and how to continue to do impactful research in this area. I believe the outcomes appear promising, and we can re-imagine graph learning in 2025 and beyond!
Massive work with 12 authors (everybody actually contributed): Maya Bechler-Speicher, Ben Finkelshtein, Fabrizio Frasca, Luis Müller, Jan Tönshoff, Antoine Siraudin, Viktor Zaverkin, Michael Bronstein, Mathias Niepert, Bryan Perozzi, and Christopher Morris (Chris you should create a LinkedIn account finally ;)
Graph Learning Will Lose Relevance Due To Poor Benchmarks
What if your LLM is… a graph?
What if your LLM is… a graph?
A few days ago, Petar Veličković from Google DeepMind gave one of the most interesting and thought provoking conference I've seen in a while, "Large Language Models as Graph Neural Networks". Once you start seeing LLM as graph neural network, many structural oddities suddenly falls into place.
For instance, OpenAI currently recommends to put the instructions at the top of a long prompt. Why is that so? Because due to the geometry of attention graphs, LLM are counter-intuitively biased in favors of the first tokens: they travel constinously through each generation steps, are internally repeated a lot and end up "over-squashing" the latter ones. Models then use a variety of internal metrics/transforms like softmax to moderate this bias and better ponderate distribution, but this is a late patch that cannot solve long time attention deficiencies, even more so for long context.
The most interesting aspect of the conference from an applied perspective: graph/geometric representations directly affect accuracy and robustness. As the generated sequence grow and deal with sequences of complex reasoning steps, cannot build solid expert system when attention graphs have single point of failures. Or at least, without extrapolating this information in the first place and providing more detailed accuracy metrics.
I do believe LLM explainability research is largely underexploited right now, despite being accordingly a key component of LLM devops in big labs. If anything, this is literal "prompt engineering", seeing models as nearly physical structure under stress and providing the right feedback loops to make them more reliable. | 30 comments on LinkedIn
What if your LLM is… a graph?
Towards Mechanistic Interpretability of Graph Transformers via Attention Graphs
Our first attempts at mechanistic interpretability of Transformers from the perspective of network science and graph theory! Check out our preprint: arxiv.org/abs/2502.12352
A wonderful collaboration with superstar MPhil students Batu El, Deepro Choudhury, as well as Pietro Lio' as part of the Geometric Deep Learning class last year at University of Cambridge Department of Computer Science and Technology
We were motivated by Demis Hassabis calling AlphaFold and other AI systems for scientific discovery as ‘engineering artifacts’. We need new tools to interpret the underlying mechanisms and advance our scientific understanding. Graph Transformers are a good place to start.
The key ideas are:
- Attention across multi-heads and layers can be seen as a heterogenous, dynamically evolving graph.
- Attention graphs are complex systems represent information flow in Transformers.
- We can use network science to extract mechanistic insights from them!
More to come on the network science perspective to understanding LLMs next! | 13 comments on LinkedIn
The evolution of graph learning
LLMs as Graph Neural Networks | Petar Veličković @ GLOW
Join our slack and come to the next Graph Learning on Wednesdays (GLOW) session.https://sites.google.com/view/graph-learning-on-wedsOn March 26th, 2025, we h...
Unifying Text Semantics and Graph Structures for Temporal Text-attributed Graphs with Large Language Models
LLMs are taking Graph Neural Networks to the next level:
While we've been discussing LLMs for natural language, they're quietly changing how we represent…
Unifying Text Semantics and Graph Structures for Temporal Text-attributed Graphs with Large
Large-Scale Graph Neural Networks
GiGL: Large-Scale Graph Neural Networks at Snapchat
Recent advances in graph machine learning (ML) with the introduction of Graph Neural Networks (GNNs) have led to a widespread interest in applying these approaches to business applications at...
GiGL: Large-Scale Graph Neural Networks at Snapchat
KET-RAG: Turbocharging AI Agents with 10x Cheaper, Smarter Knowledge Retrieval
KET-RAG: Turbocharging AI Agents with 10x Cheaper, Smarter Knowledge Retrieval
This Multi-Granular Graph Framework uses PageRank and Keyword-Chunk Graph to have the Best Cost-Quality Tradeoff
﹌﹌﹌﹌﹌﹌﹌﹌﹌
》The Problem: Knowledge Graphs Are Expensive (and Clunky)
AI agents need context to answer complex questions—like connecting “COVID vaccines” to “myocarditis risks” across research papers. But today’s solutions face two nightmares:
✸ Cost: Building detailed knowledge graphs with LLMs can cost $33,000 for a 5GB legal case.
✸ Quality: Cheap methods (like KNN graphs) miss key relationships, leading to 32% worse answers.
☆ Imagine training an AI doctor that either bankrupts you or misdiagnoses patients. Ouch.
﹌﹌﹌﹌﹌﹌﹌﹌﹌
》The Fix: KET-RAG’s Two-Layer Brain
KET-RAG merges precision (knowledge graphs) and efficiency (keyword-text maps) into one system:
✸ Layer 1: Knowledge Graph Skeleton
☆ Uses PageRank to find core text chunks (like “vaccine side effects” in medical docs).
☆ Builds a sparse graph only on these chunks with LLMs—saving 80% of indexing costs.
✸ Layer 2: Keyword-Chunk Bipartite Graph
☆ Links keywords (e.g., “myocarditis”) to all related text snippets—no LLM needed.
☆ Acts as a “fast lane” for retrieving context without expensive entity extraction.
﹌﹌﹌﹌﹌﹌﹌﹌﹌
》Results: Beating Microsoft’s Graph-RAG with Pennies
On HotpotQA and MuSiQue benchmarks, KET-RAG:
✸ Retrieves 81.6% of critical info vs. Microsoft’s 74.6%—with 10x lower cost.
✸ Boosts answer accuracy (F1 score) by 32.4% while cutting indexing bills by 20%.
✸ Scales to terabytes of data without melting budgets.
☆ Think of it as a Tesla Model 3 outperforming a Lamborghini at 1/10th the price.
﹌﹌﹌﹌﹌﹌﹌﹌﹌
》Why AI Agents Need This
AI agents aren’t just chatbots—they’re problem solvers for medicine, law, and customer service. KET-RAG gives them:
✸ Real-time, multi-hop reasoning: Connecting “drug A → gene B → side effect C” in milliseconds.
✸ Cost-effective scalability: Deploying agents across millions of documents without going broke.
✸ Adaptability: Mixing precise knowledge graphs (for critical data) with keyword maps (for speed).
Paper in comments
≣≣≣≣≣≣≣≣≣≣≣≣≣≣≣≣≣≣≣≣≣≣≣≣≣≣
》Build Your Own Supercharged AI Agent?
🔮 Join My 𝐇𝐚𝐧𝐝𝐬-𝐎𝐧 𝐀𝐈 𝐀𝐠𝐞𝐧𝐭𝐬 𝐓𝐫𝐚𝐢𝐧𝐢𝐧𝐠 TODAY!
and Learn Building AI Agent with Langgraph/Langchain, CrewAI and OpenAI Swarm + RAG Pipelines
𝐄𝐧𝐫𝐨𝐥𝐥 𝐍𝐎𝐖 [34% discount]:
👉 https://lnkd.in/eGuWr4CH | 10 comments on LinkedIn
KET-RAG: Turbocharging AI Agents with 10x Cheaper, Smarter Knowledge Retrieval
Adaptive Graph of Thoughts (AGoT), a test-time framework that replaces rigid prompting strategies (like Chain/Tree of Thought) with dynamic directed acyclic graphs
Dynamic Reasoning Graphs + LLMs = 🤝
Large Language Models (LLMs) often stumble on complex tasks when confined to linear reasoning.
What if they could dynamically restructure their thought process like humans?
A new paper introduces Adaptive Graph of Thoughts (AGoT), a test-time framework that replaces rigid prompting strategies (like Chain/Tree of Thought) with dynamic directed acyclic graphs (DAGs).
Instead of forcing fixed reasoning steps, AGoT recursively decomposes problems into sub-tasks, selectively expanding only the most critical pathways.
This is crucial for industries like scientific research or legal analysis, where problems demand non-linear, nested reasoning.
The key innovation lies in complexity checks: AGoT assesses each reasoning node, spawning sub-graphs for intricate subtasks while resolving simpler ones directly.
This mirrors how experts allocate mental effort—drilling into uncertainties while streamlining obvious steps.
The framework achieved 46.2% improvement on GPQA (a notoriously hard science QA benchmark), rivaling gains from compute-heavy fine-tuning.
By unifying chain, tree, and graph paradigms, AGoT retains CoT’s clarity, ToT’s exploration, and GoT’s flexibility without manual tuning.
The result? LLMs that self-adapt their reasoning depth based on problem complexity—no architectural changes needed.
For AI practitioners, AGoT’s DAG structure offers a principled interface to scale reasoning modularly.
↓
𝐖𝐚𝐧𝐧𝐚 𝐤𝐧𝐨𝐰 𝐰𝐡𝐚𝐭 𝐲𝐨𝐮 𝐦𝐢𝐬𝐬𝐞𝐝? Join my newsletter with 50k+ readers that breaks down all you need to know about the latest LLM research: llmwatch.com 💡
Adaptive Graph of Thoughts (AGoT), a test-time framework that replaces rigid prompting strategies (like Chain/Tree of Thought) with dynamic directed acyclic graphs
GFM-RAG: The First Graph Foundation Model for Retrieval-Augmented Generation
🚀 Introducing GFM-RAG: The First Graph Foundation Model for Retrieval-Augmented Generation!
We’re excited to share our latest research: GFM-RAG: Graph… | 20 comments on LinkedIn
GFM-RAG: The First Graph Foundation Model for Retrieval-Augmented Generation
Dynamic Reasoning Graphs + LLMs = 🤝 Adaptive graph of thoughts
Dynamic Reasoning Graphs + LLMs = 🤝
Large Language Models (LLMs) often stumble on complex tasks when confined to linear reasoning.
What if they could… | 10 comments on LinkedIn
Dynamic Reasoning Graphs + LLMs = 🤝
feedforward graphs (i.e. graphs w/o back edges)
And so we set out to understand _feedforward_ graphs (i.e. graphs w/o back edges) ⏩
Turns out these graphs are rather understudied for how often they are…
feedforward_ graphs (i.e. graphs w/o back edges)
The Evolution of Intelligent Recommendations with Agentic Graph Systems
The Evolution of Intelligent Recommendations with Agentic Graph Systems ➿ Agentic graph systems for recommendation represent a sophisticated fusion of…
The Evolution of Intelligent Recommendations with Agentic Graph Systems
GraphAgent — An innovative AI agent that efficiently integrates structured and unstructured data
🚀 Excited to Share Our Recent Work! 🌟 GraphAgent — An innovative AI agent that efficiently integrates structured and unstructured data! 📚 👉 Paper link:…
GraphAgent — An innovative AI agent that efficiently integrates structured and unstructured data
Predictions for Knowledge Graphs in 2025
Here are My Predictions for Knowledge Graphs in 2025! 🔵 GraphRAG via Ontologies: A range of GraphRAG frameworks and products will emerge, offering…
Predictions for Knowledge Graphs in 2025
Introduction to Graph Neural Networks
Want to catch up on Graph Neural Networks? Now's the time! Graph Neural Networks (GNNs) have become a popular solution for problems that include network data,…
Graph Neural Networks
Context-based Graph Neural Network
❓How Can Graph Neural Networks Enhance Recommendation Systems by Incorporating Contextual Information? Traditional recommendation systems often leverage a…
Context-based Graph Neural Network