Exploring Network-Knowledge Graph Duality: A Case Study in Agentic Supply Chain Risk Analysis
Exploring Network-Knowledge Graph Duality: A Case Study in
Agentic Supply Chain Risk Analysis ...
What happens when you ask an AI about supply chain vulnerabilities and it misses the most critical dependencies?
Most AI systems treat business relationships like isolated facts in a database. They might know Apple uses lithium batteries, but they miss the web of connections that create real risk.
👉 The Core Problem
Standard AI retrieval treats every piece of information as a standalone point. But supply chain risk lives in the relationships between companies, products, and locations. When conflict minerals from the DRC affect smartphone production, it's not just about one supplier - it's about cascading effects through interconnected networks.
Vector similarity search finds related documents but ignores the structural dependencies that matter most for risk assessment.
👉 A Different Approach
New research from UC Berkeley and MSCI demonstrates how to solve this by treating supply chains as both networks and knowledge graphs simultaneously.
The key insight: economic relationships like "Company A produces Product B" are both structural network links and semantic knowledge graph triples. This duality lets you use network science to find the most economically important paths.
👉 How It Works
Instead of searching for similar text, the system:
- Maps supply chains as networks with companies, products, and locations as nodes
- Uses centrality measures to identify structurally important paths
- Wraps quantitative data in descriptive language so AI can reason about what numbers actually mean
- Retrieves specific relationship paths rather than generic similar content
When asked about cobalt risks, it doesn't just find articles about cobalt. It traces the actual path from DRC mines through battery manufacturers to final products, revealing hidden dependencies.
The system generates risk narratives that connect operational disruptions to financial impacts without requiring specialized training or expensive graph databases.
This approach shows how understanding the structure of business relationships - not just their content - can make AI genuinely useful for complex domain problems.
Exploring Network-Knowledge Graph Duality: A Case Study inAgentic Supply Chain Risk Analysis
Protocols move bits. Semantics move value.
The reports on agents are starting to sound samey: go vertical not horizontal; redesign workflows end-to-end; clean your data; stop doing pilots that automate inefficiencies; price for outcomes when the agent does the work.
All true. All necessary. All needing repetition ad nauseam.
So it’s refreshing to see a switch-up in Bain’s Technology Report 2025: the real leverage now sits with semantics. A shared layer of meaning.
Bain notes that protocols are maturing. MCP and A2A let agents pass tool calls, tokens, and results between layers. Useful plumbing. But there’s still no shared vocabulary that says what an invoice, policy, or work order is, how it moves through states, and how it maps to APIs, tables, and approvals. Without that, cross-vendor reliability will keep stalling.
They go further: whoever lands a pragmatic semantic layer first gets winner-takes-most network effects. Define the dictionary and you steer the value flow. This isn’t just a feature. It’s a control point.
Bain frames the stack clearly:
- Systems of record (data, rules, compliance)
- Agent operating systems (orchestration, planning, memory)
- Outcome interfaces (natural language requests, user-facing actions)
The bottleneck is semantics.
And there’s a pricing twist. If agents do the work, semantics define what “done” means. That unlocks outcome-based pricing, charging for tasks completed or value delivered, not log-ons.
Bain is blunt: the open, any-to-any agent utopia will smash against vendor incentives, messy data, IP, and security. Translation: walled gardens lead first. Start where governance is clear and data is good enough, then use that traction to shape the semantics others will later adopt.
This is where I’m seeing convergence. In practice, a knowledge graph can provide that shared meaning, identity, relationships, and policy. One workable pattern: the agent plans with an LLM, resolves entities and checks rules in the graph, then acts through typed APIs, writing back as events the graph can audit.
That’s the missing vocabulary and the enforcement that protocols alone can’t cover.
Tony Seale puts it well: “Neural and symbolic systems are not rivals; they are complements… a knowledge graph provides the symbolic backbone… to ground AI in shared semantics and enforce consistency.”
To me, this is optimistic, because it moves the conversation from “make the model smarter” to “make the system understandable.”
Agents don’t need perfection if they are predictable, composable, and auditable. Semantics deliver that.
It’s also how smaller players compete with hyperscalers: you don’t need to win the model race to win the meaning race.
With semantics, agents become infrastructure.
The next few years won’t be won by who builds the biggest model.
It’ll be won by who defines the smallest shared meaning. | 27 comments on LinkedIn
Announcing the formation of a Data Façades W3C Community Group
I am excited to announce the formation of a Data Façades W3C Community Group.
Façade-X, initially introduced at SEMANTICS 2021 and successfully implemented by the SPARQL Anything project, provides a simple yet powerful, homogeneous view over diverse and heterogeneous data sources (e.g., CSV, JSON, XML, and many others). With the recent v1.0.0 release of SPARQL Anything, the time was right to work on the long-term stability and widespread adoption of this approach by developing an open, vendor-neutral technology.
The Façade-X concept was born to allow SPARQL users to query data in any structured format in plain SPARQL. Therefore, the choice of a W3C community group to lead efforts on specifications is just natural. Specifications will enhance its reliability, foster innovation, and encourage various vendors and projects—including graph database developers — to provide their own compatible implementations.
The primary goals of the Data Façades Community Group is to:
Define the core specification of the Façade-X method.
Define Standard Mappings: Formalize the required mappings and profiles for connecting Façade-X to common data formats.
Define the specification of the query dialect: Provide a reference for the SPARQL dialect, configuration conventions (like SERVICE IRIs), and the functions/magic properties used.
Establish Governance: Create a monitored, robust process for adding support for new data formats.
Foster Collaboration: Build connections with relevant W3C groups (e.g., RDF & SPARQL, Data Shapes) and encourage involvement from developers, businesses, and adopters.
Join us!
With Luigi Asprino Ivo Velitchkov Justin Dowdy Paul Mulholland Andy Seaborne Ryan Shaw ...
CG: https://lnkd.in/eSxuqsvn
Github: https://lnkd.in/dkHGT8N3
SPARQL Anything #RDF #SPARQL #W3C #FX
announce the formation of a Data Façades W3C Community Group
Today, I'd like to introduce the GitLab Knowledge Graph. This release includes a code indexing engine, written in Rust, that turns your codebase into a live, embeddable graph database for LLM RAG. You can install it with a simple one-line script, parse local repositories directly in your editor, and connect via MCP to query your workspace and over 50,000 files in under 100 milliseconds.
We also saw GKG agents scoring up to 10% higher on the SWE-Bench-lite benchmarks, with just a few tools and a small prompt added to opencode (an open-source coding agent). On average, we observed a 7% accuracy gain across our eval runs, and GKG agents were able to solve new tasks compared to the baseline agents. You can read more from the team's research here https://lnkd.in/egiXXsaE.
This release is just the first step: we aim for this local version to serve as the backbone of a Knowledge Graph service that enables you to query the entire GitLab Software Development Life Cycle—from an Issue down to a single line of code.
I am incredibly proud of the work the team has done. Thank you, Michael U., Jean-Gabriel Doyon, Bohdan Parkhomchuk, Dmitry Gruzd, Omar Qunsul, and Jonathan Shobrook. You can watch Bill Staples and I present this and more in the GitLab 18.4 release here: https://lnkd.in/epvjrhqB
Try today at: https://lnkd.in/eAypneFA
Roadmap: https://lnkd.in/eXNYQkEn
Watch more below for a complete, in-depth tutorial on what we've built: | 19 comments on LinkedIn
GraphSearch: An Agentic Deep‑Search Workflow for Graph Retrieval‑Augmented Generation
GraphSearch: An Agentic Deep‑Search Workflow for Graph Retrieval‑Augmented Generation ...
Why Current AI Search Falls Short When You Need Real Answers
What happens when you ask an AI system a complex question that requires connecting multiple pieces of information? Most current approaches retrieve some relevant documents, generate an answer, and call it done. But this single-pass strategy often misses critical evidence.
👉 The Problem with Shallow Retrieval
Traditional retrieval-augmented generation (RAG) systems work like a student who only skims the first few search results before writing an essay. They grab what seems relevant on the surface but miss deeper connections that would lead to better answers.
When researchers tested these systems on complex multi-hop questions, they found a consistent pattern: the AI would confidently provide answers based on incomplete evidence, leading to logical gaps and missing key facts.
👉 A New Approach: Deep Searching with Dual Channels
Researchers from IDEA Research and Hong Kong University of Science and Technology developed GraphSearch, which works more like a thorough investigator than a quick searcher.
The system breaks down complex questions into smaller, manageable pieces, then searches through both text documents and structured knowledge graphs. Think of it as having two different research assistants: one excellent at finding descriptive information in documents, another skilled at tracing relationships between entities.
👉 How It Actually Works
Instead of one search-and-answer cycle, GraphSearch uses six coordinated modules:
Query decomposition splits complex questions into atomic sub-questions
Context refinement filters out noise from retrieved information
Query grounding fills in missing details from previous searches
Logic drafting organizes evidence into coherent reasoning chains
Evidence verification checks if the reasoning holds up
Query expansion generates new searches to fill identified gaps
The system continues this process until it has sufficient evidence to provide a well-grounded answer.
👉 Real Performance Gains
Testing across six different question-answering benchmarks showed consistent improvements. On the MuSiQue dataset, for example, answer accuracy jumped from 35% to 51% when GraphSearch was integrated with existing graph-based systems.
The approach works particularly well under constrained conditions - when you have limited computational resources for retrieval, the iterative searching strategy maintains performance better than single-pass methods.
This research points toward more reliable AI systems that can handle the kind of complex reasoning we actually need in practice.
Paper: "GraphSearch: An Agentic Deep Searching Workflow for Graph Retrieval-Augmented Generation" by Yang et al.
GraphSearch: An Agentic Deep‑Search Workflow for Graph Retrieval‑Augmented Generation
Product management makes or breaks AI. The role of graph
Product management makes or breaks AI. That includes 𝐝𝐚𝐭𝐚.
The role of 𝐝𝐚𝐭𝐚 𝐏𝐌 is shifting.
For years, the focus was BI - dashboards, reports, warehouses.
But AI demands more: context, retrieval, real time, and integration into the flow of work.
Data PMs who understand AI requirements will define the next generation of enterprise success.
Here’s how my team thinks about BI-ready vs AI-ready data 👇
Cognee - Graph-Aware Embeddings by cognee: For Even Smarter Retrieval
Cognee introduces graph-aware embeddings: graph signals boost semantic search for faster and more precise retrievals in paid plans. Learn more and book a call.
ODKE+: Ontology-Guided Open-Domain Knowledge Extraction with LLMs
Knowledge graphs (KGs) are foundational to many AI applications, but maintaining their freshness and completeness remains costly. We present ODKE+, a production-grade system that automatically...
Snowflake Unites Industry Leaders to Unlock AI's Potential with the Open Semantic Interchange
So I am worried.
https://lnkd.in/gfpkjUNZ
A semantic exchange format in YAML?
Because there is nothing to build on already?
https://lnkd.in/gB-iEeXn
:(
Snowflake Unites Industry Leaders to Unlock AI's Potential with the Open Semantic Interchange
The $1 trillion AI problem: Why Snowflake, Tableau and BlackRock are giving away their data secrets
Snowflake, Salesforce, dbt Labs and major tech companies launch Open Semantic Interchange (OSI), an unprecedented open-source initiative to standardize business data definitions and accelerate enterprise AI adoption.
WebKnoGraph: THE FIRST FULLY transparent, AI-driven framework for improving SEO and site navigation through reproducible methods.
I had my first presentation at SEO Wiesn (SEO conference at SIXT Munich) yesterday and WOW, what an experience it has been!
This is not a sales pitch, nor a product demo: we're talking about an open source project that is rooted in science, yet applicable in practical industry scenarios as already tested. No APIs, no vendor lock-in, no tricks.
It's our duty as SEOs to produce NEW INSIGHTS, not just rewrite stuff, digest information or promote ourselves.
Big thanks goes to our sponsors WordLift and Kalicube for supporting this research and believing in me and my team to deliver WebKnoGraph: THE FIRST FULLY transparent, AI-driven framework for improving SEO and site navigation through reproducible methods.
We plan on deepening this research and iterating with additional industry & research partners. If you’d like to try this on your website, DM me.
Full project repo: https://lnkd.in/d-dvHiCc. A scientific paper will follow.
More pics and detailed experience retrospective with the amazing crew will be shared in the upcoming days too 💙💙💙
Until then, have a sneak peek at the deck.
SEO WIESN TO THE WIIIIIIN!
WebKnoGraph: THE FIRST FULLY transparent, AI-driven framework for improving SEO and site navigation through reproducible methods.
What is Entity Resolution: Techniques, Tools & Use Cases
Learn what Entity Resolution is, how it works, and why it’s essential for data quality. Explore methods, tools, and use cases in MDM, customer 360, fraud detection, and more.
LDBC to GDC: A Landmark Shift in the Graph World — GraphGeeks
The Linked Data Benchmark Council has, for over a decade, been the quiet force behind much of the progress in graph technology. Their mission is deceptively simple: to design, maintain, and promote standard benchmarks for graph data management systems. Read about the recent meeting and the announcem
ApeRAG: a production-ready RAG that combines Graph RAG, vector search, and full-text search
ApeRAG: a production-ready RAG that combines Graph RAG, vector search, and full-text search. Looks pretty cool. There are a lot of use cases where a "knowledge graph" would help a lot. I still think this is still one of the most powerful way to understand "connections" and "hierarchy" the easiest.
🔤 Github: https://lnkd.in/gdYuShgX | 18 comments on LinkedIn
ApeRAG: a production-ready RAG that combines Graph RAG, vector search, and full-text search
The field is evolving from Prompt Engineering, treating context as a single, static string, to Contextual Engineering, which views context as a dynamic system of structured components (instructions, tools, memory, knowledge) orchestrated to solve complex tasks. 🔎
Nearly all innovation is a response to the primary limitation of Transformer models: the quadratic (O(n2)) computational cost of the self-attention mechanism as the context length (n) increases.
All techniques for managing this challenge can be organized into three areas:
1. Context Generation & Retrieval (Sourcing Ingredients)
Advanced Reasoning: Chain-of-Thought (CoT), Tree-of-Thoughts (ToT).
External Knowledge: Advanced Retrieval-Augmented Generation (RAG) like GraphRAG, which uses knowledge graphs for more structured retrieval.
2. Context Processing (Cooking the Ingredients)
Refinement: Using the LLM to iterate and improve its own output (Self-Refine).
Architectural Changes: Exploring models beyond Transformers (e.g., Mamba) to escape the quadratic bottleneck.
3. Context Management (The Pantry System)
Memory: Creating stateful interactions using hierarchical memory systems (e.g., MemGPT) that manage information between the active context window and external storage.
Key Distinction: RAG is stateless I/O to the world; Memory is the agent's stateful internal history.
The most advanced applications integrate these pillars to create sophisticated agents, with an added layer of dynamic adaptation:
Tool-Integrated Reasoning: Empowering LLMs to use external tools (APIs, databases, code interpreters) to interact with the real world.
Multi-Agent Systems: Designing "organizations" of specialized LLM agents that communicate and collaborate to solve multi-faceted problems, mirroring the structure of human teams.
Adaptive Context Optimization: Leveraging Reinforcement Learning (RL) to dynamically optimize context selection and construction for specific environments and tasks, ensuring efficient and effective performance.
Contextual Engineering is the emerging science of building robust, scalable, and stateful applications by systematically managing the flow of information to and from an LLM. | 16 comments on LinkedIn
Semantic Graphs: Going beyond the semantic conflation of "metadata" and "graph" labels | LinkedIn
Semantic and graph technologies are gaining traction. Unfortunately, this rise often comes with an overload of buzzwords — and even semantic conflations across terms.
Help !! I just let the pre-order live! 😬
All of you who signed up for more information shall have received an e-mail with pre-order option now.
This is the most scariest sh*t I've done in a long time. Seeing pre-orders ticking in to something I've created---together with the most amazing guest authors---is a super weird feeling. THANK YOU! ❤️
It's been quite a process. From an idea planted in 2022, to now seeing the light at the end of the tunnel. I've spend hours and hours inside TeXworks, nerding around with LaTeX and Ti𝘬Z. Numerous moments at pubs, while waiting for someone, to edit, edit and edit. Taking vacation off work to isolate myself to write (so effective, btw!). Having a gold team of proofreaders, providing with super valuable feedback. Working with awesome SHACL practitioners to tell great SHACL stories to you! IT HAS BEEN GREAT FUN!
This week, I have been focusing on final touches (thank you Data Treehouse for letting me do this!!). Indexing like a hero.
Soon the words, bits and blobs will hit the printing press, and the first copies will ship on 𝐍𝐨𝐯𝐞𝐦𝐛𝐞𝐫 3𝐫𝐝.
If you want to pre-order, tag along to https://lnkd.in/dER72USX for more information. All pre-orders will get a tiny SHACL surprise inside their book. 😇
Btw: the final product will probably not look like this, I got good help from our mutual friend ChatGPT, but I know it will be yellow at least. 💛