AI Engineer World's Fair 2025: GraphRAG Track Spotlight
๐ฃ AI Engineer World's Fair 2025: GraphRAG Track Spotlight! ๐
So grateful to have hosted the GraphRAG Track at the Fair. The sessions were great, highlighting the depth and breadth of graph thinking for AI.
Shoutouts to...
- Mitesh Patel "HybridRAG" as a fusion of graph and vector retrieval designed to master complex data interpretation and specialized terminology for question answering
- Chin Keong Lam "Wisdom Discovery at Scale" using Knowledge Augmented Generation (KAG) in a multi agent system with n8n
- Sam Julien "When Vectors Break Down" carefully explaining how graph-based RAG architecture achieved a whopping 86.31% accuracy for dense enterprise knowledge
- Daniel Chalef "Stop Using RAG as Memory" explored temporally-aware knowledge graphs, built by the open-source Graphiti framework, to provide precise, context-rich memory for agents,
- Ola Mabadeje "Witness the power of Multi-Agent AI & Network Knowledge Graphs" showing dramatic improvements in ticket resolution efficiency and overall execution quality in network operations.
- Thomas Smoker "Beyond Documents"! casually mentioning scraping the entire internet to distill a knowledge graph focused with legal agents
- Mark Bain hosting an excellent Agentic Memory with Knowledge Graphs lunch&learn, with expansive thoughts and demos from Vasilije Markovic Daniel Chalef and Alexander Gilmore
Also, of course, huge congrats to Shawn swyx W and Benjamin Dunphy on an excellent conference. ๐ฉ
#graphrag Neo4j AI Engineer
AI Engineer World's Fair 2025: GraphRAG Track Spotlight
Want to Fix LLM Hallucination? Neurosymbolic Alone Wonโt Cut It
Want to Fix LLM Hallucination? Neurosymbolic Alone Wonโt Cut It
The Conversationโs new piece makes a clear case for neurosymbolic AIโintegrating symbolic logic with statistical learningโas the long-term fix for LLM hallucinations. Itโs a timely and necessary argument:
โNo matter how large a language model gets, it canโt escape its fundamental lack of grounding in rules, logic, or real-world structure. Hallucination isnโt a bug, itโs the default.โ
But whatโs crucialโand often glossed overโis that symbolic logic alone isnโt enough. The real leap comes from adding formal ontologies and semantic constraints that make meaning machine-computable. OWL, Shapes Constraint Language (SHACL), and frameworks like BFO, Descriptive Ontology for Linguistic and Cognitive Engineering (DOLCE), the Suggested Upper Merged Ontology (SUMO), and the Common Core Ontologies (CCO) donโt just โrepresent rulesโโthey define what exists, what can relate, and under what conditions inference is valid. Thatโs the difference between โdecoratingโ a knowledge graph and engineering one that can detect, explain, and prevent hallucinations in practice.
Iโd go further:
โข Most enterprise LLM hallucinations are just semantic errorsโmislabeling, misattribution, or class confusion that only formal ontologies can prevent.
โข Neurosymbolic systems only deliver if their symbolic half is grounded in ontological reality, not just handcrafted rules or taxonomies.
The upshot:
We need to move beyond mere integration of symbols and neurons. We need semantic scaffoldingโontologies as infrastructureโto ensure AI isnโt just fluent, but actually right.
Curious if others are layering formal ontologies (BFO, DOLCE, SUMO) into their AI stacks yet? Or are we still hoping that more compute and prompt engineering will do the trick?
#NeuroSymbolicAI #SemanticAI #Ontology #LLMs #AIHallucination #KnowledgeGraphs #AITrust #AIReasoning
Want to Fix LLM Hallucination? Neurosymbolic Alone Wonโt Cut It
Synalinks release 0.3 focuses on the Knowledge Graph layer
Your agents, multi-agent systems and LMs apps are still failing with basic logic? We got you covered.
Today we're excited to announce Synalinks 0.3 our Keras-based neuro-symbolic framework that bridges the gap between neural networks and symbolic reasoning.
Our latest release focuses entirely on the Knowledge Graph layer, delivering production-ready solutions for real-world applications:
- Fully constrained KG extraction powered by Pydantic: ensuring that relations connect to the correct entity types.
- Seamless integration with our Agents/Chain-of-Thought and Self-Critique modules.
- Automatic entity alignment with HSWN.
- KG extraction and retrieval optimizable with OPRO and RandomFewShot algorithms.
- 100% reliable Cypher query generation through logic-enhanced hybrid triplet retrieval (works with local models too!).
- We took extra care to avoid Cypher injection vulnerabilities (yes, we're looking at you, LangGraph ๐)
- The retriever don't need the graph schema, as it is included in the way we constrain the generation, avoiding context pollution (hence better accuracy).
- We also fixed Synalinks CLI for Windows users along with some minor bug fixes.
Our technology combine constrained structured output with in-context reinforcement learning, making enterprise-grade reasoning both highly efficient and cost-effective.
Currently supporting Neo4j with plans to expand to other graph databases. Built this initially for a client project, but the results were too good not to share with the community.
Want to add support for your preferred graph database? It's just one file to implement! Drop a comment and let's make it happen!
#AI #MachineLearning #KnowledgeGraphs #NeuralNetworks #Keras #Neo4j #AIAgents #TechInnovation #OpenSource
| 10 comments on LinkedIn
I'm happy to share the draft of the "Semantically Composable Architectures" mini-paper.
It is the culmination of approximately four years' work, which began with Coreless Architectures and has now evolved into something much bigger.
LLMs are impressive, but a real breakthrough will occur once we surpass the cognitive capabilities of a single human brain.
Enabling autonomous large-scale system reverse engineering and large-scale autonomous transformation with minimal to no human involvement, while still making it understandable to humans if they choose to, is a central pillar of making truly groundbreaking changes.
We hope the ideas we shared will be beneficial to humanity and advance our civilization further.
It is not final and will require some clarification and improvements, but the key concepts are present. Happy to hear your thoughts and feedback.
Some of these concepts underpin the design of the Product X system.
Part of the core team + external contribution:
Andrew Barsukov Andrey Kolodnitsky Sapta Girisa N Keith E. Glendon Gurpreet Sachdeva Saurav Chandra Mike Diachenko Oleh Sinkevych | 13 comments on LinkedIn
Want to explore the Anthropic Transformer-Circuit's as a queryable graph?
Want to explore the Anthropic Transformer-Circuit's as a queryable graph?
Wrote a script to import the graph json into Neo4j - code in Gist.
https://lnkd.in/eT4NjQgY
https://lnkd.in/e38TfQpF
Next step - write directly from the circuit-tracer library to the graph db.
https://lnkd.in/eVU_t6mS
Want to explore the Anthropic Transformer-Circuit's as a queryable graph?
Introducing FACT: Fast Augmented Context Tools (3.2x faster, 90% cost reduction vs RAG)
Introducing FACT: Fast Augmented Context Tools (3.2x faster, 90% cost reduction vs RAG)
RAG had its run, but itโs not built for agentic systems. Vectors are fuzzy, slow, and blind to context. They work fine for static data, but once you enter recursive, real-time workflows, where agents need to reason, act, and reflect. RAG collapses under its own ambiguity.
Thatโs why I built FACT: Fast Augmented Context Tools.
Traditional Approach:
User Query โ Database โ Processing โ Response (2-5 seconds)
FACT Approach:
User Query โ Intelligent Cache โ [If Miss] โ Optimized Processing โ Response (50ms)
It replaces vector search in RAG pipelines with a combination of intelligent prompt caching and deterministic tool execution via MCP. Instead of guessing which chunk is relevant, FACT explicitly retrieves structured data, SQL queries, live APIs, internal tools, then intelligently caches the result if itโs useful downstream.
The prompt caching isnโt just basic storage.
Itโs intelligent using the prompt cache from Anthropic and other LLM providers, tuned for feedback-driven loops: static elements get reused, transient ones expire, and the system adapts in real time. Some things you always want cached, schemas, domain prompts. Others, like live data, need freshness. Traditional RAG is particularly bad at this. Ask anyone force to frequently update vector DBs.
I'm also using Arcade.dev to handle secure, scalable execution across both local and cloud environments, giving FACT hybrid intelligence for complex pipelines and automatic tool selection.
If you're building serious agents, skip the embeddings. RAG is a workaround. FACT is a foundation. Itโs cheaper, faster, and designed for how agents actually work: with tools, memory, and intent.
To get started point your favorite coding agent at: https://lnkd.in/gek_akem | 38 comments on LinkedIn
Introducing FACT: Fast Augmented Context Tools (3.2x faster, 90% cost reduction vs RAG)
A-MEM Transforms AI Agent Memory with Zettelkasten Method, Atomic Notes, Dynamic Linking & Continuous Evolution
๐ฏ๐ A-MEM Transforms AI Agent Memory with Zettelkasten Method, Atomic Notes, Dynamic Linking & Continuous Evolution!
This Novel Memory fixes rigid structures with adaptable, evolving, and interconnected knowledge networks, delivering 2x performance in complex reasoning tasks.
๐ง๐ต๐ถ๐ ๐ถ๐ ๐๐ต๐ฎ๐ ๐ ๐น๐ฒ๐ฎ๐ฟ๐ป๐ฒ๐ฑ:
๏น๏น๏น๏น๏น๏น๏น๏น๏น
ใ ๐ช๐ต๐ ๐ง๐ฟ๐ฎ๐ฑ๐ถ๐๐ถ๐ผ๐ป๐ฎ๐น ๐ ๐ฒ๐บ๐ผ๐ฟ๐ ๐๐ฎ๐น๐น ๐ฆ๐ต๐ผ๐ฟ๐
Most AI agents today rely on simplistic storage and retrieval but break down when faced with complex, multi-step reasoning tasks.
โธ Common Limitations:
โ Fixed schemas: Conventional memory systems require predefined structures that limit flexibility.
โ Limited adaptability: When new information arises, old memories remain static and disconnected, reducing an agentโs ability to build on past experiences.
โ Ineffective long-term retention: AI agents often struggle to recall relevant past interactions, leading to redundant processing and inefficiencies.
๏น๏น๏น๏น๏น๏น๏น๏น๏น
ใ๐-๐ ๐๐ : ๐๐๐ผ๐บ๐ถ๐ฐ ๐ป๐ผ๐๐ฒ๐ ๐ฎ๐ป๐ฑ ๐๐๐ป๐ฎ๐บ๐ถ๐ฐ ๐น๐ถ๐ป๐ธ๐ถ๐ป๐ด
A-MEM organizes knowledge in a way that mirrors how humans create and refine ideas over time.
โธ How it Works:
โ Atomic notes: Information is broken down into small, self-contained knowledge units, ensuring clarity and easy integration with future knowledge.
โ Dynamic linking: Instead of relying on static categories, A-MEM automatically creates connections between related knowledge, forming a network of interrelated ideas.
๏น๏น๏น๏น๏น๏น๏น๏น๏น
ใ ๐ฃ๐ฟ๐ผ๐๐ฒ๐ป ๐ฃ๐ฒ๐ฟ๐ณ๐ผ๐ฟ๐บ๐ฎ๐ป๐ฐ๐ฒ ๐๐ฑ๐๐ฎ๐ป๐๐ฎ๐ด๐ฒ
A-MEM delivers measurable improvements.
โธ Empirical results demonstrate:
โ Over 2x performance improvement in complex reasoning tasks, where AI must synthesize multiple pieces of information across different timeframes.
โ Superior efficiency across top foundation models, including GPT, Llama, and Qwenโproving its versatility and broad applicability.
๏น๏น๏น๏น๏น๏น๏น๏น๏น
ใ ๐๐ป๐๐ถ๐ฑ๐ฒ ๐-๐ ๐๐
โธ Note Construction:
โ AI-generated structured notes that capture essential details and contextual insights.
โ Each memory is assigned metadata, including keywords and summaries, for faster retrieval.
โธ Link Generation:
โ The system autonomously connects new memories to relevant past knowledge.
โ Relationships between concepts emerge naturally, allowing AI to recognize patterns over time.
โธ Memory Evolution:
โ Older memories are continuously updated as new insights emerge.
โ The system dynamically refines knowledge structures, mimicking the way human memory strengthens connections over time.
โฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃ
โซธ๊ Want to build Real-World AI agents?
Join My ๐๐ฎ๐ป๐ฑ๐-๐ผ๐ป ๐๐ ๐๐ด๐ฒ๐ป๐ ๐ฐ-๐ถ๐ป-๐ญ ๐ง๐ฟ๐ฎ๐ถ๐ป๐ถ๐ป๐ด TODAY! ๐ฐ๐ด๐ฌ+ already Enrolled.
โ Build Real-World AI Agents for Healthcare, Finance,Smart Cities,Sales
โ Learn 4 Framework: LangGraph | PydanticAI | CrewAI | OpenAI Swarm
โ Work with Text, Audio, Video and Tabular Data
๐๐๐ป๐ฟ๐ผ๐น๐น ๐ก๐ข๐ช (๐ฐ๐ฑ% ๐ฑ๐ถ๐๐ฐ๐ผ๐๐ป๐):
https://lnkd.in/eGuWr4CH
| 27 comments on LinkedIn
A-MEM Transforms AI Agent Memory with Zettelkasten Method, Atomic Notes, Dynamic Linking & Continuous Evolution
RAG vs Graph RAG, explained visually.
(it's a popular LLM interview question)
Imagine you have a long document, say a biography, about an individual (X) who has accomplished several things in this life.
โณ Chapter 1: Talks about Accomplishment-1.
โณ Chapter 2: Talks about Accomplishment-2.
...
โณ Chapter 10: Talks about Accomplishment-10.
Summarizing all these accomplishments via RAG might never be possible since...
...it must require the entire context...
...but one might only be fetching the top-k relevant chunks from the vector db.
Moreover, since traditional RAG systems retrieve each chunk independently, this can often leave the LLM to infer the connections between them (provided the chunks are retrieved).
Graph RAG solves this.
The idea is to first create a graph (entities & relationships) from the documents and then do traversal over that graph during the retrieval phase.
See how Graph RAG solves the above problems.
- First, a system (typically an LLM) will create the graph by understanding the biography.
- This will produce a full graph of nodes entities & relationships, and a subgraph will look like this:
โณ X โ โ Accomplishment-1.
โณ X โ โ Accomplishment-2.
...
โณ X โ โ Accomplishment-N.
When summarizing these accomplishments, the retrieval phase can do a graph traversal to fetch all the relevant context related to X's accomplishments.
This context, when passed to the LLM, will produce a more coherent and complete answer as opposed to traditional RAG.
Another reason why Graph RAG systems are so effective is because LLMs are inherently adept at reasoning with structured data.
Graph RAG instills that structure into them with their retrieval mechanism.
๐ Over to you: What are some other issues with traditional RAG systems that Graph RAG solves?
____
Find me โย Avi Chawla
Every day, I share tutorials and insights on DS, ML, LLMs, and RAGs. | 24 comments on LinkedIn
Graph RAG open source stack to generate and visualize knowledge graphs
A serious knowledge graph effort is much more than a bit of Github, but customers and adventurous minds keep asking me if there is an easy to use (read: POC click-and-go solution) graph RAG open source stack they can use to generate knowledge graphs.
So, here is my list of projects I keep an eye on. Mind, there is nothing simple if you venture into graphs, despite all the claims and marketing. Things like graph machine learning, graph layout and distributed graph analytics is more than a bit of pip install.
The best solutions are hidden inside multi-nationals, custom made. Equity firms and investors sometimes ask me to evaluate innovations. It's amazing what talented people develop and never shows up in the news, or on Github.
TrustGraph - The Knowledge Platform for AI https://trustgraph.ai/ The only one with a distributed architecture and made for enterprise KG.
itext2kg - https://lnkd.in/e-eQbwV5 Clean and plain. Wrapped prompts done right.
Fast GraphRAG - https://lnkd.in/e7jZ9GZH Popular and with some basic visualization.
ZEP - https://lnkd.in/epxtKtCU Geared towards agentic memory.
Triplex - https://lnkd.in/eGV8FR56 LLM to extract triples.
GraphRAG Local with UI - https://lnkd.in/ePGeqqQE Another starting point for small KG efforts. Or to convince your investors.
GraphRAG visualizer - https://lnkd.in/ePuMmfkR Makes pretty pictures but not for drill-downs.
Neo4j's GraphRAG - https://lnkd.in/ex_A52RU A python package with a focus on getting data into Neo4j.
OpenSPG - https://lnkd.in/er4qUFJv Has a different take and more academic.
Microsoft GraphRAG - https://lnkd.in/e_a-mPum A classic but I don't think anyone is using this beyond experimentation.
yWorks - https://www.yworks.com If you are serious about interactive graph layout.
Ogma - https://lnkd.in/evwnJCBK If you are serious about graph data viz.
Orbifold Consulting - https://lnkd.in/e-Dqg4Zx If you are serious about your KG journey.
#GraphRAG #GraphViz #GraphMachineLearning #KnowledgeGraphs
graph RAG open source stack they can use to generate knowledge graphs.
LLMs generate possibilities; knowledge graphs remember what works
LLMs generate possibilities; knowledge graphs remember what works. Together, they forge the recursive memory and creative engine that enables AI systems to truly evolve themselves.
Combining neural components (like large language models) with symbolic verification creates a powerful framework for self-evolution that overcomes limitations of either approach used independently.
AlphaEvolve demonstrates that self-evolving systems face a fundamental tension between generating novel solutions and ensuring those solutions actually work.
The paper shows how AlphaEvolve addresses this through a hybrid architecture where:
Neural components (LLMs) provide creative generation of code modifications by drawing on patterns learned from vast training data
Symbolic components (code execution) provide ground truth verification through deterministic evaluation
Without this combination, a system would either generate interesting but incorrect solutions (neural-only approach) or be limited to small, safe modifications within known patterns (symbolic-only approach).
The system can operate at multiple levels of abstraction depending on the problem: raw solution evolution, constructor function evolution, search algorithm evolution, or co-evolution of intermediate solutions and search algorithms.
This capability emanates directly from the neurosymbolic integration, where:
Neural networks excel at working with continuous, high-dimensional spaces and recognizing patterns across abstraction levels
Symbolic systems provide precise representations of discrete structures and logical relationships
This enables AlphaEvolve to modify everything from specific lines of code to entire algorithmic approaches.
While AlphaEvolve currently uses an evolutionary database, a knowledge graph structure could significantly enhance self-evolution by:
Capturing evolutionary relationships between solutions
Identifying patterns of code changes that consistently lead to improvements
Representing semantic connections between different solution approaches
Supporting transfer learning across problem domains
Automated, objective evaluation is the core foundation enabling self-evolution:
The main limitation of AlphaEvolve is that it handles problems for which it is possible to devise an automated evaluator.
This evaluation component provides the "ground truth" feedback that guides evolution, allowing the system to:
Differentiate between successful and unsuccessful modifications
Create selection pressure toward better-performing solutions
Avoid hallucinations or non-functional solutions that might emerge from neural components alone.
When applied to optimize Gemini's training kernels, the system essentially improved the very LLM technology that powers it. | 12 comments on LinkedIn
LLMs generate possibilities; knowledge graphs remember what works
๐๐๐ค๐ช๐๐๐ฉ ๐๐ค๐ง ๐ฉ๐๐ ๐ฟ๐๐ฎ: What if we could encapsulate everything a person knowsโtheir entire bubble of knowledge, what Iโd call a Personal Knowledge Domain or better, our ๐๐๐ข๐๐ฃ๐ฉ๐๐ ๐๐๐ก๐, and represent it in an RDF graph? From that foundation, we could create Personal Agents that act on our behalf. Each of us would own our agent, with the ability to share or lease it for collaboration with other agents.
If we could make these agents secure, continuously updatable, and interoperable, what kind of power might we unlock for the human race?
Is this idea so far-fetched? It has solid grounding in knowledge representation, identity theory, and agent-based systems. It fits right in with current trends: AI assistants, the semantic web, Web3 identity, and digital twins. Yes, the technical and ethical hurdles are significant, but this could become the backbone of a future architecture for personalized AI and cooperative knowledge ecosystems.
Pieces of the puzzle already exist: Tim Berners-Leeโs Solid Project, digital twins for individuals, Personal AI platforms like personal.ai, Retrieval-Augmented Language Model agents (ReALM), and Web3 identity efforts such as SpruceID, architectures such as MCP and inter-agent protocols such as A2A. We see movement in human-centric knowledge graphs like FOAF and SIOC, learning analytics, personal learning environments, and LLM-graph hybrids.
What we still need is a unified architecture that:
* Employs RDF or similar for semantic richness
* Ensures user ownership and true portability
* Enables secure agent-to-agent collaboration
* Supports continuous updates and trust mechanisms
* Integrates with LLMs for natural, contextual reasoning
These are certainly not novel notions, for example:
* MyPDDL (My Personal Digital Life) and the PDS (Personal Data Store) concept from MIT and the EUโs DECODE project.
* The Human-Centric AI Group at Stanford and the Augmented Social Cognition group at PARC have also published research around lifelong personal agents and social memory systems.
However, one wonders if anyone is working on combining all of the ingredients into a fully baked cake - after which we can enjoy dessert while our personal agents do our bidding. | 21 comments on LinkedIn
NodeRAG restructures knowledge into a heterograph: a rich, layered, musical graph where each node plays a different role
NodeRAG restructures knowledge into a heterograph: a rich, layered, musical graph where each node plays a different role.
Itโs not just smarter retrieval. Itโs structured memory for AI agents.
ใ Why NodeRAG?
Most Retrieval-Augmented Generation (RAG) methods retrieve chunks of text. Good enough โ until you need reasoning, precision, and multi-hop understanding.
This is how NodeRAG solves these problems:
ใ ๐นStep 1: Graph Decomposition
NodeRAG begins by decomposing raw text into smart building blocks:
โธ Semantic Units (S): Little event nuggets ("Hinton won the Nobel Prize.")
โธ Entities (N): Key names or concepts ("Hinton", "Nobel Prize")
โธ Relationships (R): Links between entities ("awarded to")
โฉ This is like teaching your AI to recognize the actors, actions, and scenes inside any document.
ใ ๐นStep 2: Graph Augmentation
Decomposition alone isn't enough. NodeRAG augments the graph by identifying important hubs:
โธ Node Importance: Using K-Core and Betweenness Centrality to find critical nodes
โฉ Important entities get special attention โ their attributes are summarized into new nodes (A).
โธ Community Detection: Grouping related nodes into communities and summarizing them into high-level insights (H).
โฉ Each community gets a "headline" overview node (O) for quick retrieval.
It's like adding context and intuition to raw facts.
ใ ๐น Step 3: Graph Enrichment
Knowledge without detail is brittle. So NodeRAG enriches the graph:
โธ Original Text: Full chunks are linked back into the graph (Text nodes, T)
โธ Semantic Edges: Using HNSW for fast, meaningful similarity connections
โฉ Only smart nodes are embedded (not everything!) โ saving huge storage space.
โฉ Dual search (exact + vector) makes retrieval laser-sharp.
Itโs like turning a 2D map into a 3D living world.
ใ ๐น Step 4: Graph Searching
Now comes the magic.
โธ Dual Search: First find strong entry points (by name or by meaning)
โธ Shallow Personalized PageRank (PPR): Expand carefully from entry points to nearby relevant nodes.
โฉ No wandering into irrelevant parts of the graph. The search is surgical.
โฉ Retrieval includes fine-grained semantic units, attributes, high-level elements โ everything you need, nothing you don't.
Itโs like sending out agents into a city โ and they return not with everything they saw, but exactly what you asked for, summarized and structured.
ใ Results: NodeRAG's Performance
Compared to GraphRAG, LightRAG, NaiveRAG, and HyDE โ NodeRAG wins across every major domain: Tech, Science, Writing, Recreation, and Finance.
NodeRAG isnโt just a better graph. NodeRAG is a new operating system for memory.
โฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃโฃ
โซธ๊ Want to build Real-World AI agents?
Join My ๐๐ฎ๐ป๐ฑ๐-๐ผ๐ป ๐๐ ๐๐ด๐ฒ๐ป๐ ๐ง๐ฟ๐ฎ๐ถ๐ป๐ถ๐ป๐ด TODAY!
โ Build Real-World AI Agents + RAG Pipelines
โ Learn 3 Tools: LangGraph/LangChain | CrewAI | OpenAI Swarm
โ Work with Text, Audio, Video and Tabular Data
๐๐๐ป๐ฟ๐ผ๐น๐น ๐ก๐ข๐ช (๐ฏ๐ฐ% ๐ฑ๐ถ๐๐ฐ๐ผ๐๐ป๐):
https://lnkd.in/eGuWr4CH
| 20 comments on LinkedIn
NodeRAG restructures knowledge into a heterograph: a rich, layered, musical graph where each node plays a different role
Announcing general availability of Amazon Bedrock Knowledge Bases GraphRAG with Amazon Neptune Analytics | Amazon Web Services
Today, Amazon Web Services (AWS) announced the general availability of Amazon Bedrock Knowledge Bases GraphRAG (GraphRAG), a capability in Amazon Bedrock Knowledge Bases that enhances Retrieval-Augmented Generation (RAG) with graph data in Amazon Neptune Analytics. In this post, we discuss the benefits of GraphRAG and how to get started with it in Amazon Bedrock Knowledge Bases.
On the different roles of ontologies (& machine learning) | LinkedIn
In a previous post I was touching on how ontologies are foundational to many data activities, yet "obscure". As a consequence, the different roles of ontologies are not always known among people that make use of them, as they may focus only on some of the aspects relevant for specific use cases.
Lessons Learned from Evaluating NodeRAG vs Other RAG Systems
๐ Lessons Learned from Evaluating NodeRAG vs Other RAG Systems
I recently dug into the NodeRAG paper (https://lnkd.in/gwaJHP94) and it was eye-opening not just for how it performed, but for what it revealed about the evolution of RAG (Retrieval-Augmented Generation) systems.
Some key takeaways for me:
๐ NaiveRAG is stronger than you think.
Brute-force retrieval using simple vector search sometimes beats graph-based methods, especially when graph structures are too coarse or noisy.
๐ GraphRAG was an important step, but not the final answer.
While it introduced knowledge graphs and community-based retrieval, GraphRAG sometimes underperformed NaiveRAG because its communities could be too coarse, leading to irrelevant retrieval.
๐ LightRAG reduced token cost, but at the expense of accuracy.
By focusing on retrieving just 1-hop neighbors instead of traversing globally, LightRAG made retrieval cheaper โ but often missed important multi-hop reasoning paths, losing precision.
๐ NodeRAG shows what mature RAG looks like.
NodeRAG redesigned the graph structure itself:
Instead of homogeneous graphs, it uses heterogeneous graphs with fine-grained semantic units, entities, relationships, and high-level summaries โ all as nodes.
It combines dual search (exact match + semantic search) and shallow Personalized PageRank to precisely retrieve the most relevant context.
The result?
๐ Highest accuracy across multi-hop and open-ended benchmarks
๐ Lowest token retrieval (i.e., lower inference costs)
๐ Faster indexing and querying
๐ง Key takeaway:
In the RAG world, itโs no longer about retrieving more โ itโs about retrieving better.
Fine-grained, explainable, efficient retrieval will define the next generation of RAG systems.
If youโre working on RAG architectures, NodeRAGโs design principles are well worth studying!
Would love to hear how others are thinking about the future of RAG systems. ๐๐
#RAG #KnowledgeGraphs #AI #LLM #NodeRAG #GraphRAG #LightRAG #MachineLearning #GenAI #KnowledegGraphs