metaphacts unveils metis, the new Knowledge-driven AI platform for Enterprises
Introducing metis: an enterprise AI platform from metaphactory. Get trusted, context-aware, knowledge-driven AI for actionable insights & intelligent agents.
Building AI Agents with LLMs, RAG, and Knowledge Graphs: A practical guide to autonomous and modern AI agents
𝐁𝐨𝐨𝐤 𝐩𝐫𝐨𝐦𝐨𝐭𝐢𝐨𝐧 𝐛𝐞𝐜𝐚𝐮𝐬𝐞 𝐭𝐡𝐢𝐬 𝐨𝐧𝐞 𝐢𝐬 𝐰𝐨𝐫𝐭𝐡 𝐢𝐭.. 𝐀𝐠𝐞𝐧𝐭𝐢𝐜 𝐀𝐈 𝐚𝐭 𝐢𝐭𝐬 𝐛𝐞𝐬𝐭..
This masterpiece was published by Salvatore Raieli and Gabriele Iuculano, and it is available for orders from today, and it's already a 𝐁𝐞𝐬𝐭𝐬𝐞𝐥𝐥𝐞𝐫!
While many resources focus on LLMs or basic agentic workflows, what makes this book stand out is its deep dive into grounding LLMs with real-world data and action through the powerful combination of 𝘙𝘦𝘵𝘳𝘪𝘦𝘷𝘢𝘭-𝘈𝘶𝘨𝘮𝘦𝘯𝘵𝘦𝘥 𝘎𝘦𝘯𝘦𝘳𝘢𝘵𝘪𝘰𝘯 (𝘙𝘈𝘎) 𝘢𝘯𝘥 𝘒𝘯𝘰𝘸𝘭𝘦𝘥𝘨𝘦 𝘎𝘳𝘢𝘱𝘩𝘴.
This isn't just about building Agents; it's about building AI that reasons, retrieves accurate information, and acts autonomously by leveraging structured knowledge alongside advanced LLMs.
The book offers a practical roadmap, packed with concrete Python examples and real-world case studies, guiding you from concept to deployment of intelligent, robust, and hallucination-minimized AI solutions, even orchestrating multi-agent systems.
Order your copy here - https://packt.link/RpzGM
#AI #LLMs #KnowledgeGraphs #AIAgents #RAG #GenerativeAI #MachineLearning
Why Knowledge Graphs are Critical to Agent Context
How should we organize knowledge to provide the best context for agents? We show how knowledge graphs could play a key role in enhancing context for agents.
Leveraging Knowledge Graphs and Large Language Models to Track and...
This study addresses the challenges of tracking and analyzing students' learning trajectories, particularly the issue of inadequate knowledge coverage in course assessments. Traditional assessment...
AutoSchemaKG: Autonomous Knowledge Graph Construction through...
We present AutoSchemaKG, a framework for fully autonomous knowledge graph construction that eliminates the need for predefined schemas. Our system leverages large language models to simultaneously...
Building Truly Autonomous AI: A Semantic Architecture Approach | LinkedIn
I've been working on autonomous AI systems, and wanted to share some thoughts on what I believe makes them effective. The challenge isn't just making AI that follows instructions well, but creating systems that can reason, and act independently.
LLMs already contain overlapping world models. You just have to ask them right.
Ontologists reply to an LLM output, “That’s not a real ontology—it’s not a formal conceptualization.”
But that’s just the No True Scotsman fallacy dressed up in OWL. Boring. Not growth-oriented. Look forward, angel.
A foundation model is a compression of human knowledge. The real problem isn't that we "lack a conceptualization". The real problem with an FM is that they contain too many. FMs contain conceptualizations—plural. Messy? Sure. But usable.
At Stardog, we’re turning this latent structure into real ontologies using symbolic knowledge distillation. Prompt orchestration → structure extraction → formal encoding. OWL, SHACL, and friends. Shake till mixed. Rinse. Repeat. Secret sauce simmered and reduced.
This isn't theoretical hard. We avoid that. It’s merely engineering hard. We LTF into that!
But the payoff means bootstrapping rich, new ontologies at scale: faster, cheaper, with lineage. It's the intersection of FM latent space, formal ontology, and user intent expressed via CQs. We call it the Symbolic Latent Layer (SLL). Cute eh?
The future of enterprise AI isn’t just documents. It’s distilling structured symbolic knowledge from LLMs and plugging it into agents, workflows, and reasoning engines.
You don’t need a priesthood to get a formal ontology anymore. You need a good prompt and a smarter pipeline and the right EKG platform.
There's a lot more to say about this so I said it at Stardog Labs https://lnkd.in/eY5Sibed | 17 comments on LinkedIn
Graph is the new star schema. Change my mind.
Why? Your agents can't be autonomous unless your structured data is a graph.
It is really very simple.
1️⃣ To act autonomously, an agent must reason across structured data.
Every autonomous decision - human or agent - hinges on a judgment: have I done enough? “Enough" boils down to driving the probability of success over some threshold.
2️⃣ You can’t just point the agent at your structured data store.
Context windows are too small. Schema sprawl is too real.
If you think it works, you probably haven’t tried it.
3️⃣ Agent must first retrieve - with RAG - the right tables, columns, and snippets. Decision making is a retrieval problem before it’s a reasoning problem.
4️⃣ Standard RAG breaks on enterprise metadata.
The corpus is too entity-rich.
Semantic similarity is breaking on enterprise help articles - it won't perform on column descriptions.
5️⃣ To make structured RAG work, you need a graph.
Just like unstructured RAG needed links between articles, structured RAG needs links between tables, fields, and - most importantly - meaning.
Yes, graphs are painful. But so was deep learning—until the return was undeniable. Agents need reasoning over structured data. That makes graphs non-optional. The rest is just engineering.
Let’s stop modeling for reporting—and start modeling for autonomy. | 28 comments on LinkedIn
How can you turn business questions into production-ready agentic knowledge graphs?
❓ How can you turn business questions into production-ready agentic knowledge graphs?
Join Prashanth Rao and Dennis Irorere at the Agentic AI Summit to find out.
Prashanth is an AI Engineer and DevRel lead at Kùzu Inc.—the open-source graph database startup—where he blends NLP, ML, and data engineering to power agentic workflows. Dennis is a Data Engineer at Tripadvisor’s Viator Marketing Technology team and Director of Innovation at GraphGeeks, driving scalable, AI-driven graph solutions for customer growth.
In “Agentic Workflows for Graph RAG: Building Production-Ready Knowledge Graphs,” they’ll guide you through three hands-on lessons:
🔹 From Business Question to Graph Schema – Modeling your domain for downstream agents and LLMs, using live data sources like AskNews.
🔹 From Unstructured Data to Agent-Ready Graphs with BAML – Writing declarative pipelines that reliably extract entities and relationships at scale.
🔹 Agentic Graph RAG in Action – Completing the loop: translating NL queries into Cypher, retrieving graph data, and synthesizing responses—with fallback strategies when matches are missing.
If you’re building internal tools or public-facing AI agents that rely on knowledge graphs, this workshop is for you.
🗓️ Learn more & register free: https://hubs.li/Q03qHnpQ0
#AgenticAI #GraphRAG #KnowledgeGraphs #AgentWorkflows #AIEngineering #ODSC #Kuzu #Tripadvisor
How can you turn business questions into production-ready agentic knowledge graphs?
Find out how to combine a knowledge graph with RAG for GraphRAG. Provide more complete GenAI outputs.
You’ve built a RAG system and grounded it in your own data. Then you ask a complex question that needs to draw from multiple sources. Your heart sinks when the answers you get are vague or plain wrong.
How could this happen?
Traditional vector-only RAG bases its outputs on just the words you use in your prompt. It misses out on valuable context because it pulls from different documents and data structures. Basically, it misses out on the bigger, more connected picture.
Your AI needs a mental model of your data with all its context and nuances. A knowledge graph provides just that by mapping your data as connected entities and relationships. Pair it with RAG to create a GraphRAG architecture to feed your LLM information about dependencies, sequences, hierarchies, and deeper meaning.
Check out The Developer’s Guide to GraphRAG. You’ll learn how to:
Prepare a knowledge graph for GraphRAG
Combine a knowledge graph with native vector search
Implement three GraphRAG retrieval patterns
HippoRAG takes cues from the brain to improve LLM retrieval
HippoRAG is a technique inspired from the interactions between the cortex and hippocampus to improve knowledge retrieval for large language models (LLM).
AI Engineer World's Fair 2025: GraphRAG Track Spotlight
📣 AI Engineer World's Fair 2025: GraphRAG Track Spotlight! 🚀
So grateful to have hosted the GraphRAG Track at the Fair. The sessions were great, highlighting the depth and breadth of graph thinking for AI.
Shoutouts to...
- Mitesh Patel "HybridRAG" as a fusion of graph and vector retrieval designed to master complex data interpretation and specialized terminology for question answering
- Chin Keong Lam "Wisdom Discovery at Scale" using Knowledge Augmented Generation (KAG) in a multi agent system with n8n
- Sam Julien "When Vectors Break Down" carefully explaining how graph-based RAG architecture achieved a whopping 86.31% accuracy for dense enterprise knowledge
- Daniel Chalef "Stop Using RAG as Memory" explored temporally-aware knowledge graphs, built by the open-source Graphiti framework, to provide precise, context-rich memory for agents,
- Ola Mabadeje "Witness the power of Multi-Agent AI & Network Knowledge Graphs" showing dramatic improvements in ticket resolution efficiency and overall execution quality in network operations.
- Thomas Smoker "Beyond Documents"! casually mentioning scraping the entire internet to distill a knowledge graph focused with legal agents
- Mark Bain hosting an excellent Agentic Memory with Knowledge Graphs lunch&learn, with expansive thoughts and demos from Vasilije Markovic Daniel Chalef and Alexander Gilmore
Also, of course, huge congrats to Shawn swyx W and Benjamin Dunphy on an excellent conference. 🎩
#graphrag Neo4j AI Engineer
AI Engineer World's Fair 2025: GraphRAG Track Spotlight
Optimizing the Interface Between Knowledge Graphs and LLMs for...
Integrating Large Language Models (LLMs) with Knowledge Graphs (KGs) results in complex systems with numerous hyperparameters that directly affect performance. While such systems are increasingly...
Want to Fix LLM Hallucination? Neurosymbolic Alone Won’t Cut It
Want to Fix LLM Hallucination? Neurosymbolic Alone Won’t Cut It
The Conversation’s new piece makes a clear case for neurosymbolic AI—integrating symbolic logic with statistical learning—as the long-term fix for LLM hallucinations. It’s a timely and necessary argument:
“No matter how large a language model gets, it can’t escape its fundamental lack of grounding in rules, logic, or real-world structure. Hallucination isn’t a bug, it’s the default.”
But what’s crucial—and often glossed over—is that symbolic logic alone isn’t enough. The real leap comes from adding formal ontologies and semantic constraints that make meaning machine-computable. OWL, Shapes Constraint Language (SHACL), and frameworks like BFO, Descriptive Ontology for Linguistic and Cognitive Engineering (DOLCE), the Suggested Upper Merged Ontology (SUMO), and the Common Core Ontologies (CCO) don’t just “represent rules”—they define what exists, what can relate, and under what conditions inference is valid. That’s the difference between “decorating” a knowledge graph and engineering one that can detect, explain, and prevent hallucinations in practice.
I’d go further:
• Most enterprise LLM hallucinations are just semantic errors—mislabeling, misattribution, or class confusion that only formal ontologies can prevent.
• Neurosymbolic systems only deliver if their symbolic half is grounded in ontological reality, not just handcrafted rules or taxonomies.
The upshot:
We need to move beyond mere integration of symbols and neurons. We need semantic scaffolding—ontologies as infrastructure—to ensure AI isn’t just fluent, but actually right.
Curious if others are layering formal ontologies (BFO, DOLCE, SUMO) into their AI stacks yet? Or are we still hoping that more compute and prompt engineering will do the trick?
#NeuroSymbolicAI #SemanticAI #Ontology #LLMs #AIHallucination #KnowledgeGraphs #AITrust #AIReasoning
Want to Fix LLM Hallucination? Neurosymbolic Alone Won’t Cut It
AutoSchemaKG: Building Billion-Node Knowledge Graphs Without Human Schemas
AutoSchemaKG: Building Billion-Node Knowledge Graphs Without Human Schemas
👉 Why This Matters
Traditional knowledge graphs face a paradox: they require expert-crafted schemas to organize information, creating bottlenecks for scalability and adaptability. This limits their ability to handle dynamic real-world knowledge or cross-domain applications effectively.
👉 What Changed
AutoSchemaKG eliminates manual schema design through three innovations:
1. Dynamic schema induction: LLMs automatically create conceptual hierarchies while extracting entities/events
2. Event-aware modeling: Captures temporal relationships and procedural knowledge missed by entity-only approaches
3. Multi-level conceptualization: Organizes instances into semantic categories through abstraction layers
The system processed 50M+ documents to build ATLAS - a family of KGs with:
- 900M+ nodes (entities/events/concepts)
- 5.9B+ relationships
- 95% alignment with human-created schemas (zero manual intervention)
👉 How It Works
1. Triple extraction pipeline:
- LLMs identify entity-entity, entity-event, and event-event relationships
- Processes text at document level rather than sentence level for context preservation
2. Schema induction:
- Automatically groups instances into conceptual categories
- Creates hierarchical relationships between specific facts and abstract concepts
3. Scale optimization:
- Handles web-scale corpora through GPU-accelerated batch processing
- Maintains semantic consistency across 3 distinct domains (Wikipedia, academic papers, Common Crawl)
👉 Proven Impact
- Boosts multi-hop QA accuracy by 12-18% over state-of-the-art baselines
- Improves LLM factuality by up to 9% on specialized domains like medicine and law
- Enables complex reasoning through conceptual bridges between disparate facts
👉 Key Insight
The research demonstrates that billion-scale KGs with dynamic schemas can effectively complement parametric knowledge in LLMs when they reach critical mass (1B+ facts). This challenges the assumption that retrieval augmentation needs domain-specific tuning to be effective.
Question for Discussion
As autonomous KG construction becomes viable, how should we rethink the role of human expertise in knowledge representation? Should curation shift from schema design to validation and ethical oversight? | 15 comments on LinkedIn
AutoSchemaKG: Building Billion-Node Knowledge Graphs Without Human Schemas
DRAG introduces a novel distillation framework that transfers RAG capabilities from LLMs to SLMs through Evidence-based distillation and Graph-based structuring
Small Models, Big Knowledge: How DRAG Bridges the AI Efficiency-Accuracy Gap
👉 Why This Matters
Modern AI systems face a critical tension: large language models (LLMs) deliver impressive knowledge recall but demand massive computational resources, while smaller models (SLMs) struggle with factual accuracy and "hallucinations." Traditional retrieval-augmented generation (RAG) systems amplify this problem by requiring constant updates to vast knowledge bases.
👉 The Innovation
DRAG introduces a novel distillation framework that transfers RAG capabilities from LLMs to SLMs through two key mechanisms:
1. Evidence-based distillation: Filters and ranks factual snippets from teacher LLMs
2. Graph-based structuring: Converts retrieved knowledge into relational graphs to preserve critical connections
This dual approach reduces model size requirements by 10-100x while improving factual accuracy by up to 27.7% compared to prior methods like MiniRAG.
👉 How It Works
1. Evidence generation: A large teacher LLM produces multiple context-relevant facts
2. Semantic filtering: Combines cosine similarity and LLM scoring to retain top evidence
3. Knowledge graph creation: Extracts entity relationships to form structured context
4. Distilled inference: SLMs generate answers using both filtered text and graph data
The process mimics how humans combine raw information with conceptual understanding, enabling smaller models to "think" like their larger counterparts without the computational overhead.
👉 Privacy Bonus
DRAG adds a privacy layer by:
- Local query sanitization before cloud processing
- Returning only de-identified knowledge graphs
Tests show 95.7% reduction in potential personal data leakage while maintaining answer quality.
👉 Why It’s Significant
This work addresses three critical challenges simultaneously:
- Makes advanced RAG capabilities accessible on edge devices
- Reduces hallucination rates through structured knowledge grounding
- Preserves user privacy in cloud-based AI interactions
The GitHub repository provides full implementation details, enabling immediate application in domains like healthcare diagnostics, legal analysis, and educational tools where accuracy and efficiency are non-negotiable.
DRAG introduces a novel distillation framework that transfers RAG capabilities from LLMs to SLMs through two key mechanisms:1. Evidence-based distillation: Filters and ranks factual snippets from teacher LLMs2. Graph-based structuring: Converts retrieved knowledge into relational graphs to preserve critical connections
I'm happy to share the draft of the "Semantically Composable Architectures" mini-paper.
It is the culmination of approximately four years' work, which began with Coreless Architectures and has now evolved into something much bigger.
LLMs are impressive, but a real breakthrough will occur once we surpass the cognitive capabilities of a single human brain.
Enabling autonomous large-scale system reverse engineering and large-scale autonomous transformation with minimal to no human involvement, while still making it understandable to humans if they choose to, is a central pillar of making truly groundbreaking changes.
We hope the ideas we shared will be beneficial to humanity and advance our civilization further.
It is not final and will require some clarification and improvements, but the key concepts are present. Happy to hear your thoughts and feedback.
Some of these concepts underpin the design of the Product X system.
Part of the core team + external contribution:
Andrew Barsukov Andrey Kolodnitsky Sapta Girisa N Keith E. Glendon Gurpreet Sachdeva Saurav Chandra Mike Diachenko Oleh Sinkevych | 13 comments on LinkedIn
Leveraging Large Language Models for Realizing Truly Intelligent...
The number of published scholarly articles is growing at a significant rate, making scholarly knowledge organization increasingly important. Various approaches have been proposed to organize...
Want to explore the Anthropic Transformer-Circuit's as a queryable graph?
Want to explore the Anthropic Transformer-Circuit's as a queryable graph?
Wrote a script to import the graph json into Neo4j - code in Gist.
https://lnkd.in/eT4NjQgY
https://lnkd.in/e38TfQpF
Next step - write directly from the circuit-tracer library to the graph db.
https://lnkd.in/eVU_t6mS
Want to explore the Anthropic Transformer-Circuit's as a queryable graph?
Introducing FACT: Fast Augmented Context Tools (3.2x faster, 90% cost reduction vs RAG)
Introducing FACT: Fast Augmented Context Tools (3.2x faster, 90% cost reduction vs RAG)
RAG had its run, but it’s not built for agentic systems. Vectors are fuzzy, slow, and blind to context. They work fine for static data, but once you enter recursive, real-time workflows, where agents need to reason, act, and reflect. RAG collapses under its own ambiguity.
That’s why I built FACT: Fast Augmented Context Tools.
Traditional Approach:
User Query → Database → Processing → Response (2-5 seconds)
FACT Approach:
User Query → Intelligent Cache → [If Miss] → Optimized Processing → Response (50ms)
It replaces vector search in RAG pipelines with a combination of intelligent prompt caching and deterministic tool execution via MCP. Instead of guessing which chunk is relevant, FACT explicitly retrieves structured data, SQL queries, live APIs, internal tools, then intelligently caches the result if it’s useful downstream.
The prompt caching isn’t just basic storage.
It’s intelligent using the prompt cache from Anthropic and other LLM providers, tuned for feedback-driven loops: static elements get reused, transient ones expire, and the system adapts in real time. Some things you always want cached, schemas, domain prompts. Others, like live data, need freshness. Traditional RAG is particularly bad at this. Ask anyone force to frequently update vector DBs.
I'm also using Arcade.dev to handle secure, scalable execution across both local and cloud environments, giving FACT hybrid intelligence for complex pipelines and automatic tool selection.
If you're building serious agents, skip the embeddings. RAG is a workaround. FACT is a foundation. It’s cheaper, faster, and designed for how agents actually work: with tools, memory, and intent.
To get started point your favorite coding agent at: https://lnkd.in/gek_akem | 38 comments on LinkedIn
Introducing FACT: Fast Augmented Context Tools (3.2x faster, 90% cost reduction vs RAG)