Found 15 bookmarks
Custom sorting
๐—ง๐—ต๐—ฒ ๐—™๐˜‚๐˜๐˜‚๐—ฟ๐—ฒ ๐—œ๐˜€ ๐—–๐—น๐—ฒ๐—ฎ๐—ฟ: ๐—”๐—ด๐—ฒ๐—ป๐˜๐˜€ ๐—ช๐—ถ๐—น๐—น ๐—ก๐—˜๐—˜๐—— ๐—š๐—ฟ๐—ฎ๐—ฝ๐—ต ๐—ฅ๐—”๐—š
๐—ง๐—ต๐—ฒ ๐—™๐˜‚๐˜๐˜‚๐—ฟ๐—ฒ ๐—œ๐˜€ ๐—–๐—น๐—ฒ๐—ฎ๐—ฟ: ๐—”๐—ด๐—ฒ๐—ป๐˜๐˜€ ๐—ช๐—ถ๐—น๐—น ๐—ก๐—˜๐—˜๐—— ๐—š๐—ฟ๐—ฎ๐—ฝ๐—ต ๐—ฅ๐—”๐—š
๐Ÿคบ ๐—ง๐—ต๐—ฒ ๐—™๐˜‚๐˜๐˜‚๐—ฟ๐—ฒ ๐—œ๐˜€ ๐—–๐—น๐—ฒ๐—ฎ๐—ฟ: ๐—”๐—ด๐—ฒ๐—ป๐˜๐˜€ ๐—ช๐—ถ๐—น๐—น ๐—ก๐—˜๐—˜๐—— ๐—š๐—ฟ๐—ฎ๐—ฝ๐—ต ๐—ฅ๐—”๐—š Why? It combines Multi-hop reasoning, Non-Parameterized / Learning-Based Retrieval, Topology-Aware Prompting. ๏นŒ๏นŒ๏นŒ๏นŒ๏นŒ๏นŒ๏นŒ๏นŒ๏นŒ ๐Ÿคบ ๐—ช๐—ต๐—ฎ๐˜ ๐—œ๐˜€ ๐—š๐—ฟ๐—ฎ๐—ฝ๐—ต-๐—˜๐—ป๐—ต๐—ฎ๐—ป๐—ฐ๐—ฒ๐—ฑ ๐—ฅ๐—ฒ๐˜๐—ฟ๐—ถ๐—ฒ๐˜ƒ๐—ฎ๐—น-๐—”๐˜‚๐—ด๐—บ๐—ฒ๐—ป๐˜๐—ฒ๐—ฑ ๐—š๐—ฒ๐—ป๐—ฒ๐—ฟ๐—ฎ๐˜๐—ถ๐—ผ๐—ป (๐—ฅ๐—”๐—š)? โœฉ LLMs hallucinate. โœฉ LLMs forget. โœฉ LLMs struggle with complex reasoning. Graphs connect facts. They organize knowledge into neat, structured webs. So when RAG retrieves from a graph, the LLM doesn't just guess โ€” it reasons. It follows the map. ๏นŒ๏นŒ๏นŒ๏นŒ๏นŒ๏นŒ๏นŒ๏นŒ๏นŒ ๐Ÿคบ ๐—ง๐—ต๐—ฒ ๐Ÿฐ-๐—ฆ๐˜๐—ฒ๐—ฝ ๐—ช๐—ผ๐—ฟ๐—ธ๐—ณ๐—น๐—ผ๐˜„ ๐—ผ๐—ณ ๐—š๐—ฟ๐—ฎ๐—ฝ๐—ต ๐—ฅ๐—”๐—š 1๏ธโƒฃ โ€” User Query: The user asks a question. ("Tell me how Einstein used Riemannian geometry?") 2๏ธโƒฃ โ€” Retrieval Module: The system fetches the most structurally relevant knowledge from a graph. (Entities: Einstein, Grossmann, Riemannian Geometry.) 3๏ธโƒฃ โ€” Prompting Module: Retrieved knowledge is reshaped into a golden prompt โ€” sometimes as structured triples, sometimes as smart text. 4๏ธโƒฃ โ€” Output Response: LLM generates a fact-rich, logically sound answer. ๏นŒ๏นŒ๏นŒ๏นŒ๏นŒ๏นŒ๏นŒ๏นŒ๏นŒ ๐Ÿคบ ๐—ฆ๐˜๐—ฒ๐—ฝ ๐Ÿญ: ๐—•๐˜‚๐—ถ๐—น๐—ฑ ๐—š๐—ฟ๐—ฎ๐—ฝ๐—ต-๐—ฃ๐—ผ๐˜„๐—ฒ๐—ฟ๐—ฒ๐—ฑ ๐——๐—ฎ๐˜๐—ฎ๐—ฏ๐—ฎ๐˜€๐—ฒ๐˜€ โœฉ Use Existing Knowledge Graphs like Freebase or Wikidata โ€” structured, reliable, but static. โœฉ Or Build New Graphs From Text (OpenIE, instruction-tuned LLMs) โ€” dynamic, adaptable, messy but powerful. ๐Ÿคบ ๐—ฆ๐˜๐—ฒ๐—ฝ ๐Ÿฎ: ๐—ฅ๐—ฒ๐˜๐—ฟ๐—ถ๐—ฒ๐˜ƒ๐—ฎ๐—น ๐—ฎ๐—ป๐—ฑ ๐—ฃ๐—ฟ๐—ผ๐—บ๐—ฝ๐˜๐—ถ๐—ป๐—ด ๐—”๐—น๐—ด๐—ผ๐—ฟ๐—ถ๐˜๐—ต๐—บ๐˜€ โœฉ Non-Parameterized Retrieval (Deterministic, Probabilistic, Heuristic) โ˜… Think Dijkstra's algorithm, PageRank, 1-hop neighbors. Fast but rigid. โœฉ Learning-Based Retrieval (GNNs, Attention Models) โ˜… Think "graph convolution" or "graph attention." Smarter, deeper, but heavier. โœฉ Prompting Approaches: โ˜… Topology-Aware: Preserve graph structure โ€” multi-hop reasoning. โ˜… Text Prompting: Flatten into readable sentences โ€” easier for vanilla LLMs. ๐Ÿคบ ๐—ฆ๐˜๐—ฒ๐—ฝ ๐Ÿฏ: ๐—š๐—ฟ๐—ฎ๐—ฝ๐—ต-๐—ฆ๐˜๐—ฟ๐˜‚๐—ฐ๐˜๐˜‚๐—ฟ๐—ฒ๐—ฑ ๐—ฃ๐—ถ๐—ฝ๐—ฒ๐—น๐—ถ๐—ป๐—ฒ๐˜€ โœฉ Sequential Pipelines: Straightforward query โž” retrieve โž” prompt โž” answer. โœฉ Loop Pipelines: Iterative refinement until the best evidence is found. โœฉ Tree Pipelines: Parallel exploration โž” multiple knowledge paths at once. ๐Ÿคบ ๐—ฆ๐˜๐—ฒ๐—ฝ ๐Ÿฐ: ๐—š๐—ฟ๐—ฎ๐—ฝ๐—ต-๐—ข๐—ฟ๐—ถ๐—ฒ๐—ป๐˜๐—ฒ๐—ฑ ๐—ง๐—ฎ๐˜€๐—ธ๐˜€ โœฉ Knowledge Graph QA (KGQA): Answering deep, logical questions with graphs. โœฉ Graph Tasks: Node classification, link prediction, graph summarization. โœฉ Domain-Specific Applications: Biomedicine, law, scientific discovery, finance. โ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃโ‰ฃ Join my ๐—›๐—ฎ๐—ป๐—ฑ๐˜€-๐—ผ๐—ป ๐—”๐—œ ๐—”๐—ด๐—ฒ๐—ป๐˜ ๐—ง๐—ฟ๐—ฎ๐—ถ๐—ป๐—ถ๐—ป๐—ด. Skip the fluff and build real AI agents โ€” fast. ๐—ช๐—ต๐—ฎ๐˜ ๐˜†๐—ผ๐˜‚ ๐—ด๐—ฒ๐˜: โœ… Create Smart Agents + Powerful RAG Pipelines โœ… Master ๐—Ÿ๐—ฎ๐—ป๐—ด๐—–๐—ต๐—ฎ๐—ถ๐—ป, ๐—–๐—ฟ๐—ฒ๐˜„๐—”๐—œ & ๐—ฆ๐˜„๐—ฎ๐—ฟ๐—บ โ€“ all in one training โœ… Projects with Text, Audio, Video & Tabular Data ๐Ÿฐ๐Ÿฒ๐Ÿฌ+ engineers already enrolled ๐—˜๐—ป๐—ฟ๐—ผ๐—น๐—น ๐—ป๐—ผ๐˜„ โ€” ๐Ÿฏ๐Ÿฐ% ๐—ผ๐—ณ๐—ณ, ๐—ฒ๐—ป๐—ฑ๐˜€ ๐˜€๐—ผ๐—ผ๐—ป:ย https://lnkd.in/eGuWr4CH | 35 comments on LinkedIn
๐—ง๐—ต๐—ฒ ๐—™๐˜‚๐˜๐˜‚๐—ฟ๐—ฒ ๐—œ๐˜€ ๐—–๐—น๐—ฒ๐—ฎ๐—ฟ: ๐—”๐—ด๐—ฒ๐—ป๐˜๐˜€ ๐—ช๐—ถ๐—น๐—น ๐—ก๐—˜๐—˜๐—— ๐—š๐—ฟ๐—ฎ๐—ฝ๐—ต ๐—ฅ๐—”๐—š
ยทlinkedin.comยท
๐—ง๐—ต๐—ฒ ๐—™๐˜‚๐˜๐˜‚๐—ฟ๐—ฒ ๐—œ๐˜€ ๐—–๐—น๐—ฒ๐—ฎ๐—ฟ: ๐—”๐—ด๐—ฒ๐—ป๐˜๐˜€ ๐—ช๐—ถ๐—น๐—น ๐—ก๐—˜๐—˜๐—— ๐—š๐—ฟ๐—ฎ๐—ฝ๐—ต ๐—ฅ๐—”๐—š
Lessons Learned from Evaluating NodeRAG vs Other RAG Systems
Lessons Learned from Evaluating NodeRAG vs Other RAG Systems
๐Ÿ”Ž Lessons Learned from Evaluating NodeRAG vs Other RAG Systems I recently dug into the NodeRAG paper (https://lnkd.in/gwaJHP94) and it was eye-opening not just for how it performed, but for what it revealed about the evolution of RAG (Retrieval-Augmented Generation) systems. Some key takeaways for me: ๐Ÿ‘‰ NaiveRAG is stronger than you think. Brute-force retrieval using simple vector search sometimes beats graph-based methods, especially when graph structures are too coarse or noisy. ๐Ÿ‘‰ GraphRAG was an important step, but not the final answer. While it introduced knowledge graphs and community-based retrieval, GraphRAG sometimes underperformed NaiveRAG because its communities could be too coarse, leading to irrelevant retrieval. ๐Ÿ‘‰ LightRAG reduced token cost, but at the expense of accuracy. By focusing on retrieving just 1-hop neighbors instead of traversing globally, LightRAG made retrieval cheaper โ€” but often missed important multi-hop reasoning paths, losing precision. ๐Ÿ‘‰ NodeRAG shows what mature RAG looks like. NodeRAG redesigned the graph structure itself: Instead of homogeneous graphs, it uses heterogeneous graphs with fine-grained semantic units, entities, relationships, and high-level summaries โ€” all as nodes. It combines dual search (exact match + semantic search) and shallow Personalized PageRank to precisely retrieve the most relevant context. The result? ๐Ÿš€ Highest accuracy across multi-hop and open-ended benchmarks ๐Ÿš€ Lowest token retrieval (i.e., lower inference costs) ๐Ÿš€ Faster indexing and querying ๐Ÿง  Key takeaway: In the RAG world, itโ€™s no longer about retrieving more โ€” itโ€™s about retrieving better. Fine-grained, explainable, efficient retrieval will define the next generation of RAG systems. If youโ€™re working on RAG architectures, NodeRAGโ€™s design principles are well worth studying! Would love to hear how others are thinking about the future of RAG systems. ๐Ÿš€๐Ÿ“š #RAG #KnowledgeGraphs #AI #LLM #NodeRAG #GraphRAG #LightRAG #MachineLearning #GenAI #KnowledegGraphs
ยทlinkedin.comยท
Lessons Learned from Evaluating NodeRAG vs Other RAG Systems
Google Cloud & Neo4j: Teaming Up at the Intersection of Knowledge Graphs, Agents, MCP, and Natural Language Interfaces - Graph Database & Analytics
Google Cloud & Neo4j: Teaming Up at the Intersection of Knowledge Graphs, Agents, MCP, and Natural Language Interfaces - Graph Database & Analytics
Weโ€™re thrilled to announce new Text2Cypher models and Googleโ€™s MCP Toolbox for Databases from the collaboration between Google Cloud and Neo4j.
ยทneo4j.comยท
Google Cloud & Neo4j: Teaming Up at the Intersection of Knowledge Graphs, Agents, MCP, and Natural Language Interfaces - Graph Database & Analytics
Choosing the Right Format: How Knowledge Graph Layouts Impact AI Reasoning
Choosing the Right Format: How Knowledge Graph Layouts Impact AI Reasoning
Choosing the Right Format: How Knowledge Graph Layouts Impact AI Reasoning ... ๐Ÿ‘‰ Why This Matters Most AI systems blend knowledge graphs (structured data) with large language models (flexible reasoning). But thereโ€™s a hidden variable: "how" you translate the graph into text for the AI. Researchers discovered that the formatting choice alone can swing performance by up to "17.5%" on reasoning tasks. Imagine solving 1 in 5 more problems correctly just by adjusting how you present data. ๐Ÿ‘‰ What They Built KG-LLM-Bench is a new benchmark to test how language models reason with knowledge graphs. It includes five tasks: - Triple verification (โ€œDoes this fact exist?โ€) - Shortest path finding (โ€œHow are two concepts connected?โ€) - Aggregation (โ€œHow many entities meet X condition?โ€) - Multi-hop reasoning (โ€œWhich entities linked to A also have property B?โ€) - Global analysis (โ€œWhich node is most central?โ€) The team tested seven models (Claude, GPT-4o, Gemini, Llama, Nova) with five ways to โ€œtextualizeโ€ graphs, from simple edge lists to structured JSON and semantic web formats like RDF Turtle. ๐Ÿ‘‰ Key Insights 1. Format matters more than assumed: ย ย - Structured JSON and edge lists performed best overall, but results varied by task. ย ย - For example, JSON excels at aggregation tasks (data is grouped by entity), while edge lists help identify central nodes (repeated mentions highlight connections). 2. Models donโ€™t cheat: Replacing real entity names with fake ones (e.g., โ€œFranceโ€ โ†’ โ€œVerdaniaโ€) caused only a 0.2% performance drop, proving models rely on context, not memorized knowledge. 3. Token efficiency: ย ย - Edge lists used ~2,600 tokens vs. JSON-LDโ€™s ~13,500. Shorter formats free up context space for complex reasoning. ย ย - But concise โ‰  always better: structured formats improved accuracy for tasks requiring grouped data. 4. Models struggle with directionality: ย  Counting outgoing edges (e.g., โ€œWhich countries does France border?โ€) is easier than incoming ones (โ€œWhich countries border France?โ€), likely due to formatting biases. ๐Ÿ‘‰ Practical Takeaways - Optimize for your task: Use JSON for aggregation, edge lists for centrality. - Test your model: The best format depends on the LLMโ€”Claude thrived with RDF Turtle, while Gemini preferred edge lists. - Donโ€™t fear pseudonyms: Masking real names minimally impacts performance, useful for sensitive data. The benchmark is openly available, inviting researchers to add new tasks, graphs, and models. As AI handles larger knowledge bases, choosing the right โ€œdata languageโ€ becomes as critical as the reasoning logic itself. Paper: [KG-LLM-Bench: A Scalable Benchmark for Evaluating LLM Reasoning on Textualized Knowledge Graphs] Authors: Elan Markowitz, Krupa Galiya, Greg Ver Steeg, Aram Galstyan
Choosing the Right Format: How Knowledge Graph Layouts Impact AI Reasoning
ยทlinkedin.comยท
Choosing the Right Format: How Knowledge Graph Layouts Impact AI Reasoning
Graph Data Modeling Without Graph Databases
Graph Data Modeling Without Graph Databases
Graph Data Modeling Without Graph Databases: PostgreSQL and Hybrid Approaches for Agentic Systems ๐Ÿ–‡๏ธ Organizations implementing AI systems today face a practical challenge: maintaining multiple specialized databases (vector stores, graph databases, relational systems) creates significant operational complexity, increases costs, and introduces synchronization headaches. Companies like Writer (insight from a recent Waseem Alshikh interview with Harrison Chase) have tackled this problem by implementing graph-like structures directly within PostgreSQL, eliminating the need for separate graph databases while maintaining the necessary functionality. This approach dramatically simplifies infrastructure management, reduces the number of systems to monitor, and eliminates error-prone synchronization processes that can cost thousands of dollars in wasted resources. For enterprises focused on delivering business value rather than managing technical complexity, these PostgreSQL-based implementations offer a pragmatic path forward, though with important trade-offs when considering more sophisticated agentic systems. Writer implemented a subject-predicate-object triple structure directly in PostgreSQL tables rather than using dedicated graph databases. This approach maintains the semantic richness of knowledge graphs while leveraging PostgreSQL's maturity and scalability. Writer kept the conceptual structure of triples that underpin knowledge graphs implemented through a relational schema design. Instead of relying on native graph traversals, Writer developed a fusion decoder that reconstructs graph-like relationships at query time. This component serves as the bridge between the storage layer (PostgreSQL with its triple-inspired structure) and the language model, enabling sophisticated information retrieval without requiring a dedicated graph database's traversal capabilities. The approach focuses on query translation and result combination rather than storage structure optimization. Complementing the triple-based approach, PostgreSQL with extensions (PG Vector and PG Vector Scale) can function effectively as a vector database. This challenges the notion that specialized vector databases are necessary, Treating embeddings as derived data leads to a more natural and maintainable architecture. This reframes the database's role from storing independent vector embeddings to managing derived data that automatically synchronizes with its source. But a critical distinction between retrieval systems and agentic systems need to be made. While PostgreSQL-based approaches excel at knowledge retrieval tasks where the focus is on precision and relevance, agentic systems operate in dynamic environments where context evolves over time, previous actions influence future decisions, and contradictions need to be resolved. This distinction drives different architectural requirements and suggests potential complementary roles for different database approaches. | 15 comments on LinkedIn
Graph Data Modeling Without Graph Databases
ยทlinkedin.comยท
Graph Data Modeling Without Graph Databases
Knowledge graphs for LLM grounding and avoiding hallucination
Knowledge graphs for LLM grounding and avoiding hallucination
This blog post is part of a series that dives into various aspects of SAPโ€™s approach to Generative AI, and its technical underpinnings. In previous blog posts of this series, you learned about how to use large language models (LLMs) for developing AI applications in a trustworthy and reliable manner...
ยทcommunity.sap.comยท
Knowledge graphs for LLM grounding and avoiding hallucination
Build your hybrid-Graph for RAG & GraphRAG applications using the power of NLP | LinkedIn
Build your hybrid-Graph for RAG & GraphRAG applications using the power of NLP | LinkedIn
Build a graph for RAG application for a price of a chocolate bar! What is GraphRAG for you? What is GraphRAG? What does GraphRAG mean from your perspective? What if you could have a standard RAG and a GraphRAG as a combi-package, with just a query switch? The fact is, there is no concrete, universal
ยทlinkedin.comยท
Build your hybrid-Graph for RAG & GraphRAG applications using the power of NLP | LinkedIn