The Question That Changes Everything: "But This Doesn't Look Like an Ontology" | LinkedIn
After publishing my article on the Missing Semantic Center, a brilliant colleague asked me a question that gets to the heart of our technology stack: "But Tavi - this doesn't look like an OWL 2 DL ontology. What's going on here?" This question highlights a profound aspect of why systems have struggl
Building Truly Autonomous AI: A Semantic Architecture Approach | LinkedIn
I've been working on autonomous AI systems, and wanted to share some thoughts on what I believe makes them effective. The challenge isn't just making AI that follows instructions well, but creating systems that can reason, and act independently.
When Is an Ontological Approach Not the Right Fit for Sharing and Reusing System Knowledge in Design and Development?
🧠 When Is an Ontological Approach Not the Right Fit for Sharing and Reusing System Knowledge in Design and Development?
Ontologies promise knowledge integration, traceability, reuse, and machine reasoning across the full engineering system lifecycle. From functional models to field failures, ontologies offer a way to encode and connect it all.
💥 However, ontologies are not a silver bullet.
There are plenty of scenarios where an ontology is not just unnecessary, it might actually slow you down, confuse your team, or waste resources.
So when exactly does the ontological approach become more burden than benefit? Based on my understanding and current work in this space,
🚀 For engineering design, it's important to recognise situations where adopting a semantic model is not the most effective approach:
1. When tasks are highly localised and routine
If you're just tweaking part drawings, running standard FEA simulations, or updating well-established design details, then the knowledge already lives in your tools and practices. Adding an ontology might feel like installing a satellite dish to tune a local radio station.
2. When terminology is unstable or fragmented
Ontologies depend on consistent language. If every department speaks its own dialect, and no one agrees on terms, you can't build shared meaning. You’ll end up formalising confusion instead of clarifying it.
3. When speed matters more than structure
In prototyping labs, testing grounds, or urgent production lines, agility rules. Engineers solve problems fast, often through direct collaboration. Taking time to define formal semantics? Not always practical. Sometimes the best model is a whiteboard and a sharp marker.
4. When the knowledge won’t be reused
Not all projects aim for longevity or cross-team learning. If you're building something once, for one purpose, with no intention of scaling or sharing, skip the ontology. It’s like building a library catalog for a single book.
5. When the infrastructure isn't there
Ontological engineering isn’t magic. It needs tools, training, and people who understand the stack. If your team lacks the skills or platforms, even the best-designed ontology will gather dust in a forgotten folder.
Use the Right Tool for the Real Problem
Ontologies are powerful, but not sacred. They shine when you need to connect knowledge across domains, ensure long-term traceability, or enable intelligent automation. But they’re not a requirement for every task just because they’re clever.
The real challenge is not whether to use ontologies, but knowing when they genuinely improve clarity, consistency, and collaboration, and when they just complicate the obvious.
🧠 Feedback and critique are welcome; this is a living conversation.
Felician Campean
#KnowledgeManagement #SystemsEngineering #Ontology #MBSE #DigitalEngineering #RiskAnalysis #AIinEngineering #OntologyEngineering #SemanticInteroperability #SystemReliability #FailureAnalysis #KnowledgeIntegration | 11 comments on LinkedIn
When Is an Ontological Approach Not the Right Fit for Sharing and Reusing System Knowledge in Design and Development?
LLMs already contain overlapping world models. You just have to ask them right.
Ontologists reply to an LLM output, “That’s not a real ontology—it’s not a formal conceptualization.”
But that’s just the No True Scotsman fallacy dressed up in OWL. Boring. Not growth-oriented. Look forward, angel.
A foundation model is a compression of human knowledge. The real problem isn't that we "lack a conceptualization". The real problem with an FM is that they contain too many. FMs contain conceptualizations—plural. Messy? Sure. But usable.
At Stardog, we’re turning this latent structure into real ontologies using symbolic knowledge distillation. Prompt orchestration → structure extraction → formal encoding. OWL, SHACL, and friends. Shake till mixed. Rinse. Repeat. Secret sauce simmered and reduced.
This isn't theoretical hard. We avoid that. It’s merely engineering hard. We LTF into that!
But the payoff means bootstrapping rich, new ontologies at scale: faster, cheaper, with lineage. It's the intersection of FM latent space, formal ontology, and user intent expressed via CQs. We call it the Symbolic Latent Layer (SLL). Cute eh?
The future of enterprise AI isn’t just documents. It’s distilling structured symbolic knowledge from LLMs and plugging it into agents, workflows, and reasoning engines.
You don’t need a priesthood to get a formal ontology anymore. You need a good prompt and a smarter pipeline and the right EKG platform.
There's a lot more to say about this so I said it at Stardog Labs https://lnkd.in/eY5Sibed | 17 comments on LinkedIn
Semantics in use (part 1): an interview with Martin Rezk, Sr. Ontologist at Google. | LinkedIn
To highlight the different uses and impact of semantics and ontologies, I wanted to present a series of interview to professionals from different industries and roles. This is going to be a distributed post.
Add a Semantic Layer – a smart translator that sits between your data sources and your business applications
Tired of being told that silos are gone? The real value comes from connecting them.
🔄 The myth of data silos: why they never really disappear, and how to turn them into your biggest advantage.
Even after heavy IT investment, data silos never truly go away, they simply evolve. In food production, I saw this first-hand: every system (ERP, quality, IoT, POS) stored data in its own format. Sometimes, the same product ended up with different IDs across systems, batch information was fragmented, and data was dispersed in each silo.
People often say, “Break down the silos.” But in reality, that’s nearly impossible. Businesses change, new tools appear, acquisitions happen, teams shift, new processes and production lines are launched. Silos are part of digital life.
For years, I tried classic integrations. They helped a bit, but every change in one system caused more issues and even more integration work. I wish I had known then what I know now:
Stop trying to destroy silos. Start connecting them.
Here’s what makes the difference:
Add a Semantic Layer – a smart translator that sits between your data sources and your business applications. It maps different formats and names into a common language, without changing your original systems.
Put a Knowledge Graph on top and you don’t just translate – you connect. Suddenly, all your data sources, even legacy silos, become part of a single network. Products, ingredients, machines, partners, and customers are all logically linked and understood across your business.
In practice, this means:
- Production uses real sales and shelf-life data.
- Sales sees live inventory, not outdated reports.
- Forecasting is based on trustworthy, aligned data.
That’s the real shift:
Silos are not problems to kill, but assets to connect.
With a Semantic Layer and a Knowledge Graph, data silos become trusted building blocks for your business intelligence.
Better Data, Better ROI.
If you’ve ever spent hours reconciling reports, you’ll recognise this recurring pain in companies that haven’t optimised their data integration with a semantic and KG approach.
So: Do you still treat silos as problems, or could they be your next competitive advantage if you connect them the right way?
Meaningfy
#DataSilos #SemanticLayer #KnowledgeGraph #BusinessData #DigitalTransformation
Add a Semantic Layer – a smart translator that sits between your data sources and your business applications
Cellosaurus is now available in RDF format, with a triple store that supports SPARQL queries
If this sounds a bit abstract or unfamiliar…
1) RDF stands for Resource Description Framework. Think of RDF as a way to express knowledge using triplets:
Subject – Predicate – Object.
Example: HeLa (subject) – is_transformed_by (predicate) – Human papillomavirus type 18 (object)
These triplets are like little facts that can be connected together to form a graph of knowledge.
2) A triple store is a database designed specifically to store and retrieve these RDF triplets. Unlike traditional databases (tables, rows), triple stores are optimized for linked data. They allow you to navigate connections between biological entities, like species, tissues, genes, diseases, etc.
3) SPARQL is a query language for RDF data. It lets you ask complex questions, such as:
- Find all cell lines with a *RAS (HRAS, NRAS, KRAS) mutation in p.Gly12
- Find all Cell lines from animals belonging the order "carnivora"
More specifically we now offer from the Tool - API submenu 6 new options:
1) SPARQL Editor (https://lnkd.in/eF2QMsYR). The SPARQL Editor is a tool designed to assist users in developing their SPARQL queries.
2) SPARQL Service (https://lnkd.in/eZ-iN7_e). The SPARQL service is the web service that accepts SPARQL queries over HTTP and returns results from the RDF dataset.
3) Cellosaurs Ontology (https://lnkd.in/eX5ExjMe). An RDF ontology is a formal, structured representation of knowledge. It explicitly defines domain-specific concepts - such as classes and properties - enabling data to be described with meaningful semantics that both humans and machines can interpret. The Cellosaurus ontology is expressed in OWL.
4) Cellosaurus Concept Hopper (https://lnkd.in/e7CH5nj4). The Concept Hopper, is a tool that provides an alternative view of the Cellosaurus ontology. It focuses on a single concept at a time - either a class or a property - and shows how that concept is linked to others within the ontology, as well as how it appears in the data.
5) Cellosaurus dereferencing service (https://lnkd.in/eSATMhGb). The RDF dereferencing service is the mechanism that, given a URI, returns an RDF description of the resource identified by that URI, enabling clients to retrieve structured, machine-readable data about the resource from the web in different formats.
6) Cellosaurus RDF files download (https://lnkd.in/emuEYnMD). This allows you to download the Cellosaurus RDF files in Turtle (ttl) format.
How do you explain the difference between Semantic Layers and Ontologies?
How do you explain the difference between Semantic Layers and Ontologies? That’s the discussion I had yesterday with the CTO of a very large and well known organization.
📊 Semantic Layers Today: The First Stepping Stone
• Semantic layer is commonly used in data analytics/BI reporting, tied to modeling fact/dimension tables and defining measures
• DataLakehouse/Data Cloud, transformation tools, BI tools and semantic layer vendors exemplify this usage
• Provide descriptive metadata: definitions, calculations (e.g., revenue formulas), and human readable labels, to enhance the schema
• Serve as a first step toward better data understanding and governance
• Help in aligning glossary terms with tables and columns, improving metadata quality and documentation
• Typically proprietary (even if expressed in YAML) and are not broadly interoperable
• Enable “chat with your data” experiences over the warehouse
When organizations need to integrate diverse data sources beyond the data warehouse/lakehouse model, they hit the limits of fact/dimension modeling. This is where ontologies and knowledge graphs come in.
🌐 Ontologies & Knowledge Graphs: Scaling Beyond BI
• Represent complex relationships, hierarchies, synonyms, and taxonomies that go beyond rigid table structures
• Knowledge graphs bridge the gap from technical metadata to business metadata and ultimately to core business concepts
• Enable the integration of all types of data (structured, semi-structured, unstructured) because a graph is a common model
• Through open web standards such as RDF, OWL and SPARQL you get interoperability without lock in
Strategic Role in the Enterprise
• Knowledge graphs enable the creation of an enterprise brain, connecting disparate data and semantics across all systems inside an organization
• Represent the context and meaning that LLMs lack. Our research has proven this.
• They lay the groundwork for digital twins and what-if scenario modeling, powering advanced analytics and decision-making.
💡 Key Takeaway
The semantic layer is a first step, especially for BI use cases. Most organizations will start with them. This will eventually create semantic silos that are not inherently interoperable. Over time, they realize they need more than just local semantics for BI; they want to model real-world business assets and relationships across systems.
Organizations will realize they want to define semantics once and reuse them across tools and platforms. This requires semantic interoperability, so the meaning behind data is not tied to one system. Large scale enterprises operate across multiple systems, so interoperability is not optional, it’s essential.
To truly integrate and reason over enterprise data, you need ontologies and knowledge graphs with open standards. They form the foundation for enterprise-wide semantic reuse, providing the flexibility, connectivity, and context required for next-generation analytics, AI, and enterprise intelligence. | 102 comments on LinkedIn
How do you explain the difference between Semantic Layers and Ontologies?
It’s been a while since I have posted about the Ontology Pipeline.
With parts borrowed from library science, the Ontology Pipeline is a simple framework for building rich knowledge infrastructures.
Librarians are professional stewards of knowledge, and have valuable methodologies for building information and knowledge systems for human and machine information retrieval tasks.
While LinkedIn conversations seem to be wrestling with defining “what is the semantic layer”, we are failing to see the root of semantics.
Semantics matter because knowledge structures, not just layers, define semantics.
Semantics are more than labels or concept maps. Semantics lend structure and meaning through relationships, disambiguation of concepts, definitions and context.
The Ontology pipeline is an iterative build process that is focused upon ensuring data hygiene while minding domain data, information and knowledge.
I share this framework because it is how I have successfully built information and knowledge ecosystems , with or without AI.
#taxonomy #ontology #metadata #knowledgegraph #ia #ai
Some friends focused on building knowledge infrastructures Andrew Padilla Nagim Ashufta Ole Olesen-Bagneux Jérémy Ravenel Paco Nathan Adriano Vlad-Starrabba Andrea Gioia | 10 comments on LinkedIn
New workbook for the Ontology Engineering textbook
The first part of my textbook revisions is complete and there’s now a first version of the accompanying workbook (also available here)! It’s designed partially to not substantially increase the pri…
Integrating Knowledge Graphs with Symbolic AI: The Path to Interpretable Hybrid AI Systems in Medicine
In this position paper "Integrating Knowledge Graphs with Symbolic AI: The Path to Interpretable Hybrid AI Systems in Medicine" my L3S Research Center and TIB – Leibniz-Informationszentrum Technik und Naturwissenschaften und Universitätsbibliothek colleagues around Maria-Esther Vidal have nicely laid out some research challenges on the way to interpretable hybrid AI systems in medicine. However, I think the conceptual framework is broadly applicable way beyond medicine.
For example, my former colleagues and PhD students at eccenca are working on operationalizing Neuro-Symbolic AI for Enterprise Knowledge Management with eccenca's Corporate Memory. The paper outlines a compelling architecture for combining sub-symbolic models (e.g., deep learning) with symbolic reasoning systems to enable AI that is interpretable, robust, and aligned with human values. eccenca implements these principles at scale through its neuro-symbolic Enterprise Knowledge Graph platform, Corporate Memory for real-world industrial settings:
1. Symbolic Foundation via Semantic Web Standards - Corporate Memory is grounded in W3C standards (RDF, RDFS, OWL, SHACL, SPARQL), enabling formal knowledge representation, inferencing, and constraint validation. This allows to encode domain ontologies, business rules, and data governance policies in a machine-interpretable and human-verifiable manner.
2. Integration of Sub-symbolic Components - it integrates LLMs and ML models for tasks such as schema matching, natural language interpretation, entity resolution, and ontology population. These are linked to the symbolic layer via mappings and annotations, ensuring traceability and explainability.
3. Neuro-Symbolic Interfaces for Hybrid Reasoning - Hybrid workflows where symbolic constraints (e.g., SHACL shapes) guide LLM-based data enrichment. LLMs suggest schema alignments, which are verified against ontological axioms. Graph embeddings and path-based querying power semantic search and similarity.
4. Human-in-the-loop Interactions - Domain experts interact through low-code interfaces and semantic UIs that allow inspection, validation, and refinement of both the symbolic and neural outputs, promoting human oversight and continuous improvement.
Such an approach can power Industrial Applications, e.g. in digital thread integration in manufacturing, compliance automation in pharma and finance
and in general, cross-domain interoperability in data mesh architectures. Corporate Memory is a practical instantiation of neuro-symbolic AI that meets industrial-grade requirements for governance, scalability, and explainability – key tenets of Human-Centric AI. Check it out here: https://lnkd.in/evyarUsR
#NeuroSymbolicAI #HumanCentricAI #KnowledgeGraphs #EnterpriseArchitecture #ExplainableAI #SemanticWeb #LinkedData #LLM #eccenca #CorporateMemory #OntologyDrivenAI #AI4Industry
Integrating Knowledge Graphs with Symbolic AI: The Path to Interpretable Hybrid AI Systems in Medicine
The Great Divide: Why Ontology and Data Architecture Teams Are Solving the Same Problems with Different Languages | LinkedIn
In enterprise organisations today, two important disciplines are working in parallel universes, tackling nearly identical challenges whilst speaking completely different languages. Ontology architects and data architects are both wrestling with ETL processes, data modelling, transformations, referen
Everyone is talking about Semantic Layers, but what is a semantic layer?
Everyone is talking about Semantic Layers, but what is a semantic layer?
Some of the latest hot topics to get more out of your agents discuss topics such as knowledge graphs, vector search, semantics, and agent frameworks.
A new and important area that encompasses the above is the notion that we need to have a stronger semantic layer on top of our data to provide structure, definitions, discoverability and more for our agents (human or other). While a lot of these concepts are not new, they have had to evolve to be relevant in today's world and this means that there is a fair bit of confusion surrounding this whole area. Depending on your background (AI, ML, Library Sciences) and focus (LLM-first or Knowledge Graph), you likely will emphasize different aspects as being key to a semantic layer.
I come primarily from an AI/ML/LLM-first world, but have built and utilized knowledge graphs for most of my career. Given my background, I of course have my perspective on this and I tend to break things down to first principles and I like to simplify. Given this, preamble, here is what I think makes a semantic layer.
WHAT MAKES A SEMANTIC LAYER:
🟤 Scope
🟢 You should not create a semantic layer that covers everything in the world, nor even everything in your company. You can tie semantic layers together, but focus on the job to be done.
🟤 You will need to have semantics, obviously. There are two particular types semantics that are important to include.
🟢 Vectors: These encapsulate semantics at a high-dimensional space so you can easily find similar concepts in your data
🟢 Ontology (including Taxonomy): Explicitly define meaning of your data in a structured and fact-based way, including appropriate vocabulary. This complements vectors superbly.
🟤 You need to respect the data and meet it where it is at.
🟢 Structured data: For most companies, their data reside in data lakes of some sort and most of it is structured. There is power in this structure, but also noise. The semantic layer needs to understand this and map it into the semantics above.
🟢 Unstructured data: Most data is unstructured and resides all over the place. Often this is stored in object stores or databases as part of structured tables, for example. However there is a lot of information in the unstructured data that the semantic layer needs to map -- and for that you need extraction, resolution, and a number of other techniques based on the modality of the data.
🟤 You need to index the data
🟢 You will need to index all of this to make your data discoverable and retrievable. And this needs to scale.
🟢 You need to have tight integration between vectors, ontology/knowledge graph and keywords to make this seamless.
These are 4 key components that are all needed for you to have a true semantic layer.
Thoughts?
#knowledgegraph, #semanticlayer, #agent, #rag | 13 comments on LinkedIn
Everyone is talking about Semantic Layers, but what is a semantic layer?
Why AI Hallucinates: The Shallow Semantics Problem | LinkedIn
By J Bittner Part 1 in our 5-part series: From Hallucination to Reasoning—The Case for Ontology-Driven AI Welcome to “Semantically Speaking”—a new series on what makes AI systems genuinely trustworthy, explainable, and future-proof. This is Part 1 in a 5-part journey, exploring why so many AI system
OntoAligner: A Comprehensive Modular and Robust Python Toolkit for Ontology Alignment
Ontology Alignment (OA) is fundamental for achieving semantic interoperability across diverse knowledge systems. We present OntoAligner, a comprehensive, modular, and robust Python toolkit for...
Last week, I was happy to be able to attend the 22nd European Semantic Web Conference. I’m a regular at this conference and it’s great to see many friends and colleagues as well as meet…
Building more Expressive Knowledge Graph Nodes | LinkedIn
In a knowledge graph, more expressive nodes are clearly more useful, dramatically more valuable nodes – when we focus on the right nodes. This was a key lesson I learned building knowledge graphs at LinkedIn with the terrific team that I assembled.
Want to Fix LLM Hallucination? Neurosymbolic Alone Won’t Cut It
Want to Fix LLM Hallucination? Neurosymbolic Alone Won’t Cut It
The Conversation’s new piece makes a clear case for neurosymbolic AI—integrating symbolic logic with statistical learning—as the long-term fix for LLM hallucinations. It’s a timely and necessary argument:
“No matter how large a language model gets, it can’t escape its fundamental lack of grounding in rules, logic, or real-world structure. Hallucination isn’t a bug, it’s the default.”
But what’s crucial—and often glossed over—is that symbolic logic alone isn’t enough. The real leap comes from adding formal ontologies and semantic constraints that make meaning machine-computable. OWL, Shapes Constraint Language (SHACL), and frameworks like BFO, Descriptive Ontology for Linguistic and Cognitive Engineering (DOLCE), the Suggested Upper Merged Ontology (SUMO), and the Common Core Ontologies (CCO) don’t just “represent rules”—they define what exists, what can relate, and under what conditions inference is valid. That’s the difference between “decorating” a knowledge graph and engineering one that can detect, explain, and prevent hallucinations in practice.
I’d go further:
• Most enterprise LLM hallucinations are just semantic errors—mislabeling, misattribution, or class confusion that only formal ontologies can prevent.
• Neurosymbolic systems only deliver if their symbolic half is grounded in ontological reality, not just handcrafted rules or taxonomies.
The upshot:
We need to move beyond mere integration of symbols and neurons. We need semantic scaffolding—ontologies as infrastructure—to ensure AI isn’t just fluent, but actually right.
Curious if others are layering formal ontologies (BFO, DOLCE, SUMO) into their AI stacks yet? Or are we still hoping that more compute and prompt engineering will do the trick?
#NeuroSymbolicAI #SemanticAI #Ontology #LLMs #AIHallucination #KnowledgeGraphs #AITrust #AIReasoning
Want to Fix LLM Hallucination? Neurosymbolic Alone Won’t Cut It
From Dictionaries to Ontologies: Bridging Human Understanding and Machine Reasoning | LinkedIn
In the long tradition of dictionaries, the essence of meaning has always relied on two elements: a symbol (usually a word or a phrase) and a definition—an intelligible explanation composed using other known terms. This recursive practice builds a web of meanings, where each term is explained using o
In complex engineering systems, how can we ensure that design knowledge doesn’t get lost in spreadsheets, silos, or forgotten documents? One of the greatest challenges in design domain and product development isn’t a lack of data, but a lack of meaningful, connected knowledge. This is where ontologies come in.
An ontology is more than just a taxonomy or glossary. It’s a formal representation of concepts and relationships that enables shared understanding across teams, tools, and disciplines. In the design domain, ontologies serve as a semantic backbone, helping engineers and systems interpret, reuse, and reason over knowledge that would otherwise remain trapped in silos.
Why does this matter? Because design decisions are rarely made in isolation. Whether it’s integrating functional models, analysing field failures, or updating risk assessment documents, we need a way to bridge data across multiple sources and domains. Ontologies enable that integration by creating a common language and structured relationships, allowing information to flow intelligently from design to deployment.
Ontology-driven systems also support human decision-making by enhancing traceability, contextualising feedback, and enabling AI-powered insights. It’s not about replacing designers, it’s about augmenting their intuition with structured, reusable knowledge.
As we move towards more data-driven and model-based approaches in engineering, ontologies are key to unlocking collaboration, innovation, and resilience in product development.
#Ontology #KnowledgeEngineering #SystemsThinking #DesignThinking #SystemEngineering #AI #DigitalEngineering #MBSE #KnowledgeSharing #DecisionSupport
#AugmentedIntelligence | 16 comments on LinkedIn
An ontology is more than just a taxonomy or glossary. It’s a formal representation of concepts and relationships
I'm happy to share the draft of the "Semantically Composable Architectures" mini-paper.
It is the culmination of approximately four years' work, which began with Coreless Architectures and has now evolved into something much bigger.
LLMs are impressive, but a real breakthrough will occur once we surpass the cognitive capabilities of a single human brain.
Enabling autonomous large-scale system reverse engineering and large-scale autonomous transformation with minimal to no human involvement, while still making it understandable to humans if they choose to, is a central pillar of making truly groundbreaking changes.
We hope the ideas we shared will be beneficial to humanity and advance our civilization further.
It is not final and will require some clarification and improvements, but the key concepts are present. Happy to hear your thoughts and feedback.
Some of these concepts underpin the design of the Product X system.
Part of the core team + external contribution:
Andrew Barsukov Andrey Kolodnitsky Sapta Girisa N Keith E. Glendon Gurpreet Sachdeva Saurav Chandra Mike Diachenko Oleh Sinkevych | 13 comments on LinkedIn
Leveraging Large Language Models for Realizing Truly Intelligent...
The number of published scholarly articles is growing at a significant rate, making scholarly knowledge organization increasingly important. Various approaches have been proposed to organize...