Found 2133 bookmarks
Newest
PyGraft, a configurable #Python tool to generate both synthetic #schemas and #knowledgeGraphs easily, supporting several RDFS and OWL constructs
PyGraft, a configurable #Python tool to generate both synthetic #schemas and #knowledgeGraphs easily, supporting several RDFS and OWL constructs
Happy to announce PyGraft, a configurable #Python tool to generate both synthetic #schemas and #knowledgeGraphs easily, supporting several RDFS and OWL constructs. Paper: https://t.co/p1Ei3PIhVz Code: https://t.co/ID6gU3elqK (also available on PyPI) @nicolas_hubr @mdaquin
·twitter.com·
PyGraft, a configurable #Python tool to generate both synthetic #schemas and #knowledgeGraphs easily, supporting several RDFS and OWL constructs
LLMs-represent-Knowledge Graphs | LinkedIn
LLMs-represent-Knowledge Graphs | LinkedIn
On August 14, 2023, the paper Natural Language is All a Graph Needs by Ruosong Ye, Caiqi Zhang, Runhui Wang, Shuyuan Xu and Yongfeng Zhang hit the arXiv streets and made quite a bang! The paper outlines a model called InstructGLM that adds further evidence that the future of graph representation lea
·linkedin.com·
LLMs-represent-Knowledge Graphs | LinkedIn
The Memory Game: Investigating the Accuracy of AI Models in Storing and Recalling Facts. Comparing LLMs and Knowledge Graph on Factual Knowledge
The Memory Game: Investigating the Accuracy of AI Models in Storing and Recalling Facts. Comparing LLMs and Knowledge Graph on Factual Knowledge
The Memory Game: Investigating the Accuracy of AI Models in Storing and Recalling Facts … 🧠 ... Comparing LLMs and Knowledge Graph on Factual Knowledge I’m… | 18 comments on LinkedIn
·linkedin.com·
The Memory Game: Investigating the Accuracy of AI Models in Storing and Recalling Facts. Comparing LLMs and Knowledge Graph on Factual Knowledge
LLMs4OL: Large Language Models for Ontology Learning
LLMs4OL: Large Language Models for Ontology Learning
We propose the LLMs4OL approach, which utilizes Large Language Models (LLMs) for Ontology Learning (OL). LLMs have shown significant advancements in natural language processing, demonstrating their ability to capture complex language patterns in different knowledge domains. Our LLMs4OL paradigm investigates the following hypothesis: \textit{Can LLMs effectively apply their language pattern capturing capability to OL, which involves automatically extracting and structuring knowledge from natural language text?} To test this hypothesis, we conduct a comprehensive evaluation using the zero-shot prompting method. We evaluate nine different LLM model families for three main OL tasks: term typing, taxonomy discovery, and extraction of non-taxonomic relations. Additionally, the evaluations encompass diverse genres of ontological knowledge, including lexicosemantic knowledge in WordNet, geographical knowledge in GeoNames, and medical knowledge in UMLS.
·arxiv.org·
LLMs4OL: Large Language Models for Ontology Learning
There is a lot of research on establishing mappings between different ontologies, but hardly any work on how to leverage such mappings in a query federation engine
There is a lot of research on establishing mappings between different ontologies, but hardly any work on how to leverage such mappings in a query federation engine
“There is a lot of research on establishing mappings between different ontologies, but hardly any work on how to leverage such mappings in a query federation engine. With @Sijin_Cheng and @ferradest, we have embarked on changing that. Paper at @CoopIS2023 https://t.co/vF1emf9R6Z”
·twitter.com·
There is a lot of research on establishing mappings between different ontologies, but hardly any work on how to leverage such mappings in a query federation engine
OpenFact: Factuality Enhanced Open Knowledge Extraction | Transactions of the Association for Computational Linguistics | MIT Press
OpenFact: Factuality Enhanced Open Knowledge Extraction | Transactions of the Association for Computational Linguistics | MIT Press
Abstract. We focus on the factuality property during the extraction of an OpenIE corpus named OpenFact, which contains more than 12 million high-quality knowledge triplets. We break down the factuality property into two important aspects—expressiveness and groundedness—and we propose a comprehensive framework to handle both aspects. To enhance expressiveness, we formulate each knowledge piece in OpenFact based on a semantic frame. We also design templates, extra constraints, and adopt human efforts so that most OpenFact triplets contain enough details. For groundedness, we require the main arguments of each triplet to contain linked Wikidata1 entities. A human evaluation suggests that the OpenFact triplets are much more accurate and contain denser information compared to OPIEC-Linked (Gashteovski et al., 2019), one recent high-quality OpenIE corpus grounded to Wikidata. Further experiments on knowledge base completion and knowledge base question answering show the effectiveness of OpenFact over OPIEC-Linked as supplementary knowledge to Wikidata as the major KG.
·direct.mit.edu·
OpenFact: Factuality Enhanced Open Knowledge Extraction | Transactions of the Association for Computational Linguistics | MIT Press
Knowledge graphs are graph-structured collections of facts. And facts are statements that define and describe subject entities in terms of predicates and their values
Knowledge graphs are graph-structured collections of facts. And facts are statements that define and describe subject entities in terms of predicates and their values
Knowledge graphs are graph-structured collections of facts. And facts are statements that define and describe subject entities in terms of predicates and their…
Knowledge graphs are graph-structured collections of facts. And facts are statements that define and describe subject entities in terms of predicates and their values
·linkedin.com·
Knowledge graphs are graph-structured collections of facts. And facts are statements that define and describe subject entities in terms of predicates and their values
Neosemantics (n10s) reaches the first million all-time downloads
Neosemantics (n10s) reaches the first million all-time downloads
📢 📢 📢 Amazing milestone! 📢 📢 📢 Neosemantics (n10s) reaches the first million all-time downloads 🤯 Let's keep building Knowledge Graphs together! 💪… | 21 comments on LinkedIn
Neosemantics (n10s) reaches the first million all-time downloads
·linkedin.com·
Neosemantics (n10s) reaches the first million all-time downloads
**Improved** — the BFO Classifier
**Improved** — the BFO Classifier
brief description of our FOIS2023 paper entitled “a method to improve alignments between domain and foundational ontologies”, focusing on BFO-aligned ontologies
·keet.wordpress.com·
**Improved** — the BFO Classifier
pg-schema schemas for property graphs
pg-schema schemas for property graphs
Arrived to SIGMOD in Seattle and it’s an amazing honor that our joint academic/industry work on Property Graph Schema received the Best Industry Paper award.… | 14 comments on LinkedIn
·linkedin.com·
pg-schema schemas for property graphs
Knowledge graphs are becoming an increasing area of interest in respect to explainability and interpretability of AI.
Knowledge graphs are becoming an increasing area of interest in respect to explainability and interpretability of AI.
Knowledge graphs are becoming an increasing area of interest in respect to explainability and interpretability of AI. To consolidate my own research and… | 30 comments on LinkedIn
Knowledge graphs are becoming an increasing area of interest in respect to explainability and interpretability of AI.
·linkedin.com·
Knowledge graphs are becoming an increasing area of interest in respect to explainability and interpretability of AI.
Link prediction on knowledge graphs is a loosing game, IMHO. Without injecting any new info, you'll only find links similar to those you already had. That's why this work is interesting: injecting external knowledge into link prediction is the only way to find truly new links
Link prediction on knowledge graphs is a loosing game, IMHO. Without injecting any new info, you'll only find links similar to those you already had. That's why this work is interesting: injecting external knowledge into link prediction is the only way to find truly new links
“Link prediction on knowledge graphs is a loosing game, IMHO. Without injecting any new info, you'll only find links similar to those you already had. That's why this work is interesting: injecting external knowledge into link prediction is the only way to find truly new links.”
Link prediction on knowledge graphs is a loosing game, IMHO. Without injecting any new info, you'll only find links similar to those you already had. That's why this work is interesting: injecting external knowledge into link prediction is the only way to find truly new links
·twitter.com·
Link prediction on knowledge graphs is a loosing game, IMHO. Without injecting any new info, you'll only find links similar to those you already had. That's why this work is interesting: injecting external knowledge into link prediction is the only way to find truly new links
LLM Ontology-prompting for Knowledge Graph Extraction
LLM Ontology-prompting for Knowledge Graph Extraction
Prompting an LLM with an ontology to drive Knowledge Graph extraction from unstructured documents
I make no apology for saying that a graph is the best organization of structured data. However, the vast majority of data is unstructured text. Therefore, data needs to be transformed from its original format using an Extract-Transform-Load (ETL) or Extract-Load-Transform (ELT) into a Knowledge Graph format. There is no problem when the original format is structured, such as SQL tables, spreadsheets, etc, or at least semi-structured, such as tweets. However, when the source data is unstructured text the task of ETL/ELT to a graph is far more challenging.This article shows how an LLM can be prompted with an unstructured document and asked to extract a graph corresponding to a specific ontology/schema. This is demonstrated with a Kennedy ontology in conjunction with a publicly available description of the Kennedy family tree.
·medium.com·
LLM Ontology-prompting for Knowledge Graph Extraction
How we delimit and develop the concept of facts is the key to a deeper, more detailed understanding of knowledge graphs
How we delimit and develop the concept of facts is the key to a deeper, more detailed understanding of knowledge graphs
How we delimit and develop the concept of facts is the key to a deeper, more detailed understanding of knowledge graphs because facts are crucial defining…
How we delimit and develop the concept of facts is the key to a deeper, more detailed understanding of knowledge graphs
·linkedin.com·
How we delimit and develop the concept of facts is the key to a deeper, more detailed understanding of knowledge graphs
This new paper is a wonderful story on how generative AI can be used to help curriculum designers build a Knowledge Space dependency graph
This new paper is a wonderful story on how generative AI can be used to help curriculum designers build a Knowledge Space dependency graph
This new paper is a wonderful story on how generative AI can be used to help curriculum designers build a Knowledge Space dependency graph: Exploring the MIT… | 11 comments on LinkedIn
·linkedin.com·
This new paper is a wonderful story on how generative AI can be used to help curriculum designers build a Knowledge Space dependency graph