Found 3951 bookmarks
Newest
Introducing "Reasoning on Graphs (RoG)" - Unlocking Next-Level Reasoning for Large Language Models
Introducing "Reasoning on Graphs (RoG)" - Unlocking Next-Level Reasoning for Large Language Models
🚀 Exciting News: Introducing "Reasoning on Graphs (RoG)" - Unlocking Next-Level Reasoning for Large Language Models! 📊🧠 We are thrilled to unveil our… | 42 comments on LinkedIn
Introducing "Reasoning on Graphs (RoG)" - Unlocking Next-Level Reasoning for Large Language Models
·linkedin.com·
Introducing "Reasoning on Graphs (RoG)" - Unlocking Next-Level Reasoning for Large Language Models
Concepts is All You Need: A More Direct Path to AGI
Concepts is All You Need: A More Direct Path to AGI
Little demonstrable progress has been made toward AGI (Artificial General Intelligence) since the term was coined some 20 years ago. In spite of the fantastic breakthroughs in Statistical AI such as AlphaZero, ChatGPT, and Stable Diffusion none of these projects have, or claim to have, a clear path to AGI. In order to expedite the development of AGI it is crucial to understand and identify the core requirements of human-like intelligence as it pertains to AGI. From that one can distill which particular development steps are necessary to achieve AGI, and which are a distraction. Such analysis highlights the need for a Cognitive AI approach rather than the currently favored statistical and generative efforts. More specifically it identifies the central role of concepts in human-like cognition. Here we outline an architecture and development plan, together with some preliminary results, that offers a much more direct path to full Human-Level AI (HLAI)/ AGI.
·arxiv.org·
Concepts is All You Need: A More Direct Path to AGI
Knowledge Graphs Bootcamp on the O'Reilly Learning Platform
Knowledge Graphs Bootcamp on the O'Reilly Learning Platform
Three months ago, I had the privilege of hosting the Knowledge Graphs Bootcamp on the O'Reilly Learning Platform, and I'm truly grateful for the overwhelming…
Knowledge Graphs Bootcamp on the O'Reilly Learning Platform
·linkedin.com·
Knowledge Graphs Bootcamp on the O'Reilly Learning Platform
Graph Hairball
Graph Hairball
Knowledge graph system logic, the "things" and "relations between things" that  graph theory calls "vertices" (a.k.a. nodes, points, entit...
·digitalfinancialreporting.blogspot.com·
Graph Hairball
Graph Neural Prompting with Large Language Models
Graph Neural Prompting with Large Language Models
Large Language Models (LLMs) have shown remarkable generalization capability with exceptional performance in various language modeling tasks. However, they still exhibit inherent limitations in precisely capturing and returning grounded knowledge. While existing work has explored utilizing knowledge graphs to enhance language modeling via joint training and customized model architectures, applying this to LLMs is problematic owing to their large number of parameters and high computational cost. In addition, how to leverage the pre-trained LLMs and avoid training a customized model from scratch remains an open question. In this work, we propose Graph Neural Prompting (GNP), a novel plug-and-play method to assist pre-trained LLMs in learning beneficial knowledge from KGs. GNP encompasses various designs, including a standard graph neural network encoder, a cross-modality pooling module, a domain projector, and a self-supervised link prediction objective. Extensive experiments on multiple datasets demonstrate the superiority of GNP on both commonsense and biomedical reasoning tasks across different LLM sizes and settings.
·arxiv.org·
Graph Neural Prompting with Large Language Models
Scoping Knowledge Graphs | LinkedIn
Scoping Knowledge Graphs | LinkedIn
Building knowledge graphs is supposedly a huge and terrifying project, like fighting dragons or sending humans to Mars. I hear or see it time and time again: Knowledge graphs are too difficult, too time consuming, and too expensive to build.
·linkedin.com·
Scoping Knowledge Graphs | LinkedIn
Affinity-Aware Graph Networks
Affinity-Aware Graph Networks
Graph Neural Networks (GNNs) have emerged as a powerful technique for learning on relational data. Owing to the relatively limited number of message passing steps they perform -- and hence a smaller receptive field -- there has been significant interest in improving their expressivity by incorporating structural aspects of the underlying graph. In this paper, we explore the use of affinity measures as features in graph neural networks, in particular measures arising from random walks, including effective resistance, hitting and commute times. We propose message passing networks based on these features and evaluate their performance on a variety of node and graph property prediction tasks. Our architecture has lower computational complexity, while our features are invariant to the permutations of the underlying graph. The measures we compute allow the network to exploit the connectivity properties of the graph, thereby allowing us to outperform relevant benchmarks for a wide variety of tasks, often with significantly fewer message passing steps. On one of the largest publicly available graph regression datasets, OGB-LSC-PCQM4Mv1, we obtain the best known single-model validation MAE at the time of writing.
·arxiv.org·
Affinity-Aware Graph Networks
Chat with the Data Benchmark: Understanding Synergies between Large Language Models and Knowledge Graphs for Enterprise Conversations
Chat with the Data Benchmark: Understanding Synergies between Large Language Models and Knowledge Graphs for Enterprise Conversations
It was an honor to present the initial results of the Chat with the Data benchmark last week at the The Alan Turing Institute Knowledge Graph meetup (link to… | 11 comments on LinkedIn
·linkedin.com·
Chat with the Data Benchmark: Understanding Synergies between Large Language Models and Knowledge Graphs for Enterprise Conversations
Neural Graphical Models
Neural Graphical Models
Neural Graphical Models (NGMs) provide a solution to the challenges posed by traditional graphical models, offering greater flexibility, broader applicability, and improved performance in various domains. Learn more:
·microsoft.com·
Neural Graphical Models
JSON Crack - Visualize Data to Graphs
JSON Crack - Visualize Data to Graphs
Innovative and open-source visualization application that transforms various data formats, such as JSON, YAML, XML, CSV and more, into interactive graphs.
·jsoncrack.com·
JSON Crack - Visualize Data to Graphs
PyGraft, a configurable #Python tool to generate both synthetic #schemas and #knowledgeGraphs easily, supporting several RDFS and OWL constructs
PyGraft, a configurable #Python tool to generate both synthetic #schemas and #knowledgeGraphs easily, supporting several RDFS and OWL constructs
Happy to announce PyGraft, a configurable #Python tool to generate both synthetic #schemas and #knowledgeGraphs easily, supporting several RDFS and OWL constructs. Paper: https://t.co/p1Ei3PIhVz Code: https://t.co/ID6gU3elqK (also available on PyPI) @nicolas_hubr @mdaquin
·twitter.com·
PyGraft, a configurable #Python tool to generate both synthetic #schemas and #knowledgeGraphs easily, supporting several RDFS and OWL constructs
LLMs-represent-Knowledge Graphs | LinkedIn
LLMs-represent-Knowledge Graphs | LinkedIn
On August 14, 2023, the paper Natural Language is All a Graph Needs by Ruosong Ye, Caiqi Zhang, Runhui Wang, Shuyuan Xu and Yongfeng Zhang hit the arXiv streets and made quite a bang! The paper outlines a model called InstructGLM that adds further evidence that the future of graph representation lea
·linkedin.com·
LLMs-represent-Knowledge Graphs | LinkedIn
A eulogy for RedisGraph
A eulogy for RedisGraph
Less than five years after its initial release, Redis has announced that RedisGraph will be discontinued. Why?
·bloorresearch.com·
A eulogy for RedisGraph
Head-to-Tail: How Knowledgeable are Large Language Models (LLM)? A.K.A. Will LLMs Replace Knowledge Graphs?
Head-to-Tail: How Knowledgeable are Large Language Models (LLM)? A.K.A. Will LLMs Replace Knowledge Graphs?
Since the recent prosperity of Large Language Models (LLMs), there have been interleaved discussions regarding how to reduce hallucinations from LLM responses, how to increase the factuality of LLMs, and whether Knowledge Graphs (KGs), which store the world knowledge in a symbolic form, will be replaced with LLMs. In this paper, we try to answer these questions from a new angle: How knowledgeable are LLMs? To answer this question, we constructed Head-to-Tail, a benchmark that consists of 18K question-answer (QA) pairs regarding head, torso, and tail facts in terms of popularity. We designed an automated evaluation method and a set of metrics that closely approximate the knowledge an LLM confidently internalizes. Through a comprehensive evaluation of 14 publicly available LLMs, we show that existing LLMs are still far from being perfect in terms of their grasp of factual knowledge, especially for facts of torso-to-tail entities.
·arxiv.org·
Head-to-Tail: How Knowledgeable are Large Language Models (LLM)? A.K.A. Will LLMs Replace Knowledge Graphs?
Ontology Enrichment from Texts: A Biomedical Dataset for Concept Discovery and Placement
Ontology Enrichment from Texts: A Biomedical Dataset for Concept Discovery and Placement
Mentions of new concepts appear regularly in texts and require automated approaches to harvest and place them into Knowledge Bases (KB), e.g., ontologies and taxonomies. Existing datasets suffer from three issues, (i) mostly assuming that a new concept is pre-discovered and cannot support out-of-KB mention discovery; (ii) only using the concept label as the input along with the KB and thus lacking the contexts of a concept label; and (iii) mostly focusing on concept placement w.r.t a taxonomy of atomic concepts, instead of complex concepts, i.e., with logical operators. To address these issues, we propose a new benchmark, adapting MedMentions dataset (PubMed abstracts) with SNOMED CT versions in 2014 and 2017 under the Diseases sub-category and the broader categories of Clinical finding, Procedure, and Pharmaceutical / biologic product. We provide usage on the evaluation with the dataset for out-of-KB mention discovery and concept placement, adapting recent Large Language Model based methods.
·arxiv.org·
Ontology Enrichment from Texts: A Biomedical Dataset for Concept Discovery and Placement
The Memory Game: Investigating the Accuracy of AI Models in Storing and Recalling Facts. Comparing LLMs and Knowledge Graph on Factual Knowledge
The Memory Game: Investigating the Accuracy of AI Models in Storing and Recalling Facts. Comparing LLMs and Knowledge Graph on Factual Knowledge
The Memory Game: Investigating the Accuracy of AI Models in Storing and Recalling Facts … 🧠 ... Comparing LLMs and Knowledge Graph on Factual Knowledge I’m… | 18 comments on LinkedIn
·linkedin.com·
The Memory Game: Investigating the Accuracy of AI Models in Storing and Recalling Facts. Comparing LLMs and Knowledge Graph on Factual Knowledge