Found 368 bookmarks
Custom sorting
pacoid (Paco Xander Nathan)
pacoid (Paco Xander Nathan)
Python open source projects; natural language meets graph technologies; graph topological transformations; graph levels of detail (abstraction layers)
·huggingface.co·
pacoid (Paco Xander Nathan)
Understand and Exploit GenAI With Gartner’s New Impact Radar
Understand and Exploit GenAI With Gartner’s New Impact Radar
Use Gartner’s impact radar for generative AI to plan investments and strategy with four key themes in mind: ☑️Model-related innovations ☑️Model performance and AI safety ☑️Model build and data-related ☑️AI-enabled applications Explore all 25 technologies and trends: https://www.gartner.com/en/articles/understand-and-exploit-gen-ai-with-gartner-s-new-impact-radar
·gartner.com·
Understand and Exploit GenAI With Gartner’s New Impact Radar
The Role of the Ontologist in the Age of LLMs
The Role of the Ontologist in the Age of LLMs
What do we mean when we say something is a kind of thing? I’ve been wrestling with that question a great deal of late, partly because I think the role of the ontologist transcends the application of knowledge graphs, especially as I’ve watched LLMs and Llamas become a bigger part of the discussion.
·ontologist.substack.com·
The Role of the Ontologist in the Age of LLMs
Knowledge Engineering Using Large Language Models
Knowledge Engineering Using Large Language Models
Knowledge engineering is a discipline that focuses on the creation and maintenance of processes that generate and apply knowledge. Traditionally, knowledge engineering approaches have focused on knowledge expressed in formal languages. The emergence of large language models and their capabilities to effectively work with natural language, in its broadest sense, raises questions about the foundations and practice of knowledge engineering. Here, we outline the potential role of LLMs in knowledge engineering, identifying two central directions: 1) creating hybrid neuro-symbolic knowledge systems; and 2) enabling knowledge engineering in natural language. Additionally, we formulate key open research questions to tackle these directions.
·drops.dagstuhl.de·
Knowledge Engineering Using Large Language Models
On to Knowledge-infused Language Models
On to Knowledge-infused Language Models
A broad and deep body of on-going research – hundreds of experiments! – has shown quite conclusively that knowledge graphs are essential to guide, complement, and enrich LLMs in systematic ways. The very wide variety of tests over domains and possible combinations of KGs and LLMs attests to the robu
·linkedin.com·
On to Knowledge-infused Language Models
Do Similar Entities have Similar Embeddings?
Do Similar Entities have Similar Embeddings?
Knowledge graph embedding models (KGEMs) developed for link prediction learn vector representations for graph entities, known as embeddings. A common tacit assumption is the KGE entity similarity assumption, which states that these KGEMs retain the graph's structure within their embedding space, i.e., position similar entities close to one another. This desirable property make KGEMs widely used in downstream tasks such as recommender systems or drug repurposing. Yet, the alignment of graph similarity with embedding space similarity has rarely been formally evaluated. Typically, KGEMs are assessed based on their sole link prediction capabilities, using ranked-based metrics such as Hits@K or Mean Rank. This paper challenges the prevailing assumption that entity similarity in the graph is inherently mirrored in the embedding space. Therefore, we conduct extensive experiments to measure the capability of KGEMs to cluster similar entities together, and investigate the nature of the underlying factors. Moreover, we study if different KGEMs expose a different notion of similarity. Datasets, pre-trained embeddings and code are available at: https://github.com/nicolas-hbt/similar-embeddings.
·arxiv.org·
Do Similar Entities have Similar Embeddings?
Graph Learning Meets Artificial Intelligence
Graph Learning Meets Artificial Intelligence
By request, here are the slides from our #neurips2023 presentation yesterday! We really enjoyed the opportunity to present the different aspects of the work… | 18 comments on LinkedIn
·linkedin.com·
Graph Learning Meets Artificial Intelligence
Language, Graphs, and AI in Industry
Language, Graphs, and AI in Industry
Here's the video for my talk @ K1st World Symposium 2023 about the intersections of KGs and LLMs: https://lnkd.in/gugB8Yjj and also the slides, plus related…
Language, Graphs, and AI in Industry
·linkedin.com·
Language, Graphs, and AI in Industry
knowledge graph based RAG (retrieval-augmentation) consistently improves language model accuracy, this time in biomedical questions
knowledge graph based RAG (retrieval-augmentation) consistently improves language model accuracy, this time in biomedical questions
The evidence for the massive impact of KGs in NLQ keeps piling up - Here's one more paper that shows that knowledge graph based RAG (retrieval-augmentation)…
knowledge graph based RAG (retrieval-augmentation) consistently improves language model accuracy, this time in biomedical questions
·linkedin.com·
knowledge graph based RAG (retrieval-augmentation) consistently improves language model accuracy, this time in biomedical questions
Large Language Models on Graphs: A Comprehensive Survey
Large Language Models on Graphs: A Comprehensive Survey
Large language models (LLMs), such as ChatGPT and LLaMA, are creating significant advancements in natural language processing, due to their strong text encoding/decoding ability and newly found emergent capability (e.g., reasoning). While LLMs are mainly designed to process pure texts, there are many real-world scenarios where text data are associated with rich structure information in the form of graphs (e.g., academic networks, and e-commerce networks) or scenarios where graph data are paired with rich textual information (e.g., molecules with descriptions). Besides, although LLMs have shown their pure text-based reasoning ability, it is underexplored whether such ability can be generalized to graph scenarios (i.e., graph-based reasoning). In this paper, we provide a systematic review of scenarios and techniques related to large language models on graphs. We first summarize potential scenarios of adopting LLMs on graphs into three categories, namely pure graphs, text-rich graphs, and text-paired graphs. We then discuss detailed techniques for utilizing LLMs on graphs, including LLM as Predictor, LLM as Encoder, and LLM as Aligner, and compare the advantages and disadvantages of different schools of models. Furthermore, we mention the real-world applications of such methods and summarize open-source codes and benchmark datasets. Finally, we conclude with potential future research directions in this fast-growing field. The related source can be found at https://github.com/PeterGriffinJin/Awesome-Language-Model-on-Graphs.
·arxiv.org·
Large Language Models on Graphs: A Comprehensive Survey
Convert your text into an interactive Knowledge Graph
Convert your text into an interactive Knowledge Graph
When reading lengthy or intricate texts, keeping an overview of different dependencies within the context is crucial. Traditionally, humans achieve this through note-taking or mentally creating a concept map. Now imagine having AI at hand which generates such a map for you. Even better, the…
·ai-readiness.ch·
Convert your text into an interactive Knowledge Graph