GraphNews

#AI
pacoid (Paco Xander Nathan)
pacoid (Paco Xander Nathan)
Python open source projects; natural language meets graph technologies; graph topological transformations; graph levels of detail (abstraction layers)
·huggingface.co·
pacoid (Paco Xander Nathan)
Understand and Exploit GenAI With Gartner’s New Impact Radar
Understand and Exploit GenAI With Gartner’s New Impact Radar
Use Gartner’s impact radar for generative AI to plan investments and strategy with four key themes in mind: ☑️Model-related innovations ☑️Model performance and AI safety ☑️Model build and data-related ☑️AI-enabled applications Explore all 25 technologies and trends: https://www.gartner.com/en/articles/understand-and-exploit-gen-ai-with-gartner-s-new-impact-radar
·gartner.com·
Understand and Exploit GenAI With Gartner’s New Impact Radar
The Role of the Ontologist in the Age of LLMs
The Role of the Ontologist in the Age of LLMs
What do we mean when we say something is a kind of thing? I’ve been wrestling with that question a great deal of late, partly because I think the role of the ontologist transcends the application of knowledge graphs, especially as I’ve watched LLMs and Llamas become a bigger part of the discussion.
·ontologist.substack.com·
The Role of the Ontologist in the Age of LLMs
Neural algorithmic reasoning without intermediate supervision
Neural algorithmic reasoning without intermediate supervision
Neural algorithmic reasoning focuses on building models that can execute classic algorithms. It allows one to combine the advantages of neural networks, such as handling raw and noisy input data, with theoretical guarantees and strong generalization of algorithms. Assuming we have a neural network capable of solving a classic algorithmic task, we can incorporate it into a more complex pipeline and train end-to-end. For instance, if we have a neural solver aligned to the shortest path problem, it can be used as a building block for a routing system that accounts for complex and dynamically changing traffic conditions. In our work [ref1], we study algorithmic reasoners trained only from input-output pairs, in contrast to current state-of-the-art approaches that utilize the trajectory of a given algorithm. We propose several architectural modifications and demonstrate how standard contrastive learning techniques can regularize intermediate computations of the models without appealing to any predefined algorithm’s trajectory.
·research.yandex.com·
Neural algorithmic reasoning without intermediate supervision
Knowledge Engineering Using Large Language Models
Knowledge Engineering Using Large Language Models
Knowledge engineering is a discipline that focuses on the creation and maintenance of processes that generate and apply knowledge. Traditionally, knowledge engineering approaches have focused on knowledge expressed in formal languages. The emergence of large language models and their capabilities to effectively work with natural language, in its broadest sense, raises questions about the foundations and practice of knowledge engineering. Here, we outline the potential role of LLMs in knowledge engineering, identifying two central directions: 1) creating hybrid neuro-symbolic knowledge systems; and 2) enabling knowledge engineering in natural language. Additionally, we formulate key open research questions to tackle these directions.
·drops.dagstuhl.de·
Knowledge Engineering Using Large Language Models
On to Knowledge-infused Language Models
On to Knowledge-infused Language Models
A broad and deep body of on-going research – hundreds of experiments! – has shown quite conclusively that knowledge graphs are essential to guide, complement, and enrich LLMs in systematic ways. The very wide variety of tests over domains and possible combinations of KGs and LLMs attests to the robu
·linkedin.com·
On to Knowledge-infused Language Models
Do Similar Entities have Similar Embeddings?
Do Similar Entities have Similar Embeddings?
Knowledge graph embedding models (KGEMs) developed for link prediction learn vector representations for graph entities, known as embeddings. A common tacit assumption is the KGE entity similarity assumption, which states that these KGEMs retain the graph's structure within their embedding space, i.e., position similar entities close to one another. This desirable property make KGEMs widely used in downstream tasks such as recommender systems or drug repurposing. Yet, the alignment of graph similarity with embedding space similarity has rarely been formally evaluated. Typically, KGEMs are assessed based on their sole link prediction capabilities, using ranked-based metrics such as Hits@K or Mean Rank. This paper challenges the prevailing assumption that entity similarity in the graph is inherently mirrored in the embedding space. Therefore, we conduct extensive experiments to measure the capability of KGEMs to cluster similar entities together, and investigate the nature of the underlying factors. Moreover, we study if different KGEMs expose a different notion of similarity. Datasets, pre-trained embeddings and code are available at: https://github.com/nicolas-hbt/similar-embeddings.
·arxiv.org·
Do Similar Entities have Similar Embeddings?
Data gauging, covariance and equivariance | Maurice Weiler
Data gauging, covariance and equivariance | Maurice Weiler
The numerical representation of data is often ambiguous. This leads to a gauge theoretic view on data, requiring covariant or equivariant neural networks which are reviewed in this blog post.
·maurice-weiler.gitlab.io·
Data gauging, covariance and equivariance | Maurice Weiler
Graph Learning Meets Artificial Intelligence
Graph Learning Meets Artificial Intelligence
By request, here are the slides from our #neurips2023 presentation yesterday! We really enjoyed the opportunity to present the different aspects of the work… | 18 comments on LinkedIn
·linkedin.com·
Graph Learning Meets Artificial Intelligence
Language, Graphs, and AI in Industry
Language, Graphs, and AI in Industry
Here's the video for my talk @ K1st World Symposium 2023 about the intersections of KGs and LLMs: https://lnkd.in/gugB8Yjj and also the slides, plus related…
Language, Graphs, and AI in Industry
·linkedin.com·
Language, Graphs, and AI in Industry
knowledge graph based RAG (retrieval-augmentation) consistently improves language model accuracy, this time in biomedical questions
knowledge graph based RAG (retrieval-augmentation) consistently improves language model accuracy, this time in biomedical questions
The evidence for the massive impact of KGs in NLQ keeps piling up - Here's one more paper that shows that knowledge graph based RAG (retrieval-augmentation)…
knowledge graph based RAG (retrieval-augmentation) consistently improves language model accuracy, this time in biomedical questions
·linkedin.com·
knowledge graph based RAG (retrieval-augmentation) consistently improves language model accuracy, this time in biomedical questions