Found 112 bookmarks
Custom sorting
Knowledge Graph-Augmented Language Models for Knowledge-Grounded Dialogue Generation
Knowledge Graph-Augmented Language Models for Knowledge-Grounded Dialogue Generation
Language models have achieved impressive performances on dialogue generation tasks. However, when generating responses for a conversation that requires factual knowledge, they are far from perfect, due to an absence of mechanisms to retrieve, encode, and reflect the knowledge in the generated responses. Some knowledge-grounded dialogue generation methods tackle this problem by leveraging facts from Knowledge Graphs (KGs); however, they do not guarantee that the model utilizes a relevant piece of knowledge from the KG. To overcome this limitation, we propose SUbgraph Retrieval-augmented GEneration (SURGE), a framework for generating context-relevant and knowledge-grounded dialogues with the KG. Specifically, our SURGE framework first retrieves the relevant subgraph from the KG, and then enforces consistency across facts by perturbing their word embeddings conditioned by the retrieved subgraph. Then, we utilize contrastive learning to ensure that the generated texts have high similarity to the retrieved subgraphs. We validate our SURGE framework on OpendialKG and KOMODIS datasets, showing that it generates high-quality dialogues that faithfully reflect the knowledge from KG.
Β·arxiv.orgΒ·
Knowledge Graph-Augmented Language Models for Knowledge-Grounded Dialogue Generation
πΊπ‘Ÿπ‘Žπ‘β„ŽπΈπ‘…: 𝐴 π‘†π‘‘π‘Ÿπ‘’π‘π‘‘π‘’π‘Ÿπ‘’-π‘Žπ‘€π‘Žπ‘Ÿπ‘’ 𝑇𝑒π‘₯𝑑-π‘‘π‘œ-πΊπ‘Ÿπ‘Žπ‘β„Ž π‘€π‘œπ‘‘π‘’π‘™ π‘“π‘œπ‘Ÿ 𝐸𝑛𝑑𝑖𝑑𝑦 π‘Žπ‘›π‘‘ π‘…π‘’π‘™π‘Žπ‘‘π‘–π‘œπ‘› 𝐸π‘₯π‘‘π‘Ÿπ‘Žπ‘π‘‘π‘–π‘œπ‘›
πΊπ‘Ÿπ‘Žπ‘β„ŽπΈπ‘…: 𝐴 π‘†π‘‘π‘Ÿπ‘’π‘π‘‘π‘’π‘Ÿπ‘’-π‘Žπ‘€π‘Žπ‘Ÿπ‘’ 𝑇𝑒π‘₯𝑑-π‘‘π‘œ-πΊπ‘Ÿπ‘Žπ‘β„Ž π‘€π‘œπ‘‘π‘’π‘™ π‘“π‘œπ‘Ÿ 𝐸𝑛𝑑𝑖𝑑𝑦 π‘Žπ‘›π‘‘ π‘…π‘’π‘™π‘Žπ‘‘π‘–π‘œπ‘› 𝐸π‘₯π‘‘π‘Ÿπ‘Žπ‘π‘‘π‘–π‘œπ‘›
Our paper "πΊπ‘Ÿπ‘Žπ‘β„ŽπΈπ‘…: 𝐴 π‘†π‘‘π‘Ÿπ‘’π‘π‘‘π‘’π‘Ÿπ‘’-π‘Žπ‘€π‘Žπ‘Ÿπ‘’ 𝑇𝑒π‘₯𝑑-π‘‘π‘œ-πΊπ‘Ÿπ‘Žπ‘β„Ž π‘€π‘œπ‘‘π‘’π‘™ π‘“π‘œπ‘Ÿ 𝐸𝑛𝑑𝑖𝑑𝑦 π‘Žπ‘›π‘‘ π‘…π‘’π‘™π‘Žπ‘‘π‘–π‘œπ‘›β€¦ | 34 comments on LinkedIn
πΊπ‘Ÿπ‘Žπ‘β„ŽπΈπ‘…: 𝐴 π‘†π‘‘π‘Ÿπ‘’π‘π‘‘π‘’π‘Ÿπ‘’-π‘Žπ‘€π‘Žπ‘Ÿπ‘’ 𝑇𝑒π‘₯𝑑-π‘‘π‘œ-πΊπ‘Ÿπ‘Žπ‘β„Ž π‘€π‘œπ‘‘π‘’π‘™ π‘“π‘œπ‘Ÿ 𝐸𝑛𝑑𝑖𝑑𝑦 π‘Žπ‘›π‘‘ π‘…π‘’π‘™π‘Žπ‘‘π‘–π‘œπ‘› 𝐸π‘₯π‘‘π‘Ÿπ‘Žπ‘π‘‘π‘–π‘œπ‘›
Β·linkedin.comΒ·
πΊπ‘Ÿπ‘Žπ‘β„ŽπΈπ‘…: 𝐴 π‘†π‘‘π‘Ÿπ‘’π‘π‘‘π‘’π‘Ÿπ‘’-π‘Žπ‘€π‘Žπ‘Ÿπ‘’ 𝑇𝑒π‘₯𝑑-π‘‘π‘œ-πΊπ‘Ÿπ‘Žπ‘β„Ž π‘€π‘œπ‘‘π‘’π‘™ π‘“π‘œπ‘Ÿ 𝐸𝑛𝑑𝑖𝑑𝑦 π‘Žπ‘›π‘‘ π‘…π‘’π‘™π‘Žπ‘‘π‘–π‘œπ‘› 𝐸π‘₯π‘‘π‘Ÿπ‘Žπ‘π‘‘π‘–π‘œπ‘›
The Era of Semantic Decoding
The Era of Semantic Decoding
Recent work demonstrated great promise in the idea of orchestrating collaborations between LLMs, human input, and various tools to address the inherent limitations of LLMs. We propose a novel perspective called semantic decoding, which frames these collaborative processes as optimization procedures in semantic space. Specifically, we conceptualize LLMs as semantic processors that manipulate meaningful pieces of information that we call semantic tokens (known thoughts). LLMs are among a large pool of other semantic processors, including humans and tools, such as search engines or code executors. Collectively, semantic processors engage in dynamic exchanges of semantic tokens to progressively construct high-utility outputs. We refer to these orchestrated interactions among semantic processors, optimizing and searching in semantic space, as semantic decoding algorithms. This concept draws a direct parallel to the well-studied problem of syntactic decoding, which involves crafting algorithms to best exploit auto-regressive language models for extracting high-utility sequences of syntactic tokens. By focusing on the semantic level and disregarding syntactic details, we gain a fresh perspective on the engineering of AI systems, enabling us to imagine systems with much greater complexity and capabilities. In this position paper, we formalize the transition from syntactic to semantic tokens as well as the analogy between syntactic and semantic decoding. Subsequently, we explore the possibilities of optimizing within the space of semantic tokens via semantic decoding algorithms. We conclude with a list of research opportunities and questions arising from this fresh perspective. The semantic decoding perspective offers a powerful abstraction for search and optimization directly in the space of meaningful concepts, with semantic tokens as the fundamental units of a new type of computation.
Β·arxiv.orgΒ·
The Era of Semantic Decoding
Neurosymbolic AI for Reasoning over Knowledge Graphs: A Survey
Neurosymbolic AI for Reasoning over Knowledge Graphs: A Survey
Neurosymbolic AI is an increasingly active area of research that combines symbolic reasoning methods with deep learning to leverage their complementary benefits. As knowledge graphs are becoming a popular way to represent heterogeneous and multi-relational data, methods for reasoning on graph structures have attempted to follow this neurosymbolic paradigm. Traditionally, such approaches have utilized either rule-based inference or generated representative numerical embeddings from which patterns could be extracted. However, several recent studies have attempted to bridge this dichotomy to generate models that facilitate interpretability, maintain competitive performance, and integrate expert knowledge. Therefore, we survey methods that perform neurosymbolic reasoning tasks on knowledge graphs and propose a novel taxonomy by which we can classify them. Specifically, we propose three major categories: (1) logically-informed embedding approaches, (2) embedding approaches with logical constraints, and (3) rule learning approaches. Alongside the taxonomy, we provide a tabular overview of the approaches and links to their source code, if available, for more direct comparison. Finally, we discuss the unique characteristics and limitations of these methods, then propose several prospective directions toward which this field of research could evolve.
Β·arxiv.orgΒ·
Neurosymbolic AI for Reasoning over Knowledge Graphs: A Survey
An Online Spatial-Temporal Graph Trajectory Planner for Autonomous Vehicles
An Online Spatial-Temporal Graph Trajectory Planner for Autonomous Vehicles
The autonomous driving industry is expected to grow by over 20 times in the coming decade and, thus, motivate researchers to delve into it. The primary focus of their research is to ensure safety, comfort, and efficiency. An autonomous vehicle has several modules responsible for one or more of the aforementioned items. Among these modules, the trajectory planner plays a pivotal role in the safety of the vehicle and the comfort of its passengers. The module is also responsible for respecting kinematic constraints and any applicable road constraints. In this paper, a novel online spatial-temporal graph trajectory planner is introduced to generate safe and comfortable trajectories. First, a spatial-temporal graph is constructed using the autonomous vehicle, its surrounding vehicles, and virtual nodes along the road with respect to the vehicle itself. Next, the graph is forwarded into a sequential network to obtain the desired states. To support the planner, a simple behavioral layer is also presented that determines kinematic constraints for the planner. Furthermore, a novel potential function is also proposed to train the network. Finally, the proposed planner is tested on three different complex driving tasks, and the performance is compared with two frequently used methods. The results show that the proposed planner generates safe and feasible trajectories while achieving similar or longer distances in the forward direction and comparable comfort ride.
Β·arxiv.orgΒ·
An Online Spatial-Temporal Graph Trajectory Planner for Autonomous Vehicles
DST-GTN: Dynamic Spatio-Temporal Graph Transformer Network for Traffic Forecasting
DST-GTN: Dynamic Spatio-Temporal Graph Transformer Network for Traffic Forecasting
Accurate traffic forecasting is essential for effective urban planning and congestion management. Deep learning (DL) approaches have gained colossal success in traffic forecasting but still face challenges in capturing the intricacies of traffic dynamics. In this paper, we identify and address this challenges by emphasizing that spatial features are inherently dynamic and change over time. A novel in-depth feature representation, called Dynamic Spatio-Temporal (Dyn-ST) features, is introduced, which encapsulates spatial characteristics across varying times. Moreover, a Dynamic Spatio-Temporal Graph Transformer Network (DST-GTN) is proposed by capturing Dyn-ST features and other dynamic adjacency relations between intersections. The DST-GTN can model dynamic ST relationships between nodes accurately and refine the representation of global and local ST characteristics by adopting adaptive weights in low-pass and all-pass filters, enabling the extraction of Dyn-ST features from traffic time-series data. Through numerical experiments on public datasets, the DST-GTN achieves state-of-the-art performance for a range of traffic forecasting tasks and demonstrates enhanced stability.
Β·arxiv.orgΒ·
DST-GTN: Dynamic Spatio-Temporal Graph Transformer Network for Traffic Forecasting
GLaM: Fine-Tuning Large Language Models for Domain Knowledge Graph Alignment via Neighborhood Partitioning and Generative Subgraph Encoding
GLaM: Fine-Tuning Large Language Models for Domain Knowledge Graph Alignment via Neighborhood Partitioning and Generative Subgraph Encoding
Integrating large language models (LLMs) with knowledge graphs derived from domain-specific data represents an important advancement towards more powerful and factual reasoning. As these models grow more capable, it is crucial to enable them to perform multi-step inferences over real-world knowledge graphs while minimizing hallucination. While large language models excel at conversation and text generation, their ability to reason over domain-specialized graphs of interconnected entities remains limited. For example, can we query a LLM to identify the optimal contact in a professional network for a specific goal, based on relationships and attributes in a private database? The answer is no--such capabilities lie beyond current methods. However, this question underscores a critical technical gap that must be addressed. Many high-value applications in areas such as science, security, and e-commerce rely on proprietary knowledge graphs encoding unique structures, relationships, and logical constraints. We introduce a fine-tuning framework for developing Graph-aligned LAnguage Models (GLaM) that transforms a knowledge graph into an alternate text representation with labeled question-answer pairs. We demonstrate that grounding the models in specific graph-based knowledge expands the models' capacity for structure-based reasoning. Our methodology leverages the large-language model's generative capabilities to create the dataset and proposes an efficient alternate to retrieval-augmented generation styled methods.
Β·arxiv.orgΒ·
GLaM: Fine-Tuning Large Language Models for Domain Knowledge Graph Alignment via Neighborhood Partitioning and Generative Subgraph Encoding
GraphGPT
GraphGPT
🌟GraphGPT🌟 (385 stars in GitHub) is accepted by 🌟SIGIR'24🌟 (only 20.1% acceptance rate)! Thank Yuhao Yang, wei wei, and other co-authors for their precious…
GraphGPT
Β·linkedin.comΒ·
GraphGPT
Exploring the Potential of Large Language Models in Graph Generation
Exploring the Potential of Large Language Models in Graph Generation
Large language models (LLMs) have achieved great success in many fields, and recent works have studied exploring LLMs for graph discriminative tasks such as node classification. However, the abilities of LLMs for graph generation remain unexplored in the literature. Graph generation requires the LLM to generate graphs with given properties, which has valuable real-world applications such as drug discovery, while tends to be more challenging. In this paper, we propose LLM4GraphGen to explore the ability of LLMs for graph generation with systematical task designs and extensive experiments. Specifically, we propose several tasks tailored with comprehensive experiments to address key questions regarding LLMs' understanding of different graph structure rules, their ability to capture structural type distributions, and their utilization of domain knowledge for property-based graph generation. Our evaluations demonstrate that LLMs, particularly GPT-4, exhibit preliminary abilities in graph generation tasks, including rule-based and distribution-based generation. We also observe that popular prompting methods, such as few-shot and chain-of-thought prompting, do not consistently enhance performance. Besides, LLMs show potential in generating molecules with specific properties. These findings may serve as foundations for designing good LLMs based models for graph generation and provide valuable insights and further research.
Β·arxiv.orgΒ·
Exploring the Potential of Large Language Models in Graph Generation
OpenGraph: Open-Vocabulary Hierarchical 3D Graph Representation in Large-Scale Outdoor Environments
OpenGraph: Open-Vocabulary Hierarchical 3D Graph Representation in Large-Scale Outdoor Environments
Environment maps endowed with sophisticated semantics are pivotal for facilitating seamless interaction between robots and humans, enabling them to effectively carry out various tasks. Open-vocabulary maps, powered by Visual-Language models (VLMs), possess inherent advantages, including multimodal retrieval and open-set classes. However, existing open-vocabulary maps are constrained to closed indoor scenarios and VLM features, thereby diminishing their usability and inference capabilities. Moreover, the absence of topological relationships further complicates the accurate querying of specific instances. In this work, we propose OpenGraph, a representation of open-vocabulary hierarchical graph structure designed for large-scale outdoor environments. OpenGraph initially extracts instances and their captions from visual images using 2D foundation models, encoding the captions with features to enhance textual reasoning. Subsequently, 3D incremental panoramic mapping with feature embedding is achieved by projecting images onto LiDAR point clouds. Finally, the environment is segmented based on lane graph connectivity to construct a hierarchical graph. Validation results from real public dataset SemanticKITTI demonstrate that, even without fine-tuning the models, OpenGraph exhibits the ability to generalize to novel semantic classes and achieve the highest segmentation and query accuracy. The source code of OpenGraph is publicly available at https://github.com/BIT-DYN/OpenGraph.
Β·arxiv.orgΒ·
OpenGraph: Open-Vocabulary Hierarchical 3D Graph Representation in Large-Scale Outdoor Environments
Personalized Audiobook Recommendations at Spotify Through Graph Neural Networks
Personalized Audiobook Recommendations at Spotify Through Graph Neural Networks
In the ever-evolving digital audio landscape, Spotify, well-known for its music and talk content, has recently introduced audiobooks to its vast user base. While promising, this move presents significant challenges for personalized recommendations. Unlike music and podcasts, audiobooks, initially available for a fee, cannot be easily skimmed before purchase, posing higher stakes for the relevance of recommendations. Furthermore, introducing a new content type into an existing platform confronts extreme data sparsity, as most users are unfamiliar with this new content type. Lastly, recommending content to millions of users requires the model to react fast and be scalable. To address these challenges, we leverage podcast and music user preferences and introduce 2T-HGNN, a scalable recommendation system comprising Heterogeneous Graph Neural Networks (HGNNs) and a Two Tower (2T) model. This novel approach uncovers nuanced item relationships while ensuring low latency and complexity. We decouple users from the HGNN graph and propose an innovative multi-link neighbor sampler. These choices, together with the 2T component, significantly reduce the complexity of the HGNN model. Empirical evaluations involving millions of users show significant improvement in the quality of personalized recommendations, resulting in a +46% increase in new audiobooks start rate and a +23% boost in streaming rates. Intriguingly, our model's impact extends beyond audiobooks, benefiting established products like podcasts.
Β·arxiv.orgΒ·
Personalized Audiobook Recommendations at Spotify Through Graph Neural Networks
Artificial Intelligence for Complex Network: Potential, Methodology and Application
Artificial Intelligence for Complex Network: Potential, Methodology and Application
Complex networks pervade various real-world systems, from the natural environment to human societies. The essence of these networks is in their ability to transition and evolve from microscopic disorder-where network topology and node dynamics intertwine-to a macroscopic order characterized by certain collective behaviors. Over the past two decades, complex network science has significantly enhanced our understanding of the statistical mechanics, structures, and dynamics underlying real-world networks. Despite these advancements, there remain considerable challenges in exploring more realistic systems and enhancing practical applications. The emergence of artificial intelligence (AI) technologies, coupled with the abundance of diverse real-world network data, has heralded a new era in complex network science research. This survey aims to systematically address the potential advantages of AI in overcoming the lingering challenges of complex network research. It endeavors to summarize the pivotal research problems and provide an exhaustive review of the corresponding methodologies and applications. Through this comprehensive survey-the first of its kind on AI for complex networks-we expect to provide valuable insights that will drive further research and advancement in this interdisciplinary field.
Β·arxiv.orgΒ·
Artificial Intelligence for Complex Network: Potential, Methodology and Application
Knowledge Engineering using Large Language Models
Knowledge Engineering using Large Language Models
Knowledge engineering is a discipline that focuses on the creation and maintenance of processes that generate and apply knowledge. Traditionally, knowledge engineering approaches have focused on knowledge expressed in formal languages. The emergence of large language models and their capabilities to effectively work with natural language, in its broadest sense, raises questions about the foundations and practice of knowledge engineering. Here, we outline the potential role of LLMs in knowledge engineering, identifying two central directions: 1) creating hybrid neuro-symbolic knowledge systems; and 2) enabling knowledge engineering in natural language. Additionally, we formulate key open research questions to tackle these directions.
Β·arxiv.orgΒ·
Knowledge Engineering using Large Language Models
Knowledge Graphs Achieve Superior Reasoning versus Vector Search alone for Retrieval Augmentation
Knowledge Graphs Achieve Superior Reasoning versus Vector Search alone for Retrieval Augmentation
Knowledge Graphs Achieve Superior Reasoning versus Vector Search alone for Retrieval Augmentation πŸ”— As artificial intelligence permeates business… | 29 comments on LinkedIn
Knowledge Graphs Achieve Superior Reasoning versus Vector Search alone for Retrieval Augmentation
Β·linkedin.comΒ·
Knowledge Graphs Achieve Superior Reasoning versus Vector Search alone for Retrieval Augmentation
Knowledge Engineering Using Large Language Models
Knowledge Engineering Using Large Language Models
Knowledge engineering is a discipline that focuses on the creation and maintenance of processes that generate and apply knowledge. Traditionally, knowledge engineering approaches have focused on knowledge expressed in formal languages. The emergence of large language models and their capabilities to effectively work with natural language, in its broadest sense, raises questions about the foundations and practice of knowledge engineering. Here, we outline the potential role of LLMs in knowledge engineering, identifying two central directions: 1) creating hybrid neuro-symbolic knowledge systems; and 2) enabling knowledge engineering in natural language. Additionally, we formulate key open research questions to tackle these directions.
Β·drops.dagstuhl.deΒ·
Knowledge Engineering Using Large Language Models