Found 34 bookmarks
Custom sorting
cuGraph and Graph RAG
cuGraph and Graph RAG
**!!!! Great Talk with Bradley Rees NVIDIA RAPIDS cuGraph lead at KDD 24 Conference !!** We had an excellent discussion about the cuGraph user experience in…
cuGraph
·linkedin.com·
cuGraph and Graph RAG
Plan Like a Graph
Plan Like a Graph
An easy trick to improve your LLM results without fine-tuning. Many people know "Few-Shot prompting" or "Chain of Thought prompting". A new (better) method was… | 77 comments on LinkedIn
Plan Like a Graph
·linkedin.com·
Plan Like a Graph
GraphGPT
GraphGPT
🌟GraphGPT🌟 (385 stars in GitHub) is accepted by 🌟SIGIR'24🌟 (only 20.1% acceptance rate)! Thank Yuhao Yang, wei wei, and other co-authors for their precious…
GraphGPT
·linkedin.com·
GraphGPT
Exploring the Potential of Large Language Models in Graph Generation
Exploring the Potential of Large Language Models in Graph Generation
Large language models (LLMs) have achieved great success in many fields, and recent works have studied exploring LLMs for graph discriminative tasks such as node classification. However, the abilities of LLMs for graph generation remain unexplored in the literature. Graph generation requires the LLM to generate graphs with given properties, which has valuable real-world applications such as drug discovery, while tends to be more challenging. In this paper, we propose LLM4GraphGen to explore the ability of LLMs for graph generation with systematical task designs and extensive experiments. Specifically, we propose several tasks tailored with comprehensive experiments to address key questions regarding LLMs' understanding of different graph structure rules, their ability to capture structural type distributions, and their utilization of domain knowledge for property-based graph generation. Our evaluations demonstrate that LLMs, particularly GPT-4, exhibit preliminary abilities in graph generation tasks, including rule-based and distribution-based generation. We also observe that popular prompting methods, such as few-shot and chain-of-thought prompting, do not consistently enhance performance. Besides, LLMs show potential in generating molecules with specific properties. These findings may serve as foundations for designing good LLMs based models for graph generation and provide valuable insights and further research.
·arxiv.org·
Exploring the Potential of Large Language Models in Graph Generation
Why do LangChain and Autogen use graphs? Here are the top reasons
Why do LangChain and Autogen use graphs? Here are the top reasons
LLM frameworks like LangChain are moving towards a graph-based approach for handling their workflows. This represents the initial steps of a much larger… | 90 comments on LinkedIn
Why do LangChain and Autogen use graphs? Here are the top reasons
·linkedin.com·
Why do LangChain and Autogen use graphs? Here are the top reasons
Language, Graphs, and AI in Industry
Language, Graphs, and AI in Industry
Over the past 5 years, news about AI has been filled with amazing research – at first focused on graph neural networks (GNNs) and more recently about large language models (LLMs). Understand that business tends to use connected data – networks, graphs – whether you’re untangling supply networks in Manufacturing, working on drug discovery for Pharma, or mitigating fraud in Finance. Starting from supplier agreements, bill of materials, internal process docs, sales contracts, etc., there’s a graph inside nearly every business process, one that is defined by language. This talk addresses how to leverage both natural language and graph technologies together for AI applications in industry. We’ll look at how LLMs get used to build and augment graphs, and conversely how graph data gets used to ground LLMs for generative AI use cases in industry – where a kind of “virtuous cycle” is emerging for feedback loops based on graph data. Our team has been engaged, on the one hand, with enterprise use cases in manufacturing. On the other hand we’ve worked as intermediaries between research teams funded by enterprise and open source projects needed by enterprise – particularly in the open source ecosystem for AI models. Also, there are caveats; this work is not simple. Translating from latest research into production-ready code is especially complex and expensive. Let’s examine caveats which other teams should understand, and look toward practical examples.
·derwen.ai·
Language, Graphs, and AI in Industry
The Intersection of Graphs and Language Models
The Intersection of Graphs and Language Models
The Intersection of Graphs and Language Models 🔲 ⚫ Large language models (LLMs) have rapidly advanced, displaying impressive abilities in comprehending… | 27 comments on LinkedIn
The Intersection of Graphs and Language Models
·linkedin.com·
The Intersection of Graphs and Language Models
LangGraph: Multi-Agent Workflows
LangGraph: Multi-Agent Workflows
Links * Python Examples * JS Examples * YouTube Last week we highlighted LangGraph - a new package (available in both Python and JS) to better enable creation of LLM workflows containing cycles, which are a critical component of most agent runtimes. As a part of the launch, we highlighted two simple runtimes:
a second set of use cases for langgraph - multi-agent workflows. In this blog we will cover:What does "multi-agent" mean?Why are "multi-agent" workflows interesting?Three concrete examples of using LangGraph for multi-agent workflowsTwo examples of third-party applications built on top of LangGraph using multi-agent workflows (GPT-Newspaper and CrewAI)Comparison to other frameworks (Autogen and CrewAI)
·blog.langchain.dev·
LangGraph: Multi-Agent Workflows
🦜🕸️LangGraph | 🦜️🔗 Langchain
🦜🕸️LangGraph | 🦜️🔗 Langchain
⚡ Building language agents as graphs ⚡
🦜🕸️LangGraph⚡ Building language agents as graphs ⚡Overview​LangGraph is a library for building stateful, multi-actor applications with LLMs, built on top of (and intended to be used with) LangChain. It extends the LangChain Expression Language with the ability to coordinate multiple chains (or actors) across multiple steps of computation in a cyclic manner. It is inspired by Pregel and Apache Beam. The current interface exposed is one inspired by NetworkX.The main use is for adding cycles to your LLM application. Crucially, this is NOT a DAG framework. If you want to build a DAG, you should use just use LangChain Expression Language.Cycles are important for agent-like behaviors, where you call an LLM in a loop, asking it what action to take next.
·python.langchain.com·
🦜🕸️LangGraph | 🦜️🔗 Langchain
pacoid (Paco Xander Nathan)
pacoid (Paco Xander Nathan)
Python open source projects; natural language meets graph technologies; graph topological transformations; graph levels of detail (abstraction layers)
·huggingface.co·
pacoid (Paco Xander Nathan)
Graph Learning Meets Artificial Intelligence
Graph Learning Meets Artificial Intelligence
By request, here are the slides from our #neurips2023 presentation yesterday! We really enjoyed the opportunity to present the different aspects of the work… | 18 comments on LinkedIn
·linkedin.com·
Graph Learning Meets Artificial Intelligence
A Survey of Graph Meets Large Language Model: Progress and Future Directions
A Survey of Graph Meets Large Language Model: Progress and Future Directions
Graph plays a significant role in representing and analyzing complex relationships in real-world applications such as citation networks, social networks, and biological data. Recently, Large Language Models (LLMs), which have achieved tremendous success in various domains, have also been leveraged in graph-related tasks to surpass traditional Graph Neural Networks (GNNs) based methods and yield state-of-the-art performance. In this survey, we first present a comprehensive review and analysis of existing methods that integrate LLMs with graphs. First of all, we propose a new taxonomy, which organizes existing methods into three categories based on the role (i.e., enhancer, predictor, and alignment component) played by LLMs in graph-related tasks. Then we systematically survey the representative methods along the three categories of the taxonomy. Finally, we discuss the remaining limitations of existing studies and highlight promising avenues for future research. The relevant papers are summarized and will be consistently updated at: https://github.com/yhLeeee/Awesome-LLMs-in-Graph-tasks.
·arxiv.org·
A Survey of Graph Meets Large Language Model: Progress and Future Directions