Found 30 bookmarks
Newest
ATOM is finally here! A scalable and fast approach that can build and continuously update temporal knowledge graphs, inspired by atomic bonds.
ATOM is finally here! A scalable and fast approach that can build and continuously update temporal knowledge graphs, inspired by atomic bonds.
Alhamdulillah, ATOM is finally here! A scalable and fast approach that can build and continuously update temporal knowledge graphs, inspired by atomic bonds. Just as matter is formed from atoms, and galaxies are formed from stars, knowledge is likely to be formed from atomic knowledge graphs. Atomic knowledge graphs were born from our intention to solve a common problem in LLM-based KG construction methods: exhaustivity and stability. Often, these methods produce unstable KGs that change when rerunning the construction process, even without changing anything. Moreover, they fail to capture all facts in the input documents and usually overlook the temporal and dynamic aspects of real-world data. What is the solution? Atomic facts that are temporally aware. Instead of constructing knowledge graphs from raw documents, we split them into atomic facts, which are self-contained and concise propositions. Temporal atomic KGs are constructed from each atomic fact. Then, we defined how the temporal atomic KGs would be merged at the atomic level and how the temporal aspects would be handled. We designed a binary merge algorithm that combines two TKGs and a parallel merge process that merges all TKGs simultaneously. The entire architecture operates in parallel. ATOM employs dual-time modeling that distinguishes observation time from validity time and has 3 main modules: - Module 1 (Atomic Fact Decomposition) splits input documents observed at time t into atomic facts using LLM-based prompting, where each temporal atomic fact is a short, self-contained snippet that conveys exactly one piece of information. - Module 2 (Atomic TKGs Construction) extracts 5-tuples in parallel from each atomic fact to construct atomic temporal KGs, while embedding nodes and relations and addressing temporal resolution during extraction. - Module 3 (Parallel Atomic Merge) employs a binary merge algorithm to merge pairs of atomic TKGs through iterative pairwise merging in parallel until convergence, with three resolution phases: (1) entity resolution, (2) relation name resolution, and (3) temporal resolution that merges observation and validity time sets for relations with similar (e_s, r_p, e_o). The resulting TKG snapshot is then merged with the previous DTKG to yield the updated DTKG. Results: Empirical evaluations demonstrate that ATOM achieves ~18% higher exhaustivity, ~17% better stability, and over 90% latency reduction compared to baseline methods (including iText2KG), demonstrating strong scalability potential for dynamic TKG construction. Check our ATOM's architecture and code: Preprint Paper: https://lnkd.in/dsJzDaQc Code: https://lnkd.in/drZUyisV Website: (coming soon) Example use cases: (coming soon) Special thanks to the dream team: Ludovic Moncla, Khalid Benabdeslem, RΓ©my Cazabet, Pierre ClΓ©au πŸ“šπŸ“‘ | 14 comments on LinkedIn
ATOM is finally here! A scalable and fast approach that can build and continuously update temporal knowledge graphs, inspired by atomic bonds.
Β·linkedin.comΒ·
ATOM is finally here! A scalable and fast approach that can build and continuously update temporal knowledge graphs, inspired by atomic bonds.
ATOM: AdapTive and OptiMized dynamic temporal knowledge graph construction using LLMs
ATOM: AdapTive and OptiMized dynamic temporal knowledge graph construction using LLMs
βœ… Some state-of-the-art methods for knowledge graph (KG) construction that implement incrementality build a graph from around 3k atomic facts in 4–7 hours, while ATOM achieves the same in just 20 minutes using only 8 parallel threads and a batch size of 40 for asynchronous LLM API calls. ❓ What’s the secret behind this performance? πŸ‘‰ The architecture. The parallel design. ❌ Incrementality in KG construction was key, but it significantly limits scalability. This is because the method must first build the KG and compare it with the previous one before moving on to the next chunk. That’s why we eliminated this in iText2KG. ❓ Why is scalability so important? The short answer: real-time analytics. Fast dynamic TKG construction enables LLMs to reason over them and generate responses instantly, in real time. Discover more secrets behind this parallel architecture by reading the full paper (link in the first comment).
ATOM: AdapTive and OptiMized dynamic temporal knowledge graph construction using LLMs
Β·linkedin.comΒ·
ATOM: AdapTive and OptiMized dynamic temporal knowledge graph construction using LLMs
Integrating Knowledge Graphs into the Debian Ecosystem | Alexander Belikov
Integrating Knowledge Graphs into the Debian Ecosystem | Alexander Belikov
In an era where software systems are increasingly complex and interconnected, effectively managing the relationships between packages, maintainers, dependencies, and vulnerabilities is both a challenge and a necessity. This paper explores the integration of knowledge graphs into the Debian ecosystem as a powerful means to bring structure, semantics, and coherence to diverse sources of package-related data. By unifying information such as package metadata, security advisories, and reproducibility reports into a single graph-based representation, we enable richer visibility into the ecosystem's structure and behavior. Beyond constructing the DebKG graph, we demonstrate how it supports practical, high-impact applications β€” such as tracing vulnerability propagation and identifying gaps between community needs and development activity β€” thereby offering a foundation for smarter, data-informed decision-making within Debian.
Β·alexander-belikov.github.ioΒ·
Integrating Knowledge Graphs into the Debian Ecosystem | Alexander Belikov
Bottom-up Domain-specific Superintelligence: A Reliable Knowledge Graph is What We Need
Bottom-up Domain-specific Superintelligence: A Reliable Knowledge Graph is What We Need
Instead of just pulling facts, the system samples multi-step paths within the graph, such as a causal chain from a disease to a symptom, and translates these paths into natural language reasoning tasks complete with a step-by-step thinking trace
Β·kg-bottom-up-superintelligence.github.ioΒ·
Bottom-up Domain-specific Superintelligence: A Reliable Knowledge Graph is What We Need
The Dataverse Project: 750K FAIR Datasets and a Living Knowledge Graph
The Dataverse Project: 750K FAIR Datasets and a Living Knowledge Graph
"I'm Ukrainian and I'm wearing a suit, so no complaints about me from the Oval Office" - that's the start of my lecture about building Artificial Intelligence with Croissant ML in the Dataverse data platform, for the Bio x AI Hackathon kick-off event in Berlin. https://lnkd.in/ePYHCfJt * 750,000+ FAIR datasets across the world forcing the innovation of the whole data landscape. * A knowledge graph with 50M+ triples. * AI-ready metadata exports. * Qdrant as a vector storage, Google Meta Mistral AI as LLM model providers. * Adrian Gschwend Qlever as fastest triple store for Dataverse knowledge graphs Multilingual, machine-readable, queryable scientific data at scale. If you're interested, you can also apply for the 2-month #BioAgentHack online hackathon: β€’Β $125K+ prizes β€’Β Mentorship from Biotech and AI leaders β€’Β Build alongside top open-science researchers & devs More info: https://lnkd.in/eGhvaKdH
Β·linkedin.comΒ·
The Dataverse Project: 750K FAIR Datasets and a Living Knowledge Graph
SimGRAG is a novel method for knowledge graph driven RAG, transforms queries into graph patterns and aligns them with candidate subgraphs using a graph semantic distance metric
SimGRAG is a novel method for knowledge graph driven RAG, transforms queries into graph patterns and aligns them with candidate subgraphs using a graph semantic distance metric
SimGRAG is a novel method for knowledge graph driven RAG, transforms queries into graph patterns and aligns them with candidate subgraphs using a graph…
SimGRAG is a novel method for knowledge graph driven RAG, transforms queries into graph patterns and aligns them with candidate subgraphs using a graph semantic distance metric
Β·linkedin.comΒ·
SimGRAG is a novel method for knowledge graph driven RAG, transforms queries into graph patterns and aligns them with candidate subgraphs using a graph semantic distance metric
TGB 2.0: A Benchmark for Learning on Temporal Knowledge Graphs and Heterogeneous Graphs
TGB 2.0: A Benchmark for Learning on Temporal Knowledge Graphs and Heterogeneous Graphs
🌟 TGB 2.0 @NeurIPS 2024 🌟 We are very happy to share that our paper TGB 2.0: A Benchmark for Learning on Temporal Knowledge Graphs and Heterogeneous Graphs… | 11 comments on LinkedIn
TGB 2.0: A Benchmark for Learning on Temporal Knowledge Graphs and Heterogeneous Graphs
Β·linkedin.comΒ·
TGB 2.0: A Benchmark for Learning on Temporal Knowledge Graphs and Heterogeneous Graphs
GitHub - Orange-OpenSource/noria-ontology: The NORIA-O project is a data model for IT networks, events and operations information. The ontology is developed using web technologies (e.g. RDF, OWL, SKOS) and is intended as a structure for realizing an IT Service Management (ITSM) Knowledge Graph (KG) for Anomaly Detection (AD) and Risk Management applications. The model has been developed in collaboration with operational teams, and in connection with third parties linked vocabularies.
GitHub - Orange-OpenSource/noria-ontology: The NORIA-O project is a data model for IT networks, events and operations information. The ontology is developed using web technologies (e.g. RDF, OWL, SKOS) and is intended as a structure for realizing an IT Service Management (ITSM) Knowledge Graph (KG) for Anomaly Detection (AD) and Risk Management applications. The model has been developed in collaboration with operational teams, and in connection with third parties linked vocabularies.
The NORIA-O project is a data model for IT networks, events and operations information. The ontology is developed using web technologies (e.g. RDF, OWL, SKOS) and is intended as a structure for rea...
Β·github.comΒ·
GitHub - Orange-OpenSource/noria-ontology: The NORIA-O project is a data model for IT networks, events and operations information. The ontology is developed using web technologies (e.g. RDF, OWL, SKOS) and is intended as a structure for realizing an IT Service Management (ITSM) Knowledge Graph (KG) for Anomaly Detection (AD) and Risk Management applications. The model has been developed in collaboration with operational teams, and in connection with third parties linked vocabularies.
πΊπ‘Ÿπ‘Žπ‘β„ŽπΈπ‘…: 𝐴 π‘†π‘‘π‘Ÿπ‘’π‘π‘‘π‘’π‘Ÿπ‘’-π‘Žπ‘€π‘Žπ‘Ÿπ‘’ 𝑇𝑒π‘₯𝑑-π‘‘π‘œ-πΊπ‘Ÿπ‘Žπ‘β„Ž π‘€π‘œπ‘‘π‘’π‘™ π‘“π‘œπ‘Ÿ 𝐸𝑛𝑑𝑖𝑑𝑦 π‘Žπ‘›π‘‘ π‘…π‘’π‘™π‘Žπ‘‘π‘–π‘œπ‘› 𝐸π‘₯π‘‘π‘Ÿπ‘Žπ‘π‘‘π‘–π‘œπ‘›
πΊπ‘Ÿπ‘Žπ‘β„ŽπΈπ‘…: 𝐴 π‘†π‘‘π‘Ÿπ‘’π‘π‘‘π‘’π‘Ÿπ‘’-π‘Žπ‘€π‘Žπ‘Ÿπ‘’ 𝑇𝑒π‘₯𝑑-π‘‘π‘œ-πΊπ‘Ÿπ‘Žπ‘β„Ž π‘€π‘œπ‘‘π‘’π‘™ π‘“π‘œπ‘Ÿ 𝐸𝑛𝑑𝑖𝑑𝑦 π‘Žπ‘›π‘‘ π‘…π‘’π‘™π‘Žπ‘‘π‘–π‘œπ‘› 𝐸π‘₯π‘‘π‘Ÿπ‘Žπ‘π‘‘π‘–π‘œπ‘›
Our paper "πΊπ‘Ÿπ‘Žπ‘β„ŽπΈπ‘…: 𝐴 π‘†π‘‘π‘Ÿπ‘’π‘π‘‘π‘’π‘Ÿπ‘’-π‘Žπ‘€π‘Žπ‘Ÿπ‘’ 𝑇𝑒π‘₯𝑑-π‘‘π‘œ-πΊπ‘Ÿπ‘Žπ‘β„Ž π‘€π‘œπ‘‘π‘’π‘™ π‘“π‘œπ‘Ÿ 𝐸𝑛𝑑𝑖𝑑𝑦 π‘Žπ‘›π‘‘ π‘…π‘’π‘™π‘Žπ‘‘π‘–π‘œπ‘›β€¦ | 34 comments on LinkedIn
πΊπ‘Ÿπ‘Žπ‘β„ŽπΈπ‘…: 𝐴 π‘†π‘‘π‘Ÿπ‘’π‘π‘‘π‘’π‘Ÿπ‘’-π‘Žπ‘€π‘Žπ‘Ÿπ‘’ 𝑇𝑒π‘₯𝑑-π‘‘π‘œ-πΊπ‘Ÿπ‘Žπ‘β„Ž π‘€π‘œπ‘‘π‘’π‘™ π‘“π‘œπ‘Ÿ 𝐸𝑛𝑑𝑖𝑑𝑦 π‘Žπ‘›π‘‘ π‘…π‘’π‘™π‘Žπ‘‘π‘–π‘œπ‘› 𝐸π‘₯π‘‘π‘Ÿπ‘Žπ‘π‘‘π‘–π‘œπ‘›
Β·linkedin.comΒ·
πΊπ‘Ÿπ‘Žπ‘β„ŽπΈπ‘…: 𝐴 π‘†π‘‘π‘Ÿπ‘’π‘π‘‘π‘’π‘Ÿπ‘’-π‘Žπ‘€π‘Žπ‘Ÿπ‘’ 𝑇𝑒π‘₯𝑑-π‘‘π‘œ-πΊπ‘Ÿπ‘Žπ‘β„Ž π‘€π‘œπ‘‘π‘’π‘™ π‘“π‘œπ‘Ÿ 𝐸𝑛𝑑𝑖𝑑𝑦 π‘Žπ‘›π‘‘ π‘…π‘’π‘™π‘Žπ‘‘π‘–π‘œπ‘› 𝐸π‘₯π‘‘π‘Ÿπ‘Žπ‘π‘‘π‘–π‘œπ‘›