Found 511 bookmarks
Newest
turbopuffer: fast search on object storage
turbopuffer: fast search on object storage
turbopuffer is a vector database built on top of object storage, which means 10x-100x cheaper, usage-based pricing, and massive scalability
·turbopuffer.com·
turbopuffer: fast search on object storage
Extrinsic Hallucinations in LLMs
Extrinsic Hallucinations in LLMs
Hallucination in large language models usually refers to the model generating unfaithful, fabricated, inconsistent, or nonsensical content. As a term, hallucination has been somewhat generalized to cases when the model makes mistakes. Here, I would like to narrow down the problem of hallucination to be when the model output is fabricated and not grounded by either the provided context or world knowledge. There are two types of hallucination: In-context hallucination: The model output should be consistent with the source content in context.
·lilianweng.github.io·
Extrinsic Hallucinations in LLMs
How to self-host and hyperscale AI with Nvidia NIM
How to self-host and hyperscale AI with Nvidia NIM
Try out Nvidia NIM in the free playground https://nvda.ws/4avifodLearn how to build a futuristic workforce of AI agents, then self-host and scale them for an...
·youtube.com·
How to self-host and hyperscale AI with Nvidia NIM
Consistency Large Language Models: A Family of Efficient Parallel Decoders
Consistency Large Language Models: A Family of Efficient Parallel Decoders
TL;DR: LLMs have been traditionally regarded as sequential decoders, decoding one token after another. In this blog, we show pretrained LLMs can be easily taught to operate as efficient parallel decoders. We introduce Consistency Large Language Models (CLLMs), a new family of parallel decoders capable of reducing inference latency by efficiently decoding an $n$-token sequence per inference step. Our research shows this process – mimicking human cognitive process of forming complete sentences in mind before articulating word by word – can be effectively learned by simply finetuning pretrained LLMs.
·hao-ai-lab.github.io·
Consistency Large Language Models: A Family of Efficient Parallel Decoders
The AI Backend
The AI Backend
The AI Backend * work in progress, please provide feedback so we can improve Just like in 1995 it was obvious that every business needs an internet presence to stay competitive, in 2024 it's obvious that every software needs intelligence to stay competitive. Software products generally have 3 c...
·docs.google.com·
The AI Backend
The Surprising Power of Next Word Prediction: Large Language Models Explained, Part 1 | Center for Security and Emerging Technology
The Surprising Power of Next Word Prediction: Large Language Models Explained, Part 1 | Center for Security and Emerging Technology
Large language models (LLMs), the technology that powers generative artificial intelligence (AI) products like ChatGPT or Google Gemini, are often thought of as chatbots that predict the next word. But that isn't the full story of what LLMs are and how they work. This is the first blog post in a three-part series explaining some key elements of how LLMs function. This blog post covers pre-training—the process by which LLMs learn to predict the next word—and why it’s so surprisingly powerful.
·cset.georgetown.edu·
The Surprising Power of Next Word Prediction: Large Language Models Explained, Part 1 | Center for Security and Emerging Technology
A Survey of Techniques for Maximizing LLM Performance
A Survey of Techniques for Maximizing LLM Performance
Join us for a comprehensive survey of techniques designed to unlock the full potential of Language Model Models (LLMs). Explore strategies such as fine-tunin...
·youtube.com·
A Survey of Techniques for Maximizing LLM Performance
OpenAI Platform
OpenAI Platform
Explore developer resources, tutorials, API docs, and dynamic examples to get the most out of OpenAI's platform.
·platform.openai.com·
OpenAI Platform
metaskills/experts at labnotes.org
metaskills/experts at labnotes.org
Experts.js is the easiest way to create and deploy OpenAI's Assistants and link them together as Tools to create advanced Multi AI Agent Systems with expanded memory and attention to detail. - ...
·github.com·
metaskills/experts at labnotes.org
RAG - jxnl.co
RAG - jxnl.co
Notes about my hobbies and other things I find interesting.
·jxnl.co·
RAG - jxnl.co
Why I'm Staying Away from Crew AI: My Honest Opinion
Why I'm Staying Away from Crew AI: My Honest Opinion
Crew AI is not suitable for production use cases. I’ll be going through why I believe this is the case and what you should do instead when building your own ...
·youtube.com·
Why I'm Staying Away from Crew AI: My Honest Opinion