DeepSeek (Chinese: 深度求索; pinyin: Shēndù Qiúsuǒ) is a series of open source Large Language Models (LLMs) from DeepSeek, an AI firm funded solely by Chinese hedge fund High-Flyer based in Hangzhou.
What is the R1 effect in LLM development as of Jan 2025?
The release of DeepSeek R1 in January 2025 has created significant disruption in the LLM landscape, particularly in the realm of reasoning models. Here's a...
Today I added an infinite-nonsense honeypot to my web site just to fuck with LLM scrapers, based on a "spicy autocomplete" program I wrote about 30 years ago. Well-behaved web crawlers will ignore it, but those "AI" people.... well, you know how they are. I'm intentionally not linking to the honeypot from here, for reasons, but I'll bet you can find it pretty easily (and without guessing ...
I thought this was a fascinating post by Simon Willison: Things We Learned About LLMs in 2024
This increase in efficiency and reduction in price is my single favourite trend from 2024. I want the utility of LLMs at a fraction of the energy cost and it looks like that’
A lot has happened in the world of Large Language Models over the course of 2024. Here’s a review of things we figured out about the field in the past …
In How should you adopt LLMs?, we explore how a theoretical ride sharing company,
Theoretical Ride Sharing, should adopt Large Language Models (LLMs).
Part of that strategy’s diagnosis depends on understanding the expected evolution of
the LLM ecosystem, which we’ve build a Wardley map to better explore.
This map of the LLM space is interested in how product companies should address the
proliferation of model providers such as Anthropic, Google and OpenAI,
as well as the proliferation of LLM product patterns like agentic workflows, Retrieval Augmented Generation (RAG),
and running evals to maintain performance as models change.
Vector databases are quite popular right now, especially for building recommendation systems, adding context to chatbots and LLMs, or comparing content based on similarity. In this guide, I'll explain what vector databases are, how they work, and when to use them.
More-than-human aesthetics ⊗ Enchanted knowledge objects in LLM UI ⊗ Native Americans guarded against tyranny
No.333 — With AI, the future of Augmented Reality is in your ears ⊗ Why every company needs a futurist-in-residence ⊗ AI isn’t about unleashing our imaginations ⊗ Bringing life to L.A.’s infrastructure
Arboreal codes ⊗ Conceptual models of space colonization ⊗ AI companies trying to build god
No.329 — It feels like 2004 again ⊗ Three Future Frames ⊗ What is futures literacy ⊗ Cracks in LLMs’ “reasoning” capabilities ⊗ Trees and land absorbed almost no CO2 last year
Uber Creates GenAI Gateway Mirroring OpenAI API to Support Over 60 LLM Use Cases
Uber created a unified platform for serving large language models (LLMs) from external vendors and self-hosted ones and opted to mirror OpenAI API to help with internal adoption. GenAI Gateway provides a consistent and efficient interface and serves over 60 distinct LLM use cases across many areas.
Why Copilot is Making Programmers Worse at Programming
Over the past few years, the evolution of AI-driven tools like GitHub’s Copilot and other large language models (LLMs) has promised to revolutionise programming. By leveraging deep learning, these tools can generate code, suggest solutions, and even troubleshoot issues in real-time, saving developers hours of work. While these tools have obvious benefits in terms of productivity, there’s a growing concern that they may also have unintended consequences on the quality and skillset of programmers.
At Foundation for Economic Education (“The Ego vs. the Machine,” February 24), self-described “techno-optimist” Dylan Allman dismisses recent controversies over AI as a simple matter of wounded egos. “They feel, on some instinctual level, that if machines can do what they do — only better, faster, and more efficiently — then what value do they...
From RAGs to Riches: An In-Depth Look at Retrieval-Augmented Generation
Dive into how Retrieval-Augmented Generation (RAG) can supercharge LLMs. Find out how RAG boosts LLM accuracy and get the scoop on the evolution and limitations of today's LLMs.
Building a Graph RAG System from Scratch with LangChain: A Comprehensive Tutorial – News from generation RAG
Setting up the Development EnvironmentBuilding the Graph RAG SystemIndexing Data in Neo4jImplementing Retrieval and GenerationCode Walkthrough and ExamplesDeploying and Scaling the Graph RAG SystemConclusion and Future Directions Graph RAG (Retrieval Augmented Generation) is an innovative technique that combines the power of knowledge graphs with large language models (LLMs) to enhance the retrieval and generation of