Reproducibility is a bedrock of scientific progress. However, it’s remarkably difficult to get reproducible results out of large language models.
For example, you might observe that asking ChatGPT the same question multiple times provides different results. This by itself is not surprising, since getting a result from a language model involves “sampling”, a process that converts the language model’s output into a probability distribution and probabilistically selects a token.
What might be more surprising is that even when we adjust the temperature down to 0This means that the LLM always chooses the highest probability token, which is called greedy sampling. (thus making the sampling theoretically deterministic), LLM APIs are still not deterministic in practice (see past discussions here, here, or here). Even when running inference on your own hardware with an OSS inference library like vLLM or SGLang, sampling still isn’t deterministic (see here or here).
Agentic Design Patterns A Hands-On Guide to Building Intelligent Systems, Antonio Gulli Table of Contents - total 424 pages = 1+2+1+1+4+9+103+61+34+114+74+5+4 11 Dedication, 1 page Acknowledgment, 2 pages [final, last read done] Foreword, 1 page [final, last read done] A Thought Leader's ...
CaMeL offers a promising new direction for mitigating prompt injection attacks
In the two and a half years that we’ve been talking about prompt injection attacks I’ve seen alarmingly little progress towards a robust solution. The new paper Defeating Prompt Injections …
I may be late to the party, but LangGraph lets you build complex workflow architectures and codify them as powerful automations. Also LLMs, if you want. But you don’t have to!
Posts in the 'LLM from scratch' category on Giles Thomas’s blog. Insights on AI, startups, software development, and technical projects, drawn from 30 years of experience.
Enhanced Agentic-RAG: What If Chatbots Could Deliver Near-Human Precision? | Uber Blog
Genie is Uber’s internal on-call copilot, designed to provide real-time support for thousands of queries across multiple help channels in Slack®. It enables users to receive prompt responses with proper citations from Uber’s internal documentation. It also improves the productivity of on-call engineers and subject matter experts (SMEs) by reducing the effort required to address common, ad-hoc queries. While Genie streamlines the development of an LLM-powered on-call Slack bot, ensuring the accuracy and relevance of its responses remains a significant challenge. This blog details our efforts to improve Genie’s answer quality to near-human precision, allowing SMEs to rely on it for most queries without concern over potential misinformation in the engineering security and privacy domain.
LLM function calls don't scale; code orchestration is simpler, more effective.
One common practice for working with MCP tools calls is to put the outputs from a tool back into the LLM as a message, and ask the LLM for the next step. ...
Check out Recraft’s newest featues for free – https://go.recraft.ai/codingsloth. Use my code SLOTH11 for $11 off any paid plan.
AI Agents are kinda crazy ngl.
// SLOTH ARTISTS //
Pixel Art Sloth: https://www.behance.net/harveydentmustdie
// NEWSLETTER //
Sloth Bytes: https://slothbytes.beehiiv.com/subscribe
// BUSINESS INQUIRIES //
For business: thecodingsloth@smoothmedia.co
For brand partnerships: https://tally.so/r/mZVvKa
// SOCIALS //
Twitter: https://twitter.com/TheCodingSloth1
TikTok: https://www.tiktok.com/@thecodingsloth
Discord: https://discord.gg/2ByMHqTNca
// TOOLS/THINGS I REALLY LIKE //
If you wanna build 10x developer level projects check out CodeCrafters
https://app.codecrafters.io/join?via=TheCodingSloth
If you want to build an awesome newsletter like Sloth Bytes I use beehiiv
https://www.beehiiv.com?via=the-coding-sloth
If you want to make nice looking apps: https://mobbin.com/?via=sloth
(some of these links are affiliates, so I'll earn some money which supports the channel!)
Learn how to design, build and ship AI agents with this interactive step by step guide in 2025. We also have resources and short descriptions attached to the roadmap items so you can get everything you want to learn in one place.