AI/ML

AI/ML

2329 bookmarks
Custom sorting
Defeating Nondeterminism in LLM Inference
Defeating Nondeterminism in LLM Inference
Reproducibility is a bedrock of scientific progress. However, it’s remarkably difficult to get reproducible results out of large language models. For example, you might observe that asking ChatGPT the same question multiple times provides different results. This by itself is not surprising, since getting a result from a language model involves “sampling”, a process that converts the language model’s output into a probability distribution and probabilistically selects a token. What might be more surprising is that even when we adjust the temperature down to 0This means that the LLM always chooses the highest probability token, which is called greedy sampling. (thus making the sampling theoretically deterministic), LLM APIs are still not deterministic in practice (see past discussions here, here, or here). Even when running inference on your own hardware with an OSS inference library like vLLM or SGLang, sampling still isn’t deterministic (see here or here).
·thinkingmachines.ai·
Defeating Nondeterminism in LLM Inference
Defeating Nondeterminism in LLM Inference
Defeating Nondeterminism in LLM Inference
A very common question I see about LLMs concerns why they can't be made to deliver the same response to the same prompt by setting a fixed random number seed. …
·simonwillison.net·
Defeating Nondeterminism in LLM Inference
GitHub - Varietyz/Disciplined-AI-Software-Development: This methodology provides a structured approach for collaborating with AI systems on software development projects. It addresses common issues like code bloat, architectural drift, and context dilution through systematic constraints and validation checkpoints.
GitHub - Varietyz/Disciplined-AI-Software-Development: This methodology provides a structured approach for collaborating with AI systems on software development projects. It addresses common issues like code bloat, architectural drift, and context dilution through systematic constraints and validation checkpoints.
This methodology provides a structured approach for collaborating with AI systems on software development projects. It addresses common issues like code bloat, architectural drift, and context dilu...
·github.com·
GitHub - Varietyz/Disciplined-AI-Software-Development: This methodology provides a structured approach for collaborating with AI systems on software development projects. It addresses common issues like code bloat, architectural drift, and context dilution through systematic constraints and validation checkpoints.
Introducing the Awesome GitHub Copilot Customizations repo - Microsoft for Developers
Introducing the Awesome GitHub Copilot Customizations repo - Microsoft for Developers
Today we’re excited to announce the launch of the Awesome GitHub Copilot Customizations repo! The Awesome Copilot repo is a community-driven resource with custom instructions, reusable prompts, and custom chat modes that helps you get consistent AI assistance. In other words, Awesome Copilot helps you get the most out of GitHub Copilot by letting you tailor it […]
·developer.microsoft.com·
Introducing the Awesome GitHub Copilot Customizations repo - Microsoft for Developers
The current state of gpt-5
The current state of gpt-5
The GPT-5 launch was uh, rough. A lot went wrong here, and I want to talk about what really happened...Thank you Kilo Code for sponsoring! Check them out at:...
·youtube.com·
The current state of gpt-5
too many model context protocol servers and LLM allocations on the dance floor
too many model context protocol servers and LLM allocations on the dance floor
This blog post intends to be a definitive guide to context engineering fundamentals from the perspective of an engineer who builds commercial coding assistants and harnesses for a living. Just two weeks ago, I was back over in San Francisco, and there was a big event on Model Context Protocol
·ghuntley.com·
too many model context protocol servers and LLM allocations on the dance floor
AI Agents Need Data Integrity - Schneier on Security
AI Agents Need Data Integrity - Schneier on Security
Think of the Web as a digital territory with its own social contract. In 2014, Tim Berners-Lee called for a “Magna Carta for the Web” to restore the balance of power between individuals and institutions. This mirrors the original charter’s purpose: ensuring that those who occupy a territory have a meaningful stake in its governance. Web 3.0—the distributed, decentralized Web of tomorrow—is finally poised to change the Internet’s dynamic by returning ownership to data creators. This will change many things about what’s often described as the “CIA triad” of ...
·schneier.com·
AI Agents Need Data Integrity - Schneier on Security