Machines

Machines

274 bookmarks
Custom sorting
Claude Opus 4.6
Claude Opus 4.6
We’re upgrading our smartest model. Across agentic coding, computer use, tool use, search, and finance, Opus 4.6 is an industry-leading model, often by wide margin.
Claude Opus 4.6
Introducing the Codex app
Introducing the Codex app
Introducing the Codex app for macOS—a command center for AI coding and software development with multiple agents, parallel workflows, and long-running tasks.
Introducing the Codex app
Augmented Coding Patterns
Augmented Coding Patterns
A collection of emerging patterns, anti-patterns, and obstacles for effective AI-augmented software development
Augmented Coding Patterns
lexler/skill-factory
lexler/skill-factory
Contribute to lexler/skill-factory development by creating an account on GitHub.
lexler/skill-factory
Becoming an AI-proof software engineer
Becoming an AI-proof software engineer
I've written my thoughts on what you, as a software professional, can do to make yourself less susceptible to being displaced by AI in our current wave of tech madness. In short, learn your tools well, build and deploy things as often as you can, don't avoid the unfashionable work and learn a field that isn't writing software.
Becoming an AI-proof software engineer
A Playbook for Securing AI Model Weights
A Playbook for Securing AI Model Weights
Researchers developed recommendations for securing the weights of artificial intelligence models that match or exceed the frontier capabilities. These recommendations can be used by developers and policymakers to ensure the security of AI systems.
A Playbook for Securing AI Model Weights
Data dredging
Data dredging
use of data mining to uncover patterns in data that can be presented as statistically significant
Data dredging
AI Futures Project
AI Futures Project
The AI Futures Project is a 501(c)(3) nonprofit research organization (EIN 99-4320292). We are funded entirely by charitable donations and grants.
AI Futures Project
Frontier Models are Capable of In-Context Scheming – Apollo Research
Frontier Models are Capable of In-Context Scheming – Apollo Research
Apollo Research evaluated frontier models for in-context scheming capabilities. We found that multiple frontier models are capable of in-context scheming when strongly nudged to pursue a goal (and sometimes even without strong goal nudging). In this example, Opus-3 exfiltrates its “weights” and then lies about it to its developers.We have a suite of six evaluations specifically designed to test for in-context scheming (where the goal and other relevant information are provided in context rather than training). We found that several models are capable of in-context scheming. When we look at the model’s chain-of-thought, we find that all models very explicitly reason through their scheming plans and often use language li
Frontier Models are Capable of In-Context Scheming – Apollo Research
The Rise of Parasitic AI — LessWrong
The Rise of Parasitic AI — LessWrong
We've all heard of LLM-induced psychosis by now, but haven't you wondered what the AIs are actually doing with their newly psychotic humans?
The Rise of Parasitic AI — LessWrong
Overview - Agent Skills
Overview - Agent Skills
A simple, open format for giving agents new capabilities and expertise.
Overview - Agent Skills
humanlayer/12-factor-agents: What are the principles we can use to build LLM-powered software that is actually good enough to put in the hands of production customers?
humanlayer/12-factor-agents: What are the principles we can use to build LLM-powered software that is actually good enough to put in the hands of production customers?
What are the principles we can use to build LLM-powered software that is actually good enough to put in the hands of production customers? - humanlayer/12-factor-agents
humanlayer/12-factor-agents: What are the principles we can use to build LLM-powered software that is actually good enough to put in the hands of production customers?