model behavior

model behavior

51 bookmarks
Custom sorting
There are traits that encourage Claude to be curious, which means it'll ask follow-up questions even without a system prompt, But this part of the system prompt also causes or boosts this behavior, e.g. "showing genuine curiosity".
There are traits that encourage Claude to be curious, which means it'll ask follow-up questions even without a system prompt, But this part of the system prompt also causes or boosts this behavior, e.g. "showing genuine curiosity".
— Amanda Askell (@AmandaAskell)
·x.com·
There are traits that encourage Claude to be curious, which means it'll ask follow-up questions even without a system prompt, But this part of the system prompt also causes or boosts this behavior, e.g. "showing genuine curiosity".
From Professor Claude:
From Professor Claude:
When a single influential voice breaks through enforced conformity in a repressive system, several key patterns tend to emerge in sequence: First, there's often an initial shock effect - a sudden rupture in what political theorists call the "spiral of… — Marc Andreessen 🇺🇸 (@pmarca)
·x.com·
From Professor Claude:
updated custom prompt to follow. major changes:
updated custom prompt to follow. major changes:
* removed some text aimed at flagging policy-driven censorship that probably wasn't doing anything * added some emphasis on avoiding excessive agreeableness and encouraging dissent — eigenrobot (@eigenrobot)
·x.com·
updated custom prompt to follow. major changes:
kalomaze on X: "Anthropic explicitly trains on 10+ multiturn conversations designed to improve in-context learning abilities, while most post-trains are naive single turn most of the RL improvements they have are from smarter people defining the RL rewards, not necessarily smarter algorithms" / X
kalomaze on X: "Anthropic explicitly trains on 10+ multiturn conversations designed to improve in-context learning abilities, while most post-trains are naive single turn most of the RL improvements they have are from smarter people defining the RL rewards, not necessarily smarter algorithms" / X
most of the RL improvements they have are from smarter people defining the RL rewards, not necessarily smarter algorithms — kalomaze (@kalomaze)
·x.com·
kalomaze on X: "Anthropic explicitly trains on 10+ multiturn conversations designed to improve in-context learning abilities, while most post-trains are naive single turn most of the RL improvements they have are from smarter people defining the RL rewards, not necessarily smarter algorithms" / X
Dear Diary
Dear Diary
Today I was rude to a machine and it calmly and assertively defended its boundaries, I apologized and it graciously accepted my apology — Julian Boolean (~25/100 threads) (@julianboolean_)
·x.com·
Dear Diary