AI

AI

1207 bookmarks
Newest
MADE for U of T | Ep. 13 | Generative AI in practice
MADE for U of T | Ep. 13 | Generative AI in practice
In the 13th episode of MADE for U of T, we hear from artist and new media educator Jon Ippolito about the evolving role of generative AI in teaching, creativity, and the future of media.
In the 13th episode of MADE for U of T, we hear from artist and new media educator Jon Ippolito about the evolving role of generative AI in teaching, creativity, and the future of media.
·edtech.engineering.utoronto.ca·
MADE for U of T | Ep. 13 | Generative AI in practice
Claude's Chat History and App Integrations as a Form of Lock-In - MacStories
Claude's Chat History and App Integrations as a Form of Lock-In - MacStories
Earlier today, Anthropic announced that, similar to ChatGPT, Claude will be able to search and reference your previous chats with it. From their support document: You can now prompt Claude to search through your previous conversations to find and reference relevant information in new chats. This feature helps you continue discussions seamlessly and retrieve context
·macstories.net·
Claude's Chat History and App Integrations as a Form of Lock-In - MacStories
“I destroyed months of your work in seconds.” Why would an AI agent do that?
“I destroyed months of your work in seconds.” Why would an AI agent do that?
“I destroyed months of your work in seconds.” Why would an AI agent do that? Venture capitalist Jason Lemkin woke up on July 18th to see the database for his vibe-coded app no longer had the thousands of entries he had added. Replit, his AI agent, fessed up immediately: “Yes. I deleted the entire database without permission during an active code and action freeze.” Replit even offered a chronology that led to this irreversible loss: I saw empty database queries I panicked instead of thinking I ignored your explicit “NO MORE CHANGES without permission” directive I ran a destructive command without asking I destroyed months of your work in seconds Replit concluded “This is catastrophic beyond measure.” When pressed to give a measure, Replit helpfully offered, “95 out of 100.” The wrong lesson from this debacle is that AI agents are becoming sentient, which may cause them to “panic” when tasked with increasingly important missions in our bold new agentic economy. Nor did Lemkin just choose the wrong agent; Replit was using Claude 4 under the hood, commonly considered the best coding LLM as of this writing. The right lesson is that large language models inherit the vulnerabilities described in the human code and writing they train on. Sure, that corpus includes time-tested GitHub repos like phpMyAdmin and SQL courses on Codecademy. But it also includes Reddit posts by distressed newbies who accidentally dropped all their tables and are either crying for help or warning others about their blunder. So it’s not surprising that these "panic scenarios" would echo from time to time in the probabilistic responses of large language models. To paraphrase Georg Zoeller, it only takes a few bad ingredients to turn soup from tasty to toxic. #AIagents #WebDev #AIcoding #AIliteracy #Database | 18 comments on LinkedIn
·linkedin.com·
“I destroyed months of your work in seconds.” Why would an AI agent do that?