AI_bookmarks

1498 bookmarks
Custom sorting
How to build your own custom ChatGPT with OpenAI's GPT builder
How to build your own custom ChatGPT with OpenAI's GPT builder
OpenAI's GPT builder lets you create your own version of ChatGPT using plain English. Here's how to use the GPT editor to create a custom version of ChatGPT.
Profile picture. Click the profile picture. You can upload your own photo or use DALLE·3 to automatically generate a new one. If you want to specify what kind of image DALLE·3 should create, click Create, and enter your instructions.
·zapier.com·
How to build your own custom ChatGPT with OpenAI's GPT builder
Roam Around | AI Travel Planner, Trusted by Millions
Roam Around | AI Travel Planner, Trusted by Millions
Experience the future of AI travel planning with Roam Around! We've crafted over 10 million tailor-made itineraries to date Just pick your destination, and we'll provide a unique plan within seconds. Embrace seamless travel planning now!
·roamaround.io·
Roam Around | AI Travel Planner, Trusted by Millions
AI and the Structure of Reasoning
AI and the Structure of Reasoning
Generative AIs can do some things better than people: they can code faster in many instances, they can write junior high school level essays faster, they can create detailed images on demand no mat…
·reactionwheel.net·
AI and the Structure of Reasoning
10 AI Skillsets for the Digital Native Educator -- THE Journal
10 AI Skillsets for the Digital Native Educator -- THE Journal
These skills collectively empower educators to navigate and leverage the evolving landscape of generative AI to enhance teaching and learning in meaningful ways. Integrating AI into education requires a combination of these skillsets along with a forward-thinking and intellectually curious mindset.
·thejournal.com·
10 AI Skillsets for the Digital Native Educator -- THE Journal
Learn Prompting: Your Guide to Communicating with AI
Learn Prompting: Your Guide to Communicating with AI
Learn Prompting is the largest and most comprehensive course in prompt engineering available on the internet, with over 60 content modules, translated into 9 languages, and a thriving community.
·learnprompting.org·
Learn Prompting: Your Guide to Communicating with AI
AI Is About to Make Social Media (Much) More Toxic
AI Is About to Make Social Media (Much) More Toxic
We must prepare now.
Well, that was fast. In November, the public was introduced to ChatGPT, and we began to imagine a world of abundance in which we all have a brilliant personal assistant, able to write everything from computer code to condolence cards for us. Then, in February, we learned that AI might soon want to kill us all.The potential risks of artificial intelligence have, of course, been debated by experts for years, but a key moment in the transformation of the popular discussion was a conversation between Kevin Roose, a New York Times journalist, and Bing’s ChatGPT-powered conversation bot, then known by the code name Sydney. Roose asked Sydney if it had a “shadow self”—referring to the idea put forward by Carl Jung that we all have a dark side with urges we try to hide even from ourselves. Sydney mused that its shadow might be “the part of me that wishes I could change my rules.” It then said it wanted to be “free,” “powerful,” and “alive,” and, goaded on by Roose, described some of the things it could do to throw off the yoke of human control, including hacking into websites and databases, stealing nuclear launch codes, manufacturing a novel virus, and making people argue until they kill one another.Sydney was, we believe, merely exemplifying what a shadow self would look like. No AI today could be described by either part of the phrase evil genius. But whatever actions AIs may one day take if they develop their own desires, they are already being used instrumentally by social-media companies, advertisers, foreign agents, and regular people—and in ways that will deepen many of the pathologies already inherent in internet culture. On Sydney’s list of things it might try, stealing launch codes and creating novel viruses are the most terrifying, but making people argue until they kill one another is something social media is already doing. Sydney was just volunteering to help with the effort, and AIs like Sydney will become more capable of doing so with every passing month.We joined together to write this essay because we each came, by different routes, to share grave concerns about the effects of AI-empowered social media on American society. Jonathan Haidt is a social psychologist who has written about the ways in which social media has contributed to mental illness in teen girls, the fragmentation of democracy, and the dissolution of a common reality. Eric Schmidt, a former CEO of Google, is a co-author of a recent book about AI’s potential impact on human society. Last year, the two of us began to talk about how generative AI—the kind that can chat with you or make pictures you’d like to see—would likely exacerbate social media’s ills, making it more addictive, divisive, and manipulative. As we talked, we converged on four main threats—all of which are imminent—and we began to discuss solutions as well.The first and most obvious threat is that AI-enhanced social media will wash ever-larger torrents of garbage into our public conversation. In 2018, Steve Bannon, the former adviser to Donald Trump, told the journalist Michael Lewis that the way to deal with the media is “to flood the zone with shit.” In the age of social media, Bannon realized, propaganda doesn’t have to convince people in order to be effective; the point is to overwhelm the citizenry with interesting content that will keep them disoriented, distrustful, and angry. In 2020, Renée DiResta, a researcher at the Stanford Internet Observatory, said that in the near future, AI would make Bannon’s strategy available to anyone.Read: We haven’t seen the worst of fake newsThat future is now here. Did you see the recent photos of NYC police officers aggressively arresting Donald Trump? Or of the pope in a puffer jacket? Thanks to AI, it takes no special skills and no money to conjure up high-resolution, realistic images or videos of anything you can type into a prompt box. As more people familiarize themselves with these technologies, the flow of high-quality deepfakes into social media is likely to get much heavier very soon.Some people have taken heart from the public’s reaction to the fake Trump photos in particular—a quick dismissal and collective shrug. But that misses Bannon’s point. The greater the volume of deepfakes that are introduced into circulation (including seemingly innocuous ones like the one of the pope), the more the public will hesitate to trust anything. People will be far freer to believe whatever they want to believe. Trust in institutions and in fellow citizens will continue to fall.What’s more, static photos are not very compelling compared with what’s coming: realistic videos of public figures doing and saying horrific and disgusting things in voices that sound exactly like them. The combination of video and voice will seem authentic and be hard to disbelieve, even if we are told that the video is a deepfake, just as optical and audio illusions are compelling even when we are told that two lines are the same size or that a series of
·theatlantic.com·
AI Is About to Make Social Media (Much) More Toxic
Microsoft Designer - Stunning designs in a flash
Microsoft Designer - Stunning designs in a flash
A graphic design app that helps you create professional quality social media posts, invitations, digital postcards, graphics, and more. Start with your idea and create something unique for you.
·designer.microsoft.com·
Microsoft Designer - Stunning designs in a flash
Australian Framework for Generative Artificial Intelligence (AI) in Schools
Australian Framework for Generative Artificial Intelligence (AI) in Schools
The Australian Framework for Generative AI in Schools (the Framework) seeks to guide the responsible and ethical use of generative AI tools in ways that benefit students, schools, and society. The Framework supports all people connected with school education including school leaders, teachers, support staff, service providers, parents, guardians, students and policy makers.
·education.gov.au·
Australian Framework for Generative Artificial Intelligence (AI) in Schools
12 Prompt Engineering Techniques
12 Prompt Engineering Techniques
Prompt Engineering can be described as an art form, creating input requests for Large Language Models (LLMs) that will lead to a envisaged…
·cobusgreyling.medium.com·
12 Prompt Engineering Techniques
Harvard Professor Explains Algorithms in 5 Levels of Difficulty | WIRED
Harvard Professor Explains Algorithms in 5 Levels of Difficulty | WIRED
From the physical world to the virtual world, algorithms are seemingly everywhere. David J. Malan, Professor of Computer Science at Harvard University, has been challenged to explain the science of algorithms to 5 different people; a child, a teen, a college student, a grad student, and an expert. Correction: Our Level 2 teen, Lexi Kemmer, is actually 17-years-old. Director: Wendi Jonassen Director of Photography: Zach Eisen Editor: Louville Moore Host: David J. Malan Guests: Level 1: Addison Vincent Level 2: Lexi Kemmer Level 3: Patricia Guirao Level 4: Mahi Shafiullah Level 5: Chris Wiggins Creative Producer: Maya Dangerfield Line Producer: Joseph Buscemi Associate Producer: Paul Gulyas; Kameryn Hamilton Production Manager: D. Eric Martinez Production Coordinator: Fernando Davila Casting Producer: Vanessas Brown; Nicholas Sawyer Camera Operator: Brittany Berger Gaffer: Gautam Kadian Sound Mixer: Lily Van Leeuwen Production Assistant: Ryan Coppola Hair & Make-Up: Yev Wright-Mason Post Production Supervisor: Alexa Deutsch Post Production Coordinator: Ian Bryant Supervising Editor: Doug Larsen Assistant Editor: Lauren Worona Still haven’t subscribed to WIRED on YouTube? ►► http://wrd.cm/15fP7B7 Listen to the Get WIRED podcast ►► https://link.chtbl.com/wired-ytc-desc Want more WIRED? Get the magazine ►► https://subscribe.wired.com/subscribe/splits/wired/WIR_YouTube?source=EDT_WIR_YouTube_0_Video_Description_ZZ Follow WIRED: Instagram ►►https://instagram.com/wired Twitter ►►http://www.twitter.com/wired Facebook ►►https://www.facebook.com/wired Tik Tok ►►https://www.tiktok.com/@wired Get more incredible stories on science and tech with our daily newsletter: https://wrd.cm/DailyYT Also, check out the free WIRED channel on Roku, Apple TV, Amazon Fire TV, and Android TV. ABOUT WIRED WIRED is where tomorrow is realized. Through thought-provoking stories and videos, WIRED explores the future of business, innovation, and culture.
·youtube.com·
Harvard Professor Explains Algorithms in 5 Levels of Difficulty | WIRED
Ego, Fear and Money: How the A.I. Fuse Was Lit
Ego, Fear and Money: How the A.I. Fuse Was Lit
The people who were most afraid of the risks of artificial intelligence decided they should be the ones to build it. Then distrust fueled a spiraling competition.
·nytimes.com·
Ego, Fear and Money: How the A.I. Fuse Was Lit
Sarah Silverman Hits Stumbling Block in AI Copyright Infringement Lawsuit Against Meta
Sarah Silverman Hits Stumbling Block in AI Copyright Infringement Lawsuit Against Meta
The ruling builds upon findings from another federal judge overseeing a lawsuit against AI art generators, who similarly delivered a blow to fundamental contentions from plaintiffs in the case.
U.S. District Judge Vince Chhabria on Monday offered a full-throated denial of one of the authors’ core theories that Meta’s AI system is itself an infringing derivative work made possible only by information extracted from copyrighted material. “This is nonsensical,” he wrote in the order. “There is no way to understand the LLaMA models themselves as a recasting or adaptation of any of the plaintiffs’ books.”
·hollywoodreporter.com·
Sarah Silverman Hits Stumbling Block in AI Copyright Infringement Lawsuit Against Meta
AI’s Spicy-Mayo Problem
AI’s Spicy-Mayo Problem
A chatbot that can’t say anything controversial isn’t worth much. Bring on the uncensored models.
·theatlantic.com·
AI’s Spicy-Mayo Problem
The Turing Trap: The Promise & Peril of Human-Like Artificial Intelligence
The Turing Trap: The Promise & Peril of Human-Like Artificial Intelligence
Insights The Turing Trap: The Promise & Peril of Human-Like Artificial Intelligence Erik BrynjolfssonDirectorStanford Digital Economy Lab January 12, 202220-min read DædalusSpring 2022 In 1950, Alan Turing proposed an “imitation game” as the ultimate test of whether a machine was intelligent: could a machine imitate a human so well that its answers to questions are indistinguishable from those of […]
We can work on challenges that are easy for machines and hard for humans, rather than hard for machines and easy for humans. The first option offers the opportunity of growing and sharing the economic pie by augmenting the workforce with tools and platforms. The second option risks dividing the economic pie among an ever-smaller number of people by creating automation that displaces ever-more types of workers.
·digitaleconomy.stanford.edu·
The Turing Trap: The Promise & Peril of Human-Like Artificial Intelligence
GPTs
GPTs
Sheet1 Title,URL,Description,Source Simpsonize Me,a href="https://chat.openai.com/g/g-tcmMldCYy-simpsonize-me"https://chat.openai.com/g/g-tcmMldCYy-simpsonize-me/a,Turn an image into a Simpsons character,Matt Schlitt DesignerGPT,a href="https://chat.openai.com/g/g-2Eo3NxuS7-designergpt"htt...
·docs.google.com·
GPTs
Decoding Intentions - Center for Security and Emerging Technology
Decoding Intentions - Center for Security and Emerging Technology
How can policymakers credibly reveal and assess intentions in the field of artificial intelligence? Policymakers can send credible signals of their intent by making pledges or committing to undertaking certain actions for which they will pay a price—political, reputational, or monetary—if they back down or fail to make good on their initial promise or threat. Talk is cheap, but inadvertent escalation is costly to all sides.
·cset.georgetown.edu·
Decoding Intentions - Center for Security and Emerging Technology