Research

112 bookmarks
Custom sorting
Who on Earth Is Using Generative AI ? (English)
Who on Earth Is Using Generative AI ? (English)
Leveraging unconventional data, including website traffic data and Google Trends, this paper unveils the real-time usage patterns of generative artificial intelligence .
·documents.worldbank.org·
Who on Earth Is Using Generative AI ? (English)
A.I. Chatbots Defeated Doctors at Diagnosing Illness
A.I. Chatbots Defeated Doctors at Diagnosing Illness
A small study found ChatGPT outdid human physicians when assessing medical case histories, even when those doctors were using a chatbot.
Dr. Chen said he noticed that when he peered into the doctors’ chat logs, “they were treating it like a search engine for directed questions: ‘Is cirrhosis a risk factor for cancer? What are possible diagnoses for eye pain?’”“It was only a fraction of the doctors who realized they could literally copy-paste in the entire case history into the chatbot and just ask it to give a comprehensive answer to the entire question,” Dr. Chen added.“Only a fraction of doctors actually saw the surprisingly smart and comprehensive answers the chatbot was capable of producing.”
·nytimes.com·
A.I. Chatbots Defeated Doctors at Diagnosing Illness
GPT-Poetry.pdf
GPT-Poetry.pdf
·efaidnbmnnnibpcajpcglclefindmkaj·
GPT-Poetry.pdf
Simple techniques to bypass GenAI text detectors: implications for inclusive education - International Journal of Educational Technology in Higher Education
Simple techniques to bypass GenAI text detectors: implications for inclusive education - International Journal of Educational Technology in Higher Education
This study investigates the efficacy of six major Generative AI (GenAI) text detectors when confronted with machine-generated content modified to evade detection (n = 805). We compare these detectors to assess their reliability in identifying AI-generated text in educational settings, where they are increasingly used to address academic integrity concerns. Results show significant reductions in detector accuracy (17.4%) when faced with simple techniques to manipulate the AI generated content. The varying performances of GenAI tools and detectors indicate they cannot currently be recommended for determining academic integrity violations due to accuracy limitations and the potential for false accusation which undermines inclusive and fair assessment practices. However, these tools may support learning and academic integrity when used non-punitively. This study aims to guide educators and institutions in the critical implementation of AI text detectors in higher education, highlighting the importance of exploring alternatives to maintain inclusivity in the face of emerging technologies.
By mimicking these imperfections, AI-generated content can effectively mislead detectors into classifying them as human-authored content.
·educationaltechnologyjournal.springeropen.com·
Simple techniques to bypass GenAI text detectors: implications for inclusive education - International Journal of Educational Technology in Higher Education
Black Students Are More Likely to Be Falsely Accused of Using AI to Cheat
Black Students Are More Likely to Be Falsely Accused of Using AI to Cheat
Report notes why this is a problem that educators need to pay closer attention to.
Black students are more than twice as likely as their white or Hispanic peers to have their writing incorrectly flagged as the work of artificial intelligence tools, concludes a report released Sept. 18 by Common Sense Media, a nonprofit that examines the impact of technology on young people.Overall, about 10 percent of teens of any background said they had their work inaccurately identified as generated by an AI tool, Common Sense found. But 20 percent of Black teens were falsely accused of using AI to complete an assignment, compared with 7 percent of white and 10 percent of Latino teens. See Also Classroom Technology Should It Stay or Should It Go? Schools Trim Number of Tech Tools They Use Ed-tech leaders are culling the wide variety of digital tools teachers embraced over the past two years. This may be at least partially due to flaws in AI detection software. About 79 percent of teens who had their assignments incorrectly flagged by a teacher also had their work submitted to AI detection software, while 27 percent said their work had not been submitted.AI detection software has already been shown to have problematic biases, even though secondary school teachers commonly use the technology.More than two-thirds—68 percent—of teachers report using an AI detection tool regularly, according to a survey of 460 6th to 12th grade public school teachers conducted for the Center for Democracy & Technology, a nonprofit organization that aims to shape technology policy.But the tools often reflect societal biases. Researchers ran essays written by Chinese students for the Test of English as a Foreign Language, or TOEFL, through seven widely-used detectors. They did the same with a sample of essays written by U.S. 8th graders who were native English speakers. The tools incorrectly labeled more than half of the TOEFL essays as AI-generated, while accurately classifying the 8th grade essays as human-crafted.Common Sense Media’s findings on Black students could be due to either unfairness in AI detection tools or biases in educators themselves, according to experts.“We know that AI is putting out incredibly biased content,” said Amanda Lenhart, the head of research at Common Sense. “Humans come in with biases and preconceived notions about students in their classroom. AI is just another place in which unfairness is being laid upon students of color.”Put another way, even though AI tools aren’t human themselves, they reflect people’s prejudices, even unconscious ones. “AI is not going to walk us out of our pre-existing biases,” Lenhart said.If a teacher does suspect a student used AI to cheat on an assignment, it’s best to have a conversation with the student before jumping to punitive measures, educators and experts say. Schools also need to craft clear policies on when and how it’s acceptable to use AI to complete schoolwork.The Common Sense report is based on a nationally representative survey conducted from March to May of 1,045 adults in the United States who are the parents or guardians of one or more teens aged 13 to 18, and responses from one of their teenage children. All 18-year-old respondents were still in high school when surveyed.
·edweek.org·
Black Students Are More Likely to Be Falsely Accused of Using AI to Cheat
RAID: A Shared Benchmark for Robust Evaluation of...
RAID: A Shared Benchmark for Robust Evaluation of...
Many commercial and open-source models claim to detect machine-generated text with extremely high accuracy (99% or more). However, very few of these detectors are evaluated on shared benchmark...
·arxiv.org·
RAID: A Shared Benchmark for Robust Evaluation of...
The 5 Percent Problem
The 5 Percent Problem
Online mathematics programs may benefit most the kids who need it least
·educationnext.org·
The 5 Percent Problem
BOND
BOND
BOND is a global technology investment firm that supports visionary founders throughout their entire life cycle of innovation & growth.
·bondcap.com·
BOND
Research: Using AI at Work Makes Us Lonelier and Less Healthy
Research: Using AI at Work Makes Us Lonelier and Less Healthy
The promise of AI is alluring — optimized productivity, lightning-fast data analysis, and freedom from mundane tasks — and both companies and workers alike are fascinated (and more than a little dumbfounded) by how these tools allow them to do more and better work faster than ever before. Yet in fervor to keep pace with competitors and reap the efficiency gains associated with deploying AI, many organizations have lost sight of their most important asset: the humans whose jobs are being fragmented into tasks that are increasingly becoming automated. Across four studies, employees who use it as a core part of their jobs reported feeling lonelier, drinking more, and suffering from insomnia more than employees who don’t.
·hbr.org·
Research: Using AI at Work Makes Us Lonelier and Less Healthy
Two-Sigma Tutoring: Separating Science Fiction from Science Fact
Two-Sigma Tutoring: Separating Science Fiction from Science Fact
An experimental intervention in the 1980s raised certain test scores by two standard deviations. It wasn’t just tutoring, and it’s never been replicated, but it continues to inspire.
·educationnext.org·
Two-Sigma Tutoring: Separating Science Fiction from Science Fact
189 innovative school leaders: Teacher staffing, AI, mental health top ed issues | LA School Report
189 innovative school leaders: Teacher staffing, AI, mental health top ed issues | LA School Report
A common set of problems are keeping education leaders up at night: Will there be enough teachers to staff America’s schools? Can artificial intelligence enhance learning without deepening inequality? How can educators address the mental health crisis among young people? None of these have easy answers. New data confirm that these issues are top of...
·laschoolreport.com·
189 innovative school leaders: Teacher staffing, AI, mental health top ed issues | LA School Report
Researchers from Stanford and OpenAI Introduce 'Meta-Prompting': An Effective Scaffolding Technique Designed to Enhance the Functionality of Language Models in a Task-Agnostic Manner
Researchers from Stanford and OpenAI Introduce 'Meta-Prompting': An Effective Scaffolding Technique Designed to Enhance the Functionality of Language Models in a Task-Agnostic Manner
Language models (LMs), such as GPT-4, are at the forefront of natural language processing, offering capabilities that range from crafting complex prose to solving intricate computational problems. Despite their advanced functionalities, these models need fixing, sometimes yielding inaccurate or conflicting outputs. The challenge lies in enhancing their precision and versatility, particularly in complex, multi-faceted tasks. A key issue with current language models is their occasional inaccuracy and limitation in handling diverse and complex tasks. While these models excel in many areas, their efficacy could improve when confronted with tasks that demand nuanced understanding or specialized knowledge beyond their general capabilities.
·marktechpost.com·
Researchers from Stanford and OpenAI Introduce 'Meta-Prompting': An Effective Scaffolding Technique Designed to Enhance the Functionality of Language Models in a Task-Agnostic Manner
Meta-Prompting: Enhancing Language Models with Task-Agnostic Scaffolding
Meta-Prompting: Enhancing Language Models with Task-Agnostic Scaffolding
We introduce meta-prompting, an effective scaffolding technique designed to enhance the functionality of language models (LMs). This approach transforms a single LM into a multi-faceted conductor, adept at managing and integrating multiple independent LM queries. By employing high-level instructions, meta-prompting guides the LM to break down complex tasks into smaller, more manageable subtasks. These subtasks are then handled by distinct "expert" instances of the same LM, each operating under specific, tailored instructions. Central to this process is the LM itself, in its role as the conductor, which ensures seamless communication and effective integration of the outputs from these expert models. It additionally employs its inherent critical thinking and robust verification processes to refine and authenticate the end result. This collaborative prompting approach empowers a single LM to simultaneously act as a comprehensive orchestrator and a panel of diverse experts, significantly enhancing its performance across a wide array of tasks. The zero-shot, task-agnostic nature of meta-prompting greatly simplifies user interaction by obviating the need for detailed, task-specific instructions. Furthermore, our research demonstrates the seamless integration of external tools, such as a Python interpreter, into the meta-prompting framework, thereby broadening its applicability and utility. Through rigorous experimentation with GPT-4, we establish the superiority of meta-prompting over conventional scaffolding methods: When averaged across all tasks, including the Game of 24, Checkmate-in-One, and Python Programming Puzzles, meta-prompting, augmented with a Python interpreter functionality, surpasses standard prompting by 17.1%, expert (dynamic) prompting by 17.3%, and multipersona prompting by 15.2%.
·arxiv.org·
Meta-Prompting: Enhancing Language Models with Task-Agnostic Scaffolding
Introducing General World Models.
Introducing General World Models.
An early look into the future of AI with General World Models.
·research.runwayml.com·
Introducing General World Models.
State of AI Report 2023
State of AI Report 2023
The State of AI Report analyses the most interesting developments in AI. Read and download here.
·stateof.ai·
State of AI Report 2023