Research

102 bookmarks
Custom sorting
RAID: A Shared Benchmark for Robust Evaluation of...
RAID: A Shared Benchmark for Robust Evaluation of...
Many commercial and open-source models claim to detect machine-generated text with extremely high accuracy (99% or more). However, very few of these detectors are evaluated on shared benchmark...
·arxiv.org·
RAID: A Shared Benchmark for Robust Evaluation of...
The 5 Percent Problem
The 5 Percent Problem
Online mathematics programs may benefit most the kids who need it least
·educationnext.org·
The 5 Percent Problem
BOND
BOND
BOND is a global technology investment firm that supports visionary founders throughout their entire life cycle of innovation & growth.
·bondcap.com·
BOND
Research: Using AI at Work Makes Us Lonelier and Less Healthy
Research: Using AI at Work Makes Us Lonelier and Less Healthy
The promise of AI is alluring — optimized productivity, lightning-fast data analysis, and freedom from mundane tasks — and both companies and workers alike are fascinated (and more than a little dumbfounded) by how these tools allow them to do more and better work faster than ever before. Yet in fervor to keep pace with competitors and reap the efficiency gains associated with deploying AI, many organizations have lost sight of their most important asset: the humans whose jobs are being fragmented into tasks that are increasingly becoming automated. Across four studies, employees who use it as a core part of their jobs reported feeling lonelier, drinking more, and suffering from insomnia more than employees who don’t.
·hbr.org·
Research: Using AI at Work Makes Us Lonelier and Less Healthy
Two-Sigma Tutoring: Separating Science Fiction from Science Fact
Two-Sigma Tutoring: Separating Science Fiction from Science Fact
An experimental intervention in the 1980s raised certain test scores by two standard deviations. It wasn’t just tutoring, and it’s never been replicated, but it continues to inspire.
·educationnext.org·
Two-Sigma Tutoring: Separating Science Fiction from Science Fact
189 innovative school leaders: Teacher staffing, AI, mental health top ed issues | LA School Report
189 innovative school leaders: Teacher staffing, AI, mental health top ed issues | LA School Report
A common set of problems are keeping education leaders up at night: Will there be enough teachers to staff America’s schools? Can artificial intelligence enhance learning without deepening inequality? How can educators address the mental health crisis among young people? None of these have easy answers. New data confirm that these issues are top of...
·laschoolreport.com·
189 innovative school leaders: Teacher staffing, AI, mental health top ed issues | LA School Report
Researchers from Stanford and OpenAI Introduce 'Meta-Prompting': An Effective Scaffolding Technique Designed to Enhance the Functionality of Language Models in a Task-Agnostic Manner
Researchers from Stanford and OpenAI Introduce 'Meta-Prompting': An Effective Scaffolding Technique Designed to Enhance the Functionality of Language Models in a Task-Agnostic Manner
Language models (LMs), such as GPT-4, are at the forefront of natural language processing, offering capabilities that range from crafting complex prose to solving intricate computational problems. Despite their advanced functionalities, these models need fixing, sometimes yielding inaccurate or conflicting outputs. The challenge lies in enhancing their precision and versatility, particularly in complex, multi-faceted tasks. A key issue with current language models is their occasional inaccuracy and limitation in handling diverse and complex tasks. While these models excel in many areas, their efficacy could improve when confronted with tasks that demand nuanced understanding or specialized knowledge beyond their general capabilities.
·marktechpost.com·
Researchers from Stanford and OpenAI Introduce 'Meta-Prompting': An Effective Scaffolding Technique Designed to Enhance the Functionality of Language Models in a Task-Agnostic Manner
Meta-Prompting: Enhancing Language Models with Task-Agnostic Scaffolding
Meta-Prompting: Enhancing Language Models with Task-Agnostic Scaffolding
We introduce meta-prompting, an effective scaffolding technique designed to enhance the functionality of language models (LMs). This approach transforms a single LM into a multi-faceted conductor, adept at managing and integrating multiple independent LM queries. By employing high-level instructions, meta-prompting guides the LM to break down complex tasks into smaller, more manageable subtasks. These subtasks are then handled by distinct "expert" instances of the same LM, each operating under specific, tailored instructions. Central to this process is the LM itself, in its role as the conductor, which ensures seamless communication and effective integration of the outputs from these expert models. It additionally employs its inherent critical thinking and robust verification processes to refine and authenticate the end result. This collaborative prompting approach empowers a single LM to simultaneously act as a comprehensive orchestrator and a panel of diverse experts, significantly enhancing its performance across a wide array of tasks. The zero-shot, task-agnostic nature of meta-prompting greatly simplifies user interaction by obviating the need for detailed, task-specific instructions. Furthermore, our research demonstrates the seamless integration of external tools, such as a Python interpreter, into the meta-prompting framework, thereby broadening its applicability and utility. Through rigorous experimentation with GPT-4, we establish the superiority of meta-prompting over conventional scaffolding methods: When averaged across all tasks, including the Game of 24, Checkmate-in-One, and Python Programming Puzzles, meta-prompting, augmented with a Python interpreter functionality, surpasses standard prompting by 17.1%, expert (dynamic) prompting by 17.3%, and multipersona prompting by 15.2%.
·arxiv.org·
Meta-Prompting: Enhancing Language Models with Task-Agnostic Scaffolding
Introducing General World Models.
Introducing General World Models.
An early look into the future of AI with General World Models.
·research.runwayml.com·
Introducing General World Models.
State of AI Report 2023
State of AI Report 2023
The State of AI Report analyses the most interesting developments in AI. Read and download here.
·stateof.ai·
State of AI Report 2023
Decoding Intentions - Center for Security and Emerging Technology
Decoding Intentions - Center for Security and Emerging Technology
How can policymakers credibly reveal and assess intentions in the field of artificial intelligence? Policymakers can send credible signals of their intent by making pledges or committing to undertaking certain actions for which they will pay a price—political, reputational, or monetary—if they back down or fail to make good on their initial promise or threat. Talk is cheap, but inadvertent escalation is costly to all sides.
·cset.georgetown.edu·
Decoding Intentions - Center for Security and Emerging Technology
The Future of AI in Education: 13 Things We Can Do to Minimize the Damage
The Future of AI in Education: 13 Things We Can Do to Minimize the Damage
We may already be in the era of ‘peak humanity’, a time where we have the greatest levels of education, reasoning, rationality, and creativity – spread out amongst the greatest number of us. A brilliant result of the massification of universal basic education and the power of the university. But with the rapid advancement of Artificial Intelligence (AI) that can already replicate and even exceed many of our reasoning capabilities – there may soon be less incentive for us to learn and grow. The grave risk is that we then become de-educated and de-coupled from the driving seat to the future. In all the hype about AI, we need to properly assess these risks to collectively decide whether the AI upsides are worth it and whether we should ‘stick or twist’. This paper aims to catalyse the debate and reduce the probability that we sleepwalk to a destination that we don’t want and can’t reverse back out of. We also make 13 clear recommendations about how AI developments could be regulated - to slow things down a little and give time for informed choices about the best future for humanity. Those potential long-term futures include: (1) AI Curtailment; (2) Fake Work; (3) Transhumanism; and (4) Universal Basic Income – each with very different implications for the future of education.
·osf.io·
The Future of AI in Education: 13 Things We Can Do to Minimize the Damage
44% of Teens Intend to Have AI Do Their Schoolwork This Fall, and 60% Consider This 'Cheating' -- THE Journal
44% of Teens Intend to Have AI Do Their Schoolwork This Fall, and 60% Consider This 'Cheating' -- THE Journal
In a survey conducted for the Junior Achievement organization in July 2023, of the 1,006 respondents between the ages of 13 and 17 who were polled, nearly half of them said they intend to use AI this fall to do their classwork for them. But most teens consider doing this to be “cheating.”
·thejournal.com·
44% of Teens Intend to Have AI Do Their Schoolwork This Fall, and 60% Consider This 'Cheating' -- THE Journal
Wikipedia’s Moment of Truth
Wikipedia’s Moment of Truth
Can the online encyclopedia help teach A.I. chatbots to get their facts right — without destroying itself in the process?
·nytimes.com·
Wikipedia’s Moment of Truth
Researchers populated a tiny virtual town with AI
Researchers populated a tiny virtual town with AI
What happens if you fill a virtual town with AIs and set them loose? Turns out, they brush their teeth and are very nice to one another!
·techcrunch.com·
Researchers populated a tiny virtual town with AI