ai

ai

225 bookmarks
Custom sorting
The era of predictive AI Is almost over
The era of predictive AI Is almost over
“Language models are just glorified autocomplete” has been the critics’ refrain — but reinforcement learning is proving them wrong. New breakthroughs could follow.
·thenewatlantis.com·
The era of predictive AI Is almost over
KI: Kollaps droht wegen KI-generierter Trainingsdaten
KI: Kollaps droht wegen KI-generierter Trainingsdaten
Forscher fanden heraus, dass sich KI-Modelle selbst sabotieren, indem sie KI-generierte Daten zum Training verwenden. Sie produzieren dann immer mehr Müll.
·heise.de·
KI: Kollaps droht wegen KI-generierter Trainingsdaten
The “ChatGPT and Friends” Collection
The “ChatGPT and Friends” Collection
This page contains an annotated list of my publications/explanations about the 2023 surge of attention (hype is not always a wrong word here) society paid to Artificial Intelligence, mostly because…
·ea.rna.nl·
The “ChatGPT and Friends” Collection
When ChatGPT summarises, it actually does nothing of the kind.
When ChatGPT summarises, it actually does nothing of the kind.
One of the use cases I thought was reasonable to expect from ChatGPT and Friends (LLMs) was summarising. It turns out I was wrong. What ChatGPT isn’t summarising at all, it only looks like it…
·ea.rna.nl·
When ChatGPT summarises, it actually does nothing of the kind.
Making AI Less "Thirsty": Uncovering and Addressing the Secret Water Footprint of AI Models
Making AI Less "Thirsty": Uncovering and Addressing the Secret Water Footprint of AI Models
The growing carbon footprint of artificial intelligence (AI) models, especially large ones such as GPT-3, has been undergoing public scrutiny. Unfortunately, however, the equally important and enormous water (withdrawal and consumption) footprint of AI models has remained under the radar. For example, training GPT-3 in Microsoft's state-of-the-art U.S. data centers can directly evaporate 700,000 liters of clean freshwater, but such information has been kept a secret. More critically, the global AI demand may be accountable for 4.2 -- 6.6 billion cubic meters of water withdrawal in 2027, which is more than the total annual water withdrawal of 4 -- 6 Denmark or half of the United Kingdom. This is very concerning, as freshwater scarcity has become one of the most pressing challenges shared by all of us in the wake of the rapidly growing population, depleting water resources, and aging water infrastructures. To respond to the global water challenges, AI models can, and also must, take social responsibility and lead by example by addressing their own water footprint. In this paper, we provide a principled methodology to estimate the water footprint of AI models, and also discuss the unique spatial-temporal diversities of AI models' runtime water efficiency. Finally, we highlight the necessity of holistically addressing water footprint along with carbon footprint to enable truly sustainable AI.
·arxiv.org·
Making AI Less "Thirsty": Uncovering and Addressing the Secret Water Footprint of AI Models
LLMs can solve any word problem! As long as they can crib the answer
LLMs can solve any word problem! As long as they can crib the answer
AI companies keep claiming that LLMs do reasoning. The AI can think! The AI can figure out word puzzles! The AI can pass the LSAT! You’re fired! So our good friend Diz tested this by throwing logic…
·pivot-to-ai.com·
LLMs can solve any word problem! As long as they can crib the answer
Data Workers‘ Inquiry: Die versteckten Arbeitskräfte hinter der KI erzählen ihre Geschichten
Data Workers‘ Inquiry: Die versteckten Arbeitskräfte hinter der KI erzählen ihre Geschichten
Ohne Millionen Datenarbeiter:innen würden weder sogenannte Künstliche Intelligenz noch Content-Moderation funktionieren. In einem neuen Projekt erzählen sie ihre Geschichten: von Plattformarbeiter:innen in Venezuela und Syrien über Angestellte von Outsourcing-Firmen in Kenia bis zu Content-Moderator:innen in Deutschland.
·netzpolitik.org·
Data Workers‘ Inquiry: Die versteckten Arbeitskräfte hinter der KI erzählen ihre Geschichten
Bianca Kastl (@bkastl@mastodon.social)
Bianca Kastl (@bkastl@mastodon.social)
Wichtiger Beitrag über die digitale Drecksarbeit von Content-Moderator*innen. https://www.spiegel.de/netzwelt/web/digitale-drecksarbeit-die-belastende-arbeit-der-content-moderatoren-in-afrika-a-7dd3f260-8bb3-40e3-95a3-adf4137eaad3
·mastodon.social·
Bianca Kastl (@bkastl@mastodon.social)
AI scaling myths
AI scaling myths
Scaling will run out. The question is when.
·aisnakeoil.com·
AI scaling myths
There are three Rs! - interrogation post - Imgur
There are three Rs! - interrogation post - Imgur
Discover topics like interrogation, picard, chatgpt, and the magic of the internet at Imgur, a community powered entertainment destination. Lift your spirits with funny jokes, trending memes, entertaining gifs, inspiring stories, viral videos, and so much more from users like zenoshogun.
·imgur.com·
There are three Rs! - interrogation post - Imgur
ᴺⁱˡᶻ 🍸 (@nilz@norden.social)
ᴺⁱˡᶻ 🍸 (@nilz@norden.social)
Attached: 1 image #Strawberry 🍓 #AI #KI 💩 ( https://imgur.com/gallery/rVHerTG )
·norden.social·
ᴺⁱˡᶻ 🍸 (@nilz@norden.social)
Brands Are Beginning to Turn Against AI
Brands Are Beginning to Turn Against AI
Brands and companies are growing increasingly suspicious of AI, and promising customers they won't use it in their ads or products.
·rollingstone.com·
Brands Are Beginning to Turn Against AI
Datasets for Large Language Models: A Comprehensive Survey
Datasets for Large Language Models: A Comprehensive Survey
This paper embarks on an exploration into the Large Language Model (LLM) datasets, which play a crucial role in the remarkable advancements of LLMs. The datasets serve as the foundational infrastructure analogous to a root system that sustains and nurtures the development of LLMs. Consequently, examination of these datasets emerges as a critical topic in research. In order to address the current lack of a comprehensive overview and thorough analysis of LLM datasets, and to gain insights into their current status and future trends, this survey consolidates and categorizes the fundamental aspects of LLM datasets from five perspectives: (1) Pre-training Corpora; (2) Instruction Fine-tuning Datasets; (3) Preference Datasets; (4) Evaluation Datasets; (5) Traditional Natural Language Processing (NLP) Datasets. The survey sheds light on the prevailing challenges and points out potential avenues for future investigation. Additionally, a comprehensive review of the existing available dataset resources is also provided, including statistics from 444 datasets, covering 8 language categories and spanning 32 domains. Information from 20 dimensions is incorporated into the dataset statistics. The total data size surveyed surpasses 774.5 TB for pre-training corpora and 700M instances for other datasets. We aim to present the entire landscape of LLM text datasets, serving as a comprehensive reference for researchers in this field and contributing to future studies. Related resources are available at: https://github.com/lmmlzn/Awesome-LLMs-Datasets.
·arxiv.org·
Datasets for Large Language Models: A Comprehensive Survey
Kris (@isotopp@chaos.social)
Kris (@isotopp@chaos.social)
Angehängt: 1 Bild In https://ddinstagram.com/sloan_spencer_author/p/C48JN_TrvxO/?img_index=1 heißt es: "Warning to all writers who use Google Docs. In a Discord with other indie authors last night, one romance author explained that all her work was suspended by Google for violating their rules of sharing explicit content with her beta/alpha readers. She has no acces to these files or her other works. Please read the following screenshots and go back up your workj now on a different plattform."
·chaos.social·
Kris (@isotopp@chaos.social)
elle (@ElleGray@mstdn.social)
elle (@ElleGray@mstdn.social)
Angehängt: 1 Bild Omg any day now computers will be able to add
·mstdn.social·
elle (@ElleGray@mstdn.social)