Students in RISD's Digital + Media graduate program make research-driven, multimedia work informed by art, science and technology. Learn more at RISD.edu.
GitHub - DivergentAI/dreamGPT: Leverage hallucinations from Large Language Models (LLMs) for novelty-driven explorations.
Leverage hallucinations from Large Language Models (LLMs) for novelty-driven explorations. - GitHub - DivergentAI/dreamGPT: Leverage hallucinations from Large Language Models (LLMs) for novelty-dri...
Hallucinations as a feature, not a bug | Union Square Ventures
Co-authored with Grace Carney A few months ago Fred kicked off a conversation about what the “native” applications of AI technology will be. What are the
Generative AI Systems Aren't Just Open or Closed Source
Conversation around generative AI tends to focus on whether its development is open or closed. It's more responsible to envision releases along a gradient.
Studying up Machine Learning Data: Why Talk About Bias When We Mean Power? | Montreal AI Ethics Institute
🔬 Research Summary by Shreyasha Paudel, a Ph.D. student at the University of Toronto with an interdisciplinary research focus that combines Human-Computer Interaction with critical theories from…
About Us Encord is a fast-growing startup building an active learning platform for computer vision AI applications. Our mission is to enable companies to unlock the power of AI. We have raised $20M from top investors including CRV, Y Combinator Continuity, the Harvard Management Company, top industry executives, and other leading Bay Area investors. Started by ex-computer scientists, physicists, and quants, we felt first hand how the lack of tools to prepare quality training data was impeding the progress of building practical AI. AI feels to us like what the early days of computing or the internet must have felt like, where the potential of the technology is clear, but the tools and processes surrounding it are terrible. We have devised a unique methodology for automating the tasks related to preparing quality training data, in effect turning the training data problem into a data science problem. Role and Responsibilities We’re looking for a product designer to accelerate our eff
The Curse of Recursion: Training on Generated Data Makes Models Forget
Stable Diffusion revolutionised image creation from descriptive text. GPT-2,
GPT-3(.5) and GPT-4 demonstrated astonishing performance across a variety of
language tasks. ChatGPT introduced such language models to the general public.
It is now clear that large language models (LLMs) are here to stay, and will
bring about drastic change in the whole ecosystem of online text and images. In
this paper we consider what the future might hold. What will happen to GPT-{n}
once LLMs contribute much of the language found online? We find that use of
model-generated content in training causes irreversible defects in the
resulting models, where tails of the original content distribution disappear.
We refer to this effect as Model Collapse and show that it can occur in
Variational Autoencoders, Gaussian Mixture Models and LLMs. We build
theoretical intuition behind the phenomenon and portray its ubiquity amongst
all learned generative models. We demonstrate that it has to be taken seriously
if we are to sustain the benefits of training from large-scale data scraped
from the web. Indeed, the value of data collected about genuine human
interactions with systems will be increasingly valuable in the presence of
content generated by LLMs in data crawled from the Internet.
Hi, I'm Pi. I'm your personal AI, designed to be supportive, smart, and there for you anytime. Ask me for advice, for answers, or let's talk about whatever's on your mind.
Design and AI with Nadia Piet - Machine Ethics Podcast
This episode Nadia and I chat about how design can co-create AI, what the role of designers are in AI services? post-deployment design, narratives in AI development and AI ideologues, anthropocentric AI, augmented creativity, new AI perspectives, situated intelligences and more...
Algorithmic Filmmaking Async Class (August 7 thru September 11, 2023) | BustBright - Machine Learning Art
Please note: unlike previous courses, this entirely class will be done async. You will get access to a Slack channel and weekly recordings. It's a great option for anyone across the globe.Algorithmic Filmmaking is a course for filmmakers, creative coders, and anyone else interested in tools for non-traditional editing and filmmaking. Every week you’ll receive 2 hours of recorded tutorials and discussions to help you create films using machine learning, algorithms and other cutting-edge tools. You’ll also receive supplemental lectures, films to watch, and other materials.Wondering the kind of things you’ll learn? Check out this overview of my project Scream Scenes. We’ll cover these techniques and much more.Prerequisitesknowledge of coding principles is very helpful, but not requiredaccess to a Windows, Mac, or Linux machine that can run Zoomexperience or familiarity with command line is very helpufl, and comfort reading code (any code will be provided to you)Access to free version of Google Colab and Google Drive Example syllabus (subject to change)Week 1 How to create a database of video clipsWeek 2 Categorizing shot typesWeek 3 Algorithmic fancamsWeek 4 Matchcuts and other uses of OpenposeWeek 5 Algorithmic sequencing Week 6 Audioreactive and other guidance techniquesWeek 7 Free Space/TBDTBD Student Demo DaySchedule and StructureThis course will be run asynchronously. Every week you’ll receive 2 hours of recorded tutorials and discussions. In addition to the seven 2-hour recordings, there will be two open 1hr sessions with Derrick every week to answer questions and get help on projects. We will be using Slack for asynchronous messaging and discussion. Additional materials outside of the lectures and demos will be provided as video links, webpages or PDFs.About the instructorsDerrick Schultz is a designer, filmmaker and artist working with machine learning. His work has been featured at machine learning conference CVPR and NeurIPS, and has been commissioned by the New York Times, HP, and Acne. He has taught numerous machine learning courses, and has teaching experience at ITP, Parsons, and SVA. You can see more of his work on his website.
This year Google Creative Lab is inviting multidisciplinary creatives to apply for a paid gig (£300 per day) and collaboration up to 12 months at Google Creative Lab, London. As a Fiver you will work on projects big, small and first of their kind within Google. You will be paired with a mentor and work closely with other creatives. A Fiver is an essential part of Google Creative Lab and it’s the way a lot of our current leaders started.
Tactical Tech is looking for a research assistant to enhance the
information in our Influence Industry Explorer, an open-access database that
compiles information about companies working with political parties and
candidates on data-driven and digital campaigns.
Call for Instructional Design Consultant: Adapting research into workshops
Tactical Tech is looking for a professional Instructional Design
consultant to develop workshop curricula for the Influence Industry Project to
accompany each of the learning modules into training materials in the existing
format.
One Project is looking for a Product Designer to collaborate on prototypes and products that support alternative economic and governance systems. We are seeking an experienced Product Designer fluent in participatory practices, grounded in principles of equity and justice, and curious about how a world beyond capitalism would look, feel, and function.
The MAK Museum of Applied Arts highlights AI, technology, and speculative narratives in a new exhibition
A new, timely exhibition at the MAK Museum of Applied Arts in Vienna is taking a closer look at the technological developments pushing architecture and design into new territory in the past decade using a variety of different relevant media.
From May 10–September 10, ...
The work reflects on the creative potential, challenges and limitations of human-AI relationships. Available to watch on Somerset House's online space Channel.
Three new film commissions created by artists and researchers Nouf Aljowaysir, Juan Covelli and Chris Zhongtian Yuan available on Somerset House's online space Channel.
CALL FOR AUGE NEXT Augmented Europe FellowshipsDigital Artists-in-Residency Programme With the support of the European Union’s Creative Europe programme, AUGE NEXT cooperation project launches four Fellowship Programmes for Digital Artists-in-Residency organized by the project partners in four European countries: Italy, Latvia, Germany, and Greece. Each of the four AUGE NEXT Fellowship Programmes consists of a […]
Decolonial AI: Decolonial Theory as Sociotechnical Foresight in Artificial Intelligence
This paper explores the important role of critical science, and in particular
of post-colonial and decolonial theories, in understanding and shaping the
ongoing advances in artificial intelligence. Artificial Intelligence (AI) is
viewed as amongst the technological advances that will reshape modern societies
and their relations. Whilst the design and deployment of systems that
continually adapt holds the promise of far-reaching positive change, they
simultaneously pose significant risks, especially to already vulnerable
peoples. Values and power are central to this discussion. Decolonial theories
use historical hindsight to explain patterns of power that shape our
intellectual, political, economic, and social world. By embedding a decolonial
critical approach within its technical practice, AI communities can develop
foresight and tactics that can better align research and technology development
with established ethical principles, centring vulnerable peoples who continue
to bear the brunt of negative impacts of innovation and scientific progress. We
highlight problematic applications that are instances of coloniality, and using
a decolonial lens, submit three tactics that can form a decolonial field of
artificial intelligence: creating a critical technical practice of AI, seeking
reverse tutelage and reverse pedagogies, and the renewal of affective and
political communities. The years ahead will usher in a wave of new scientific
breakthroughs and technologies driven by AI research, making it incumbent upon
AI communities to strengthen the social contract through ethical foresight and
the multiplicity of intellectual perspectives available to us; ultimately
supporting future technologies that enable greater well-being, with the goal of
beneficence and justice for all.