Machine Learning & Artificial Intelligence

Machine Learning & Artificial Intelligence

531 bookmarks
Newest
XLM-RoBERTa
XLM-RoBERTa
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
·huggingface.co·
XLM-RoBERTa
RoBERTa: An optimized method for pretraining self-supervised NLP systems
RoBERTa: An optimized method for pretraining self-supervised NLP systems
Facebook AI’s RoBERTa is a new training recipe that improves on BERT, Google’s self-supervised method for pretraining natural language processing systems. By training longer, on more data, and dropping BERT’s next-sentence prediction RoBERTa topped the GLUE leaderboard.
·ai.meta.com·
RoBERTa: An optimized method for pretraining self-supervised NLP systems
AI21 Studio
AI21 Studio
A powerful language model, with an API that makes you smile
·ai21.com·
AI21 Studio
GPT-NeoX
GPT-NeoX
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
·huggingface.co·
GPT-NeoX
GPT-J
GPT-J
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
·huggingface.co·
GPT-J
Product
Product
Our API platform offers our latest models and guides for safety best practices.
·openai.com·
Product
Meta Open-Sources 175 Billion Parameter AI Language Model OPT
Meta Open-Sources 175 Billion Parameter AI Language Model OPT
Meta AI Research released Open Pre-trained Transformer (OPT-175B), a 175B parameter AI language model. The model was trained on a dataset containing 180B tokens and exhibits performance comparable with GPT-3, while only requiring 1/7th GPT-3's training carbon footprint.
·infoq.com·
Meta Open-Sources 175 Billion Parameter AI Language Model OPT
OPT
OPT
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
·huggingface.co·
OPT
Google GLaM | Discover AI use cases
Google GLaM | Discover AI use cases
The Generalist Language Model GLaM is a mixture of experts (MoE) model, a type of model that can be thought of as having different submodels (or expert...
·gpt3demo.com·
Google GLaM | Discover AI use cases
Gopher by DeepMind | Discover AI use cases
Gopher by DeepMind | Discover AI use cases
280 Billion Parameters Language model DeepMind’s language model, which it calls Gopher, is significantly more accurate than these existing ultra-large ...
·gpt3demo.com·
Gopher by DeepMind | Discover AI use cases
Turing-NLG: A 17-billion-parameter language model by Microsoft - Microsoft Research
Turing-NLG: A 17-billion-parameter language model by Microsoft - Microsoft Research
This figure was adapted from a similar image published in DistilBERT (opens in new tab). Turing Natural Language Generation (T-NLG) is a 17 billion parameter language model by Microsoft that outperforms the state of the art on many downstream NLP tasks. We present a demo of the model, including its freeform generation, question answering, and […]
·microsoft.com·
Turing-NLG: A 17-billion-parameter language model by Microsoft - Microsoft Research
Generate
Generate
Harness the power of Cohere's API to access pre-trained language models for AI-driven text generation. Create engaging content effortlessly. Get started for free.
·cohere.com·
Generate
NVIDIA NeMo
NVIDIA NeMo
This easy, efficient, and cost-effective framework helps developers build, train, and deploy large language models (LLMs) faster for enterprise application development.
·nvidia.com·
NVIDIA NeMo
Introducing Claude
Introducing Claude
After working with key partners for the past few months, we’re opening up access to Claude, our AI assistant.
·anthropic.com·
Introducing Claude