Large language models like GPT-3 work by representing words as vectors of numbers and using neural networks with attention and transformer layers.
Word vectors allow language models to perform operations and reason about words in ways that strings of letters cannot.
Attention heads allow words to share contextual information with each other, helping the model resolve ambiguities and predict the next word.
Feed-forward layers act as a database of facts that the model has learned, enabling it to make predictions based on that knowledge.
Language models are trained by trying to predict the next word in text, requiring huge amounts of training data.
The performance of language models scales dramatically with their size, the amount of training data, and the compute used for training.
As language models get larger, they develop the ability to perform more complex reasoning and tasks requiring abstract thought.
Researchers do not fully understand how language models accomplish their abilities, and fully explaining them remains a huge challenge.
Language models appear to spontaneously develop capabilities like theory of mind as a byproduct of increasing language ability.
There is debate over whether language models truly "understand" language in the same sense that humans do.
griptape-ai/griptape: Python framework for AI workflows and pipelines with chain of thought reasoning, external tools, and memory. Griptape is an enterprise grade alternative to LangChain.
Python framework for AI workflows and pipelines with chain of thought reasoning, external tools, and memory. Griptape is an enterprise grade alternative to LangChain
Building an LLM Open-Source Search Engine in 100 Lines
Non-determinism in GPT-4 is caused by Sparse MoE - 152334H
How to Chat With Any PDFs and Image Files Using Large Language Models — With Code | by Zoumana Keita | Aug, 2023 | Towards Data Science
GitHub - hegelai/prompttools: Open-source tools for prompt testing and experimentation, with support for both LLMs (e.g. OpenAI, LLaMA) and vector databases (e.g. Chroma, Weaviate).
Open-source tools for prompt testing and experimentation, with support for both LLMs (e.g. OpenAI, LLaMA) and vector databases (e.g. Chroma, Weaviate). - GitHub - hegelai/prompttools: Open-source t...
Political Milestones for AI - Schneier on Security
The Need for Trustworthy AI - Schneier on Security
AI-supported mammogram screening increases breast cancer detection by 20%, study finds | CNN
2D OCR
Glasgow university spinout raises $43mn to ‘digitise chemistry’ | Financial Times
Patterns for Building LLM-based Systems & Products
How ChatGPT changed my plans for the fall
We Talked to The Researchers Studying How AI-Generated Content Breaks AI's Brain
We talked to Rice University scientists studying how training AI with synthetic, AI-generated content causes a destructive inbreeding effect.
Run Llama 2 on your own Mac using LLM and Homebrew
LLM-CLI
Pythia: A Suite of 16 LLMs for In-Depth Research - KDnuggets
How Sweden's biggest daily uses ChatGPT in its newsroom
Show HN: Khoj – Chat offline with your second brain using Llama 2 | Hacker News
Khoj Install
khoj-ai/khoj: An AI personal assistant for your digital brain
Decipad — Make sense of numbers
Greg Rutkowski Was Removed From Stable Diffusion, But AI Artists Brought Him Back - Decrypt
chenhunghan/ialacol: 🪶 Lightweight 🦄 Self hosted, private, 🐟 scalable, 🤑 commercially usable, 💬 LLM chat streaming service with 1-click Kubernetes cluster installation on any cloud
Anthropic, Google, Microsoft and OpenAI launch Frontier Model Forum
Universal and Transferable Attacks on Aligned Language Models
[P] Run Llama 2 Locally in 7 Lines! (Apple Silicon Mac) : r/MachineLearning
mlc-ai/mlc-llm: Enable everyone to develop, optimize and deploy AI models natively on everyone's devices.
Running PyTorch on the M1 GPU
Sparta Is No Model for U.S. Soldiers
Guide to running Llama 2 locally | Hacker News
Large language models, explained with a minimum of math and jargon