AI/ML

AI/ML

2386 bookmarks
Custom sorting
griptape-ai/griptape: Python framework for AI workflows and pipelines with chain of thought reasoning, external tools, and memory. Griptape is an enterprise grade alternative to LangChain.
griptape-ai/griptape: Python framework for AI workflows and pipelines with chain of thought reasoning, external tools, and memory. Griptape is an enterprise grade alternative to LangChain.
Python framework for AI workflows and pipelines with chain of thought reasoning, external tools, and memory. Griptape is an enterprise grade alternative to LangChain
·github.com·
griptape-ai/griptape: Python framework for AI workflows and pipelines with chain of thought reasoning, external tools, and memory. Griptape is an enterprise grade alternative to LangChain.
GitHub - hegelai/prompttools: Open-source tools for prompt testing and experimentation, with support for both LLMs (e.g. OpenAI, LLaMA) and vector databases (e.g. Chroma, Weaviate).
GitHub - hegelai/prompttools: Open-source tools for prompt testing and experimentation, with support for both LLMs (e.g. OpenAI, LLaMA) and vector databases (e.g. Chroma, Weaviate).
Open-source tools for prompt testing and experimentation, with support for both LLMs (e.g. OpenAI, LLaMA) and vector databases (e.g. Chroma, Weaviate). - GitHub - hegelai/prompttools: Open-source t...
·github.com·
GitHub - hegelai/prompttools: Open-source tools for prompt testing and experimentation, with support for both LLMs (e.g. OpenAI, LLaMA) and vector databases (e.g. Chroma, Weaviate).
Large language models, explained with a minimum of math and jargon
Large language models, explained with a minimum of math and jargon
  • Large language models like GPT-3 work by representing words as vectors of numbers and using neural networks with attention and transformer layers.

  • Word vectors allow language models to perform operations and reason about words in ways that strings of letters cannot.

  • Attention heads allow words to share contextual information with each other, helping the model resolve ambiguities and predict the next word.

  • Feed-forward layers act as a database of facts that the model has learned, enabling it to make predictions based on that knowledge.

  • Language models are trained by trying to predict the next word in text, requiring huge amounts of training data.

  • The performance of language models scales dramatically with their size, the amount of training data, and the compute used for training.

  • As language models get larger, they develop the ability to perform more complex reasoning and tasks requiring abstract thought.

  • Researchers do not fully understand how language models accomplish their abilities, and fully explaining them remains a huge challenge.

  • Language models appear to spontaneously develop capabilities like theory of mind as a byproduct of increasing language ability.

  • There is debate over whether language models truly "understand" language in the same sense that humans do.

·understandingai.org·
Large language models, explained with a minimum of math and jargon