Improving ChatGPT's Ability to Understand Ambiguous Prompts
Prompt engineering technique helps large language models (LLMs) handle pronouns and other complex coreferences in retrieval augmented generation (RAG) systems.
Humane finally actually announced their first product: Ai Pin. Here's their introduction video.
I love technology, and I think I'm generally enthusiastic about new tech that pushes things forwards. I'm enthusiastic about the ways LLMs are creating so many new things that weren't even possible literally one year ago, and
Here’s number 12 in the series on LLM-assisted coding over at The New Stack: Let’s Talk: Conversational Software Development I keep coming back to the theme of the first article in this serie…
How to Think Computationally about AI, the Universe and Everything
In his TED Talk, Stephen Wolfram covers the emergence of space by the application of computational rules to spacetime, gravity and quantum mechanics to AI and LLMs. Computational irreducibility and the ruliad.
AutoGen: Enabling next-generation large language model applications
Microsoft researchers are introducing AutoGen, a framework for simplifying the orchestration, optimization, and automation of workflows for large language model (LLM) applications—potentially transforming and extending what LLMs can do. Learn more.
Critics of LLM-based products like ChatGPT, Claude, Midjourney, and other such products like to brush them off as just this year’s version of NFTs. They’re crypto bullshit being peddled by the same jokers who are just out there to stow disinformation and make a quick buck.
I won’
There are many benefits of running a private LLM for your company or product, but it all boils down to being able to provide real-time data in context.
ChatGPT and LLMs can do anything, so what can you do with them? How do you know? Do we move to chat bots as a magical general-purpose interface, or do we unbundle them back into single-purpose software?
Will AI hamper our ability to crawl the web for useful data?
As websites start to block Common Crawl, and as the project leans in to its role in training LLMs, will it become harder to use data from the web for other purposes?
Prompt Engineering, also known as In-Context Prompting, refers to methods for how to communicate with LLM to steer its behavior for desired outcomes without updating the model weights. It is an empirical science and the effect of prompt engineering methods can vary a lot among models, thus requiring heavy experimentation and heuristics.
This post only focuses on prompt engineering for autoregressive language models, so nothing with Cloze tests, image generation or multimodality models.