The ultimate AI productivity app that protects your privacy. Bring all your apps and data into one AI-powered search and assistant. Get it for you and for your teams today.
GitHub - thiswillbeyourgithub/WDoc: Summarize and query from a lot of heterogeneous documents. Any LLM provider, any filetype, scalable, under developpement
Summarize and query from a lot of heterogeneous documents. Any LLM provider, any filetype, scalable, under developpement - thiswillbeyourgithub/WDoc
How streaming LLM APIs work | Simon Willison’s TILs
I decided to have a poke around and see if I could figure out how the HTTP streaming APIs from the various hosted LLM providers actually worked. Here are my notes so far.
GitHub - ictnlp/LLaMA-Omni: LLaMA-Omni is a low-latency and high-quality end-to-end speech interaction model built upon Llama-3.1-8B-Instruct, aiming to achieve speech capabilities at the GPT-4o level.
LLaMA-Omni is a low-latency and high-quality end-to-end speech interaction model built upon Llama-3.1-8B-Instruct, aiming to achieve speech capabilities at the GPT-4o level. - ictnlp/LLaMA-Omni
Sourcetable is an AI spreadsheet that helps you analyze data and create reports. Chat with your data, create charts and graphs, build financial models, + more.
Here's an interesting new embedding/RAG technique, described by Anthropic but it should work for any embedding model against any other LLM. One of the big challenges in implementing semantic search …
Reader-LM: Small Language Models for Cleaning and Converting HTML to Markdown
Reader-LM-0.5B and Reader-LM-1.5B are two novel small language models inspired by Jina Reader, designed to convert raw, noisy HTML from the open web into clean markdown.
OpenAI released two major new preview models today: o1-preview and o1-mini (that mini one is also a preview, despite the name)—previously rumored as having the codename “strawberry”. There’s a lot …
New version of my `files-to-prompt` CLI tool for turning a bunch of files into a prompt suitable for piping to an LLM, [described here previously](https://simonwillison.net/2024/Apr/8/files-to-prompt/). It now has a `-c/--cxml` …
Yes, the rumours are true! Kagi has been thoughtfully integrating AI into our search experience, creating a smarter, faster, and more intuitive search.
New hosted API for Llama running at absurdly high speeds: "1,800 tokens per second for Llama3.1 8B and 450 tokens per second for Llama3.1 70B". How are they running so …