Found 511 bookmarks
Newest
AI Machine Learning Roadmap: Self Study AI!
AI Machine Learning Roadmap: Self Study AI!
Unlock the secrets to mastering Artificial Intelligence (AI) quickly with this self-study roadmap, based on the prestigious Stanford AI Graduate Certificate ...
·youtube.com·
AI Machine Learning Roadmap: Self Study AI!
Previously, RAG systems were the standard method for retrieving information from documents. However, if you are not repeatedly querying the same document, it may be more convenient and effective to just use long-context LLMs. For example, Llama 3.1 8B and Llama 3.2 1B/3B now…
Previously, RAG systems were the standard method for retrieving information from documents. However, if you are not repeatedly querying the same document, it may be more convenient and effective to just use long-context LLMs. For example, Llama 3.1 8B and Llama 3.2 1B/3B now…
— Sebastian Raschka (@rasbt)
·x.com·
Previously, RAG systems were the standard method for retrieving information from documents. However, if you are not repeatedly querying the same document, it may be more convenient and effective to just use long-context LLMs. For example, Llama 3.1 8B and Llama 3.2 1B/3B now…
(3) LlamaIndex 🦙 on X: "Check out this video from @thesourabhd on how to build AI agents using LlamaCloud plus @qdrant_engine! This deep dive covers: ➡️ Implementing semantic caching in agent systems to improve speed and efficiency ➡️ Advanced agent techniques like query routing, query decomposition, https://t.co/DVfK0FE0bD" / X
(3) LlamaIndex 🦙 on X: "Check out this video from @thesourabhd on how to build AI agents using LlamaCloud plus @qdrant_engine! This deep dive covers: ➡️ Implementing semantic caching in agent systems to improve speed and efficiency ➡️ Advanced agent techniques like query routing, query decomposition, https://t.co/DVfK0FE0bD" / X
This deep dive covers: ➡️ Implementing semantic caching in agent systems to improve speed and efficiency ➡️ Advanced agent techniques like query routing, query decomposition,… — LlamaIndex 🦙 (@llama_index)
·x.com·
(3) LlamaIndex 🦙 on X: "Check out this video from @thesourabhd on how to build AI agents using LlamaCloud plus @qdrant_engine! This deep dive covers: ➡️ Implementing semantic caching in agent systems to improve speed and efficiency ➡️ Advanced agent techniques like query routing, query decomposition, https://t.co/DVfK0FE0bD" / X
Nexa AI - The On-Device AI Open Source Community Building The Future. Explore Quantized AI Models On Edge | Nexa AI Model Hub For NLP, Computer Vision, Multimodality & On-Device AI
Nexa AI - The On-Device AI Open Source Community Building The Future. Explore Quantized AI Models On Edge | Nexa AI Model Hub For NLP, Computer Vision, Multimodality & On-Device AI
Nexa AI On-Device Model Hub: LLaMA, Stable Diffusion, Whisper & more. Pre-trained AI models for NLP, vision, multimodality.
·nexa.ai·
Nexa AI - The On-Device AI Open Source Community Building The Future. Explore Quantized AI Models On Edge | Nexa AI Model Hub For NLP, Computer Vision, Multimodality & On-Device AI
Transformers Inference Optimization Toolset
Transformers Inference Optimization Toolset
Large Language Models are pushing the boundaries of artificial intelligence, but their immense size poses significant computational challenges. As these models grow, so does the need for smart optimization techniques to keep them running efficiently on modern hardware. In this post, we’ll explore key optimization strategies that are making LLMs faster and more memory-efficient. We’ll start with a brief look at GPU memory hierarchy, which forms the foundation for many of these techniques. Then, we’ll explore algorithms that allow LLMs to process information more quickly and handle longer contexts. Understanding these techniques offers valuable insights helping to unlock the full potential of Large Language Models.
·astralord.github.io·
Transformers Inference Optimization Toolset
Build an end-to-end RAG solution using Knowledge Bases for Amazon Bedrock and the AWS CDK | Amazon Web Services
Build an end-to-end RAG solution using Knowledge Bases for Amazon Bedrock and the AWS CDK | Amazon Web Services
In this post, we demonstrate how to seamlessly automate the deployment of an end-to-end RAG solution using Knowledge Bases for Amazon Bedrock and the AWS Cloud Development Kit (AWS CDK), enabling organizations to quickly set up a powerful question answering system.
·aws.amazon.com·
Build an end-to-end RAG solution using Knowledge Bases for Amazon Bedrock and the AWS CDK | Amazon Web Services
Best practices for building robust generative AI applications with Amazon Bedrock Agents – Part 1 | Amazon Web Services
Best practices for building robust generative AI applications with Amazon Bedrock Agents – Part 1 | Amazon Web Services
In this post, we show you how to create accurate and reliable agents. Agents helps you accelerate generative AI application development by orchestrating multistep tasks. Agents use the reasoning capability of foundation models (FMs) to break down user-requested tasks into multiple steps.
·aws.amazon.com·
Best practices for building robust generative AI applications with Amazon Bedrock Agents – Part 1 | Amazon Web Services
A helpful approach to navigating the SEO AI shift
A helpful approach to navigating the SEO AI shift
Explore AI's impact on SEO strategies and content creation. Find out how to balance AI tools with human expertise for better search rankings.
·builder.io·
A helpful approach to navigating the SEO AI shift
Don’t Build AI Products The Way Everyone Else Is Doing It
Don’t Build AI Products The Way Everyone Else Is Doing It
More and more, people are building products as thin wrappers over other models, especially LLMs. This can cause major issues, and there is a better way.
·builder.io·
Don’t Build AI Products The Way Everyone Else Is Doing It
(10) Career Advice For A World After AI - YouTube
(10) Career Advice For A World After AI - YouTube
Enjoy the videos and music that you love, upload original content and share it all with friends, family and the world on YouTube.
·youtube.com·
(10) Career Advice For A World After AI - YouTube
18 months of pgvector learnings in 47 minutes
18 months of pgvector learnings in 47 minutes
📌 Github with all Code, Slides, Resources ⇒ https://tsdb.co/busy-dev-resources ​​PostgreSQL is catching fire as the default database choice for AI applications. But how can you best leverage all that PostgreSQL has to offer for AI applications? What are best practices to follow, common pitfalls to avoid, and tools that will accelerate your development? ​​​ Avthar Sewrathan, PM AI and Vector @ Timescale shares his learning from helping developers build AI applications with PostgreSQL over the past 18 months. Avthar covers the state of the union of developing AI applications in 2024, and why PostgreSQL can save you time and headaches now and in the future. 🛠 𝗥𝗲𝗹𝗲𝘃𝗮𝗻𝘁 𝗥𝗲𝘀𝗼𝘂𝗿𝗰𝗲𝘀 📌 Free Trial of Timescale ⇒ https://tsdb.co/busy-dev-signup 📌 pgai ⇒ https://tsdb.co/pgai 📌 pgvectorscale ⇒ https://tsdb.co/pgvectorscale 🐯 𝗔𝗯𝗼𝘂𝘁 𝗧𝗶𝗺𝗲𝘀𝗰𝗮𝗹𝗲 At Timescale, we see a world made better via innovative technologies, and we are dedicated to serving software developers and businesses worldwide, enabling them to build the next wave of computing. Timescale is a remote-first company with a global workforce backed by top-tier investors with a track record of success in the industry. 💻 𝗙𝗶𝗻𝗱 𝗨𝘀 𝗢𝗻𝗹𝗶𝗻𝗲! 🔍 Website ⇒ https://tsdb.co/homepage 🔍 Slack ⇒ https://slack.timescale.com 🔍 GitHub ⇒ https://github.com/timescale 🔍 Twitter ⇒ https://twitter.com/timescaledb 🔍 Twitch ⇒ https://www.twitch.tv/timescaledb 🔍 LinkedIn ⇒ https://www.linkedin.com/company/timescaledb 🔍 Timescale Blog ⇒ https://tsdb.co/blog 🔍 Timescale Documentation ⇒ https://tsdb.co/docs 📚 𝗖𝗵𝗮𝗽𝘁𝗲𝗿𝘀 00:00 What You’ll Learn Today 02:03 Foundations of AI with Postgres 03:58 Understanding Vector Data and Databases 06:25 Types of AI Applications with Postgres 08:53 Postgres Extensions for AI 14:25 Demo: Using pgVector and pgAI 21:42 Vector Search Indexes in Postgres 25:14 Vector Compression and Indexing Options 26:13 Introducing Streaming Disk Index 28:12 Creating Vector Search Indices in Postgres 30:22 Advanced Topics in AI Systems 32:14 Evaluation Driven Development 35:51 Filtered Vector Search 40:32 Hybrid Search Techniques 42:31 Multi-Tenancy in RAG Applications 44:41 Text to SQL in AI Applications 45:33 Conclusion and Resources
·m.youtube.com·
18 months of pgvector learnings in 47 minutes
Production RAG with a Postgres Vector Store and Open-Source Models
Production RAG with a Postgres Vector Store and Open-Source Models
Explore Retrieval Augmented Generation (RAG) with Postgres Vector Store for sophisticated search functionalities in Django applications, leveraging the power of open-source models.
·christophergs.com·
Production RAG with a Postgres Vector Store and Open-Source Models
Some advice and good practices when integrating an LLM in your application
Some advice and good practices when integrating an LLM in your application
When integrating an LLM into your applicaton to extend it and make it smarter, it’s important to be aware of the pitfalls and best practices you need to follow to avoid some common problems and integrate them successfully. This article will guide you through some key best practices that I’ve come across. Understanding the Challenges of Implementing LLMs in Real-World Applications One of the first challenges is that LLMs are constantly being improved.
·glaforge.dev·
Some advice and good practices when integrating an LLM in your application
The Unreasonable Effectiveness of Prompt "Engineering"
The Unreasonable Effectiveness of Prompt "Engineering"
Check out the FREE non-technical guide for using AI in your business here: https://clickhubspot.com/fuu7 This video imma be yapping about why prompt engineering is unreasonable and everything that is a bit sus about it. The sequel of this video will probably be me roasting o1 (hopefully not) check out my newsletter https://mail.bycloud.ai/ Special thanks to LDJ for helping with this video OpenAI o1 [Announcement] https://openai.com/index/introducing-openai-o1-preview/ [Blogs] https://openai.com/o1/ Claude 3.5 Sonnet [Announcement] https://www.anthropic.com/news/claude-3-5-sonnet [AntThinking] https://gist.github.com/dedlim/6bf6d81f77c19e20cd40594aa09e3ecd Language Models (Mostly) Know What They Know [Paper] https://arxiv.org/abs/2207.05221 Plan-and-Solve Prompting [Paper] https://arxiv.org/abs/2305.04091 Llama-3.1 System Card [Documentation] https://www.llama.com/docs/model-cards-and-prompt-formats/meta-llama-3/ Think Before You Speak [Paper] https://arxiv.org/abs/2310.02226 Future Lens [Paper] https://arxiv.org/abs/2311.04897 This video is supported by the kind Patrons & YouTube Members: 🙏Andrew Lescelius, alex j, Chris LeDoux, Alex Maurice, Miguilim, Deagan, FiFaŁ, Robert Zawiasa, Owen Ingraham, Daddy Wen, Tony Jimenez, Panther Modern, Jake Disco, Demilson Quintao, Penumbraa, Shuhong Chen, Hongbo Men, happi nyuu nyaa, Carol Lo, Mose Sakashita, Miguel, Bandera, Gennaro Schiano, gunwoo, Ravid Freedman, Mert Seftali, Mrityunjay, Richárd Nagyfi, Timo Steiner, Henrik G Sundt, projectAnthony, Brigham Hall, Kyle Hudson, Kalila, Jef Come, Jvari Williams, Tien Tien, BIll Mangrum, owned, Janne Kytölä, SO, Richárd Nagyfi, Hector, Drexon, Claxvii 177th, Inferencer, Michael Brenner, Akkusativ, Oleg Wock, FantomBloth, Thipok Tham, Clayton Ford, Theo, Handenon, Diego Silva, mayssam [Discord] https://discord.gg/NhJZGtH [Twitter] https://twitter.com/bycloudai [Patreon] https://www.patreon.com/bycloud [Music 1] Spirit Blossom - RomanBelov [Music 2] Tie Me Down - PremiumMusicOdyssey [Music 3] Dimmed Lights - PremiumMusicOdyssey [Music 4] Wrapped in your Love - PremiumMusicOdyssey [Profile & Banner Art] https://twitter.com/pygm7 [Video Editor] @bhargavesque 0:00 Intro 4:21 The prooompting basics 6:36 How companies make prompting works 12:11 Does how you prompt REALLY matter?
·youtube.com·
The Unreasonable Effectiveness of Prompt "Engineering"
AI should do chores, not the fun stuff
AI should do chores, not the fun stuff
What’s the *right* use for AI? Laurie Voss thinks it’s great at doing boring chores, and in this episode we learn what that means and how we can put the robo...
·youtube.com·
AI should do chores, not the fun stuff
LLM University (LLMU)
LLM University (LLMU)
Welcome to LLM University, your premier learning destination for mastering Enterprise AI technologies. Designed for developers and technical professionals, our hub offers comprehensive resources, expert-led courses, and step-by-step guides to help you start building quickly and stay ahead in the rapidly evolving AI landscape.
·cohere.com·
LLM University (LLMU)
Fine-tuning | How-to guides
Fine-tuning | How-to guides
Full parameter fine-tuning is a method that fine-tunes all the parameters of all the layers of the pre-trained model.
·llama.com·
Fine-tuning | How-to guides