Learn AI

559 bookmarks
Custom sorting
Supercharging LLM Application Development with LLM-Kit
Supercharging LLM Application Development with LLM-Kit
Discover how Grab's LLM-Kit enhances AI app development by addressing scalability, security, and integration challenges. This article discusses the challenges faced in LLM app building, the solution, the architecture of the LLM-Kit as well as the future plans of the LLM-Kit.
·engineering.grab.com·
Supercharging LLM Application Development with LLM-Kit
A guide to Amazon Bedrock Model Distillation (preview)
A guide to Amazon Bedrock Model Distillation (preview)
This post introduces the workflow of Amazon Bedrock Model Distillation. We first introduce the general concept of model distillation in Amazon Bedrock, and then focus on the important steps in model distillation, including setting up permissions, selecting the models, providing input dataset, commencing the model distillation jobs, and conducting evaluation and deployment of the student models after model distillation.
·aws.amazon.com·
A guide to Amazon Bedrock Model Distillation (preview)
AI SDK 4.0 - Vercel
AI SDK 4.0 - Vercel
Introducing PDF support, computer use, and an xAI Grok provider
·vercel.com·
AI SDK 4.0 - Vercel
AI Engineer Roadmap
AI Engineer Roadmap
Learn to become an AI Engineer using this roadmap. Community driven, articles, resources, guides, interview questions, quizzes for modern backend development.
·roadmap.sh·
AI Engineer Roadmap
Weights & Biases
Weights & Biases
Weights & Biases, developer tools for machine learning
·wandb.ai·
Weights & Biases
The Complete RAG Course - Learn AI Skills
The Complete RAG Course - Learn AI Skills
Use code YOUTUBE to get an extra 20% off my AI courses here:https://www.jointakeoff.com/This is the RAG course from Takeoff. We're making the full videos fro...
·youtube.com·
The Complete RAG Course - Learn AI Skills
before you code, learn how computers work
before you code, learn how computers work
People hop on stream all the time and ask me, what is the fastest way to learn about the lowest level? How do I learn about how computers work. Check out this video to find out. Code: https://pastebin.com/raw/TpHbB91G 🏫 COURSES 🏫 Learn to code in C at https://lowlevel.academy 📰 NEWSLETTER 📰 Sign up for our newsletter at https://mailchi.mp/lowlevel/the-low-down 🛒 GREAT BOOKS FOR THE LOWEST LEVEL🛒 Blue Fox: Arm Assembly Internals and Reverse Engineering: https://amzn.to/4394t87 Practical Reverse Engineering: x86, x64, ARM, Windows Kernel, Reversing Tools, and Obfuscation : https://amzn.to/3C1z4sk Practical Malware Analysis: The Hands-On Guide to Dissecting Malicious Software : https://amzn.to/3C1daFy The Ghidra Book: The Definitive Guide: https://amzn.to/3WC2Vkg 🔥🔥🔥 SOCIALS 🔥🔥🔥 Low Level Merch!: https://lowlevel.store/ Follow me on Twitter: https://twitter.com/LowLevelTweets Follow me on Twitch: https://twitch.tv/lowlevellearning Join me on Discord!: https://discord.gg/gZhRXDdBYY
·youtube.com·
before you code, learn how computers work
Adding payments to your LLM agentic workflows
Adding payments to your LLM agentic workflows
This post discusses integrating the Stripe agent toolkit with large language models (LLMs) to enhance automation workflows, enabling financial services access, metered billing, and streamlined operations across agent frameworks.
·stripe.dev·
Adding payments to your LLM agentic workflows
NVIDIA AI Learning Essentials
NVIDIA AI Learning Essentials
Build skills, get certified, and learn from NVIDIA experts through hands-on self-paced courses and instructor-led workshops.
·nvidia.com·
NVIDIA AI Learning Essentials
AI Machine Learning Roadmap: Self Study AI!
AI Machine Learning Roadmap: Self Study AI!
Unlock the secrets to mastering Artificial Intelligence (AI) quickly with this self-study roadmap, based on the prestigious Stanford AI Graduate Certificate ...
·youtube.com·
AI Machine Learning Roadmap: Self Study AI!
Previously, RAG systems were the standard method for retrieving information from documents. However, if you are not repeatedly querying the same document, it may be more convenient and effective to just use long-context LLMs. For example, Llama 3.1 8B and Llama 3.2 1B/3B now…
Previously, RAG systems were the standard method for retrieving information from documents. However, if you are not repeatedly querying the same document, it may be more convenient and effective to just use long-context LLMs. For example, Llama 3.1 8B and Llama 3.2 1B/3B now…
— Sebastian Raschka (@rasbt)
·x.com·
Previously, RAG systems were the standard method for retrieving information from documents. However, if you are not repeatedly querying the same document, it may be more convenient and effective to just use long-context LLMs. For example, Llama 3.1 8B and Llama 3.2 1B/3B now…
(3) LlamaIndex 🦙 on X: "Check out this video from @thesourabhd on how to build AI agents using LlamaCloud plus @qdrant_engine! This deep dive covers: ➡️ Implementing semantic caching in agent systems to improve speed and efficiency ➡️ Advanced agent techniques like query routing, query decomposition, https://t.co/DVfK0FE0bD" / X
(3) LlamaIndex 🦙 on X: "Check out this video from @thesourabhd on how to build AI agents using LlamaCloud plus @qdrant_engine! This deep dive covers: ➡️ Implementing semantic caching in agent systems to improve speed and efficiency ➡️ Advanced agent techniques like query routing, query decomposition, https://t.co/DVfK0FE0bD" / X
This deep dive covers: ➡️ Implementing semantic caching in agent systems to improve speed and efficiency ➡️ Advanced agent techniques like query routing, query decomposition,… — LlamaIndex 🦙 (@llama_index)
·x.com·
(3) LlamaIndex 🦙 on X: "Check out this video from @thesourabhd on how to build AI agents using LlamaCloud plus @qdrant_engine! This deep dive covers: ➡️ Implementing semantic caching in agent systems to improve speed and efficiency ➡️ Advanced agent techniques like query routing, query decomposition, https://t.co/DVfK0FE0bD" / X
Nexa AI - The On-Device AI Open Source Community Building The Future. Explore Quantized AI Models On Edge | Nexa AI Model Hub For NLP, Computer Vision, Multimodality & On-Device AI
Nexa AI - The On-Device AI Open Source Community Building The Future. Explore Quantized AI Models On Edge | Nexa AI Model Hub For NLP, Computer Vision, Multimodality & On-Device AI
Nexa AI On-Device Model Hub: LLaMA, Stable Diffusion, Whisper & more. Pre-trained AI models for NLP, vision, multimodality.
·nexa.ai·
Nexa AI - The On-Device AI Open Source Community Building The Future. Explore Quantized AI Models On Edge | Nexa AI Model Hub For NLP, Computer Vision, Multimodality & On-Device AI
Transformers Inference Optimization Toolset
Transformers Inference Optimization Toolset
Large Language Models are pushing the boundaries of artificial intelligence, but their immense size poses significant computational challenges. As these models grow, so does the need for smart optimization techniques to keep them running efficiently on modern hardware. In this post, we’ll explore key optimization strategies that are making LLMs faster and more memory-efficient. We’ll start with a brief look at GPU memory hierarchy, which forms the foundation for many of these techniques. Then, we’ll explore algorithms that allow LLMs to process information more quickly and handle longer contexts. Understanding these techniques offers valuable insights helping to unlock the full potential of Large Language Models.
·astralord.github.io·
Transformers Inference Optimization Toolset
Build an end-to-end RAG solution using Knowledge Bases for Amazon Bedrock and the AWS CDK | Amazon Web Services
Build an end-to-end RAG solution using Knowledge Bases for Amazon Bedrock and the AWS CDK | Amazon Web Services
In this post, we demonstrate how to seamlessly automate the deployment of an end-to-end RAG solution using Knowledge Bases for Amazon Bedrock and the AWS Cloud Development Kit (AWS CDK), enabling organizations to quickly set up a powerful question answering system.
·aws.amazon.com·
Build an end-to-end RAG solution using Knowledge Bases for Amazon Bedrock and the AWS CDK | Amazon Web Services