GenAI

GenAI

416 bookmarks
Newest
Fine Tune DeepSeek R1 | Build a Medical Chatbot
Fine Tune DeepSeek R1 | Build a Medical Chatbot
In this video, we show you how to fine-tune DeepSeek R1, an open-source reasoning model, using LoRA (Low-Rank Adaptation). We'll also be using Kaggle, Hugging Face and Weights & Biases. We walk you through data preparation, model configuration, and optimization, including advanced techniques like four-bit quantization for efficient training on consumer GPUs. By the end of this tutorial, you’ll be equipped with the skills to customize DeepSeek R1 for your own specialized tasks, such as medical reasoning. 🔗 Resources & Tutorials Kaggle Notebook: https://www.kaggle.com/code/aan1994/fine-tuning-deepseek-r1-reasoning-model-youtube How Transformers Work: https://www.datacamp.com/tutorial/how-transformers-work Fine-Tuning DeepSeek R1 Reasoning Model: https://www.datacamp.com/tutorial/fine-tuning-deepseek-r1-reasoning-model DeepSeek R1 Blog Overview: https://www.datacamp.com/blog/deepseek-r1 Understanding Janus Pro: https://www.datacamp.com/blog/janus-pro DeepSeek R1 Project Walkthrough: https://www.datacamp.com/tutorial/deepseek-r1-project DeepSeek vs ChatGPT: https://www.datacamp.com/blog/deepseek-vs-chatgpt Qwen-2.5 MAX Model: https://www.datacamp.com/blog/qwen-2-5-max DeepSeek R1 Ollama Tutorial: https://www.datacamp.com/tutorial/deepseek-r1-ollama 📕 Chapters 00:00 Introduction 00:30 Why Fine-Tuning DeepSeek Matters 02:30 LoRA Explained with a PS5 Factory Analogy 05:20 Tools & Setup Overview 09:00 Loading DeepSeek R1 Model and Tokenizer 16:10 Formatting Data for Fine-Tuning 23:00 Applying LoRA for Efficient Updates 34:00 Configuring Training Parameters 43:15 Running the Fine-Tuning Process on Kaggle 46:00 Comparing Model Performance After Fine-Tuning 47:50 Final Thoughts on Future Models 📱 Follow Us on Social Media Facebook: https://www.facebook.com/datacampinc/ Twitter: https://twitter.com/datacamp LinkedIn: https://www.linkedin.com/school/datacampinc/ Instagram: https://www.instagram.com/datacamp/ #deepseek #DeepSeekR1 #FineTuningAI #LearnAI #MachineLearning #Transformers #HuggingFace #Kaggle #WeightsAndBiases #LoRA #LargeLanguageModels #DeepSeekTutorial #AIResearch #AIOptimization #DataScience
·youtu.be·
Fine Tune DeepSeek R1 | Build a Medical Chatbot
transformerlab/transformerlab-app: Open Source Application for Advanced LLM Engineering: interact, train, fine-tune, and evaluate large language models on your own computer.
transformerlab/transformerlab-app: Open Source Application for Advanced LLM Engineering: interact, train, fine-tune, and evaluate large language models on your own computer.
Open Source Application for Advanced LLM Engineering: interact, train, fine-tune, and evaluate large language models on your own computer. - transformerlab/transformerlab-app
·t.co·
transformerlab/transformerlab-app: Open Source Application for Advanced LLM Engineering: interact, train, fine-tune, and evaluate large language models on your own computer.
Agents
Agents
Foundation models enable many new application interfaces, but one that has especially grown in popularity is the conversational interface, such as with chatbots and assistants. The conversational interface makes it easier for users to give feedback but harder for developers to extract signals. This post will discuss what conversational AI feedback looks like and how to design a system to collect the right feedback without hurting user experience.
·huyenchip.com·
Agents
Knowledge Graph-Enhanced RAG
Knowledge Graph-Enhanced RAG
Upgrade your RAG applications with the power of knowledge graphs./b Retrieval Augmented Generation (RAG) is a great way to harness the power of generative AI for information not contained in a LLM’s training data and to avoid depending on LLM for factual information. However, RAG only works when you can quickly identify and supply the most relevant context to your LLM. Knowledge Graph-Enhanced RAG/i shows you how to use knowledge graphs to model your RAG data and deliver better performance, accuracy, traceability, and completeness. Inside Knowledge Graph-Enhanced RAG/i you’ll learn: The benefits of using Knowledge Graphs in a RAG system/li How to implement a GraphRAG system from scratch/li The process of building a fully working production RAG system/li Constructing knowledge graphs using LLMs/li Evaluating performance of a RAG pipeline/li /ul Knowledge Graph-Enhanced RAG/i is a practical guide to empowering LLMs with RAG. You’ll learn to deliver vector similarity-based approaches to find relevant information, as well as work with semantic layers, and generate Cypher statements to retrieve data from a knowledge graph.
·manning.com·
Knowledge Graph-Enhanced RAG