The Practical AI Digest de Mo Bhuiyan via NotebookLM
Mo Bhuiyan via NotebookLM
Categorias: Tecnologia
Adicionar à Minha Lista
Ouvir último episódio:
This episode dives into strategies for fine-tuning gigantic AI models without needing massive compute. We explain parameter-efficient fine-tuning methods like LoRA (Low-Rank Adaptation), which freezes the original model and trains only small adapter weights, and QLoRA, which goes a step further by quantizing model parameters to 4-bit precision. You’ll learn why techniques like these have become essential for customizing large language models on modest hardware, how they preserve full performance, and what recent results (like fine-tuning a 65B model on a single GPU) mean for practitioners.
Episódios anteriores
-
12 - Efficient Fine-Tuning: Adapting Large Models on a Budget Tue, 03 Feb 2026
-
11 - Diffusion Models: AI Image Generation Through Noise Tue, 20 Jan 2026
-
10 - Graph Neural Networks: Learning from Connections, Not Just Data Tue, 30 Sep 2025
-
9 - Neuro-Symbolic AI: Combining Learning With Logic Tue, 16 Sep 2025
-
8 - LLMs in Chip Design: How AI Is Entering the Hardware Workflow Tue, 02 Sep 2025
-
7 - How Embeddings and Vector Databases Power Generative AI Tue, 19 Aug 2025
-
6 - Agentic AI: What Happens When Models Start Acting Tue, 05 Aug 2025
-
5 - Understanding Attention: Why Transformers Actually Work Tue, 22 Jul 2025
-
4 - Markov Chains, Monte Carlo, and HMC: A Deep Dive Tue, 08 Jul 2025
-
3 - The Model Context Protocol (MCP): Making LLMs Actually Useful Tue, 24 Jun 2025
-
2 - Generative Adversarial Networks (GANs) Explained: From DL Basics to Real-World Training Tips Tue, 10 Jun 2025
-
1 - Bayesian vs. Frequentist Thinking in Marketing Mix Modeling Tue, 27 May 2025
Mostrar mais episódios
5