Tags
    Author
      Technology
        Rating
        Pricing
        Sort By
        Video
        Results To Show
        Most Recent
        Most Popular
        Highest Rated
        Reset
        https://image.mux.com/7Nrk00Iu01uMR00DuMTIkcxZR4Yb100eXPPc8A5pGdUlVUM/thumbnail.png?time=0

        lesson

        Advanced RAGPower AI course

        - Intro to RAG and Why LLMs Need External Knowledge - LLM Limitations and How Retrieval Fixes Hallucinations - How RAG Combines Search + Generation Into One System - Fresh Data Retrieval to Overcome Frozen Training Cutoffs - Context Engineering for Giving LLMs the Right Evidence - Multi-Agent RAG and Routing Queries to the Right Tools - Retrieval Indexes: Vector DBs, APIs, SQL, and Web Search - Query Routing With Prompts and Model-Driven Decision Logic - API Calls vs RAG: When You Need Data vs Full Answers - Tool Calling for Weather, Stocks, Databases, and More - Chunking Long Documents Into Searchable Units - Chunk Size Trade-offs for Precision vs Broad Context - Metadata Extraction to Link Related Chunks Together - Semantic Search Using Embeddings for Nearest-Neighbor Retrieval - Image and Multimodal Handling for RAG Pipelines - Text-Based Image Descriptions vs True Image Embeddings - Query Rewriting for Broad, Vague, or Ambiguous Questions - Hybrid Retrieval Using Metadata + Embeddings Together - Rerankers to Push the Correct Chunk to the Top - Vector Databases and How They Index Embeddings at Scale - Term-Based vs Embedding-Based vs Hybrid Search - Multi-Vector RAG and When to Use Multiple Embedding Models - Retrieval Indexes Beyond Vector DBs: APIs, SQL, Search Engines - Generation Stage: Stitching Evidence Into Final Answers - Tool Calling With Multiple Retrieval Sources for Complex Tasks - Synthetic Data for Stress-Testing Retrieval Quality Early - RAG vs Fine-Tuning: When to Retrieve and When to Update the Model - Prompt Patterns for Retrieval-Driven Generation - Evaluating Retrieval: Recall, Relevance, and Chunk Quality - Building End-to-End RAG Systems for Real Applications

        https://image.mux.com/jz23JtpqMe9q9TUzCELOn6KH01AN6DTOr3CJB01lgrw02A/thumbnail.png?time=0

        lesson

        Advanced Prompt EngineeringPower AI course

        - Intro to Prompt Engineering and Why It Shapes Every LLM Response - How Prompts Steer the Probability Space of an LLM - Context Engineering for Landing in the Right “Galaxy” of Meaning - Normal Prompts vs Engineered Prompts and Why Specificity Wins - Components of a High-Quality Prompt: Instruction, Style, Output Format - Role-Based Prompting for Business, Coding, Marketing, and Analysis Tasks - Few-Shot Examples for Teaching Models How to Behave - Synthetic Data for Scaling Better Prompts and Personalization - Choosing the Right Model Using Model Cards and Targeted Testing - When to Prompt First vs When to Reach for RAG or Fine-Tuning - Zero-Shot, Few-Shot, and Chain-of-Thought Prompting Techniques - PAL and Code-Assisted Prompting for Higher Accuracy - Multi-Prompt Reasoning: Self-Consistency, Prompt Chaining, and Divide-and-Conquer - Tree-of-Thought and Branching Reasoning for Hard Problems - Tool-Assisted Prompting and External Function-Calling - DSPy for Automatic Prompt Optimization With Reward Functions - Understanding LLM Limitations: Hallucinations, Fragile Reasoning, Memory Gaps - Temperature, Randomness, and How to Control Output Stability - Defensive Prompting to Resist Prompt Injection and Attacks - Blocklists, Allowlists, and Instruction Defense for Safer Outputs - Sandwiching and Random Enclosure for Better Security - XML and Structured Tagging for Reliable, Parseable AI Output - Jailbreak Prompts and How Attackers Trick Models - Production-Grade Prompts for Consistency, Stability, and Deployment - LLM-as-Judge for Evaluating Prompt Quality and Safety - Cost Optimization: How Better Prompts Reduce Token Usage

        https://image.mux.com/7Nrk00Iu01uMR00DuMTIkcxZR4Yb100eXPPc8A5pGdUlVUM/thumbnail.png?time=0

        lesson

        RAG

        - Intro to RAG and Why LLMs Need External Knowledge - LLM Limitations and How Retrieval Fixes Hallucinations - How RAG Combines Search + Generation Into One System - Fresh Data Retrieval to Overcome Frozen Training Cutoffs - Context Engineering for Giving LLMs the Right Evidence - Multi-Agent RAG and Routing Queries to the Right Tools - Retrieval Indexes: Vector DBs, APIs, SQL, and Web Search - Query Routing With Prompts and Model-Driven Decision Logic - API Calls vs RAG: When You Need Data vs Full Answers - Tool Calling for Weather, Stocks, Databases, and More - Chunking Long Documents Into Searchable Units - Chunk Size Trade-offs for Precision vs Broad Context - Metadata Extraction to Link Related Chunks Together - Semantic Search Using Embeddings for Nearest-Neighbor Retrieval - Image and Multimodal Handling for RAG Pipelines - Text-Based Image Descriptions vs True Image Embeddings - Query Rewriting for Broad, Vague, or Ambiguous Questions - Hybrid Retrieval Using Metadata + Embeddings Together - Rerankers to Push the Correct Chunk to the Top - Vector Databases and How They Index Embeddings at Scale - Term-Based vs Embedding-Based vs Hybrid Search - Multi-Vector RAG and When to Use Multiple Embedding Models - Retrieval Indexes Beyond Vector DBs: APIs, SQL, Search Engines - Generation Stage: Stitching Evidence Into Final Answers - Tool Calling With Multiple Retrieval Sources for Complex Tasks - Synthetic Data for Stress-Testing Retrieval Quality Early - RAG vs Fine-Tuning: When to Retrieve and When to Update the Model - Prompt Patterns for Retrieval-Driven Generation - Evaluating Retrieval: Recall, Relevance, and Chunk Quality - Building End-to-End RAG Systems for Real Applications

        https://image.mux.com/jz23JtpqMe9q9TUzCELOn6KH01AN6DTOr3CJB01lgrw02A/thumbnail.png?time=0

        lesson

        Prompt Engineering

        - Intro to Prompt Engineering and Why It Shapes Every LLM Response - How Prompts Steer the Probability Space of an LLM - Context Engineering for Landing in the Right “Galaxy” of Meaning - Normal Prompts vs Engineered Prompts and Why Specificity Wins - Components of a High-Quality Prompt: Instruction, Style, Output Format - Role-Based Prompting for Business, Coding, Marketing, and Analysis Tasks - Few-Shot Examples for Teaching Models How to Behave - Synthetic Data for Scaling Better Prompts and Personalization - Choosing the Right Model Using Model Cards and Targeted Testing - When to Prompt First vs When to Reach for RAG or Fine-Tuning - Zero-Shot, Few-Shot, and Chain-of-Thought Prompting Techniques - PAL and Code-Assisted Prompting for Higher Accuracy - Multi-Prompt Reasoning: Self-Consistency, Prompt Chaining, and Divide-and-Conquer - Tree-of-Thought and Branching Reasoning for Hard Problems - Tool-Assisted Prompting and External Function-Calling - DSPy for Automatic Prompt Optimization With Reward Functions - Understanding LLM Limitations: Hallucinations, Fragile Reasoning, Memory Gaps - Temperature, Randomness, and How to Control Output Stability - Defensive Prompting to Resist Prompt Injection and Attacks - Blocklists, Allowlists, and Instruction Defense for Safer Outputs - Sandwiching and Random Enclosure for Better Security - XML and Structured Tagging for Reliable, Parseable AI Output - Jailbreak Prompts and How Attackers Trick Models - Production-Grade Prompts for Consistency, Stability, and Deployment - LLM-as-Judge for Evaluating Prompt Quality and Safety - Cost Optimization: How Better Prompts Reduce Token Usage

        https://image.mux.com/sqQ02pCMF902bHm014NVUyXjoKcq8V4hyA02geiJLqIHBJg/thumbnail.png?time=0

        lesson

        Synthetic DataPower AI course

        - Intro to Synthetic Data and Why It Matters in Modern AI - What Synthetic Data Really Is vs Common Misconceptions - How Synthetic Data Fills Gaps When Real Data Is Limited or Unsafe - The Synthetic Data Flywheel: Generate → Evaluate → Iterate - Using Synthetic Data Across Pretraining, Finetuning, and Evaluation - Synthetic Data for RAG: How It Stress-Tests Retrieval Systems - Fine-Tuning with Synthetic Examples to Update Model Behavior - When to Use RAG vs Fine-Tuning for Changing Information - Building RAG Systems Like Lego: LLM + Vector DB + Retrieval - How Vector Databases Reduce Hallucinations and Improve Accuracy - Generating Edge Cases, Adversarial Queries, and Hard Negatives - Control Knobs for Diversity: Intent, Persona, Difficulty, Style - Guardrails and Bias Control Using Prompt Engineering and DPO - Privacy Engineering with Synthetic Data for Safe Testing - Debugging AI Apps Using Synthetic Data Like a Developer Debugs Code - LLM-as-Judge for Fast, Cheap, Scalable Data Quality Checks - Axial Coding: Turning Model Failures Into Actionable Error Clusters - Evaluation-First Loops: The Only Way to Improve Synthetic Data Quality - Components of High-Quality Prompts for Synthetic Data Generation - User Query Generators for Realistic Customer Support Scenarios - Chatbot Response Generators for Complete and Partial Solutions - Error Analysis to Catch Hallucinations, Bias, and Structure Failures - Human + LLM Evaluation: Combining Experts With Automated Judges - Model Cards and Benchmarks for Understanding Model Capabilities

        https://image.mux.com/HvEMAiBHLB02ObLhq43qdyjQxDRrwdzraYkIdUzZGFBc/thumbnail.png?time=0

        lesson

        Neural Network FundamentalsPower AI course

        - Feedforward networks as transformer core - Linear layers for learned projections - Nonlinear activations enable expressiveness - SwiGLU powering modern FFN blocks - MLPs refine token representations - LayerNorm stabilizes deep training - Dropout prevents co-adaptation overfitting - Skip connections preserve information flow - Positional encoding injects word order - NLL loss guides probability learning - Encoder vs decoder architectures explained - FFNN + attention form transformer blocks

        https://image.mux.com/KvCgBJ4EblbZZ6nQ8PF9t49MS8vqJkB41q7p5wz3U4s/thumbnail.png?time=0

        lesson

        Introduction to Building an LLMPower AI course

        - Intuition for decoder-only LLMs - Tokens, embeddings, transformer pipeline - Autoregressive next-token generation - Generative AI modalities overview - Diffusion vs transformer model families - Inference flow and prompt processing - Build a real LLM inference API - Architecture: attention, context, decoding - Training phases: pretrain to RLHF - Vertical vs generic LLM design - Distillation, quantization, efficient scaling - Reasoning models: Chain of Thought and Test Time Compute - Hands on Exercises

        https://image.mux.com/K001j00cUK01HQowjP963j6FoTiW25kjlz007KaypIPM9nw/thumbnail.png?time=0

        lesson

        Technical Orientation (Python, Numpy, Probability, Statistics, Tensors)Power AI course

        - How AI Thinks in Numbers: Dot Products and Matrix Logic - NumPy Power-Tools: The Math Engine Behind Modern AI - Introduction To Machine Learning Libraries - Two and Three Dimensional Arrays - Data as Fuel: Cleaning, Structuring, and Transforming with Pandas - Normalization in Data Processing: Teaching Models to Compare Apples to Apples - Probability Foundations: How Models Reason About the Unknown - The Bell Curve in AI: Detecting Outliers and Anomalies - Evaluating Models Like a Scientist: Bootstrapping, T-Tests, Confidence Intervals - Transformers: The Architecture That Gave AI Its Brain - Diffusion Models: How AI Creates Images, Video, and Sound - Activation Functions: Teaching Models to Make Decisions - Vectors and Tensors: The Language of Deep Learning - GPUs, Cloud, and APIs: How AI Runs in the Real World

        https://image.mux.com/rL01dNFNmokUH02TE01D8yr5o6BAW1pDUE01B2lwW1C583Q/thumbnail.png?time=0

        lesson

        Attention LayerPower AI course

        - Why context is fundamental in LLMs - Limits of n-grams, RNNs, embeddings - Self-attention solves long-range context - QKV: query–key–value mechanics - Dynamic contextual embeddings per token - Attention weights determine word relevance - Multi-head attention = parallel perspectives - GQA reduces attention compute cost - Mixture-of-experts for specialized attention - Editing and modifying transformer layers - Decoder-only vs encoder–decoder framing - Building context-aware prediction systems

        https://image.mux.com/tW83E873KgzEG56rX7ZttQdCt9dfI01X00fNQEKICRqVc/thumbnail.png?time=0

        lesson

        Multimodal EmbeddingsPower AI course

        - Foundations of multimodal representation learning - Text, image, audio, video embeddings - Contrastive learning for cross-modal alignment - Shared latent spaces across modalities - Vision encoders and patch tokenization - Transformer encoders for text meaning - Audio preprocessing and spectral features - Time-series tokenization via SAX or VQ - Fusion modules for modality alignment - Cross-attention for integrated reasoning - Zero-shot retrieval and multimodal search - Real-world multimodal applications overview