Tags
Language
Tags
May 2025
Su Mo Tu We Th Fr Sa
27 28 29 30 1 2 3
4 5 6 7 8 9 10
11 12 13 14 15 16 17
18 19 20 21 22 23 24
25 26 27 28 29 30 31
    Attention❗ To save your time, in order to download anything on this site, you must be registered 👉 HERE. If you do not have a registration yet, it is better to do it right away. ✌

    ( • )( • ) ( ͡⚆ ͜ʖ ͡⚆ ) (‿ˠ‿)
    SpicyMags.xyz

    RAG-LLM Evaluation & Test Automation for Beginners

    Posted By: Sigha
    RAG-LLM Evaluation & Test Automation for Beginners

    RAG-LLM Evaluation & Test Automation for Beginners
    2025-02-03
    MP4 | Video: h264, 1920x1080 | Audio: AAC, 44.1 KHz
    Language: English (US) | Size: 4.24 GB | Duration: 8h 43m

    Understand, Evaluate & Test RAG - LLM's (AI based Systems) from Scratch using RAGAS-Python-Pytest Framework

    What you'll learn
    How Custom Large Language Models (LLM) are designed using Retrieval Augmented Generation (RAG) Architecture
    Common Benchmarks/Metrics used in Evaluating RAG based LLM’s
    Introduction to RAGAS Evaluation framework for evaluating/test LLM’s
    Practical Scripts generation to automate and assert the Metrics Score of LLM’s.
    Automate Scenarios such as Single turn interactions and Multi turn interactions with LLM’s using RAGAS Framework
    Generate Test Data for evaluating the Metrics of LLM using RAGAS Framework.
    Create RAGAS Pytest Evaluation Framework to assert the Metrics of RAG- (Custom) LLM’s

    Requirements
    Python, PyTest basics are required. But wait! we have 2 dedicated sections at the end of this course which gives you necessary knowledge on Python & Pytest required to follow the course.
    Basic knowledge on API Testing

    Description
    LLMs are everywhere! Every business is building its own custom AI-based RAG-LLMs to improve customer service. But how are engineers testing them? Unlike traditional software testing, AI-based systems need a special methodology for evaluation.This course starts from the ground up, explaining the architecture of how AI systems (LLMs) work behind the scenes. Then, it dives deep into LLM evaluation metrics.This course shows you how to effectively use the RAGAS framework library to evaluate LLM metrics through scripted examples. This allows you to use Pytest assertions to check metric benchmark scores and design a robust LLM Test/evaluation automation framework.What will you learn from the course?High level overview on Large Language Models (LLM)Understand how Custom LLM’s are built using Retrieval Augmented Generation (RAG) ArchitectureCommon Benchmarks/Metrics used in Evaluating RAG based LLM’sIntroduction to RAGAS Evaluation framework for evaluating/test LLM’sPractical Scripts generation to automate and assert the Metrics Score of LLM’s.Automate Scenarios such as Single turn interactions and Multi turn interactions with LLM’s using RAGAS FrameworkGenerate Test Data for evaluating the Metrics of LLM using RAGAS Framework.By end of the course, you will be able to create RAGAS Pytest Evaluation Framework to assert the Metrics of RAG- (Custom) LLM’sImportant Note:This course covers Top 7 Metrics which are commonly used to Evaluate and test the LLM’s. Same logic can be applied to rest of any other metric evaluations.Handson Experience:Course provides Practice RAG -LLM for you for Handson, but at scripting phase, you need a basic subscription of Open AI to access their API’s (Minimal 10$ credit will suffice)Course Prerequisites:Python, PyTest basics are required to understand the Framework.We have 2 dedicated sections at the end of this course which gives you necessary knowledge on Python & Pytest required to follow the course.Basic knowledge on API Testing.

    Who this course is for:
    Software Engineers, Quality Assurance Engineers, Software Testers


    RAG-LLM Evaluation & Test Automation for Beginners


    For More Courses Visit & Bookmark Your Preferred Language Blog
    From Here: English - Français - Italiano - Deutsch - Español - Português - Polski - Türkçe - Русский