Tags
Language
Tags
November 2025
Su Mo Tu We Th Fr Sa
26 27 28 29 30 31 1
2 3 4 5 6 7 8
9 10 11 12 13 14 15
16 17 18 19 20 21 22
23 24 25 26 27 28 29
30 1 2 3 4 5 6
    Attention❗ To save your time, in order to download anything on this site, you must be registered 👉 HERE. If you do not have a registration yet, it is better to do it right away. ✌

    ( • )( • ) ( ͡⚆ ͜ʖ ͡⚆ ) (‿ˠ‿)
    SpicyMags.xyz

    Data Processing with Spark Kafka (Data Engineering Vol2 AWS)

    Posted By: lucky_aut
    Data Processing with Spark Kafka (Data Engineering Vol2 AWS)

    Data Processing with Spark Kafka (Data Engineering Vol2 AWS)
    Published 4/2025
    Duration: 3h 35m | .MP4 1280x720, 30 fps(r) | AAC, 44100 Hz, 2ch | 1.28 GB
    Genre: eLearning | Language: English

    Batch & Stream Processing using Spark and Kafka on AWS

    What you'll learn
    - Deep dive on Spark and Kafka using AWS EMR, Glue, MSK
    - Understand Data Engineering (Volume 2) on AWS using Spark and Kafka
    - Batch and Stream processing using Spark and Kafka
    - Production level projects and hands-on to help candidates provide on-job-like training
    - Get access to datasets of size 100 GB - 200 GB and practice using the same
    - Learn Python for Data Engineering with HANDS-ON (Functions, Arguments, OOP (class, object, self), Modules, Packages, Multithreading, file handling etc.
    - Learn SQL for Data Engineering with HANDS-ON (Database objects, CASE, Window Functions, CTE, CTAS, MERGE, Materialized View etc.)
    - AWS Data Analytics services - S3, EMR, Glue, MSK

    Requirements
    - Good to have AWS and SQL knowledge

    Description
    This isVolume 2 of Data Engineeringcourse. In this course I will talk about Open Source Data Processing technologies -Spark and Kafka, which are the most used and most popular data processing frameworks forBatch & Stream Processing. In this course you will learnSpark from Level 100 to Level 400 with real-life hands on and projects.I will also introduce you to Data Lake on AWS (that is S3) & Data Lakehouse usingApache Iceberg.

    I will use AWS as the hosting platform and talk about AWS Services likeEMR, S3, Glue and MSK. I will also show you Spark integration with other services likeAWS RDS, Redshift and DynamoDB.

    You will get opportunities to do hands-on using large datasets (100 GB - 300 GB or more of data).This course will provide you hands-on exercises that match with real-time scenarios like Spark batch processing, stream processing, performance tuning, streaming ingestion, Window functions, ACID transactions on Iceberg etc.

    Some other highlights:

    5 Projects with different datasets. Total dataset size of 250 GB or more.

    Contains training of data modelling -Normalization & ER Diagramfor OLTP systems.Dimensional modellingfor OLAP/DWH systems.

    Other technologies covered - EC2, EBS, VPC and IAM.

    Optional Python Course

    Who this course is for:
    - Python developers, Application Developers, Big Data Developers
    - Data Engineers, Data Scientists, Data Analysts
    - Database Administrators, Big Data Administrators
    - Data Engineering Aspirants
    - Solutions Architect, Cloud Architect, Big Data Architect
    - Technical Managers, Engineering Managers, Project Managers
    More Info

    Please check out others courses in your favourite language and bookmark them
    English - German - Spanish - French - Italian
    Portuguese