* Cantinho Satkeys

Refresh History
  • okapa:
    24 de Dezembro de 2025, 19:01
  • sacana10: A todos um feliz natal
    24 de Dezembro de 2025, 17:57
  • cereal killa: dgtgtr passei por ca para vos desejar feliz natal e familias  :smiles_natal:
    24 de Dezembro de 2025, 15:46
  • bruno mirandela: deso a todos um feliz natal
    24 de Dezembro de 2025, 14:31
  • FELISCUNHA: ghyt74   :34rbzg9:  e bom natal  :13arvoresnatalmagiagifs:
    24 de Dezembro de 2025, 10:15
  • tgh12: mikrotik
    24 de Dezembro de 2025, 07:49
  • tgh12: Spanish for Beginners: Spanish from 0 to Conversational
    24 de Dezembro de 2025, 04:57
  • JPratas: try65hytr Pessoal  4tj97u<z
    24 de Dezembro de 2025, 03:03
  • m1957: Para toda a equipa e membros deste fórum, desejo um Natal feliz e que o novo ano de 2026, seja muito próspero a todos os níveis.
    24 de Dezembro de 2025, 00:47
  • FELISCUNHA: Bom dia pessoal   :34rbzg9:
    22 de Dezembro de 2025, 10:35
  • j.s.: :13arvoresnatalmagiagifs:
    21 de Dezembro de 2025, 19:01
  • j.s.: try65hytr a todos  :smiles_natal: :smiles_natal:
    21 de Dezembro de 2025, 19:01
  • FELISCUNHA: ghyt74  49E09B4F  e bom fim de semana  4tj97u<z
    20 de Dezembro de 2025, 11:20
  • JPratas: try65hytr Pessoal  2dgh8i k7y8j0 classic dgf64y
    19 de Dezembro de 2025, 05:26
  • cereal killa: ghyt74 e boa semana de chuva e frio  RGG45wj erfb57j
    15 de Dezembro de 2025, 11:26
  • FELISCUNHA: Votos de um santo domingo para todo o auditório  4tj97u<z
    14 de Dezembro de 2025, 09:28
  • j.s.: tenham um excelente fim de semana com muitas comprinhas  :13arvoresnatalmagiagifs: sdfgsdg
    13 de Dezembro de 2025, 14:58
  • j.s.: dgtgtr a todos  :smiles_natal:
    13 de Dezembro de 2025, 14:57
  • FELISCUNHA: dgtgtr   49E09B4F  e bom fim de semana   :34rbzg9:
    13 de Dezembro de 2025, 12:29
  • JPratas: try65hytr Pessoal  4tj97u<z 2dgh8i classic bve567o+
    12 de Dezembro de 2025, 05:34

Autor Tópico: Writing production-ready ETL pipelines in Python / Pandas  (Lida 154 vezes)

0 Membros e 1 Visitante estão a ver este tópico.

Offline mitsumi

  • Sub-Administrador
  • ****
  • Mensagens: 129146
  • Karma: +0/-0
Writing production-ready ETL pipelines in Python / Pandas
« em: 17 de Julho de 2021, 18:27 »

MP4 | Video: h264, 1280x720 | Audio: AAC, 44.1 KHz, 2 Ch
Genre: eLearning | Language: English + srt | Duration: 78 lectures (7h 3m) | Size: 2.43 GB
Learn how to write professional ETL pipelines using best practices in Python and Data Engineering

What you'll learn:
How to write professional ETL pipelines in Python.
Steps to write production level Python code.
How to apply functional programming in Data Engineering.
How to do a proper object oriented code design.
How to use a meta file for job control.
Coding best practices for Python in ETL/Data Engineering.
How to implement a pipeline in Python extracting data from an AWS S3 source, transforming and loading the data to another AWS S3 target.

Requirements
Basic Python and Pandas knowledge is desirable.
Basic ETL and AWS S3 knowledge is desirable.

Description
This course will show each step to write an ETL pipeline in Python from scratch to production using the necessary tools such as Python 3.9, Jupyter Notebook, Git and Github, Visual Studio Code, Docker and Docker Hub and the Python packages Pandas, boto3, pyyaml, awscli, jupyter, pylint, moto, coverage and the memory-profiler.

Two different approaches how to code in the Data Engineering field will be introduced and applied - functional and object oriented programming.

Best practices in developing Python code will be introduced and applied:

design principles

clean coding

virtual environments

project/folder setup

configuration

logging

exeption handling

linting

dependency management

performance tuning with profiling

unit testing

integration testing

dockerization

What is the goal of this course?

In the course we are going to use the Xetra dataset. Xetra stands for Exchange Electronic Trading and it is the trading platform of the Deutsche Börse Group. This dataset is derived near-time on a minute-by-minute basis from Deutsche Börse's trading system and saved in an AWS S3 bucket available to the public for free.

The ETL Pipeline we are going to create will extract the Xetra dataset from the AWS S3 source bucket on a scheduled basis, create a report using transformations and load the transformed data to another AWS S3 target bucket.

The pipeline will be written in a way that it can be deployed easily to almost any production environment that can handle containerized applications. The production environment we are going to write the ETL pipeline for consists of a GitHub Code repository, a DockerHub Image Repository, an execution platform such as Kubernetes and an Orchestration tool such as the container-native Kubernetes workflow engine Argo Workflows or Apache Airflow.

So what can you expect in the course?

You will receive primarily practical interactive lessons where you have to code and implement the pipeline and theory lessons when needed. Furthermore you will get the python code for each lesson in the course material, the whole project on GitHub and the ready to use docker image with the application code on Docker Hub.

There will be power point slides for download for each theoretical lesson and useful links for each topic and step where you find more information and can even dive deeper.

Who this course is for
Data engineers, scientists and developers who want to write professional production-ready data pipelines in Python.
Everyone who is interested in writing data pipelines in Python that are ready for production.


Download link:
Só visivel para registados e com resposta ao tópico.

Only visible to registered and with a reply to the topic.

Links are Interchangeable - No Password - Single Extraction