Example end to end data engineering project.
- Updated
Dec 8, 2022 - Python
Example end to end data engineering project.
Code for "Efficient Data Processing in Spark" Course
Ready to run docker-compose configuration for ML Flow with Mysql and Minio S3
Sample project to demonstrate data engineering best practices
Nyc_Taxi_Data_Pipeline - DE Project
Minio Backend for Django
Unified cloud storage API for storage services.
A self-contained, ready to run Airflow ELT project. Can be run locally or within codespaces.
Ansible role for installing and configuring Minio
Arquitetura CRM de Baixo Custo com Gen AI, projetada para startups que precisam processar e analisar dados de vendas de forma eficiente.
📈 A scalable, production-ready data pipeline for real-time streaming & batch processing, integrating Kafka, Spark, Airflow, AWS, Kubernetes, and MLflow. Supports end-to-end data ingestion, transformation, storage, monitoring, and AI/ML serving with CI/CD automation using Terraform & GitHub Actions.
Full stack data engineering tools and infrastructure set-up
📡 Real-time data pipeline with Kafka, Flink, Iceberg, Trino, MinIO, and Superset. Ideal for learning data systems.
Creation of a data lakehouse and an ELT pipeline to enable the efficient analysis and use of data
Maternal Health Risk prediction MLOps pipeline
Add a description, image, and links to the minio topic page so that developers can more easily learn about it.
To associate your repository with the minio topic, visit your repo's landing page and select "manage topics."