Cloud Data Warehouse of Sparkify Data using Redshift
-
Updated
Jun 16, 2020 - Python
Cloud Data Warehouse of Sparkify Data using Redshift
FAA Airline On-Time Performance Data
Udacity Data Engineering Nanodegree
Data Warehouses and Business Intelligence Course Project @ FMI, Sofia University.
applying data warehouses tools and AWS to build an ETL pipeline for a database hosted on Redshift. loading data from AWS S3 bucket to staging tables on Redshift and executing SQL statements that create the analytics tables from these staging tables.
Applying data modeling to a NoSQL database with Apache Cassandra and build an ETL pipeline using Python. And modeling the data by creating tables in Apache Cassandra to run queries.
Loading data from Amazon S3 bucket to staging tables on Amazon Redshift then extracting Analytics tables from them.
This project demonstrates Snowflake table cloning and swapping techniques. It covers creating original and cloned tables, loading data from S3, verifying cloned data, and performing table swaps to efficiently exchange data between staging and production tables.
Add a description, image, and links to the staging-tables topic page so that developers can more easily learn about it.
To associate your repository with the staging-tables topic, visit your repo's landing page and select "manage topics."