Client: https://www.deepchannel.com/

Role: Senior Data Pipeline Engineer

Objective: Developed and optimized a custom data pipeline using Python and Prefect to extract, transform, and load data from various sources into a Redshift data warehouse.

Technologies and Tools:

Key Accomplishments:

  1. Enhanced and maintained a custom data pipeline to extract files from a file system, perform data manipulations and standardization, and load the processed data into a Redshift warehouse.
  2. Created a modular data-processor component that could be built into a Docker image and deployed to AWS EC2 instances using a CI/CD pipeline.
  3. Ingested data from diverse data warehouses, including SharePoint, Snowflake, and Redshift, to feed the data pipeline.
  4. Built transformation functions to map and convert data from source systems to staging and aggregation layers.