ETL Developer, VM
Job Description:
• Develop and automate ETL workflows using AWS Glue (Python/PySpark) to ingest, transform, and load data into Redshift and S3.
• Monitor and troubleshoot pipelines, ensuring data accuracy and SLA adherence.
• Optimize S3 storage (partitioning, lifecycle policies) for cost-effective data lake management.
• Collaborate on data modeling (star schema, dimensional modeling) for analytics-ready datasets.
• Document data flows, ERDs, and lineage for clarity and governance.
• Profile and validate data to catch issues early and maintain trust in reporting.
• Support BI teams by resolving data discrepancies and improving pipeline efficiency.
• Conduct ad hoc analysis and exploratory data work as needed.
• Contribute to the documentation of system specifications, testing procedures, and operational support materials.
• Participate in code reviews and continuous improvement of development standards.
• Provide production support and assist with troubleshooting and resolving data issues.
Requirements:
• Bachelor's degree in Computer Science or a related field, or equivalent work experience.
• 3+ years of related work experience as an ETL Developer, with AWS cloud experience (Glue, Redshift, S3).
• Strong SQL & Python (PySpark preferred).
• Experience building batch ETL pipelines (not just drag-and-drop tools).
• Knowledge of data warehousing (ODS, dimensional modeling, slowly changing dimensions).
• Ability to troubleshoot pipeline failures and optimize performance/cost.
• Clear communicator who can explain technical details to non-engineers.
• Excellent problem-solving and analytical skills.
• Strong organizational skills and attention to detail.
Benefits:
• Salary in USD
• Long-term
• Flexible schedule (within US Time zones)
• 100% Remote
Apply tot his job
Apply To this Job