Tech Stack
AirflowAmazon RedshiftAWSETLJavaPHPPythonScalaSQLTerraform
About the role
- Design, build, and maintain scalable, reusable data pipelines and architectures
- Ensure data quality, governance, and reliability across the organization
- Provide strategic guidance on data solutions to support business growth
- Lead high-impact projects and cross-functional collaborations with engineers, product managers, and stakeholders
- Own architecture and design of data systems and drive adoption of new tools and technologies
- Establish best practices, reusable frameworks, and high standards for data engineering
- Mentor engineers, conduct code reviews, and foster technical excellence
- Work on projects ranging from building new data solutions to improving existing infrastructure
Requirements
- Strong expertise in SQL and data modeling
- Proficiency in at least one programming language: Python, Scala, or Java
- Deep experience with AWS data services (Redshift, Glue, EMR, Athena, S3, Lambda, ECS)
- Hands-on experience with distributed data processing and workflow orchestration (e.g., Airflow, Step Functions)
- Proven track record designing ETL/ELT pipelines, data warehouses, and CDC solutions
- Knowledge of infrastructure-as-code (Terraform, CDK) a plus
- Ability to solve complex data challenges at scale, including real-time analytics
- Strong architecture, design, and problem-solving skills
- Excellent collaboration and communication skills with technical and non-technical stakeholders
- Experience mentoring engineers and conducting code reviews
- Prior experience with ThoughtSpot as an analytics platform (desirable)