Tech Stack
Amazon RedshiftAWSAzureCloudNumpyPandasPySparkPythonSQL
About the role
- Build, optimize, and maintain data pipelines and workflows in the cloud
- Design and implement datasets and models for analytics and BI consumption
- Ensure data governance, security, and quality across diverse datasets
- Automate data platform provisioning, deployment, and operations
- Implement and maintain CI/CD workflows to support reliable and scalable data operations
- Partner with teams to develop dashboards, reports, and insights for end users
- Collaborate on AI/agentic projects by enabling advanced data solutions
Requirements
- Expert-level SQL: queries, performance tuning, data modeling
- Expert-level Python: data engineering libraries (pandas, numpy, pyspark, pyarrow, etc.)
- Real experience with modern data platforms Databricks strongly preferred Snowflake or Redshift acceptable
- Proven, hands-on experience with CI/CD automation (GitHub Actions or similar tools)
- Strong background in cloud platforms (AWS or Azure), including IAM policies, Lambda, storage services
- Understanding of data governance, pipeline development, and optimization practices
- Excellent communication skills and ability to collaborate across teams
- Production experience with DBT (Data Build Tool) (strongly preferred)
- Dashboard/reporting development experience (QuickSight, Power BI, Sigma, etc.) (strongly preferred)
- RESTful service API development experience (building or integrating APIs for data services) (strongly preferred)
- Exposure to AI/agentic workflows (strongly preferred)