Tech Stack
AirflowApacheCloudETLPostgresSQL
About the role
- Design, build, and maintain scalable data pipelines to integrate data from various business units into a unified data lakehouse
- Automate data workflows to ensure accurate, timely, and efficient data availability for reporting and analytics
- Optimize database performance, ensuring efficient data retrieval and fast query execution across large datasets
- Develop robust ETL/ELT processes to support internal analytical needs and external client data feeds
- Work closely with analysts, data scientists, and business stakeholders to deliver optimized standard reports and custom data feeds
- Maintain and enforce consistent metric definitions and canonical queries across reports and analytics
- Collaborate with cross-functional teams to maximize use of tools like Power BI, Snowflake, ClickHouse, and Looker
- Evaluate and recommend new technologies, tools, and best practices for data storage, transformation, and reporting
- Ensure data integrity, security, and governance protocols are upheld across all data assets and pipelines
Requirements
- Strong experience in SQL and database management (e.g., PostgreSQL, ClickHouse, Snowflake)
- Proficiency in data pipeline development and ETL/ELT processes
- Experience with workflow automation and orchestration tools (e.g., Apache Airflow, dbt, Airbyte, or similar)
- Familiarity with business intelligence tools such as Power BI, Looker, or other BI platforms
- Experience working with cloud-native data warehouse and lakehouse architectures
- Knowledge of data modeling and schema design
- Ability to work both independently and collaboratively in a fast-paced, evolving environment
- Strong problem-solving skills and attention to detail