Tech Stack
AirflowApacheAWSAzureCloudDockerGrafanaOraclePostgresPrometheusPythonSparkSQL
About the role
- Design, develop, and maintain robust, scalable, and secure data pipelines and architectures.
- Collaborate with the data team and business stakeholders to ensure data availability and reliability.
- Develop, orchestrate, and monitor data pipelines using Apache Airflow.
- Implement monitoring and alerting routines (structured logs, metrics, and notifications).
- Model and implement bronze, silver, and gold layers in a Databricks-based data lake.
- Build and maintain data integrations from REST APIs, relational databases, CSV/Excel/Parquet files, and message queues.
- Ensure pipeline observability and quality with unit and integration tests and up-to-date technical documentation.
Requirements
- Proficiency in Python for developing pipelines and automations.
- Advanced SQL knowledge (complex queries, tuning, indexes, CTEs, partitioning, materialized views).
- Strong experience with Apache Airflow (orchestration, version control, monitoring).
- Experience with PostgreSQL and data modeling for data lakes.
- Familiarity with object storage (S3 or equivalent) and optimized file formats (Parquet, ORC, GZIP).
- Experience with Git and good code versioning practices.
- Knowledge of Oracle/PLSQL is a plus.
- Experience with Databricks or similar environments (Spark, Delta Lake) is desirable.
- Familiarity with Docker, DBT, and monitoring with Prometheus/Grafana is desirable.
- Bachelor's degree in Computer Science, Software Engineering, Information Systems, or related fields.
- Cloud or data engineering certifications (AWS, Azure, Databricks, etc.) are a plus.
- Health insurance with low co-payment.
- Company-paid dental plan.
- Company-paid life insurance for protection and peace of mind.
- Flexible benefit for meals or groceries.
- Gympass and Zenklub included in our benefits.
- Day off on your birthday.
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard skills
PythonSQLApache AirflowPostgreSQLDatabricksDockerDBTPrometheusGrafanadata modeling
Certifications
AWSAzureDatabricks