Salary
💰 $138,900 - $203,900 per year
Tech Stack
AirflowAmazon RedshiftAWSCloudDistributed SystemsETLPySparkPythonSparkSQL
About the role
- Contribute to the design and growth of our Data Products and Data Warehouses around Content Performance and Content Engagement data.
- Design and develop scalable data warehousing solutions, building ETL pipelines in Big Data environments (cloud, on-prem, hybrid)
- Help architect data solutions/frameworks and define data models for the underlying data warehouse and data marts
- Collaborate with Data Product Managers, Data Architects and Data Engineers to design, implement, and deliver successful data solutions
- Maintain detailed documentation of your work and changes to support data quality and data governance
- Ensure high operational efficiency and quality of your solutions to meet SLAs and support commitment to our customers (Data Science, Data Analytics teams)
- Be an active participant and advocate of agile/scrum practice to ensure health and process improvements for your team
Requirements
- 5+ years of data engineering experience developing large data pipelines.
- Strong understanding of data modeling principles including Dimensional modeling, data normalization principles.
- Experience using analytic SQL, working with traditional relational databases and/or distributed systems (Snowflake or Redshift), required.
- Good understanding of SQL Engines and able to conduct advanced performance tuning.
- Ability to think strategically, analyze and interpret market and consumer information.
- Strong communication skills – written and verbal presentations.
- Comfortable working in a fast-paced and highly collaborative environment.
- Experience programming languages (e.g. Python, Pyspark), preferred.
- Experience with data orchestration/ETL tools (Airflow, Nifi), preferred.
- Experience with Snowflake, Databricks/EMR/Spark & Airflow a plus.
- Bachelor's Degree in Computer Science, Information Systems or related field, or equivalent work experience.