Tech Stack
AirflowAmazon RedshiftApacheAWSETLHadoopKafkaPythonSparkSQL
About the role
- Design, build, and maintain scalable, efficient data infrastructures and pipelines.
- Develop and optimize processes for collecting, storing, processing, and analyzing large volumes of data.
- Ensure data quality, integrity, and governance by implementing cleansing, validation, and monitoring practices.
- Collaborate with data science, engineering, and analytics teams to provide consistent and accessible data for advanced analyses.
- Explore and apply new technologies and approaches to solve complex data challenges.
- Maintain and develop prompts, contributing to the use of Generative AI in data solutions and automation.
Requirements
- Strong knowledge of SQL and experience with programming languages such as Python.
- Experience with Power BI.
- Experience with ETL tools, big data technologies (Hadoop, Spark, Kafka), and data warehouses (e.g., Amazon Redshift).
- Familiarity with cloud computing concepts and experience with AWS.
- Proficiency with tools such as Glue, Athena, Apache Airflow, or equivalents for data processing and orchestration.
- Desirable: knowledge of and interest in Generative AI and prompt engineering.
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard skills
SQLPythonETLHadoopSparkKafkaAmazon RedshiftAWSGlueApache Airflow
Soft skills
collaborationdata governancedata qualityproblem-solvinganalytical thinking