Tech Stack
AirflowApacheCloudDockerElasticSearchETLGoogle Cloud PlatformKubernetesLinuxPythonScalaSparkSQLTerraform
About the role
- Support migration of the on-premise Cloudera-based data lake to a modern cloud-native architecture on Google Cloud Platform (GCP)
- Leverage containerization, managed services such as Dataproc, and infrastructure-as-code practices
- Contribute directly to delivery as part of the DataOps DCA Tribe
- Build a sustainable and compliant data platform to meet new compliance and regulatory requirements
- Work with Spark, NiFi, Airflow, Dataproc and related DataOps tooling
- Implement CI/CD pipelines, container orchestration, and monitoring/observability
Requirements
- Profile/Role: DataOps / DevOps
- Expected years of experience: 4-8 years
- Language skills: English required, Swedish is meritorious
- Utilization: 100%
- Start date: ASAP
- Apache Spark
- Cloud experience
- Containerization (Docker, Kubernetes)
- Infrastructure as Code (Terraform)
- CI/CD pipelines
- Python
- SQL
- Database management and ETL processes
- Meritorious qualifications: Apache NiFi, Apache Airflow, Cloudera ecosystem, Scala
- Cloud & Modern Stack Experience: Google Cloud Platform (GCP), especially Dataproc
- Observability and monitoring tools
- YAML
- Linux, Windows
- GCP Event Hubs, Elasticsearch, Kibana
- Command-line interface (CLI) proficiency
- Flexibility
- Strong English communication skills
- Delivery-oriented mindset