
Data Engineer
OLX Group
full-time
Posted on:
Location Type: Remote
Location: Poland
Visit company websiteExplore more
About the role
- - Enable efficient consumption of domain data as a product by delivering and promoting strategically designed actionable datasets and data models
- - Build, maintain, and improve rock-solid data pipelines using a broad range of technologies like AWS Redshift, Trino, Spark, Airflow, and Kafka streaming for real-time processing
- - Build data pipelines in our Lakehouse using Medallion architecture
- - Support teams without data engineers in building decentralised data solutions and product integrations, for example, around DynamoDB
- - Act as a data ambassador, promoting the value of data and our data platform among engineering teams and enabling cooperation
- - Cooperate with product teams to streamline source data so that data is cleaner earlier in the process.
- - Follow and promote data mesh principles around decentralised data ownership and ‘data as a product’
- - Enforce privacy and security standards (privacy by design)
- - Increase data literacy of the engineering organisation by hosting training and mentoring other engineers (both less experienced data engineers and software engineers)
Requirements
- - +3 years of experience building complex data pipelines, including extracting and processing data from external sources via API or real-time streaming tools
- - Entrepreneurial spirit and ability to think creatively; highly driven and self-motivated
- - Master of SQL language regardless of the underlying data technology. Writes efficient, easy-to-read, and crisp SQL code.
- - Knowledge of real-time data processing frameworks, such as Apache Kafka, AWS Kinesis or Apache Flink
- - Experience with Medallion architecture
- - Knowledge of Lakehouse architectures and Apache Iceberg format.
- - Experience with designing for scale and data growth in terms of storage and performance
- - Business and product understanding (why over what) and the ability to deal with ambiguity
- - Consistent track record of successful data projects delivered
- - Build rock-solid Python applications - writes understandable and maintainable code
- - Experience with cloud data technologies (AWS, GCP, or Azure Cloud), such as Redshift, Athena, Spark
- - Knowledge and curiosity about other big data technologies and tools
- - Experience with analytics and defining metrics
- - Experience with measuring data quality
- - Excellent written and verbal proficiency in English
- - You must be based in Poland
Benefits
- - Permanent contract (we are not offering the B2B option)
- - Salary + annual bonus up to 15% of your annual gross salary
- - Work with and impact all OLX services in Europe
- - Train and conference budget and free access to Udemy courses
- - Health insurance and life insurance
- - MacBook Pro along with the necessary accessories to make your work comfortable
- - 100% remote work (and the possibility to work from our offices in Warsaw and Poznan)
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard skills
data pipelinesSQLreal-time data processingMedallion architectureLakehouse architectureApache KafkaAWS RedshiftPythondata quality measurementanalytics
Soft skills
entrepreneurial spiritcreativityself-motivatedbusiness understandingdealing with ambiguitymentoringcommunicationdata literacy promotioncooperationleadership