
Mid/Senior Data Engineer
CodiLime
contract
Posted on:
Location Type: Remote
Location: Poland
Visit company websiteExplore more
Salary
💰 PLN 16,500 - PLN 25,000 per month
Job Level
About the role
- Designing, building, and maintaining scalable, end-to-end data pipelines for ingesting, cleaning, transforming, and integrating large structured and semi-structured datasets
- Optimizing data collection, processing, and storage workflows
- Conducting periodic data refresh processes (through data pipelines)
- Building a robust ETL infrastructure using SQL technologies
- Assisting with data migration to the new platform
- Automating manual workflows and optimizing data delivery
- Developing data transformation logic using SQL and DBT for Snowflake
- Designing and implementing scalable and high-performance data models
- Creating matching logic to deduplicate and connect entities across multiple sources
- Ensuring data quality, consistency, and performance to support downstream applications
- Orchestrating data workflows using Apache Airflow, running on Kubernetes
- Monitoring and troubleshooting data pipeline performance and operations
- Enabling integration of 3rd-party and pre-cleaned data into a unified schema with rich metadata and hierarchical relationships
- Writing data processing logic in Python
- Applying software engineering best practices: version control (Git), CI/CD pipelines (GitHub Actions), DevOps workflows
- Ensuring code quality using tools like SonarQube
- Documenting data processes and workflows
- Preparing the platform for future integrations (e.g., REST APIs, LLM/agentic AI)
Requirements
- Strong experience with Snowflake and dbt (must-have)
- Strong SQL skills, including experience with query optimization
- Experience with orchestration tools like Apache Airflow, Azure Data Factory (ADF), or similar
- Experience with Docker, Kubernetes, and CI/CD practices for data workflows
- Experience in working with large-scale datasets
- Very good understanding of data pipeline design concepts and best practices
- Very good coding skills in Python
- Ability to write clean, scalable, and testable code (including unit tests)
- Understanding and applying object-oriented programming (OOP)
- Experience with version control systems: Git
- Good knowledge of English (minimum C1 level)
Benefits
- Flexible working hours and approach to work: fully remote, in the office, or hybrid
- Professional growth supported by internal training sessions and a training budget
- Solid onboarding with a hands-on approach to give you an easy start
- A great atmosphere among professionals who are passionate about their work
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard Skills & Tools
SQLDBTPythonETLdata pipeline designquery optimizationobject-oriented programmingdata transformationdata quality assurancedata modeling
Soft Skills
problem-solvingattention to detailcommunicationcollaborationdocumentation