Design, develop, and maintain batch ETL/ELT pipelines and data workflows for large-scale datasets in AWS.
Implement and optimize data lakehouse architectures using Apache Iceberg on S3, ensuring schema evolution, partitioning strategies, and table maintenance.
Build and tune distributed data processing jobs with Apache Spark (PySpark or Spark) for performance and cost efficiency.
Orchestrate workflows using Apache Airflow, including DAG design, scheduling, and SLA monitoring.
Apply best practices in code quality, version control (Git/GitHub), and CI/CD for data engineering projects.
Ensure data quality, security, and compliance through validation, monitoring, and governance frameworks (Glue Catalog, IAM, encryption).
Collaborate with cross-functional teams (data scientists, analysts, architects) to deliver scalable and reliable data solutions.
Contribute to the development and optimization of analytics applications, ensuring they are powered by well-structured, high-quality data pipelines.
Continuously evaluate and adopt emerging technologies and AI-powered tools to improve productivity and maintain technical excellence.
Requirements
Bachelor's Degree in Computer Science, Engineering, or a related field.
5+ Years of experience working in data engineering, designing and developing large-scale data solutions.
5+ Years of experience working with Python for data manipulation, scripting, and integration tasks.
5+ Years of experience working with SQL & DBMS (e.g., PostgreSQL, MySQL, SQL Server), data modeling (Star Schema, Snowflake Schema) and query tuning and optimization.
3+ Years of experience working with Apache Spark (PySpark preferred).
3+ Years of experience building batch ETL/ELT pipelines for large-scale datasets.
3+ Years of experience working with AWS data services (S3, Athena/Presto, Glue, Lambda, CloudWatch) in production.
2+ Years of experience working with Apache Iceberg (table design, partition strategies, schema evolution, maintenance, ingestion pipelines into Apache Iceberg on S3).
2+ Years of experience working with AWS EMR as the execution platform for big data workloads.
2+ Years of experience orchestrating data workflows with Apache Airflow (DAG design, scheduling, backfills, SLAs).
2+ Years of experience working with Git/GitHub (branching strategies, Pull request reviews, CI/CD for data pipelines).
Experience designing efficient ingestion pipelines into analytical systems.
Proficiency in logging, auditing, and monitoring for data pipelines.
Experience with data cleansing, validation, and transformation for analytical/reporting systems.
Familiarity with data security and privacy practices.
Solid understanding of cloud-native analytics architectures (data lake/lakehouse, ELT patterns).
Proven ability to leverage AI-powered assistants (e.g., GitHub Copilot, AI coding agents) as part of the engineering workflow, demonstrating comfort and efficiency in using these tools to accelerate delivery and maintain code quality.
Advanced English Level is required for this role, as you will work with US clients. Effective communication in English is essential to deliver the best solutions to our clients and expand your horizons.
Benefits
Competitive USD salary💲 – We value your skills and contributions!
100% remote work🏢 – While you can work from anywhere, you’re always welcome to connect with teammates and grow your network at our coworking spaces across LATAM!
Paid time off – Take the time you need according to your country’s regulations, all while receiving your full salary. Rest, recharge, and come back stronger!
National Holidays celebrated🌴 – Take time off to celebrate important events and traditions with loved ones, fully embracing your culture.
Sick leave – Focus on your health without the stress. Take the necessary time to recover and feel better.
Refundable Annual Credit – Spend it on the perks you love to enhance your work-life balance!
Team-building activities – Join us for coffee breaks, tech talks, and after-work gatherings to bond with your Nearsure family and feel part of our vibrant community.
Birthday day off🎂 – Enjoy an extra day off during your birthday week to celebrate in style with friends and family!
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.