Trafilea Tech E-commerce Group

Senior Data Engineer, BI

Trafilea Tech E-commerce Group

full-time

Posted on:

Location Type: Remote

Location: Argentina

Visit company website

Explore more

AI Apply
Apply

Job Level

About the role

  • Analysis, Design, implementation, and maintenance pipelines that produce business-critical data reliably and efficiently using cloud technologies.
  • Develop new ETLs (Extract, Transform, Load), using the current Apache Airflow.
  • Propose new initiatives to improve performance, scalability, reliability, and overall robustness.
  • Collect, process, and clean data from different sources using Python & SQL.
  • Work side by side with the main Architects and Developers to create and ensure best practices and guidelines are being used properly by all projects.
  • Assess and communicate effectively the effort for required developments.
  • Discover new data sources to improve new and existing data pipelines.
  • Be in charge of building and maintaining data pipelines and data models for new and existing projects.
  • Maintain detailed documentation of your work and changes to support data quality and governance.
  • Provide feedback and expert points of view as needed to help all data initiatives in the company.
  • Improve the quality of existing and new data processes (ETL), incorporating statistical process control, and creating alerts when anomalies are received from data sources at every step of the pipeline.
  • Create benchmark control of execution times for every pipeline, to control and identify potential availability issues.

Requirements

  • 5+ years of experience as a Data Engineer, ML Engineer, or similar.
  • 5+ years of experience using Python object-oriented programming and scripting.
  • Strong experience creating data pipelines and ETL processes in modern technologies using Apache Airflow as the main workflow management tool.
  • 3+ years of experience with the AWS data ecosystem (S3, Glue, Athena, Redshift, Lambda, EC2, RDS ...)
  • Experience with large-scale data and query optimization techniques using SQL.
  • Experience managing and monitoring data lakes systems.
  • Strong experience using source control management and CI/CD (GitHub actions/Gitlab pipelines).
  • Nice to have: Experience with containerized (Docker, Kubernetes) environments in AWS using EKS/ECS and ECR.
  • Knowledge and hands-on on cloud stream-processing systems.
  • Hand-on IAC solutions like Terraform or CloudFormation.
Benefits
  • Collaborate with world-class talents in a data-driven, dynamic, energetic work environment.
  • Opportunity to grow and develop both professionally and personally.
  • Safe space to be who you truly are, with a commitment to diversity, equity, and inclusion.
  • Openness to new ideas and initiatives.
  • Great benefits package including remote work, 15 working days of paid holidays, Learning subsidy, and more!
Applicant Tracking System Keywords

Tip: use these terms in your resume and cover letter to boost ATS matches.

Hard Skills & Tools
PythonSQLETLApache AirflowAWSdata pipelinesdata modelingdata qualityquery optimizationIAC
Soft Skills
communicationcollaborationproblem-solvinginitiativedocumentationfeedbackperformance improvementscalabilityreliabilityrobustness