Unilever

CD Excellence Data Engineering Associate

Unilever

full-time

Posted on:

Location Type: Remote

Location: United States

Visit company website

Explore more

AI Apply
Apply

Salary

💰 $71,400 - $107,200 per year

About the role

  • Design, build, and maintain scalable data pipelines and solutions on Microsoft Azure or similar cloud platforms.
  • Develop and optimize ETL/ELT workflows to support high-volume, high-velocity data ingestion.
  • Implement robust data models and structures that support analytics, reporting, and machine learning workloads.
  • Integrate new data sources—internal and external—into the enterprise data ecosystem to expand data availability and unlock new business insights.
  • Partner with product, engineering, business and global teams to identify opportunities for new datasets and ensure seamless onboarding.
  • Establish scalable frameworks for data discovery, cataloging, and lineage to support enterprise wide data growth.
  • Automate data workflows, quality checks, and monitoring using Cloud native tools and Databricks capabilities.
  • Collaborate with data scientists to operationalize AI models using Databricks, Azure Machine Learning, or similar platforms.
  • Ensure data readiness, reliability, and accessibility to accelerate AI adoption and experimentation.
  • Contribute to the development of an enterprise AI strategy by identifying data gaps, opportunities, and scalable patterns.
  • Work closely with cross functional teams to translate business requirements into scalable data solutions.
  • Provide technical guidance and best practices on Azure and Databricks to engineering and analytics teams.
  • Participate in code reviews, architecture discussions, and continuous improvement initiatives.
  • Assist in identifying & defining new systems functionality within the Go To Market technology stack.
  • Improve processes/workflows within the Sales organization in regard to Sales applications and system support.
  • User support - troubleshooting, identifying problems and working with Local & Global IT to resolve technical issues and work with users to provide proper training.
  • Assist in the analysis of underlying system issues arising from investigations into requirements and problems, and identify available solutions for consideration.

Requirements

  • Technical aptitude and the ability to drive business value through focused technology solutions.
  • Deep hands-on experience with cloud services such as Azure data services (e.g., Data Factory, Databricks, ADLS, Synapse, Azure SQL).
  • Strong proficiency in building scalable ETL/ELT pipelines using Databricks (APIs, PySpark, Spark SQL, Delta Lake).
  • Solid understanding of distributed computing, data lakehouse architecture, Unity Catalog and modern data engineering patterns.
  • Ability to design and optimize data models that support analytics, reporting, and machine learning workloads.
  • Strong SQL and Python skills, with the ability to write clean, efficient, production ready code.
  • Proven ability to onboard new data sources, integrate APIs, and work with structured, semi-structured data.
  • Experience designing frameworks for data ingestion, metadata management, and data lineage.
  • Comfort working with large-scale datasets and evolving data ecosystems.
  • Experience automating data workflows, quality checks, and monitoring using Azure-native tools.
  • Familiarity with CI/CD practices for data engineering (e.g., GitHub Actions, Azure DevOps).
  • Ability to build resilient, self-healing pipelines that minimize manual intervention.
  • Strong focus on performance tuning, cost optimization, and operational reliability.
  • Ability to build and maintain feature pipelines that support ML and AI initiatives.
  • Experience collaborating with data scientists to operationalize models in Databricks or Azure ML.
  • Understanding of how data quality, structure, and availability impact AI outcomes.
  • Curiosity and initiative to identify new data opportunities that unlock AI use cases.
  • Strong ability to turning data into insights and communicate actionable business narratives.
  • Ability to quickly adapt new AI technologies around LLMs, MCPs, prompt engineering and advanced data modelling.
  • Interest in developing a deep understanding of the Foodservice industry and its go to market use cases.
  • Ability to lead and execute multiple projects simultaneously.
  • Strong business partnering and communication skills.
Benefits
  • Health insurance (including prescription drug, dental, and vision coverage)
  • Retirement savings benefits
  • Life insurance and disability benefits
  • Parental leave
  • Sick leave
  • Paid vacation and holidays
  • Access to numerous voluntary benefits
Applicant Tracking System Keywords

Tip: use these terms in your resume and cover letter to boost ATS matches.

Hard Skills & Tools
ETLELTdata pipelinesdata modelsSQLPythonDatabricksAzure Machine LearningAPIsPySpark
Soft Skills
technical aptitudebusiness valuecollaborationcommunicationcuriosityinitiativeadaptabilityleadershipproblem-solvingproject management