Ford Motor Company

Data Engineer

Ford Motor Company

full-time

Posted on:

Location Type: Hybrid

Location: Naucalpan de Juárez • 🇲🇽 Mexico

Visit company website
AI Apply
Apply

Job Level

Mid-LevelSenior

Tech Stack

ApacheBigQueryCloudDockerETLGoogle Cloud PlatformJavaKafkaMicroservicesPySparkPythonSQLTerraform

About the role

  • Design and build production data engineering solutions on Google Cloud Platform (GCP) using services such as BigQuery, Dataflow, DataForm, Astronomer, Data Fusion, DataProc, Cloud Composer/Air Flow, Cloud SQL, Compute Engine, Cloud Functions, Cloud Run, Artifact Registry, GCP APIs, Cloud Build, App Engine, and real-time data streaming platforms like Apache Kafka and GCP Pub/Sub.
  • Design new solutions to better serve AI/ML needs.
  • Lead teams to expand our AI-enabled services.
  • Partner with governance teams to tackle key business needs.
  • Collaborate with stakeholders and cross-functional teams to gather and define data requirements and ensure alignment with business objectives.
  • Partner with analytics teams to understand how value is created using data.
  • Partner with central teams to leverage existing solutions to drive future products.
  • Design and implement batch, real-time streaming, scalable, and fault-tolerant solutions for data ingestion, processing, and storage.
  • Create insights into existing data to fuel the creation of new data products.
  • Perform necessary data mapping, impact analysis for changes, root cause analysis, and data lineage activities, documenting information flows.
  • Implement and champion an enterprise data governance model.
  • Actively promote data protection, sharing, reuse, quality, and standards to ensure data integrity and confidentiality.
  • Develop and maintain documentation for data engineering processes, standards, and best practices.
  • Ensure knowledge transfer and ease of system maintenance.
  • Utilize GCP monitoring and logging tools to proactively identify and address performance bottlenecks and system failures.
  • Provide production support by addressing production issues as per SLAs.
  • Optimize data workflows for performance, reliability, and cost-effectiveness on the GCP infrastructure.
  • Work within an agile product team.
  • Deliver code frequently using Test-Driven Development (TDD), continuous integration, and continuous deployment (CI/CD).
  • Continuously enhance your domain knowledge.
  • Stay current on the latest data engineering practices.
  • Contribute to the company's technical direction while maintaining a customer-centric approach.

Requirements

  • GCP certified Professional Data Engineer
  • Successfully designed and implemented data warehouses and ETL processes for over five years, delivering high-quality data solutions.
  • 5+ years of complex SQL development experience
  • 2+ experience with programming languages such as Python, Java, or Apache Beam.
  • Experienced cloud engineer with 3+ years of GCP expertise, specializing in managing cloud infrastructure and applications to production-scale solutions.
  • In-depth understanding of GCP’s underlying architecture and hands-on experience of crucial GCP services, especially those related to data processing (Batch/Real Time) leveraging Terraform, Big Query, Dataflow, Pub/Sub, Data form, astronomer, Data Fusion, DataProc, Pyspark, Cloud Composer/Air Flow, Cloud SQL, Compute Engine, Cloud Functions, Cloud Run, Cloud build and App Engine, alongside and storage including Cloud Storage
  • DevOps tools such as Tekton, GitHub, Terraform, Docker.
  • Expert in designing, optimizing, and troubleshooting complex data pipelines.
  • Experience developing and deploying microservices architectures leveraging container orchestration frameworks
  • Experience in designing pipelines and architectures for data processing.
  • Passion and self-motivation to develop/experiment/implement state-of-the-art data engineering methods/techniques.
  • Self-directed, work independently with minimal supervision, and adapts to ambiguous environments.
  • Evidence of a proactive problem-solving mindset and willingness to take the initiative.
  • Strong prioritization, collaboration & coordination skills, and ability to simplify and communicate complex ideas with cross-functional teams and all levels of management.
  • Proven ability to juggle multiple responsibilities and competing demands while maintaining a high level of productivity.
  • Master’s degree in computer science, software engineering, information systems, Data Engineering, or a related field.
  • Data engineering or development experience gained in a regulated financial environment.
  • Experience in coaching and mentoring Data Engineers
  • Project management tools like Atlassian JIRA
  • Experience working in an implementation team from concept to operations, providing deep technical subject matter expertise for successful deployment.
  • Experience with data security, governance, and compliance best practices in the cloud.
  • Experience using data science concepts on production datasets to generate insights
Benefits
  • GCP certified Professional Data Engineer
  • Successfully designed and implemented data warehouses and ETL processes for over five years, delivering high-quality data solutions.
  • 5+ years of complex SQL development experience
  • 2+ experience with programming languages such as Python, Java, or Apache Beam.
  • Experienced cloud engineer with 3+ years of GCP expertise, specializing in managing cloud infrastructure and applications to production-scale solutions.
  • In-depth understanding of GCP’s underlying architecture and hands-on experience of crucial GCP services, especially those related to data processing (Batch/Real Time) leveraging Terraform, Big Query, Dataflow, Pub/Sub, Data form, astronomer, Data Fusion, DataProc, Pyspark, Cloud Composer/Air Flow, Cloud SQL, Compute Engine, Cloud Functions, Cloud Run, Cloud build and App Engine, alongside and storage including Cloud Storage
  • DevOps tools such as Tekton, GitHub, Terraform, Docker.
  • Expert in designing, optimizing, and troubleshooting complex data pipelines.
  • Experience developing and deploying microservices architectures leveraging container orchestration frameworks
  • Experience in designing pipelines and architectures for data processing.
  • Passion and self-motivation to develop/experiment/implement state-of-the-art data engineering methods/techniques.
  • Self-directed, work independently with minimal supervision, and adapts to ambiguous environments.
  • Evidence of a proactive problem-solving mindset and willingness to take the initiative.
  • Strong prioritization, collaboration & coordination skills, and ability to simplify and communicate complex ideas with cross-functional teams and all levels of management.
  • Proven ability to juggle multiple responsibilities and competing demands while maintaining a high level of productivity.
  • Master’s degree in computer science, software engineering, information systems, Data Engineering, or a related field.
  • Data engineering or development experience gained in a regulated financial environment.
  • Experience in coaching and mentoring Data Engineers
  • Project management tools like Atlassian JIRA
  • Experience working in an implementation team from concept to operations, providing deep technical subject matter expertise for successful deployment.
  • Experience with data security, governance, and compliance best practices in the cloud.
  • Experience using data science concepts on production datasets to generate insights

Applicant Tracking System Keywords

Tip: use these terms in your resume and cover letter to boost ATS matches.

Hard skills
data engineeringSQLPythonJavaApache BeamGCP servicesdata processingdata pipelinesETL processesmicroservices architectures
Soft skills
problem-solvingself-motivationindependenceadaptabilitycollaborationcommunicationprioritizationcoachingmentoringproject management
Certifications
GCP certified Professional Data EngineerMaster’s degree in computer scienceMaster’s degree in software engineeringMaster’s degree in information systemsMaster’s degree in Data Engineering