
Data Engineer
Ford Motor Company
full-time
Posted on:
Location Type: Hybrid
Location: Naucalpan de Juárez • 🇲🇽 Mexico
Visit company websiteJob Level
Mid-LevelSenior
Tech Stack
ApacheBigQueryCloudDockerETLGoogle Cloud PlatformJavaKafkaMicroservicesPySparkPythonSQLTerraform
About the role
- Design and build production data engineering solutions on Google Cloud Platform (GCP) using services such as BigQuery, Dataflow, DataForm, Astronomer, Data Fusion, DataProc, Cloud Composer/Air Flow, Cloud SQL, Compute Engine, Cloud Functions, Cloud Run, Artifact Registry, GCP APIs, Cloud Build, App Engine, and real-time data streaming platforms like Apache Kafka and GCP Pub/Sub.
- Design new solutions to better serve AI/ML needs.
- Lead teams to expand our AI-enabled services.
- Partner with governance teams to tackle key business needs.
- Collaborate with stakeholders and cross-functional teams to gather and define data requirements and ensure alignment with business objectives.
- Partner with analytics teams to understand how value is created using data.
- Partner with central teams to leverage existing solutions to drive future products.
- Design and implement batch, real-time streaming, scalable, and fault-tolerant solutions for data ingestion, processing, and storage.
- Create insights into existing data to fuel the creation of new data products.
- Perform necessary data mapping, impact analysis for changes, root cause analysis, and data lineage activities, documenting information flows.
- Implement and champion an enterprise data governance model.
- Actively promote data protection, sharing, reuse, quality, and standards to ensure data integrity and confidentiality.
- Develop and maintain documentation for data engineering processes, standards, and best practices.
- Ensure knowledge transfer and ease of system maintenance.
- Utilize GCP monitoring and logging tools to proactively identify and address performance bottlenecks and system failures.
- Provide production support by addressing production issues as per SLAs.
- Optimize data workflows for performance, reliability, and cost-effectiveness on the GCP infrastructure.
- Work within an agile product team.
- Deliver code frequently using Test-Driven Development (TDD), continuous integration, and continuous deployment (CI/CD).
- Continuously enhance your domain knowledge.
- Stay current on the latest data engineering practices.
- Contribute to the company's technical direction while maintaining a customer-centric approach.
Requirements
- GCP certified Professional Data Engineer
- Successfully designed and implemented data warehouses and ETL processes for over five years, delivering high-quality data solutions.
- 5+ years of complex SQL development experience
- 2+ experience with programming languages such as Python, Java, or Apache Beam.
- Experienced cloud engineer with 3+ years of GCP expertise, specializing in managing cloud infrastructure and applications to production-scale solutions.
- In-depth understanding of GCP’s underlying architecture and hands-on experience of crucial GCP services, especially those related to data processing (Batch/Real Time) leveraging Terraform, Big Query, Dataflow, Pub/Sub, Data form, astronomer, Data Fusion, DataProc, Pyspark, Cloud Composer/Air Flow, Cloud SQL, Compute Engine, Cloud Functions, Cloud Run, Cloud build and App Engine, alongside and storage including Cloud Storage
- DevOps tools such as Tekton, GitHub, Terraform, Docker.
- Expert in designing, optimizing, and troubleshooting complex data pipelines.
- Experience developing and deploying microservices architectures leveraging container orchestration frameworks
- Experience in designing pipelines and architectures for data processing.
- Passion and self-motivation to develop/experiment/implement state-of-the-art data engineering methods/techniques.
- Self-directed, work independently with minimal supervision, and adapts to ambiguous environments.
- Evidence of a proactive problem-solving mindset and willingness to take the initiative.
- Strong prioritization, collaboration & coordination skills, and ability to simplify and communicate complex ideas with cross-functional teams and all levels of management.
- Proven ability to juggle multiple responsibilities and competing demands while maintaining a high level of productivity.
- Master’s degree in computer science, software engineering, information systems, Data Engineering, or a related field.
- Data engineering or development experience gained in a regulated financial environment.
- Experience in coaching and mentoring Data Engineers
- Project management tools like Atlassian JIRA
- Experience working in an implementation team from concept to operations, providing deep technical subject matter expertise for successful deployment.
- Experience with data security, governance, and compliance best practices in the cloud.
- Experience using data science concepts on production datasets to generate insights
Benefits
- GCP certified Professional Data Engineer
- Successfully designed and implemented data warehouses and ETL processes for over five years, delivering high-quality data solutions.
- 5+ years of complex SQL development experience
- 2+ experience with programming languages such as Python, Java, or Apache Beam.
- Experienced cloud engineer with 3+ years of GCP expertise, specializing in managing cloud infrastructure and applications to production-scale solutions.
- In-depth understanding of GCP’s underlying architecture and hands-on experience of crucial GCP services, especially those related to data processing (Batch/Real Time) leveraging Terraform, Big Query, Dataflow, Pub/Sub, Data form, astronomer, Data Fusion, DataProc, Pyspark, Cloud Composer/Air Flow, Cloud SQL, Compute Engine, Cloud Functions, Cloud Run, Cloud build and App Engine, alongside and storage including Cloud Storage
- DevOps tools such as Tekton, GitHub, Terraform, Docker.
- Expert in designing, optimizing, and troubleshooting complex data pipelines.
- Experience developing and deploying microservices architectures leveraging container orchestration frameworks
- Experience in designing pipelines and architectures for data processing.
- Passion and self-motivation to develop/experiment/implement state-of-the-art data engineering methods/techniques.
- Self-directed, work independently with minimal supervision, and adapts to ambiguous environments.
- Evidence of a proactive problem-solving mindset and willingness to take the initiative.
- Strong prioritization, collaboration & coordination skills, and ability to simplify and communicate complex ideas with cross-functional teams and all levels of management.
- Proven ability to juggle multiple responsibilities and competing demands while maintaining a high level of productivity.
- Master’s degree in computer science, software engineering, information systems, Data Engineering, or a related field.
- Data engineering or development experience gained in a regulated financial environment.
- Experience in coaching and mentoring Data Engineers
- Project management tools like Atlassian JIRA
- Experience working in an implementation team from concept to operations, providing deep technical subject matter expertise for successful deployment.
- Experience with data security, governance, and compliance best practices in the cloud.
- Experience using data science concepts on production datasets to generate insights
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard skills
data engineeringSQLPythonJavaApache BeamGCP servicesdata processingdata pipelinesETL processesmicroservices architectures
Soft skills
problem-solvingself-motivationindependenceadaptabilitycollaborationcommunicationprioritizationcoachingmentoringproject management
Certifications
GCP certified Professional Data EngineerMaster’s degree in computer scienceMaster’s degree in software engineeringMaster’s degree in information systemsMaster’s degree in Data Engineering