Tech Stack
BigQueryCloudETLGoogle Cloud PlatformNumpyPandasPythonPyTorchScikit-LearnSQLTableauTensorflow
About the role
- Design, develop, and maintain scalable data pipelines and ETL/ELT processes on Google Cloud Platform to ingest, process, and store data from diverse sources, ensuring data quality and reliability.
- Leverage expert-level SQL and Python for complex data extraction, transformation, cleansing, and analysis to prepare data for analytical and machine learning applications.
- Develop, deploy, and monitor machine learning models and AI-driven solutions (predictive analytics, NLP for chatbots, anomaly detection) using Vertex AI and other AI APIs.
- Architect and manage data structures and tables within GCP (BigQuery, Cloud SQL, Spanner), optimizing for performance, cost, and accessibility.
- Create and automate insightful dashboards and reports (Looker, Power BI, Qlik Sense) for CX performance tracking and ad-hoc analyses.
- Collaborate with CX Performance Managers, Analysts, Product Owners, and stakeholders to define data requirements and deliver solutions.
- Explore and analyze large datasets, including vehicle maintenance data, to uncover trends and insights informing CX strategy and product development.
- Lead and support data projects using AI APIs (Dialogflow, Natural Language API, Vision API) to build CX tools and functionalities.
- Ensure data governance, security, and compliance best practices for all data solutions, particularly customer data and PII.
- Provide technical guidance and mentorship on data engineering best practices, GCP services, and AI/ML techniques.
- Stay current with emerging technologies in data engineering, data science, AI/ML, and GCP ecosystem and advocate adoption where beneficial.
- Document data architectures, data flows, model specifications, and processes to ensure clarity and maintainability.
Requirements
- Degree in Computer Science, Data Science, Engineering, Statistics, Mathematics, or comparable quantitative field (Bachelor's degree required; Master's or PhD desirable).
- Proven hands-on experience as a Data Engineer or Data Scientist, with projects in data pipeline development, data modeling, and cloud deployments.
- Expert proficiency in SQL and Python, including Pandas, NumPy, Scikit-learn, TensorFlow, or PyTorch.
- In-depth knowledge and practical experience with GCP data services: BigQuery, Dataflow, Dataproc, Pub/Sub, Cloud Storage, Cloud Functions, Vertex AI, Looker.
- Strong understanding of data warehousing principles, database design, and data modeling techniques (relational and non-relational).
- Experience developing and deploying machine learning models into production; familiarity with MLOps.
- Proficiency building and automating dashboards and reports using Looker, Power BI, Tableau, or Qlik Sense.
- Experience working with APIs for data ingestion and AI services.
- Familiarity with connected vehicle data, dealer invoice data, automotive industry data, or CX metrics (NPS, CSAT) is a plus.
- Excellent analytical and problem-solving skills; ability to translate complex business problems into technical solutions.
- Strong communication and interpersonal skills for collaboration with technical and non-technical stakeholders.
- Self-motivated, proactive; able to work independently and in an agile team.
- Knowledge of data governance and data security best practices in cloud contexts.
- Background checks are required pending successful interview.