Tech Stack
AirflowAmazon RedshiftKafkaPythonSQL
About the role
- Design, develop, and optimize the ingestion of large volumes of structured and unstructured data from diverse sources.
- Support data architecture transformation initiatives on Databricks, ensuring scalable and efficient systems.
- Guide the design and implementation of platform components for the training, deployment, and monitoring of ML models in production environments.
- Provide expertise on industry best practices, tools, and technologies in ML engineering.
- Drive the continuous improvement of data and ML workflows through automation and innovative solutions.
- Take full ownership of projects, ensuring successful delivery from planning to execution.
- Collaborate with cross-functional teams to gather business requirements and translate them into effective technical solutions.
Requirements
- 5+ years of software development experience, specializing in data and ML
- Expertise in SQL, Python, and Databricks
- Experience with Airflow, Kafka, and Redshift
- Proficiency in data modeling and database design.
- Proficient in building and deploying machine learning models (including language models)
- Proficient in building model serving pipelines for batch, streaming, and real-time inference
- Driving the operationalization of ML models, ensuring seamless deployment, monitoring, and continuous integration/continuous delivery (CI/CD) pipelines for ML systems.
- Strong problem-solving skills with a strategic mindset and attention to detail
- Effective communicator and collaborator across technical and non-technical stakeholders