
Senior Data Engineer
ThetaRay
full-time
Posted on:
Location Type: Office
Location: Hod HaSharon • 🇮🇱 Israel
Visit company websiteJob Level
Senior
Tech Stack
AirflowApacheAzureDockerHadoopHDFSJenkinsKubernetesLinuxMicroservicesPySparkScalaSparkSQL
About the role
- Implement and maintain data pipeline flows in production within the ThetaRay system
- Design and implement solution-based data flows for specific use cases, enabling the applicability of implementations within the ThetaRay product
- Building a Machine Learning data pipeline
- Create data tools for analytics and data scientist team members that assist them in building and optimizing our product
- Work with product, R&D, data, and analytics experts to strive for greater functionality in our systems
- Train customer data scientists and engineers to maintain and amend data pipelines within the product
- Travel to customer locations both domestically and abroad
- Build and manage technical relationships with customers and partners
Requirements
- 4+ years of experience as a data engineer - must
- Hands-on experience working with Apache Spark with Pyspark or Scala - must
- Hands-on experience with SQL - must
- Hands-on experience with version-control tools such as GIT - must
- Hands-on experience with data transformation, validations, cleansing, and ML feature engineering - must
- BSc degree or higher in Computer Science, Statistics, Informatics, Information Systems, Engineering, or another quantitative field - must
- Strong analytic skills related to working with structured and semi-structured datasets - must
- Business-oriented and able to work with external customers and cross-functional teams - must
- Fluent in English - must
- Experience with Linux - nice to have
- Experience in building Machine Learning pipeline - nice to have
- Experience with Jupyter - nice to have
- Experience with workflow automation platforms such as Jenkins or Apache Airflow - nice to have
- Experience with Microservices architecture components, including Docker and Kubernetes - nice to have
- Experience in GitHub Copilot - nice to have
- Hands-on experience with Apache Hadoop Ecosystem including Hive, Impala, Hue, HDFS, Sqoop etc. - nice to have
- Experience working with and optimizing big data pipelines, architectures, and data sets - nice to have
- Familiarity with Azure SaaS solutions - nice to have
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard skills
data engineeringApache SparkPysparkScalaSQLdata transformationdata validationdata cleansingML feature engineeringbig data pipelines
Soft skills
analytical skillsbusiness-orientedcustomer relationship managementcross-functional teamworkcommunication skills
Certifications
BSc degree in Computer ScienceBSc degree in StatisticsBSc degree in InformaticsBSc degree in Information SystemsBSc degree in Engineering