ThetaRay

Senior Data Engineer

ThetaRay

full-time

Posted on:

Location Type: Office

Location: Hod HaSharon • 🇮🇱 Israel

Visit company website
AI Apply
Apply

Job Level

Senior

Tech Stack

AirflowApacheAzureDockerHadoopHDFSJenkinsKubernetesLinuxMicroservicesPySparkScalaSparkSQL

About the role

  • Implement and maintain data pipeline flows in production within the ThetaRay system
  • Design and implement solution-based data flows for specific use cases, enabling the applicability of implementations within the ThetaRay product
  • Building a Machine Learning data pipeline
  • Create data tools for analytics and data scientist team members that assist them in building and optimizing our product
  • Work with product, R&D, data, and analytics experts to strive for greater functionality in our systems
  • Train customer data scientists and engineers to maintain and amend data pipelines within the product
  • Travel to customer locations both domestically and abroad
  • Build and manage technical relationships with customers and partners

Requirements

  • 4+ years of experience as a data engineer - must
  • Hands-on experience working with Apache Spark with Pyspark or Scala - must
  • Hands-on experience with SQL - must
  • Hands-on experience with version-control tools such as GIT - must
  • Hands-on experience with data transformation, validations, cleansing, and ML feature engineering - must
  • BSc degree or higher in Computer Science, Statistics, Informatics, Information Systems, Engineering, or another quantitative field - must
  • Strong analytic skills related to working with structured and semi-structured datasets - must
  • Business-oriented and able to work with external customers and cross-functional teams - must
  • Fluent in English - must
  • Experience with Linux - nice to have
  • Experience in building Machine Learning pipeline - nice to have
  • Experience with Jupyter - nice to have
  • Experience with workflow automation platforms such as Jenkins or Apache Airflow - nice to have
  • Experience with Microservices architecture components, including Docker and Kubernetes - nice to have
  • Experience in GitHub Copilot - nice to have
  • Hands-on experience with Apache Hadoop Ecosystem including Hive, Impala, Hue, HDFS, Sqoop etc. - nice to have
  • Experience working with and optimizing big data pipelines, architectures, and data sets - nice to have
  • Familiarity with Azure SaaS solutions - nice to have

Applicant Tracking System Keywords

Tip: use these terms in your resume and cover letter to boost ATS matches.

Hard skills
data engineeringApache SparkPysparkScalaSQLdata transformationdata validationdata cleansingML feature engineeringbig data pipelines
Soft skills
analytical skillsbusiness-orientedcustomer relationship managementcross-functional teamworkcommunication skills
Certifications
BSc degree in Computer ScienceBSc degree in StatisticsBSc degree in InformaticsBSc degree in Information SystemsBSc degree in Engineering