Seagate Technology

Big Data Engineer, Intern

Seagate Technology

internship

Posted on:

Location Type: Office

Location: Wuxi • 🇨🇳 China

Visit company website
AI Apply
Apply

Job Level

Entry Level

Tech Stack

AirflowAWSCloudDockerETLHadoopJavaKafkaKubernetesPythonScalaSpark

About the role

  • Part of 10-12 Platform Engineers that are the crux for developing and maintaining Big Data (Data Lake, Data Warehouse and Data Integration) and advanced analytics platforms at SeaGate.
  • Apply your hands-on subject matter expertise in the Architecture of and administration of Big Data platforms.
  • Develop and manage SPARK ETL Frameworks, Data orchestration with Airflow and support building PRESTO/Trino queries for the key stakeholders.
  • Design, scale and deploy Machine Learning pipelines.
  • Collaborate with Application Architects and Business SMEs to design and develop end-to-end data pipelines and supporting infrastructure.
  • Establish and maintain productive relationships with peer organizations, partners, and software vendors.

Requirements

  • Excellent coding skills in any language with deep desire to learn new skills and technologies.
  • You’re a passionate professional who is up to the challenge of blending the fast-changing technology landscape of Big Data analytics with the complex and high-impact space of HiTech and Manufacturing analytics.
  • As a motivated self-starter, you have the experience working in a dynamic environment.
  • Exceptional data engineering skills in large, high-scale Data platforms and applications using cloud and big data technologies like Hadoop ecosystem and Spark.
  • Strong appetite for constant learning, thinking out of the box, questioning the problems & solutions with the intent to understand and solve better.
  • Excellent interpersonal skills to develop relationships with different teams and peers in the organization.
  • Big data processing frameworks knowledge: Spark, Hadoop, Hive, Kafka, EMR.
  • Big data solutions on cloud (AWS or Other).
  • Advanced experience and hands-on architecture and administration experience on big data platforms.
  • Data Warehouse Appliances, Hadoop (AWS EMR), Data Lake Technologies (AWS S3/GCS/Other) and experience with ML and Data Science platforms (Spark ML , H2O , KNIME).
  • Python, Java, Scala.
  • DevOps, Continuous Delivery, and Agile development.
  • Creating a culture of technical excellence by leading code and design reviews, promoting mentorship, and identifying and promoting educational opportunities for engineers.
  • Strong understanding of Micro-services and container-based development using Docker and Kubernetes ecosystem is a BIG plus.
  • Experience working in a Software Product Development environment is a BIG plus.
Benefits
  • On site you can grab breakfast, lunch, dinner and snacks at our canteen, grab-and-go market and coffee shop.
  • We offer basketball, badminton, yoga clubs and group exercise classes.
  • We also have music, dance, photography, and literature clubs galore, along with an active Toastmasters International chapter, and frequently have on-site festivals, celebrations and community volunteer opportunities.

Applicant Tracking System Keywords

Tip: use these terms in your resume and cover letter to boost ATS matches.

Hard skills
SPARKETLData orchestrationPRESTOMachine LearningHadoopHiveKafkaPythonJava
Soft skills
interpersonal skillsself-starterconstant learningproblem-solvingcollaborationmentorshipleadershipcommunicationadaptabilitycreativity