Build and optimize data pipelines, managing data storage and processing systems
Collaborate with cross-functional teams, including data scientists, software engineers, and business stakeholders to understand data requirements and deliver efficient high-quality data solutions
Design modern schemas, data models, and pipelines that enable advanced analytics, machine learning, and real-time reporting
Implement and manage real-time data streaming architectures using AWS services
Ensure solutions facilitate secure data analysis and reporting
Perform data profiling, cleansing, and transformation tasks
Implement data security and privacy measures
Requirements
Strong programming skills in languages like Python, Java, or Scala
Proven experience as a data engineer, with experience in designing and building large-scale data processing systems
Strong understanding of data modeling concepts and data management principles
In-depth knowledge of SQL and experience working with relational and non-relational databases
Knowledge of Data Mesh principles and experience designing and implementing data architectures following Data Mesh concepts within the AWS ecosystem
Experience with real-time data streaming architectures using AWS services like Amazon Kinesis or Apache Kafka
Familiarity with AWS cloud services, such as AWS Sagemaker Unified Studio, AWS Glue, AWS Lambda, AWS EMR, AWS S3, Amazon Redshift, and their data-related features and functionalities
Familiarity with AWS security services and features for data security and privacy
Bachelor's or Master's degree in Computer Science, Information Systems, or a related field
Benefits
Flexible work arrangements
Supportive policies for work-life balance
Professional development opportunities
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.