skillventory - A Leading Talent Research Firm

Principal AI/ML Engineer

skillventory - A Leading Talent Research Firm

full-time

Posted on:

Location Type: Hybrid

Location: Durham • New Jersey, North Carolina, Texas • 🇺🇸 United States

Visit company website
AI Apply
Apply

Salary

💰 $107,000 - $216,000 per year

Job Level

Lead

Tech Stack

ApacheAWSAzureCloudDistributed SystemsDockerDynamoDBETLJavaJenkinsKafkaNoSQLNumpyPythonScalaScikit-LearnSparkSQLTensorflowTerraformUnix

About the role

  • create frameworks to support large-scale ML infrastructure and pipelines, including tools for the containerization and deployment of ML models
  • Collaborating with Data Scientists, you will develop advanced analytics and machine learning platforms to enable the prediction and optimization of models
  • extend existing ML platforms for scaling model training and deployment
  • partner with various business and engineering teams to drive the adoption and integration of model outputs
  • building advanced cloud and software solutions in collaboration with Data Scientists to support packaging, deployment, and scaling of AI/ML Models in production

Requirements

  • Has bachelor’s or master’s Degree in a technology related field (e.g. Engineering, Computer Science, etc.)
  • 8+ years of proven experience in developing and implementing Python-based cloud applications and/or machine learning solutions
  • 2+ years of experience in developing ML infrastructure and MLOps in the Cloud using AWS Sagemaker
  • 5+ years of experience in building cloud-native applications using a range of AWS services, including but not limited to SageMaker AI, Bedrock, S3, CloudFormation (CFT), SNS, SQS, Lambda, AWS Batch, Step Functions, EventBridge, and CloudWatch
  • Familiarity with both Azure Cognitive Services, particularly for deploying OpenAI models, and Google Compute Vertex is beneficial
  • Extensive experience working with machine learning models with respect to deployment, inference, tuning, and measurement required
  • Experience in Object Oriented Programming (Java, Scala, Python), SQL, Unix scripting or related programming languages and exposure to some of Python’s ML ecosystem (numpy, panda, sklearn, tensorflow, etc.)
  • Experience with building data pipelines in getting the data required to build and evaluate ML models, using tools like Apache Spark or other distributed data processing frameworks
  • Data movement technologies (ETL/ELT), Messaging/Streaming Technologies (AWS SQS, Kinesis/Kafka), Relational and NoSQL databases (DynamoDB, EKS, Graph database), API and in-memory technologies
  • Strong knowledge of developing highly scalable distributed systems using Open-source technologies
  • Strong experience with CI/CD tools, particularly Jenkins, for automating and streamlining the software development pipeline
  • Proficient in using version control systems like Git for effective code management and collaboration
  • Hands-on experience with containerization technologies such as Docker for building and deploying applications
  • Expertise in infrastructure as code (IaC) services, including AWS CloudFormation and tools like Terraform or OpenTofu, for managing and provisioning cloud resources
  • Solid experience in Agile methodologies (Kanban and SCRUM)
Benefits
  • comprehensive health care coverage and emotional well-being support
  • market-leading retirement
  • generous paid time off and parental leave
  • charitable giving employee match program
  • educational assistance including student loan repayment, tuition reimbursement, and learning resources to develop your career

Applicant Tracking System Keywords

Tip: use these terms in your resume and cover letter to boost ATS matches.

Hard skills
PythonMLOpsAWS SageMakerAWS servicesApache SparkETLCI/CDDockerInfrastructure as CodeAgile methodologies
Soft skills
collaborationcommunicationproblem-solvingleadershiporganizational skills