Design, build, and operate highly reliable Node.js/TypeScript services on AWS to enable generative‑AI capabilities across products and internal workflows
Create scalable REST/GraphQL APIs to serve generative‑AI features such as chat, summarization, and content generation
Build and maintain AWS‑native architectures using Lambda, API Gateway, ECS/Fargate, DynamoDB, S3, and Step Functions
Integrate and orchestrate LLM services (Amazon Bedrock, OpenAI, HuggingFace, self‑hosted models) and vector databases (Aurora pgvector, Pinecone, Chroma) to power RAG pipelines
Create secure, observable, and cost‑efficient infrastructure as code (CDK/Terraform) and automate CI/CD with GitHub Actions or AWS CodePipeline
Implement monitoring, tracing, and logging (CloudWatch, X‑Ray, OpenTelemetry) to track latency, cost, and output quality of AI endpoints
Collaborate with ML engineers, product managers, and front‑end teams in agile sprints; participate in design reviews and knowledge‑sharing sessions
Establish best practices for prompt engineering, model evaluation, and data governance to ensure responsible AI usage
Participate in on‑call/operational activities to ensure production reliability and cost optimization of AI workloads
Requirements
Available working some US hours
Proficient in Hebrew and English both written and verbal
4+ years professional experience building production services with Node.js/TypeScript
3+ years hands‑on with AWS, including Lambda, API Gateway, DynamoDB, and at least one container service (ECS, EKS, or Fargate)
Experience building Retrieval‑Augmented Generation (RAG) systems or knowledge‑base chatbots
Hands‑on with vector databases such as Pinecone, Chroma, or pgvector on Postgres/Aurora