
Gen AI Data Engineer II
Dynatron Software, Inc.
full-time
Posted on:
Location Type: Remote
Location: Remote • 🇺🇸 United States
Visit company websiteSalary
💰 $110,000 - $135,000 per year
Job Level
Mid-LevelSenior
Tech Stack
AirflowAmazon RedshiftAWSCloudPythonSQL
About the role
- Engineer Generative AI Data Systems
- Design and maintain data pipelines for training, fine-tuning, and retrieval-augmented generation (RAG) use cases.
- Build ingestion frameworks using AWS Glue, Lambda, Kinesis, and Step Functions to support large-scale AI workloads.
- Develop embedding pipelines, feature stores, and vector database integrations (Pinecone, FAISS, Chroma, Amazon OpenSearch) to power semantic retrieval.
- Transform unstructured data–documents, text, images, logs–into AI-ready assets for LLM applications.
- Integrate & Orchestrate LLM Architectures
- Build end-to-end GenAI pipelines connecting enterprise data with LLMs including Anthropic Claude, Amazon Titan, OpenAI GPT, and Llama 3.
- Use LangChain, LlamaIndex, and Bedrock Agents to deliver context-rich RAG, prompt-chaining, and conversational intelligence.
- Develop LLM-powered APIs enabling natural language querying, summarization, search, and generative workflows.
- Optimize prompts, context windows, model evaluation, and response quality.
- Scale AI Infrastructure & MLOps
- Deploy, monitor, and optimize LLM workflows on AWS Bedrock and other cloud AI platforms.
- Implement CI/CD pipelines for GenAI systems using Airflow, Prefect, GitHub Actions, or AWS CodePipeline.
- Establish data and model observability frameworks to track drift, accuracy, latency, and performance.
- Partner with Data Science and MLOps teams to streamline fine-tuning, deployment, and scalable model operations.
- Champion Governance, Security & Responsible AI
- Implement data lineage, access controls, encryption, and governance for AI datasets.
- Enforce Responsible AI practices, ensuring transparency, risk mitigation, and ethical use of LLMs.
- Maintain prompt logs, telemetry, and audit documentation supporting SOC2, GDPR, and CCPA compliance.
Requirements
- Bachelor’s or Master’s degree in Computer Science, Data Engineering, or related field.
- 5+ years of data engineering experience, including 2+ years developing GenAI or LLM-based solutions.
- Strong proficiency in: AWS Bedrock, SageMaker, or Vertex AI
- LangChain or LlamaIndex
- Snowflake, Redshift, or Databricks
- Python, SQL, and API integrations
- Vector databases (Pinecone, FAISS, Chroma, OpenSearch)
- Proven experience building RAG pipelines, embeddings, and prompt-chaining architectures.
- Deep understanding of data modeling, orchestration, and MLOps best practices.
- Ability to integrate LLM capabilities into enterprise SaaS products and data platforms.
Benefits
- Comprehensive health, vision, and dental insurance
- Employer-paid short- and long-term disability and life insurance
- 401(K) with competitive company match
- Flexible vacation policy and 9 paid holidays
- Remote-first culture
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard skills
data engineeringGenerative AILLM-based solutionsAWS GlueAWS LambdaAWS KinesisAWS Step FunctionsPythonSQLAPI integrations
Soft skills
collaborationcommunicationproblem-solvingleadershiporganizational skills
Certifications
Bachelor’s degree in Computer ScienceMaster’s degree in Computer Science