Dynatron Software, Inc.

Gen AI Data Engineer II

Dynatron Software, Inc.

full-time

Posted on:

Location Type: Remote

Location: Remote • 🇺🇸 United States

Visit company website
AI Apply
Apply

Salary

💰 $110,000 - $135,000 per year

Job Level

Mid-LevelSenior

Tech Stack

AirflowAmazon RedshiftAWSCloudPythonSQL

About the role

  • Engineer Generative AI Data Systems
  • Design and maintain data pipelines for training, fine-tuning, and retrieval-augmented generation (RAG) use cases.
  • Build ingestion frameworks using AWS Glue, Lambda, Kinesis, and Step Functions to support large-scale AI workloads.
  • Develop embedding pipelines, feature stores, and vector database integrations (Pinecone, FAISS, Chroma, Amazon OpenSearch) to power semantic retrieval.
  • Transform unstructured data–documents, text, images, logs–into AI-ready assets for LLM applications.
  • Integrate & Orchestrate LLM Architectures
  • Build end-to-end GenAI pipelines connecting enterprise data with LLMs including Anthropic Claude, Amazon Titan, OpenAI GPT, and Llama 3.
  • Use LangChain, LlamaIndex, and Bedrock Agents to deliver context-rich RAG, prompt-chaining, and conversational intelligence.
  • Develop LLM-powered APIs enabling natural language querying, summarization, search, and generative workflows.
  • Optimize prompts, context windows, model evaluation, and response quality.
  • Scale AI Infrastructure & MLOps
  • Deploy, monitor, and optimize LLM workflows on AWS Bedrock and other cloud AI platforms.
  • Implement CI/CD pipelines for GenAI systems using Airflow, Prefect, GitHub Actions, or AWS CodePipeline.
  • Establish data and model observability frameworks to track drift, accuracy, latency, and performance.
  • Partner with Data Science and MLOps teams to streamline fine-tuning, deployment, and scalable model operations.
  • Champion Governance, Security & Responsible AI
  • Implement data lineage, access controls, encryption, and governance for AI datasets.
  • Enforce Responsible AI practices, ensuring transparency, risk mitigation, and ethical use of LLMs.
  • Maintain prompt logs, telemetry, and audit documentation supporting SOC2, GDPR, and CCPA compliance.

Requirements

  • Bachelor’s or Master’s degree in Computer Science, Data Engineering, or related field.
  • 5+ years of data engineering experience, including 2+ years developing GenAI or LLM-based solutions.
  • Strong proficiency in: AWS Bedrock, SageMaker, or Vertex AI
  • LangChain or LlamaIndex
  • Snowflake, Redshift, or Databricks
  • Python, SQL, and API integrations
  • Vector databases (Pinecone, FAISS, Chroma, OpenSearch)
  • Proven experience building RAG pipelines, embeddings, and prompt-chaining architectures.
  • Deep understanding of data modeling, orchestration, and MLOps best practices.
  • Ability to integrate LLM capabilities into enterprise SaaS products and data platforms.
Benefits
  • Comprehensive health, vision, and dental insurance
  • Employer-paid short- and long-term disability and life insurance
  • 401(K) with competitive company match
  • Flexible vacation policy and 9 paid holidays
  • Remote-first culture

Applicant Tracking System Keywords

Tip: use these terms in your resume and cover letter to boost ATS matches.

Hard skills
data engineeringGenerative AILLM-based solutionsAWS GlueAWS LambdaAWS KinesisAWS Step FunctionsPythonSQLAPI integrations
Soft skills
collaborationcommunicationproblem-solvingleadershiporganizational skills
Certifications
Bachelor’s degree in Computer ScienceMaster’s degree in Computer Science