
AI Engineer, Data Pipeline
Coupa Software
full-time
Posted on:
Location Type: Remote
Location: India
Visit company websiteExplore more
About the role
- Build data ingestion pipelines to extract and transform enterprise data.
- Implement data cleansing and normalization routines.
- Write and maintain ETL jobs using Spark/PySpark on cloud infrastructure.
- Implement data validation and quality checks at each pipeline stage.
- Build automated data export jobs for model training datasets.
- Support feature extraction from enterprise schemas.
- Monitor pipeline health, troubleshoot failures, and optimize performance.
- Document data lineage, schemas, and transformation logic.
Requirements
- 3+ years of software engineering experience.
- Experience with Python and data processing (pandas, PySpark, or equivalent).
- Familiarity with SQL and relational databases (MySQL, PostgreSQL).
- Experience with cloud data services (object storage, managed Spark, managed ETL, or equivalent).
- Understanding of ETL/ELT patterns and data pipeline design.
- Experience with data formats (Parquet, JSON, Avro).
- Strong attention to data quality and testing.
- BS in Computer Science or equivalent experience.
Benefits
- Pioneering Technology: At Coupa, we're at the forefront of innovation, leveraging the latest technology to empower our customers with greater efficiency and visibility in their spend.
- Collaborative Culture: We value collaboration and teamwork, and our culture is driven by transparency, openness, and a shared commitment to excellence.
- Global Impact: Join a company where your work has a global, measurable impact on our clients, the business, and each other.
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard Skills & Tools
ETLdata ingestiondata cleansingdata normalizationSparkPySparkPythonSQLdata processingdata formats
Soft Skills
attention to detailtroubleshootingoptimizationdocumentation
Certifications
BS in Computer Science