Tech Stack
AirflowAWSAzureCloudETLGoogle Cloud PlatformJavaScriptPythonSQL
About the role
- Clean, normalize, and QA incoming datasets to ensure integrity, usability, and consistency across systems.
- Design, maintain, and optimize SQL queries, views, and data pipelines for analytics and model inputs.
- Contribute to the evolution of our data model to better support reporting, analytics, and AI applications.
- Partner with the PM and engineering team to identify high-value opportunities for AI/ML integration.
- Build business intelligence reports and insights leveraging Sigma.
- Build and test lightweight models, agents, or prototypes that demonstrate potential AI applications.
- Develop pipelines that make data more “AI-ready” (structured, labeled, and enriched).
- Create clear, reliable documentation for data sources, transformations, and model-ready datasets.
- Proactively identify gaps in data quality, structure, or process, and drive improvements.
Requirements
- 2–3 years of professional experience in data analytics, data engineering, or a similar role.
- Strong proficiency in SQL (joins, CTEs, window functions, optimization) in Snowflake.
- Experience with ETL/ELT pipelines and data transformation tools (e.g., dbt, Airflow, Fivetran) and Snowflake orchestration (tasks, streams, dynamic tables, Snowpipe).
- Familiarity with data modeling concepts (star schema, normalization, dimensional modeling).
- Exposure to Python (or R) for scripting, data manipulation, REST/JSON API ingestion and light ML/AI tasks.
- Understanding of cloud data environments (AWS, GCP, or Azure) and cloud object storage (S3).
- Growth mindset, and an eagerness to learn every day.
- Excellent problem-solving skills, with a focus on data integrity and AI-readiness.
- Strong communicator, comfortable collaborating with product and engineering teams.