
Senior Data Engineer – Databricks
MediaRadar, Inc.
full-time
Posted on:
Location Type: Remote
Location: Remote • 🇮🇳 India
Visit company websiteJob Level
Senior
Tech Stack
AirflowApacheAzureCloudETLKubernetesPostgresPySparkPythonSparkSQL
About the role
- Involve in Design, development, and maintenance of scalable ETL/ELT pipelines on Azure Databricks using Apache Spark (PySpark/Spark SQL).
- Design and implement both batch and real-time data ingestion and transformation processes.
- Build and manage Delta Lake tables, schemas, and data models to support efficient querying and analytics.
- Consolidate and process large-scale datasets from various structured and semi-structured sources (e.g., JSON, Parquet, Avro).
- Write optimized SQL queries for large datasets using Spark SQL and PostgreSQL.
- Develop, schedule, and monitor workflows using Databricks Workflows, Airflow or similar orchestration tools.
- Design, build, and deploy cloud-native, containerized applications on Azure Kubernetes Service (AKS) and integrate with Azure services.
- Ensure data quality, governance, and compliance through validation, documentation, and secure practices.
- Collaborate with data analysts, data architects, and business stakeholders to translate requirements into technical solutions.
- Contribute to and enforce best practices in data engineering, including version control (Git), CI/CD pipelines, and coding standards.
- Continuously enhance data systems for improved performance, reliability, and scalability.
- Mentor junior engineers and help evolve team practices and documentation.
- Stay up to date on emerging trends, technologies, and best practices in the data engineering space.
- Work effectively within an agile, cross-functional project team.
Requirements
- Proven experience as a Data Engineer, with a strong focus on Azure Databricks and Apache Spark.
- Proficiency in Python, PySpark, Spark SQL, and working with large-scale datasets in different data formats.
- Strong experience designing and building ETL/ELT workflows in both batch and streaming environments.
- Solid understanding of data lakehouse architectures and Delta Lake.
- Experience in Azure Kubernetes Service (AKS) is desired.
- Proficient in SQL and experience with PostgreSQL or similar relational databases.
- Experience with workflow orchestration tools (e.g., Databricks Workflows, Airflow, Azure Data Factory).
- Familiarity with data governance, quality control, and security best practices.
- Strong problem-solving skills and attention to detail.
- Excellent communication and collaboration skills, with a track record of working cross-functionally.
- Experience mentoring junior engineers and leading by example.
- Comfortable working in agile development environments and using tools like Git, CI/CD, and issue trackers (e.g., Jira).
Benefits
- At MediaRadar, we are committed to creating an inclusive and accessible workplace where everyone can thrive.
- We believe that diversity of backgrounds, perspectives, and experiences makes us stronger and more innovative.
- We are proud to be an Equal Opportunity Employer and make employment decisions without regard to race, color, religion, sex (including pregnancy, sexual orientation, or gender identity), national origin, age, disability, genetic information, or any other legally protected status.
- This is a full-time exempt role with base salary plus benefits.
- Final compensation will depend on location, skill level, and experience.
Applicant Tracking System Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard skills
ETLELTAzure DatabricksApache SparkPySparkSpark SQLSQLPostgreSQLDelta Lakedata lakehouse architecture
Soft skills
problem-solvingattention to detailcommunicationcollaborationmentoringleadershipagile developmentcross-functional teamwork