Collaborate in defining the overall architecture of the solution, with experience in designing, building, and maintaining reusable data products using Data Bricks, Delta Live Tables (DLT), Pyspark, and SQL. Migration of existing data pipelines to modern frameworks and ensure scalability and efficiency.
Develop the data infrastructure, pipeline architecture, and integration solutions while actively contributing to hands-on implementation.
Build and maintain scalable, efficient data processing pipelines and solutions for data-driven applications.
Monitor and ensure adherence to data security, privacy regulations, and compliance standards.
Troubleshoot and resolve complex data-related challenges and incidents in a timely manner.
Stay at the forefront of emerging trends and technologies in data engineering and advocate for their integration when relevant.
Requirements
Proven expertise in Data Bricks, Delta Live Tables, SQL, and Pyspark for processing and managing large data volumes.
Strong experience in designing and implementing dimensional models and medallion architecture.
Strong experience in designing and migrating existing databricks workspaces and models to Unity Catalog enabled workspaces.
Strong Experience creatinging and managing group Access Control Lists (ACL) and compute and governance policies in Databricks Unity Catalog.
Hands-on experience with modern data pipeline tools (e.g. AWS Glue, Azure Data Factory) and cloud platforms (e.g. Databricks).
Knowledge of cloud data lakes (e.g., Data Bricks Delta Lake, Azure Storage and/or AWS S3).
Demonstrated experience applying DevOps principles using Version Control and CICD for IaC and code base deployments(e.g. AzureDevops, Git, CI/CD) to data engineering projects.
Strong experience with batch and streaming data processing techniques and file compactization strategies.
Familiarity with architectural best practices for building data lakes.
Hands on with additional Azure Services including, Message Queues, Service Bus, Cloud Storage, Virtual Cloud, Serverless ~ Compute, CloudSQL, OOP Languages and Frameworks.
Experience with BI tools (e.g., Power BI, Tableau) and deploying data models.
Experience with Databricks Unity Catalog, i.e. configuring and managing data governance and access controls in a Delta Lake environment.
ATS Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard skills
Data BricksDelta Live TablesPysparkSQLAWS GlueAzure Data FactoryDevOpsCICDbatch processingstreaming data processing
Soft skills
collaborationtroubleshootingproblem-solvingadvocacy for emerging technologies