RBC

Data Engineer

RBC

full-time

Posted on:

Location Type: Hybrid

Location: Newcastle upon TyneUnited Kingdom

Visit company website

Explore more

AI Apply
Apply

Tech Stack

About the role

  • Responsible for the development and ongoing maintenance of the Data Lakehouse platform infrastructure using the Microsoft Azure technology stack, including Databricks and Data Factory.
  • Manage data pipelines consisting of a series of stages through which data flows (for example, from data sources or endpoints of acquisition to integration to consumption for specific use cases).
  • Architecting, creating and maintaining data pipelines will be the primary responsibility of the data engineer.
  • Create new and modify existing Notebooks, Functions and Workflows to support efficient reporting and analytics to the business.
  • Create, maintain, and develop Dev, UAT and Production environments ensuring consistency.
  • Champion for the DevOps process to ensure the latest techniques are being used and that implementation methodologies involving new or changes to existing source code or data structures follow the agreed development and release processes and that all productionised code is adequately documented, reviewed and unit tested where appropriate.
  • Identify, source, stage, and model internal process improvements to automate manual processes and optimise data delivery for greater scalability, as part of the end-to-end data lifecycle.

Requirements

  • Proven experience working within Data Engineering and Data Management architectures like Data Warehouse, Data Lake, Data Hub and the supporting processes like Data Integration, Governance, Metadata Management.
  • Proven experience working in cross-functional teams and collaborating with business stakeholders in support of a departmental and/or multi-departmental data management and analytics initiative.
  • Strong experience with popular database programming languages for relational databases (SQL, T-SQL).
  • Experience working on a cloud data platform such as Databricks or Snowflake.
  • Adept in agile methodologies, and capable of applying DevOps and DataOps principles to data pipelines.
  • Basic experience in working with data governance, data quality and data security teams.
  • Good understanding of datasets, Data Lakehouses, modelling, database design and programming.
  • Knowledge of Data Lakehousing techniques, solutions and methodologies.
  • Strong experience supporting and working with cross-functional teams in a dynamic business environment.
  • Required to be highly creative and collaborative working closely with business teams and IT teams to define the business problem, refine the requirements, and design and develop data deliverables accordingly.
Benefits
  • Opportunities to work with the best in the field
  • Ability to make a difference and lasting impact
  • Work in a dynamic, collaborative, progressive, and high-performing team
  • Leaders who support your development through coaching and managing opportunities
Applicant Tracking System Keywords

Tip: use these terms in your resume and cover letter to boost ATS matches.

Hard Skills & Tools
Data LakehouseDatabricksData FactorySQLT-SQLData IntegrationData GovernanceData QualityData SecurityDatabase Design
Soft Skills
collaborationcreativitycommunicationproblem-solvingcross-functional teamworkadaptabilityleadershiporganizational skillsanalytical thinkingstakeholder engagement