RELX

Architect

RELX

full-time

Posted on:

Location Type: Hybrid

Location: Cape Town • 🇿🇦 South Africa

Visit company website
AI Apply
Apply

Job Level

Mid-LevelSenior

Tech Stack

AWSAzureCloudElasticSearchETLGoogle Cloud PlatformPySparkScalaSparkSQLTableauUnity

About the role

  • Designing and implementing cloud-native data architectures using Databricks and technologies such as Delta Lake, Spark, and MLflow.
  • Developing and maintaining robust data pipelines, including batch and streaming workloads, to support data ingestion, processing, and consumption.
  • Collaborating with business stakeholders and analytics teams to define data requirements, data models, and data integration strategies.
  • Ensuring data architecture solutions are secure, scalable, and high performing, adhering to enterprise standards and best practices.
  • Leading technical efforts in data quality, metadata management, data cataloguing, and governance (including Unity Catalogue if applicable).
  • Providing technical guidance to junior engineers and analysts in the adoption of modern data architecture patterns.
  • Evaluating and recommending emerging tools and frameworks within the Databricks ecosystem and broader data engineering space.
  • Having a solid understanding of analytics engines and columnar databases to support performance-optimised data solutions.
  • Experience with full-text search platforms is highly desirable; familiarity with technologies like Elasticsearch or Solr is a strong advantage.

Requirements

  • Bachelor’s or master’s degree in computer science, Information Systems, Engineering, or a related field.
  • Hands-on experience in data architecture, data engineering, or a similar role.
  • Deep expertise in Databricks, including Spark (PySpark/Scala), Delta Lake, and orchestration within Databricks workflows.
  • Strong understanding of cloud infrastructure and data services on at least one major cloud platform (Azure preferred, but AWS or GCP also accepted).
  • Proficiency in data modelling, SQL, data warehousing, and ETL frameworks.
  • Hands-on experience with CI/CD pipelines, version control (Git), and DevOps practices.
  • Solid understanding of data governance, privacy, and security best practices.
  • Databricks certifications (e.g., Data Engineer Associate/Professional) are nice to have.
  • Familiarity with Unity Catalogue, MLflow, and Lakehouse architecture principles is a plus.
  • Experience working in regulated industries (e.g., financial services, healthcare) is preferred.
  • Exposure to BI tools (e.g., Power BI, Tableau) and data virtualisation platforms is a plus.
Benefits
  • Working flexible hours - flexing the times when you work in the day to help you fit everything in and work when you are the most productive.
  • numerous wellbeing initiatives
  • shared parental leave
  • study assistance
  • sabbaticals

Applicant Tracking System Keywords

Tip: use these terms in your resume and cover letter to boost ATS matches.

Hard skills
DatabricksDelta LakeSparkMLflowdata architecturedata engineeringdata modellingSQLETL frameworksCI/CD
Soft skills
collaborationtechnical guidanceleadership
Certifications
Data Engineer AssociateData Engineer Professional