GSK

Senior Scientific Knowledge Engineer

GSK

full-time

Posted on:

Location Type: Hybrid

Location: CambridgeCaliforniaMarylandUnited States

Visit company website

Explore more

AI Apply
Apply

Salary

💰 $145,200 - $242,000 per year

Job Level

About the role

  • Definition of schemas and data models of scientific information required for the creation of value adding data products.
  • This includes accountability for the quality control and mapping specifications to be industrialized by data engineering and maintained in platform provisioned tooling.
  • Accountable for the quality control (through validation and verification) of mapping specifications to be industrialized by data engineering and maintained in platform provisioned tooling – e.g., models, schemas, controlled vocab.
  • Working with Product managers/engineers confidently convert business need into defined deliverable business requirements to enable the integration of large-scale biology data to predict, model, and stabilize therapeutically relevant protein complex and antigen conformations for drug and vaccine discovery.
  • Collaborate with external groups to align GSK data standards with industry/ academic ontologies ensuring that data standards are defined with usage/analytics in mind.
  • Support effective ingestion of data by GSK through understanding the entry requirements required by platform engineering teams and ensuring that the “barrier for entry” is met e.g. Scientific information has the appropriate metadata to be indexed, structured, integrated and standardised as needed.
  • Provides bespoke subject matter expertise for R&D data to translate deep science into data for actionable insights
  • Champion data lineage, data quality, and FAIR data principles across the Onyx platform, working with engineering and product teams to embed governance and quality frameworks into data pipelines
  • Contribute to and maintain documentation of data standards, ontology decisions, and mapping rationale to support organizational knowledge transfer and auditability
  • Support self-service data enablement by ensuring metadata and knowledge products are accessible, well-documented, and usable by scientists and analysts without requiring bespoke engineering support

Requirements

  • Masters degree in Bioinformatics, Biomedical Science, Biomedical Engineering, Molecular Biology, or Computer Science (with a life science application focus)
  • 6+ years of relevant work experience
  • Experience contributing to Knowledge Graph development efforts, including entity modeling, relationship design, and schema governance
  • Experience in operating and leading across organizational boundaries a matrixed team
  • Experience with industry standard data management / metadata platforms e.g. Collibra, Datahub, Datum, Informatica
  • Proficiency in at least one programming language — preferably Python — for scripting vocabulary mappings, building data models, automating QC, and prototyping pipelines
  • Experience with bioinformatics pipelines and workflow management systems (e.g., Nextflow)
Benefits
  • health care and other insurance benefits (for employee and family)
  • retirement benefits
  • paid holidays
  • vacation
  • paid caregiver/parental and medical leave
Applicant Tracking System Keywords

Tip: use these terms in your resume and cover letter to boost ATS matches.

Hard Skills & Tools
data modelingquality controlmapping specificationsbioinformatics pipelinesprogramming languagePythonworkflow management systemsentity modelingschema governancedata ingestion
Soft Skills
collaborationcommunicationleadershipaccountabilitysubject matter expertiseorganizational skillsknowledge transferproblem-solvingadaptabilitystakeholder engagement
Certifications
Masters degree in BioinformaticsMasters degree in Biomedical ScienceMasters degree in Biomedical EngineeringMasters degree in Molecular BiologyMasters degree in Computer Science