Develop big data applications for Synchrony in Hadoop ecosystem
Participate in the agile development process including backlog grooming, coding, code reviews, testing and deployment
Work with team members to achieve business results in a fast paced and quickly changing environment
Pair with experienced data engineers to develop analytic applications leveraging technologies such as: Hadoop, NoSQL, In-memory Data Grids, Kafka, Spark, Ab Initio
Provide data analysis for Synchrony’s data ingestion, standardization and curation efforts ensuring all data is understood from a business context
Identify enablers and level of effort required to properly ingest and transform data for the data lake.
Profile data to assist with defining the data elements, propose business term mappings, and define data quality rules
Work with the Data Office to ensure that data dictionaries for all ingested and created data sets are properly documented in data dictionary repository
Ensure the lineage of all data assets are properly documented in the appropriate enterprise metadata repositories
Assist with the creation and implementation of data quality rules
Ensure the proper identification of sensitive data elements and critical data elements
Requirements
Bachelor's degree in a quantitative field (such as Engineering, Computer Science, Statistics, Econometrics);
In lieu of degree, High School Diploma/GED and minimum 2 years of Information Technology experience.