Salary
💰 $95,000 - $115,000 per year
Tech Stack
AWSAzureCassandraCloudDistributed SystemsETLHadoopKafkaMongoDBNeo4jNoSQLSparkSQLTypeScript
About the role
- Develop, maintain, and test infrastructures for data generation to transform data from various structured and unstructured data sources.
- Develop complex queries to ensure accessibility while optimizing the performance of NoSQL and or big data infrastructure. Create and maintain optimal data pipeline architecture.
- Build and maintain the infrastructure to support extraction, transformation, and loading (ETL) of data from a wide variety of data sources. Extract data from multiple data sources, relational SQL and NoSQL databases, and other platform APIs, for data ingestion and integration.
- Configure and manage data analytic frameworks and pipelines using databases and tools such as NoSQL, SQL, HDInsight, MongoDB, Cassandra, Neo4j, GraphDB, OrientDB, Spark, Hadoop, Kafka, Hive, and Pig.
- Apply distributed systems concepts and principles such as consistency and availability, liveness and safety, durability, reliability, fault-tolerance, consensus algorithms.
- Administrate cloud computing and CI/CD pipelines to include Azure, Google, and Amazon Web Service (AWS).
- Coordinate with stakeholders, including product, data and design teams to assist with data-related technical issues and support their data infrastructure needs.
Requirements
- Active TS/SCI Clearance required
- Minimum of 1 year of experience required
- Bachelors degree in a STEM field with preference towards Computer Science and Software Engineering
- Verifiable work experience working with data structures, database management, distributed computing, and API driven architectures using SQL and No-SQL engines
- A Certified Data Management Professional certification is preferred
- Proficient in modeling frameworks like Universal Modeling Language (UML), Agile Development, and Git Operations
- Proficient in multiple programming languages and deep knowledge of SQL database design
- Basic knowledge in machine learning to build efficient and accurate data pipelines
- Experience with NoSQL and big data infrastructure (HDInsight, MongoDB, Cassandra, Neo4j, GraphDB, OrientDB)
- Experience with big data and distributed processing tools (Spark, Hadoop, Kafka, Hive, Pig)
- Experience administrating cloud computing and CI/CD pipelines including Azure, Google, and Amazon Web Service (AWS)
- Verifiable experience integrating APIs and building ETL pipelines
- The ACI Group is unable to sponsor H1B Visas