Tech Stack
ApacheAWSAzureCassandraCloudDistributed SystemsDockerGoogle Cloud PlatformGrafanaIoTJavaKafkaKubernetesMicroservicesMongoDBNoSQLPostgresPrometheusPulsarPythonRabbitMQScalaSparkSQL
About the role
- Design, develop, and maintain real-time data streaming pipelines using Kafka, Flink, and Spark Streaming.
- Architect event-driven and microservices-based solutions for real-time analytics and processing.
- Implement data ingestion, transformation, and enrichment workflows across distributed systems.
- Optimize performance and scalability of streaming jobs for high-throughput, low-latency environments.
- Ensure data quality, governance, and fault tolerance within the streaming infrastructure.
- Integrate streaming solutions with data warehouses, data lakes, and cloud platforms (AWS, Azure, GCP).
- Collaborate with data engineers, data scientists, and application teams to deliver business-critical real-time insights.
- Monitor, troubleshoot, and improve the reliability and resilience of streaming systems.
- Participate in system design, code reviews, and best practices development.
Requirements
- 5+ years of experience in data engineering with at least 3+ years focused on real-time streaming.
- Strong expertise in Apache Kafka (producers, consumers, Connect, Streams, schema registry).
- Hands-on experience with Apache Flink or Spark Streaming for real-time data processing.
- Solid understanding of event-driven architectures, pub/sub systems, and distributed computing.
- Strong programming skills in Java, Scala, or Python.
- Proficiency in SQL and experience with databases (relational and NoSQL: PostgreSQL, Cassandra, MongoDB, etc.).
- Familiarity with cloud-native streaming solutions (AWS Kinesis, Azure Event Hubs, GCP Pub/Sub).
- Knowledge of CI/CD, containerization (Docker, Kubernetes), and monitoring tools (Prometheus, Grafana, ELK).
- Strong problem-solving and debugging skills, with experience in large-scale production environments.
- Language Requirements: Fluent English.
- Nice to Have: Knowledge of data lakehouse architectures (Delta Lake, Iceberg, Hudi).
- Nice to Have: Experience with machine learning pipelines on streaming data.
- Nice to Have: Familiarity with message brokers (RabbitMQ, Pulsar, ActiveMQ).
- Nice to Have: Background in industries like fintech, telecom, IoT, or e-commerce where real-time data is critical.
- Nice to Have: Contributions to open-source streaming projects.
ATS Keywords
Tip: use these terms in your resume and cover letter to boost ATS matches.
Hard skills
data engineeringreal-time streamingApache KafkaApache FlinkSpark StreamingJavaScalaPythonSQLCI/CD
Soft skills
problem-solvingdebuggingcollaborationcommunicationreliabilityresilience